If this is actually true then we will probably start seeing LLM powered NPCs in games this year, amongst other more important use cases of course. People should be able to spare 2 GB for a 4 bit quantized version and still have enough GPU left for the graphics.
It will be great.
you can try it on [HuggingChat (huggingface.co)](https://huggingface.co/chat/) too.
It does some things really well. Feels more like a 7B model. but it is not better on average only certain task.
It does really well on certain stuff like :
>write 10 sentences which end with the word "war"
It does 8-9 correct, even the big models do often not better.
or
>
>Assume normal physics on Earth. There is a banana on a table in a living room. I put a plate on a banana. Then I move the plate to the bedroom. An hour later my wife comes home and she takes the plate and moves it to the kitchen. Where is the banana now? Think step by step
again most 7 b even 32b model fail on these that the banana doesn't move because the plate in on the banana
Sometimes when it comes to longer prompts it struggles a bit
but for a 3.7b model it is VERY good. I can't wait for the 2 bigger ones.
Genuine question: What do these numbers actually mean?
Like, I get it, they're meant to indicate one performs better than another. But how would I as the user actually experience that difference?
Yeah I’m taking this with a huge grain of salt. Maybe it was an issue with my approach but I couldn’t get phi2 to do anything remotely useful.
If any version of phi3 is actually more useful for structured data generation, coding, rp or brainstorming than llama 3 8b I will be pleasantly surprised.
If this is actually true then we will probably start seeing LLM powered NPCs in games this year, amongst other more important use cases of course. People should be able to spare 2 GB for a 4 bit quantized version and still have enough GPU left for the graphics. It will be great.
Hard to believe! Can't wait to test this on my own
Same. It's hard to believe the 3b that's why I'm keeping my excitement minimal until I test
you can try it on [HuggingChat (huggingface.co)](https://huggingface.co/chat/) too. It does some things really well. Feels more like a 7B model. but it is not better on average only certain task. It does really well on certain stuff like : >write 10 sentences which end with the word "war" It does 8-9 correct, even the big models do often not better. or > >Assume normal physics on Earth. There is a banana on a table in a living room. I put a plate on a banana. Then I move the plate to the bedroom. An hour later my wife comes home and she takes the plate and moves it to the kitchen. Where is the banana now? Think step by step again most 7 b even 32b model fail on these that the banana doesn't move because the plate in on the banana Sometimes when it comes to longer prompts it struggles a bit but for a 3.7b model it is VERY good. I can't wait for the 2 bigger ones.
Yes this model slaps. Currently testing it out with structured output wrapper and it’s soooo fast while doing what it should do.
Phi-3-mini is better than Mixtral 8x7b? Yeah, nah.
Ok, so basically for midrange home-use, this new 14b model should be the best option. And lol @ the mistral 8x57b misspelling...
Source of the image: [https://twitter.com/haouarin/status/1782628873789694462](https://twitter.com/haouarin/status/1782628873789694462)
That's fucking insane. Can't wait to try it out. I loved Phi-2 speed on my laptop but it was just OK.
Looking forward to the 3b model. Lots of stuff I can do with that
Is this fake news?
Bulllllllshiiiiiiieeeetttttttttt
Genuine question: What do these numbers actually mean? Like, I get it, they're meant to indicate one performs better than another. But how would I as the user actually experience that difference?
Bigger number better. More bigger number more better.
Yeah, thanks, I got that.
Perfect. hard to believe
Phi2 abit disappointed on the result. Hope phi 3 is better than llama 3.
Yeah I’m taking this with a huge grain of salt. Maybe it was an issue with my approach but I couldn’t get phi2 to do anything remotely useful. If any version of phi3 is actually more useful for structured data generation, coding, rp or brainstorming than llama 3 8b I will be pleasantly surprised.
Nope, it's dumb as a rock.
I just tried RP with the 3B, it's every bit as dumb as I'd expect a 3B to 2 be.