How does the Nemo 12B compare to the Llama 3.1 8B?
Apparently I am an idiot and read the wrong paper. The previous paper mentioned that “comparable with the 8-bit models”
They said their’s is “comparable with the 8-bit models”. Its all tradeoffs. It isn’t clear to me where you allocate your compute/memory budget. I’ve noticed that full 7b 16 bit models often produce better results for me than some much larger quantied models. It will be interesting to find the sweet spot.
So are more bits less important than more paramters? Would a higher paramter or higher bit count matter more if the models ended up the same size?
I’m so glad I work for a medium-small company. We moved to a smaller office and only require to go in twice a month
deleted by creator
deleted by creator
Thank you! I had no idea this existed.
I am surprised reddit hasn’t removed this post yet.
I got an account banned for saying “lemmy dot world” when someone asked “Are there even any good alternatives?”
He’s opinionated but a pretty good science communicator.
Sony choose to not offer refunds. Sony knew the contract when they agreed to sell the content. When something gets pulled from steam I can still download and install it.
I got banned because I said “lemmy dot world” in reply to someone who said they wished reddit had an alternative.
Its a disturbing tend and more evidence to own physical copies.
I hope Mastodon succeeds but I don’t see it taking off like twitter did.
At work we pay Microsoft thousands for developer licensees, have windows pro, and there is still ads in the OS… Windows exists to service them not us.
Jokes on you. I’m off for 4 days!
Stuff like this will be used in the anti monopoly cases going on world wide.
I love this channel so much
Gamemaker is still focused on 2D. I also think its easier for people without programming experience and the design guides you more but is less flexible.
Idk man I’ve yet to know anyone who died from drinking magma.