Playing with Llama 2 this morning - several interesting results. Overall conclusion: probably the best model to run locally available now. Haven't fully tested but my hypothesis is that a 4-bit quantised of the 13B version is going to be the sweet spot for local inference for now. A few interesting results below...
- 1 reply
- 0 recasts
- 0 reactions
Can I post to Purple if I don’t have the NFT?
- 0 replies
- 0 recasts
- 0 reactions
Interesting to finally get a glimpse into the inner workings of GPT-4. TL;DR: GPT-4 is made up of 16 'expert' models, each of which are ~110B parameters and make for ~1.8 trillion total parameters (more than 10x the 175B parameters of GPT-3.5). https://www.semianalysis.com/p/gpt-4-architecture-infrastructure
- 1 reply
- 0 recasts
- 0 reactions
