I did a quick research. ChatGPT still has way, way more parameters.

- ChatGPT-3.5, and GPT-4 are Large Language Models (LLMs) ChatGPT has 175 billion parameters

> Parameters including weights and biases.

>The numbers of parameters in a neural network is directly related to the number of neurons and the number of connections between them.

>Needs supercomputers for training and inference

- LLaMA (by Facebook): Open and Efficient Foundation Language Models

>[llama.cpp](https://github.com/ggerganov/llama.cpp)

> a collection of foundation language models ranging from 7B to 65B parameters

> Facebook says LLaMA-13B outperforms GPT-3 while being more than 10x smaller

> If it is 10x smaller then we don't need a supercomputer

- alpaca.cpp (Stanford Alpaca is an instruction-following LLaMA model)

> It is a fine-tuned version of Facebook's LLaMA 7B model

> It is trained on 52K instructions-following demonstrations.

> uses around 4GB of RAM.

Reply to this note

Please Login to reply.

Discussion

That’s exactly what makes LLaMA the better model. It has way less parameters and still gets GPT-3.5 level outputs.

Parameters aren’t a measure of how good a model is, just how large.