Replying to Avatar StellarStoic

I did a quick research. ChatGPT still has way, way more parameters.

- ChatGPT-3.5, and GPT-4 are Large Language Models (LLMs) ChatGPT has 175 billion parameters

> Parameters including weights and biases.

>The numbers of parameters in a neural network is directly related to the number of neurons and the number of connections between them.

>Needs supercomputers for training and inference

- LLaMA (by Facebook): Open and Efficient Foundation Language Models

>[llama.cpp](https://github.com/ggerganov/llama.cpp)

> a collection of foundation language models ranging from 7B to 65B parameters

> Facebook says LLaMA-13B outperforms GPT-3 while being more than 10x smaller

> If it is 10x smaller then we don't need a supercomputer

- alpaca.cpp (Stanford Alpaca is an instruction-following LLaMA model)

> It is a fine-tuned version of Facebook's LLaMA 7B model

> It is trained on 52K instructions-following demonstrations.

> uses around 4GB of RAM.

That’s exactly what makes LLaMA the better model. It has way less parameters and still gets GPT-3.5 level outputs.

Parameters aren’t a measure of how good a model is, just how large.

Reply to this note

Please Login to reply.

Discussion

No replies yet.