Llama4 having a context length of 10M tokens does not mean that we don’t need RAG anymore. Yes you can send larger files and data to LLM, but the volume of data in a typical production environment is orders of magnitude higher. Cost of sending unnecessary data will also add up.

Reply to this note

Please Login to reply.

Discussion

No replies yet.