A detailed walkthrough of building a budget-friendly AI workstation with 48GB VRAM for running local LLMs, costing around 1700 euros using second-hand Tesla P40 GPUs. The setup enables running various AI models locally, achieving 5-15 tokens per second depending on model size, while maintaining independence from cloud-based AI services.
https://ewintr.nl/posts/2025/building-a-personal-private-ai-computer-on-a-budget/
#aihardware #diycomputing #llms #costoptimization #technicalguide