GroveAI
Free Tool

AI Hardware Calculator

Find out exactly what hardware you need to run AI models locally. Adjust the inputs below to match your use case.

Your Configuration

Model Size

Quantisation Level

Lower quantisation = less VRAM, slightly lower quality

2K128K
110

Priority

Recommended Build

Required VRAM

5.0 GB

Recommended GPU

RTX 4060 (8 GB)

System RAM

32 GB

CPU

AMD Ryzen 5 / Intel i5

Storage

1 TB NVMe SSD

PSU

550W

Estimated Build Cost

£500 – £800

Build Tier

Entry

What You Can Run

  • 7B chatbot (Llama 3.2, Mistral 7B)
  • Code completion assistant
  • Document summariser
  • Local embeddings for RAG

Reference

VRAM Requirements by Model & Quantisation

ModelFP16Q8Q5Q4Q3
7B14.0 GB7.0 GB4.4 GB3.5 GB2.6 GB
13B26.0 GB13.0 GB8.1 GB6.5 GB4.9 GB
30B60.0 GB30.0 GB18.8 GB15.0 GB11.3 GB
70B140.0 GB70.0 GB43.8 GB35.0 GB26.3 GB
120B+240.0 GB120.0 GB75.0 GB60.0 GB45.0 GB

Base VRAM only. Add KV cache overhead for long contexts and multiple concurrent users.

Software Stack

Recommended Software

Ollama

Easiest way to download and run models. One-command setup with built-in API.

llama.cpp

Maximum performance inference engine. Powers most local AI tools under the hood.

vLLM

Production serving with batching, OpenAI-compatible API. Best for multi-user setups.

Open WebUI

ChatGPT-like interface for local models. Works with Ollama out of the box.

LM Studio

GUI model manager for Mac and Windows. Great for non-technical users.

text-generation-webui

Feature-rich web interface with model loading, fine-tuning, and extensions.

Need Help With Your Local AI Setup?

We help UK businesses deploy AI on their own infrastructure — from hardware specification to production deployment.