Tiny Weights is a blog focused on small language models — the Gemmas, SmolLMs, Phis, and Mistrals of the world.

The goal is simple: cut through the hype and focus on what actually runs well on modest hardware. Every post is written with the assumption that you have a laptop, a cheap VPS, or a Raspberry Pi — not a data centre.

Topics covered:

  • Model releases and benchmarks
  • Running models locally with Ollama, llama.cpp, and LlamaEdge
  • Quantization and efficiency
  • Fine-tuning at small scale
  • Edge and on-device inference