TinyLlama
1.1B-parameter Llama-architecture model trained on ~3T tokens—Apache-2.0 weights for fast experiments and teaching.
Why it is included
Tiny open checkpoint for CI, education, and edge prototypes without huge VRAM.
Best for
Students and engineers testing pipelines before scaling parameters.
Strengths
- Small
- Permissive
- Llama-compatible tooling
Limitations
- Capability ceiling vs 7B+ models
Good alternatives
SmolLM · Phi · Gemma
Related tools
AI & Machine Learning
Microsoft Phi
Small language model family (Phi-3/4 lineage) emphasizing strong quality per parameter; weights on Hugging Face under Microsoft licenses per release.
AI & Machine Learning
llama.cpp
Plain C/C++ inference for LLaMA-class models with broad community backends.
AI & Machine Learning
SmolLM
Hugging Face TB small LM family (135M–1.7B) with Apache-2.0 weights aimed at on-device and edge quality per size.
AI & Machine Learning
MLC LLM
Universal deployment stack compiling models to Vulkan, Metal, CUDA, and WebGPU via TVM/Unity for phones, browsers, and servers.
AI & Machine Learning
Google Gemma
Google’s smaller open **weights** Gemma line (Gemma 2/3, etc.) with Gemma license terms, plus `gemma.cpp` for lightweight CPU inference.
AI & Machine Learning
Falcon
Technology Innovation Institute Falcon open weights (7B–180B era) under Apache-2.0 weights for many releases—landmark UAE-led open model line.
