Cottus Runtime is a custom inference engine built from scratch for Llama architectures, prioritizing low-latency and strict memory management. It implements its own Transformer execution pipeline, KV cache management, and attention kernels.
- Core: Custom C++20 Transformer implementation.
- Memory: PagedAttention with BlockAllocator for efficient KV cache management.
- Compute: CUDA-accelerated kernels for Attention, RoPE, and GEMM (cuBLAS).
- Parity: Exact token matching with HuggingFace Transformers (verified).
- Interface: Clean Python API via PyBind11.
Cottus Runtime currently supports the following architectures with verified exact parity to HuggingFace Transformers.
| Architecture | Models | Precision | Status |
|---|---|---|---|
| Llama | Llama 2 (7B/13B/70B), Llama 3 (8B/70B) | FP16 / FP32 | IN PRODUCTION |
| TinyLlama | TinyLlama-1.1B | FP16 / FP32 | IN PRODUCTION |
| Mistral | Mistral 7B | FP16 | IN PROGRESS |
Here is what I have planned to add to the project in later iterations. Send a PR if you want to contribute to the project.
- Multi‑GPU & Distributed Execution – Enable scaling across multiple GPUs and clusters for larger models.
- Expanded Model Support – Add native support for Mistral, Falcon, and other non‑LLaMA families.
- Optimized CPU Backend – Introduce a high‑performance CPU path (vectorized kernels, OpenMP) and enable CPU‑only inference.
- Quantization & INT8 – Provide post‑training quantization pipelines and INT8 kernels for reduced memory and faster inference.
- FlashAttention‑style Kernels – Integrate memory‑efficient, block‑sparse attention kernels to cut latency and improve throughput.
- Plugin System – Allow community‑contributed extensions (custom ops, alternative KV‑cache strategies).
- Better Tooling – CLI utilities for model conversion, benchmarking, and profiling.
- NVIDIA GPU with CUDA 11/12 (Recommended)
- C++ Compiler (GCC 10+ or Clang 12+)
- CMake 3.18+
- Python 3.8+
Using pip:
pip install cottusUsing uv (recommended for speed):
uv add cottuspython examples/1_basic_inference.py --device cudaThere is a CPU fallback here. Wouldn't recommebd=
python examples/2_cpu_inference.pyRequires ~2.2GB download.
python examples/3_tinyllama_real.pyThe best way to get started is to look at the examples/ directory, which contains complete scripts for various use cases.
from cottus import Engine, EngineConfig
from cottus.model import load_hf_model
# 1. Load Model Weights
weights, _, _, tokenizer, _ = load_hf_model("TinyLlama/TinyLlama-1.1B-Chat-v1.0", device="cuda")
# 2. Config
config = EngineConfig()
config.model_type = "llama"
config.hidden_dim = 2048
config.num_layers = 22
config.num_heads = 32
config.num_kv_heads = 4
config.head_dim = 64
config.intermediate_dim = 5632
config.device = "cuda"
# 3. Helpers
engine = Engine(config, weights)
input_ids = tokenizer.encode("Hello!")
# 4. Generate
output_ids = engine.generate(input_ids, max_new_tokens=20)
print(tokenizer.decode(output_ids))Cottus Runtime is licensed under the Apache License Version 2.0. By contributing to the project, you agree to the license and copyright terms therein and release your contribution under these terms.