Compile classical ML models to native C. Serve them in microseconds.
Documentation · Changelog · PyPI · Technical Paper · Agent Skill
Timber takes a trained ML model — XGBoost, LightGBM, scikit-learn, CatBoost, ONNX (tree ensembles, linear models, SVMs, k-NN, Naive Bayes, GPR, Isolation Forest), or a URDF robot description — runs it through a multi-pass optimizing compiler, and emits a self-contained C99 inference artifact with zero runtime dependencies. A built-in HTTP server (Ollama-compatible API) lets you serve any model in one command.
For hardware-accelerated or safety-critical deployments, the bundled timber accel backend emits AVX2/AVX-512/NEON/SVE/RVV SIMD, CUDA/Metal/OpenCL GPU, Xilinx/Intel FPGA HLS, and Cortex-M/ESP32/STM32 embedded variants — plus WCET analysis, DO-178C/ISO-26262/IEC-62304 certification reports, Ed25519 artifact signing, AES-256-GCM encryption, air-gapped deployment bundles, and ROS 2 / PX4 / gRPC server generators. Everything ships in one pip install.
~2 µs single-sample inference · ~336× faster than Python XGBoost · ~48 KB artifact · zero runtime dependencies
Point Timber at any URL and it downloads, compiles, and serves in one command — no setup required.
$ pip install timber-compiler
$ timber serve https://raw.githubusercontent.com/kossisoroyce/timber/main/examples/breast_cancer_model.json
✓ Downloaded breast_cancer_model.json
✓ Format detected xgboost
✓ Parsed model 50 trees · 30 features · binary:logistic
✓ Optimized 3/5 passes applied
✓ Generated C99 169 lines
✓ Compiled binary 47.9 KB
Serving breast_cancer_model
Endpoint http://localhost:11434
Framework xgboost · 50 trees · 30 features
POST http://localhost:11434/api/predict
GET http://localhost:11434/api/models
GET http://localhost:11434/api/healthPredict immediately:
$ curl -s http://localhost:11434/api/predict \
-H 'Content-Type: application/json' \
-d '{"model": "breast_cancer_model", "inputs": [[1.799e+01, 1.038e+01, 1.228e+02, 1.001e+03, 0.1184, 0.2776, 0.3001, 0.1471, 0.2419, 0.07871, 1.095, 0.9053, 8.589, 153.4, 0.006399, 0.04904, 0.05373, 0.01587, 0.03003, 0.006193, 2.538e+01, 1.733e+01, 1.846e+02, 2.019e+03, 0.1622, 0.6656, 0.7119, 0.2654, 0.4601, 0.1189]]}'
{"model": "breast_cancer_model", "outputs": [[0.9971]], "n_samples": 1}Or load from a local file and serve by name:
$ timber load fraud_model.json --name fraud-detector
$ timber serve fraud-detector- Who is this for?
- How it works
- Quick Start
- Supported Formats
- Hardware Acceleration & Safety
- Performance
- Runtime Comparison
- API Reference
- CLI Reference
- Examples
- Limitations
- Roadmap
- Development
- Citation
- Community & Governance
- License
Timber is built for teams that need fast, predictable, and portable inference:
- Fraud & risk teams — run classical models in sub-millisecond transaction paths without Python overhead
- Edge & IoT deployments — ship a ~48 KB C artifact to gateways, microcontrollers, or ARM Cortex-M targets
- Regulated industries — finance, healthcare, and automotive teams that need deterministic, auditable inference artifacts
- Platform & infra teams — eliminate the Python model-serving stack from your critical path entirely
┌─────────────────────────────────────────────────────────┐
│ timber load │
│ │
│ Model file ──► Parser ──► Timber IR ──► Optimizer│
│ (.json/.pkl/ (typed AST) (dead-leaf │
│ .txt/.onnx) elim, quant, │
│ branch-sort) │
│ │ │
│ ▼ │
│ C99 Emitter │
│ │ │
│ ┌────────────────┼────────────────┐ │
│ ▼ ▼ ▼ │
│ model.c model.h model_data.c│
│ (inference) (public API) (tree data) │
│ │ │
│ └──► gcc / clang ──► model.so │
└─────────────────────────────────────────────────────────┘
│
▼
timber serve <name>
http://localhost:11434/api/predict
The compiler pipeline:
- Parse — reads the native model format into a framework-agnostic Timber IR
- Optimize — dead-leaf elimination, threshold quantization, constant-feature folding, branch sorting
- Emit — generates deterministic, portable C99 with no dynamic allocation and no recursion
- Compile —
gcc/clangproduces a shared library loaded viactypes - Serve — an Ollama-compatible HTTP API wraps the binary for drop-in integration
pip install timber-compilerServe any model directly from a URL — no pre-download step:
timber serve https://yourhost.com/models/fraud_model.jsonOr load a local model and serve by name:
timber load fraud_model.json --name fraud-detector
timber serve fraud-detectorPredict:
curl -s http://localhost:11434/api/predict \
-H "Content-Type: application/json" \
-d '{"model": "fraud-detector", "inputs": [[1.2, 0.4, 3.1, 0.9]]}'{"model": "fraud-detector", "outputs": [[0.031]], "latency_us": 1.8}That's it. No model server configuration, no Python runtime in the hot path.
Or serve a robot's forward kinematics from a URDF file:
timber serve robot.urdfcurl -s http://localhost:11434/api/predict \
-H "Content-Type: application/json" \
-d '{"model": "robot", "inputs": [[0.1, 0.2, 0.3, -0.4, 0.5, -0.6, 0.7]]}'{"model": "robot", "outputs": [[0.082, -0.982, 0.170, 0.310, 0.959, 0.031, -0.283, 0.053, 0.272, 0.187, 0.944, 1.180, 0.0, 0.0, 0.0, 1.0]], "latency_us": 65.0}| Framework | File format | Notes |
|---|---|---|
| XGBoost | .json |
All objectives; multiclass, binary, regression; XGBoost 3.1+ per-class base_score |
| LightGBM | .txt, .model, .lgb |
All objectives including multiclass |
| scikit-learn | .pkl, .pickle |
GradientBoosting, RandomForest, ExtraTrees, DecisionTree, Linear/Logistic, SVM, OneClassSVM, IsolationForest, GaussianNB, KNeighborsClassifier/Regressor, GaussianProcessRegressor, Pipeline |
| ONNX | .onnx |
TreeEnsembleClassifier/Regressor, LinearClassifier/Regressor, SVMClassifier/Regressor, Normalizer, Scaler |
| CatBoost | .json |
JSON export (save_model(..., format='json')) |
| URDF | .urdf |
Robot description → forward kinematics; outputs 4×4 homogeneous transform; inputs are joint angles |
Every pip install timber-compiler also installs the timber-accel CLI and the timber.accel Python package — no separate install, no paid tier.
| Category | Backends |
|---|---|
| SIMD | AVX2, AVX-512, ARM NEON, ARM SVE, RISC-V V (RVV) |
| GPU | CUDA (SM 7.5 / 8.6), Apple Metal (M1+), OpenCL |
| FPGA / HLS | Xilinx Vitis HLS, Intel FPGA SDK (OpenCL) |
| Embedded | ARM Cortex-M4 / M7, ESP32, STM32 (no-heap, static buffers) |
18 target profiles ship as TOML files under timber/accel/targets/ — load one by name or point at a custom .toml.
- WCET analysis (
timber-accel wcet) — worst/average-case cycle counts per IR stage for Cortex-M4/M7, x86_64, AArch64, RISC-V64 with configurable safety margin; advisory output (not a replacement foraiT/RapiTime). - Deterministic & constant-time IR passes — eliminates data-dependent branches in generated C.
- Certification reports (
timber-accel certify) — structured JSON reports for DO-178C (levels A–D), ISO 26262 (ASIL A–D), IEC 62304 (classes A/B/C); heuristic static analysis plus optional embedded WCET block.
- Ed25519 signing & verification (
timber-accel sign/verify) — per-artifact detached signatures; keypair generation built in. - AES-256-GCM encryption (
timber-accel encrypt/decrypt) — symmetric protection for models in transit / on-disk. - TPM hooks — Linux TPM 2.0 and software emulator paths.
- Air-gapped deployment bundles (
timber-accel bundle) — tar.gz with model artifact, optional source, optional cert report, manifest for offline deployment.
- C++ gRPC inference server (
timber-accel serve-native) — wraps the compiled model in a gRPC (or plain HTTP) service you can drop into a production C++ stack. - ROS 2 node package — launch file +
rclpynode that calls the compiled model on a ROS topic. - PX4 autopilot module — PX4-style uORB module skeleton for flight-controller deployment.
- Sensor preprocessing — radar, RF, and telemetry front-ends (C callable from compiled code).
# Compile with AVX2 SIMD and sign the output
timber-accel compile --model fraud.pkl --target x86_64_avx2_simd --sign --out ./dist
# WCET for a Cortex-M4 flight controller at 168 MHz
timber-accel wcet --model anomaly.pkl --arch cortex-m4 --clock-mhz 168
# DO-178C Level B certification report with embedded WCET
timber-accel certify --model model.pkl --profile do_178c --include-wcet -o cert.json
# Air-gapped deployment bundle
timber-accel bundle --model model.pkl --include-cert -o deploy.tar.gzFull CLI reference: timber-accel --help · Docs: docs/accel.md
Advisory notice — WCET, MISRA-C, and certification features are heuristic / regex-based and explicitly documented as advisory only. For safety-critical certification, use certified tooling (LDRA, Polyspace, Astrée, aiT, RapiTime) and verify independently with a qualified DER.
Benchmarks run on Apple M2 Pro · 16 GB RAM · macOS · XGBoost binary classifier · 50 trees · max depth 4 · 30 features (sklearn
breast_cancer) · 10,000 timed iterations after 1,000 warmup.
| Runtime | Single-sample latency | Throughput | Speedup vs Python |
|---|---|---|---|
| Timber (native C) | ~2 µs | ~500,000 / sec | 336× |
| ONNX Runtime | ~80–150 µs | ~10,000 / sec | ~5× |
| Treelite (compiled) | ~10–30 µs | ~50,000 / sec | ~20× |
| Python XGBoost | ~670 µs | ~1,500 / sec | 1× (baseline) |
| Python scikit-learn | ~900 µs | ~1,100 / sec | 0.7× |
Latency is in-process (not HTTP round-trip). Network overhead adds ~50–200 µs depending on your stack.
python benchmarks/run_benchmarks.py --output benchmarks/results.json
python benchmarks/render_table.py --input benchmarks/results.jsonSee benchmarks/ for full methodology, hardware capture script, and optional ONNX Runtime / Treelite / lleaves comparisons.
| Timber | Python serving | ONNX Runtime | Treelite | lleaves | |
|---|---|---|---|---|---|
| Latency | ~2 µs | 100s of µs–ms | ~100 µs | ~10–30 µs | ~50 µs |
| Runtime deps | None | Python + framework | ONNX Runtime libs | Treelite runtime | Python + LightGBM |
| Artifact size | ~48 KB | 50–200+ MB process | MBs | MB-scale | Python env |
| Formats | 5 (trees + linear + SVM) | Each framework only | ONNX only | GBDTs | LightGBM only |
| C export | Yes (C99) | No | No | Yes | No |
| LLVM IR export | Yes | No | No | No | No |
| Edge / embedded | Yes (Cortex-M4/M33, RISC-V) | No | Partial | Partial | No |
| MISRA-C output | Yes | No | No | No | No |
| Differential privacy | Yes | No | No | No | No |
Timber's server exposes an Ollama-compatible REST API on http://localhost:11434 by default.
| Endpoint | Method | Body / Params | Description |
|---|---|---|---|
/api/predict |
POST | {"model": str, "inputs": [[float]]} |
Run inference |
/api/generate |
POST | same as /api/predict |
Ollama alias |
/api/models |
GET | — | List all loaded models |
/api/model/:name |
GET | — | Model metadata & schema |
/api/health |
GET | — | Health check |
Example — batch inference:
curl -s http://localhost:11434/api/predict \
-H "Content-Type: application/json" \
-d '{
"model": "fraud-detector",
"inputs": [
[1.2, 0.4, 3.1, 0.9],
[0.1, 2.3, 1.0, 4.4]
]
}'timber load <path> --name <name> Compile and register a model
timber serve <name> [--port N] Start the inference server
timber list List registered models
timber inspect <name> Show model IR summary and schema
timber validate <name> Run numerical validation against source
timber bench <name> Benchmark latency and throughput
timber pull <url> --name <name> Download and compile from URL
timber remove <name> Remove a model from the registry
timber-accel compile Compile with SIMD/GPU/HLS/embedded target + optional safety transforms
timber-accel wcet Worst-case execution time analysis (Cortex-M4/M7, x86_64, AArch64, RISC-V64)
timber-accel certify Generate DO-178C / ISO 26262 / IEC 62304 certification report
timber-accel sign Ed25519-sign an artifact (generates keypair on demand)
timber-accel verify Verify an Ed25519 signature
timber-accel encrypt AES-256-GCM encrypt a model artifact
timber-accel decrypt AES-256-GCM decrypt a model artifact
timber-accel bundle Create an air-gapped deployment tar.gz
timber-accel serve-native Generate a C++ gRPC / HTTP inference server
Runnable end-to-end examples live in examples/:
python examples/quickstart_xgboost.py # trains, compiles, and benchmarks
python examples/quickstart_lightgbm.py
python examples/quickstart_sklearn.pyEach script trains a model, saves it, runs timber load, and validates predictions against the source framework.
- ONNX — supports
TreeEnsemble,LinearClassifier/Regressor,SVMClassifier/Regressor,Normalizer,Scaler; other operators (e.g., neural network layers) are not yet supported - CatBoost — requires JSON export (
save_model(..., format='json')); native binary format not supported - scikit-learn — major estimators and
Pipelinewrappers are supported; uncommon custom estimators may require a custom front-end - Pickle — follow standard pickle security hygiene; only load artifacts from trusted sources
- XGBoost — JSON model format is the primary path; binary booster format is not supported
- LLVM IR — currently emitted as text (
.ll); requires a local LLVM/Clang installation to produce native code from it - MISRA-C — the built-in compliance checker covers the rules most relevant to generated code; it is not a substitute for a certified static analysis tool
| Status | Item |
|---|---|
| ✅ | XGBoost, LightGBM, scikit-learn, CatBoost, ONNX front-ends |
| ✅ | Multi-pass IR optimizer (dead-leaf, quantization, branch sort, scaler fusion) |
| ✅ | C99 emitter with WebAssembly target |
| ✅ | Ollama-compatible HTTP inference server with multi-worker FastAPI |
| ✅ | PyPI packaging with OIDC trusted publishing |
| ✅ | ONNX Linear/SVM/Normalizer/Scaler operator support |
| ✅ | ARM Cortex-M4/M33 and RISC-V rv32imf/rv64gc embedded deployment profiles |
| ✅ | MISRA-C:2012 compliant output mode with built-in compliance checker |
| ✅ | LLVM IR backend with configurable target triples |
| ✅ | Differential privacy (Laplace + Gaussian) inference mode |
| ✅ | Richer bench reports: P50/P95/P99/P999, CV%, JSON + HTML output |
| ✅ | URDF forward-kinematics frontend — robot FK as a compiled C function |
| ✅ | 5 new sklearn primitives: IsolationForest, OneClassSVM, GaussianNB, GPR, k-NN |
| ✅ | SIMD codegen (AVX2, AVX-512, NEON, SVE, RVV) via timber.accel |
| ✅ | GPU codegen (CUDA, Metal, OpenCL) via timber.accel |
| ✅ | FPGA HLS codegen (Xilinx Vitis, Intel FPGA SDK) via timber.accel |
| ✅ | WCET analysis and DO-178C / ISO 26262 / IEC 62304 certification reports |
| ✅ | Ed25519 signing, AES-256-GCM encryption, air-gapped deployment bundles |
| ✅ | ROS 2 / PX4 / gRPC server code generators |
| 🔄 | Remote model registry (timber pull from hosted model library) |
| 🔲 | Neural network operator support (MLPClassifier) |
| 🔲 | ONNX export path (Timber IR → ONNX) |
| 🔲 | Rust backend emitter |
git clone https://github.com/kossisoroyce/timber.git
cd timber
pip install -e ".[dev]"
pytest tests/ -v # 650+ tests (core + accel)
ruff check timber/ # lintingThe test suite covers: parsers (sklearn, ONNX, XGBoost, LightGBM, CatBoost, URDF), IR layer (serialization, deep_copy, all stage types including the 5 new primitives), optimizer passes (correctness, idempotency, pipeline fusion math), C99/WASM/MISRA-C/LLVM IR emitters (compile + numeric accuracy), differential privacy (statistical correctness, all dtypes), timber.accel (WCET analysis, SIMD codegen, Ed25519 signing, certification reports, SIMD/GPU/HLS target profiles), and full end-to-end pipelines.
See CONTRIBUTING.md for the full development guide.
If you use Timber in research or production, please cite the accompanying technical paper:
@misc{royce2026timber,
title = {Timber: Compiling Classical Machine Learning Models to Native Inference Binaries},
author = {Kossiso Royce},
year = {2026},
howpublished = {GitHub repository and technical paper},
institution = {Electricsheep Africa},
url = {https://github.com/kossisoroyce/timber}
}The full paper is available at paper/timber_paper.pdf.
- Contributing:
CONTRIBUTING.md - Code of conduct:
CODE_OF_CONDUCT.md - Security policy:
SECURITY.md - Changelog:
CHANGELOG.md
Bugs and feature requests: open an issue. Questions: start a discussion.
Timber is developed and maintained by Electricsheep Africa. If Timber saves your team engineering time, consider supporting continued development:
We're also building a hosted library of compiled models — if you have high-performance models you'd like to contribute, get in touch.
Apache-2.0 — see LICENSE for the full text.