projects
Raijin
Raijin is a CPU inference engine built in Rust using the ONNX runtime. Named after the Japanese god of thunder and lightning, it aims to deliver blazing-fast local model execution.
Vision
- ONNX runtime: leverage Microsoft’s ONNX Runtime for optimized CPU inference
- Pure Rust: safe, fast, memory-efficient
- Local-first: all inference happens on your machine
- Cross-platform: runs anywhere Rust and ONNX runtime compile
Why ONNX?
While llama.cpp excels at GGUF-based transformer inference, ONNX runtime provides a complementary path for models exported in the ONNX format. This is particularly useful for smaller, task-specific models (classification, embedding, summarization) that don’t need the full generative pipeline.
Status
Raijin is an incomplete, early-stage project. The repository exists with initial structure but the implementation is not yet fully realized. It is on the roadmap for future development as part of the broader gods ecosystem.