Resources
Explore AI & ML resources curated by the community.
Explore AI & ML resources curated by the community.
A lightweight, high-performance C/C++ implementation for running LLaMA‑style large language models locally, supporting CPU and GPU inference with minimal dependencies.
Open Resource