A precision-tuned 1 billion-parameter Llama 3.2 model optimized for computer-engineering tasks.

Key Specifications:


Model Size: 1 B parameters

Frameworks: PyTorch & Hugging Face Transformers

Deployment: On-device inference & cloud

Quantization: 8-bit

Why It Matters

Trained on textbooks, datasheets, and Logisim projects

Lightweight & Fast: Ideal for embedding in development tools or running on edge devices

For non-commercial use under Meta’s license (see GitHub)

Quick start example

Get Started

🔗 GitHub Repository

For more details on how to use it and for downloading it you can find the model on GitHub which you can find below