First Inference

Run a language model in 3 commands:

# 1. Install
cargo install aprender

# 2. Download a model from HuggingFace
apr pull qwen2.5-coder-1.5b

# 3. Run inference
apr run qwen2.5-coder-1.5b "Explain quicksort in Rust"

From a Local File

# GGUF format (most common)
apr run model.gguf "What is 2+2?"

# SafeTensors format
apr run model.safetensors --prompt "Hello, world"

# APR format (native)
apr run model.apr "Translate to Spanish: Good morning"

Options

# Control generation
apr run model.gguf "prompt" --max-tokens 100 --temperature 0.7 --top-p 0.9

# JSON output
apr run model.gguf "prompt" --json

# Verbose (show timing, tokens/sec)
apr run model.gguf "prompt" --verbose

# GPU acceleration
apr run model.gguf "prompt" --gpu

Interactive Chat

apr chat model.gguf
# > What is Rust?
# Rust is a systems programming language...
# > How does ownership work?
# ...