Contents
Tags
We built Runyard because we kept running into the same problem: you find an interesting LLM, download it, and then discover it doesn't fit in your VRAM — or barely fits but runs at 2 tokens per second. Runyard solves that before you waste the time.

Runyard's Model Radar is a hardware-aware model browser. You enter your specs — CPU, RAM, GPU, and VRAM — and it filters the entire catalog of major open-source LLMs to show only what will actually run on your machine. Each model comes with a performance tier, recommended quantization, and an expected tokens-per-second range.

The sunburst is a zoomable radial chart. Every arc is a model — organized by category in the inner rings (Chat, Code, Vision, Reasoning) and individual models in the outer arcs. Colors update live: green means comfortable fit, yellow is tight, red means it won't run on your current VRAM.

Below the sunburst is the Tier List — the same data in a ranked linear format. Every model is scored across three factors: VRAM headroom (40%), memory bandwidth utilization (35%), and benchmark quality (25%). The result is a composite 0–100 score that slots each model into a tier.
Use the Sunburst for exploration — it shows the whole catalog at a glance. Switch to the Tier List when you're ready to decide — it puts your best options at the top. Both update instantly when you change your hardware specs at runyard.dev.
Runyard is free to use. We're building the infrastructure layer for the local AI ecosystem — starting with the tool that helps you figure out where to start. Go to runyard.dev, enter your specs, and find out what's possible on your hardware today.
Tools
Find AI models that fit your exact hardware. Enter your specs and get a ranked list instantly.
Newsletter