A cutting-edge, full-stack agnostic neural orchestration system that intelligently adapts to ANY model format, ANY runtime environment, and ANY deployment scenario.
Seamlessly load and manage GGUF, ONNX, Safetensors, HuggingFace models, and future formats with zero configuration.
Harness the power of WebGPU for maximum performance, WASM for universal compatibility, and Node.js for server deployments.
Advanced algorithms automatically select the optimal model based on quality metrics, cost constraints, speed requirements, or custom strategies.
Experience lightning-fast token generation with async generators and real-time streaming capabilities for responsive AI applications.
Built-in cost analysis and optimization algorithms ensure maximum efficiency while staying within budget constraints.
Intuitive APIs designed for both simple quick-start scenarios and complex enterprise-grade implementations.
Get started with LLM Runner Router in just a few lines of code
// Simple Mode - For Rapid Prototyping import { quick } from 'llm-runner-router'; const response = await quick("Explain quantum computing to a goldfish"); console.log(response.text); // Advanced Mode - For Production Systems import LLMRouter from 'llm-runner-router'; const router = new LLMRouter({ strategy: 'quality-first', fallbacks: ['gpt-3.5', 'local-llama'], cacheEnabled: true }); // Load multiple models await router.load('huggingface:meta-llama/Llama-2-7b'); await router.load('local:./models/mistral-7b.gguf'); // Intelligent routing in action const response = await router.advanced({ prompt: "Write a haiku about JavaScript", temperature: 0.8, maxTokens: 50, streaming: true }); // Stream tokens in real-time for await (const token of response) { process.stdout.write(token); }
Intelligent model selection algorithms designed for every use case
Everything you need to master LLM Runner Router
Comprehensive guides, API reference, and tutorials for all skill levels
Ready-to-use examples for common use cases and advanced implementations
Get up and running in 5 minutes with our step-by-step installation guide
Platform-specific integration guides for React, Node.js, Docker, and more
Optimization strategies, benchmarking tools, and performance best practices
Join our community, get help, and contribute to the project's development