Now in beta

Hybrid Cloud-Edge ML Inference

Run ML models on-device or in the cloud with intelligent routing based on device capabilities. Built for Flutter, Rust, and native platforms.

Everything you need for edge ML

From speech recognition to text-to-speech, run powerful ML models anywhere.

🚀

Edge-First

Run ASR and TTS models directly on-device for low latency and offline support.

☁️

Cloud Fallback

Seamlessly route to cloud APIs when device capabilities are insufficient.

🔗

Pipeline DSL

Chain models together with YAML pipelines: ASR → LLM → TTS in one config.

📱

Cross-Platform

Flutter, iOS, Android, and Rust SDKs with unified API.

NPU Acceleration

Leverage CoreML, QNN, and Metal for maximum performance.

📊

Telemetry

Track inference metrics and device capabilities across your fleet.

Simple, powerful API

Run inference in just a few lines of code.

main.dart
final xybrid = Xybrid.simple(apiKey: 'sk_...');

// Load a pipeline
final pipeline = await xybrid.loadPipeline('voice-assistant.yaml');

// Run inference
final result = await pipeline.run(
  Envelope.audio(audioBytes)
);

// Get the response audio
final responseAudio = result.audio;

Ready to get started?

Join developers building the next generation of voice-enabled applications.