Roadmap#
FuriosaAI regularly publishes its software with new features, performance improvements, and expanded hardware support. This page shows the forward-looking roadmap of ongoing & upcoming projects and when they are expected to land, broken down by areas on our software stack.
Note
The latest release is 2026.1.0. You can find the release notes here.
Upcoming Releases 2026 Q1#
π¨ Qwen3 MoE, GPT-OSS, K-EXAONE model support
π¨ Qwen3 VL and multi-modal model support
π¨ KV cache offloading support
π¨ OpenResponse support
π¨ Speculative decoding support
π¨ PyTorch eager mode support
2025 Q3 - Q4#
Furiosa-LLM#
β Hybrid batching support (i.e., chunked prefill or inflight-batching)
β Exaone4, Qwen3 support
β Guided-decoding support (libguidance, xgrammar backends)
β Tool-calling support
β Prefix-caching support
β Pooling Model support (embedding, score, and rank)
β Fine-tuned model support
β Tensor Parallelism support Phase 2: Inter-chip
β Hugging Face Hub support
β Pre-compiled artifacts on Hugging Face Hub
β Qwen2 and Qwen2.5 model support
β EXAONE3 model support
β API Key based authentication support
β Harmony response format support
Quantization#
β Fine-grained FP8 Quantization (dynamic quantization, mixed quantization)
Distributed & Scalable Inference#
β llm-d integration
β NPU operator support for Kubernetes
β DRA (Dynamic Resource Allocation) support for Kubernetes
2025 Q1 - Q2#
β Tool-calling support in Furiosa-LLM (2025.1.0 release)
β Device remapping support (e.g., /dev/rngd/npu2pe0-3 -> /dev/rngd/npu0pe0-3) for container (2025.1.0 release)
β Automatic configuration for the maximum KV-cache memory allocation (2025.1.0 release)
β Min-p sampling support (2025.1.0 release)
β Chunked Prefill support in Furiosa-LLM (planned for 2025.2.0 release)
β Chat API support in Furiosa-LLM (planned for 2025.2.0 release)
β Reasoning parser support (2025.2.0 release)
β Torch 2.5.1 support (2025.2.0 release)
β Python 3.11 and 3.12 support (2025.2.0 release)
β Support for building bfloat16, float16, and float32 models to model artifact without quantization (2025.2.0 release)
β Metrics endpoint (
/metrics/) support in Furiosa-LLM (2025.2.0 release)β Model artifact support in Huggingface Hub (2025.2.0 release)
β Sampling parameter βlogprobsβ support (2025.2.0 release)
β Container Runtime and Container Interface Device (CDI) support (2025.2.0 release)
2024 Q4#
β Language Model Support: CodeLLaMA2, Vicuna, Solar, EXAONE-3.0 (2024.2.0 release)
β Vision Model Support: MobileNetV1, MobileNetV2, ResNet152, ResNet50, EfficientNet, YOLOv8m, etc (2024.2.0 release)
β Tensor Parallelism support Phase 1: Intra-chip (2024.2.0 release)
β Torch 2.4.1 support (2024.2.0)
β Huggingface Optimum integration (2024.2.0 release)