Distributed KV-cache store & transfer engine that decouples prefilling from decoding to scale vLLM serving clusters.
Volcano Engine Reinforcement Learning library for efficient LLM post-training—open-sourced HybridFlow.
RAGFlow is InfiniFlow’s open-source Retrieval-Augmented Generation engine focused on deep-document understanding and scalable multi-format ingestion.
vLLM-project’s control-plane that orchestrates cost-efficient, plug-and-play LLM inference infrastructure.
NVIDIA Dynamo is an open-source, high-throughput, low-latency inference framework that scales generative-AI and reasoning models across large, multi-node GPU clusters.