llm-serving

vllm

A high-throughput and memory-efficient inference and serving engine for LLMs

ray

Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.