ilumiere - Overview
Navigation Menu
Pinned Loading
-
Forked from InternLM/lmdeploy
LMDeploy is a toolkit for compressing, deploying, and serving LLMs.
Python
-
Forked from ollama/ollama
Get up and running with Llama 3.1, Mistral, Gemma 2, and other large language models.
Go
-
Forked from sgl-project/sglang
SGLang is yet another fast serving framework for large language models and vision language models.
Python
-
Forked from vllm-project/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
Python