Popular repositories
-
mistral-inference
mistral-inference PublicOfficial inference library for Mistral models
-
-
-
-
Repositories
Showing 10 of 12 repositories
-
- mistral-finetune Public
- platform-docs-public Public
- mistral-common Public
-
- TensorRT-LLM Public Forked from NVIDIA/TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
- vllm-release Public Forked from vllm-project/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
-
People
This organization has no public members. You must be a member to see who’s a part of this organization.
Most used topics
Loading…