An image recognizer that recognizes 20 different shapes of Pasta.
-
Updated
May 14, 2024 - Jupyter Notebook
An image recognizer that recognizes 20 different shapes of Pasta.
A high-performance inference system for large language models, designed for production environments.
🤗 Optimum Intel: Accelerate inference with Intel optimization tools
A high-throughput and memory-efficient inference and serving engine for LLMs
High-efficiency floating-point neural network inference operators for mobile, server, and Web
Port of OpenAI's Whisper model in C/C++
Large Language Model Text Generation Inference
A fast, easy-to-use, production-ready inference server for computer vision supporting deployment of many popular model architectures and fine-tuned models.
A universal scalable machine learning model deployment solution
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
OpenVINO™ is an open-source toolkit for optimizing and deploying AI inference
The Triton Inference Server provides an optimized cloud and edge inferencing solution.
Woodwork is a Python library that provides robust methods for managing and communicating data typing information.
Triton Model Analyzer is a CLI tool to help with better understanding of the compute and memory requirements of the Triton Inference Server models.
Utilities to use the Hugging Face Hub API
TypeDB: the polymorphic database powered by types
A large-scale simulation framework for LLM inference
Seamlessly integrate with top LLM APIs for speedy, robust, and scalable querying. Ideal for developers needing quick, reliable AI-powered responses.
ncnn is a high-performance neural network inference framework optimized for the mobile platform
Add a description, image, and links to the inference topic page so that developers can more easily learn about it.
To associate your repository with the inference topic, visit your repo's landing page and select "manage topics."