Cross-platform, customizable ML solutions for live and streaming media.
-
Updated
Jan 9, 2023 - C++
Cross-platform, customizable ML solutions for live and streaming media.
ncnn is a high-performance neural network inference framework optimized for the mobile platform
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
Colossal-AI: A Unified Deep Learning System for Big Model Era
Example
NVIDIA® TensorRT™, an SDK for high-performance deep learning inference, includes a deep learning inference optimizer and runtime that delivers low latency and high throughput for inference applications.
Hello AI World guide to deploying deep-learning inference networks and deep vision primitives with TensorRT and NVIDIA Jetson.
Runtime type system for IO decoding/encoding
Port of OpenAI's Whisper model in C/C++
The Triton Inference Server provides an optimized cloud and edge inferencing solution.
OpenVINO™ Toolkit repository
TNN: developed by Tencent Youtu Lab and Guangying Lab, a uniform deep learning inference framework for mobile、desktop and server. TNN is distinguished by several outstanding features, including its cross-platform capability, high performance, model compression and code pruning. Based on ncnn and Rapidnet, TNN further strengthens the support and …
An easy to use PyTorch to TensorRT converter
Pre-trained Deep Learning models and demos (high quality and extremely fast)
Adversarial Robustness Toolbox (ART) - Python Library for Machine Learning Security - Evasion, Poisoning, Extraction, Inference - Red and Blue Teams
TypeDB: a strongly-typed database
LightSeq: A High Performance Library for Sequence Processing and Generation
TensorFlow template application for deep learning
Add a description, image, and links to the inference topic page so that developers can more easily learn about it.
To associate your repository with the inference topic, visit your repo's landing page and select "manage topics."