SDK for TEE AI Stick (includes model training script, inference library, examples)
-
Updated
Feb 15, 2019 - Python
SDK for TEE AI Stick (includes model training script, inference library, examples)
Describing How to Enable OpenVINO Execution Provider for ONNX Runtime
Latte is a convolutional neural network (CNN) inference engine written in C++ and uses AVX to vectorize operations. The engine runs on Windows 10, Linux and macOS Sierra.
Rust library managing long conversations with any LLM
Unified JavaScript API for scoring via various DL framework
Experimental Python implementation of OpenVINO Inference Engine (very slow, limited functionality). All codes are written in Python. Easy to read and modify.
NodeJS binding for Menoh DNN inference library
Local Generative AI, unleashed. It signals a fundamental shift: no more cloud lock‑in or opaque black‑box services—now you run powerful generative AI models entirely on your own machine, giving you ultimate control over latency, privacy, and customization.
Add a description, image, and links to the inference-library topic page so that developers can more easily learn about it.
To associate your repository with the inference-library topic, visit your repo's landing page and select "manage topics."