Substrate Python SDK
-
Updated
Jun 1, 2024 - Python
Substrate Python SDK
Search, Knowledge, Uncertainty, Optimization, Learning, Neural Networks and Language.
A high-throughput and memory-efficient inference and serving engine for LLMs
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
High-efficiency floating-point neural network inference operators for mobile, server, and Web
Cross-platform, customizable ML solutions for live and streaming media.
OpenVINO™ is an open-source toolkit for optimizing and deploying AI inference
Template designed to kickstart your machine learning training projects in Python
Large Language Model Text Generation Inference
PyTorch/XLA integration with JetStream (https://github.com/google/JetStream) for LLM inference"
Substrate TypeScript SDK
TensorRT C++ API Tutorial
Study materials for taking the Harvard Biostatistics PhD Qualifying Exam, Summer 2024
A universal scalable machine learning model deployment solution
Example 📓 Jupyter notebooks that demonstrate how to build, train, and deploy machine learning models using 🧠 Amazon SageMaker.
JetStream is a throughput and memory optimized engine for LLM inference on XLA devices, starting with TPUs (and GPUs in future -- PRs welcome).
🏗️ Fine-tune, build, and deploy open-source LLMs easily!
MIVisionX toolkit is a set of comprehensive computer vision and machine intelligence libraries, utilities, and applications bundled into a single toolkit. AMD MIVisionX also delivers a highly optimized open-source implementation of the Khronos OpenVX™ and OpenVX™ Extensions.
A lightweight, fast, parallel inference server for Llama
Add a description, image, and links to the inference topic page so that developers can more easily learn about it.
To associate your repository with the inference topic, visit your repo's landing page and select "manage topics."