Search "AI Inference"
6 results
A Python library that lets anyone easily use top-tier AI models for text, images, and audio without training from scratch.
Tencent's open-source mobile AI inference framework, runs deep learning models offline on phones with high speed and no dependencies.
A mobile AI app by Google that lets you experience large language models and generative AI features offline, ensuring privacy and speed.
Apache TVM is an open-source compiler framework that compiles AI models into efficient code, making them run faster and more compatible on various hardware like CPUs and GPUs.
A high-performance computing library for NVIDIA GPUs that significantly accelerates large model training and inference, boosting underlying code efficiency.
A tool that speeds up large model generation by predicting blocks of content to reduce waiting time.