COLLECTED BY
Organization:
Internet Archive
Focused crawls are collections of frequently-updated webcrawl data from narrow (as opposed to broad or wide) web crawls, often focused on a single domain or subdomain.
The Wayback Machine - https://web.archive.org/web/20200726161432/https://github.com/topics/inference
Here are
439 public repositories
matching this topic...
ncnn is a high-performance neural network inference framework optimized for the mobile platform
MediaPipe is the simplest way for researchers and developers to build world-class ML solutions and applications for mobile, edge, cloud and the web.
YOLOv3 in PyTorch > ONNX > CoreML > iOS
Updated
Jul 26, 2020
Jupyter Notebook
💎 1MB lightweight face detection model (1MB轻量级人脸检测模型)
Updated
Jun 20, 2020
Python
Runtime type system for IO decoding/encoding
Updated
Jul 21, 2020
TypeScript
Hello AI World guide to deploying deep-learning inference networks and deep vision primitives with TensorRT and NVIDIA Jetson.
Grakn Core: The Knowledge Graph
Updated
Jul 24, 2020
Java
TensorFlow template application for deep learning
Updated
Jan 3, 2019
Python
OpenVINO™ Toolkit repository
An easy to use PyTorch to TensorRT converter
Updated
Jul 8, 2020
Python
Acceleration package for neural networks on multi-core CPUs
The Triton Inference Server provides a cloud inferencing solution optimized for NVIDIA GPUs.
DELTA is a deep learning based natural language and speech processing platform.
Updated
Jul 20, 2020
Python
A curated list of research in machine learning system. I also summarize some papers if I think they are really interesting.
A uniform interface to run deep learning models from multiple frameworks
Multi Model Server is a tool for serving neural net models for inference
Updated
Jul 22, 2020
Java
Pytorch-Named-Entity-Recognition-with-BERT
Updated
Jan 24, 2020
Python
The challenge projects for Inferencing machine learning models on iOS
Updated
Nov 26, 2019
Python
TensorFlow models accelerated with NVIDIA TensorRT
Updated
Nov 22, 2019
Python
Embedded and mobile deep learning research resources
High-efficiency floating-point neural network inference operators for mobile, server, and Web
Shape inference for PyTorch (like Keras) & new layers
Updated
May 30, 2020
Python
a fast and user-friendly runtime for transformer inference (Bert, Albert, GPT2, Decoders, etc) on CPU and GPU.
A curated list of awesome TensorFlow Lite models, samples, tutorials, tools and learning resources.
A REST API for Caffe using Docker and Go
TensorFlow examples in C, C++, Go and Python without bazel but with cmake and FindTensorFlow.cmake
Updated
Aug 18, 2019
CMake
Lua Language Server coded by Lua
A scalable inference server for models optimized with OpenVINO™
Updated
Jul 23, 2020
Python
Fast implementation of BERT inference directly on NVIDIA (CUDA, CUBLAS) and Intel MKL
Improve this page
Add a description, image, and links to the
inference
topic page so that developers can more easily learn about it.
Curate this topic
Add this topic to your repo
To associate your repository with the
inference
topic, visit your repo's landing page and select "manage topics."
Learn more
You can’t perform that action at this time.
You signed in with another tab or window. Reload to refresh your session.
You signed out in another tab or window. Reload to refresh your session.