#计算机科学#Turn any computer or edge device into a command center for your computer vision projects.
#大语言模型#The goal of RamaLama is to make working with AI boring.
#计算机科学#The simplest way to serve AI/ML models in production
#计算机科学#An open-source computer vision framework to build and deploy apps in minutes
#自然语言处理#Python + Inference - Model Deployment library in Python. Simplest model inference server ever.
#计算机科学#A REST API for Caffe using Docker and Go
#计算机科学#This is a repository for an nocode object detection inference API using the Yolov3 and Yolov4 Darknet framework.
#计算机科学#This is a repository for an nocode object detection inference API using the Yolov4 and Yolov3 Opencv.
Work with LLMs on a local environment using containers
#计算机科学#This is a repository for an object detection inference API using the Tensorflow framework.
Serving AI/ML models in the open standard formats PMML and ONNX with both HTTP (REST API) and gRPC endpoints
#计算机科学#ONNX Runtime Server: The ONNX Runtime Server is a server that provides TCP and HTTP/HTTPS REST APIs for ONNX inference.
#计算机科学#Orkhon: ML Inference Framework and Server Runtime
K3ai is a lightweight, fully automated, AI infrastructure-in-a-box solution that allows anyone to experiment quickly with Kubeflow pipelines. K3ai is perfect for anything from Edge to laptops.
#计算机科学#Deploy DL/ ML inference pipelines with minimal extra code.
#计算机科学#A standalone inference server for trained Rubix ML estimators.
#大语言模型#Friendli: the fastest serving engine for generative AI
#下载器#Wingman is the fastest and easiest way to run Llama models on your PC or Mac.
Advanced inference pipeline using NVIDIA Triton Inference Server for CRAFT Text detection (Pytorch), included converter from Pytorch -> ONNX -> TensorRT, Inference pipelines (TensorRT, Triton server -...
#计算机科学#Fullstack machine learning inference template