#计算机科学#Turn any computer or edge device into a command center for your computer vision projects.
#大语言模型#The goal of RamaLama is to make working with AI boring.
#计算机科学#The simplest way to serve AI/ML models in production
#计算机科学#An open-source computer vision framework to build and deploy apps in minutes
#自然语言处理#Python + Inference - Model Deployment library in Python. Simplest model inference server ever.
#计算机科学#A REST API for Caffe using Docker and Go
#计算机科学#This is a repository for an nocode object detection inference API using the Yolov3 and Yolov4 Darknet framework.
#计算机科学#This is a repository for an nocode object detection inference API using the Yolov4 and Yolov3 Opencv.
Work with LLMs on a local environment using containers
#计算机科学#This is a repository for an object detection inference API using the Tensorflow framework.
Serving AI/ML models in the open standard formats PMML and ONNX with both HTTP (REST API) and gRPC endpoints
#计算机科学#ONNX Runtime Server: The ONNX Runtime Server is a server that provides TCP and HTTP/HTTPS REST APIs for ONNX inference.
#计算机科学#Orkhon: ML Inference Framework and Server Runtime
K3ai is a lightweight, fully automated, AI infrastructure-in-a-box solution that allows anyone to experiment quickly with Kubeflow pipelines. K3ai is perfect for anything from Edge to laptops.
#计算机科学#Deploy DL/ ML inference pipelines with minimal extra code.
#计算机科学#A standalone inference server for trained Rubix ML estimators.
#下载器#Wingman is the fastest and easiest way to run Llama models on your PC or Mac.
#大语言模型#Friendli: the fastest serving engine for generative AI
Advanced inference pipeline using NVIDIA Triton Inference Server for CRAFT Text detection (Pytorch), included converter from Pytorch -> ONNX -> TensorRT, Inference pipelines (TensorRT, Triton server -...
#计算机科学#Fullstack machine learning inference template