Use your locally running AI models to assist you in your web browsing
#大语言模型#A generalized information-seeking agent system with Large Language Models (LLMs).
#自然语言处理#[ICML 2024] SqueezeLLM: Dense-and-Sparse Quantization
Model swapping for llama.cpp (or any local OpenAPI compatible server)
#自然语言处理#[NeurIPS 2024] KVQuant: Towards 10 Million Context Length LLM Inference with KV Cache Quantization
A nifty little library for working with Ollama in Elixir.
The .NET library to consume 100+ APIs: OpenAI, Anthropic, Google, DeepSeek, Cohere, Mistral, Azure, xAI, Perplexity, Groq, Ollama, LocalAi, and many more!
The PyVisionAI Official Repo
#大语言模型#Run Open Source/Open Weight LLMs locally with OpenAI compatible APIs
#大语言模型#MVP of an idea using multiple local LLM models to simulate and play D&D
Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on different ports and loading/unloading them on demand
#大语言模型#MVP of an idea using multiple local LLM models to simulate and play D&D
#大语言模型#Chat with your pdf using your local LLM, OLLAMA client.(incomplete)
#安全#The client for the Symmetry peer-to-peer inference network. Enabling users to connect with each other, share computational resources, and collect valuable machine learning data.
A local chatbot for managing docs