🥷 Run AI-agents with an API
-
Updated
May 23, 2024 - TypeScript
🥷 Run AI-agents with an API
A high-throughput and memory-efficient inference and serving engine for LLMs
Build high-quality LLM apps - from prototyping, testing to production deployment and monitoring.
Text Embedding for Retrieval, Rerank and RAG
LMDeploy is a toolkit for compressing, deploying, and serving LLMs.
A collection of prompts to challenge the reasoning abilities of large language models
🤖 The free, Open Source OpenAI alternative. Self-hosted, community-driven and local-first. Drop-in replacement for OpenAI running on consumer-grade hardware. No GPU required. Runs gguf, transformers, diffusers and many more models architectures. It allows to generate Text, Audio, Video, Images. Also with voice cloning capabilities.
A cloud-native vector database, storage for next generation AI applications
Learn all how to run Ollama in GitHub Codespaces for free
RAGFlow is an open-source RAG (Retrieval-Augmented Generation) engine based on deep document understanding.
Test your prompts, models, and RAGs. Catch regressions and improve prompt quality. LLM evals for OpenAI, Azure, Anthropic, Gemini, Mistral, Llama, Bedrock, Ollama, and other local & private models with CI/CD integration.
Too Long, Didn't Watch(TL/DW): Your Personal Research Multi-Tool
🔍 AI search engine - self-host with local or cloud LLMs
Drop-in, local AI alternative to the OpenAI stack. Multi-engine (llama.cpp, TensorRT-LLM). Powers 👋 Jan
📰 Must-read papers and blogs on LLM based Long Context Modeling 🔥
A NLQ(Natural Language Query) demo using Amazon Bedrock, Amazon OpenSearch with RAG technique.
A LLM client for use from the command line or IDE. 一个在命令行或者IDE中使用的大语言模型客户端
Add a description, image, and links to the llm topic page so that developers can more easily learn about it.
To associate your repository with the llm topic, visit your repo's landing page and select "manage topics."