Large Language Model Text Generation Inference
-
Updated
Jun 8, 2024 - Python
Large Language Model Text Generation Inference
🤘 TT-NN operator library, and TT-Metalium low level kernel programming model.
Run any open-source LLMs, such as Llama 2, Mistral, as OpenAI compatible API endpoint in the cloud.
The simplest way to serve AI/ML models in production
This is the official implementation of "LLM-QBench: A Benchmark Towards the Best Practice for Post-training Quantization of Large Language Models", and it is also an efficient LLM compression tool with various advanced compression methods, supporting multiple inference backends.
Toolkit for fine-tuning, ablating and unit-testing open-source LLMs.
A friendly library for parsing HTTP request arguments, with built-in support for popular web frameworks, including Flask, Django, Bottle, Tornado, Pyramid, webapp2, Falcon, and aiohttp.
Complete set up guide for Humio now as Falcon Logscale on single node self hosted server.
PowerShell for CrowdStrike's OAuth2 APIs
🌸 Run LLMs at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading
Java Examples using LangChain4J for Generative AI using ChatGPT LLM, RAG and other open source LLMs. Sentiment Analysis, Application Context based ChatBots. Custom Data Handling. LLMs - GPT 3.5 / 4o, Gemini, Claude 3, Llama3, Phi-3, Gemma, Falcon 2, Mistral, Wizard Math
TQ42 Cryptography is a comprehensive suite of quantum-resistant and asymmetric post-quantum encryption algorithms and key management functions offered under the open source AGPLv3 license and available under a Commercial license from Terra Quantum.
Bouncy Castle adoption to wan24-Crypto
中文nlp解决方案(大模型、数据、模型、训练、推理)
Add a description, image, and links to the falcon topic page so that developers can more easily learn about it.
To associate your repository with the falcon topic, visit your repo's landing page and select "manage topics."