Large Language Model Text Generation Inference
-
Updated
Apr 15, 2025 - Python
Large Language Model Text Generation Inference
🌸 Run LLMs at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading
🩹Editing large language models within 10 seconds⚡
[NeurIPS 2023] LLM-Pruner: On the Structural Pruning of Large Language Models. Support Llama-3/3.1, Llama-2, LLaMA, BLOOM, Vicuna, Baichuan, TinyLlama, etc.
Fast Inference Solutions for BLOOM
💬 Chatbot web app + HTTP and Websocket endpoints for LLM inference with the Petals client
LLMs4OL: Large Language Models for Ontology Learning
Train llm (bloom, llama, baichuan2-7b, chatglm3-6b) with deepspeed pipeline mode. Faster than zero/zero++/fsdp.
Okapi: Instruction-tuned Large Language Models in Multiple Languages with Reinforcement Learning from Human Feedback
一套代码指令微调大模型
Finetuning BLOOM on a single GPU using gradient-accumulation
Generate README.md with GPT-3 few-shot learning
Scrapy Redis with Bloom Filter,support redis sentinel and cluster
Easy-to-use framework for evaluating cross-lingual consistency of factual knowledge (Supported LLaMA, BLOOM, mT5, RoBERTa, etc.) Paper here: https://aclanthology.org/2023.emnlp-main.658/
Add a description, image, and links to the bloom topic page so that developers can more easily learn about it.
To associate your repository with the bloom topic, visit your repo's landing page and select "manage topics."