-
The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits
Paper β’ 2402.17764 β’ Published β’ 608 -
Qwen2.5 Technical Report
Paper β’ 2412.15115 β’ Published β’ 345 -
Phi-3 Technical Report: A Highly Capable Language Model Locally on Your Phone
Paper β’ 2404.14219 β’ Published β’ 256 -
LLM in a flash: Efficient Large Language Model Inference with Limited Memory
Paper β’ 2312.11514 β’ Published β’ 259
Collections
Discover the best community collections!
Collections including paper arxiv:2307.09288
-
black-forest-labs/FLUX.1-dev
Text-to-Image β’ Updated β’ 1.6M β’ β’ 8.65k -
openai/whisper-large-v3-turbo
Automatic Speech Recognition β’ Updated β’ 6.84M β’ β’ 1.93k -
meta-llama/Llama-3.2-11B-Vision-Instruct
Image-Text-to-Text β’ Updated β’ 1.86M β’ β’ 1.3k -
deepseek-ai/DeepSeek-V2.5
Text Generation β’ Updated β’ 3.87k β’ 693
-
Qwen2.5 Technical Report
Paper β’ 2412.15115 β’ Published β’ 345 -
Qwen2.5-Coder Technical Report
Paper β’ 2409.12186 β’ Published β’ 140 -
Qwen2.5-Math Technical Report: Toward Mathematical Expert Model via Self-Improvement
Paper β’ 2409.12122 β’ Published β’ 3 -
DeepSeek LLM: Scaling Open-Source Language Models with Longtermism
Paper β’ 2401.02954 β’ Published β’ 43
-
Self-Play Preference Optimization for Language Model Alignment
Paper β’ 2405.00675 β’ Published β’ 27 -
FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness
Paper β’ 2205.14135 β’ Published β’ 13 -
Attention Is All You Need
Paper β’ 1706.03762 β’ Published β’ 50 -
FlashAttention-2: Faster Attention with Better Parallelism and Work Partitioning
Paper β’ 2307.08691 β’ Published β’ 8
-
The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits
Paper β’ 2402.17764 β’ Published β’ 608 -
Phi-3 Technical Report: A Highly Capable Language Model Locally on Your Phone
Paper β’ 2404.14219 β’ Published β’ 256 -
Llama 2: Open Foundation and Fine-Tuned Chat Models
Paper β’ 2307.09288 β’ Published β’ 244 -
LLM in a flash: Efficient Large Language Model Inference with Limited Memory
Paper β’ 2312.11514 β’ Published β’ 259