-
Rho-1: Not All Tokens Are What You Need
Paper • 2404.07965 • Published • 90 -
LLM2Vec: Large Language Models Are Secretly Powerful Text Encoders
Paper • 2404.05961 • Published • 65 -
Compression Represents Intelligence Linearly
Paper • 2404.09937 • Published • 27 -
Multi-Head Mixture-of-Experts
Paper • 2404.15045 • Published • 60
Peter Liu
Aozaki-Shinji
·
AI & ML interests
None yet
Recent Activity
upvoted
a
paper
8 days ago
Visual-RFT: Visual Reinforcement Fine-Tuning
updated
a collection
about 1 month ago
Interesting Papers
updated
a collection
about 2 months ago
Interesting Papers
Organizations
None yet
Collections
1
models
None public yet
datasets
None public yet