Pantograph: A Machine-to-Machine Interaction Interface for Advanced Theorem Proving, High Level Reasoning, and Data Extraction in Lean 4 Paper • 2410.16429 • Published Oct 21, 2024 • 5
Multi-Draft Speculative Sampling: Canonical Architectures and Theoretical Limits Paper • 2410.18234 • Published Oct 23, 2024 • 5
Data Scaling Laws in Imitation Learning for Robotic Manipulation Paper • 2410.18647 • Published Oct 24, 2024 • 6
ZIP-FIT: Embedding-Free Data Selection via Compression-Based Alignment Paper • 2410.18194 • Published Oct 23, 2024 • 6
Should We Really Edit Language Models? On the Evaluation of Edited Language Models Paper • 2410.18785 • Published Oct 24, 2024 • 7
The Nature of Mathematical Modeling and Probabilistic Optimization Engineering in Generative AI Paper • 2410.18441 • Published Oct 24, 2024 • 7
Asynchronous RLHF: Faster and More Efficient Off-Policy RL for Language Models Paper • 2410.18252 • Published Oct 23, 2024 • 7
Value Residual Learning For Alleviating Attention Concentration In Transformers Paper • 2410.17897 • Published Oct 23, 2024 • 8
ADEM-VL: Adaptive and Embedded Fusion for Efficient Vision-Language Tuning Paper • 2410.17779 • Published Oct 23, 2024 • 9
Stable Consistency Tuning: Understanding and Improving Consistency Models Paper • 2410.18958 • Published Oct 24, 2024 • 10
Robust Watermarking Using Generative Priors Against Image Editing: From Benchmarking to Advances Paper • 2410.18775 • Published Oct 24, 2024 • 10
DeCoRe: Decoding by Contrasting Retrieval Heads to Mitigate Hallucinations Paper • 2410.18860 • Published Oct 24, 2024 • 11
CCI3.0-HQ: a large-scale Chinese dataset of high quality designed for pre-training large language models Paper • 2410.18505 • Published Oct 24, 2024 • 11
WAFFLE: Multi-Modal Model for Automated Front-End Development Paper • 2410.18362 • Published Oct 24, 2024 • 13
MotionCLR: Motion Generation and Training-free Editing via Understanding Attention Mechanisms Paper • 2410.18977 • Published Oct 24, 2024 • 15
Skywork-Reward: Bag of Tricks for Reward Modeling in LLMs Paper • 2410.18451 • Published Oct 24, 2024 • 18
Why Does the Effective Context Length of LLMs Fall Short? Paper • 2410.18745 • Published Oct 24, 2024 • 18