[FEEDBACK] Daily Papers

Note that this is not a post about adding new papers, it's about feedback on the Daily Papers community update feature.
How to submit a paper to the Daily Papers, like @akhaliq (AK)?
- Submitting is available to paper authors
- Only recent papers (less than 7d) can be featured on the Daily
Then drop the arxiv id in the form at https://huggingface.co/papers/submit
- Add medias to the paper (images, videos) when relevant
- You can start the discussion to engage with the community
Please check out the documentation
We are excited to share our recent work on MLLM architecture design titled "Ovis: Structural Embedding Alignment for Multimodal Large Language Model".
Paper: https://arxiv.org/abs/2405.20797
Github: https://github.com/AIDC-AI/Ovis
Model: https://huggingface.co/AIDC-AI/Ovis-Clip-Llama3-8B
Data: https://huggingface.co/datasets/AIDC-AI/Ovis-dataset
we are excited to share our work titled "Hierarchical Prompting Taxonomy: A Universal Evaluation Framework for Large Language Models" : https://arxiv.org/abs/2406.12644
Hello AK and HF Team,
We would like to add our paper "GenConViT: Deepfake Video Detection Using Generative Convolutional Vision Transformer"
paper: https://arxiv.org/abs/2307.07036
code: https://github.com/erprogs/GenConViT
Thank you!,
Deressa
doi:10.57967/hf/4791
π ScrapeGoat Music Models: Now Available! π΅
After our recent work on HCF-Aware Training to optimize memory efficiency in language models (especially beneficial for specialized models like ours), we're excited to announce that all training and weights for the ScrapeGoat Music project are now available in our Hugging Face repositories:
ALL training python codes and weights available in the repo on HF.
Dear AK and HF Team,
Exciting Update! We're thrilled to share WritingBench: A Comprehensive Framework for Evaluating Generative Writing
π [Paper] β’ π [Github Repo] β’ π [Critic Model] β’ βοΈ [Writing Model]
π‘WritingBench is a comprehensive benchmark for evaluating LLMs' writing capabilities across 1,239 real-world queries, spanning:
- 6 primary domains and 100 fine-grained subdomains
- 3 core writing requirements: Style / Format / Length
- 1,546 avg. tokens per query, integrating diverse sources of materials
- Each query is paired with 5 instance-specific criteria
Regards,
Yuning
Hello AK and HF Team,
We would to add our recent paper "Simple Semi-supervised Knowledge Distillation from Vision-Language Models via Dual-Head Optimization" in HF daily papers.
I'm putting request here because I don't have paper claimed in HF daily papers yet.
Paper: https://arxiv.org/pdf/2503.02424
Regards,
Seongjae

Hello AK and HF Team,
We would to add our recent paper "MM-Agent: LLM as Agents for Real-world Mathematical Modeling Problem"
Mathematical modeling is more than reasoning β it requires open-ended analysis, abstraction, and principled formulation. This work introduces MM-Agent, a large language model (LLM)-powered agent framework designed to tackle real-world mathematical modeling tasks end-to-end.
Key highlights:
π Proposes MM-Bench: 111 curated problems from MCM/ICM (2000β2025), across physics, biology, economics, etc.
π§© MM-Agent decomposes modeling into 4 expert-inspired stages:
Problem Analysis β Model Formulation β Problem Solving β Report Generation
π Outperforms baselines by 11.88% over expert-written solutions using GPT-4o, while costing just $0.88 and 15 minutes per task.
π Helped two undergrad teams win Finalist Award (top 2%) in MCM/ICM 2025.
π Paper: https://arxiv.org/abs/2505.14148
π» Code: https://github.com/usail-hkust/LLM-MM-Agent
Hi, everyone! We are happy to share with you our work SVD-Free Low-Rank Adaptive Gradient Optimization for Large Language Models.
We focus on the low-rank compression of optimizer states and propose replacing the expensive SVD decomposition with a fixed orthogonal matrix that comes from the Discrete Consine Transformation (DCT).
In our work we couple the DCT matrix with a theoretically-justified approach to choose the most appropriate columns from the DCT matrix that minimize the reconstruction error for each gradient matrix G and obtain a dynamic projection matrix tailored to each gradient G.
Our numerical results show that DCT matrix not only recovers the performance of existing low-rank optimizers, but also reduces the running time by 20% and memory usage for large models, both for pretraining and finetuning.
π Paper: https://arxiv.org/pdf/2505.17967
π Code: soon to appear in https://github.com/IST-DASLab/ISTA-DASLab-Optimizers via pip