Feynman Innovations

ajibawa-2023

AI & ML interests

LLM, RL, DL, ML, AGI. Developing LLMs (preferably fully fine tuned ) for various use cases.

Recent Activity

reacted to singhsidhukuldeep's post with πŸ”₯ 12 days ago
Exciting Research Alert: Revolutionizing Complex Information Retrieval! A groundbreaking paper from researchers at MIT, AWS AI, and UPenn introduces ARM (Alignment-Oriented LLM-based Retrieval Method), a novel approach to tackle complex information retrieval challenges. >> Key Innovations Information Alignment The method first decomposes queries into keywords and aligns them with available data using both BM25 and embedding similarity, ensuring comprehensive coverage of information needs. Structure Alignment ARM employs a sophisticated mixed-integer programming solver to identify connections between data objects, exploring relationships beyond simple semantic matching. Self-Verification The system includes a unique self-verification mechanism where the LLM evaluates and aggregates results from multiple retrieval paths, ensuring accuracy and completeness. >> Performance Highlights The results are impressive: - Outperforms standard RAG by up to 5.2 points in execution accuracy on Bird dataset - Achieves 19.3 points higher F1 scores compared to existing approaches on OTT-QA - Reduces the number of required LLM calls while maintaining superior retrieval quality >> Technical Implementation The system uses a three-step process: 1. N-gram indexing and embedding computation for all data objects 2. Constrained beam decoding for information alignment 3. Mixed-integer programming optimization for structure exploration This research represents a significant step forward in making complex information retrieval more efficient and accurate. The team's work demonstrates how combining traditional optimization techniques with modern LLM capabilities can solve challenging retrieval problems.
View all activity

Organizations

Stanford AI's profile picture AI FILMS's profile picture Giskard's profile picture AI vulnerability Database (AVID)'s profile picture FreedomAI's profile picture Ontocord's M*DEL's profile picture The Waifu Research Department's profile picture Media Party 2023's profile picture Keynote Technology's profile picture ZeroGPU Explorers's profile picture Aurora-M's profile picture Ontocord.AI's profile picture One-Man-Army's profile picture MLX Community's profile picture Social Post Explorers's profile picture Cognitive Computations's profile picture M4-ai's profile picture Data Is Better Together Contributor's profile picture

ajibawa-2023's activity

reacted to singhsidhukuldeep's post with πŸ”₯ 12 days ago
view post
Post
3572
Exciting Research Alert: Revolutionizing Complex Information Retrieval!

A groundbreaking paper from researchers at MIT, AWS AI, and UPenn introduces ARM (Alignment-Oriented LLM-based Retrieval Method), a novel approach to tackle complex information retrieval challenges.

>> Key Innovations

Information Alignment
The method first decomposes queries into keywords and aligns them with available data using both BM25 and embedding similarity, ensuring comprehensive coverage of information needs.

Structure Alignment
ARM employs a sophisticated mixed-integer programming solver to identify connections between data objects, exploring relationships beyond simple semantic matching.

Self-Verification
The system includes a unique self-verification mechanism where the LLM evaluates and aggregates results from multiple retrieval paths, ensuring accuracy and completeness.

>> Performance Highlights

The results are impressive:
- Outperforms standard RAG by up to 5.2 points in execution accuracy on Bird dataset
- Achieves 19.3 points higher F1 scores compared to existing approaches on OTT-QA
- Reduces the number of required LLM calls while maintaining superior retrieval quality

>> Technical Implementation

The system uses a three-step process:
1. N-gram indexing and embedding computation for all data objects
2. Constrained beam decoding for information alignment
3. Mixed-integer programming optimization for structure exploration

This research represents a significant step forward in making complex information retrieval more efficient and accurate. The team's work demonstrates how combining traditional optimization techniques with modern LLM capabilities can solve challenging retrieval problems.
reacted to Tonic's post with πŸ”₯ 12 days ago
view post
Post
2067
πŸ™‹πŸ»β€β™‚οΈhey there folks ,

Goedel's Theorem Prover is now being demo'ed on huggingface : Tonic/Math

give it a try !
reacted to hba123's post with πŸ”₯ 12 days ago
view post
Post
1721
We developed a method that ensures almost-sure safety (i.e., safety with probability approaching 1). We proved this result. We then, present a practical implementation which we call InferenceGuard. InferenceGuard has impressive practical results: 91.04% on Alpaca-7B and 100% safety results on Beaver 7B-v3.

Now, it is easy to get high safety results like those if we want a dumb model, e.g., just don't answer or answer with EOS and so on. However, our goal is not to only have safe results, but also to make sure that the rewards are high - we want a good trade-off between safety and rewards! That's exactly, what we show. InferenceGuard achieves that!

Check it out: Almost Surely Safe Alignment of Large Language Models at Inference-Time (2502.01208)
reacted to davanstrien's post with πŸ‘ 19 days ago
reacted to DawnC's post with ❀️ about 1 month ago
view post
Post
2277
🌟 PawMatchAI: Making Breed Selection More Intuitive! πŸ•
Excited to share the latest update to this AI-powered companion for finding your perfect furry friend! I've made significant architectural improvements to enhance breed recognition accuracy and feature detection.

✨ What's New?
Enhanced breed recognition through advanced morphological feature analysis:
- Implemented a sophisticated feature extraction system that analyzes specific characteristics like body proportions, head features, tail structure, fur texture, and color patterns
- Added an intelligent attention mechanism that dynamically focuses on the most relevant features for each image
- Improved multi-dog detection capabilities through enhanced spatial feature analysis
- Achieved better precision in distinguishing subtle breed characteristics

🎯 Key Features:
Smart breed recognition powered by advanced AI architecture
Visual matching scores with intuitive color indicators
Detailed breed comparisons with interactive tooltips
Lifestyle-based recommendations tailored to your needs

πŸ’­ Project Vision
Combining my passion for AI and pets, this project represents another step toward creating meaningful AI applications. Each update aims to make the breed selection process more accessible while improving the underlying technology.

πŸ‘‰ Try it now: DawnC/PawMatchAI

Your likes ❀️ on this space fuel this project's growth!

#AI #MachineLearning #DeepLearning #Pytorch #ComputerVision #TechForLife
  • 2 replies
Β·
replied to DawnC's post about 1 month ago
reacted to clem's post with ❀️ about 1 month ago
view post
Post
4310
Cool to see @ylecun joining the top 10 of most followed on HF!

(and leaderboard by @mvaloatto is here: mvaloatto/TCTF)
  • 2 replies
Β·
reacted to AkimfromParis's post with πŸ‘€ about 1 month ago
view post
Post
1692
πŸ’΅ Polymarket is leveraging β€œChatbot Arena LLM Leaderboard” on HuggingFace for online gambling on the β€œTop AI model on January 31?”. πŸ€—

As of January 3rd, 2025:
-1./ Gemini (83%) -2./ ChatGPT (13%) -3./ Other (2%) -4./ Claude (2%) -5./ Grok (1%) -6./ Llama (<1%)

πŸ‡ΊπŸ‡Έ The market opinion is following historical data. It's clearly bias towards US historical AI giants, yet Polymarket is forbidden in the USA and for US citizens.

πŸ‡¨πŸ‡³ In the β€œOther”, you might have Chinese AI labs that are probably the future AI leaders (Qwen, DeepSeek, Yi).

βš–οΈ In the market resolution, if two models are tied in the evaluation, they will take the alphabetical order. (e.g. if both were tied, β€œGoogle” would resolve to β€œYes”, and β€œxAI” would resolve to β€œNo”). πŸ™ƒ

That might be illegal usage of the Chatbot Arena policy? And maybe HuggingFace? @clem
Or maybe authors and contributors should get a cut each month as β€œmarket markers”.Β  @weichiang @angelopoulos
  • 1 reply
Β·
reacted to cfahlgren1's post with πŸ‘ about 1 month ago
view post
Post
2245
You'll notice the AI in the SQL Console is much better at working with chatml conversations:

Here's example of unnesting the cfahlgren1/react-code-instructions in less than 10 seconds by asking it. Check it out here: cfahlgren1/react-code-instructions

- "show me the average assistant response length"
- "extract user, system, and assistant messages into separate columns"

It's super easy to work with conversational datasets now with natural language πŸ—£οΈ





  • 2 replies
Β·
reacted to Kseniase's post with πŸ‘ about 2 months ago
view post
Post
3100
**15 Agentic Systems and Frameworks of 2024**

This year, we started our β€œAI Agents and Agentic Workflows” series (https://www.turingpost.com/t/AI-Agents) to explore everything about AI agents step by step: all the vocabulary, how they work, and how to build them.
The huge interest in this series and the large number of studies conducted on agents showed that it was one of the most popular and important themes of the year. In 2025, most likely, agents will reach new highs – we will be covering that for you. Now, let’s review the agentic systems that have emerged this year.

Here is a list of 15 agentic systems and frameworks of 2024:

1. GUI Agents: A Survey (2412.13501)

2. Large Language Models Orchestrating Structured Reasoning Achieve Kaggle Grandmaster Level (2411.03562)

3. The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery (2408.06292)

4. MALT: Improving Reasoning with Multi-Agent LLM Training (2412.01928)

5. Agent S: An Open Agentic Framework that Uses Computers Like a Human (2410.08164)

6. Automated Design of Agentic Systems (2408.08435)

7. AgentInstruct: Toward Generative Teaching with Agentic Flows (2407.03502)

8. AgentStore: Scalable Integration of Heterogeneous Agents As Specialized Generalist Computer Assistant (2410.18603)

9. WALL-E: World Alignment by Rule Learning Improves World Model-based LLM Agents (2410.07484)

10. Generative Agent Simulations of 1,000 People (2411.10109)

11. DynaSaur: Large Language Agents Beyond Predefined Actions (2411.01747)

12. PRefLexOR: Preference-based Recursive Language Modeling for Exploratory Optimization of Reasoning and Agentic Thinking (2410.12375)

13. Generative World Explorer (2411.11844)

14. Bel Esprit: Multi-Agent Framework for Building AI Model Pipelines (2412.14684)

15. AutoKaggle: A Multi-Agent Framework for Autonomous Data Science Competitions (2410.20424)

Thanks for reading Turing Post!
Subscribe to receive new posts straight into your inbox -> https://www.turingpost.com/subscribe
reacted to sayakpaul's post with πŸ”₯ about 2 months ago
view post
Post
4343
Commits speak louder than words πŸ€ͺ

* 4 new video models
* Multiple image models, including SANA & Flux Control
* New quantizers -> GGUF & TorchAO
* New training scripts

Enjoy this holiday-special Diffusers release πŸ€—
Notes: https://github.com/huggingface/diffusers/releases/tag/v0.32.0
replied to MoritzLaurer's post about 2 months ago
reacted to MoritzLaurer's post with πŸ”₯ about 2 months ago
replied to di-zhang-fdu's post about 2 months ago
reacted to di-zhang-fdu's post with πŸ€— about 2 months ago
reacted to davidberenstein1957's post with πŸ”₯ 3 months ago
view post
Post
1593
πŸ”₯ Dataset Drop - Open Image Preferences

BlackForest Labs Flux Dev VS. Stability AI Stable Diffusion Large 3.5

Together with the ⁠data-is-better-together community, we've worked on an Apache 2.0 licensed open image preference dataset based on the fal ai imgsys prompts dataset. Thanks to the awesome community, we have managed to get 5K preference pairs in less than 2 days. The annotation alignment among annotators is great too.

Aashish Kumar won a month of Hugging Face Pro by making the most contributions! Congrats from the entire team πŸ₯‡

The best thing?! We are not done yet! Let's keep the annotations coming for 5K more in the second part of the sprint! (with more prices to go around).

Dataset: https://huggingface.co/datasets/data-is-better-together/image-preferences-results
reacted to MohamedRashad's post with πŸš€ 3 months ago
reacted to di-zhang-fdu's post with πŸ‘ 3 months ago
view post
Post
6405
LLaMA-O1: Open Large Reasoning Model Frameworks For Training, Inference and Evaluation With PyTorch and HuggingFace
Large Reasoning Models powered by Monte Carlo Tree Search (MCTS), Self-Play Reinforcement Learning, PPO, AlphaGo Zero's dua policy paradigm and Large Language Models!
https://github.com/SimpleBerry/LLaMA-O1/

What will happen when you compound MCTS ❀ LLM ❀ Self-Play ❀RLHF?
Just a little bite of strawberry!πŸ“

Past related works:
LLaMA-Berry: Pairwise Optimization for O1-like Olympiad-Level Mathematical Reasoning (2410.02884)
Accessing GPT-4 level Mathematical Olympiad Solutions via Monte Carlo Tree Self-refine with LLaMa-3 8B (2406.07394)
  • 2 replies
Β·
reacted to Jaward's post with πŸ”₯ 3 months ago
view post
Post
2115
It's work like this that in some way signal the eventual β€œdominance” of AI over all the sciences.

β€œWe train our model on the six-dimensional N-body phase space, predicting particle velocities as the time derivative of the model’s displacement outputs”

The emulator is capable of predicting
the nonlinear displacement and velocity fields for 128^3 particles in half a second on a single GPU🀯
  • 1 reply
Β·
replied to their post 4 months ago