Moritz Laurer
banner
moritzlaurer.bsky.social
Moritz Laurer
@moritzlaurer.bsky.social
Machine Learning Engineer @hf.co Hugging Face
—without GPT-4-based data distillation.
💾 While we wait for the release of code and datasets, you can already download the prompts they used from the HF Hub!

Details here 👇
January 15, 2025 at 12:31 PM
🤖 A Process Preference Model (PPM) enables fine-grained evaluation of intermediate steps, improving training data quality.
🧪 The system underwent four rounds of self-evolution, progressively refining both the policy and reward models to tackle Olympiad-level math problems
January 15, 2025 at 12:31 PM
📏 The paper introduces rStar-Math, which claims to rival OpenAI o1's math reasoning capabilities by integrating Monte Carlo Tree Search (MCTS) with step-by-step verified reasoning trajectories.
January 15, 2025 at 12:31 PM
💾 You can now download and reuse these prompt templates via the prompt-templates library!

🔄 The library simplifies sharing prompt templates on the HF hub or locally via standardized YAML files. Let’s make LLM work more transparent and reproducible by sharing more templates like this!

Links 👇
January 11, 2025 at 11:14 AM
🧪 The authors tested different prompt templates on held-out data to ensure their generalization.

📚 It's highly educational to read these templates to learn how frontier labs design prompts and understand their limitations.
January 11, 2025 at 11:14 AM
📏 The paper introduces the FACTS Grounding benchmark for evaluating the factuality of LLM outputs.

🤖 Fact-checking is automated by an ensemble of LLM judges that verify if a response is fully grounded in a factual reference document.
January 11, 2025 at 11:14 AM
January 9, 2025 at 1:05 PM
⚖️ Mixture of judges: The new AllTrueJudge combines decisions from multiple binary judges for more nuanced evaluation.

Read the release notes and other resources here 👇
January 9, 2025 at 1:05 PM
🛠️ Tool call support: TRL preprocessing now supports tool integration, laying the groundwork for agent fine-tuning with examples like dynamic temperature fetching in prompts.
January 9, 2025 at 1:05 PM
Perfect for tasks like stepwise reasoning.
🔀 Model merging: A new callback leverages mergekit to merge models during training, improving performance by blending reference and policy models - optionally pushing merged models to the Hugging Face Hub.
January 9, 2025 at 1:05 PM
on revenue of $3.7 billion last year, with ChatGPT alone once costing an estimated $700,000 per day to operate. 💸🔥
- They build strong models and do great research. Whether this business model will work in the long run is one of the biggest questions in the AI economy.

Source with the numbers 👇
OpenAI is losing money on its pricey ChatGPT Pro plan, CEO Sam Altman says | TechCrunch
OpenAI CEO Sam Altman says that the company is currently losing money on its $200-per-month plan because people use it more than expected.
techcrunch.com
January 7, 2025 at 11:12 AM
Great work by @answerdotai !

If you’re looking for a high-speed zeroshot classifier, give it a try!

📄 Resources below: 👇
January 6, 2025 at 4:40 PM
- 💡 What’s next? I’m preparing a newer version trained on better + longer synthetic data to fully leverage the 8k context window and improve upon the training mix of my older zeroshot-v2.0 models. I also hope that there will be a multilingual variant in the future.
January 6, 2025 at 4:40 PM
- 📉 Performance tradeoff: It performs slightly worse than DeBERTav3 on average across my zeroshot classification task collection
- 🧠 Use cases: I recommend using it for scenarios requiring speed and a larger context window (8k).
January 6, 2025 at 4:40 PM
Congrats @answerdotai, @LightOnIO and collaborators like @tomaarsen.com !

Paper and models here 👇https://huggingface.co/collections/answerdotai/modernbert-67627ad707a4acbf33c41deb
December 20, 2024 at 2:21 PM