Mariusz Kurman
mkurman.bsky.social
Mariusz Kurman
@mkurman.bsky.social
from bedside to byte_side, MD to AI, 🇵🇱
- Data Preview: Have data but unsure what's inside? Explore it directly!
- Verifier View: evaluate generated data, remove duplicates, assign ratings

and many more!

github.com/mkurman/synt...
GitHub - mkurman/synthlabs
Contribute to mkurman/synthlabs development by creating an account on GitHub.
github.com
January 11, 2026 at 10:44 AM
- Multi-turn Support: pass one DEEP run, let the model ask follow-up questions, and choose who should respond using SYNTH-like thinking
- Firebase/Firestore: download your data directly as a JSONL file or upload it to your Firestore (production mode)
January 11, 2026 at 10:44 AM
- Generator: create your own dataset from scratch
- Converter: use existing datasets (Hugging Face support) with reasoning traces to match our SYNTH style
- DEEP Mode: multiple agents working together in various configurations
January 11, 2026 at 10:44 AM
It has a minor bug that requires further fine-tuning (sometimes it starts with the <|python_tag|> instead of <Thought>.
January 4, 2025 at 1:17 PM
I will definitely give it a try!
December 4, 2024 at 10:28 PM
PS. You can find the model mentioned here: huggingface.co/meditsolutio...

License: Apache 2.0​​​​​​​​​​​​​​​​ 4/4
meditsolutions/SmolLM2-MedIT-Upscale-2B · Hugging Face
We’re on a journey to advance and democratize artificial intelligence through open source and open science.
huggingface.co
December 2, 2024 at 2:54 PM
the model achieved better results in IFEval and a higher overall average score in Open LLM Leaderboard

I consider this a big success 😇, since surpassing the original in metrics is often very time-consuming, generates high costs, and doesn't always work out. 3/4
meditsolutions/SmolLM2-MedIT-Upscale-2B · Hugging Face
We’re on a journey to advance and democratize artificial intelligence through open source and open science.
huggingface.co
December 2, 2024 at 2:54 PM
In total, not much really, since we don't have the original trained under the same conditions as our upscale. However...

1. We scaled up the model without losing its quality
2. We confirmed that the method we devised works
3. After extremely short fine-tuning, 2/4
meditsolutions/SmolLM2-MedIT-Upscale-2B · Hugging Face
We’re on a journey to advance and democratize artificial intelligence through open source and open science.
huggingface.co
December 2, 2024 at 2:54 PM
Polish has a nice set of open data from the SpeakLeash initiative.

Link: speakleash.org
SpeakLeash | Spichlerz
Najnowsze wiadomości O nas Otwarty projekt, którego celem jest zbudowanie zestawu danych dla Dużego Modelu Językowego, o rozmiarze co najmniej 1TB, składającego się z różnorodnych tekstów w języku pol...
speakleash.org
November 27, 2024 at 9:29 PM