Leonardo Cotta
@cottascience.bsky.social
1K followers 250 following 55 posts
scaling lawver @ EIT from BH🔺🇧🇷 http://cottascience.github.io
Posts Media Videos Starter Packs
Reposted by Leonardo Cotta
thematterlab.bsky.social
We're excited to present our latest article in Nature Machine Intelligence - Boosting the predictive power of protein representations with a corpus of text annotations.

Link: www.nature.com/articles/s42...
[1/4]
cottascience.bsky.social
I’d add data/task understanding as a separate mid layer. Most papers I know break in the transition of high to mid.
cottascience.bsky.social
This is why I personally love TMLR. If it's correct and well-written let's publish. The interesting papers are the ones the community actively recognizes in their work, e.g. citing, extending, turning into products, etc. (independent process of publication).
cottascience.bsky.social
I agree with most of your thread, but classifying "uninteresting work" is quite hard nowadays. Papers became this "hype-seeking" game, where out of the 10 hyped papers of the month, at most 1 survives further investigation of the results. And even if we think we're immune to this, what is interest?
cottascience.bsky.social
I loved this new preprint by Lourie/Hu/ @kyunghyuncho.bsky.social . If you really wanna convince someone youre training a foundation model, or proposing better methodology, loss scaling laws aren't enough. It has to be tied w/ downstream performance. it shouldn't be vibes
arxiv.org/abs/2507.00885
Scaling Laws Are Unreliable for Downstream Tasks: A Reality Check
Downstream scaling laws aim to predict task performance at larger scales from pretraining losses at smaller scales. Whether this prediction should be possible is unclear: some works demonstrate that t...
arxiv.org
cottascience.bsky.social
We're at ICML, drop us a line if you're excited about this direction.

📄 Paper: arxiv.org/abs/2507.02083
💻 Code: github.com/h4duan/SciGym
🌍 Website: h4duan.github.io/scigym-bench...
🗂️ Dataset: huggingface.co/datasets/h4d...
cottascience.bsky.social
I'm very excited about our new work: SciGym. How can we scale scientific agents' evaluation?
TLDR; Systems biologists have spent decades encoding biochemical networks (metabolic pathways, gene regulation, etc.) into machine-runnable systems. We can use these as "dry labs" to test AI agents!
cottascience.bsky.social
Also, I see ITCS more like a “out of the box”, “bold” idea or even new area, I don’t see the papers having simplicity as a goal, but just my experience.
cottascience.bsky.social
Mhm, I agree with the idealistic part, I certainly have seen the same. But I know quite a few papers that are aligned w the call, tbh this happens in any venue. I think the message and the openness to this kind of paper is important though
cottascience.bsky.social
I wish we had an ML equivalent of SOSA (Symposium On Simplicity in Algorithms). "simpler algorithms manifest a better understanding of the problem at hand; they are more likely to be implemented and trusted by practitioners; they are more easily taught" www.siam.org/conferences-....
cottascience.bsky.social
this is not my area, but if you think of it in terms of a randomized algorithm (BPP,PP), the hard part is usually the generation, at least for the algorithms we tend to design. e.g. Schwartz-Zippel Lemma. (Although in theory you can have the "hard part" in verification for any problem)
cottascience.bsky.social
It takes 1 terrible paper for knowledgeable people to stop reading all your papers, this risk is often not accounted for
cottascience.bsky.social
Maybe check Cat s22, it gives you the basics, eg whatsapp+gps and nothing else
Reposted by Leonardo Cotta
quaidmorris.bsky.social
Please check out our new approach to modeling somatic mutation signatures.

DAMUTA has independent Damage and Misrepair signatures whose activities are more interpretable and more predictive of DNA repair defects, than COSMIC SBS signatures 🧬🖥️🧪

www.biorxiv.org/content/10.1...
Damage and Misrepair Signatures: Compact Representations of Pan-cancer Mutational Processes
Mutational signatures of single-base substitutions (SBSs) characterize somatic mutation processes which contribute to cancer development and progression. However, current mutational signatures do not ...
www.biorxiv.org
cottascience.bsky.social
it just sounds like "see you three times" ;) it's like some people named "Sinho" that is often confused with portuguese/brazilians; but from what I heard it's a variation of Singh (not sure though)
cottascience.bsky.social
One simple way to reason about this: treatment assignment guarantees you have the right P(T|X). Self-selection changes P(X), a different quantity. Looking at your IPW estimator you can see that changing P(X) will bias regardless of P(T|X).
cottascience.bsky.social
I haven't been up to date with the model collapse literature, but it's crazy the amount of papers that consider the case where people only reuse data from the model distribution. This never happens, there's always some human curation or conditioning that yields some type of "real-world, new, data".
cottascience.bsky.social
this general idea of using an external world/causal model given by a human and using the LM only for inference is really cool ---it's also the insight behind our work in NATURAL. Do you guys think it's possible to write a more general software for the interface DAG->LLM_inference->estimate?
Reposted by Leonardo Cotta
dkthomp.bsky.social
Unbelievable news.

Pancreatic is one of the deadliest cancers.

New paper shows personalized mRNA vaccines can induce durable T cells that attack pancreatic cancer, with 75% of patients cancer free at three years—far, far better than standard of care.

www.nature.com/articles/s41...
cottascience.bsky.social
Oh gotcha. I think it’s just super cheesy to quote feynman at this point haha but it’s a good philosophy to embrace
cottascience.bsky.social
In what contexts do you think it’s misused? Just curious, I’m a big fan and might be overusing it 😅
Reposted by Leonardo Cotta
thomwolf.bsky.social
After 6+ months in the making and over a year of GPU compute, we're excited to release the "Ultra-Scale Playbook": hf.co/spaces/nanot...

A book to learn all about 5D parallelism, ZeRO, CUDA kernels, how/why overlap compute & coms with theory, motivation, interactive plots and 4000+ experiments!
The Ultra-Scale Playbook - a Hugging Face Space by nanotron
The ultimate guide to training LLM on large GPU Clusters
hf.co
cottascience.bsky.social
if you're feeling uninspired and getting nan's everywhere, you can give your codebase, describe the problem and ask for suggestions to try or debug. I think of it more as a debugger assistant than a code generator.