Flaviu Cipcigan
@flaviucipcigan.bsky.social
6.4K followers 460 following 350 posts
Building AIs for scientific discovery. Discovered antibiotics and materials for carbon capture. Tango dancer. See more at flaviucipcigan.com. Opinions my own.
Posts Media Videos Starter Packs
Pinned
flaviucipcigan.bsky.social
One of my big motivations is accelerating science with AI.

Every discovery project had a beautiful aha moment, such as the structure of antibiotics emerging in the latent space of a model or a GFlowNet proposing new carbon capture materials.

Here's some of the threads I've wrote on this topic.
flaviucipcigan.bsky.social
there's a bunch of scripts that can migrate old posts through the api, such as this one: github.com/marcomaroni-...

i don't know if timestamps would migrate, but I've seen folks who have posts with older timestamps than the start of the bluesky
GitHub - marcomaroni-github/twitter-to-bluesky: Import all tweets exported from X/Twitter to a Bluesky account.
Import all tweets exported from X/Twitter to a Bluesky account. - marcomaroni-github/twitter-to-bluesky
github.com
flaviucipcigan.bsky.social
Super interesting application of program search

Goals are mapped to programs which are embedded in a latent space.

A fitness metric is assigned to the programs and program search is done to synthesise new human-like goals.
flaviucipcigan.bsky.social
Thanks! Not sure, I'll try it 🤔
flaviucipcigan.bsky.social
flaviucipcigan.bsky.social
What is large for a language model? Is it 400B, 70B or maybe 1T?

I think focus on raw number of parameters is a less useful frame than thinking about inference speed, cost and location of inference (on-device vs cloud).
flaviucipcigan.bsky.social
flaviucipcigan.bsky.social
ARC-AGI is one of the most interesting benchmarks in ML.

o3 achieving human-level on the semi-private eval feels like a significant breakthrough.

Calibrating, I'd say o3 is a GPT-1 or GPT-2 moment. The direction for improvement is getting clear, with more of the research fog getting lifted.
flaviucipcigan.bsky.social
flaviucipcigan.bsky.social
I've been reflecting today about OpenAI's five levels to measure progress in AI.

GPT-4 was at Level 1, conversational AI: a model competent at 0.1-1s tasks, like holding a conversation.

O1 / R1 reached Level 2, reasoners: a model solving 1-10min tasks such as basic coding tasks and math.
flaviucipcigan.bsky.social
flaviucipcigan.bsky.social
A critique I hear often of LLMs is that they don't have a notion of truth, that they are BS machines, in Frankfurt's sense.

I don't think that's quite right.

Here's two papers that helped me have a more nuanced view of this question.
flaviucipcigan.bsky.social
flaviucipcigan.bsky.social
I've been posting a lot about AI lately, so wanted to also share some of my work on the bio/chem side.

Antimicrobial peptides are proteins that kill bacteria. Most do so by making circular holes in their membranes.

In this fun to write paper, we showed fractal pores in bacterial membranes.
flaviucipcigan.bsky.social
flaviucipcigan.bsky.social
MetaGFN preprint is out 🥳

1/ When building AIs for science, it's important for the algorithms to discover beyond things we already know. This is why effective, open-ended exploration is important. Here we propose MetaGFN, an algorithm to effectively find distant modes in probability distributions.
MetaGFN: Exploring Distant Modes with Adapted Metadynamics for...
Generative Flow Networks (GFlowNets) are a class of generative models that sample objects in proportion to a specified reward function through a learned policy. They can be trained either...
arxiv.org
flaviucipcigan.bsky.social
flaviucipcigan.bsky.social
One of the exciting things happening in AI for Science has been the growth in lab automation.

Automated labs coupled with active learning are a super exciting area with lots of opportunities for progress.

I promised @cpaxton.bsky.social a short thread on this, so here it goes!

🧪
flaviucipcigan.bsky.social
flaviucipcigan.bsky.social
If you're interested in foundation models for materials and molecules, check out our repo: github.com/IBM/materials

We have three models released based on SMILES, SELFIES and molecular graphs.

More to come shortly - we aim to have a unified collection of state-of art models across all modalities.
GitHub - IBM/materials: Foundation Model for Materials - FM4M
Foundation Model for Materials - FM4M. Contribute to IBM/materials development by creating an account on GitHub.
github.com
flaviucipcigan.bsky.social
flaviucipcigan.bsky.social
There's a lot of enthusiasm in the community about transformers trained on chemical or biological data.

Here's some interesting results and some thoughts on future directions.
timkellogg.me
this is nuts

a new 7B llama-style LLM for embedding of genomes & detection of pathogens in wastewater

i’ve had a hunch that LLMs could lead to some big bio breakthroughs, since it feels like genes & proteins are a lot like a language
flaviucipcigan.bsky.social
One of my big motivations is accelerating science with AI.

Every discovery project had a beautiful aha moment, such as the structure of antibiotics emerging in the latent space of a model or a GFlowNet proposing new carbon capture materials.

Here's some of the threads I've wrote on this topic.
flaviucipcigan.bsky.social
Wanna try to guess which of those gets parsed as a string and which as a number? Answer in alt text.

YAML parsing in python is weird.
{'lol': ['5.0E6',
  '5.0e6',
  '5.E6',
  '5.e6',
  '5E6',
  '5e6',
  5e-06,
  5e-06,
  5e-06,
  5e-06,
  '5E-6',
  '5e-6',
  5000000.0,
  5000000.0,
  5000000.0,
  5000000.0,
  '5E+6',
  '5e+6']}
flaviucipcigan.bsky.social
Interesting idea to generate responses using diffusion rather than left-to-right auto-regressive models
flaviucipcigan.bsky.social
Supercomputers - large computer clusters - allow you to work a number of years ahead.

Creating the GUI at PARC seemed like a "waste of FLOPs" but revolutionized computing.
From here https://www.youtube.com/watch?v=dZQ7x0-MZcI
flaviucipcigan.bsky.social
Where do large compute clusters come into play in this case?

Alan Kay talked about the Wayne Gretzky game, a hockey player famous for his quote about skating where the puck will be.
flaviucipcigan.bsky.social
Similarly, the benchmark scores of a model with a given number of parameters increases each generation due to better data and training algorithms, caveated by dataset leakage.
flaviucipcigan.bsky.social
For each generation, for a fixed parameter count, the speed of training & inferring decreases due to hardware and software advances, like flash attention and multi-head latent attention.

At each generation, larger and larger number of parameters can be ran locally.
flaviucipcigan.bsky.social
My first computer used a processor in the Intel 8086 generation, which had about 29k transistors.

Today, an Apple M4 has 28B transistors, meaning I experienced a scale-up of 1,000,000x in my lifetime.

I expect a similar scale-up for language models.
flaviucipcigan.bsky.social
What is large for a language model? Is it 400B, 70B or maybe 1T?

I think focus on raw number of parameters is a less useful frame than thinking about inference speed, cost and location of inference (on-device vs cloud).