Ishika Agarwal
@wonderingishika.bsky.social
890 followers 400 following 22 posts
CS PhD @ UIUC | Data Efficiency NLP | Conversational AI | agarwalishika.github.io | same handle on twitter
Posts Media Videos Starter Packs
Pinned
wonderingishika.bsky.social
I'm so excited to share my latest paper called DELIFT along with Krishnateja Killamsetty, Lucian Popa, and Marina Danilevksy at IBM Research 🎉

We tackle expensive fine-tuning by selecting a small subset of informative data that targets a model's weaknesses.
wonderingishika.bsky.social
6/6 For more details, see:

Paper: arxiv.org/pdf/2502.09969
Code: github.com/agarwalishik...

Thank you so much to @dilekh.bsky.social and @convai-uiuc.bsky.social for their guidance and support during this project 🎉🎉
arxiv.org
wonderingishika.bsky.social
5/6 Finally, using our influence values, we pick a small subset & fine-tune the model. In our evaluation, we use 4 SOTA influence functions -- NN-CIFT achieves the same performance while using a model 34,000x smaller!
wonderingishika.bsky.social
4/6 Second, we train the InfluenceNetwork using basic mini-batch gradient descent, then let it estimate the influence for the remaining data. It has a very low error of 0.067!
wonderingishika.bsky.social
3/6 First, the neural network (called the “InfluenceNetwork”) needs to be trained. We compute influence values using existing methods -- but only for a tiny fraction of data (just 0.25%-5%).
wonderingishika.bsky.social
2/6 Estimating the value of data is expensive.

Past works use LLMs to estimate the influence of data -- we use small neural networks to *learn to estimate* influence, instead. This reduces costs and adapts to new data without heavy recomputation.

Here’s how it works:
wonderingishika.bsky.social
🚀Very excited about my new paper!

NN-CIFT slashes data valuation costs by 99% using tiny neural nets (205k params, just 0.0027% of 8B LLMs) while maintaining top-tier performance!
wonderingishika.bsky.social
Elated to announce that DELIFT has been accepted to ICLR'25 🎉 Looking forward to discussing it in Singapore!
wonderingishika.bsky.social
I'm so excited to share my latest paper called DELIFT along with Krishnateja Killamsetty, Lucian Popa, and Marina Danilevksy at IBM Research 🎉

We tackle expensive fine-tuning by selecting a small subset of informative data that targets a model's weaknesses.
wonderingishika.bsky.social
Thank you Guneet! Would love to hear more about these stress tests :)
wonderingishika.bsky.social
Hey! Would love to be added :)
Reposted by Ishika Agarwal
pkargupta.bsky.social
Can LLMs make us critical thinkers?

TreeInstruct reorients assistant-like LLMs to be instructors that guide students towards understanding their mistakes, without providing direct/indirect answers.

Check out aclanthology.org/2024.finding... (w/ @wonderingishika.bsky.social) to learn more!
wonderingishika.bsky.social
All around the theme of data-efficient NLP:

(1) using influence functions to improve language model performance from less data
(2) enabling language models to generate queries for things it doesn't know
lastpositivist.bsky.social
Bluesky academics, lets get to know each other! Quote this & tell me: 1) a project you are working on & 2) an odd idea/theory you aren’t working on but keep thinking about

1. I came to hate my work and thinking so don't do it anymore.
2.
etvpod.bsky.social
Bluesky academics, lets get to know each other! Quote this & tell me: 1) a project you are working on & 2) an odd idea/theory you aren’t working on but keep thinking about

1. Convincing everyone that everything is luck, all the way down.

2. LLM’s can reason and understand in the external sense.
wonderingishika.bsky.social
For more details, see:
Paper: arxiv.org/pdf/2411.04425
Code: github.com/agarwalishik...

Thank you so much to Krishnateja, Lucian, and Marina for their help, mentorship, and guidance during this project! 🎉🎉
arxiv.org
wonderingishika.bsky.social
3. Continual fine-tuning: given a fine-tuned model, enabling it to integrate new and complementary information while mitigating catastrophic forgetting. We find that reducing the dataset helps remove samples that hinder performance, surpassing the performance of the full dataset.
wonderingishika.bsky.social
2. Task-specific fine-tuning: given an instruction-tuned model, refining the LLM's expertise in specific domains. We find that pruning the dataset removes noise and keeps relevant examples, achieving better performance than fine-tuning on the full dataset.
wonderingishika.bsky.social
1. Instruction tuning: given a base model, fine-tuning a model to follow general instructions. We find that performance drops are minimal when reducing the dataset by 70%.
wonderingishika.bsky.social
DELIFT quantifies the information present in a sample wrt an LLM's capabilities. Using submodular functions, DELIFT can automatically adapt the chosen subset based on the objectives in the 3 stages of language model fine-tuning:
wonderingishika.bsky.social
I'm so excited to share my latest paper called DELIFT along with Krishnateja Killamsetty, Lucian Popa, and Marina Danilevksy at IBM Research 🎉

We tackle expensive fine-tuning by selecting a small subset of informative data that targets a model's weaknesses.
wonderingishika.bsky.social
TreeInstruct is preferred 78.43% of the time. It solves 14.09% more bugs across all settings, and our questions are 14.18% better at addressing bugs, maintaining relevance, and ensuring logical conversation flow. TreeInstruct also adapts to human students of varying backgrounds.
wonderingishika.bsky.social
TreeInstruct estimates the knowledge a student needs to debug their code and devises a conversation plan. It then dynamically constructs a question tree based on its interactions with the student, navigating the knowledge state space till the student comprehends & fixes all bugs.
wonderingishika.bsky.social
I'd love to be added - thank you!!