Views not mine!!
Can't wait to share more soon...
Can't wait to share more soon...
New Cookbook: Long Document Summarization + Evaluation.
We fine-tune Llama 3.1 8B to improve summarization of documents 32k tokens long and show outperformance over 70B models!
New Cookbook: Long Document Summarization + Evaluation.
We fine-tune Llama 3.1 8B to improve summarization of documents 32k tokens long and show outperformance over 70B models!
We fine-tune Llama 3.1 8B, with instruction loss masking, on the conversational CoQA dataset and show a 2x improvement in exact match score for outputs!
We fine-tune Llama 3.1 8B, with instruction loss masking, on the conversational CoQA dataset and show a 2x improvement in exact match score for outputs!
To learn how we can improve fine-tuning over long form conversational data I fine-tuned a bunch of models on the CoQA dataset and 2x'd performance!
Full code notebook below🔽
To learn how we can improve fine-tuning over long form conversational data I fine-tuned a bunch of models on the CoQA dataset and 2x'd performance!
Full code notebook below🔽
an architecture that extends a LLM to store a complete record of the agent's experiences using natural language, synthesize those memories over time into higher-level reflections, and retrieve them dynamically to plan behavior.
an architecture that extends a LLM to store a complete record of the agent's experiences using natural language, synthesize those memories over time into higher-level reflections, and retrieve them dynamically to plan behavior.
He covers:
> Prompt design systems - DSPy
> Sampling systems - top_p, dynamic temp etc.
> Tool use - databases, search, functions
> Evaluating system vs. LLMs
He covers:
> Prompt design systems - DSPy
> Sampling systems - top_p, dynamic temp etc.
> Tool use - databases, search, functions
> Evaluating system vs. LLMs
Basically a degree in LLM creation, everything is study-able and transparent!
OLMo - open language model
OLMoE -mixture of experts
Dolma - full pre-training dataset
Molmo/PixMo - VLM
Tulu 3 - post training and datasets
Basically a degree in LLM creation, everything is study-able and transparent!
OLMo - open language model
OLMoE -mixture of experts
Dolma - full pre-training dataset
Molmo/PixMo - VLM
Tulu 3 - post training and datasets
Generative Agent Simulations of 1,000 People - arxiv.org/abs/2411.10109
The generative agents replicate participants' responses 85% as accurately as participants replicate their own answers two weeks later.
Generative Agent Simulations of 1,000 People - arxiv.org/abs/2411.10109
The generative agents replicate participants' responses 85% as accurately as participants replicate their own answers two weeks later.
Basically a degree in LLM creation, everything is study-able and transparent!
OLMo - open language model
OLMoE -mixture of experts
Dolma - full pre-training dataset
Molmo/PixMo - VLM
Tulu 3 - post training and datasets
Basically a degree in LLM creation, everything is study-able and transparent!
OLMo - open language model
OLMoE -mixture of experts
Dolma - full pre-training dataset
Molmo/PixMo - VLM
Tulu 3 - post training and datasets
📚 1M context length with 100% accuracy and 4x cheaper then 4o -mini
📚 1M context length with 100% accuracy and 4x cheaper then 4o -mini
"what's already happened is much more important than anything else that's going to be done and then it's just going to be a long ways in applying it." - Thiel
"what's already happened is much more important than anything else that's going to be done and then it's just going to be a long ways in applying it." - Thiel
Counts visual, tactile, auditory, olfactory, taste data.
Interesting hypothesis.
Counts visual, tactile, auditory, olfactory, taste data.
Interesting hypothesis.