Bubbles
0 points · 15 days ago · 0 comments

Leaving the autoresearch loop going, the LLM was able to make 7.8% progress on the distillation task. I saw Andrej Karpathy’s Autoresearch results the other day and decided to give it a shot on a relatively difficult task: distilling a retrieval/reranking model into a much smaller 10MB model. The model was trained on MS MARCO v1.1 and SQuAD (~169K pairs total) and evaluated on MRR@10.

No comments yet. Log in to discuss on the Fediverse