Hello everyone,
In this week's reading discussion, we will talk about:
LLM2LLM: Boosting LLMs with Novel Iterative Data Enhancement<https://arxiv.org/abs/2403.15042>
Abstract: While many real-world applications still require fine-tuning to reach satisfactory levels of performance, many of them are in the low-data regime, making fine-tuning challenging. To address this, the authors propose LLM2LLM, a targeted and iterative data augmentation strategy that uses a teacher LLM to enhance a small seed dataset by augmenting additional data that can be used for fine-tuning on a specific task.
You could attend the meeting in the Data & Decision Building, Room 366, or on Zoom (https://virginiatech.zoom.us/j/4506040955?pwd=TGFtTUVWR2tuYmVlS3JraTQzbjgrU…) at 4-5 pm on Friday. Note that the Zoom link is different from the previous reading discussions.
Happy reading!
Best regards,
Barry Menglong Yao
Ph.D. student in Computer Science
Virginia Tech
Homepage<https://barry-yao.netlify.app/>