MLGym: A New Framework and Benchmark for Advancing AI Research Agents
AutoML is dead an LLMs have killed it? MLGym is a benchmark and framework testing this theory. Roberta Raileanu and Deepak Nathani discuss how well current LLMs are doing at solving ML tasks, what the biggest roadblocks are, and what that means for AutoML generally.Check out the paper: https://arxiv.org/pdf/2502.14499More on Roberta: https://rraileanu.github.io/More on Deepak: https://dnathani.net/
--------
1:28:33
--------
1:28:33
Leverage Foundational Models for Black-Box Optimization
Where and how can we use foundation models in AutoML? Richard Song, researcher at Google DeepMind, has some answers. Starting off from his position paper on leveraging foundation models for optimization, we chat about what makes foundation models valuable for AutoML, how the next steps could look like, but also why the community is not currently embracing the topic as much as it could.Paper Link: https://arxiv.org/abs/2405.03547Richard's website: https://xingyousong.github.io/
--------
56:48
--------
56:48
Nyckel - Building an AutoML Startup
Oscar Beijbom is talking about what it's like to run an AutoML startup: Nyckel. Beyond that, we chat about the differences between academia and industry, what truly matters in application and more.Check out Nyckel at: https://www.nyckel.com/
--------
1:20:59
--------
1:20:59
Neural Architecture Search: Insights from 1000 Papers
Colin White, head of research at Abacus AI, takes us on a tour of Neural Architecture Search: its origins, important paradigms and the future of NAS in the age of LLMs. If you're looking for a broad overview of NAS, this is the podcast for you!
--------
1:15:44
--------
1:15:44
Quick-Tune: Quickly Learning Which Pretrained Model to Finetune and How
There are so many great foundation models in many different domains - but how do you choose one for your specific problem? And how can you best finetune it? Sebastian Pineda has an answer: Quicktune can help select the best model and tune it for specific use cases. Listen to find out when this will be a Huggingface feature and if hyperparameter optimization is even important in finetuning models (spoiler: very much so)!