![The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence) artwork](https://is1-ssl.mzstatic.com/image/thumb/Podcasts113/v4/39/58/c6/3958c6ce-86e4-3b80-bfb9-840e1dfd7e4b/mza_491361902049110775.png/100x100bb.jpg)
Beyond Accuracy: Behavioral Testing of NLP Models with Sameer Singh - #406
The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence)
English - September 03, 2020 19:10 - 41 minutes - ★★★★★ - 323 ratingsTechnology News Tech News machinelearning artificialintelligence datascience samcharrington tech technology thetwimlaipocast thisweekinmachinelearning twiml twimlaipodcast Homepage Download Apple Podcasts Google Podcasts Overcast Castro Pocket Casts RSS feed
Today we’re joined by Sameer Singh, an assistant professor in the department of computer science at UC Irvine.
Sameer’s work centers on large-scale and interpretable machine learning applied to information extraction and natural language processing. We caught up with Sameer right after he was awarded the best paper award at ACL 2020 for his work on Beyond Accuracy: Behavioral Testing of NLP Models with CheckList.
In our conversation, we explore CheckLists, the task-agnostic methodology for testing NLP models introduced in the paper. We also discuss how well we understand the cause of pitfalls or failure modes in deep learning models, Sameer’s thoughts on embodied AI, and his work on the now famous LIME paper, which he co-authored alongside Carlos Guestrin.
The complete show notes for this episode can be found at twimlai.com/go/406.
Today we’re joined by Sameer Singh, an assistant professor in the department of computer science at UC Irvine.
Sameer’s work centers on large-scale and interpretable machine learning applied to information extraction and natural language processing. We caught up with Sameer right after he was awarded the best paper award at ACL 2020 for his work on Beyond Accuracy: Behavioral Testing of NLP Models with CheckList.
In our conversation, we explore CheckLists, the task-agnostic methodology for testing NLP models introduced in the paper. We also discuss how well we understand the cause of pitfalls or failure modes in deep learning models, Sameer’s thoughts on embodied AI, and his work on the now famous LIME paper, which he co-authored alongside Carlos Guestrin.
The complete show notes for this episode can be found at twimlai.com/go/406.