Artwork

Elder Research에서 제공하는 콘텐츠입니다. 에피소드, 그래픽, 팟캐스트 설명을 포함한 모든 팟캐스트 콘텐츠는 Elder Research 또는 해당 팟캐스트 플랫폼 파트너가 직접 업로드하고 제공합니다. 누군가가 귀하의 허락 없이 귀하의 저작물을 사용하고 있다고 생각되는 경우 여기에 설명된 절차를 따르실 수 있습니다 https://ko.player.fm/legal.
Player FM -팟 캐스트 앱
Player FM 앱으로 오프라인으로 전환하세요!

Debate: One Metric to Rule Them All? AUC for Classification Models

30:42
 
공유
 

Manage episode 407983676 series 3338095
Elder Research에서 제공하는 콘텐츠입니다. 에피소드, 그래픽, 팟캐스트 설명을 포함한 모든 팟캐스트 콘텐츠는 Elder Research 또는 해당 팟캐스트 플랫폼 파트너가 직접 업로드하고 제공합니다. 누군가가 귀하의 허락 없이 귀하의 저작물을 사용하고 있다고 생각되는 경우 여기에 설명된 절차를 따르실 수 있습니다 https://ko.player.fm/legal.

When considering evaluation metrics for classification models, is it possible for one metric to rule them all? Join us for a lively debate between Aric LaBarr, Associate Professor of Analytics at NC State's Institute for Advanced Analytics, and Robert Robison, Elder Research Senior Data Scientist.

During the debate Robert champions AUC’s comprehensive measure of model performance, while Aric advocates for a broader perspective, emphasizing the importance of business context in metric selection. Tune in as host Evan Wimpey moderates the discussion, and gain valuable insight on what really matters when it comes to machine learning model evaluation. We hope you enjoy the conversation!

In this episode you will learn:

⛛ The importance of exploring various metrics to evaluate model performance

⛛ Why metrics should align with business objectives

⛛ The need for data science teams to invest time in feature engineering

⛛ Why a model's success relies not only on its performance but also on stakeholders' ability to understand and trust the insights it provides
Quotes
💬
“There's a difference between communicating the value of a model and distinguishing between which models are better.” –Robert Robison
💬 “If you can't explain the model to your stakeholders or business users, then it's not going to get implemented.” –Aric LaBarr
Featured in This Episode
Aric LaBarr | Associate Professor of Analytics, Institute for Advanced Analytics

LinkedIn: https://www.linkedin.com/in/ariclabarr/
Robert Robison | Senior Data Scientist, Elder Research
LinkedIn: https://www.linkedin.com/in/robert-robison/
Evan Wimpey, Director of Analytics Strategy, Elder Research
LinkedIn: linkedin.com/in/evan-wimpey
Chapters
00:00
Evan introduces the debate topic and guests, Aric LaBarr and Robert Robison.
01:37 Robert begins his argument by defining AUC (Area Under the Curve) and its significance as a metric for classification models.
06:11 Aric begins his rebuttal, challenging the notion that AUC is the only metric to consider.
09:26 Robert provides a rebuttal to Aric's points.
11:48 Aric starts his rebuttal, focusing on communicating models to business users.
14:41 Robert responds to Aric’s points.
16:18 Evan asks Robert if certain cases may require metrics other than AUC.
17:03 Robert responds to Evan’s question.
17:53 Aric weighs in on the question.
19:37 Evan asks Aric if focusing solely on AUC may save time and costs.
20:30 Aric responds to Evan’s question.
22:20 Evan gives time for the debaters to ask each other questions.
25:30 The debaters share closing remarks, summarizing their positions.
29:14 Evan wraps up the show.

Find more show notes, transcripts, & more episodes at:
https://www.elderresearch.com/resource/podcasts/

  continue reading

34 에피소드

Artwork
icon공유
 
Manage episode 407983676 series 3338095
Elder Research에서 제공하는 콘텐츠입니다. 에피소드, 그래픽, 팟캐스트 설명을 포함한 모든 팟캐스트 콘텐츠는 Elder Research 또는 해당 팟캐스트 플랫폼 파트너가 직접 업로드하고 제공합니다. 누군가가 귀하의 허락 없이 귀하의 저작물을 사용하고 있다고 생각되는 경우 여기에 설명된 절차를 따르실 수 있습니다 https://ko.player.fm/legal.

When considering evaluation metrics for classification models, is it possible for one metric to rule them all? Join us for a lively debate between Aric LaBarr, Associate Professor of Analytics at NC State's Institute for Advanced Analytics, and Robert Robison, Elder Research Senior Data Scientist.

During the debate Robert champions AUC’s comprehensive measure of model performance, while Aric advocates for a broader perspective, emphasizing the importance of business context in metric selection. Tune in as host Evan Wimpey moderates the discussion, and gain valuable insight on what really matters when it comes to machine learning model evaluation. We hope you enjoy the conversation!

In this episode you will learn:

⛛ The importance of exploring various metrics to evaluate model performance

⛛ Why metrics should align with business objectives

⛛ The need for data science teams to invest time in feature engineering

⛛ Why a model's success relies not only on its performance but also on stakeholders' ability to understand and trust the insights it provides
Quotes
💬
“There's a difference between communicating the value of a model and distinguishing between which models are better.” –Robert Robison
💬 “If you can't explain the model to your stakeholders or business users, then it's not going to get implemented.” –Aric LaBarr
Featured in This Episode
Aric LaBarr | Associate Professor of Analytics, Institute for Advanced Analytics

LinkedIn: https://www.linkedin.com/in/ariclabarr/
Robert Robison | Senior Data Scientist, Elder Research
LinkedIn: https://www.linkedin.com/in/robert-robison/
Evan Wimpey, Director of Analytics Strategy, Elder Research
LinkedIn: linkedin.com/in/evan-wimpey
Chapters
00:00
Evan introduces the debate topic and guests, Aric LaBarr and Robert Robison.
01:37 Robert begins his argument by defining AUC (Area Under the Curve) and its significance as a metric for classification models.
06:11 Aric begins his rebuttal, challenging the notion that AUC is the only metric to consider.
09:26 Robert provides a rebuttal to Aric's points.
11:48 Aric starts his rebuttal, focusing on communicating models to business users.
14:41 Robert responds to Aric’s points.
16:18 Evan asks Robert if certain cases may require metrics other than AUC.
17:03 Robert responds to Evan’s question.
17:53 Aric weighs in on the question.
19:37 Evan asks Aric if focusing solely on AUC may save time and costs.
20:30 Aric responds to Evan’s question.
22:20 Evan gives time for the debaters to ask each other questions.
25:30 The debaters share closing remarks, summarizing their positions.
29:14 Evan wraps up the show.

Find more show notes, transcripts, & more episodes at:
https://www.elderresearch.com/resource/podcasts/

  continue reading

34 에피소드

모든 에피소드

×
 
Loading …

플레이어 FM에 오신것을 환영합니다!

플레이어 FM은 웹에서 고품질 팟캐스트를 검색하여 지금 바로 즐길 수 있도록 합니다. 최고의 팟캐스트 앱이며 Android, iPhone 및 웹에서도 작동합니다. 장치 간 구독 동기화를 위해 가입하세요.

 

빠른 참조 가이드