Artwork

Everyday AI에서 제공하는 콘텐츠입니다. 에피소드, 그래픽, 팟캐스트 설명을 포함한 모든 팟캐스트 콘텐츠는 Everyday AI 또는 해당 팟캐스트 플랫폼 파트너가 직접 업로드하고 제공합니다. 누군가가 귀하의 허락 없이 귀하의 저작물을 사용하고 있다고 생각되는 경우 여기에 설명된 절차를 따르실 수 있습니다 https://ko.player.fm/legal.
Player FM -팟 캐스트 앱
Player FM 앱으로 오프라인으로 전환하세요!

EP 628: What’s the best LLM for your team? 7 Steps to evaluate and create ROI for AI

38:52
 
공유
 

Manage episode 512576138 series 3470198
Everyday AI에서 제공하는 콘텐츠입니다. 에피소드, 그래픽, 팟캐스트 설명을 포함한 모든 팟캐스트 콘텐츠는 Everyday AI 또는 해당 팟캐스트 플랫폼 파트너가 직접 업로드하고 제공합니다. 누군가가 귀하의 허락 없이 귀하의 저작물을 사용하고 있다고 생각되는 경우 여기에 설명된 절차를 따르실 수 있습니다 https://ko.player.fm/legal.

How can you measure ROI on GenAI for your team? 🤔

Internal evaluations and intentionality.

We've helped thousands of orgs put LLMs to work and ACTUALLY save time. On today's show, we're dishing the 7 steps you need to follow.

What’s the best LLM for your team? 7 Steps to evaluate and create ROI for AI -- An Everyday AI chat with Jordan Wilson

Newsletter: Sign up for our free daily newsletter
More on this Episode: Episode Page
Join the discussion on LinkedIn: Thoughts on this? Join the convo on LinkedIn and connect with other AI leaders.
Upcoming Episodes: Check out the upcoming Everyday AI Livestream lineup
Website: YourEverydayAI.com
Email The Show: [email protected]
Connect with Jordan on LinkedIn
Topics Covered in This Episode:

  1. Choosing the Right Large Language Model
  2. Evaluating LLMs for Business ROI
  3. Front-End AI Operating Systems Explained
  4. Common Traps in AI Model Evaluation
  5. Public Benchmarks for LLM Evaluation
  6. Seven-Step LLM Evaluation Framework
  7. Measuring Pre-GenAI Human Baselines
  8. Building Realistic AI Test Datasets
  9. Calculating ROI for GenAI Implementation
  10. Monthly Retesting and AI Model Updates

Timestamps:
00:00 Choosing the Right AI Model

07:02 Adapting Workflows for AI Integration

10:58 "Gemini's Versatile Modes Overview"

14:30 Avoiding AI Shiny Object Syndrome

15:36 AI Evaluation for Reliability and Improvement

20:36 "Data Testing Guide Essentials"

25:15 Realistic and Messy Data Essentials

26:06 "Building Effective AI Workspaces"

31:08 AI Evaluation and ROI Calculation

34:11 Human Oversight in AI Testing

35:52 Evaluating GenAI Use Cases

39:00 "NotebookLM: AI-Powered Idea Organizer"

Keywords:
Large Language Model, LLM, generative AI, AI operating system, front end AI models, AI evaluation, model ROI, model evaluation steps, AI benchmarks, scientific benchmarks, API connection, enterprise AI, ChatGPT, Claude, Gemini, Copilot, team AI adoption, knowledge worker AI, operating system choice, productivity modes, connectors, deep research mode, agent mode, image generation, web search, Canvas mode, advanced voice mode, business process automation, workflow evaluation, change management, AI training,

Send Everyday AI and Jordan a text message. (We can't reply back unless you leave contact info)

  continue reading

633 에피소드

Artwork
icon공유
 
Manage episode 512576138 series 3470198
Everyday AI에서 제공하는 콘텐츠입니다. 에피소드, 그래픽, 팟캐스트 설명을 포함한 모든 팟캐스트 콘텐츠는 Everyday AI 또는 해당 팟캐스트 플랫폼 파트너가 직접 업로드하고 제공합니다. 누군가가 귀하의 허락 없이 귀하의 저작물을 사용하고 있다고 생각되는 경우 여기에 설명된 절차를 따르실 수 있습니다 https://ko.player.fm/legal.

How can you measure ROI on GenAI for your team? 🤔

Internal evaluations and intentionality.

We've helped thousands of orgs put LLMs to work and ACTUALLY save time. On today's show, we're dishing the 7 steps you need to follow.

What’s the best LLM for your team? 7 Steps to evaluate and create ROI for AI -- An Everyday AI chat with Jordan Wilson

Newsletter: Sign up for our free daily newsletter
More on this Episode: Episode Page
Join the discussion on LinkedIn: Thoughts on this? Join the convo on LinkedIn and connect with other AI leaders.
Upcoming Episodes: Check out the upcoming Everyday AI Livestream lineup
Website: YourEverydayAI.com
Email The Show: [email protected]
Connect with Jordan on LinkedIn
Topics Covered in This Episode:

  1. Choosing the Right Large Language Model
  2. Evaluating LLMs for Business ROI
  3. Front-End AI Operating Systems Explained
  4. Common Traps in AI Model Evaluation
  5. Public Benchmarks for LLM Evaluation
  6. Seven-Step LLM Evaluation Framework
  7. Measuring Pre-GenAI Human Baselines
  8. Building Realistic AI Test Datasets
  9. Calculating ROI for GenAI Implementation
  10. Monthly Retesting and AI Model Updates

Timestamps:
00:00 Choosing the Right AI Model

07:02 Adapting Workflows for AI Integration

10:58 "Gemini's Versatile Modes Overview"

14:30 Avoiding AI Shiny Object Syndrome

15:36 AI Evaluation for Reliability and Improvement

20:36 "Data Testing Guide Essentials"

25:15 Realistic and Messy Data Essentials

26:06 "Building Effective AI Workspaces"

31:08 AI Evaluation and ROI Calculation

34:11 Human Oversight in AI Testing

35:52 Evaluating GenAI Use Cases

39:00 "NotebookLM: AI-Powered Idea Organizer"

Keywords:
Large Language Model, LLM, generative AI, AI operating system, front end AI models, AI evaluation, model ROI, model evaluation steps, AI benchmarks, scientific benchmarks, API connection, enterprise AI, ChatGPT, Claude, Gemini, Copilot, team AI adoption, knowledge worker AI, operating system choice, productivity modes, connectors, deep research mode, agent mode, image generation, web search, Canvas mode, advanced voice mode, business process automation, workflow evaluation, change management, AI training,

Send Everyday AI and Jordan a text message. (We can't reply back unless you leave contact info)

  continue reading

633 에피소드

Semua episode

×
 
Loading …

플레이어 FM에 오신것을 환영합니다!

플레이어 FM은 웹에서 고품질 팟캐스트를 검색하여 지금 바로 즐길 수 있도록 합니다. 최고의 팟캐스트 앱이며 Android, iPhone 및 웹에서도 작동합니다. 장치 간 구독 동기화를 위해 가입하세요.

 

빠른 참조 가이드

탐색하는 동안 이 프로그램을 들어보세요.
재생