
Player FM 앱으로 오프라인으로 전환하세요!
SPOTLIGHT. Unified Flash Memory and Reduced HBM are Reshaping AI Training and Inference with Phison
Manage episode 506695477 series 2434477
AI will need less HBM (high bandwidth memory) because flash memory unification is changing training and inference. This episode of the Tech Field Day podcast features Sebastien Jean from Phison, Max Mortillaro, Brian Martin, and Alastair Cooke. Training, fine-tuning, and inference with Large Language Models traditionally use GPUs with high bandwidth memory to hold entire data models and data sets. Phison’s aiDaptiv+ framework offers the ability to trade lower cost of infrastructure against training speed or allow larger data sets (context) for inference. This approach enables users to balance cost, compute, and memory needs, making larger models accessible without requiring top-of-the-line GPUs, and giving smaller companies more access to generative AI.
Learn more about Phison's solutions here.
Phsion Representative: Sebastien Jean, CTO of Phison Electronics
Host
Alastair Cooke, Tech Field Day Event Lead
Panelists
Brian Martin, VP of AI and Datacenter Performance at Signal65
Max Mortillaro, Chief Research Officer at Osmium Group
Follow the Tech Field Day Podcast on X/Twitter or on Bluesky and use the Hashtag #TFDPodcast to join the discussion. Listen to more episodes on the podcast page of the website.
Follow Tech Field Day for more information on upcoming and current event coverage on X/Twitter, on Bluesky, and on LinkedIn, or visit our website.
328 에피소드
Manage episode 506695477 series 2434477
AI will need less HBM (high bandwidth memory) because flash memory unification is changing training and inference. This episode of the Tech Field Day podcast features Sebastien Jean from Phison, Max Mortillaro, Brian Martin, and Alastair Cooke. Training, fine-tuning, and inference with Large Language Models traditionally use GPUs with high bandwidth memory to hold entire data models and data sets. Phison’s aiDaptiv+ framework offers the ability to trade lower cost of infrastructure against training speed or allow larger data sets (context) for inference. This approach enables users to balance cost, compute, and memory needs, making larger models accessible without requiring top-of-the-line GPUs, and giving smaller companies more access to generative AI.
Learn more about Phison's solutions here.
Phsion Representative: Sebastien Jean, CTO of Phison Electronics
Host
Alastair Cooke, Tech Field Day Event Lead
Panelists
Brian Martin, VP of AI and Datacenter Performance at Signal65
Max Mortillaro, Chief Research Officer at Osmium Group
Follow the Tech Field Day Podcast on X/Twitter or on Bluesky and use the Hashtag #TFDPodcast to join the discussion. Listen to more episodes on the podcast page of the website.
Follow Tech Field Day for more information on upcoming and current event coverage on X/Twitter, on Bluesky, and on LinkedIn, or visit our website.
328 에피소드
All episodes
×플레이어 FM에 오신것을 환영합니다!
플레이어 FM은 웹에서 고품질 팟캐스트를 검색하여 지금 바로 즐길 수 있도록 합니다. 최고의 팟캐스트 앱이며 Android, iPhone 및 웹에서도 작동합니다. 장치 간 구독 동기화를 위해 가입하세요.