July 18th, 2023 - How about AI Evolution: The Dawn of Supercharged Transformers, Retentive Networks and Visionary Language Models
Fetch error
Hmmm there seems to be a problem fetching this series right now. Last successful fetch was on September 26, 2024 07:56 ()
What now? This series will be checked again in the next day. If you believe it should be working, please verify the publisher's feed link below is valid and includes actual episode links. You can contact support to request the feed be immediately fetched.
Manage episode 371591074 series 3485608
- FlashAttention-2: Faster Attention with Better Parallelism and Work Partitioning
- Retentive Network: A Successor to Transformer for Large Language Models
- AlpaGasus: Training A Better Alpaca with Fewer Data
- CoTracker: It is Better to Track Together
- Diffusion Models Beat GANs on Image Classification
- Planting a SEED of Vision in Large Language Model
챕터
1. Intro (00:00:00)
2. FlashAttention-2: Faster Attention with Better Parallelism and Work Partitioning (00:01:55)
3. Retentive Network: A Successor to Transformer for Large Language Models (00:04:37)
4. AlpaGasus: Training A Better Alpaca with Fewer Data (00:07:39)
5. CoTracker: It is Better to Track Together (00:09:00)
6. Diffusion Models Beat GANs on Image Classification (00:10:14)
7. Planting a SEED of Vision in Large Language Model (00:12:57)
75 에피소드