arxiv preprint - Quiet-STaR: Language Models Can Teach Themselves to Think Before Speaking
AI Breakdown

arxiv preprint - Quiet-STaR: Language Models Can Teach Themselves to Think Before Speaking

2024-03-22
In this episode, we discuss Quiet-STaR: Language Models Can Teach Themselves to Think Before Speaking by Eric Zelikman, Georges Harik, Yijia Shao, Varuna Jayasiri, Nick Haber, Noah D. Goodman. The paper presents Quiet-STaR, an advancement of Self-Taught Reasoner (STaR), which teaches Language Models to generate internal rationales to enhance text predictions. By introducing a tokenwise parallel sampling algorithm, learnable tokens for marking thoughts, and extending teacher-forcing, the...
View more
Comments (3)

More Episodes

All Episodes>>

Get this podcast on your phone, Free

Create Your Podcast In Minutes

  • Full-featured podcast site
  • Unlimited storage and bandwidth
  • Comprehensive podcast stats
  • Distribute to Apple Podcasts, Spotify, and more
  • Make money with your podcast
Get Started
It is Free