Podbean logo
  • Discover
  • Podcast Features

    Your all-in-one podcasting solution.

    Podcast Studio

    Easy-to-use audio recorder app.

  • Livestream

    High-performing audio live, without limits.

  • Podcast App

    The best podcast player & podcast app.

  • Ads Marketplace

    Join Ads Marketplace to earn money
    through sponsorship on your podcast.

    PodAds

    Manage your ads with dynamic ad insertion capability.

  • Patron & Paid Content

    The seamless way for fans to support you directly
    from your podcast.

  •  
  • All Arts Business Comedy Education
  • Fiction Government Health & Fitness History Kids & Family
  • Leisure Music News Religion & Spirituality Science
  • Society & Culture Sports Technology True Crime TV & Film
  • Live
  • How to Start a Podcast
  • How to Start a Live Podcast
  • How to Monetize a podcast
  • How to Promote Your Podcast
  • How to Use Group Recording
  • Log in
  • Start your podcast for free
  • Podcasting
    • Podcast Features
    • Live Stream
    • PodAds
    • Podcast App
    • Podcast Studio
  • Monetization
    • Premium
    • Patron
    • Ads Marketplace
  • Enterprise
  • Pricing
  • Discover
  • Log in
    Sign up free
Papers Read on AI

Papers Read on AI

News:Tech News

LoRA: Low-Rank Adaptation of Large Language Models

LoRA: Low-Rank Adaptation of Large Language Models

2023-05-17
Download
The dominant paradigm of natural language processing consists of large-scale pre-training on general domain data and adaptation to particular tasks or domains. As we pre-train larger models, conventional fine-tuning, which retrains all model parameters, becomes less feasible. Using GPT-3 175B as an example, deploying many independent instances of fine-tuned models, each with 175B parameters, is extremely expensive. We propose Low-Rank Adaptation, or LoRA, which freezes the pre-trained model weights and injects trainable rank decomposition matrices into each layer of the Transformer architecture, greatly reducing the number of trainable parameters for downstream tasks. For GPT-3, LoRA can reduce the number of trainable parameters by 10,000 times and the computation hardware requirement by 3 times compared to full fine-tuning. LoRA performs on-par or better than fine-tuning in model quality on both GPT-3 and GPT-2, despite having fewer trainable parameters, a higher training throughput, and no additional inference latency. We also provide an empirical investigation into rank-deficiency in language model adaptations, which sheds light on the efficacy of LoRA. We release our implementation in GPT-2 at https://github.com/microsoft/LoRA. 2021: Edward J. Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Weizhu Chen https://arxiv.org/pdf/2106.09685v2.pdf
view more

More Episodes

SpQR: A Sparse-Quantized Representation for Near-Lossless LLM Weight Compression
2023-06-09 63
Orca: Progressive Learning from Complex Explanation Traces of GPT-4
2023-06-08 79
Hiera: A Hierarchical Vision Transformer without the Bells-and-Whistles
2023-06-06 62
Let’s Verify Step by Step
2023-06-05 79
Large Language Models as Tool Makers
2023-06-04 87
Gorilla: Large Language Model Connected with Massive APIs
2023-06-02 93
CodeT5+: Open Code Large Language Models for Code Understanding and Generation
2023-05-31 104
VanillaNet: the Power of Minimalism in Deep Learning
2023-05-29 104
QLoRA: Efficient Finetuning of Quantized LLMs
2023-05-26 169
SpeechGPT: Empowering Large Language Models with Intrinsic Cross-Modal Conversational Abilities
2023-05-25 113
LLM-Pruner: On the Structural Pruning of Large Language Models
2023-05-24 111
Tree of Thoughts: Deliberate Problem Solving with Large Language Models
2023-05-23 116
Training language models to follow instructions with human feedback
2023-05-19 135
Language Models Trained on Media Diets Can Predict Public Opinion
2023-05-18 88
Pretraining Without Attention
2023-05-15 113
ImageBind: One Embedding Space To Bind Them All
2023-05-12 121
ZipIt! Merging Models from Different Tasks without Training
2023-05-10 125
Chain of Thought Prompting Elicits Reasoning in Large Language Models
2023-05-09 129
CodeGen2: Lessons for Training LLMs on Programming and Natural Languages
2023-05-08 119
  • ←
  • 1
  • 2
  • 3
  • 4
  • 5
  • 6
  • 7
  • 8
  • 9
  • 10
  • →
0123456789101112131516171819

Get this podcast on your
phone, FREE

Download Podbean app on App Store Download Podbean app on Google Play

Create your
podcast in
minutes

  • Full-featured podcast site
  • Unlimited storage and bandwidth
  • Comprehensive podcast stats
  • Integrate with iTunes and Google
    store
  • Make money with your podcast
Get started

It is Free

  • Podcast Services

    • Podcast Features
    • Pricing
    • Enterprise Solution
    • Private Podcast
    • The Podcast App
    • Live Stream
    • Audio Recorder
    • Remote Recording
  •  
    • Create a Podcast
    • Video Podcast
    • Start Podcasting
    • Start Radio Talk Show
    • Education Podcast
    • Church Podcast
    • Nonprofit Podcast
    • Get Sermons Online
    • Free Audiobooks
  • MONETIZATION & MORE

    • Podcast Advertising
    • Dynamic Ads Insertion
    • Patron Program
    • Affiliate Program
    • Switch to Podbean
    • Submit Your Podcast
    • Podbean Plugins
    • Developers
    • Badges
  • KNOWLEDGE BASE

    • How to Start a Podcast
    • How to Start a Live Podcast
    • How to Monetize a podcast
    • How to Promote Your Podcast
    • How to Use Group Recording
  • Support

    • Support Center
    • Podbean Blog
    • What’s New
    • Free Webinars
    • Podcast Events
    • Podbean Academy
    • Podcasting Smarter
    • Resources
  • Podbean

    • About Us
    • Careers
    • Press and Media
    • Green Initiative
    • Contact Us
  • Privacy Policy
  • Cookie Policy
  • Terms of Use
  • Consent Preferences
  • Copyright © 2006-2023 Podbean.com