Monday, December 4, 2023
No Result
View All Result
AI CRYPTO BUZZ
  • Home
  • Bitcoins
  • Crypto
    • Altcoins
    • Ethereum
    • Crypto Exchanges
  • NFT
  • Blockchain
  • AI
  • ML
  • Cyber Security
  • Web3
  • Metaverse
  • DeFi
  • Analysis
Marketcap
  • Home
  • Bitcoins
  • Crypto
    • Altcoins
    • Ethereum
    • Crypto Exchanges
  • NFT
  • Blockchain
  • AI
  • ML
  • Cyber Security
  • Web3
  • Metaverse
  • DeFi
  • Analysis
Marketcap
No Result
View All Result
AI CRYPTO BUZZ
No Result
View All Result

Meet LoftQ: LoRA-Fine-Tuning-Aware Quantization for Large Language Models

October 25, 2023
in Artificial Intelligence
Reading Time: 4 mins read
0 0
A A
0
Home Artificial Intelligence
Share on FacebookShare on Twitter


The introduction of Pre-trained Language Fashions (PLMs) has signified a transformative shift within the discipline of Pure Language Processing. They’ve demonstrated distinctive proficiency in performing a variety of language duties, together with Pure Language Understanding (NLU) and Pure Language Technology (NLG). These fashions usually incorporate tens of millions and even billions of parameters, resulting in substantial computational and reminiscence necessities. Nonetheless, the appreciable computational and reminiscence wants of those fashions current vital challenges, as acknowledged by the analysis group.

On this paper, the authors introduce a novel quantization framework referred to as LoRA-Superb-Tuning-aware Quantization (LoftQ). This framework is particularly tailor-made for pre-trained fashions that necessitate quantization and LoRA fine-tuning. The framework actively combines low-rank approximation, working at the side of quantization to collectively approximate the unique high-precision pre-trained weights.

The above picture demonstrates QLoRA efficiency with completely different bits. Left: QLoRA initialization of LLAMA-2-13b on WikiText-2. Proper: Apply QLoRA to LLAMA-2-13b on WikiText-2 language modelling process. Smaller perplexity signifies higher efficiency. 

Quantization Strategies. We apply two quantization strategies to display LoftQ is appropriate with completely different quantization features:

• Uniform quantization is a traditional quantization methodology. It uniformly divides a steady interval into 2N classes and shops an area most absolute worth for dequantization.

• NF4 and its 2-bit variant NF2 are quantization strategies utilized in QLoRA. They assume that the high-precision values are drawn from a Gaussian distribution and map these values to discrete slots which have equal likelihood.

We carry out 2-bit and 4-bit quantization on all fashions, attaining compression ratios of 25-30% and 15-20% on the 4-bit and 2-bit ranges, respectively. All of the experiments are performed on NVIDIA A100 GPUs.

The analysis of their quantization framework is carried out by means of in depth experiments on varied downstream duties, together with NLU, query answering, summarization, and NLG. The outcomes of those experiments display that LoftQ persistently surpasses QLoRA throughout all precision ranges. For instance, with 4-bit quantization, they attain a 1.1 and 0.8 enchancment in Rouge-1 for XSum and CNN/DailyMail, respectively. As the sphere of NLP continues to advance, it’s anticipated that additional improvements and optimizations will assist bridge the hole between the immense potential of PLMs and their sensible deployment, benefiting a variety of functions and customers.

Take a look at the Paper. All Credit score For This Analysis Goes To the Researchers on This Challenge. Additionally, don’t overlook to hitch our 31k+ ML SubReddit, 40k+ Fb Neighborhood, Discord Channel, and Electronic mail Publication, the place we share the newest AI analysis information, cool AI tasks, and extra.

When you like our work, you’ll love our e-newsletter..

We’re additionally on WhatsApp. Be a part of our AI Channel on Whatsapp..

Janhavi Lande, is an Engineering Physics graduate from IIT Guwahati, class of 2023. She is an upcoming knowledge scientist and has been working on this planet of ml/ai analysis for the previous two years. She is most fascinated by this ever altering world and its fixed demand of people to maintain up with it. In her pastime she enjoys touring, studying and writing poems.

▶️ Now Watch AI Analysis Updates On Our Youtube Channel [Watch Now]



Source link

Tags: languageLargeLoftQLoRAFineTuningAwareMeetmodelsQuantization
Previous Post

How to Optimize Revenues Using Dynamic Pricing?

Next Post

Bitcoin Bounce Above $35,000 Puts Holder Profitability At Yearly Highs

Related Posts

Text analytics: A recipe for food safety success
Artificial Intelligence

Text analytics: A recipe for food safety success

December 4, 2023
CMU Researchers Discover Key Insights into Neural Network Behavior: The Interplay of Heavy-Tailed Data and Network Depth in Shaping Optimization Dynamics
Artificial Intelligence

CMU Researchers Discover Key Insights into Neural Network Behavior: The Interplay of Heavy-Tailed Data and Network Depth in Shaping Optimization Dynamics

December 4, 2023
Courage to Learn ML: Decoding Likelihood, MLE, and MAP | by Amy Ma | Dec, 2023
Artificial Intelligence

Courage to Learn ML: Decoding Likelihood, MLE, and MAP | by Amy Ma | Dec, 2023

December 4, 2023
Microsoft Researchers Propose MAIRA-1: A Radiology-Specific Multimodal Model for the Task of Generating Radiological Reports from Chest X-rays (CXRs)
Artificial Intelligence

Microsoft Researchers Propose MAIRA-1: A Radiology-Specific Multimodal Model for the Task of Generating Radiological Reports from Chest X-rays (CXRs)

December 3, 2023
Implementing Soft Nearest Neighbor Loss in PyTorch | by Abien Fred Agarap | Nov, 2023
Artificial Intelligence

Implementing Soft Nearest Neighbor Loss in PyTorch | by Abien Fred Agarap | Nov, 2023

December 2, 2023
Introduction to Mathematical Optimisation in Python | by Zolzaya Luvsandorj | Dec, 2023
Artificial Intelligence

Introduction to Mathematical Optimisation in Python | by Zolzaya Luvsandorj | Dec, 2023

December 3, 2023
Next Post
Bitcoin Bounce Above $35,000 Puts Holder Profitability At Yearly Highs

Bitcoin Bounce Above $35,000 Puts Holder Profitability At Yearly Highs

Microsoft Shuts Down Industrial Metaverse AI Projects

Microsoft Shuts Down Industrial Metaverse AI Projects

AWS FedRAMP Revision 5 baselines transition update

AWS FedRAMP Revision 5 baselines transition update

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Facebook Twitter Instagram Youtube RSS
AI CRYPTO BUZZ

The latest news and updates about the Cryptocurrency and AI Technology around the world... The AI Crypto Buzz keeps you in the loop.

CATEGORIES

  • Altcoins
  • Analysis
  • Artificial Intelligence
  • Bitcoins
  • Blockchain
  • Crypto Exchanges
  • Cyber Security
  • DeFi
  • Ethereum
  • Machine Learning
  • Metaverse
  • NFT
  • Web3

SITE MAP

  • Disclaimer
  • Privacy Policy
  • DMCA
  • Cookie Privacy Policy
  • Terms and Conditions
  • Contact us

Copyright © 2023 AI Crypto Buzz.
AI Crypto Buzz is not responsible for the content of external sites.

No Result
View All Result
  • Home
  • Bitcoins
  • Crypto
    • Altcoins
    • Ethereum
    • Crypto Exchanges
  • NFT
  • Blockchain
  • AI
  • ML
  • Cyber Security
  • Web3
  • Metaverse
  • DeFi
  • Analysis

Copyright © 2023 AI Crypto Buzz.
AI Crypto Buzz is not responsible for the content of external sites.

Welcome Back!

Login to your account below

Forgotten Password?

Retrieve your password

Please enter your username or email address to reset your password.

Log In