𝒴-Tuning: An Efficient Tuning Paradigm for Large-Scale Pre-Trained Models via Label Representation Learning
en-GBde-DEes-ESfr-FR

𝒴-Tuning: An Efficient Tuning Paradigm for Large-Scale Pre-Trained Models via Label Representation Learning

30.09.2024 Frontiers Journals

Large-scale pre-trained models (PTMs) have been extensively utilized as backbone models for numerous natural language processing downstream tasks. Recently, various lightweight-tuning paradigms have emerged and achieved comparable performance with fine-tuning in a more parameter-efficient manner. Nonetheless, these works still necessitate the computation and storage of gradients, leading to high training expenses.
To address these issues, a research team led by Xipeng Qiu from Fudan University published their latest findings on 15 August 2024 in Frontiers of Computer Science co-published by Higher Education Press and Springer Nature.

𝒴-Tuning, for large-scale pre-trained language models. The tuning paradigm learns dense representations for labels 𝒴 defined in a given task and aligns them to fixed feature representation generated frozen pre-trained models. By avoiding the computation of text encoder’s gradients at training phrase, 𝒴-Tuning is not only parameter-efficient but also training-efficient.Illustration of 𝒴-Tuning and other tuning paradigms.
𝒴-Tuning achieves close performance with full fine-tuning while significantly increasing training speed. Experimental results demonstrate that for DeBERTaXXL which has 1.6 billion parameters, 𝒴-Tuning achieves performance over 96% of full fine-tuning performance on GLUE Benchmark with only 2% tunable parameters and substantially reduces training costs. Furthermore, 𝒴-Tuning exhibits better model-robustness than baselines, as the label matching mechanism employed in 𝒴-Tuning is less sensitive to feature perturbation.

Future research could explore enhancing tuning performance by incorporating label prior information and utilizing reparameterization techniques to accelerate inference. This would enable a more comprehensive understanding and application of the 𝒴-Tuning paradigm in various natural language processing tasks.
DOI: 10.1007/s11704-023-3131-8
Angehängte Dokumente
  • Illustration of 𝒴-Tuning and other tuning paradigms.
  • Training speed-up ratio over fine-tuning on various methods
30.09.2024 Frontiers Journals
Regions: Asia, China
Keywords: Applied science, Computing

Disclaimer: AlphaGalileo is not responsible for the accuracy of news releases posted to AlphaGalileo by contributing institutions or for the use of any information through the AlphaGalileo system.

Referenzen

We have used AlphaGalileo since its foundation but frankly we need it more than ever now to ensure our research news is heard across Europe, Asia and North America. As one of the UK’s leading research universities we want to continue to work with other outstanding researchers in Europe. AlphaGalileo helps us to continue to bring our research story to them and the rest of the world.
Peter Dunn, Director of Press and Media Relations at the University of Warwick
AlphaGalileo has helped us more than double our reach at SciDev.Net. The service has enabled our journalists around the world to reach the mainstream media with articles about the impact of science on people in low- and middle-income countries, leading to big increases in the number of SciDev.Net articles that have been republished.
Ben Deighton, SciDevNet
AlphaGalileo is a great source of global research news. I use it regularly.
Robert Lee Hotz, LA Times

Wir arbeiten eng zusammen mit...


  • BBC
  • The Times
  • National Geographic
  • The University of Edinburgh
  • University of Cambridge
  • iesResearch
Copyright 2024 by DNN Corp Terms Of Use Privacy Statement