This AI paper from Google DeepMind introduces enhanced learning capabilities with many-shot in-context learning

Machine Learning


https://arxiv.org/abs/2404.11018

In-context learning (ICL) of large-scale language models (LLMs) leverages input and output samples to adapt to new tasks without changing the underlying model architecture. This method has transformed the way models handle different tasks by learning from first-hand examples provided during inference. The current problem is the limitations of several-shot ICLs in handling complex tasks. These tasks often require deep understanding that few-shot learning cannot provide, as they operate under the constraints of minimal input data. This scenario may be suitable for applications that require detailed analysis and decision-making based on extensive datasets, such as advanced inference and language translation.

Existing research in the ICL field has mainly focused on the few-shot learning capabilities of models like GPT-3 that adapt to new tasks with a limited set of examples. Research has investigated the performance limits of these models within small context windows, highlighting task complexity and scalability constraints. The development of models with larger context windows, such as Gemini 1.5 Pro, which supports up to 1 million tokens, represents a major evolution. This enhancement enables multi-shot ICL exploration and greatly enhances the model's ability to process and learn from larger datasets.

Researchers at Google Deepmind leveraged the larger context window of models such as the Gemini 1.5 Pro to introduce the move to multi-shot ICLs. This transition from few-shot to many-shot learning takes advantage of the increased number of input examples and significantly improves model performance and adaptability across complex tasks. What is unique about this methodology is the integration of enhanced and unsupervised ICL, reducing reliance on human-generated content by using only model-generated data and domain-specific inputs.

In terms of methodology, the Gemini 1.5 Pro model was adopted to handle an expanded input/output sample, supporting up to 1 million tokens in the context window. This has enabled the exploration of enhanced ICL, where the model generates and evaluates evidence for correctness, and unsupervised ICL, which requires the model to operate without explicit evidence. Experiments were conducted across a variety of domains including machine translation, summarization, and complex reasoning tasks, using datasets such as MATH for mathematical problem solving and FLORES for machine translation tasks, using a multi-shot ICL framework. Tested and verified effectiveness.

Significant performance improvements were demonstrated as a result of implementing the many-shot ICL. For machine translation tasks, the Gemini 1.5 Pro model outperformed previous benchmarks, delivering 4.5% more accurate Kurdish translations and 1.5% more accurate Tamil translations compared to previous models. For mathematical problem solving, we found that using a multi-shot setup increased the accuracy of solutions on the MATH dataset by 35%. These quantitative results validate the effectiveness of multi-shot ICL in improving model adaptability and accuracy across diverse and complex cognitive tasks.

In conclusion, this study represents an important advance in ICL by moving from a few-shot ICL to a multi-shot ICL using the Gemini 1.5 Pro model. By extending the context window and integrating innovative methodologies such as enhanced ICL and unsupervised ICL, this study successfully improves model performance across a variety of tasks, including machine translation and mathematical problem solving. Did. These advances not only improve the adaptability and efficiency of large-scale language models, but also pave the way for more sophisticated applications in AI.


Please check paper. All credit for this research goes to the researchers of this project.Don't forget to follow us twitter.Please join us telegram channel, Discord channeland linkedin groupsHmm.

If you like what we do, you'll love Newsletter..

Don't forget to join us 40,000+ ML subreddits

Nikhil is an intern consultant at Marktechpost. He is pursuing an integrated dual degree in materials from the Indian Institute of Technology, Kharagpur. Nikhil is an AI/ML enthusiast and is constantly researching applications in areas such as biomaterials and biomedicine. With a strong background in materials science, he explores new advances and creates opportunities to contribute.

🐝 Join the fastest growing AI research newsletter from researchers at Google + NVIDIA + Meta + Stanford + MIT + Microsoft and more…





Source link

Leave a Reply

Your email address will not be published. Required fields are marked *