Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
×
Dec 11, 2023 · In this paper, we empirically study the optimization dynamics of multi-task learning, particularly focusing on those that govern a collection ...
In this paper, we introduce a simple training scheme that combines the best of static sampling and transfer learning: pre-train on a high-resource task and fine ...
This paper investigates the strategies on how to best deal with data imbalance in multilingual learning. Their proposed method enables high-resource tasks to ...
Dec 11, 2023 · We show that in the presence of large data imbalance, the order at which tasks are introduced has significant impact on overall performance. We ...
This paper introduces a new method called pre-train joint fine-tuning for multi-task learning, especially for tasks with imbalanced data.
Dec 11, 2023 · We present a simple yet effective method of pre-training on high-resource tasks, followed by fine-tuning on a mixture of high/low-resource tasks ...
Order Matters in the Presence of Dataset Imbalance for Multilingual Learning ... learning training system using data echoing. D Choi, AT Passos, CJ Shallue ...
A Loss Curvature Perspective on Training Instability in Deep Learning. J ... Order Matters in the Presence of Dataset Imbalance for Multilingual Learning.
Unlock the magic of AI with handpicked models, awesome datasets, papers, and mind-blowing Spaces from stereoplegic.