Deep Long-Tailed Learning: A Survey.

IEEE Trans Pattern Anal Mach Intell

Published: September 2023

Deep long-tailed learning, one of the most challenging problems in visual recognition, aims to train well-performing deep models from a large number of images that follow a long-tailed class distribution. In the last decade, deep learning has emerged as a powerful recognition model for learning high-quality image representations and has led to remarkable breakthroughs in generic visual recognition. However, long-tailed class imbalance, a common problem in practical visual recognition tasks, often limits the practicality of deep network based recognition models in real-world applications, since they can be easily biased towards dominant classes and perform poorly on tail classes. To address this problem, a large number of studies have been conducted in recent years, making promising progress in the field of deep long-tailed learning. Considering the rapid evolution of this field, this article aims to provide a comprehensive survey on recent advances in deep long-tailed learning. To be specific, we group existing deep long-tailed learning studies into three main categories (i.e., class re-balancing, information augmentation and module improvement), and review these methods following this taxonomy in detail. Afterward, we empirically analyze several state-of-the-art methods by evaluating to what extent they address the issue of class imbalance via a newly proposed evaluation metric, i.e., relative accuracy. We conclude the survey by highlighting important applications of deep long-tailed learning and identifying several promising directions for future research.

Download full-text PDF

Source
http://dx.doi.org/10.1109/TPAMI.2023.3268118DOI Listing

Publication Analysis

Top Keywords

deep long-tailed
24
long-tailed learning
24
visual recognition
12
deep
9
learning
8
large number
8
long-tailed class
8
class imbalance
8
long-tailed
7
recognition
5

Similar Publications

Mixed Mutual Transfer for Long-Tailed Image Classification.

Entropy (Basel)

October 2024

College of Computer Science and Technology, Harbin Engineering University, Nantong Street, Harbin 150001, China.

Real-world datasets often follow a long-tailed distribution, where a few majority (head) classes contain a large number of samples, while many minority (tail) classes contain significantly fewer samples. This imbalance creates an information disparity between head and tail classes, which can negatively impact the performance of deep networks. Some transfer knowledge techniques attempt to mitigate this gap by generating additional minority samples, either through oversampling the tail classes or transferring knowledge from the head classes to the tail classes.

View Article and Find Full Text PDF

ChatDiff: A ChatGPT-based diffusion model for long-tailed classification.

Neural Netw

January 2025

School of Technology, Beijing Forestry University, Beijing, 100083, PR China; Research Center for Biodiversity Intelligent Monitoring, Beijing Forestry University, Beijing, 100083, PR China; State Key Laboratory of Efficient Production of Forest Resources, Beijing Forestry University, Beijing, 100083, PR China. Electronic address:

Long-tailed data distributions have been a major challenge for the practical application of deep learning. Information augmentation intends to expand the long-tailed data into uniform distribution, which provides a feasible way to mitigate the data starvation of underrepresented classes. However, most existing augmentation methods face two significant challenges: (1) limited diversity in generated samples, and (2) the adverse effect of generated negative samples on downstream classification performance.

View Article and Find Full Text PDF

Real-world data often follows a long-tailed distribution, where a few head classes occupy most of the data and a large number of tail classes only contain very limited samples. In practice, deep models often show poor generalization performance on tail classes due to the imbalanced distribution. To tackle this, data augmentation has become an effective way by synthesizing new samples for tail classes.

View Article and Find Full Text PDF

When an extra rejection class meets out-of-distribution detection in long-tailed image classification.

Neural Netw

October 2024

State Key Laboratory for Novel Software Technology, Nanjing University, Nanjing, 210023, China; Department of Computer Science and Technology, Nanjing University, Nanjing, 210023, China. Electronic address:

Detecting Out-of-Distribution (OOD) inputs is essential for reliable deep learning in the open world. However, most existing OOD detection methods have been developed based on training sets that exhibit balanced class distributions, making them susceptible when confronted with training sets following a long-tailed distribution. To alleviate this problem, we propose an effective three-branch training framework, which demonstrates the efficacy of incorporating an extra rejection class along with auxiliary outlier training data for effective OOD detection in long-tailed image classification.

View Article and Find Full Text PDF
Article Synopsis
  • Researchers are studying how viruses infect bats in Puerto Rico to learn more about viral epidemics and improve public health.
  • They collected samples from 1086 bats and found a lot of different herpesviruses, which they categorized into groups for further study.
  • The results showed that most herpesviruses prefer to infect just one type of bat rather than spreading to many different species, and they explained how this can help us understand virus communities better.
View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!