Multi-label text classification (MLTC) assigns multiple relevant labels to a text. While deep learning models have achieved cutting-edge results in this area, they require large amounts of labeled data, which is expensive and time-consuming. Active learning helps optimize this process by selecting the most informative unlabeled samples for annotation, reducing labeling effort. However, most existing active learning methods are designed for traditional single-label models and do not directly apply to deep multi-label models. Given the complexity of multi-label tasks and the high cost of annotations, there is a need for active learning techniques tailored to deep multi-label classification.
Active learning allows a model to request labels for more informative unlabeled samples, reducing annotation costs. Common active learning approaches include membership query synthesis, stream-based selective sampling, and group-based sampling, focusing on the latter in this work. Uncertainty-based sampling is often used in multi-label classification, but challenges still need to be resolved when applying active learning to multi-label deep models. While Bayesian deep learning methods have shown promise for uncertainty estimation, most research has focused on single-label tasks.
Researchers from the Institute of Automation, Chinese Academy of Sciences and other institutions propose BEAL, a deep active learning method for MLTC. BEAL uses Bayesian deep learning with dropout to infer the posterior predictive distribution of the model and introduces a new confidence-based expected acquisition function to select uncertain samples. Experiments with a BERT-based MLTC model on benchmark datasets such as AAPD and StackOverflow show that BEAL improves training efficiency and achieves convergence with fewer labeled samples. This method can be extended to other multi-label classification tasks and significantly reduces labeled data requirements compared to existing methods.
The methodology introduces a batch-mode active learning framework for deep multi-label text classification. Starting with a small set of labeled data, the framework iteratively selects unlabeled samples for annotation based on an acquisition function. This function chooses samples with the lowest expected confidence, as measured by the model's predictive uncertainty. Bayesian deep learning calculates the posterior predictive distribution using Monte Carlo dropout, approximating model confidence. The acquisition function selects a batch of samples with the lowest expected confidence for labeling, which improves model efficiency by reducing the need for labeled data. The process continues until the model performance converges.
In this study, the authors evaluate the BEAL method for deep multi-label text classification using two benchmark datasets: AAPD and StackOverflow. The process is compared to several active learning strategies, including random sampling, BADGE, BALD, Core-Set, and the full data approach. BEAL outperforms these methods by selecting the most informative samples based on the posterior predictive distribution, reducing the need for labeled data. The results show that BEAL achieves the highest performance with fewer labeled samples than others, requiring only 64% of the labeled samples in AAPD and 40% in StackOverflow. An ablation study highlights the advantage of using Bayesian deep learning in BEAL.
In conclusion, the study presents BEAL, an active learning method for deep MLTC models. BEAL uses Bayesian deep learning to infer the posterior predictive distribution and defines a confidence-based expected acquisition function to select uncertain samples for training. Experimental results show that BEAL outperforms other active learning methods, allowing for more efficient model training with fewer labeled samples. This is valuable in real-world applications where it is difficult to obtain labeled data at a large scale. Future work will explore the integration of diversity-based methods to further reduce the labeled data required for effective training of MLTC models.
look at the Paper. All credit for this research goes to the researchers of this project. Also, don't forget to follow us on <a target="_blank" href="https://twitter.com/Marktechpost”>twitter and join our Telegram channel and LinkedIn Grabove. If you like our work, you will love our information sheet.. Don't forget to join our SubReddit over 55,000ml.
(<a target="_blank" href="https://landing.deepset.ai/webinar-implementing-idp-with-genai-in-financial-services?utm_campaign=2411%20-%20webinar%20-%20credX%20-%20IDP%20with%20GenAI%20in%20Financial%20Services&utm_source=marktechpost&utm_medium=newsletter” target=”_blank” rel=”noreferrer noopener”>FREE WEBINAR on ai) <a target="_blank" href="https://landing.deepset.ai/webinar-implementing-idp-with-genai-in-financial-services?utm_campaign=2411%20-%20webinar%20-%20credX%20-%20IDP%20with%20GenAI%20in%20Financial%20Services&utm_source=marktechpost&utm_medium=newsletter” target=”_blank” rel=”noreferrer noopener”>Implementation of intelligent document processing with GenAI in financial services and real estate transactions– <a target="_blank" href="https://landing.deepset.ai/webinar-implementing-idp-with-genai-in-financial-services?utm_campaign=2411%20-%20webinar%20-%20credX%20-%20IDP%20with%20GenAI%20in%20Financial%20Services&utm_source=marktechpost&utm_medium=banner-ad-desktop” target=”_blank” rel=”noreferrer noopener”>From framework to production
Sana Hassan, a consulting intern at Marktechpost and a dual degree student at IIT Madras, is passionate about applying technology and artificial intelligence to address real-world challenges. With a strong interest in solving practical problems, he brings a new perspective to the intersection of ai and real-life solutions.
<script async src="//platform.twitter.com/widgets.js” charset=”utf-8″>