TY - GEN
T1 - Exploring Large Language Models for Low-Resource IT Information Extraction
AU - Bhavya, Bhavya
AU - Isaza, Paulina Toro
AU - Deng, Yu
AU - Nidd, Michael
AU - Azad, Amar Prakash
AU - Shwartz, Larisa
AU - Zhai, Cheng Xiang
N1 - Publisher Copyright:
© 2023 IEEE.
PY - 2023
Y1 - 2023
N2 - Information Extraction (IE) in IT is an important foundational task that is needed for many AIOps applications. A major challenge of IE in IT is that we often do not have sufficient labelled data for training machine learning algorithms since acquiring labels is labor-intensive and costly. In this paper, we propose to leverage Large Language Models (LLMs) to address this challenge of low resources and study two data augmentation strategies, i.e., using LLMs to generate pseudo labels and generate synthetic data. We use multiple IE tasks and datasets, including a new Semantic Troubleshooting-Segment Extraction task and Named Entity Recognition, to evaluate the benefits of LLMs. Our experiment results suggest that data augmentation using LLMs, specifically, using SeqMix model that combines active labeling with synthetic data samples generated in the embedding-vector space, is a promising approach for IT domain IE. Our study also shows that although data augmentation and direct labeling with the state-of-the-art, ChatGPT model achieves a high performance on general domain IE, there is a need to adapt it for IE from IT text data. Moreover, our initial exploration of two label weighting and selection strategies (confidence and consistency-based) suggests that they could be used to improve data augmentation with ChatGPT for IT domain IE. Finally, we also suggest directions for future research on the new STSE task, including developing better evaluation metrics.
AB - Information Extraction (IE) in IT is an important foundational task that is needed for many AIOps applications. A major challenge of IE in IT is that we often do not have sufficient labelled data for training machine learning algorithms since acquiring labels is labor-intensive and costly. In this paper, we propose to leverage Large Language Models (LLMs) to address this challenge of low resources and study two data augmentation strategies, i.e., using LLMs to generate pseudo labels and generate synthetic data. We use multiple IE tasks and datasets, including a new Semantic Troubleshooting-Segment Extraction task and Named Entity Recognition, to evaluate the benefits of LLMs. Our experiment results suggest that data augmentation using LLMs, specifically, using SeqMix model that combines active labeling with synthetic data samples generated in the embedding-vector space, is a promising approach for IT domain IE. Our study also shows that although data augmentation and direct labeling with the state-of-the-art, ChatGPT model achieves a high performance on general domain IE, there is a need to adapt it for IE from IT text data. Moreover, our initial exploration of two label weighting and selection strategies (confidence and consistency-based) suggests that they could be used to improve data augmentation with ChatGPT for IT domain IE. Finally, we also suggest directions for future research on the new STSE task, including developing better evaluation metrics.
KW - information extraction
KW - IT domain
KW - LLM
UR - http://www.scopus.com/inward/record.url?scp=85186140312&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85186140312&partnerID=8YFLogxK
U2 - 10.1109/ICDMW60847.2023.00157
DO - 10.1109/ICDMW60847.2023.00157
M3 - Conference contribution
AN - SCOPUS:85186140312
T3 - IEEE International Conference on Data Mining Workshops, ICDMW
SP - 1203
EP - 1212
BT - Proceedings - 23rd IEEE International Conference on Data Mining Workshops, ICDMW 2023
A2 - Wang, Jihe
A2 - He, Yi
A2 - Dinh, Thang N.
A2 - Grant, Christan
A2 - Qiu, Meikang
A2 - Pedrycz, Witold
PB - IEEE Computer Society
T2 - 23rd IEEE International Conference on Data Mining Workshops, ICDMW 2023
Y2 - 1 December 2023 through 4 December 2023
ER -