IRMA-International.org: Creator of Knowledge
Information Resources Management Association
Advancing the Concepts & Practices of Information Resources Management in Modern Organizations

Improving Logging Prediction on Imbalanced Datasets: A Case Study on Open Source Java Projects

Improving Logging Prediction on Imbalanced Datasets: A Case Study on Open Source Java Projects
View Sample PDF
Author(s): Sangeeta Lal (Jaypee Institute of Information Technology Noida, Department of CSE & IT, Noida, Uttar-Pradesh, India), Neetu Sardana (Jaypee Institute of Information Technology Noida, Department of CSE & IT, Noida, Uttar-Pradesh, India)and Ashish Sureka (ABB Corporate Research Center, Bangalore, India)
Copyright: 2020
Pages: 33
Source title: Cognitive Analytics: Concepts, Methodologies, Tools, and Applications
Source Author(s)/Editor(s): Information Resources Management Association (USA)
DOI: 10.4018/978-1-7998-2460-2.ch039

Purchase

View Improving Logging Prediction on Imbalanced Datasets: A Case Study on Open Source Java Projects on the publisher's website for pricing and purchasing information.

Abstract

Logging is an important yet tough decision for OSS developers. Machine-learning models are useful in improving several steps of OSS development, including logging. Several recent studies propose machine-learning models to predict logged code construct. The prediction performances of these models are limited due to the class-imbalance problem since the number of logged code constructs is small as compared to non-logged code constructs. No previous study analyzes the class-imbalance problem for logged code construct prediction. The authors first analyze the performances of J48, RF, and SVM classifiers for catch-blocks and if-blocks logged code constructs prediction on imbalanced datasets. Second, the authors propose LogIm, an ensemble and threshold-based machine-learning model. Third, the authors evaluate the performance of LogIm on three open-source projects. On average, LogIm model improves the performance of baseline classifiers, J48, RF, and SVM, by 7.38%, 9.24%, and 4.6% for catch-blocks, and 12.11%, 14.95%, and 19.13% for if-blocks logging prediction.

Related Content

Jaime Salvador, Zoila Ruiz, Jose Garcia-Rodriguez. © 2020. 12 pages.
Stavros Pitoglou. © 2020. 11 pages.
Mette L. Baran. © 2020. 13 pages.
Yingxu Wang, Victor Raskin, Julia M. Rayz, George Baciu, Aladdin Ayesh, Fumio Mizoguchi, Shusaku Tsumoto, Dilip Patel, Newton Howard. © 2020. 15 pages.
Yingxu Wang, Lotfi A. Zadeh, Bernard Widrow, Newton Howard, Françoise Beaufays, George Baciu, D. Frank Hsu, Guiming Luo, Fumio Mizoguchi, Shushma Patel, Victor Raskin, Shusaku Tsumoto, Wei Wei, Du Zhang. © 2020. 18 pages.
Nayem Rahman. © 2020. 24 pages.
Amir Manzoor. © 2020. 27 pages.
Body Bottom