Automating Software Size Measurement With Language Models: Insights From Industrial Case Studies
No Thumbnail Available
Date
2026
Journal Title
Journal ISSN
Volume Title
Publisher
Elsevier Inc.
Open Access Color
OpenAIRE Downloads
OpenAIRE Views
Abstract
Objective software size measurement is critical for accurate effort estimation, yet many organizations avoid it due to high costs, required expertise, and time-consuming manual effort. This often leads to vague predictions, poor planning, and project overruns. To address this challenge, we investigate the use of pre-trained language models — BERT and SE-BERT — to automate size measurement based on textual requirements using COSMIC and MicroM methods. We constructed one heterogeneous dataset and two industrial datasets, each manually measured by experienced analysts. Models were evaluated in three settings: (i) generic model evaluation, where the models are trained and tested on heterogeneous data, (ii) internal evaluation, where the models are trained and tested on organization-specific data, and (iii) external evaluation, where generic models were tested on organization-specific data. Results show that organization-specific models significantly outperform generic models, indicating that aligning training data with the target organization's requirement style is critical for accuracy. SE-BERT, a domain-adapted variant of BERT, improves performance, particularly in low-resource settings. These findings highlight the practical potential of tailoring training data for broader adoption and cost-effective software size measurement in industrial contexts. © 2025 Elsevier B.V., All rights reserved.
Description
Keywords
BERT, Case Study, Cosmic, Microm, Natural Language Processing, NLP, Software Size Measurement
Turkish CoHE Thesis Center URL
Fields of Science
Citation
WoS Q
Q1
Scopus Q
Q1
Source
Journal of Systems and Software
Volume
231