Exploring the efficacy of transfer learning in mining image-based software artifacts

Abstract Background Transfer learning allows us to train deep architectures requiring a large number of learned parameters, even if the amount of available data is limited, by leveraging existing models previously trained for another task. In previous attempts to classify image-based software artifa...

Full description

Bibliographic Details
Main Authors: Natalie Best, Jordan Ott, Erik J. Linstead
Format: Article
Language:English
Published: SpringerOpen 2020-08-01
Series:Journal of Big Data
Subjects:
Online Access:http://link.springer.com/article/10.1186/s40537-020-00335-4
_version_ 1828488825845841920
author Natalie Best
Jordan Ott
Erik J. Linstead
author_facet Natalie Best
Jordan Ott
Erik J. Linstead
author_sort Natalie Best
collection DOAJ
description Abstract Background Transfer learning allows us to train deep architectures requiring a large number of learned parameters, even if the amount of available data is limited, by leveraging existing models previously trained for another task. In previous attempts to classify image-based software artifacts in the absence of big data, it was noted that standard off-the-shelf deep architectures such as VGG could not be utilized due to their large parameter space and therefore had to be replaced by customized architectures with fewer layers. This proves to be challenging to empirical software engineers who would like to make use of existing architectures without the need for customization. Findings Here we explore the applicability of transfer learning utilizing models pre-trained on non-software engineering data applied to the problem of classifying software unified modeling language (UML) diagrams. Our experimental results show training reacts positively to transfer learning as related to sample size, even though the pre-trained model was not exposed to training instances from the software domain. We contrast the transferred network with other networks to show its advantage on different sized training sets, which indicates that transfer learning is equally effective to custom deep architectures in respect to classification accuracy when large amounts of training data is not available. Conclusion Our findings suggest that transfer learning, even when based on models that do not contain software engineering artifacts, can provide a pathway for using off-the-shelf deep architectures without customization. This provides an alternative to practitioners who want to apply deep learning to image-based classification but do not have the expertise or comfort to define their own network architectures.
first_indexed 2024-12-11T10:15:38Z
format Article
id doaj.art-70ee77d769df445ab8d8103313e59019
institution Directory Open Access Journal
issn 2196-1115
language English
last_indexed 2024-12-11T10:15:38Z
publishDate 2020-08-01
publisher SpringerOpen
record_format Article
series Journal of Big Data
spelling doaj.art-70ee77d769df445ab8d8103313e590192022-12-22T01:11:37ZengSpringerOpenJournal of Big Data2196-11152020-08-017111010.1186/s40537-020-00335-4Exploring the efficacy of transfer learning in mining image-based software artifactsNatalie Best0Jordan Ott1Erik J. Linstead2Fowler School of Engineering, Chapman UniversitySchool of Information and Computer Science, University of California, IrvineFowler School of Engineering, Chapman UniversityAbstract Background Transfer learning allows us to train deep architectures requiring a large number of learned parameters, even if the amount of available data is limited, by leveraging existing models previously trained for another task. In previous attempts to classify image-based software artifacts in the absence of big data, it was noted that standard off-the-shelf deep architectures such as VGG could not be utilized due to their large parameter space and therefore had to be replaced by customized architectures with fewer layers. This proves to be challenging to empirical software engineers who would like to make use of existing architectures without the need for customization. Findings Here we explore the applicability of transfer learning utilizing models pre-trained on non-software engineering data applied to the problem of classifying software unified modeling language (UML) diagrams. Our experimental results show training reacts positively to transfer learning as related to sample size, even though the pre-trained model was not exposed to training instances from the software domain. We contrast the transferred network with other networks to show its advantage on different sized training sets, which indicates that transfer learning is equally effective to custom deep architectures in respect to classification accuracy when large amounts of training data is not available. Conclusion Our findings suggest that transfer learning, even when based on models that do not contain software engineering artifacts, can provide a pathway for using off-the-shelf deep architectures without customization. This provides an alternative to practitioners who want to apply deep learning to image-based classification but do not have the expertise or comfort to define their own network architectures.http://link.springer.com/article/10.1186/s40537-020-00335-4Deep learningTransfer learningUML
spellingShingle Natalie Best
Jordan Ott
Erik J. Linstead
Exploring the efficacy of transfer learning in mining image-based software artifacts
Journal of Big Data
Deep learning
Transfer learning
UML
title Exploring the efficacy of transfer learning in mining image-based software artifacts
title_full Exploring the efficacy of transfer learning in mining image-based software artifacts
title_fullStr Exploring the efficacy of transfer learning in mining image-based software artifacts
title_full_unstemmed Exploring the efficacy of transfer learning in mining image-based software artifacts
title_short Exploring the efficacy of transfer learning in mining image-based software artifacts
title_sort exploring the efficacy of transfer learning in mining image based software artifacts
topic Deep learning
Transfer learning
UML
url http://link.springer.com/article/10.1186/s40537-020-00335-4
work_keys_str_mv AT nataliebest exploringtheefficacyoftransferlearninginminingimagebasedsoftwareartifacts
AT jordanott exploringtheefficacyoftransferlearninginminingimagebasedsoftwareartifacts
AT erikjlinstead exploringtheefficacyoftransferlearninginminingimagebasedsoftwareartifacts