Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Transfer Learning for Mining Feature Requests and Bug Reports from Tweets and App Store Reviews
Technical University of Munich, DEU.
Qualicen GmbH, DEU.
Qualicen GmbH, DEU.
Blekinge Institute of Technology, Faculty of Computing, Department of Software Engineering.ORCID iD: 0000-0003-3995-6125
Show others and affiliations
2021 (English)In: Proceedings of the IEEE International Conference on Requirements Engineering / [ed] Yue T., Mirakhorli M., IEEE Computer Society , 2021, p. 80-86Conference paper, Published paper (Refereed)
Abstract [en]

Identifying feature requests and bug reports in user comments holds great potential for development teams. However, automated mining of RE-related information from social media and app stores is challenging since (1) about 70% of user comments contain noisy, irrelevant information, (2) the amount of user comments grows daily making manual analysis unfeasible, and (3) user comments are written in different languages. Existing approaches build on traditional machine learning (ML) and deep learning (DL), but fail to detect feature requests and bug reports with high Recall and acceptable Precision which is necessary for this task. In this paper, we investigate the potential of transfer learning (TL) for the classification of user comments. Specifically, we train both monolingual and multilingual BERT models and compare the performance with state-of-the-art methods. We found that monolingual BERT models outperform existing baseline methods in the classification of English App Reviews as well as English and Italian Tweets. However, we also observed that the application of heavyweight TL models does not necessarily lead to better performance. In fact, our multilingual BERT models perform worse than traditional ML methods. © 2021 IEEE.

Place, publisher, year, edition, pages
IEEE Computer Society , 2021. p. 80-86
Keywords [en]
Data Driven Requirements Engineering, Natural Language Processing, Social Media Analytics, Deep learning, E-learning, Learning algorithms, Natural language processing systems, Social networking (online), App stores, Bug reports, Data driven, Data driven requirement engineering, Development teams, Feature requests, Performance, Requirement engineering, Transfer learning, Requirements engineering
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:bth-22376DOI: 10.1109/REW53955.2021.00019ISI: 000788547300011Scopus ID: 2-s2.0-85118462487ISBN: 9781665418980 (print)OAI: oai:DiVA.org:bth-22376DiVA, id: diva2:1612794
Conference
29th IEEE International Requirements Engineering Conference Workshops, REW 2021, Virtual, Notre Dame, 20 September 2021 through 24 September 2021
Part of project
SERT- Software Engineering ReThought, Knowledge Foundation
Funder
Knowledge Foundation, 20180010
Note

open access

Available from: 2021-11-19 Created: 2021-11-19 Last updated: 2022-05-30Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopusarXiv.org

Authority records

Frattini, Julian

Search in DiVA

By author/editor
Frattini, Julian
By organisation
Department of Software Engineering
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar

doi
isbn
urn-nbn

Altmetric score

doi
isbn
urn-nbn
Total: 58 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf