System disruptions
We are currently experiencing disruptions on the search portals due to high traffic. We are working to resolve the issue, you may temporarily encounter an error message.
Change search
Link to record
Permanent link

Direct link
Publications (10 of 43) Show all publications
Yu, L., Alégroth, E., Chatzipetrou, P. & Gorschek, T. (2025). Experience with Large Language Model Applications for Information Retrieval from Enterprise Proprietary Data. In: Dietmar Pfahl, Javier Gonzalez Huerta, Jil Klünder, Hina Anwar (Ed.), Product-Focused Software Process Improvement: . Paper presented at 25th International Conference on Product-Focused Software Process Improvement, PROFES 2024, Tartu, Dec 2-4, 2024 (pp. 92-107). Springer, 15452
Open this publication in new window or tab >>Experience with Large Language Model Applications for Information Retrieval from Enterprise Proprietary Data
2025 (English)In: Product-Focused Software Process Improvement / [ed] Dietmar Pfahl, Javier Gonzalez Huerta, Jil Klünder, Hina Anwar, Springer, 2025, Vol. 15452, p. 92-107Conference paper, Published paper (Refereed)
Abstract [en]

Large Language Models (LLMs) offer promising capabilities for information retrieval and processing. However, the LLM deployment for querying proprietary enterprise data poses unique challenges, particularly for companies with strict data security policies. This study shares our experience in setting up a secure LLM environment within a FinTech company and utilizing it for enterprise information retrieval while adhering to data privacy protocols. 

We conducted three workshops and 30 interviews with industrial engineers to gather data and requirements. The interviews further enriched the insights collected from the workshops. We report the steps to deploy an LLM solution in an industrial sandboxed environment and lessons learned from the experience. These lessons contain LLM configuration (e.g., chunk_size and top_k settings), local document ingestion, and evaluating LLM outputs.

Our lessons learned serve as a practical guide for practitioners seeking to use private data with LLMs to achieve better usability, improve user experiences, or explore new business opportunities. © The Author(s), under exclusive license to Springer Nature Switzerland AG 2025.

Place, publisher, year, edition, pages
Springer, 2025
Series
Lecture Notes in Computer Science (LNCS), ISSN 0302-9743, E-ISSN 1611-3349 ; 15452
Keywords
AI, Artificial intelligence, Data security, Information retrieval, Large Language Model, LLM, Sandbox environment, Data privacy, Fintech, Enterprise data, Language model, Model application, Modeling environments, Privacy protocols, Security policy, Structured Query Language
National Category
Software Engineering
Identifiers
urn:nbn:se:bth-27326 (URN)10.1007/978-3-031-78386-9_7 (DOI)001423664600007 ()2-s2.0-85211960724 (Scopus ID)9783031783852 (ISBN)
Conference
25th International Conference on Product-Focused Software Process Improvement, PROFES 2024, Tartu, Dec 2-4, 2024
Funder
Knowledge Foundation, 20180010
Available from: 2024-12-28 Created: 2024-12-28 Last updated: 2025-03-14Bibliographically approved
Coppola, R., Feldt, R., Nass, M. & Alégroth, E. (2025). Ranking approaches for similarity-based web element location. Journal of Systems and Software, 222, Article ID 112286.
Open this publication in new window or tab >>Ranking approaches for similarity-based web element location
2025 (English)In: Journal of Systems and Software, ISSN 0164-1212, E-ISSN 1873-1228, Vol. 222, article id 112286Article in journal (Refereed) Published
Abstract [en]

Context: GUI-based tests for web applications are frequently broken by fragility, i.e. regression tests fail due to changing properties of the web elements. The most influential factor for fragility are the locators used in the scripts, i.e. the means of identifying the elements of the GUI.

Objective: We extend a state-of-the-art Multi-Locator solution that considers 14 locators from the DOM model of a web application, and identifies overlapping nodes in the DOM tree (VON-Similo). We augment the approach with standard Machine Learning and Learning to Rank (LTR) approaches to aid the location of web elements.

Method: We document an experiment with a ground truth of 1163 web element pairs, taken from different releases of 40 web applications, to compare the robustness of the algorithms to locator weight change, and the performance of LTR approaches in terms of MeanRank and PctAtN.

Results: Using LTR algorithms, we obtain a maximum probability of finding the correct target at the first position of 88.4% (lowest 82.57%), and among the first three positions of 94.79% (lowest 91.86%). The best mean rank of the correct candidate is 1.57.

Conclusion: The similarity-based approach proved to be highly dependable in the context of web application testing, where a low percentage of matching errors can still be accepted.

Place, publisher, year, edition, pages
Elsevier, 2025
Keywords
GUI testing, Test automation, Test case robustness, Web element locators, XPath locators, Learning to rank, Mean-ranks, Ranking approach, Test case, WEB application, Web applications, Web element locator, Xpath locator, Contrastive Learning
National Category
Software Engineering
Identifiers
urn:nbn:se:bth-27257 (URN)10.1016/j.jss.2024.112286 (DOI)001375573600001 ()2-s2.0-85211062465 (Scopus ID)
Funder
Knowledge Foundation, 20180010
Available from: 2024-12-17 Created: 2024-12-17 Last updated: 2024-12-27Bibliographically approved
Bauer, A., Frattini, J. & Alégroth, E. (2024). Augmented Testing to support Manual GUI-based Regression Testing: An Empirical Study. Empirical Software Engineering, 29(6), Article ID 140.
Open this publication in new window or tab >>Augmented Testing to support Manual GUI-based Regression Testing: An Empirical Study
2024 (English)In: Empirical Software Engineering, ISSN 1382-3256, E-ISSN 1573-7616, Vol. 29, no 6, article id 140Article in journal (Refereed) Published
Abstract [en]

Context: Manual graphical user interface (GUI) software testing presents a substantial part of the overall practiced testing efforts, despite various research efforts to further increase test automation. Augmented Testing (AT), a novel approach for GUI testing, aims to aid manual GUI-based testing through a tool-supported approach where an intermediary visual layer is rendered between the system under test (SUT) and the tester, superimposing relevant test information.

Objective: The primary objective of this study is to gather empirical evidence regarding AT's efficiency compared to manual GUI-based regression testing. Existing studies involving testing approaches under the AT definition primarily focus on exploratory GUI testing, leaving a gap in the context of regression testing. As a secondary objective, we investigate AT's benefits, drawbacks, and usability issues when deployed with the demonstrator tool, Scout.

Method: We conducted an experiment involving 13 industry professionals, from six companies, comparing AT to manual GUI-based regression testing. These results were complemented by interviews and Bayesian data analysis (BDA) of the study's quantitative results.

Results: The results of the Bayesian data analysis revealed that the use of AT shortens test durations in 70% of the cases on average, concluding that AT is more efficient.When comparing the means of the total duration to perform all tests, AT reduced the test duration by 36% in total. Participant interviews highlighted nine benefits and eleven drawbacks of AT, while observations revealed four usability issues.

Conclusion: This study makes an empirical contribution to understanding Augmented Testing, a promising approach to improve the efficiency of GUI-based regression testing in practice. Furthermore, it underscores the importance of continual refinements of AT.

Place, publisher, year, edition, pages
Springer, 2024
Keywords
GUI-based testing, GUI testing, Augmented Testing, manual teting, Bayesian data analysis
National Category
Software Engineering
Research subject
Systems Engineering
Identifiers
urn:nbn:se:bth-25391 (URN)10.1007/s10664-024-10522-z (DOI)001292331700002 ()2-s2.0-85201391671 (Scopus ID)
Funder
Knowledge Foundation, 20180010
Available from: 2023-09-18 Created: 2023-09-18 Last updated: 2024-08-30Bibliographically approved
Fucci, D., Alégroth, E., Felderer, M. & Johannesson, C. (2024). Evaluating software security maturity using OWASP SAMM: Different approaches and stakeholders perceptions. Journal of Systems and Software, 214, Article ID 112062.
Open this publication in new window or tab >>Evaluating software security maturity using OWASP SAMM: Different approaches and stakeholders perceptions
2024 (English)In: Journal of Systems and Software, ISSN 0164-1212, E-ISSN 1873-1228, Vol. 214, article id 112062Article in journal (Refereed) Published
Abstract [en]

Background: Recent years have seen a surge in cyber-attacks, which can be prevented or mitigated using software security activities. OWASP SAMM is a maturity model providing a versatile way for companies to assess their security posture and plan for improvements. Objective: We perform an initial SAMM assessment in collaboration with a company in the financial domain. Our objective is to assess a holistic inventory of the company security-related activities, focusing on how different roles perform the assessment and how they perceive the instrument used in the process. Methodology: We perform a case study to collect data using SAMM in a lightweight and novel manner through assessment using an online survey with 17 participants and a focus group with seven participants. Results: We show that different roles perceive maturity differently and that the two assessments deviate only for specific practices making the lightweight approach a viable and efficient solution in industrial practice. Our results indicate that the questions included in the SAMM assessment tool are answered easily and confidently across most roles. Discussion: Our results suggest that companies can productively use a lightweight SAMM assessment. We provide nine lessons learned for guiding industrial practitioners in the evaluation of their current security posture as well as for academics wanting to utilize SAMM as a research tool in industrial settings. Editor's note: Open Science material was validated by the Journal of Systems and Software Open Science Board. © 2024 The Author(s)

Place, publisher, year, edition, pages
Elsevier, 2024
Keywords
Industry-academia collaboration, OWASP SAMM, Software security, Cybersecurity, Industrial research, Petroleum reservoir evaluation, Cyber-attacks, Evaluating software, Financial domains, Maturity model, Open science, Security activities, Stakeholder perception, Network security
National Category
Software Engineering
Identifiers
urn:nbn:se:bth-26188 (URN)10.1016/j.jss.2024.112062 (DOI)001237888500001 ()2-s2.0-85192019707 (Scopus ID)
Funder
Knowledge Foundation, 20180010
Available from: 2024-05-13 Created: 2024-05-13 Last updated: 2024-06-19Bibliographically approved
Nass, M., Alégroth, E. & Feldt, R. (2024). Improving Web Element Localization by Using a Large Language Model. Software testing, verification & reliability, 34(7)
Open this publication in new window or tab >>Improving Web Element Localization by Using a Large Language Model
2024 (English)In: Software testing, verification & reliability, ISSN 0960-0833, E-ISSN 1099-1689, Vol. 34, no 7Article in journal (Refereed) Published
Abstract [en]

Web-based test automation heavily relies on accurately finding web elements. Traditional methods compare attributes but don't grasp the context and meaning of elements and words. The emergence of Large Language Models (LLMs) like GPT-4, which can show human-like reasoning abilities on some tasks, offers new opportunities for software engineering and web element localization. This paper introduces and evaluates VON Similo LLM, an enhanced web element localization approach. Using an LLM, it selects the most likely web element from the top-ranked ones identified by the existing VON Similo method, ideally aiming to get closer to human-like selection accuracy. An experimental study was conducted using 804 web element pairs from 48 real-world web applications. We measured the number of correctly identified elements as well as the execution times, comparing the effectiveness and efficiency of VON Similo LLM against the baseline algorithm. In addition, motivations from the LLM were recorded and analyzed for all instances where the original approach failed to find the right web element. VON Similo LLM demonstrated improved performance, reducing failed localizations from 70 to 39 (out of 804), a 44 percent reduction. Despite its slower execution time and additional costs of using the GPT-4 model, the LLMs human-like reasoning showed promise in enhancing web element localization. LLM technology can enhance web element identification in GUI test automation, reducing false positives and potentially lowering maintenance costs. However, further research is necessary to fully understand LLMs capabilities, limitations, and practical use in GUI testing.

Place, publisher, year, edition, pages
John Wiley & Sons, 2024
Keywords
GUI Testing, Test Automation, Test Case Robustness, Web Element Locators, Large Language Models
National Category
Computer Systems
Research subject
Software Engineering
Identifiers
urn:nbn:se:bth-25637 (URN)10.1002/stvr.1893 (DOI)001290853000001 ()2-s2.0-85201296537 (Scopus ID)
Funder
Knowledge Foundation, 20180010
Available from: 2023-11-22 Created: 2023-11-22 Last updated: 2025-01-03Bibliographically approved
Coppola, R., Fulcini, T., Ardito, L., Torchiano, M. & Alégroth, E. (2024). On Effectiveness and Efficiency of Gamified Exploratory GUI Testing. IEEE Transactions on Software Engineering, 50(2), 322-337
Open this publication in new window or tab >>On Effectiveness and Efficiency of Gamified Exploratory GUI Testing
Show others...
2024 (English)In: IEEE Transactions on Software Engineering, ISSN 0098-5589, E-ISSN 1939-3520, Vol. 50, no 2, p. 322-337Article in journal (Refereed) Published
Abstract [en]

Context: Gamification appears to improve enjoyment and quality of execution of software engineering activities, including software testing. Though commonly employed in industry, manual exploratory testing of web application GUIs was proven to be mundane and expensive. Gamification applied to that kind of testing activity has the potential to overcome its limitations, though no empirical research has explored this area yet.

Goal: Collect preliminary insights on how gamification, when performed by novice testers, affects the effectiveness, efficiency, test case realism, and user experience in exploratory testing of web applications.

Method: Common gamification features augment an existing exploratory testing tool: Final Score with Leaderboard, Injected Bugs, Progress Bar, and Exploration Highlights. The original tool and the gamified version are then compared in an experiment involving 144 participants. User experience is elicited using the Technology Acceptance Model (TAM) questionnaire instrument.

Results: Statistical analysis identified several significant differences for metrics that represent the effectiveness and efficiency of tests showing an improvement in coverage when they were developed with gamification. Additionally, user experience is improved with gamification.

Conclusions: Gamification of exploratory testing has a tangible effect on how testers create test cases for web applications. While the results are mixed, the effects are most beneficial and interesting and warrant more research in the future. Further research shall be aimed at confirming the presented results in the context of state-of-the-art testing tools and real-world development environments. 

Place, publisher, year, edition, pages
IEEE Computer Society, 2024
Keywords
Games, Gamification, Graphical user interfaces, Manuals, Software, Software testing, Task analysis, User experience, Web Application Testing, Application programs, Efficiency, Job analysis, Exploratory testing, Game, Manual, Software testings, Users' experiences
National Category
Software Engineering
Identifiers
urn:nbn:se:bth-25929 (URN)10.1109/TSE.2023.3348036 (DOI)001167516600010 ()2-s2.0-85181571816 (Scopus ID)
Available from: 2024-01-29 Created: 2024-01-29 Last updated: 2024-04-04Bibliographically approved
Yu, L., Alégroth, E., Chatzipetrou, P. & Gorschek, T. (2024). Visualizing CI’s role in software quality attribute evaluation: A Roadmap for Using Continuous Integration Environments. Communications of the ACM, 67(6), 82-90
Open this publication in new window or tab >>Visualizing CI’s role in software quality attribute evaluation: A Roadmap for Using Continuous Integration Environments
2024 (English)In: Communications of the ACM, ISSN 0001-0782, E-ISSN 1557-7317, Vol. 67, no 6, p. 82-90Article in journal (Refereed) Published
Abstract [en]

Quality attributes of software systems, also known as system qualities, such as performance, security, and scalability, continue to grow in importance in industrial practice. The evaluation of quality attributes is critical to software development since optimizing a software system’s core attributes can provide marketing advantage and set a product apart from its competitors. Many existing studies of unsuccessful development projects report that lack of quality attribute evaluation is often a contributing factor of project failure. Therefore, continuous quality attribute evaluation, throughout the development process, is needed to ensure customers’ expectations and demands are met.

Manual evaluation of software attributes is common in many software development companies, but it has proven to be insufficient in meeting the demands of rapid releases and high-quality expectations from customers. Automated practices have therefore gained widespread popularity as a solution to enhance efficiency, reduce costs, and increase accuracy compared to manual evaluation.

One way to automate the evaluation is using continuous integration (CI) environments. The CI environment provides several benefits, such as fast feedback on code quality, early detection of quality defects, and visualization of system quality trends. As such, these environments inherently offer organizations the opportunity to continuously monitor the quality of their software systems. However, an immature automation process can result in negative outcomes, such as cost and schedule overruns, slow feedback loops, and delayed releases.

To improve the evaluation process, prior studies have investigated different key areas, including knowledge, processes, tools, and metrics. While leveraging these areas can have a positive impact on quality evaluation, to the best of our knowledge, there is a lack of frameworks that link CI environment knowledge, metrics, and evolution together.

In this article, we aim to fill this gap by presenting the state-of-practice of using CI environments for the evaluation of quality attributes. This is achieved through an industrial study at four partner companies. Study results show that metrics acquired from CI components have a positive effect on evaluating quality requirements. Through analyzing these results, we propose a model by providing guidelines to mature existing CI environments that organizations can use for quality improvements.

As such, we claim the following contributions of this study:

A generic model of how CI environments contribute to quality attribute evaluation.

Empirical evidence that demonstrates how CI components can be used to produce data supporting the evaluation of quality attributes with metrics.

A model, derived from the study results, which provides decision support to evolve software quality evaluation through CI environments over time. © 2024 Owner/Author.

Place, publisher, year, edition, pages
Association for Computing Machinery (ACM), 2024
Keywords
Attribute evaluation, Continuous integrations, Integration environments, Roadmap, Software quality attributes, Computer software selection and evaluation
National Category
Software Engineering
Identifiers
urn:nbn:se:bth-26367 (URN)10.1145/3631519 (DOI)001240956100025 ()2-s2.0-85194381501 (Scopus ID)
Available from: 2024-06-17 Created: 2024-06-17 Last updated: 2024-08-05Bibliographically approved
Zabardast, E., Gonzalez-Huerta, J., Gorschek, T., Šmite, D., Alégroth, E. & Fagerholm, F. (2023). A taxonomy of assets for the development of software-intensive products and services. Journal of Systems and Software, 202, Article ID 111701.
Open this publication in new window or tab >>A taxonomy of assets for the development of software-intensive products and services
Show others...
2023 (English)In: Journal of Systems and Software, ISSN 0164-1212, E-ISSN 1873-1228, Vol. 202, article id 111701Article in journal (Refereed) Published
Abstract [en]

Context:Developing software-intensive products or services usually involves a plethora of software artefacts. Assets are artefacts intended to be used more than once and have value for organisations; examples include test cases, code, requirements, and documentation. During the development process, assets might degrade, affecting the effectiveness and efficiency of the development process. Therefore, assets are an investment that requires continuous management.

Identifying assets is the first step for their effective management. However, there is a lack of awareness of what assets and types of assets are common in software-developing organisations. Most types of assets are understudied, and their state of quality and how they degrade over time have not been well-understood.

Methods:We performed an analysis of secondary literature and a field study at five companies to investigate and identify assets to fill the gap in research. The results were analysed qualitatively and summarised in a taxonomy.

Results:We present the first comprehensive, structured, yet extendable taxonomy of assets, containing 57 types of assets.

Conclusions:The taxonomy serves as a foundation for identifying assets that are relevant for an organisation and enables the study of asset management and asset degradation concepts.

Place, publisher, year, edition, pages
Elsevier, 2023
Keywords
Assets in software engineering, Asset management in software engineering, Assets for software-intensive products or services, Taxonomy
National Category
Software Engineering
Identifiers
urn:nbn:se:bth-24426 (URN)10.1016/j.jss.2023.111701 (DOI)000984121100001 ()2-s2.0-85152899759 (Scopus ID)
Funder
Knowledge Foundation, 20170176Knowledge Foundation, 20180010
Available from: 2023-04-11 Created: 2023-04-11 Last updated: 2023-06-02Bibliographically approved
Yu, L., Alégroth, E., Chatzipetrou, P. & Gorschek, T. (2023). Automated NFR testing in Continuous Integration Environments: a multi-case study of Nordic companies. Empirical Software Engineering, 28(6), Article ID 144.
Open this publication in new window or tab >>Automated NFR testing in Continuous Integration Environments: a multi-case study of Nordic companies
2023 (English)In: Empirical Software Engineering, ISSN 1382-3256, E-ISSN 1573-7616, Vol. 28, no 6, article id 144Article in journal (Refereed) Published
Abstract [en]

Context: Non-functional requirements (NFRs) (also referred to as system qualities) are essential for developing high-quality software.Notwithstanding its importance, NFR testing remains challenging, especially in terms of automation.Compared to manual verification, automated testing shows the potential to improve the efficiency and effectiveness of quality assurance, especially in the context of  Continuous Integration (CI).However, studies on how companies manage automated NFR testing through CI are limited.

Objective: This study examines how automated NFR testing can be enabledand supported using CI environments in software development companies.

Method: We performed a multi-case study at four companies by conducting 22 semi-structured interviews with industrial practitioners.

Results: Maintainability, reliability, performance, security and scalability, were found to be evaluated with automated tests in CI environments.Testing practices, quality metrics, and challenges for measuring NFRs were reported.

Conclusions: This study presents an empirically derived model that shows how data produced by CI environments can be used for evaluation and monitoring of implemented NFR quality. Additionally, the manuscript presents explicit metrics, CI components, tools, and challenges that shall be considered while performing NFR testing in practice.

Place, publisher, year, edition, pages
Springer, 2023
Keywords
Automated testing, Case study, CI, Continuous integration, Metrics, NFR, Non-functional requirements, Automation, Integration, Integration testing, Quality control, Software design, Case-studies, Continuous integrations, Integration environments, Metric, Nordic companies, System quality, Quality assurance
National Category
Software Engineering
Identifiers
urn:nbn:se:bth-24400 (URN)10.1007/s10664-023-10356-1 (DOI)001087927600001 ()2-s2.0-85174862814 (Scopus ID)
Funder
Knowledge Foundation, 20180010Knowledge Foundation, 20170213
Available from: 2023-03-29 Created: 2023-03-29 Last updated: 2024-08-07Bibliographically approved
Bauer, A., Coppola, R., Alégroth, E. & Gorschek, T. (2023). Code review guidelines for GUI-based testing artifacts. Information and Software Technology, 163, Article ID 107299.
Open this publication in new window or tab >>Code review guidelines for GUI-based testing artifacts
2023 (English)In: Information and Software Technology, ISSN 0950-5849, E-ISSN 1873-6025, Vol. 163, article id 107299Article, review/survey (Refereed) Published
Abstract [en]

Context: Review of software artifacts, such as source or test code, is a common practice in industrial practice. However, although review guidelines are available for source and low-level test code, for GUI-based testing artifacts, such guidelines are missing. Objective: The goal of this work is to define a set of guidelines from literature about production and test code, that can be mapped to GUI-based testing artifacts. Method: A systematic literature review is conducted, using white and gray literature to identify guidelines for source and test code. These synthesized guidelines are then mapped, through examples, to create actionable, and applicable, guidelines for GUI-based testing artifacts. Results: The results of the study are 33 guidelines, summarized in nine guideline categories, that are successfully mapped as applicable to GUI-based testing artifacts. Of the collected literature, only 10 sources contained test-specific code review guidelines. These guideline categories are: perform automated checks, use checklists, provide context information, utilize metrics, ensure readability, visualize changes, reduce complexity, check conformity with the requirements and follow design principles and patterns. Conclusion: This pivotal set of guidelines provides an industrial contribution in filling the gap of general guidelines for review of GUI-based testing artifacts. Additionally, this work highlights, from an academic perspective, the need for future research in this area to also develop guidelines for other specific aspects of GUI-based testing practice, and to take into account other facets of the review process not covered by this work, such as reviewer selection. © 2023 The Author(s)

Place, publisher, year, edition, pages
Elsevier, 2023
Keywords
Code review, GUI testing, GUI-based testing, Guidelines, Modern code review, Practices, Software testing, Graphical user interfaces, Guideline, Practice, Software testings, Source codes, Test code
National Category
Software Engineering
Identifiers
urn:nbn:se:bth-25235 (URN)10.1016/j.infsof.2023.107299 (DOI)001051358500001 ()2-s2.0-85165535690 (Scopus ID)
Funder
Knowledge Foundation, 20180010
Available from: 2023-08-08 Created: 2023-08-08 Last updated: 2023-09-18Bibliographically approved
Projects
M.E.T.A. – Modelling Efficient Test Architectures [20180102]; Blekinge Institute of Technology; Publications
Alégroth, E., Petersén, E. & Tinnerholm, J. (2021). A Failed attempt at creating Guidelines for Visual GUI Testing: An industrial case study. In: Proceedings - 2021 IEEE 14th International Conference on Software Testing, Verification and Validation, ICST 2021: . Paper presented at 14th IEEE International Conference on Software Testing, Verification and Validation, ICST 2021, 12 April 2021 through 16 April 2021 (pp. 340-350). Institute of Electrical and Electronics Engineers Inc., Article ID 9438551.
Organisations
Identifiers
ORCID iD: ORCID iD iconorcid.org/0000-0001-7526-3727

Search in DiVA

Show all publications