Modelling Assistants Under the Lens: What the Current Landscape and Gaps Mean for Digital Health Software Development using Low-Code/No-Code
Knowledge database Structures & processes Data management & digitalisation System selection & implementation Human Training & digital expertise C.3: Artificial intelligence-based software factory for MedTech applicationsModelling assistants can accelerate Low-Code/No-Code development, but over 50% fail to report key information like limitations and evaluation metrics. What does this mean for teams adopting them in LCNC software development? What are the risks in domains like digital health, where software quality is a MUST? And what actions should we take as a community?
Problem Description, Research Question, Relevance
Modelling assistants in Low-code/No-code tools—also known as Model-Driven Software Engineering (MDSE) tools—hold the potential to streamline even more the software development by supporting software developers, domain experts, and designers while modelling..
However, for teams operating in quality-critical domains like digital health software development, the decision to adopt a modelling assistant is far from trivial. The first question teams face is: How can we benefit from a modelling assistant, and what do we need to know to make an informed decision for adopting them? To explore this, we investigate the following research questions:
- RQ1: What goals and limitations do existing modelling assistance proposals report?
- RQ2: Which evaluation metrics do existing modelling assistance proposals consider?
Methods and Approach in the Project
We conducted a systematic mapping [1] following the guidelines proposed by Petersen et al. [3] to review the state of the art and answer RQ1 and RQ2 (among other questions). Our study collected and analysed data from both the literature and practice as follows:
- To gather data from the literature, three reviewers selected 58 proposals from a set of 3.176 records from 1985 to 2024 using inclusion/exclusion criteria. Having 58 proposals selected with a substantial inclusion agreement [4]—i.e., having a 0.8 > K-statistic > 0.61. We clustered the data to answer the research questions.
- To gather data from the practice, we explored the 17 LCNC tools that were mentioned in the Gartner Magic Quadrant for enterprise low-code/no-code application platforms [2] and extracted quotes about goals, limitations, and evaluation metrics from modelling assistants.
Results and Findings
After executing our study, we found that software development teams can benefit from modelling assistants in several key areas, such as software model evolution, vulnerability detection, consistency checking, and change propagation, among others.
However, many modelling assistants come with limitations:
- Accuracy: Recommendations may be inconsistent or irrelevant; automatic extractions may lack precision.
- Effort: Assistants may increase modelling overhead through additional steps or manual interventions.
- Generality: Some are narrowly tailored to specific domains, refactorings, or training datasets.
- Learnability: Users face steep learning curves or must adopt intermediate modelling languages.
- Scope: Key assistant features may be missing, limiting expected benefits.
- Usability: Poor user interaction design can hinder adoption by users.
Evaluation metrics, when provided, tend to focus on:
- Effectiveness: Fault detection rates, accuracy, recall, precision, and accepted suggestions.
- Efficiency: Modelling time, performance, computational cost, and recommendation latency.
- User perception: Perceived usefulness and potential for industrial adoption.
Despite their importance, this data is often missing.
In literature:
- 50% of proposals fail to report any limitations.
- 48.6% do not include any evaluation metrics.
In practice, the gap is wider:
- 75% of modelling assistants lack documented limitations.
- 68.8% provide no evaluation data.
This is especially problematic in domains as digital health software development: the risk of adopting an assistant that generates flawed recommendations/editions could directly affect the resulting digital health software and, in consequence, the patient and practitioner experience with it. The absence of clear information makes an already difficult decision even more uncertain—and riskier. What can we do, as a community, to bridge this gap? (See next section.)
Recommendations for Practice
- For modelling assistant designers/developers/engineers: While technical features are essential, transparency is just as crucial. Report not only what your assistant can do, but also its limitations and how well it performs. Without this, LCNC development teams in quality-critical domains like digital health cannot make informed decisions. Fill the gaps—report fully.
- For researchers working on modelling assistance: We play a key role in shaping trustworthy modelling assistants. This requires clear methods and frameworks that prevent underreporting and promote transparent designs from modelling assistant developers. Existing initiatives like the Framework for Designing Intelligent Modelling Assistance (FR-IMA) [5] and the Emerging Framework for Modelling Assistance [6] are good starting points. The next steps include proposing tailorable methods that help design and adopt modelling assistance in context.
- And finally, don’t forget the users: In this knowledge capsule, we discuss goals, limitations, and evaluation metrics. But users are also important. Check our last post titled “It’s Not Just Developers”: Why Modelling Assistance Must Fit All Roles in Low-Code/No-Code Software Development for Digital Health [7]!
Literature and Other Sources
[1] Mosquera, D., Ruiz, M., Pastor, O., & Spielberger, J. (2024). Understanding the landscape of software modelling assistants for MDSE tools: A systematic mapping. Information and Software Technology, 173, 107492. https://doi.org/10.1016/j.infsof.2024.107492
[2] Gartner: gartner magic quadrant for enterprise low-code application platforms, https://powerapps.microsoft.com/en-us/blog/microsoft-named-a-leader-in-2023-gartner-magic-quadrant-for-enterprise-low-code-application-platforms/, last accessed 2024/05/22.
[3] K. Petersen, S. Vakkalanka, L. Kuzniarz. Guidelines for conducting systematic mapping studies in software engineering: an update. Inf. Softw. Technol., 64 (2015), pp. 1-18, 10.1016/j.infsof.2015.03.007
[4] J.R. Landis, G.G. Koch. The measurement of observer agreement for categorical data. Biometrics, 33 (1977), pp. 159-174, 10.2307/2529310
[5] G. Mussbacher, B. Combemale, S. Abrahão, N. Bencomo, L. Burgueño, G. Engels,
J. Kienzle, T. Kühn, S. Mosser, H. Sahraoui, M. Weyssow. Towards an assessment grid for intelligent modelling assistance. Proceedings of the 23rd ACM/IEEE International Conference on Model Driven Engineering Languages and Systems: Companion Proceedings, New York, NY, USA, ACM (2020), pp. 1-10, 10.1145/3417990.3421396
[6] D. Mosquera, M. Ruiz, O. Pastor, J. Spielberger. Assisted modelling requirements for model-driven development tools. Proceedings of the International Conference on Research Challenges in Information Science (2022), pp. 458-474, 10.1007/978-3-031-05760-1_27
[7] Mosquera, David & Ruiz, Marcela (2025). “It’s Not Just Developers”: Why Modelling Assistance Must Fit All Roles in Low-Code/No-Code Software Development for Digital Health. In Flagshipprojekt SHIFT. Wissensbeitrag C.3 (Nr. 2). [LINK]
