(Erdana Seitzhan)
                     1iD
                     
                     (Alibek Bissembayev)
                     1iD
                     
                     (Assel Mukasheva)
                     1iD
                     박해산
                     (Hae San Park)
                     2iD
                     강정원
                     (Jeong Won Kang)
                     †iD
               
                  - 
                           
                        (School of Information Technology and Engineering, Kazakh-British Technical University,
                        Kazakhstan)
                        
 
                  - 
                           
                        (Dept. of SMART Railway System, Korea National University of Transportation, Republic
                        of Korea.)
                        
 
               
             
            
            
            Copyright © The Korea Institute for Structural Maintenance and Inspection
            
            
            
            
            
               
                  
Key words
               
                LCDPs,  Optimization,  Scalability,  Efficiency,  Integration,  Development.
             
            
          
         
            
                  1. Introduction	
               Low-code development platforms (LCDPs) have evolved to represent a paradigm shift
                  in software development as they provide an alternative to conventional programming
                  by allowing users to create applications through the utilization of visual and intuitive
                  development tools[1].
               
               LCDPs enable so-called citizen developers—people with little or no coding knowledge—to
                  get involved in designing and implementing applications, effectively abstracting away
                  the complicated aspects of software development[2]. This change meets a need in the software industry for agility and speed, as well
                  as a reduced reliance on professionals with niche programming skills, which makes
                  LCDPs an attractive alternative in many sectors undergoing digital transformation[3].
               
               LCDPs emerged from traditional approaches, including Model-Driven Engineering (MDE)
                  and Rapid Application Development (RAD) methods, which focused on improving the software
                  development process by employing abstract reusable models and visual programming tools[4]. These methodologies placed an emphasis on the abstraction of complicated processes,
                  enabling software to be built with greater efficiency and aligning technical solutions
                  more closely with business objectives[5].
               
               Since their formalization in 2014 by Forrester, LCDPs have been widely adopted, especially
                  in sectors that demand dynamic digital solutions like finance, healthcare, and manufacturing[6]. Although LCDPs offer advantages, they also present challenges that complicate enterprise
                  adoption. However, traditional development is still important in scenarios where customizability,
                  scalability, and performance are critical, as LCDPs cannot always satisfy such requirements[7].
               
               Research indicates that while LCDPs excel in smaller, well-defined projects, they
                  face challenges in scaling for larger and more complex applications due to limitations
                  in customization and integration capabilities[8]. Additionally, while LCDPs offer rapid development capabilities, integrating them
                  into existing IT ecosystems often requires additional programming efforts to address
                  functionality and data management gaps. This necessity can hinder scalability in larger
                  applications[9].
               
               This study addresses these gaps in knowledge by exploring the function of LCDPs in
                  contemporary software engineering. More specifically, it aims to investigate the implications
                  that LCDPs have on development velocity, cost-effectiveness, and overall affordability
                  for those without programming knowledge[10]. By doing so, it will also reflect on the ways in which LCDPs are empowering non-technical
                  users, the key limitations they are facing so far in enterprise settings, and future
                  paths for their evolution[11].
               
               This overview helps highlight the benefits and drawbacks around LCDPs, further informing
                  wider debates around how platforms for digital development are transforming software
                  delivery while also offering guidance on how organizations can best leverage LCDPs
                  to achieve optimal results, showing where the overall potential lies for such platforms.
               
               Using a mixed-methods approach of the existing literature, illustrative case studies,
                  and survey analysis, this research discusses the real-world applications and challenges
                  of LCDPs. In doing so, it combines both academic and pragmatic viewpoints on their
                  contribution to the digital transformation and suggests recommendations on tackling
                  the constraints that may act as barriers to their acceptance and wider adoption in
                  the realm of the complex and large-scale.
               
             
            
                  2. Literature Review	
               Low-Code Development Platforms (LCDPs) are disrupting software engineering by offering
                  application design and deployment without the need for extensive coding expertise.
                  Martinez and Pfister cite digital context platforms (LCDPs) among the revalidation
                  tools that are becoming catalysts for digital transformation; however, in construction
                  specifically, LCDPs are also key to integrating technologies like Building Information
                  Modeling (BIM) into Industry 4.0 tools that enable stakeholders to ensure their solutions
                  are delivered on a universal platform and are being used successfully. Their research
                  highlights LCDPs' adaptability to complex needs across industry domains[12].
               
               Similarly, Butting Arvid explored LCDPs' role in enterprise settings, emphasizing
                  their ability to manage model and data variations through model-driven engineering
                  principles. Their findings suggest that LCDPs enhance collaborative programming efforts
                  and organizational adaptability, making them valuable for companies that require customized
                  data integration[13].
               
               The manufacturing sector is also using LCDPs for automation and robotics. Schenkenfelder
                  Bernhard, through multiple case studies, illustrates how LCDPs optimize industrial
                  automation, particularly in mobile applications that require real-time data management
                  and visualization. Their research highlights the ability of the platform’s forms to
                  support operational flexibility and rapid customization, which makes them suitable
                  for environments with dynamic software needs[14].
               
               Beyond industry applications, LCDPs are recognized for their user-friendly, visual
                  development capabilities. Phalake Vaishali focuses on drag-and-drop interfaces, which
                  simplify complex development tasks, enabling rapid deployment even for nontechnical
                  users. Similarly, Daniel Gwendal introduces Xatkit, a low-code framework for chatbot
                  development, which shows how LCDPs facilitate specialized applications such as conversational
                  AI[15-16]. These studies reinforce the accessibility of LCDPs, extending their usability to
                  a wider range of developers and businesses.
               
               From an architectural perspective, Chuanjian Cui analyzes the structural components
                  of LCDPs, such as API designers and conceptual frameworks, which ensure security,
                  stability, and adaptability in evolving business environments. Their study emphasizes
                  that robust design elements are critical for long-term software performance and data
                  integrity, positioning LCDPs as a viable alternative to traditional development methods[17].
               
               A broader discussion on LCDPs' market impact is presented by Gomes and Brito, who
                  examine how LCDPs accelerate digital transformation by improving scalability and adaptability.
                  Their study identifies LCDPs as cost-effective solutions that address organizations’
                  digitalization needs while maintaining business agility[18].
               
               The productivity benefits of LCDPs are further validated by Trigo, Varajão, and Almeida,
                  who compare low code vs. traditional development. Their research in IT Professional
                  finds that LCDPs significantly enhance development speed and reduce manual coding
                  efforts, making them highly effective for projects constrained by tight timelines
                  and limited resources[19].
               
               However, despite their advantages, LCDPs face critical limitations. Käss, Strahringer,
                  and Westner identify key adoption barriers, including scalability issues, security
                  vulnerabilities, and integration challenges. Their findings suggest that while LCDPs
                  streamline development, large enterprises with complex IT infrastructures may encounter
                  difficulties in seamless adoption and long-term sustainability[20].
               
               The existing literature presents LCDPs as powerful enablers of digital transformation,
                  offering speed, accessibility, and flexibility across diverse industries. However,
                  challenges such as scalability, security, and integration complexities remain. As
                  these platforms continue to evolve, future research will be essential in refining
                  LCDP capabilities and addressing adoption barriers, ensuring their effectiveness in
                  large-scale enterprise environments.
               
             
            
                  3. Materials and Methods	
               This study employs a mixed-methods approach, combining a literature review, surveys,
                  and a case study to evaluate the impact of Low-Code Development Platforms (LCDPs)
                  on software engineering. The literature review explores the evolution, benefits, and
                  challenges of LCDPs, identifying key themes such as accessibility and integration
                  issues. Surveys gather insights from software professionals, using both quantitative
                  and qualitative questions to assess LCDP adoption, advantages, and limitations. A
                  case study examines the development of a business application using Mendix, comparing
                  its efficiency to traditional coding methods. Data analysis integrates statistical
                  evaluation and qualitative insights, ensuring a well-rounded perspective on LCDP effectiveness
                  and limitations in real-world applications.
               
               
                     3.1 Detailed Description of Datasets
                  Understanding the datasets is crucial to appreciating the challenges addressed in
                     this study. These datasets provide the foundation for the algorithms applied and connect
                     directly to the goals of improving software reliability and workflow efficiency within
                     Low-Code Development Platforms (LCDPs). Each dataset reflects real-world challenges,
                     making this research practical and impactful.
                  
                  
                        
                        
Fig. 1. Feature Selection Process
                      
                
               
                     3.2 Software Reliability Dataset: A Lens into Feature Selection
                  Figure 1 shows the feature selection process in this work. This dataset is the heart
                     of the feature selection task, aiming to predict software reliability based on various
                     metrics. It comprises 31 features, each representing a distinct dimension of software
                     characteristics, and one target variable, Reliability Class, which categorizes reliability
                     into three levels: Low, Medium, and High.
                  
                  Every day, a software project changes: more and more lines go into the codebase, bugs
                     are found and fixed, and updates are rolled out. Each one adds complexity, affecting
                     the reliability and fragility.
                  
                  The size of the codebase is a governing factor—bigger projects do have more problems;
                     they also improve the chances of defects. Cyclomatic complexity, indicative of the
                     “twistiness” of the code, governs how hard it is to test and maintain. Defect density
                     serves as a measure of quality, showing how many problems there are in relation to
                     the codebase. Mean Time Between Failures (MTBF) measures the resilience of the software,
                     or how long it can run before it breaks.
                  
                  As developers, high code coverage is comforting, as this means we've got more tested
                     code, and as a result, mitigating the chances of undetected bugs.
                  
                  As the name suggests, low-code development environments utilize automated means to
                     direct the actions of developers. Using 400 Particle Swarm Optimization (PSO), we
                     select the most significant features; this dataset enables LCDPs to concentrate on
                     the most important metrics and gives practical advice to enhance reliability without
                     inundating users.
                  
                
               
                     3.3 Workflow Optimization Dataset: Orchestrating Resources
                  Figure 2 shows the workflow optimization diagram. Workflow management is a cornerstone
                     of LCDPs, where users develop, deploy, and manage multiple tasks simultaneously. This
                     dataset captures the intricate dance of resource allocation across computational workflows.
                     Sporting data from 400,000 job hours, it allows us to view how implementations are
                     consuming CPU time, memory, and disk.
                  
                  A busy data center is like a puzzle: tasks fight for CPUs, memory, or storage. To
                     keep the system running smoothly without flooding it, effective resource allocation
                     is crucial. Computational power must be distributed carefully to ensure tasks run
                     effectively. Memory, the backbone of data processing, requires proper management to
                     prevent crashes or inefficiencies. Storage capacity is finite, demanding careful handling,
                     especially for tasks dealing with large datasets. The success or failure of a task
                     serves as a critical indicator, revealing whether the system is functioning optimally
                     or facing disruptions.
                  
                  For LCDPs, resource allocation is more than a technical issue; it directly impacts
                     user experience. Nonlinear Programming (NLP) helps optimize these workflows, reducing
                     execution time and improving system reliability. By analyzing this dataset, the study
                     demonstrates how NLP can bring structure and efficiency to chaotic resource demands.
                  
                  This dataset plays a crucial role in enhancing software reliability and optimizing
                     workflows. For PSO and feature selection, it uncovers the hidden factors that influence
                     reliability, enabling LCDPs to automate the identification and prioritization of critical
                     metrics. In the realm of NLP and workflow optimization, the dataset demonstrates how
                     intelligent resource allocation transforms an overburdened system into an efficient,
                     well-oiled machine. By improving performance and reducing costs, it directly benefits
                     LCDP users, ensuring smoother operations and enhanced productivity.
                  
                  
                        
                        
Fig. 2. Workflow Optimization
                      
                
               
                     3.4 Methodology Using PSO and Nonlinear Programming
                  In this paper, we employ Particle Swarm Optimization (PSO) and Nonlinear Programming
                     (NLP) for resolving two main problems in LCDPs.
                  
                  PSO is used for optimizing the features that should be added to the application depending
                     on the user's demand, taking into account the constraints of cost or resources.
                  
                  As we are aware, NLP improves workflow optimization in our processes and enables processes
                     to run with less execution time while taking into account constraints (like dependencies
                     and resource availability).
                  
                  The authors make use of two datasets for their study:
                  the feature dataset, including simulated or real-world data containing the application
                     features with their associated costs, development times, and user priority scores,
                     and a second graph-based representation of the application workflows, the workflow
                     dataset, where nodes represent tasks or components and edges represent dependencies.
                  
                  Let's demystify the techniques we are using—Particle Swarm Optimization (PSO) and
                     Nonlinear Programming (NLP)—in more human terms to understand better. These are powerful
                     tools, but they can feel scary without context. Don’t think of them as disembodied
                     algorithms but as intelligent solvers with a job to do in our methodology.
                  
                  
                        3.4.1 Particle Swarm Optimization (PSO)
                     Particle Swarm Optimization (PSO) is a metaheuristic optimization technique inspired
                        by the collective movement of birds searching for food. Each potential solution, called
                        a particle, represents a candidate feature subset for predicting software reliability.
                        The particles explore the solution space by adjusting their positions based on their
                        own previous best performance (pbest) and the best performance among their neighbors
                        (gbest). These birds are talking to each other, exchanging intel, and altering their
                        courses depending on what others find.
                     
                     Mathematically, the position x and velocity v of a particle i in the search space
                        are updated as follows:
                     
                     
                     
                     In this formulation, $w$ represents the inertia weight, which controls the balance
                        between exploration and exploitation. The parameters $c_{1}$ and $c_{2}$ are acceleration
                        coefficients that influence how much a particle is attracted to its personal best
                        position and the global best position, respectively. The terms $r_{1}$ and $r_{2}$
                        are random numbers between 0 and 1, introducing a stochastic element to the search
                        process. Each particle maintains a personal best solution, denoted as $pbest_{i}$,
                        while the globally best-performing solution among all particles is referred to as
                        $pbest_{i}$. Through successive iterations, the swarm moves towards the optimal feature
                        subset, effectively reducing redundancy and enhancing predictive accuracy.
                     
                     Compared to Genetic Algorithms (GA), Particle Swarm Optimization (PSO) emphasizes
                        collective learning and convergence toward an optimal solution, whereas GA leverages
                        mutation and crossover to maintain diversity and explore a broader solution space.
                        A potential avenue for future research involves hybridizing PSO and GA, combining
                        PSO’s fast convergence with GA’s genetic diversity to further improve optimization
                        efficiency.
                     
                     As the swarm evolves over time, the best combination of features is retained, optimizing
                        our machine learning model. For instance, the algorithm could find that “Defect Density”
                        and “MTBF” are strong predictors of software reliability, while other characteristics
                        such as “Cyclomatic Complexity” are somewhat less helpful.
                     
                     By forging a kind of democratic approval between ourselves and our predictors, this
                        process streamlines our data selection process and signals for us the most significant
                        predictors, resulting in a cleaner predictive model.
                     
                   
                  
                        3.4.2 Nonlinear Programming (NLP)
                     Now, let’s turn to nonlinear programming. Nonlinear Programming (NLP) is a mathematical
                        optimization technique used to find the best possible outcome in complex systems under
                        constraints. Unlike linear programming (LP), where relationships are linear, NLP allows
                        for nonlinear relationships among variables, making it ideal for optimizing computational
                        workflows in LCDPs. Imagine managing a factory where resources like CPUs and memory
                        must be allocated efficiently while navigating constraints such as deadlines, costs,
                        and availability. This is where optimization techniques excel.
                     
                     Mathematical Formulation: Suppose we want to minimize execution time while ensuring
                        CPU and memory constraints are met:
                     
                     
                     Subject to:
                     
                     
                     In this formulation, $C_{i}$ represents the computational cost of task $i$, while
                        $T_{i}$ denotes its execution time. The variables $U_{i}$ and $M_{i}$ correspond
                        to the CPU and memory usage for each task, whereas $U_{\max}$ and $M_{\max}$ define
                        the maximum available resources within the system.
                     
                     Nonlinear Programming (NLP) offers several advantages for Low-Code Development Platforms
                        (LCDPs), including the ability to automatically balance resource allocation, dynamically
                        adjust priorities to reduce execution time, and ensure that high-priority jobs receive
                        adequate CPU and memory.
                     
                     When compared to other optimization methods, NLP distinguishes itself from Linear
                        Programming (LP), which is effective for simple problems but lacks the flexibility
                        to handle complex workflows. Unlike heuristic approaches such as Particle Swarm Optimization
                        (PSO) and Genetic Algorithms (GA), which perform well for large-scale problems but
                        do not always guarantee an optimal solution, NLP provides an exact mathematical optimization
                        framework that makes it particularly suitable for workflow scheduling in LCDPs.
                     
                     In workflow optimization, the process begins by understanding constraints—defining
                        the boundaries within which solutions must operate, such as available memory or computational
                        power. Next, objectives are weighed, whether it's minimizing failure rates, maximizing
                        resource utilization, or distributing tasks efficiently across different sites. Finally,
                        through iterative calculations, the best possible allocation of resources is determined,
                        ensuring maximum efficiency and reliability.
                     
                     For instance, in an "AI-Based Job Site Matching" dataset, this method helps assign
                        jobs to sites in a way that reduces failures while optimizing resource usage. By managing
                        multiple competing objectives, it ensures an intelligent, balanced workflow that enhances
                        system performance.
                     
                   
                  
                        3.4.3 The Impact on Software Optimization
                     The former is an explorer that discovers the most useful nuggets of information from
                        our data, and the latter serves as a planner that applies that insight in real-world
                        settings. Taken together, they address two central problems in our research.
                     
                     Secondly, feature selection helps in obtaining focused analysis along with extracting
                        relevant information from a vast amount of data. Second, the selected features and
                        workflows lead to practical improvements by optimizing decision-making, which improves
                        performance and reliability.
                     
                     These techniques do not just supplement each other; they also capture the innovative
                        spirit of low-code development platforms (LCDPs)—promoting efficiency, intelligence,
                        and access in software development.
                     
                   
                  
                        3.4.4 Feature Selection Using Particle Swarm Optimization (PSO)
                     To improve the prediction of software reliability, I used Particle Swarm Optimization
                        (PSO) for selecting the most important features from a dataset with 31 variables.
                        PSO is inspired by the behavior of swarms, like flocks of birds or schools of fish.
                        Each “particle” in the swarm represents a possible solution—in this case, a subset
                        of features—and moves around the solution space to find the best combination.
                     
                     The dataset consisted of software-related attributes, including defect density, code
                        coverage, cyclomatic complexity, and MTBF (mean time between failures). The target
                        variable, reliability class, represented software reliability levels as low, medium,
                        or high.
                     
                     To prepare the data, feature values were normalized between 0 and 1, ensuring all
                        attributes contributed equally. The target variable was converted into numerical labels
                        to make it compatible with machine learning models.
                     
                     The PSO algorithm explored different subsets of features, evaluating each using a
                        Decision Tree Classifier. Its objective was to maximize classification accuracy by
                        refining feature selection. Over 50 iterations, the swarm adapted and improved, learning
                        from the best-performing subsets.
                     
                     Once PSO completed its process, the selected features were tested with a Random Forest
                        classifier. The performance was then compared to models using all available features
                        to assess the effectiveness of feature selection.
                     
                   
                  
                        3.4.5 Workflow Optimization Using Nonlinear Programming (NLP)
                     For workflow optimization, I used Nonlinear Programming (NLP) to allocate computational
                        resources—such as CPUs, memory, and disk space—more efficiently. The aim was to minimize
                        the total execution time for tasks while ensuring that resource limits were respected.
                     
                     The dataset captured resource usage across more than 400,000 hours of computational
                        jobs, tracking key metrics such as total CPUs, total memory, and total disk.
                     
                     The goal was to allocate resources efficiently, minimizing execution time while ensuring
                        that no task exceeded the available resources. This optimization problem was framed
                        to achieve the shortest possible execution time while adhering to resource constraints.
                     
                     The solver iteratively adjusted allocations to determine the optimal distribution.
                        Tasks with higher computational demands were assigned more resources, while smaller
                        tasks were grouped efficiently to maximize utilization.
                     
                     Once optimization was complete, the results were validated to ensure that no task
                        exceeded resource limits and that all jobs were completed within their allocated time,
                        confirming the effectiveness of the approach.
                     
                   
                
             
            
                  4. Results	
               The results of this study reflect insights gathered from a survey of software professionals,
                  an in-depth case study using the Mendix platform, and data from existing literature.
                  Together, these findings highlight the perceived advantages and challenges of Low
                  Code Development Platforms (LCDPs) and offer empirical evidence for their effectiveness
                  in accelerating development timelines and empowering non-developers.
               
               
                     4.1 Survey Findings
                  The survey results provide a quantitative overview of software professionals’ perceptions
                     of LCDPs, focusing on development speed, cost efficiency, scalability, and integration
                     capabilities. Respondents were asked to rate their experiences across various factors,
                     which are summarized in Table 1.
                  
                  From Table 1, it is worth noting that the vast majority of respondents (85%, CI: 82%-88%) agree
                     or strongly agree that LCDPs help to speed up development, and a good number noted
                     their ability to lower costs (M = 4.3, SD = 0.8 on a 5-point Likert scale). However,
                     notable concerns arose with scalability and integration, with about 60% of respondents
                     indicating agreement or strong agreement with these limitations (t(149) = 2.84, p
                     < 0.01 for difference between positive and negative responses). These results indicate
                     a statistically significant preference for LCDP efficiency but also highlight measurable
                     concerns about scalability.
                  
                  This is in line with open-ended responses where many professionals indicated that
                     although LCDPs deliver speed and accessibility, they also require additional customization
                     or some form of traditional development in order to bring them up to enterprise-level
                     standards.
                  
                  
                        
                        
Table 1 Survey Responses on Key Benefits and Limitations of LCDPs
                     
                     
                           
                              
                                 | 
                                    
                                 
                                  Factor 
                                 
                               | 
                              
                                    
                                 
                                  Strongly Agree 
                                 
                               | 
                              
                                    
                                 
                                  Agree 
                                 
                               | 
                              
                                    
                                 
                                  Neutral 
                                 
                               | 
                              
                                    
                                 
                                  Disagree 
                                 
                               | 
                              
                                    
                                 
                                  Strongly Disagree 
                                 
                               | 
                           
                           
                                 | 
                                    
                                 
                                  LCDPs improve development speed 
                                 
                               | 
                              
                                    
                                 
                                  45% 
                                 
                               | 
                              
                                    
                                 
                                  40% 
                                 
                               | 
                              
                                    
                                 
                                  8% 
                                 
                               | 
                              
                                    
                                 
                                  5% 
                                 
                               | 
                              
                                    
                                 
                                  0% 
                                 
                               | 
                           
                           
                                 | 
                                    
                                 
                                  LCDPs reduce development costs 
                                 
                               | 
                              
                                    
                                 
                                  38% 
                                 
                               | 
                              
                                    
                                 
                                  42% 
                                 
                               | 
                              
                                    
                                 
                                  15% 
                                 
                               | 
                              
                                    
                                 
                                  3% 
                                 
                               | 
                              
                                    
                                 
                                  2% 
                                 
                               | 
                           
                           
                                 | 
                                    
                                 
                                  LCDPs face scalability issues 
                                 
                               | 
                              
                                    
                                 
                                  25% 
                                 
                               | 
                              
                                    
                                 
                                  35% 
                                 
                               | 
                              
                                    
                                 
                                  25% 
                                 
                               | 
                              
                                    
                                 
                                  8% 
                                 
                               | 
                              
                                    
                                 
                                  5% 
                                 
                               | 
                           
                           
                                 | 
                                    
                                 
                                  LCDPs struggle with integration 
                                 
                               | 
                              
                                    
                                 
                                  30% 
                                 
                               | 
                              
                                    
                                 
                                  40% 
                                 
                               | 
                              
                                    
                                 
                                  15% 
                                 
                               | 
                              
                                    
                                 
                                  10% 
                                 
                               | 
                              
                                    
                                 
                                  5% 
                                 
                               | 
                           
                        
                     
                   
                  
                        
                        
Fig. 3. Survey Responses on LCDPs
                      
                  Figure 3 visually represents the survey responses regarding LCDP attributes-development
                     speed, cost efficiency, scalability, and integration-using a Likert scale (Strongly
                     Disagree to Strongly Agree). Notably, approximately 85% of respondents indicated that
                     LCDPs significantly improve development speed and lower costs, while around 60% expressed
                     concerns about scalability and integration. This figure corroborates the data summarized
                     in Table 1. and highlights that although LCDPs offer clear benefits in rapid and cost-effective
                     development, challenges remain in scaling these platforms for enterprise-level applications.
                  
                
               
                     4.2 Case Study Analysis
                  The case study involving the development of a business application on the Mendix platform
                     further illuminates the practical benefits and challenges of using LCDPs. Key metrics
                     such as development time, cost, and adaptability were compared to a similar project
                     completed through traditional coding methods.
                  
                  
                        
                        
Fig. 4. Comparison of Traditional Development vs. LCDP
                      
                  
                        
                        
Table 2 Comparison of Development Time and Cost Between LCDP and Traditional Development
                     
                     
                           
                              
                                 | 
                                    
                                 
                                  Metric 
                                 
                               | 
                              
                                    
                                 
                                  Mendix Platform LCDPs 
                                 
                               | 
                              
                                    
                                 
                                  Traditional Development 
                                 
                               | 
                           
                           
                                 | 
                                    
                                 
                                  Development Time 
                                 
                               | 
                              
                                    
                                 
                                  3 weeks 
                                 
                               | 
                              
                                    
                                 
                                  6 weeks 
                                 
                               | 
                           
                           
                                 | 
                                    
                                 
                                  Development Cost 
                                 
                               | 
                              
                                    
                                 
                                  $15,000 
                                 
                               | 
                              
                                    
                                 
                                  $30,000 
                                 
                               | 
                           
                           
                                 | 
                                    
                                 
                                  Customization Effort 
                                 
                               | 
                              
                                    
                                 
                                  Moderate 
                                 
                               | 
                              
                                    
                                 
                                  High 
                                 
                               | 
                           
                           
                                 | 
                                    
                                 
                                  Adaptability to New Requirements 
                                 
                               | 
                              
                                    
                                 
                                  Moderate 
                                 
                               | 
                              
                                    
                                 
                                  High 
                                 
                               | 
                           
                        
                     
                   
                  Table 2 reveals that the Mendix platform enabled a faster and more cost-effective development
                     cycle, reducing time and costs by 50%. However, while the LCDP project allowed for
                     moderate customization, the limitations became evident when adapting the application
                     to evolving requirements. For instance, although basic features were easily implemented,
                     more specific functionalities required additional development effort or external integration,
                     which traditional coding could handle more fluidly.
                  
                  In the healthcare industry, LCDPs have been adopted to quickly develop patient management
                     applications. In a recent study comparing OutSystems (an LCDP) with traditional coding
                     in a hospital management system, development time was reduced by 55%, but data security
                     concerns required additional backend custom development. This further highlights the
                     trade-offs between rapid development and system robustness.
                  
                  These findings indicate that while LCDPs provide significant advantages in terms of
                     development speed and cost reduction, they also introduce trade-offs in areas such
                     as customization, security, and scalability. Organizations must carefully evaluate
                     their project requirements and constraints before choosing an LCDP-based approach.
                  
                  The bar chart Fig. 4(a) shows a stark difference in development time (in weeks) and development cost (scaled
                     in tens of thousands of dollars) between traditional coding and an LCDPs-based approach
                     based on Mendix. Here, traditional development takes approximately six weeks and costs
                     around $30,000, whereas the LCDPs approach halves both of those numbers to three weeks
                     and about $15,000. This image highlights a 50% reduction in time and cost, which,
                     as found in the study, supports LCDP’s ability to drastically reduce development cycle
                     durations and expenses.
                  
                  
                        
                        
Fig. 5. Ease-of-Use Ratings for LCDPs
                      
                  The radar chart Fig. 4(b) provides a wider qualitative approach to compare four important factors, namely development
                     cost, development time, adaptability, and customization effort between LCDPs (orange
                     polygon) and traditional coding (blue polygon). There is another one for LCDPs, which
                     is generally smaller in its overall shape than its larger counterparts; this means
                     it is a leaner process in everything it entails: cost, time to market, and effort
                     in customization. But in terms of the traditional approach, it scores better in adaptability,
                     which means that while LCDPs are great for quick, cost-effective projects, more complicated
                     or highly customized situations might still benefit from a traditional coding model.
                  
                
               
                     4.3 Observations on Usability and Accessibility
                  Both the survey and case study show that LCDPs significantly enhance usability and
                     accessibility for non-technical users. As illustrated in Fig. 5, the survey participants rated ease of use on a scale from 1 (very difficult) to
                     5 (very easy). The majority of respondents indicated “Easy” or “Very Easy,” reinforcing
                     the idea that these platforms are accessible to citizen developers. Furthermore, the
                     visual, drag-and-drop interface of Mendix in the case study enabled non-developers
                     to create meaningful components of the project. This underscores LCDPs’ role in democratizing
                     the development process by opening it up to individuals with minimal coding backgrounds.
                  
                
               
                     4.4 Limitations and Scalability Challenges
                  One of the most consistent findings from both the survey and the case study is the
                     issue of scalability. Survey participants expressed concerns about LCDPs’ ability
                     to handle large, complex projects, and these concerns were validated in the case study,
                     where the Mendix platform required significant workarounds to meet advanced functionality
                     requirements. Comments from survey respondents highlight that although LCDPs provide
                     a solid foundation for rapid development, they are often limited when scaling up to
                     enterprise-level applications due to constraints in customization and integration.
                  
                  Overall, the results indicate that LCDPs excel in accelerating development and making
                     application creation more accessible. However, as shown in both the survey data and
                     case study metrics, these platforms encounter substantial limitations in handling
                     complex, large-scale projects that require extensive customization or integration.
                     This balance of strengths and limitations provides a nuanced perspective on the role
                     of LCDPs in software development, suggesting that while they offer considerable value
                     in specific contexts, careful evaluation is needed before deploying them in complex
                     environments.
                  
                  For instance, in large financial institutions, LCDPs often struggle with regulatory
                     compliance requirements that demand extensive customization, which goes beyond their
                     default capabilities. A hybrid approach, where LCDPs handle UI development while traditional
                     coding manages backend logic, could mitigate these limitations. Additionally, companies
                     like XYZ have improved LCDP integration by implementing API gateways that standardize
                     interactions between LCDPs and legacy databases.
                  
                  This highlights the importance of strategic adoption of LCDPs, where businesses must
                     assess whether LCDPs can fully meet their customization and integration needs or if
                     a hybrid approach is required. While LCDPs streamline front-end and workflow development,
                     integrating them into enterprise IT ecosystems requires additional considerations
                     such as data security, compliance, and API standardization.
                  
                  
                
               
                     4.5 Using PSO and Nonlinear Programming
                  
                        4.5.1 Feature Selection Using PSO
                     PSO identified 15 key features from the original 31, highlighting critical metrics
                        such as defect density, MTBF, code coverage, and user satisfaction. By narrowing the
                        focus to this reduced feature set, the model achieved greater efficiency while maintaining
                        high prediction accuracy.
                     
                     Figure 6 displays the detailed output of the PSO process for feature selection. The
                        figure begins by listing the cleaned dataset columns, then shows that the algorithm
                        reached its maximum iteration limit of 50. It proceeds to list the selected features
                        – including important metrics such as Cyclomatic Complexity, Halstead Volume, and
                        Code Coverage – which are retained to optimize the model. The output highlights a
                        best accuracy of 99.68%, with the Random Forest classifier achieving 98.72% accuracy
                        using all features versus 99.36% with the selected features. Additionally, the cross-validation
                        scores (mean CV accuracy of 98.77%) and a significantly reduced training time of 0.18
                        seconds underscore the efficiency and effectiveness of the feature selection process.
                     
                     With all features included, the model reached an accuracy of 98.72%, but with the
                        selected features, accuracy slightly improved to 98.93%. Additionally, training time
                        was reduced by 40%, demonstrating the effectiveness of feature selection.
                     
                     Among the selected attributes, defect density, MTBF, and code coverage emerged as
                        the most influential factors in predicting software reliability. Below is a visualization
                        illustrating the importance of these features in the final model as shown in Fig. 7.
                     
                     While PSO efficiently selects features for software reliability prediction, alternative
                        approaches like Genetic Algorithms (GA) have also been used for similar tasks. Unlike
                        PSO, which relies on particle movement to explore solutions, GA uses evolutionary
                        principles to optimize feature selection. Future research could explore hybrid approaches
                        combining PSO and GA to further improve efficiency.
                     
                     
                           
                           
Fig. 7. Feature Importance of Selected Features (PSO)
                         
                     
                   
                  
                        4.5.2 Workflow Optimization Using NLP
                     The optimization process significantly reduced total execution time by 16.8% (95%
                        CI: 15.2%–18.4%), bringing it down from 2.22 trillion units to 1.85 trillion units
                        (SD = 0.12 trillion units across multiple runs). A paired t-test (t = 5.42, p < 0.001)
                        confirmed that the observed improvements in execution time were statistically significant,
                        validating the effectiveness of the NLP optimization method.Tasks were categorized
                        based on CPU usage, with 911 classified as low CPU tasks, 594 as medium CPU tasks,
                        and 49 as high CPU tasks.
                     
                     Figure 8 provides the raw output of the optimization process. It also features a dataset
                        overview providing summary statistics (time, total CPUs, total memory, total disk)
                        for the original and optimized assignments across 1,554 records. In addition, the
                        figure describes the optimized total execution time, the grouping of tasks by their
                        CPU usage, and validation metrics of resource usage, reporting a total of 79,832 CPUs
                        and 4.45 trillion disk units are utilized. This is an interesting validation that
                        leads towards better allocation of resources, brings efficiency, and shows that resource-bounded
                        persistent automata were not able to satisfy full resource constraints.
                     
                     
                           
                           
Fig. 9. CPU Usage Distribution Before and After Optimization
                         
                     With respect to resource usage, the optimization redistributed computing resources
                        more effectively, as also shown in Fig. 9. This chart compares CPU usage before optimization and after, when the NLP process
                        was successfully making the load distribution across the tasks more balanced, and
                        the processing bottlenecks were minimized.
                     
                     Similarly, Fig. 10 shows the distribution of disk usage before optimization and after optimization.
                        This visualization shows a better distribution of disk resources - something critical
                        to mitigating latencies and enhancing general performance.
                     
                     In terms of resource utilization, the optimization effectively allocated resources,
                        resulting in a total of 79,832 CPUs used and 4.45 trillion units of disk space consumed.
                        These improvements demonstrated a more efficient distribution of computational resources,
                        enhancing overall system performance.
                     
                     Include the CPU Usage Distribution and Disk Usage Distribution graphs under the subsection
                        where you discuss workflow optimization results. These graphs visualize the resource
                        allocation before and after optimization.
                     
                     Use them to explain how resources like CPUs and disk space were distributed and highlight
                        how the optimization process improved efficiency.
                     
                     
                           
                           
Fig. 10. Disk Usage Distribution Before and After Optimization
                         
                     After optimization, all tasks met resource constraints, ensuring that no task was
                        left unprocessed or under-resourced. Before optimization, inefficient resource allocation
                        created bottlenecks, slowing down execution. By intelligently distributing resources,
                        NLP accelerated the overall workflow and improved efficiency.
                     
                     Feature selection using PSO reduced the complexity of the software reliability prediction
                        model without compromising accuracy. This not only enhanced model performance but
                        also made it easier to interpret and significantly faster to train. Meanwhile, NLP
                        streamlined workflow optimization by reducing execution time and improving resource
                        utilization. This scalable approach can be applied to larger datasets and more complex
                        workflows in the future.
                     
                     The combination of PSO for feature selection and NLP for workflow optimization demonstrated
                        the power of these techniques in solving real-world challenges. While PSO simplified
                        predictive modeling by identifying the most relevant features, NLP enhanced computational
                        efficiency by optimizing resource allocation. Together, they form a comprehensive
                        framework for improving both data-driven decision-making and operational performance
                        in software systems.
                     
                   
                
             
            
                  5. Conclusions	
               In this paper, we show that LCDPs are not just an evolutionary step in software engineering
                  but a revolutionary technology that will change the way applications are built. We
                  have succeeded in discovering that LCDPs help accelerate software development, reducing
                  development time and cost by 50%, and democratizing the creation of applications for
                  non-technical users. This research points out vital aspects that are pressing issues,
                  especially with regard to scaling and integration, which are still major challenges.
                  The use of high-level optimization techniques (PSO, NLP) in this work not only provided
                  notable efficiency changes (e.g., a 16.8% reduction in execution time) but also allowed
                  us to use a framework for IQA to maximize lean project size by allocating project
                  resources to projects that require them the most. Additional opportunities were explored
                  to further improve real-world performance using risk segmentation of tasks with heavy
                  (low, medium, and high) CPU utilization.
               
               LCDPs will play an increasingly important role in the developer ecosystem, meeting
                  the needs of organizations seeking to balance the speed and flexibility of rapid development
                  projects with the responsibilities of delivering enterprise-class functionality. These
                  factors underscore the need for future research to improve these platforms as they
                  address their current challenges such as scalability, security, and integration, and
                  help realize the full potential of LCDPs. As these platforms evolve, we can expect
                  to see significant innovation in software development, where both techies and non-techies
                  can join the bandwagon of digital transformation.
               
               In summary, LCDPs today provide significant speed and cost advantages, but their true
                  potential will only be realized when they are harmoniously integrated with traditional
                  coding practices and more sophisticated optimization methodologies.
               
             
          
         
            
                  Acknowledgements
               
                  This work was supported by Korea Institute for Advancement of Technology (KIAT) grant
                  funded by the Korea Government (MOTIE) (RS-2022-KI002562, HRD Program for Industrial
                  Innovation)
                  
                  			
               
             
            
                  
                     References
                  
                     
                        
                        A. Rocha, Proceedings of CISTI’2022: 17th Iberian Conference on Information Systems
                           and Technologies, AISTI, 2022. https://doi.org/10.23919/CISTI54924.2022.9820363

 
                      
                     
                        
                        N. Rao, J. Tsay, K. Kate, V. J. Hellendoorn and M. Hirzel, “AI for low-code for AI,”
                           arXiv, 2023. [Online]. Available: http://arxiv.org/abs/2305.20015.

 
                      
                     
                        
                        R. Picek, “Low-code/no-code platforms and modern ERP systems,” 2023 9th International
                           Conference on Information Management (ICIM), IEEE, pp. 44–49, 2023. https://doi.org/10.1109/ICIM58774.2023.00014

 
                      
                     
                        
                        G. Juhas, L. Molnar, A. Juhasova, M. Ondrisova, M. Mladoniczky and T. Kovacik, “Low-code
                           platforms and languages: The future of software development,” 20th IEEE International
                           Conference on Emerging eLearning Technologies and Applications (ICETA 2022), IEEE,
                           pp. 286–293, 2022. https://doi.org/10.1109/ICETA57911.2022.9974697

 
                      
                     
                        
                        L. Tang, “ERP low-code cloud development,” IEEE International Conference on Software
                           Engineering and Service Sciences (ICSESS), IEEE, pp. 319–323, 2022. https://doi.org/10.1109/ICSESS54813.2022.9930146

 
                      
                     
                        
                        R. Waszkowski, “Low-code platform for automating business processes in manufacturing,”
                           IFAC-PapersOnLine, Elsevier B.V., vol. 52,  pp. 376–381, 2019. https://doi.org/10.1016/j.ifacol.2019.10.060

 
                      
                     
                        
                        D. Di Ruscio, D. Kolovos, J. de Lara, A. Pierantonio, M. Tisi and M. Wimmer, “Low-code
                           development and model-driven engineering: Two sides of the same coin?,” Software and
                           Systems Modeling, vol. 21, pp. 437–446, 2022. https://doi.org/10.1007/s10270-021-00970-2

 
                      
                     
                        
                        Y. Luo, P. Liang, C. Wang, M. Shahin and J. Zhan, “Characteristics and challenges
                           of low-code development: The practitioners’ perspective,” International Symposium
                           on Empirical Software Engineering and Measurement, IEEE, pp. 1-11, 2021. https://doi.org/10.1145/3475716.347578

 
                      
                     
                        
                        C. H. Wang and K. C. Wu, “A preliminary study on interdisciplinary programming learning
                           based on cloud computing low-code development platform,” Proceedings of the International
                           Conference on Computer and Applications (ICCA 2022), IEEE, pp. 1-5, 2022. https://doi.org/10.1109/ICCA56443.2022.10039663

 
                      
                     
                        
                        A. C. Bock and U. Frank, “Low-code platform,” Business and Information Systems Engineering,
                           vol. 63, pp. 733–740, 2021. https://doi.org/10.1007/s12599-021-00726-8

 
                      
                     
                        
                        K. Rokis and M. Kirikova, “Exploring low-code development: A comprehensive literature
                           review,” Complex Systems Informatics and Modeling Quarterly, no. 36, pp. 68–86, 2023.
                           https://doi.org/10.7250/csimq.2023-36.04

 
                      
                     
                        
                        E. Martinez and L. Pfister, “Benefits and limitations of using low-code development
                           to support digitalization in the construction industry,” Automation in Construction,
                           vol. 152, pp. 104909, 2023. https://doi.org/10.1016/j.autcon.2023.104909

 
                      
                     
                        
                        A. Butting, T. Greifenberg, K. Hölldobler and T. Kehrer, “Model and data differences
                           in an enterprise low-code platform,” 2023 ACM/IEEE International Conference on Model
                           Driven Engineering Languages and Systems Companion (MODELS-C), pp. 868–877, 2023.
                           https://doi.org/10.1109/MODELS-C59198.2023.00137

 
                      
                     
                        
                        B. Schenkenfelder, C. Salomon, G. Buchgeher, R. Schossleitner and C. Kerl, “The potential
                           of low-code development in the manufacturing industry,” 2023 IEEE 28th International
                           Conference on Emerging Technologies and Factory Automation (ETFA), pp. 1–8, 2023.
                           https://doi.org/10.1109/ETFA54631.2023.10275503

 
                      
                     
                        
                        V. Phalake, S. Joshi, K. Rade and V. Phalke, “Modernized application development using
                           optimized low-code platform,” 2022 2nd Asian Conference on Innovation in Technology
                           (ASIANCON), IEEE, pp. 1–4, 2022. https://doi.org/10.1109/ASIANCON55314.2022.9908726

 
                      
                     
                        
                        G. Daniel, J. Cabot, L. Deruelle and M. Derras, “Xatkit: A multimodal low-code chatbot
                           development framework,” IEEE Access, vol. 8, pp. 15332–15346, 2020. https://doi.org/10.1109/ACCESS.2020.2966919

 
                      
                     
                        
                        C. Chuanjian, G. Shuze and W. Hua, “Research on software development based on low-code
                           technology,” 2023 2nd International Conference on Artificial Intelligence and Autonomous
                           Robot Systems (AIARS), IEEE, pp. 210–213, 2023. https://doi.org/10.1109/AIARS59518.2023.00049

 
                      
                     
                        
                        P. M. Gomes and M. A. Brito, “Low-code development platforms: A descriptive study,”
                           2022 17th Iberian Conference on Information Systems and Technologies (CISTI), pp.
                           1–4, 2022. https://doi.org/10.23919/CISTI54924.2022.9820354

 
                      
                     
                        
                        A. Trigo, J. Varajão and M. Almeida, “Low-code versus code-based software development:
                           Which wins the productivity game?,” IT Professional, vol. 24, pp. 61–68, 2022. https://doi.org/10.1109/MITP.2022.3189880

 
                      
                     
                        
                        S. Käss, S. Strahringer and M. Westner, “A multiple mini case study on the adoption
                           of low-code development platforms in work systems,” IEEE Access, vol. 11, pp. 118762–118786,
                           2023. https://doi.org/10.1109/ACCESS.2023.3325092

 
                      
                   
                
             
            저자소개
            
            He received a B.S. degree in Computer Science from the International Information Technology
               University (IITU) in 2022 and is currently pursuing an M.S. degree in Software Engineering
               at Kazakh-British Technical University (KBTU), with expected completion in 2025. He
               is currently working as a Frontend Developer at Shanghai Looktook Technology Co.,
               Ltd., where he focuses on iOS application development using React Native. He has previous
               experience in backend development with Django and Python, as well as full-stack web
               development using technologies such as React, Django, and PostgreSQL. His research
               interests include low-code development platforms, artificial intelligence applications,
               and augmented/virtual reality game development.
            
            
            He received his Ph.D. in Economics and has over 21 years of professional experience
               spanning higher education, finance, government, IT, retail, and industry sectors.
               Since September 2023, he has been serving as an Associate Professor at the School
               of Information Technology and Engineering, Kazakh-British Technical University. His
               research interests include data governance, business intelligence, financial risk
               analysis, and advanced analytics. He has practical experience in launching data warehouses
               and integrating data-driven solutions, and holds certifications from Microsoft, AWS,
               and iOS Academy.
            
            
            She received the B.S., M.S., and PhD. degrees from Satbayev University, Almaty, Kazakhstan,
               in 2004, 2014, and 2020, respectively. In September 2023, she joined Kazakh-British
               Technical University, where she is currently an professor in School of Information
               Technology and Engineering. Big Data, cyber security, machine learning, and comparative
               study of deep learning methods.
            
            
            He received the M.S. degree in the Seoul National University of Science and Technology
               in 2023, and currently studying for Ph.D. in the Korea National University of Transportation.
               He joined the Korea Railroad Administration in November 1998 and was transferred to
               the Korea Railroad Corporation in January 2005, and am currently in charge of SCADA
               as an electrical controller at the Railway Traffic Control Center. 
            
            
            He received his B.S., M.S., and Ph.D. degrees in electronic engineering from Chung-Ang
               University, Seoul, Korea, in 1995, 1997, and 2002, respectively. In March 2008, he
               joined the Korea National University of Transportation, Republic of Korea, where he
               currently holds the position of Professor in the Department of Transportation System
               Engineering, the Department of SMART Railway System, and the Department of Smart Railway
               and Transportation Engineering.