LoCoML
LoCoML
analysis for ASR + TTS task, and how it scales as the number of Machine Learning Designer enables users to create pipelines
models increases. Our initial results demonstrate that LoCoML’s via a drag-and-drop interface, integrating seamlessly within
overhead increases linearly with the number of models, scaling Azure’s ecosystem. Similarly, AWS SageMaker Pipelines
from 52.57ms for a single MT model to 845.38ms for 16 offers a complete suite for creating, deploying, and managing
models, while remaining negligible at only 1.8% of the total workflows, tightly coupled with AWS-native services. However,
runtime. This linear overhead growth, coupled with the fact these platforms often face challenges when dealing with custom
that 98.2% of the execution time is spent on model inference, models from external sources, primarily because they are largely
indicates that our platform introduces minimal performance designed to operate within their respective ecosystems [22]
impact while effectively managing complex ML pipelines. [23]. It is too tedious to have to port custom models into the
specific input-output constraints as prescribed by these platforms.
V. R ELATED W ORK In contrast, our framework addresses this gap by providing
Recent advances in MDE have highlighted the role of low- a unified and adaptable solution capable of accommodating
code platforms in simplifying ML development and deployment. diverse, partner-contributed models, ensuring compatibility and
Naveed et al. [5] recommend researchers and practitioners to seamless integration—capabilities that are currently absent in
develop low-code platforms for systems with ML components to these existing platforms.
make ML capabilities more accessible to non-experts, as these
platforms can significantly reduce development complexity and VI. C ONCLUSION AND F UTURE W ORK
time to deployment. Similarly, Iyer et al. [9] introduced Trinity, a
no-code platform specifically designed to handle complex spatial To conclude, we have introduced LoCoML, a low code
datasets, highlighting the versatility and scalability low-code framework designed to streamline the development of ML
solutions bring to ML applications. In addition, Esposito et inference pipelines. LoCoML has been successfully integrated
al. [13] emphasize the importance of user-configurable controls into the Bhashini Project with a drag-and-drop interface to
within AI systems, proposing that a balance between automation create pipelines, where it operates in a real-world production
and manual adjustments can address diverse user needs environment, supporting users in building and managing inference
effectively. Sahay et al. [14] provide a detailed survey of various pipelines efficiently. The framework’s simple interface allows
low-code development platforms, identifying key features such users, including those without extensive coding skills, to connect
as graphical interfaces, interoperability, and scalability as critical and control various ML models seamlessly. Our evaluation across
for decision-makers evaluating such platforms. multiple scenarios, including TTS, MT, and ASR, indicate that
LoCoML builds on these principles, offering a flexible, LoCoML has significantly simplified the process of constructing
user-friendly platform that enables non-expert users to construct complex, multimodel workflows, making ML pipeline develop-
and customize ML pipelines, facilitating tasks like data ment more accessible and practical for a diverse range of users.
preprocessing, model training, and inference without extensive Looking ahead, we aim to expand LoCoML’s capabilities in
coding knowledge [15]. This approach empowers users to response to evolving requirements within the Bhashini Project.
iteratively reconfigure workflows, bridging adaptability gaps Stakeholders of this project are exploring the potential of
noted in previous studies and enhancing both accessibility and extending the framework to support model training, thus creating
control [16] [17]. Unlike traditional ML systems, LoCoML an end-to-end solution covering both training and inference within
allows users to adjust pipelines dynamically, aligning with the same pipeline. Additionally, we plan to conduct further studies
recommendations for integrating user-centric features and to assess LoCoML’s effectiveness in terms of user experience,
configurable controls within ML platforms [18] [19]. usability, and performance. These future enhancements will
Existing platforms like Azure Machine Learning Designer [20] ensure that LoCoML continues to evolve as a versatile and robust
and AWS SageMaker Pipelines [21] provide robust solutions framework, meeting the growing demands of ML practitioners
for building and managing machine learning workflows. Azure and researchers within the Bhashini Project and beyond.
ACKNOWLEDGMENT [21] Amazon, “Amazon SageMaker Introduction.” https://docs.aws.amazon.
com/sagemaker/latest/dg/use-auto-ml.html?icmpid=docs sagemaker lp/
The authors acknowledge the anonymous reviewers for index.html, 2024. Accessed: 2024-11-16.
their valuable feedback. The authors thank Ayush Agrawal, [22] Microsoft, “How to Deploy Models in Azure Machine Learning
Designer.” https://learn.microsoft.com/en-us/azure/machine-learning/
Harshit Karwal, Mukta Chanda, Rohan Chowdary, Shashwat how-to-deploy-model-designer?view=azureml-api-1, 2024. Accessed:
Dash, Siddharth Mavani, and Supreeeth S Karan, for their 2024-11-16.
assistance in developing the code artifacts necessary to build [23] Amazon, “Bring Your Own Model with Amazon SageMaker
Script Mode.” https://aws.amazon.com/blogs/machine-learning/
this framework. The authors would also like to acknowledge bring-your-own-model-with-amazon-sagemaker-script-mode/, 2021.
Bhashini Engineering Unit 3 team for the support. Accessed: 2024-11-16.
R EFERENCES
[1] G. A. Lewis, H. Muccini, I. Ozkaya, K. Vaidhyanathan, R. Weiss, and
L. Zhu, “Software architecture and machine learning (dagstuhl seminar
23302),” 2024.
[2] S. Amershi, A. Begel, C. Bird, R. Deline, H. Gall, E. Kamar, N. Nagappan,
B. Nushi, and T. Zimmermann, “Software engineering for machine
learning: A case study,” pp. 291–300, 05 2019.
[3] H. Muccini and K. Vaidhyanathan, “Software architecture for ml-based
systems: What exists and what lies ahead,” in 2021 IEEE/ACM 1st
Workshop on AI Engineering-Software Engineering for AI (WAIN),
pp. 121–128, IEEE, 2021.
[4] M. Brambilla, J. Cabot, and M. Wimmer, Model-driven software
engineering in practice. Morgan & Claypool Publishers, 2017.
[5] H. Naveed, C. Arora, H. Khalajzadeh, J. Grundy, and O. Haggag, “Model
driven engineering for machine learning components: A systematic
literature review,” Information and Software Technology, p. 107423, 2024.
[6] D. Kreuzberger, N. Kühl, and S. Hirschl, “Machine learning operations
(mlops): Overview, definition, and architecture,” 2022.
[7] A. C. Bock and U. Frank, “Low-code platform,” Business & Information
Systems Engineering, vol. 63, pp. 733–740, 2021.
[8] J. Cabot, “Positioning of the low-code movement within the field of
model-driven engineering,” in Proceedings of the 23rd ACM/IEEE
International Conference on Model Driven Engineering Languages and
Systems: Companion Proceedings, pp. 1–3, 2020.
[9] C. V. K. Iyer, F. Hou, H. Wang, Y. Wang, K. Oh, S. Ganguli, and V. Pandey,
“Trinity: A no-code ai platform for complex spatial datasets,” 2021.
[10] G. A. Lewis, I. Ozkaya, and X. Xu, “Software architecture challenges
for ml systems,” in 2021 IEEE International Conference on Software
Maintenance and Evolution (ICSME), pp. 634–638, IEEE, 2021.
[11] D. C. Schmidt et al., “Model-driven engineering,” Computer-IEEE
Computer Society-, vol. 39, no. 2, p. 25, 2006.
[12] C. Di Sipio, D. Di Ruscio, and P. T. Nguyen, “Democratizing the develop-
ment of recommender systems by means of low-code platforms,” in Pro-
ceedings of the 23rd ACM/IEEE international conference on model driven
engineering languages and systems: companion proceedings, pp. 1–9, 2020.
[13] A. Esposito, M. Calvano, A. Curci, G. Desolda, R. Lanzilotti, C. Lorusso,
and A. Piccinno, “End-user development for artificial intelligence: A
systematic literature review,” in International Symposium on End User
Development, pp. 19–34, Springer, 2023.
[14] A. Sahay, A. Indamutsa, D. Di Ruscio, and A. Pierantonio, “Supporting
the understanding and comparison of low-code development platforms,” in
2020 46th Euromicro Conference on Software Engineering and Advanced
Applications (SEAA), pp. 171–178, IEEE, 2020.
[15] J. Bosch, H. H. Olsson, and I. Crnkovic, “Engineering ai systems: A
research agenda,” Artificial intelligence paradigms for smart cyber-physical
systems, pp. 1–19, 2021.
[16] D. De Silva and D. Alahakoon, “An artificial intelligence life cycle: From
conception to production,” Patterns, vol. 3, no. 6, 2022.
[17] G. Giray, “A software engineering perspective on engineering machine
learning systems: State of the art and challenges,” Journal of Systems
and Software, vol. 180, p. 111031, 2021.
[18] M. Steidl, M. Felderer, and R. Ramler, “The pipeline for the continuous
development of artificial intelligence models—current state of research
and practice,” Journal of Systems and Software, vol. 199, p. 111615, 2023.
[19] D. Xin, E. Y. Wu, D. J.-L. Lee, N. Salehi, and A. Parameswaran, “Whither
automl? understanding the role of automation in machine learning
workflows,” in Proceedings of the 2021 CHI Conference on Human
Factors in Computing Systems, pp. 1–16, 2021.
[20] Microsoft, “Azure Machine Learning Designer (v2) Introduction.”
https://learn.microsoft.com/en-us/azure/machine-learning/
concept-designer?view=azureml-api-2, 2024. Accessed: 2024-11-16.
3 https://bhashini.gov.in/sahyogi/anushandhan-mitra/15