Abstract
In this paper, we present the results of research in constructing criteria for qualitative assessment of domain specific languages. Proposed compound metric consists of three parts: classic grammar size metrics, for evaluation programming language as a software and for evaluating conciseness. Each part of compound metric evaluates different characteristics needed for assessment of modern domain specific languages. To process evaluation automated software was developed. Research also contains explanation of marks for each component and their influence on the quality of accessing programming language. The analysis of assessment results provided required information for further improvements which will be presented as a part of system for computer-aided design of domain specific languages. Achieved compound metric could be used in assessment of domain specific programming languages of any types and can eliminate human provided mistakes in some areas of design.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Alaca, O.F., Tezel, B.T., Challenger, M., Goulão, M., Amaral, V., Kardas, G.: Agentdsm-eval: a framework for the evaluation of domain-specific modeling languages for multi-agent systems. Comput. Stand. Interf. 76, 103513 (2021). https://doi.org/10.1016/j.csi.2021.103513
Fernau, H., Kuppusamy, L., Oladele, R.O., Raman, I.: Improved descriptional complexity results on generalized forbidding grammars. Disc. Appl. Math. (2021). https://doi.org/10.1016/j.dam.2020.12.027
Holmes, J., Ahmed, I., Brindescu, C., Gopinath, R., Zhang, H., Groce, A.: Using relative lines of code to guide automated test generation for python. ACM Trans. Softw. Eng. Methodol. (TOSEM) 29(4), 1–38 (2020). https://doi.org/10.1145/3408896
Munir, K., Sheraz Anjum, M.: The use of ontologies for effective knowledge modelling and information retrieval. Appl. Comput. Inf. 14(2), 116–126 (2018)
Muñoz Barón, M., Wyrich, M., Wagner, S.: An empirical validation of cognitive complexity as a measure of source code understandability, pp. 1–12 (2020). https://doi.org/10.1145/3382494.3410636
Parr, T., Fisher, K.: Ll(*): The foundation of the antlr parser generator. In: Proceedings of the ACM SIGPLAN Conference on Programming Language Design and Implementation (PLDI), pp. 425–436 (2011)
Peitek, N., Apel, S., Parnin, C., Brechmann, A., Siegmund, J.: Program comprehension and code complexity metrics: an fmri study, pp. 524–536 (2021). https://doi.org/10.1109/ICSE43902.2021.00056
Ruiz-Rube, I., Person, T., Dodero, J.M., Mota, J.M., Sánchez-Jara, J.M.: Applying static code analysis for domain-specific languages. Softw. Syst. Model. 19(1), 95–110 (2019). https://doi.org/10.1007/s10270-019-00729-w
Terragni, V., Salza, P., Pezzè, M.: Measuring software testability modulo test quality, pp. 241–251 (2020). https://doi.org/10.1145/3387904.3389273
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Ocheretianyi, O., Baklan, I. (2022). Assesment Model for Domain Specific Programming Language Design. In: Babichev, S., Lytvynenko, V. (eds) Lecture Notes in Computational Intelligence and Decision Making. ISDMCI 2021. Lecture Notes on Data Engineering and Communications Technologies, vol 77. Springer, Cham. https://doi.org/10.1007/978-3-030-82014-5_53
Download citation
DOI: https://doi.org/10.1007/978-3-030-82014-5_53
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-82013-8
Online ISBN: 978-3-030-82014-5
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)