Keywords

1 Introduction

The COVID-19 pandemic is expected to lead to a severe recessionary economic crisis with quite negative consequences for large numbers of firms and citizens, as containment and social distancing policies adopted in most countries will on one hand ‘flatten’ the medical curve of virus spread (which is quite positive), but on the other hand will ‘steepen’ the resulting recession curve (which will have quite negative economic and social consequences) [1]. However, this is an ‘old story’: recessionary economic crises of various intensities and durations appear repeatedly in the last 100 years in market-based economies, and are recognized as one of the most severe and threatening weaknesses of them, having quite negative consequences for the whole economy and society [2,3,4,5,6,7]. The fluctuations that economic activity usually exhibits, and also some critical events, such as banking crises (like the one that gave rise to the 2008 Global Financial Crisis), epidemics (like the COVID-19 one), large increases in the prices of important goods (e.g. oil or gas), etc., can lead to significant economic recessions, meant as serious contractions of economic activity, resulting in economic crises. These contractions of economic activity can lead to big reductions of firms’ production, procurement, investment, innovation and employment [5,6,7], and also closure of numerous firms, with serious social consequences, such as increase of unemployment, poverty and social exclusion, resulting in political unrest and instability.

For these reasons economic crises constitute one of the most severe and threatening problems that governments often face and have strong pressure to deal with them effectively in order at least to avoid catastrophic consequences. So it is imperative that governments develop effective public policies and make drastic interventions for addressing these economic crises. The most usual of these interventions is the provision of various kinds of support to the most vulnerable firms in the beginning of such crises, or even earlier, when an economic crisis is in sight (i.e. when there are serious forecasts that an economic crisis is going to occur, like the ones currently made by serious economists and economic research centers that the COVID-19 pandemic will give rise to a severe recessionary economic crisis [1]). This support usually includes the provision to firms of low-interest loans, subsidies, education or consulting services, co-finance of investments or employees’ payroll, tax reductions, etc. However, since the available government resources (e.g. financial, human, etc.) for providing such support to firms are usually limited, much below than the ones required for meeting the high demands (from the numerous firms that submit applications for receiving this support), a selection among the applicant firms has to be made; it is based on some predefined criteria, which usually concern firm’s sector and location (according to existing sectoral and regional priorities), as well as sales and profits in the last 3–5 years, market share, export activities, debts, etc. It would be very good if to these criteria could be added firm’s expected vulnerability to the crisis, as an additional criterion, aiming to provide more assistance and support to the most vulnerable firms; this can increase substantially the effectiveness of such interventions, making them more focused, and improve public value generation from them towards the reduction of the negative consequences of recessionary economic crises. For the above reasons it would be quite useful if we can make a prediction of the vulnerability of individual firms to economic crisis.

Given the high importance of firm-level economic crisis vulnerability prediction it is necessary to exploit for this purpose to the highest possible extent: a) existing firms’ data from economic crisis periods in various government agencies (such as Ministries of Finance and Statistical Authorities); and b) the most sophisticated algorithms for ‘learning’ crisis vulnerability prediction models from such data, and especially the most widely used Artificial Intelligence (AI) algorithms for such ‘predictive analytics’: the Machine Learning (ML) ones [8,9,10,11,12]. In this direction our paper presents a methodology for using existing government data in order to predict the vulnerability of individual firms to economic crisis, based on their particular characteristics (concerning strategic directions, processes, human and technological resources, structures) using ML. Furthermore, a first application of the proposed methodology is presented, based on existing data from the Greek Ministry of Finance and Statistical Authority concerning 363 firms for the economic crisis period 2009–2014, which gives encouraging results.

In the following Sect. 2 the background of our methodology is outlined, while in Sect. 3 the methodology is described, followed by its abovementioned application in Sect. 4. The final Sect. 5 summarizes conclusions and proposes directions for future research.

2 Background

2.1 Economic Crises

As mentioned in the Introduction one of the most serious problems of market-based economies are the economic recessions that repeatedly appear, which cause significant disruptions and problems to the economy and the society in general, and have to be addressed by government through effective public policies including appropriate interventions [2,3,4,5,6,7]. The economic crises have negative both short-term as well as medium- and long-term consequences for the economy and the society. The short-term consequences include reductions of the demand for many goods and services, resulting in serious decrease of firms’ sales, production and profits, leading to reductions in personnel employment (thus increasing unemployment) and materials’ procurement (thus propagating the crisis towards the suppliers’ sectors, etc., spreading the crisis to large parts of the economy). Furthermore, during economic crises firms usually reduce capital investment in production equipment, ICT, buildings, etc., and also in product, service and process innovations; this reduces the degree of renewal and improvement of their equipment, products, services and operation, as well as the exploitation of emerging new technologies, which causes serious medium- and long-term consequences for their efficiency and competitiveness.

Considerable research has been conducted concerning the negative impact of such recessionary economic crises on different aspects of firms’ activities and performance, as well as the factors that affect the magnitude of this negative impact [13,14,15,16,17,18,19]. It has concluded that the above negative consequences of the economic crises differ significantly among firms, and depend critically on their individual characteristics, such as their human and technological resources, their strategic orientations, etc. Some firms have superior human and technological resources, as well as management structures, that allow them to be more efficient and effective during difficult crisis periods; these firms are able to offer higher value-for-money products and services during the crisis, and in general have higher capacity to make the required adaptations to the crisis conditions, which make them less vulnerable and more resilient to the crisis. Such firms have less negative consequences on their sales revenue, and therefore on their employment, procurement, as well as capital investment and innovation. For some other firms the opposite holds. Therefore it would be interesting to investigate the capabilities of AI/ML algorithms to predict firm’s vulnerability to recessionary economic crisis based on its particular characteristics, e.g. concerning its human and technological resources, management structures, etc.

2.2 Artificial Intelligence

Artificial Intelligence (AI) includes a group of techniques that enable computers to perform tasks of higher intelligence, approaching the human one, by learning from their environment, and then using the knowledge they have obtained from it for taking or proposing action [20,21,22]. The most representative and widely used AI techniques are definitely the Machine Learning (ML) ones [8,9,10]. They enable exploiting historic data we possess for a number of units (e.g. individuals, firms, etc.) concerning the value of an important dependent variable (usually an outcome one), and also a set of independent variables (that might be possible causes of this outcome or factors affecting it), by processing them through various algorithms, and finally extracting knowledge from them, usually having the form of a model or a set of rules, concerning the relationships between the independent variables and the dependent one (this is usually referred to as ‘model training’, while the historic data used for this purpose as ‘training data’). This knowledge can be used then on one hand for gaining deeper understanding and insight about these relationships, and on the other hand for predicting the value of the dependent variable for new units, based on the known values of the independent variables, which can be quite useful for supporting relevant decisions or optimizing actions.

Initially AI was used successfully in the private sector, and these ‘success stories’ generated high levels of interest to exploit AI techniques in the public sector as well, in order to automate or support more sophisticated mental tasks than the simpler routine ones automated or supported by the traditional operational IS of government agencies [23,24,25,26,27]. Some first research has been conducted concerning the exploitation of AI in a variety of public sector thematic domains, such as education, policing, justice, public health, transport, etc. [28,29,30,31,32,33]. However, this research has revealed and investigated only a small part of the large potential of AI use in government. Therefore, much more research is required in order to exploit more this potential, in order to reveal and investigate more opportunities and methodologies for AI exploitation in government, in a wide variety of public sector thematic domains. Our study makes a contribution in this direction, by developing a methodology for using AI (ML algorithms) in a highly important domain of government activity: the intervention of government for addressing the quite threatening recessionary economic crises that repeatedly appear in market-based economies and cause major disruptions and problems.

3 The Proposed Methodology

As mentioned in Sect. 2.1 previous research concerning the negative impact of recessionary economic crises on firms has found that its magnitude differs significantly among firms, depending on their particular characteristics, such as their human and technological resources, management structures, etc. Furthermore, previous management science research has identified the main elements of a firm that determine its performance; so we can expect that these main elements might determine to a considerable extent firm’s performance during the crisis in coping with the difficult recession conditions as well, and therefore the degree of its resilience/vulnerability to the crisis. Several conceptualizations of the main elements of a firm have been developed, the most widely validated and used among them being definitely the ‘Leavitt’s Diamond’ framework [34]. According to it the most important elements of a firm that determine its performance are: a) task (=the strategies as well as the administrative and production processes of the firm); b) people (=the skills of firm’s human resources of the firm); c) technology (=the technologies used for implementing the above processes); and d) structure (=the organization of the firm in departments, and the communication and coordination patterns them). An extension of it has been developed subsequently, which analyses the above ‘task’ element into ‘strategy’ and ‘processes’ [35]. Therefore we expect that firm’s characteristics concerning the above five main elements (strategy, processes, people, technology and structure) might be good predictors of its crisis resilience/vulnerability. Government traditionally collects large amounts of data concerning firms’ economic performance (the Taxation Authorities) and also firms’ characteristics concerning human resources, use of various technologies (e.g. ICT or various production technologies), organization, innovation and other strategic orientations, etc. (the Statistical Authorities). It is quite useful to exploit these existing government data from economic crisis periods in order to construct firm-level crisis resilience/vulnerability prediction models based on firm’s characteristics.

In particular, using government firm-level data from an economic crisis period concerning:

  • on one hand firms’ resilience/vulnerability to the crisis (e.g. data from Taxation Authorities concerning decrease of sales revenue, profitability, employment, etc. during the crisis)

  • and on the other hand characteristics of the same firms concerning the above main elements (e.g. relevant data collected from Statistical Authorities),

We can train/construct prediction models of the resilience/vulnerability of a firm to recessionary economic crisis (to be used as dependent variable Y), based on the above characteristics of it (to be used as independent variables Xi); for this purpose we can use several different alternative ML algorithms [8,9,10], such as Decision Trees, Random Forests, Gradient Boosted Trees, Support Vector Machines, Generalized Linear Modelling, etc., and finally select among them the one providing the highest prediction accuracy. The constructed model can be used for predicting the value of the resilience/vulnerability of an individual firm to future economic crisis Y (taking into account that all economic crises, despite their differences, have in common their main characteristics: economic recession, contraction of economic activity and decrease of demand for products and services of most sectors) for which we have the values of the abovementioned characteristics Xi. The structure of the proposed methodology is shown in Fig. 1.

Fig. 1.
figure 1

Structure of the proposed methodology

Our methodology will use existing government data from two main sources:

  1. I)

    Data from Taxation Authorities for a large number of firms concerning the decrease of sales revenue, profits or other measures of economic performance during the economic crisis

  2. II)

    Data from Statistical Authorities about characteristics of the same firms concerning the above five main elements of them:

    • Strategic orientations: characteristics concerning the degree of adoption of the main strategies described in relevant strategic management literature [36], such as cost leadership, differentiation, focus, innovation, export, etc.

    • Processes: characteristics of firm’s processes, such as complexity, efficiency, formality, flexibility, etc.

    • Human Resources: characteristics concerning the general education/skills level of firm’s human resources (e.g. shares of firm’s personnel having tertiary education, vocational/technical education, etc.), as well as the possession of specific skills concerning important technologies (e.g. concerning ICTs or various production technologies), the provision of various kinds of training, etc.

    • Technology: characteristics concerning the use of the main enterpsise ICTs, such as Enterprise Resource Planning (ERP), Customer Relationships Management (CRM), Supply Chain Management (SCM), Business Intelligence/Business Analytics (BI/BA) systems, Collaboration Support (CS), e-sales, etc. systems, and also social media, cloud computing, or other technologies, or the use of various production technologies.

    • Structure: characteristics concerning various aspects of the management structure of the firm, such as its main structural design (functional, product/service based, geographic, matrix), degree of differentiation, specialization, centralization-decentralization, or the use of organic structural forms (such as teamwork or job rotation) [37, 38].

Furthermore, we can also include general data about each firm, such as sector, degree of competiveness in this sector (in comparison with the other competitor firms), etc.

4 Application

A first application of the proposed methodology has been made, using firm-level data for the period 2009–2014 from the Ministry of Finance – Taxation Authorities as well as the Statistical Authority of Greece, in order to construct a prediction model of the most important measure of firm’s vulnerability to the crisis, the reduction of sales revenue due to economic crisis, based on firm’s characteristics. In particular, we used data provided by the Ministry of Finance – Taxation Authorities concerning the percentage of sales revenue reduction due to the economic crisis in the period 2009–2014 (dependent variable) of 400 Greek firms; these data were discretized by the Ministry of Finance (in order to avoid providing detailed firm’s sales revenue data) into a variable with 13 possible discrete values (SALREV_RED): increase by more than 100%; increase by 80–100%; increase by 60–80%; increase by 40–60%; increase by 20–40%; increase by 1–20%; unchanged sales; decrease by 1–20%; decrease by 20–40%; decrease by 40–60%; decrease by 60–80%; decrease by 80–100%; decrease by more than 100%. For 363 of these firms from the Statistical Authority we found data concerning 43 characteristics of them (independent variables), which concern their strategic orientations (12 variables), human resources (9 variables), technology (16 variables), structure (1 variable) and also general characteristics (5 variables). The definitions – corresponding questions for our dependent and independent variables are shown in the Appendix. These 363 firms cover a wide range of sectors and sizes: 40.2% of them were from manufacturing sectors, 9.4% from constructions, and 50.4% from services sectors; also, 52.6% of them were small, 36.1% medium and 11.3% large ones.

Using the above data we constructed prediction models of SALREV_RED based on the above mentioned 43 firm characteristics. Since the dependent variable is ordinal, with a large number of values (13) and equal distances between them, it approaches a continuous variable, we have constructed and compared prediction models of SALREV_RED using the main alternative ML algorithms proposed by relevant literature [8,9,10] for the case that the dependent variable is continuous: the classical regression (termed as ‘Generalized Linear Model’) and also four alternative more advanced statistical methods:

  • Decision Trees, which splits the initial training set into more homogeneous sub-sets using the most appropriate independent variables, and finally construct a decision tree with internal ‘decision nodes’ concerning some of the independent variables (the most useful as predictors) – actually tests concerning these independent variables – as well as leaf ‘terminal nodes’ that include predicted values of the dependent variable [9, 40];

  • Random Forests, which creates a pool of decision trees based on sub-sets of the training data set, and then gets the prediction from each of them and finally selects the best prediction by means of ‘voting’ of these decision trees [39, 40];

  • Gradient Boosted Trees, which is based on boosting prediction accuracy through the combination of a learning algorithm in a series, in order to achieve a ‘strong learner’ from many sequentially connected ‘weak learners’; usually the weak learners are decision trees, and consequtive decision trees are constructed, with each new tree attempting to minimize the errors of the previous tree [41];

  • Support Vector Machines, which mqp each training data item as a point in n-dimensional space, where n is number of independent variables; then, the algorithm tries to find hyper-plane that best fits the available training data [40, 42].

For each of these five ML algorithms we calculated its mean absolute error, which is according to relevant ML literature the recommended measure of prediction accuracy for the case that the dependent variable is continuous, using the k-fold cross-validation procedure (we selected k = 10), which includes the following steps [8,9,10, 40]:

  1. a)

    the data set is divided randomly in k equally sized partitions,

  2. b)

    then we estimate the prediction model k times, in each of them using one of these segments as test data, and the remaining ones as training data,

  3. c)

    and finally the absolute prediction error is estimated as the mean of the prediction errors over the test sets of the above k runs.

In Fig. 2 we can see the estimated mean absolute prediction errors of these five ML algorithms. We can see that the examined algorithms have similar levels of prediction performance/error (ranging from 1.422 to 1.572), with the Decision Trees algorithm exhibiting the lowest mean absolute error (1.422). This is a satisfactory prediction performance, given the small size of the dataset we have used (data from 363 firms), so using a larger dataset (which is feasible, as Ministries of Finance – Taxation Authorities and Statistical Authorities have such data for quite large numbers of firms) can result in a smaller mean absolute error, and therefore an even more accurate prediction of the sales revenue reduction of a firm during an economic crisis.

Fig. 2.
figure 2

Mean absolute prediction errors of the five ML algorithms

5 Conclusions

In the previous sections has been described the development of a methodology for exploiting existing government data for economic crisis periods, in order to construct prediction models of individual firms’ vulnerability to economic crises, based on Artificial Intelligence (AI) Machine Learning (ML) algorithms. In particular, it uses existing data for economic crisis periods, on one hand from Taxation Authorities (concerning firms’ sales revenue, profits, employment, etc. decrease), and on the other hand from Statistical Authorities (concerning human and technological resources, structures, processes, strategic orientations, etc. of these firms); from these data prediction models can be trained/constructed, which enable the prediction of the vulnerability of a firm to future economic crises (e.g. in terms of sales revenue, profitability, employment, etc. decrease), based on firm’s characteristics (concerning human and technological resources, structures, processes, strategic orientations, etc.). The prediction of the degree of vulnerability of a firm to economic crisis can be useful as a criterion in firm support programmes of government agencies, for the selection of the applicant firms that will receive support, as part of multi-criteria evaluation of applications, in combination with other selection criteria usually used in such programmes, such as sector, geographic location, sales, profits, market share, export activities, debts, etc.

The research described in this paper has interesting implications for both research and practice. With respect to research it creates new knowledge in the emerging areas of government data analytics and government AI exploitation, concerning a highly important domain of economic government activity/intervention, aiming to address the quite threatening recessionary economic crises that repeatedly appear in market-based economies and cause major disruptions and problems. With respect to practice it proposes a methodology of using existing historic government data for constructing prediction models of the vulnerability that individual firms will exhibit to future recessionary economic crises. These prediction models can be very useful to government agencies that implement firm support programs in the beginning of economic crises, or even earlier when crises are in sight, in order to use more effectively their available financial resources.

Future research is required towards: i) further application of the proposed methodology using larger datasets, in various different national contexts experiencing economic crises of different intensities, and also at sectoral level (using also sector specific independent variables); ii) investigation of the prediction performance of other ML algorithms, and especially Deep Learning ones; iii) analysis of the legal aspects of the practical application of the proposed methodology by government agencies, and especially with respect to the new EU GDPR (as the data used by our methodology have been collected by government for quite different purposes).