A comparative study of feature selection and feature extraction methods for financial distress identification

View article
PeerJ Computer Science
This study appendices are held in a GitHub repository: https://github.com/DovileKuiziniene/FS_and_FE_for_FD_Appendices.

Main article text

 

Introduction

Literature review

Feature selection

Feature extraction

Data

  1. legal status belongs to one of the following categories: (a) a private limited liability; (b) a public limited liability; (c) an individual enterprise, (d) a small community. The frequency of the number of enterprises in the market has been considered when choosing the statuses. Moreover, all these types of enterprises share the objective of increasing it’s worth;

  2. does not belong to the following Nace sectors: K – financial and insurance, L – real estate, O – public administration and defense, compulsory social security, these sectors have been eliminated because of different accounting requirements for financial statements;

  3. enterprise’s age is ≥ 1.5 years in order to reduce the number of missing values in the data;

  4. enterprise has provided at least one financial statement from the last two years;

  5. enterprise has one or more social insured employee, avoiding inactive market participants;

  6. at least 1.5 years have passed since good status recovery. Only enterprises having additional registration of good enterprise conditions (e.g., the enterprise had a history of bankruptcy case in court, but after a change in circumstances, the enterprise’s activities continued, and its good conditions were registered in the Lithuanian register) were analyzed. This criterion has been incorporated to re-evaluate enterprises that have experienced financial difficulties in the past, yet persist in their operations.

Class

  1. a bankruptcy case is filed against the enterprise;

  2. the enterprise’s status changed to going bankrupt, bankrupt, under restructuring, under liquidation, liquidation being initiated, removed, liquidated, liquidation due to bankruptcy;

  3. the enterprise has made announcements to the register center about bankruptcy, liquidation, restructuring, insolvency, etc.;

  4. the enterprise is included in the State tax inspectorate’s lists of (1) companies temporarily exempted from submitting declarations to the STI, (2) companies that have declared temporary inactivity of the STI; (3) companies for which the STI has submitted a proposal to the State Register for deregistration in accordance with Article 2.70 of the Civil Code.

  5. enterprise had no socially insured employees for the past six months.

Features

  1. stable: information is constant, e.g., legal status, types of sectors;

  2. depending on an event: changes when the event occurs, the number of courts, the number of changes of directors, the time elapsed since the last event, etc.;

  3. periodic data (annual, quarterly, monthly): information that is updated at the indicated periodicity, e.g., financial reports, macro indicators, the number of employees.

Methodology

  1. Which feature selection or extraction method is the most efficient for financial distress identification?

  2. Which data categories are the most important in financial distress prediction?

  3. Does the composition of the final feature set depend on the feature selection technique?

  4. Which feature extraction strategy is more efficient: analyzing all features at once vs analyzing features by categories?

  5. Which machine learning model performs better in financial distress identification?

Dimensionality reduction methods

  1. Feature selection approach (FS) is used to determine a small subset of informative features from the original large range of data (Al-Tashi et al., 2020) by removing irrelevant, redundant or noisy features. Let Y = {y1y2, …, yn} denote the class labels, where n is the number of enterprises and yi ∈ {0, 1}, i =1 , …, n belongs to one of the two classes c1 – non-financial distress or c2 - financial distress. Each enterprise is defined by a number of features xij, j = 1, …, d. We assume that xij are realizations of features Xj that can be ordered according to their importance FI=(Xπ(1),Xπ(2),,Xπ(d)) where π(j), j = 1, 2, …, d is the new descending index of feature Xj.

    1. Group—Filter methods assume complete independence between the model and the data. In this case, features are selected based on their scores in various statistical tests, determining their correlation with the outcome variable.

      1. Cohen’s D (ChD) coefficient is a standardized effect size for comparing the means of Almasri (2022); Imik Tanyildizi & Tanyildizi (2022) two groups (financial distress enterprises and non-financial distressed). Cohen’s D is defined as ChD=|¯Xxc1j¯Xxc2jσ2xc1j+σ2xc2j2| where ¯Xj and σj are means and standard deviations of corresponding groups. The order of the feature’s importance is determined by the effect size, i.e., the greater the ChD, the larger the effect size.

      2. Correlation (Corr) coefficient analyzes possible linear association between variables (Cheng et al., 2020; Jiarpakdee, Tantithamthavorn & Treude, 2020; Séverin & Veganzones, 2021). Based on the type of feature, we use Pearson and point-biserial correlation: Pearson correlation=rXjXk=|ni=1(xij¯Xj)(xik¯Xk)σXjσXk| Point − biserial correlation=rpb=|¯Xxc1j¯Xxc2jσXjpb| where p and b denotes the proportion of cases in the classes of â financial distressâù and “non-financial distress”. σXj is a standard deviation of feature j for all enterprises. Correlation with the outcome variable determines the order of the feature’s relevance (higher is preferable). A high correlation (> 0.75) between features indicates the problem of multicollinearity. To deal with this problem, from the group of highly correlated features, only one is left, having the highest correlation with the outcome, while others are removed from the sample. Finally, all the remaining features are ranked by the strongest relationship with the outcome.

      3. The Kruskal–Wallis (K–W) test is used to compare medians among two or more groups. The Kruskal–Wallis test is chosen because it does not require any prior assumptions about data distribution and is good for all types of features (Devi Priya et al., 2022). The selected significance level is 0.05, i.e., smaller p values suggest that there are substantial differences between the two groups (Linja et al., 2023). The test statistic for the Kruskal–Wallis test is defined as follows: Kruskal − Wallisχ2=H=(n1)cl=1nl(¯Rl¯R)2cl=1nlj=1(Rlj¯R)2 where n – the sample size, nl is the number of observations in class l, Rlj is the rank of feature j from class l, ¯Rl the mean rank of all features in class l, ¯Rl is the mean of all the Rlj, ¯R is the mean of all the Rlj. The order of feature importance can be defined by the lowest p − value or the highest |H| value.

      4. Group—Embedded methods include the feature selection in the model fitting process. Classification methods that perform embedded feature selection include:

      1. The least absolute shrinkage and selection operator (LASSO) is a method that combines feature selection and regularization (Altman et al., 2022; Zizi et al., 2021; Huang, Wang & Kochenberger, 2017). The LASSO has the advantage of maintaining the stability of ridge regression and outperforming stepwise regression models (Li et al., 2021). The LASSO function aims to solve (Tibshirani, 1996): minβ0,βi={ni=1(yiβ0xTijβ)2}s.t.dj=1|βj|λ. where β: = (β1β2, …, βd) is a coefficient vector and λ > 0 adjusts the sparsity of the estimator by eliminating features from the model forcing βj → 0. The importance of a feature is the absolute value of the LASSO regression coefficient.

      2. Random Forest (RF) uses a permutation importance measure, which is shown to be an effective tool for feature selection (Gregorutti, Michel & Saint-Pierre, 2015). RF offers two techniques for calculating the relevance of all features, resulting in a feature importance rank. These techniques create a rank by considering both the feature cost and differentiating ability (Zhou, Zhou & Li, 2016):

        1. Mean Decreasing accuracy (MDA) (Han, Guo & Yu, 2016) is a feature importance measure based on OOB (out of bag) error. Assume ht(Xi) and ht(xij) refer to the predicted label for OOB instance Xi before and after feature permutation (Wang, Yang & Luo, 2016). The relevance of a feature Xj is assessed by MDA, which estimates the mean drop in OOB accuracy before and after feature Xj permutation: MDA(Xj)=1ntreentreet=1iOOBI(yi=ht(Xi))iOOBI(yi=ht(xij))|OOB|

        2. Mean Decreasing Gini (MDG) calculates the overall reduction in node impurity (e.g., Gini index) as a result of splitting on the feature and averaging it across all trees (Wang, Yang & Luo, 2016). The MDG is defined as follows: MDG(Xj)=1ndot[1ndotj=1Gini(l)j] where Gini(l)j is the jth Gini index of feature Xj among the ndot tree nodes. The Gini index is defined as follows: Gini=cl=1lj(1lj)=1cjl=1l2k Where lj is the probability of the feature Xj classification in the distinct class (Zhang et al., 2019).

      3. Extreme Gradient Boosting Machine (XGBoost or XGB) method assigns a relevance value to each feature based on its involvement in the outcome decision-making using boosted decision trees by Gain metrics (Zheng, Yuan & Chen, 2017) ((11)). Gain=12[(iILgi)2iILhi+γ1+(iIRgi)2iIRhi+γ1+(iIgi)2iIhi+γ1]γ2 where gi, hi indicate the first and second-order gradients, γ1 and γ2 are regularization parameters, I = ILIR, while IL and IR represent all the left and all the right nodes after each split, respectively (Jiang et al., 2023).The greater the XGBoost gain, the more useful and significant the feature (Chen et al., 2021).

      4. Group—Hybrid combines different methods together:

      1. Overlapping Features (Over_feat)–features having a rating in all feature selection methods, e.g., if LASSO excludes a feature it will not be further analyzed.

      2. Voted importance (V oted_imp)–a joint rank combining results from all feature selection methods. Each feature selection method ranks the features from the most important to the least. Based on these ranks, we can calculate the cumulative weight for each feature: V oted_imp(Xj)=mj=1FIj where FIj is the rank of the feature and m is the number of used feature selection approaches (in I and II group).

  2. Feature extraction approach (FE) maps high-dimensional data X into a new, lower-dimensional space (Z = [z1z2, …, zn] ∈ ℝn×k, where k < d) retaining as much information as possible (Ayesha, Hanif & Talib, 2020):

    1. Group—Linear methods a transformation of data, which projects data linearly:

      1. Linear Discriminant Analysis (LDA) projects the data into a new feature space by minimizing the scatter within classes (Sw) and maximizing the scatter between classes(Sb) (Djoufack Nkengfack et al., 2021). This method only creates as many linear discriminants as there are number of classes, minus one (Anowar, Sadaoui & Selim, 2021). Sw=1ncl=1nli=1(Xil¯Xg)T(Xil¯Xl) Sb=1ncl=1nl(¯Xl¯X)T(¯Xl¯X) where c is the number of classes, ¯X – the global mean of all classes (Anowar, Sadaoui & Selim, 2021; Djoufack Nkengfack et al., 2021), and eigenvectors are defined as: W=eig(S1wSb) where the largest eigenvalues correspond to k-eigenvectors of (S1wSb), and of linear discriminants forms W = (w1w2, …, wk) (Djoufack Nkengfack et al., 2021; Ayesha, Hanif & Talib, 2020) Z=XWT

      2. Principal component analysis (PCA) is the most extensively used dimensionality reduction approach summarizing a large set of features in low dimension with minimum loss of information (Ma & Park, 2022). The PCA method transforms a set of features into uncorrelated Principal Components (PCs), which are ranked and ordered by their importance, e.g., the first PC explains the most variance, the second PC explains the most variance in what is left once the effect of the first PC is removed, etc. (Ayesha, Hanif & Talib, 2020). The general form of PCA transformation can be estimated as a linear weighted combination of features: Z=XW where Z = [z1z2, …, zn] ∈ ℝn×d are principal components and W = [w1w2, …, wd] ∈ ℝd×d are eigenvectors by Ayesha, Hanif & Talib (2020). The first eigenvector is determined as follows: w1=argmax||w||=1{||Xw||2}

      3. The factorial analysis of mixed data (FAMD) was proposed for handling data containing continuous and categorical features (Ran, 2019). FAMD is a combination of PCA and MCA (multiple correspondence analysis) methods (Ma & Park, 2022). Based on the types of features, relationships between them are measured using Pearson’s correlation coefficient, the squared correlation ratio or the chi-squared test (Momenzadeh et al., 2021; Josse, 2016). Fs=argmaxFsRnpcontjcont=1r2(F1,Xjcont)+pcatjcat=1η2(F1,Xjcat) where Fs is the orthogonal condition to Fs for all s′ < s, pcont and pcat are the number of continues and categorical features, respectively (Josse, 2016). The general transformation of FAMD can be expressed as: Z=XFs

      4. Union of separate PCA & FAMD models (Union_PCA_FAMD) select the appropriate data transformation depending on a data type in the analyzed category (see Table 2). In total, there were 12 different categories analyzed, and for three of them, subcategories were created: (1) Financial statements: records and ratios; (2) Sector: class and ratios; (3) Social insurance: employees and delay of taxes (including analyzes of time and amount). The PCA or FAMD method is chosen depending on the data type (continuous or categorical). f(x)={Z,if all features belonging to the category are continuousZ,if category contains mixed type features

      5. Group – Non–linear methods reduce the dimensionality by reflecting intrinsic non–linear relations between variables.

      1. Autoencoder (AE) is a neural network which is very efficient in generating abstract features of high–dimensional data by minimizing reconstruction loss (Meng et al., 2017). This loss function sets the target value equal to the input by applying a backpropagation algorithm. The main idea of the process is the usage of the encoder–decoder. Firstly, the encoder process transforms input data to lower-dimensional space (to a bottleneck or middle layer); secondly, the decoder converts it back into high-dimensional space (Yan & Han, 2018). The minimization of reconstruction loss assures that the middle layer incorporates most of the information from the original input space (Phadikar, Sinha & Ghosh, 2023). Thus, the middle layer holds the reduced representation of the input data (Kunang et al., 2018). X=gθ(Z)=s(WZ+bZ) where f(X) is an encoder function, g(Z) is a decoder, s is an activation function, W represents weights and b is the bias for data x. The training process of the autoencoder finds parameters θ = (WbXbZ) by minimizing the reconstruction loss (Kunang et al., 2018): θ=minL(X,X)=minL(X,g((f(X))))

      2. Union of separate autoencoders (Union_AE) trains several autoencoders depending on the analyzed data category. Analyzed data categories are the same as in Union_PCA_FAMD.

Number of features

Machine learning methods

  1. Supervised machine learning techniques are used to construct predictive models. These algorithms try to model relationships and dependencies between the dependent variable and independent variables. In machine learning, dependent variables are frequently referred to as labels, targets or classes and independent variables are known as features (Algren, Fisher & Landis, 2021). This study employs boosting and neural network techniques:

    1. The Boosting technique seeks to increase the prediction accuracy by combining results from multiple classifiers (Nettleton, 2014). In this study, we focus on AdaBoost, Categorical Boosting (CatBoost) and Extreme Gradient Boosting Machine (XGBoost) techniques.

      1. AdaBoost—an ensemble approach that trains and deploys trees one after the other. It implements the boosting technique by combining multiple weak classifiers to build one strong classifier. Each weak classifier attempts to improve the classification of samples that were misclassified by the previous weak classifier (Misra & Li, 2020).

      2. Categorical Boosting (CatBoost)—a machine learning technique that is based on gradient boosting decision trees. A powerful machine learning technique called gradient boosting can handle challenges with diverse features, noisy data, and complex dependencies. Moreover, this approach can easily handle categorical features by substituting the original categorical variables with one or more numerical values. Ordered boosting is a new technique that replaces the traditional gradient estimation methods (Zhang, Zhao & Zheng, 2020).

      3. Extreme Gradient Boosting Machine (XGBoost)—a tree ensemble model that can be applied for both classification and regression problems. Its main idea is to make the target function as minimal as possible while employing the gradient descent method to produce new trees based on all previous trees. While using XGBoost to solve regression issues, incrementally new regression trees are added, and the residuals of the prior model are subsequently fitted using the newly created CART tree. The final predicted value is the sum of the outcomes of each tree (Pan et al., 2022). Selecting the loss function is a crucial step in setting XGBoost models. Loss functions are used for classification and regression problems; specifically, we use the “binary” loss function for binary classification (Brownlee, 2021).

    2. Discriminant analysis (DA)—a classification and dimension reduction technique described in ‘Dimensionality reduction methods’ (Fisher, 1936). The discriminant analysis creates linear combinations separating objects into classes, assuring that the variance within the class would be minimized and the variance between classes—maximized. These directions are known as discriminant functions, and their number is equal to the number of classes minus one (Canizo et al., 2019). In this study, we use two types of DA: Linear DA and Quadratic DA, which enables the non–linear separation of data (Fisher, 1936).

    3. Decision trees (DT)—a powerful data mining tool that’s often used for feature selection and classification tasks. A decision tree may be used in data mining as a primary model-building method, or for automated feature selection (Bunge & Judson, 2005). Decision trees assist in identifying the most important characteristics that support precise classification by assessing the relevance of each feature at each node split (Sugumaran, Muralidharan & Ramachandran, 2007).

    4. K-nearest neighbor (KNN)—probably the most well-know non-parametric classification technique (Chanal et al., 2022). This method classifies each unlabeled observation by the majority class among its k-nearest neighbors in the training set (Chanal et al., 2022). In this study, we perform experiments with k = 3, 5, 7 nearest neighbors.

    5. Logistic regression (LR)—contrary to its name, LR is a classification model rather than a regression model (Subasi, 2020). For binary and linear classification issues, logistic regression provides a faster and more efficient solution. This classification approach performs very well with linearly separable classes and is relatively simple to implement. For the LR method, the assumption of multicollinearity is fulfilled by removing features, highly correlated with other features. Furthermore, a significance level of p-value is utilized, and features with a p-value greater than 0.05 are also eliminated. For example, using the KW feature selection approach, we have identified 100 significant features that have been used for LG model development. However, to avoid the problem of multicollinearity, we have performed additional analysis and have removed highly correlated features. Thus, the final LG model has been developed using 20 statistically significant features.

    6. Naive Bayes (NB)—based on the Bayes theorem, the naive Bayes classifier is a member of the family of probabilistic classifiers. It is based on the idea that a feature’s presence in a class is independent of any other features that may also be present in that class. For example, if an enterprise has a debt of 100,000 euros to a social insurance institution, suffers financial losses, and the turnover is consistently decreasing, it can be categorized as bankrupt. All three of these independent contributing factors are considered by a naive Bayes algorithm (Krishnan, 2021).

    7. Neural networks are artificial intelligence models that are designed to mimic the human brain’s functions (Casas, 2020). We use three different neural network methods:

      1. Artificial neural networks (ANN (I-III))—the model that uses calculations and mathematics to imitate human psychology. The unique architecture format used by ANN models mimics a biological nervous system. The ANN models are made up of neurons that interact in a complex, non–linear manner, just like the human brain. Weighted links connect the neurons to one another (Malekian & Chitsaz, 2021). The hidden structure of the neuron network has been marked I-III, which indicates hidden layers between input and dense layers. After each layer (except a dense one) is implemented, a drop-out layer is excluded.

      2. Convolutional neural network (CNN (II, III, V))—a specific kind of feed-forward neural network used in AI. The input data to CNN is displayed as multidimensional arrays. It works well for a large number of labelled data. Based on the important function played by the receptive field, it gives weights to each neuron, such that it can distinguish between the relative importance of different neurons. Three different layer types create CNN’s architecture: convolution, pooling, and fully linked (Shajun Nisha & Nagoor Meeral, 2021). The hidden structure is indicated the same as in ANN. However, this is only an indication for used conv_1d and flatten layers; input, drop out, max-pooling and dense layer are not included in the hidden layers structure calculation.

      3. Extreme learning machine (ELM (100,150,200,300))—is essentially a single feed-forward neural network. The weights between inputs and hidden nodes are distributed at random throughout its single layer of hidden nodes. As a result, the parameters of the model can be calculated without the need for a learning process, and they remain constant throughout the training and prediction stages. On the contrary, the weights that connect hidden nodes to outputs can be learned exceedingly fast (Garza-Ulloa, 2022).

      4. Self-organizing map (SOM)—takes a set of input data and maps it onto neurons of a (usually) two-dimensional grid. Each neuron in the 2D grid is assigned a weight vector with the same dimensionality as that of the input vector. The weights represent the associated memory. Finding a winning neuron and modifying the weights of the winner neuron and its nearby neurons are the two fundamental steps in the SOM computational process (Köküer et al., 2007).

    8. Random Forest (RF)—an algorithm that uses classification and decision regression trees as the foundation for its ensemble machine learning approach (Xia, 2020). Using the values of the predictor variables as inputs, decision tree learning attempts to build a statistical prediction model to predict the values of the response variable(s). The model is created by establishing a value for the response variable within each of the recursively partitioned predictor variable spaces, which generates a decision tree.

    9. Support vector machine (SVM)—an approach for supervised machine learning that can be applied to both classification and regression problems. The SVM algorithm seeks a hyperplane in the data space that creates the largest minimum distance (referred to as a “margin”) between the samples belonging to different classes (Satapathy et al., 2019). In this research, SVM kernel polynomial and radial basis functions are used.

  2. Data labels are not used in unsupervised learning methods. Instead, discovering a structure within a dataset, “cleaning up” the data, or grouping data points into groups are the tasks that the algorithm is expected to complete (Algren, Fisher & Landis, 2021). In this study, three different unsupervised methods are employed:

    1. Isolation Forest (IF)—an unsupervised, non-parametric method for anomaly detection in multivariate data. This outlier detection method seeks to isolate anomalies from the rest of the data using an ensemble of decision trees. It constructs a forest of isolation trees, or iTrees. Partitioning data is required for creating iTrees. Furthermore, iTrees are binary trees that use recursive partitioning to isolate abnormalities. These anomalies, which are “rare and diverse”, are therefore “more sensitive to isolation and hence have short path lengths” (Chater et al., 2022).

    2. K-means clustering (Kmeans) divides the set of samples into K clusters, each described by the centroid, i.e., the mean of the samples in a cluster. The k-means algorithm uses an iterative approach to find the optimal cluster assignments by minimizing the sum of squared distances between observations and assigned cluster centroid (Ashour, Guo & Hawas, 2019).

    3. One-class support vector machine (One class SVM)—an approach that identifies unusual observations compared to the instances of the known class. The one-class SVM attempts to classify one class of objects and separate it from all other possible objects. The classifier can correctly classify certain objects, but the others will be classified as outliers. Thus, a one-class classifier is trained to label these observations as outliers (Seo, 2007).

Evaluation metrics

Research Results

A benchmark model

Feature selection

Feature extraction

Machine learning models

Conclusions

  1. Feature selection approach results proved the suitability of embedded and hybrid methods for financial distress identification. The best results were obtained when analyzing feature sets consisting of ≤50 features. Commonly, features in FS methods were selected from Financial statements and Social Insurance data categories.

  2. Feature extraction approach results approved the usage of the union strategy. The union strategy has been implemented separately for linear and non–linear data transformations based on the data category. This led to dimensionality-reduced data categories that are linked to a single feature space, e.g., constructing 12 unique autoencoders for each data category and then combining them into one feature set. The highest AUC score was achieved after the implementation of the union strategy in the FE approach for all ML separately.

  3. The analysis of the implementation of machine learning methods for identifying financial distress reveals that simpler (not deep) neural network methods give better results. Furthermore, in order to conduct a more comprehensive comparison of the findings, it is imperative to extend the neural network structure experiments for each dimensionality reduction technique individually in further research. Moreover, the findings indicate that the unsupervised methods that have been analyzed are not worthy of further investigation.

  1. Test set. The research results were tested based on the information available for half a year.

  2. Number of features. Feature ranking has been used after the implementation of feature selection techniques. However, it is still unclear how to select the appropriate number of features. Therefore, the number of features has been randomly selected to make the methods comparable.

  3. Stability of feature set. The stability of the chosen features over time was not examined.

  4. Balancing technique. It is unclear how different data balancing techniques would affect the results. In this study, only one balancing technique was used.

  5. Model specifications. Model parameter optimization was not performed; baseline models were used to compare results. However, small changes in the results of the neural network models were observed when changing the number of features.

Supplemental Information

Code files

Code files in R programming language.

DOI: 10.7717/peerj-cs.1956/supp-1

Additional Information and Declarations

Competing Interests

Robertas Damaševičius is an Academic Editor for PeerJ.

Author Contributions

Dovilė Kuizinienė conceived and designed the experiments, performed the experiments, analyzed the data, performed the computation work, prepared figures and/or tables, authored or reviewed drafts of the article, and approved the final draft.

Paulius Savickas conceived and designed the experiments, performed the experiments, analyzed the data, performed the computation work, prepared figures and/or tables, and approved the final draft.

Rimantė Kunickaitė conceived and designed the experiments, performed the experiments, analyzed the data, performed the computation work, prepared figures and/or tables, and approved the final draft.

Rūta Juozaitienė conceived and designed the experiments, performed the experiments, analyzed the data, performed the computation work, prepared figures and/or tables, and approved the final draft.

Robertas Damaševičius analyzed the data, prepared figures and/or tables, authored or reviewed drafts of the article, and approved the final draft.

Rytis Maskeliūnas analyzed the data, authored or reviewed drafts of the article, and approved the final draft.

Tomas Krilavičius conceived and designed the experiments, authored or reviewed drafts of the article, and approved the final draft.

Data Availability

The following information was supplied regarding data availability:

The dataset is available at Figshare: Damasevicius, Robertas (2023). Experimental data for paper “A comparative study of feature selection and feature extraction methods for financial distress identification”. figshare. Dataset. https://doi.org/10.6084/m9.figshare.23507826.v1.

Funding

This study has received funding under the Horizon Europe Widening Participation program - Teaming for Excellence 2022 (Centre of Excellence of AI for Sustainable Living and Working (SustAInLivWork) project, grant agreement No. 101059903) and from the European Union. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

1,237 Visitors 1,143 Views 86 Downloads

MIT

Your institution may have Open Access funds available for qualifying authors. See if you qualify

Publish for free

Comment on Articles or Preprints and we'll waive your author fee
Learn more

Five new journals in Chemistry

Free to publish • Peer-reviewed • From PeerJ
Find out more