SEMGROMI—a semantic grouping algorithm to identifying microservices using semantic similarity of user stories

View article
Vera-Rivera et al. @UFPSCUCUTA present #SEMGROMI—a semantic grouping algorithm to identifying microservices using semantic similarity of user stories Read the full article https://t.co/EvfTZ2uCGq #Algorithms #ArtificialIntelligence
PeerJ Computer Science

Main article text

 

Introduction

Microservices backlog

Methodology

  1. Identify and frame the problem: Review the current state of studies related to ours to identify gaps, and identify metrics to evaluate the granularity of microservices (Vera-Rivera, Gaona Cuevas & Astudillo, 2019).

  2. Design grouping algorithm: Once the metrics have been identified, the grouping algorithm is designed. The algorithm has six parts: (1) semantic similarity calculator, (2) semantic grouper, (3) call and request calculator, (4) dependency analyser, (5) coupling grouper, and (6) metric calculator.

  3. Create grouping algorithm: It was implemented in Python using the artificial intelligence and natural language processing libraries Spacy (Spacy.io, 2020) for vector algebra and distance between points.

  4. Evaluation of a case study: The evaluation was carried out in two academic projects (Cargo Tracking (Baresi, Garriga & De Renzis, 2017) and JPetStore (Jin et al., 2019)) and two industry projects (Foristom Conferences and Sinplafut (Vera-Rivera, Vera-Rivera & Gaona-Cuevas, 2019)).

  5. Compare the results of the algorithm with other methods: We compared the decompositions proposed by our algorithm with those proposed by other state-of-the-art microservices identification methods: domain-driven design (DDD), service cutter (Gysel et al., 2016), Microservices identification trough interface analysis (MITIA) (Baresi, Garriga & De Renzis, 2017), and our own genetic programming technique (Vera-Rivera, Gaona & Astudillo, 2021). We compared these approaches using coupling, cohesion, complexity, granularity, development time and performance metrics.

  6. Proposal for an algorithm to group user stories into microservices: After carrying out evaluations and appropriate adjustments, the algorithm for grouping user stories into microservices is proposed in this article.

Semgromi: semantic grouping algorithm

Algorithm parameters

  • Semantic similarity threshold: Since the semantic similarity between two texts is a value between 0 and 1 (with one being the same or very similar), this is the minimum acceptable similarity value above which stories will be grouped; its initial default value is 0.85. This value indicates the percentage of similarity that exists between user stories; they are considered to be similar if the semantic similarity value is greater than 85%. This value can be adjusted by the user or the architect. The semantic similarity threshold indicates how similar the texts must be to be grouped, a value closer to 1 indicates that the semantic similarity must be higher to be grouped into the same microservice; therefore, more microservices are obtained. Smaller values will group more stories into fewer microservices.

  • Coupling threshold: the minimum acceptable value of the degree of coupling between the stories to be grouped; its initial default value is 0.5. The microservices whose distance is greater than the coupling threshold are grouped together. If the value is close to 1.0, more microservices will be obtained because few will exceed that value. On the other hand, if its value is close to 0.0, only one microservice will be obtained, because all of them exceed that value. This parameter can be changed by the user, initially and according to the tests performed its value is 0.5.

  • Language: the natural language in which the user stories are written; currently only Spanish and English are supported.

  • Use entity lemmatization vs full text: whether to use the full text of the user stories or only entity lemmas to calculate semantic similarity.

Group user stories by semantic similarity

Semantic similarity among user stories

where:

  • DSS: the semantic similarity dictionary among user stories.

  • huj and huk: the user stories’ identifier; it is used as dictionary key, which is formed by concatenating the identifiers of the user stories that are compared.

  • ajk are the dictionary values, it is the semantic similarity value between the huj and huk user stories (obtained by Spacy), which is in the 0..1 interval: the closer to 1, the more similar the user stories are.

  • 1. Join name and description.

  • 2. We removed verbs, articles, adjectives and prepositions from the text, leaving only nouns.

  • 3. The topics of the text were identified by counting the number of times an entity is repeated.

  • 4. Calculate the semantic similarity.

Grouping user stories

where:

  • 1. SSi: semantic similarity of the ith microservice.

  • 2. c: number of comparisons made to calculate the semantic similarity of the microservice’s user stories. It is used to calculate the average semantic similarity of the microservice. For example, for a microservice that has two user stories assigned to it (hu1,hu2), the semantic similarity between hu1 and hu2 is calculated once, a single comparison (c=1). If the microservice has three user stories, it must compare (hu1,hu2), (hu1,hu3), and (hu2,hu3), corresponding to three comparisons (c=3).

  • 3. m: number of microservice’s user stories.

  • 4. ajk corresponds to the dictionary value, they are the semantic similarity values between the huj and huk.

where:

  • 1. SsT: the total semantic similarity of the application, it was the average of the semantic similarity of each microservice. To obtain a semantic similarity value between 0 and 100, we multiply the average by 100.

  • 2. SSi: semantic similarity of the ith microservice.

  • 3. n: number of microservices of the application.

Group microsevices using interdependence

Group microservices by coupling distance

Computing the coupling between microservices

where:

  • CpDij: coupling distance between microservices i and j.

  • callsij: number of times microservice i calls microservice j (i.e., i’s inputs).

  • requestij: number of times microservice j calls microservice i (i.e., i, outputs).

  • total_calls: the total number of calls among the application microservices.

Group microservices by semantic similarity

  • 1. Identify the entities of the microservices: For each microservice, the name and description of all associated user stories are combined into a global text that is lemmatized to identify its domain entities (as denoted by nouns).

  • 2. Computes the semantic similarity between the microservice entities: The frequency of each entity in each microservice is calculated, and the two entities with the highest frequency are selected to perform the semantic comparisons between the microservices. This process uses the machine learning technique of text classification, which automatically assigns tags or categories to text. The result is the semantic similarity between each pair of microservices, which is computed iteratively to create a semantic similarity matrix.

  • 3. Group microservices by semantic similarity: Microservices are grouped using the same algorithm that groups user stories, but now using the semantic similarity between microservices (as defined above). For each pair of microservices, they are joined if their semantic similarity value is above the semantic similarity grouping threshold; its initial and default value is 0.85, but can be changed by the designer.

Calculate metrics and evaluate candidate microservices

Validation

Evaluation methodology

  • 1. The state-of-the-art examples and industry projects were analyzed and described.

  • 2. The user stories of each example and project were identified to obtain the “product backlog”.

  • 3. User story dependencies were identified according to data flow, calls, invocations between user stories or business logic.

  • 4. Decompositions were obtained with Microservices Backlog (using the previous genetic algorithm and this new grouping algorithm), and the decompositions obtained with the other methods were uploaded to the system.

  • 5. The metric calculator obtained the metrics and the dependency graph of the Microservices Backlog of the candidate microservices for each decomposition.

  • 6. The metrics for each decomposition were evaluated and compared.

The evaluation data set (projects)

The evaluation metrics

  • Granularity—N: Number of microservices of the decomposition or system.

  • • Coupling (CpT): the absolute importance of the microservice (AIS), absolute dependence of the microservice (ADS), and microservices interdependence (SIY).

  • AISi is the number of clients invoking at least one operation of MSi; to calculate the total value of AIS at the system level (AisT) the vector norm is calculated. See Eq. (5).

ADSi is the number of other microservices on which the MSi depends. To calculate the total value of ADS at the system level (AdsT) the ADS vector norm is calculated (See Eq. (6)). Then:

SIYi defines the number of pairs of microservices that depend bi-directionally on each other divided by the total number of microservices (Eq. (7)).

Calculating the norm of the vector Cp we have the coupling value for the application (CpT), Eq. (8):

where Cp=[AisT,AdsT,SiyT], we amplify CpT by 10, in such a way that its dimension is like the dimension of the other metrics.

  • Cohesion—Lack of cohesion (CohT): the number of pairs of microservices not having any dependency between them, adapted from Candela et al. (2016). Lack of cohesion of a microservice is the number of pairs of microservices that have no interdependency between them. The LC of MSi has been defined by us as the number of pairs of microservices that have no interdependency between MSi. The lack of cohesion degree (Cohi) of each microservice i is the ratio of LC and the total number of microservices in the application (Eq. (9)), and CohT is the vector norm of the vector consisting of the Coh value of each microservice of the application (Eq. (10)).

where Coh=[Coh1,Coh2,,CohN]

  • • Cohesion—Total Semantic similarity (SsT): the average of the semantic similarity of each microservice (see Eqs. (2) and (3)).

  • • Granularity—Weighted Service Interface Count (WSICi): is the number of exposed interface operations of the microservice i (Hirzalla, Cleland-Huang & Arsanjani, 2009). We assume that each user story is associated with a single operation, so we adapt this metric as the number of user stories associated with a microservice, and WsicT is the highest WSICi of the system decomposition (Eq. (11)).

  • • Performance— Calls: the total number of invocations between microservices.

  • • Performance— Avg.Calls: average of calls that a microservice makes to another: Calls/N.

  • • Complexity—Story Points Max.(Pi): estimated effort needed to develop a user story; Max.(Pi) is the largest number of story points associated with any microservice.

  • • Complexity—Cognitive complexity points (CxT): estimated difficulty of developing and maintaining a microservice-based application, using its estimated story points, relationships, and dependencies among microservices (see more details in Vera-Rivera et al. (2021)).

CxT=Cx/Cx0

  • where: CxT: Cognitive complexity points of the system.

  • i: ith microservices.

  • Cgi: Pi(Callsi+Requesti), Callsi are the outputs of MSi and Requesti are the inputs of MSi.

  • Pi: Total user story points of MSi. Max(P1,,PN) Maximum Pi of the system.

  • Pfi: Number of nodes used sequentially from a call that makes a microservice to other microservices, counted from the ith microservice; A larger depth implies a greater complexity of implementing and maintaining the application.

  • Cx0: The base case where the application has one microservice, one user story with one estimated story point. Then Cg1=0, Max(P1)=1, N=1, WsicT=1, Pf1=0, SIY=0, and Cx=2. Therefore Cx0=2.

  • • Development Time— (Ti): estimated development time (in hours) for microservice i, calculated by adding the estimated time of each user story in it. The longest development time is used to compare the decompositions.

  • • Granularity— (Gm): indicator of how good or bad the system decomposition is, according to its coupling (CpT), cohesion (CohT), number of user stories associated with the microservice (WsicT), points of cognitive complexity (CxT), and semantic similarity (SsT); it is calculated as the norm of the vector with these metrics MT=[CpT,CohT,CxT,WsicT,(100SsT)].

Objective function

Evaluation results

  • Number of microservices in the system: All the methods analysed converged on almost the same number: CargoTracking three or four microservices, JPetStore to four or five microservices, Foristom Conferences to four or five microservices, and Sinplafut to nine or 11 microservices. Semgromi algorithm in some projects required microservice join operations to converge to the expected number of microservices, as did the genetic algorithm.

  • Coupling (CpT): had the lowest value for the genetic algorithm in all four projects, while Semgromi algorithm was lower than DDD and state of the art in only two projects.

  • Lack of Cohesion Metric (CohT): Both the genetic and Semgromi algorithms had similar but higher values than DDD and state-of-the-art methods.

  • Highest number of stories associated with a microservice (WsicT): In three out of four projects, the genetic algorithm had a lower value than DDD and the state-of-the-art methods, but Semgromi algorithm in two out of four projects the values are very close.

  • Calls between microservices: An indicator of application performance: the more calls between them, the more performance is affected, as algorithms have to be executed in different containers or machines, resulting in higher latency and longer execution times. In three of the four projects, the genetic algorithm gave a lower value, while in two of the four projects, Semgromi algorithm showed very small differences compared to DDD and state-of-the-art methods.

  • Cognitive complexity (CxT): it was significantly lower (Fig. 4) for the distribution obtained by the genetic algorithm in the four projects, whereas the Semgromi algorithm was lower than DDD and the state-of-the-art methods in only two projects (JPetStore and Sinplafut), and in the other two projects had similar complexity.

  • Granularity ( Gm): had results (Fig. 5) very similar to those obtained in complexity, with the genetic algorithm obtaining less complexity than DDD and the state-of-the-art methods; similarly, the Semgromi algorithm yield leses Gm in some projects (JPetStore and SSinplafut) and had a closely similar result in the others.

Discussion

Conclusions

Supplemental Information

User stories Foristom Conferences - Spanish.

DOI: 10.7717/peerj-cs.1380/supp-1

User stories Jpet Store - Spanish.

DOI: 10.7717/peerj-cs.1380/supp-2

User stories Sinplafut - Spanish.

DOI: 10.7717/peerj-cs.1380/supp-3

User Stories Cargo Tracking.

DOI: 10.7717/peerj-cs.1380/supp-4

Additional Information and Declarations

Competing Interests

The authors declare that they have no competing interests.

Author Contributions

Fredy H. Vera-Rivera conceived and designed the experiments, performed the experiments, analyzed the data, performed the computation work, prepared figures and/or tables, authored or reviewed drafts of the article, and approved the final draft.

Eduard Gilberto Puerto Cuadros conceived and designed the experiments, performed the experiments, analyzed the data, performed the computation work, authored or reviewed drafts of the article, and approved the final draft.

Boris Perez conceived and designed the experiments, performed the experiments, analyzed the data, prepared figures and/or tables, authored or reviewed drafts of the article, and approved the final draft.

Hernán Astudillo conceived and designed the experiments, performed the experiments, analyzed the data, performed the computation work, prepared figures and/or tables, authored or reviewed drafts of the article, and approved the final draft.

Carlos Gaona conceived and designed the experiments, performed the experiments, analyzed the data, authored or reviewed drafts of the article, and approved the final draft.

Data Availability

The following information was supplied regarding data availability:

The datasets are available in the Supplemental Files and the code is available at BitBucket: https://bitbucket.org/freve9/microservicesbacklog/.

Funding

This work was supported by Colombia’s Ministry of Science and Technology (Minciencias-Colciencias) through doctoral scholarship “753-Formación de capital humano de alto nivel para el departamento Norte de Santander”; by the Francisco de Paula Santander University (Cúcuta, Colombia) through the doctoral studies commission number 14 of 2016; by the Universidad del Valle (Cali, Colombia); and by ANID (Chile) through Anillo ACT210021 Aconcagua. There was no additional external funding received for this study. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

4 Citations 1,211 Views 83 Downloads

MIT

Your institution may have Open Access funds available for qualifying authors. See if you qualify

Publish for free

Comment on Articles or Preprints and we'll waive your author fee
Learn more

Five new journals in Chemistry

Free to publish • Peer-reviewed • From PeerJ
Find out more