《1. Introduction》

1. Introduction

As Industry 4.0 progresses, modern manufacturing is embracing a conversion from a traditional operation mode to an intelligent mode [1]. Within this trend, various advanced manufacturing technologies and paradigms are emerging [2–4]. As crucial enablers of smart manufacturing, Industrial Internet platforms aim to achieve service-oriented and on-demand production [5,6]. Industrial Internet platforms, which are an extension and implementation of the Internet of Things [7,8] and cloud computing [9,10] into the industrial field, are drawing great interest from both academia and industry [11–13]. With the support of an Industrial Internet platform, siloed manufacturing resources can be connected and encapsulated as manufacturing services delivered to customers to carry out manufacturing tasks, thereby empowering the unified management and utilization of discrete resources [14]. In general, the fulfillment of a manufacturing task necessitates the collaboration of multiple services [15]. After receiving a task, the platform decomposes it into several subtasks and matches each with a set of candidate services. A manufacturing service collaboration solution refers to a portfolio of services selected from candidate sets to fulfill a task. Consequently, one of the main purposes of an Industrial Internet platform is to orchestrate services in order to generate high-quality solutions—that is, to perform manufacturing service collaboration optimization.

During service collaboration optimization, both the functional requirements proposed by the customer and the required processing amount of the task must be satisfied. For example, it is common for a task submitted to an Industrial Internet platform to have more than one product to be processed. To enhance the performance of the collaboration solutions, it is rational and practical to select multiple services to fulfill a subtask in parallel. In this way, services are fully utilized, while solution efficiencies significantly increase. Within this context, two dimensions of collaboration coexist: In the vertical dimension, a cluster of services with similar functions collaborate in parallel for a subtask; meanwhile in the horizontal dimension, multiple service clusters with complementary functions collaborate in sequence for the entire task, forming a complex collaboration network. Service collaboration optimization for such a network comprises more than just the optimal selection of services. Amount distribution—that is, how to reasonably distribute the processing amount of one subtask to the multiple selected services—is indispensable. A task may have thousands of alternative service combinations; moreover, within a combination, there may be even more potential schemes for amount distribution among services, resulting in numerous collaboration solutions. To obtain high-quality solutions, neither service selection nor amount distribution can be disregarded. 

In addition, to achieve collaboration in two dimensions, an appropriate manufacturing service modeling method is essential. In existing service collaboration optimization, the service invocation time and cost are assumed to be fixed, regardless of the processing amount [16,17]. This assumption cannot be employed due to a violation of the on-demand principle when amount distribution is involved. Moreover, since a subtask is fulfilled collaboratively by multiple services, the amount distribution of the subtask largely hinges on the availability of these services. From the perspective of the entire Industrial Internet platform, manufacturing services exist at intra-enterprise and inter-enterprise levels, resulting in different granularities [18]. For example, a manufacturing service may correspond to a single resource such as a machine tool, a group of resources such as a workstation on a shop floor, or even a complex production process involving multiple resources in different subordinate companies of a conglomerate. The availability of services is limited by their component states and relationships, which changes over time. Unless this factor is taken into consideration during service collaboration optimization, it is difficult to safeguard the feasibility and efficiency of collaboration solutions. 

With the aim of generating high-quality collaboration solutions, this paper proposes a novel dual-dimensional manufacturing service collaboration optimization (DMSCO) methodology. The proposed methodology considers the multi-granularity characteristic of manufacturing services and comprises both functional collaboration in the horizontal dimension and amount collaboration in the vertical dimension. The main contributions of this paper are as follows:

(1) By analyzing the components and structures of manufacturing services, a multi-granularity service modeling method is presented, which enables the introduction of complicated services (i.e., composite services and service chains) to service collaboration optimization.

(2) A mathematical model for DMSCO integrating service selection and amount distribution is formulated under the consideration of service granularities, with the objective of optimizing the total cost, reliability, and finishing time of service collaboration solutions.

(3) A multi-objective memetic algorithm is developed to efficiently search for collaboration solutions to the model, in which a competition mechanism is embedded to adjust the selection probabilities of local search operators according to their contributions.

The rest of this paper is organized as follows. Related work is reviewed in Section 2. Section 3 models manufacturing services and formulates the DMSCO model. In Section 4, the proposed algorithm is depicted. Section 5 reports the numerical experiments, and the conclusions are presented in Section 6.

《2. Related work》

2. Related work

《2.1. Manufacturing service modeling》

2.1. Manufacturing service modeling

As the prerequisites of manufacturing service collaboration, physical resource virtualization and manufacturing service modeling have been intensively studied in the past decades. Researchers initially focused on functional and non-functional information descriptions. Shi et al. [19] proposed a manufacturing resource hierarchy model, in which resource information abstracted from the physical resource layer was encapsulated in a resource expressing layer using extensible markup language, and operations to resources were outlined in the resource interface layer in order to shield the internal information communication details of the resources. Vichare et al. [20] established a unified manufacturing resource model for a computer numerical control (CNC) machining system composed of the machine resource and additional auxiliary devices. The model was a standard information representation able to define the elements of the CNC machining system and support the automation of process planning decision-making. Ameri and McArthur [21] extended the manufacturing service description language, an ontology used for modeling service manufacturing capabilities, by means of property inference rules and classification rules, thereby enhancing the ontology semantically and enabling advanced reasoning. Wang and Wang [22] transformed the capabilities of physical resources into manufacturing services utilizing a function block mechanism, thus providing a viable integration method capable of coordinating various manufacturing resources in an interoperable and flexible way.

In fact, manufacturing services occur within a multi-granularity environment [18] and, based on research progress, the construction of multi-granularity manufacturing services has attracted the attention of researchers. The coexistence of services with different granularities empowers the efficient and flexible generation of collaboration solutions. A coarse-grained service can be invoked as a whole, quickly covering most requirements of a task [23,24]. Meanwhile, fine-grained services can flexibly meet other personalized needs. Liu et al. [25] presented multi-granularity resource virtualization strategies consisting of resource aggregation functions and resource clustering algorithms to encapsulate physical resources into cloud services, which bridged the gap between complex manufacturing tasks and underlying resources. Yu et al. [26] designed a multi-level aggregate service planning method based on data mining techniques, thereby determining the number of manufacturing services maintained at different granularities. Zhang et al. [27] proposed a new application pattern—namely, cloud manufacturing for industry alliances—for which a domaindriven development method for multi-granularity manufacturing services was designed, including the two phases of atomic and coarse-grained service development, in order to achieve the ondemand provision of manufacturing resources. Li et al. [28] presented a resource service chain composition evolutionary algorithm based on the dependency between resource services in order to construct optimal coarse-grained services, thereby enhancing the reusability and selection efficiency of the manufacturing services.

《2.2. Manufacturing service collaboration optimization》

2.2. Manufacturing service collaboration optimization

Existing research efforts on manufacturing service collaboration optimization mainly focus on optimal service selection and configuration. Abundant models have been proposed thus far. Lu and Xu [29] investigated knowledge-based service composition and adaptive resource planning in a cloud manufacturing environment, facilitating fast resource allocation for a given service request through a semantic web-based system. Ren et al. [30] proposed a service selection model considering social relationships among the manufacturing services to maximize the overall synergy effect. Five types of social relationships among services were abstracted from the service interaction and cooperation data, and a service synergy effect model was derived based on a weighted aggregation of relationship strength. Considering logistics services in service collaboration, Zhou et al. [31] analyzed and established a collaborative optimization model of logistics and processing services and took the average task delivery time as the optimization objective. Wu et al. [32] presented a multi-objective integer bilevel multi-follower programming model to concurrently optimize the sustainability (economic, environmental, and social performance) and quality of collaboration solutions. Wang et al. [33] established a multi-objective dynamic service composition reconfiguration model in which eight crucial practical constraints were introduced according to a real-life cloud manufacturing process, thereby narrowing the gap between theory and application.

Due to the non-deterministic polynomial-time hard (NP-hard) nature of optimal service selection [34], intelligent evolutionary algorithms are widely utilized to solve the problem, allowing high-quality solutions to the models to be obtained with fewer computing resources [35]. Bouzary and Chen [36] developed a grey wolf optimizer combined with evolutionary operators of a genetic algorithm; they adapted the continuous structure of the grey wolf optimizer to the optimal service selection problem and avoided local optimal stagnation during the hunting process by providing more exploration strength. Akbaripour and Houshmand [37] were the first to analyze the solution space landscape of the problem, demonstrating that the landscape was rugged and that local optima were gathered in a small area of the search space. Based on their analysis, they designed a hybrid approach integrating an imperialist competitive algorithm with a local search algorithm. Zhang et al. [38] proposed a novel genetic-based hyper-heuristic algorithm in which the genetic algorithm acted as a high-level heuristic to produce appropriate combinations of low-level heuristics operating on the solution domain in order to solve the multitask-oriented optimal service selection problem in uncertain environments. Zhou and Yao [39] proposed a multi-objective hybrid artificial bee colony algorithm to optimize the quality of service and energy consumption, in which a cuckoo search with a Lévy flight was introduced to maintain the diversity of solutions, and a comprehensive learning strategy was designed to ensure a balance between the exploitation and exploration of the algorithm. Zhang et al. [40] presented an improved non-dominated sorting genetic algorithm-II (NSGA-II) coupled with a Tabu search and enhanced K-means mechanism to optimize the long-/short-term utilities of the providers, consumers, and operators involved in the service collaboration.

To date, the exploration of dual-dimension service collaboration remains limited. Some of the relevant research only focuses on service selection [41,42]. However, distributing the processing amount to the selected services is integral to manufacturing service collaboration optimization when multiple services fulfill the same subtask in parallel. Zhu et al. [43] investigated the amount distribution but ignored the service selection phase; they distributed the amount to all the candidate services. Despite decreasing the finishing time, execution of the collaboration solution becomes difficult due to the massively involved services. Furthermore, the availability of a manufacturing service is simply assumed to be a numerical value in previous references. Although such a solution may be suitable for the field of computer science, the situation in the manufacturing industry is more complicated. The availability of a manufacturing service is affected by many factors, such as the service structure, internal component services, and scheduled occupancy. Therefore, a single numerical value cannot accurately reflect the availability of a manufacturing service.

《2.3. Overview》

2.3. Overview

As mentioned above, both manufacturing service modeling and collaboration optimization have been extensively studied. However, research on service collaboration optimization rarely considers the characteristics of the service granularity, resulting in an incoherence between the service modeling and collaboration optimization. Despite the few efforts that have been made in this area [44,45], a detailed description of the availability of services at different granularities is still missing. Moreover, existing research on service collaboration optimization primarily focuses on the selection of services to satisfy functional requirements, whereas amount-related collaboration is disregarded. Further research on the optimization model and the specific algorithm for dualdimensional service collaboration is required. To this end, this paper presents a DMSCO methodology that is multi-granularity service-based and that simultaneously optimizes service selection and amount distribution.

《3. Problem formulation》

3. Problem formulation

《3.1. Multi-granularity manufacturing service modeling》

3.1. Multi-granularity manufacturing service modeling

In the overall performance of collaboration solutions, the total cost, reliability, and finishing time are of primary concern with regard to customers, respectively corresponding to the unit cost, reliability, and processing speed of services. In addition, the feasibility of solutions is limited by the availability of services. While modeling manufacturing services, this paper primarily focuses on these attributes. It should be noted that the availability of services is measured from the perspective of time—that is, whether the services are available during certain time periods. Manufacturing services exist at both the intra-enterprise level and the interenterprise level. According to their internal components and structures, services can be modeled at three granularities: resource services, composite services, and service chains. Notations used for modeling these services are listed in Table 1, and the service expressions and attributes are shown in Table 2.

《Table 1》

Table 1 Notations used for multi-granularity manufacturing service modeling.

《Table 2》

Table 2 Modeling of multi-granularity manufacturing services.

A resource service is a simple service mapped by a single physical resource within an enterprise. Since the resource cannot be available all the time, the availability of the service is expressed as a series of available periods of time, where the element dk indicates the duration of time during which the resource is accessible.

A composite service comprises multiple interdependent resource services, which are generally provided by a single enterprise for a specific function. A set of component services V = {S1, S2, ..., Sn, ..., SN} is invoked at the same time when providing the function to the outside, where a central service S1 acts as the main implementer of the function, and the rest are auxiliary services collaborating around S1. The availability of a composite service is the intersection of the availability of its components.

A service chain is a fixed service provided by multiple correlated enterprises, such as conglomerates or alliances. It consists of a set of component (i.e., resource or composite) services V = {S1, S2, ..., Sn, ..., SN} that collaborate with specific execution constraints E = {<Sm , Sn>|Sm Sn V }. <Sm , Sn> means that Sm is invoked before Sn, and a component service will not be occupied until invoked. The processing speed and availability of a service chain are expressed as the sets of corresponding attributes of the components, respectively.

《3.2. The model for DMSCO》

3.2. The model for DMSCO

In the model for DMSCO, a manufacturing task T contains I subtasks with different function requirements, and each subtask STi can be fulfilled by a set of candidate services CSSi. Regarding the vertical dimension, a cluster of functionally equivalent services CSi = {Si1, Si2, ..., Sij, ..., } are selected from CSSi to fulfill STi. Aside from service selection, the distribution of the processing amount of STi to the services in CSi needs to be determined. In the horizontal dimension, all functionally complementary CSi compose the final collaboration solution CS = {CS1, CS2, ..., CSi, ..., CSI} for T. The following text formulates the model from the two dimensions, where the availability constraints of the services at different granularities are constructed, and the total cost, reliability, and finishing time are adopted as the optimization objectives. Notations used for DMSCO are listed in Table 3.

《Table 3》

Table 3 Notations used for the DMSCO model.

3.2.1. Collaboration in the vertical dimension

Considering the amount distribution, the total cost Cij, reliability Rij, and finishing time Fij of a single Sij are formulated first. The formulas are different depending on the service granularities. For a resource service or composit service: 

where Eq. (5) indicates that Amtij must be processed within an available duration of , ensuring the availability of Sij.

For a service chain, Cij and Rij are calculated in the same way. Fij and Fij are decided by the component services, and are formulated as follows:

where Eqs. (10) and (11) ensure the availability and correct invocation sequence of the component services, respectively.

For a service cluster CSi, the total cost Ci, reliability Ri, beginning time Bi, and finishing time Fi are formulated as follows:

where Eq. (16) ensures that the sum of the processing amount distributed to the services in CSi is equal to Amti, and Eq. (17) ensures that CSi is selected from the candidate service set CSSi.

3.2.2. Collaboration in the horizontal dimension

Regarding the horizontal dimension, the total cost C, reliability R, and finishing time F of a service collaboration solution CS are formulated as follows:

Due to the different measurement methods and units of C, R, and F, objective functions are normalized as follows to optimize the metrics at the same time:

《4. The proposed memetic algorithm for DMSCO》

4. The proposed memetic algorithm for DMSCO

Service selection has been proved to be an NP-hard problem [34]. Consequently, DMSCO incorporating both service selection and amount distribution is NP-hard. To address this problem, a competition-based multi-objective memetic algorithm (CMOMA) is developed.

《4.1. Framework of CMOMA》

4.1. Framework of CMOMA

Memetic algorithms represent a category of metaheuristics that combine evolutionary algorithms with local searches [46]. The framework of the proposed CMOMA is shown in Fig. 1, where a grey wolf optimizer is deployed for the global search, and a competition-based variable neighborhood search is embedded for the local search. For the adaption of DMSCO, a mixed-crossover operator is designed for the global search. In addition, to fully explore the search space, the population is updated through fast non-dominated sorting, and parents (leaders) for a solution are randomly selected from the population.

《Fig. 1》

Fig. 1. Framework of CMOMA.

The local search is of great significance in memetic algorithms and should be tailored to the specific problem. This paper develops two types of local search operators corresponding to service selection and amount distribution, each of which includes four operators. When applying the local search to a solution, one of the operators is selected for execution. In addition, a competition mechanism is designed to update the selection probabilities of the operators.

《4.2. Encoding and decoding》

4.2. Encoding and decoding

To accommodate DMSCO, a two-vector encoding scheme is proposed, which involves a service selection vector X and an amount distribution vector Y. Each vector consists of I segments corresponding to I subtasks. Since a subtask STi is collaboratively fulfilled by Ji services, the segment Segi in both X and Y comprises Ji elements carrying information about the services selected for STi. The element xij of X is an integer in [1, Li], representing the index of a selected service in CSSi (Li is the length of CSSi). Y is used to distribute the amount to the selected services. The element yij is a real number in [0, 1], representing the weight of the amount distributed to the allelic service on X. Fig. 2 shows an example of the encoding scheme, in which the task contains three subtasks, each subtask has 20 candidate services, and up to three of these can be selected. Based on the encoding scheme, the population of CMOMA is randomly initialized.

《Fig. 2》

Fig. 2. Example of the encoding scheme.

Prior to decoding X, elements of Y are inspected. If yij is equal to 0, the allelic service on X will not participate in the collaboration. Meanwhile, to prevent the amount distributed to a service from being too small, yij will be set to 0 when less than 0.1. If all yij belonging to Segi equal 0, one of them will be set to a random number between 0.1 and 1.0. After inspection, the xijth candidate service in CSSi is selected if the allelic yij is greater than 0. When decoding Y, the last yij greater than 0 in each Segi is marked as yi*, and the amount distributed to services is then calculated as shown in Eq. (25). Supposing that the total amount of the task in the example in Fig. 2 is 10 000, the decoding result is presented in Fig. 3.

《Fig. 3》

Fig. 3. Example of the decoding scheme.

《4.3. Global search》

4.3. Global search

The global search algorithm is extended from the grey wolf optimizer, which is an emerging evolutionary algorithm inspired by the hunting behaviors of grey wolves [47]. In each iteration of the grey wolf optimizer, solutions renew their positions under the guidance of three leaders (the first three best solutions), α, β, and δ. Compared with similar algorithms, the optimizer is characterized by a fast search speed and easy implementation. It was initially designed for solving continuous optimization problems. As there are both discrete variables X and continuous variables Y in the solutions, a mixed-crossover operator is designed in the proposed algorithm. For X, the crossover operator is formulated as Eqs. (26) and (27), where rand() is a function returning a random number in [0, 1], and  are X elements of the new solution, the original solution,α, β, and δ. Fig. 4 presents an example of the crossover operator for X.

《Fig. 4》

Fig. 4. Example of the crossover operator for X.

For Y, the original crossover operator in the grey wolf optimizer is adopted, as shown in Eqs. (28)–(31), where a linearly decreases from 2 to 0 during the iteration, and are the Y elements of the new solution, the original solution, α, β, and δ

Regarding multi-objective optimization, most research introduces an external archive into the grey wolf optimizer to preserve the non-dominated solutions obtained so far, and utilizes a roulette-wheel method to select leaders from the archive. However, because only the best solutions are preserved, the leaders selected from the archive are likely to cause the algorithm to fall into a local optimum. Therefore, this paper adopts fast nondominated sorting to select solutions for the next population. When selecting leaders, all solutions in the population have the same probability of being selected. In this way, solutions that are not in the first Pareto front have a chance to be selected, which is conducive to jumping out of the local optimal and fully exploring the search space. Details about fast non-dominated sorting can be found in Ref. [48].

《4.4. Competition-based variable neighborhood search》

4.4. Competition-based variable neighborhood search

The local search is executed by a competition-based variable neighborhood search. Two types of local search operators are individually designed for service selection and amount distribution, and each type contains four operators. When the local search is performed on a solution, the local search type is randomly determined first, and then a specific search operator from the type is selected according to the probabilities updated by a competition mechanism.

4.4.1. Local search for service selection

Since high-quality services can significantly improve collaboration solutions, the local search for service selection is designed to replace low-quality services with high-quality services. This type includes three objective-oriented operators, OS1, OS2, and OS3, and a hybrid operator, OS4. Each objective-oriented operator corresponds to one objective to be optimized and performs the same related operation on all segments in a solution. The hybrid operator randomly performs the corresponding operations of different objectives on each segment in a solution and is dedicated to maintaining a balance. The details of the operators are as follows: 

(1) OS1 for total cost : For each Segi in X, the operator finds the service with the highest unit cost in Segi and replaces it with a randomly selected service with a lower unit price from CSSi.

(2) OS2 for reliability : For each Segi in X, the operator finds the service with the lowest reliability in Segi and replaces it with a randomly selected service with higher reliability from CSSi.

(3) OS3 for finishing time : For each Segi in X, the operator finds the service with the lowest speed in Segi and replaces it with a randomly selected service with a higher speed from CSSi. More specifically, if the service is a service chain, the lowest speed of the internal component services is regarded as the speed of the service chain.

(4) OS4 for all objectives: For each Segi in X, the operator randomly selects an objective to be optimized and modifies Segi according to the above corresponding operator. In this way, different service selection operations to optimize different objectives can be performed with equal probability on a solution.

4.4.2. Local search for amount distribution

Regarding total cost and reliability, collaboration solutions will be better if the relatively low-quality service among the selected services is distributed to less processing amount of a subtask. Meanwhile, services are expected to be fully utilized to reduce the finishing time. Therefore, it is preferable to distribute the processing amount in a way that is proportional to the speeds of the services selected for a subtask. Accordingly, a local search for amount distribution is designed, including three objectiveoriented operators, OA1, OA2, and OA3, and a hybrid operator, OA4. The details of the operators are as follows:

(1) OA1 for total cost : For each Segi in Y, the operator finds the service with the highest unit cost in Segi and decreases the corresponding yij according to Eq. (32), where y' ij is the value before decreasing.

(2) OA2 for reliability : For each Segi in Y, the operator finds the service with the lowest reliability in Segi and decreases the corresponding yij according to Eq. (32).

(3) OA3 for finishing time : For each Segi in Y, the operator adjusts all yij belonging to Segi according to Eq. (33), where is the speed of the current service and is the index of each selected service in Segi. Notably, the lowest speed of the internal component services is regarded as the speed of the service chain.

(4) OA4 for all objectives: For each Segi in Y, the operator randomly selects an objective to be optimized and modifies Segi according to the above corresponding operator. As in OS4, different amount distribution operations can be performed with equal probability on a solution.

4.4.3. Competition mechanism

With the iterative execution of the algorithm, the optimization progress of different objectives is usually different. A competition mechanism is developed to assign more computing resources to the objective with more space for improvement. In the mechanism, operators with the same type compete to obtain a higher selection probability. The new probabilities are updated depending on the effects of the operators on the solutions. If an operator improves a solution, it will be incentivized to obtain a higher probability; otherwise, it will be suppressed.

To quantify the effect of operators in the current iteration, the population is updated after the local search, and solutions with the local search operator applied are identified and classified. Taking the local search for service selection as an example, the effect of OSg is formulated as Eqs. (34) and (35), where g (or h) is the index of optimization objectives. Eq. (34) is for objectiveoriented operators and takes the focused objective as the primary factor. Other objectives are included as secondary factors to avoid their excessive deterioration. η in Eq. (34) is a coordination coefficient used to control the weights among the factors; it is greater than 1/3 in order to highlight the effect on the focused objective. Eq. (35) is for the hybrid operator that considers all the objectives. is the set containing the solutions that applied OSg, CS' is the original solution before the local search operator is applied,  and  are the gth objectives of CS and CS' ,  and  are the hth objectives of CS and CS' ,  and ε is a small number preventing the denominator from being 0.

In the above formulas, will increase if OSg improves a solution and will decrease if a solution deteriorates, so that the incentives or suppressions can be achieved. To prevent from being negative or 0, a correction function H() is presented in Eq. (36), where is a small number. in Eq. (37) is the value of in the last iteration. In this paper, is set to 0.01 at the beginning of the algorithm. Since only considering effects in the current iteration may lead to drastic changes in the selection probabilities and thus result in instability of the algorithm, the historical effects of operators are introduced to alleviate the changes. As the probabilities used in the current iteration carry historical information, they are deployed to update the selection probabilities. Eq. (38) presents the formula for updating the selection probability of OSg, where is the selection probability used in the current iteration, and is the updated probability to be used in the next iteration. Without loss of generality, the selection probabilities of the operators are equal at the beginning. Similarly, the selection probability of OAg can be updated. 

《5. Implementation through numerical experiments》

5. Implementation through numerical experiments

Extensive numerical experiments were conducted to test the performance of CMOMA in solving the proposed problem. Each solution of the experiments determined the schemes of the service selection and amount distribution to the selected services, which were simultaneously optimized as the algorithm evolved. All programs for the experiments were coded with Java (JDK 14) and executed on a laptop with 32 GB RAM and a 2.2 GHz Intel Core i7- 8750H under a Windows 10 (64) system. Parallel technology was implemented in the decoding stage for speed.

《5.1. Performance metric and test instance》

5.1. Performance metric and test instance

In the experiments, four indicators were adopted as performance metrics: generational distance (GD), inverted GD (IGD), hypervolume (HV), and set coverage (SC).

(1) GD is a convergence metric defined as Eq. (39), where U is the non-dominated solution set obtained by a specific algorithm, P is the true Pareto-optimal front, and dp,u is the Euclidean distance between solutions p and u. A smaller GD value indicates a solution set with a better convergence performance.

(2) IGD is a comprehensive metric defined as Eq. (40), which simultaneously reflects the convergence and diversity of a solution set. A smaller IGD value indicates a solution set with a better comprehensive performance.

(3) HV calculates the volume covered by a solution set in objective space; it is defined as Eq. (41), where cu is the hypercube constructed by the line segment between u and the reference point as the diagonal. Since Eqs. (22)–(24) have normalized objectives, the reference point in this paper is (1,1,1). A solution set with a bigger HV value is preferable.

(4) SC measures the dominance relationship between two nondominated solution sets U and V, through which the quality of the two sets can be compared. SC is defined as Eq. (42), where C(U, V) is the proportion of solutions in V that are dominated by solutions in U, and u > means that u dominates . If C(U, V) is bigger than C(U, V), U has a higher quality.

Test instances were generated based on the principles adopted in Ref. [40]. Services with high reliability or fast processing speed were given high unit costs, and vice versa. As shown in Table 4, 21 instances were generated to fully evaluate the algorithm, where the number of subtasks came in three sizes: 15, 30, and 45. In each size, the proportion of subtasks using complicated services—that is, composite services and service chains—increased from 20% to 80%, with a step size of 10%. The available periods of resource services (including the component resource services in complicated services) were generated randomly. Each subtask has 50 candidate services in this paper, and up to three services can be selected.

《Table 4》

Table 4 The proportion of different services in test instances.

《5.2. Sensitivity analysis of the coordination coefficient η》

5.2. Sensitivity analysis of the coordination coefficient η

Deployed to update the selection probabilities of local search operators, the parameter η in the competition mechanism is the sole parameter to be tuned in CMOMA, and impacts the performance of the algorithm. Theoretically, a big η value causes the algorithm to concentrate on the effect on the oriented objective when updating the selection probabilities. In contrast, a small η value causes the algorithm to notice the effects on the other objectives.

To study the impact of η on CMOMA and select a suitable value of η, a group of experiments was conducted. First, four values, 0.4, 0.6, 0.8, and 1.0, were tested to roughly determine a reasonable scope of η. At each value, CMOMA was independently executed 20 times on the medium-sized instance 11; the running time for each execution was 15 s. After preliminary experiments, the population size was set to 200. Fig. 5 shows the means of the IGD with the standard deviations. The true Pareto-optimal front consists of the non-dominated solutions obtained by running CMOMA at all η values. As shown in Fig. 5, η > 0.8 is a desirable scope for CMOMA, where the algorithm maintains a preferable performance. In other words, when updating the selection probabilities, considering more effects on the oriented objective benefits the performance of the algorithm. To refine η, further tests were performed in [0.8, 1.0] with a step size of 0.02. Fig. 6 depicts the means of HV with the 95% confidence intervals. Although it is not apparent, the g in [0.9, 1.0] has superiorities over that in [0.8, 0.9]. However, with an increase in g, the performance of the algorithm deteriorates. In particular, at the point η = 1, where no effect of other objectives is included, noticeable performance degradation can be observed. This phenomenon suggests that the effects of both the oriented objective and other objectives are essential, and a balance between the two aspects is required. Among the tested values, when η = 0.9, CMOMA reaches a higher mean of HV with a narrow confidence interval. Therefore, η is set to 0.9 for the tradeoff. In addition, from a global perspective, the trajectory of HV shows slight fluctuations but generally remains stable when η > 0.8, illustrating that η is insensitive if it is within an appropriate scope, which supports the robustness of CMOMA.

《Fig. 5》

Fig. 5. Means of the IGD with standard deviations at different η values on test instance 11.

《Fig. 6》

Fig. 6. Means of HV with 95% confidence intervals at different η values on test instance 11.

《5.3. Performance verification of the competition mechanism》

5.3. Performance verification of the competition mechanism

To verify the effectiveness of the competition mechanism, a variant of CMOMA, fixed-probability multi-objective memetic algorithm (FMOMA), was developed for comparison; this variant removes the competition mechanism and adopts fixed equal probabilities to select local search operators. Both algorithms were repeated 20 times on each instance with the same running time for each repetition (10, 15, and 20 s for small, medium-sized, and large instances, respectively). Table 5 presents the means of the GD, IGD, and HV. To calculate the IGD, the results of three well-known multi-objective algorithms—namely, NSGA-II [48], strength Pareto evolutionary algorithm-2 (SPEA-2) [49], and multi-objective Grey Wolf optimizer (MOGWO) [50], which will be discussed in the following section in detail—were deployed to construct the true Pareto-optimal front. An independent two-sample t-test with a 0.05 significance level was performed on each instance. The symbols ‘‘+” and ‘‘=” are respectively used below to indicate that the results of CMOMA are statistically superior or similar to those of FMOMA. It is clear that CMOMA obtains the best GD values on all instances. This advantage is also confirmed from a statistical perspective, demonstrating the preferable convergence of CMOMA. For the IGD, CMOMA achieves the best values on 20 instances, among which the results on 15 instances are significantly superior to those of FMOMA. In terms of HV, CMOMA always yields better results than FMOMA, despite their similar values on seven instances in the statistical sense. To sum up, CMOMA outstrips FMOMA in terms of comprehensive metrics.

《Table 5》

Table 5 Means and two-sample t-test results of GD, IGD, and HV for CMOMA and FMOMA on all test instances.

The symbols ‘‘+” and ‘‘=” are respectively used below to indicate that the results of CMOMA are statistically superior or similar to those of FMOMA.

To assess the quality, Table 6 compares the means of SC. C(CMOMA, FMOMA) are manifestly bigger than C(FMOMA, CMOMA) on all instances, suggesting the advantage of CMOMA in searching for high-quality solutions. To obtain more intuitive insights, Fig. 7 depicts the SC results with the linear regression trend lines. The superiority of CMOMA becomes apparent as the instance size increases. In particular, on large-sized instances, approximately half of the solutions of FMOMA are dominated by those of CMOMA. On instances with the same scale, a similar conclusion is observed. As the proportion of complicated services increases, the advantage of CMOMA is enhanced. This can be primarily attributed to the competition mechanism enabling the algorithm to find appropriate selection probabilities for local search operators. Combining all the above analyses confirms that CMOMA surpasses FMOMA and that the designed competition mechanism contributes to the performance.

《Table 6》

Table 6 A comparison of the means of SC between CMOMA and FMOMA on all test instances.

《Fig. 7》

Fig. 7. SC results for CMOMA and FMOMA on all test instances.

《5.4. Comparisons with other algorithms》

5.4. Comparisons with other algorithms

To further evaluate the performance of CMOMA, comparisons were made with both classical and emerging multi-objective algorithms: namely, NSGA-II, SPEA-2, and MOGWO. The experiment settings remained the same as those described in Section 5.3. Moreover, the Taguchi method was executed to tune the parameters of the comparison algorithms for optimal performance. For NSGA-II, the crossover rate is 1, and the mutation rate is 0.02; for SPEA-2, the crossover rate is 1, and the mutation rate is 0.03. MOGWO requires no predefined parameter.

Tables 7–10 present the comparison results of the GD, IGD, HV, and SC. It can be seen that CMOMA achieves the best results in all metrics, showing impressive performance. Regarding the GD, CMOMA significantly outperforms NSGA-II on 20 out of 21 instances and significantly outperforms SPEA-2 and MOGWO on all instances. In terms of the IGD, CMOMA overwhelms the comparison algorithms. NSGA-II and SPEA-2 show a comparable performance with each other, outstripping MOGWO. However, their results are far inferior to those of CMOMA, particularly on large instances. For HV, the results of CMOMA on all instances remain at around 0.8, which means that the non-dominated solutions of CMOMA consistently cover a large volume of the objective space. Conversely, the performances of NSGA-II, SPEA-2, and MOGWO are inferior, and significant performance degradations can be seen on medium-sized and large instances. As for the quality comparison, CMOMA has tremendous advantages. The solutions of CMOMA dominate most of the solutions of SPEA-2 and MOGWO. Except for the first instance, the solutions of CMOMA are much better than those of NSGA-II, particularly on large instances.

《Table 7》

Table 7 Means and two-sample t-test results of GD for CMOMA, NSGA-II, SPEA-2, and MOGWO on all test instances.

《Table 8》

Table 8 Means and two-sample t-test results of IGD for CMOMA, NSGA-II, SPEA-2, and MOGWO on all test instances.

《Table 9》

Table 9 Means and two-sample t-test results of HV for CMOMA, NSGA-II, SPEA-2, and MOGWO on all test instances.

《Table 10》

Table 10 A comparison of the means of SC between CMOMA and NSGA-II, SPEA-2, and MOGWO on all test instances.

The above analyses reveal a phenomenon in which the superiority of CMOMA becomes more apparent as the instance size increases. To further explore this phenomenon, non-dominated solutions obtained by randomly running the algorithms on three instances of each size are shown in Fig. 8. On the small instance, NSGA-II reaches a limited Pareto front and presents a relatively good distribution, surpassing SPEA-2. However, on the mediumsized and large instances, the performance of NSGA-II degrades palpably. The solutions of NSGA-II, SPEA-2, and MOGWO are close and concentrated in a small area, failing to explore the objective space adequately. In contrast, the solutions of CMOMA on all instances occupy large spaces and cover most of the solutions of the comparison algorithms. This advantage mainly stems from the local search operators and the competition mechanism, which allow the algorithm to search broadly and efficiently.

To recap, this paper illustrates the efficacy of the proposed DMSCO methodology by preferentially investigating a sequential pattern in which the subtasks of a task are fulfilled in sequence. This is a fundamental pattern that can be effortlessly extended to other patterns. For example, to accommodate a pattern in which parallel subtasks exist, all research can be adopted with just two minor modifications. First, it is necessary to convert Eq. (20) F = FI to F and Eq. (21) BiFi1 ≥ 0 to Bi ≥ 0, where is the index of subtasks and subtask STi should be started after (as more than one precursor could exist with a parallel relationship before STi). Second, in the encoding scheme, it is necessary to ensure that a subtask is placed after all its precursor subtasks. In this way, this paper shows how DMSCO improves service collaboration solutions while being capable of providing references for other patterns and retaining extensibility.

《Fig. 8》

Fig. 8. Non-dominated solutions obtained by a random running of CMOMA, NSGA-II, SPEA-2, and MOGWO on small test instances (2, 4, and 6), medium-sized test instances (9, 11, and 13), and large test instances (16, 18, and 20).

《6. Conclusions》

6. Conclusions

With the application of Industrial Internet platforms in the manufacturing industry, high-quality manufacturing service collaboration solutions are desirable. This paper proposes a DMSCO methodology to optimize service selection and amount distribution in order to improve the solution quality. In this work, manufacturing services are modeled at three granularities, and their components and structures are discussed to ensure the feasibility of collaboration solutions. Unlike previous works, the availability of a manufacturing service is formulated as a series of accessible periods of the physical resource corresponding to the service. In particular, the availability of a complicated service is affected by the component services and their relationships. The DMSCO model is then formulated to simultaneously optimize service selection and amount distribution. A CMOMA embedding a competition mechanism is developed for searching for solutions to the model. Extensive experiments are conducted, and the results suggest the superiority of the proposed algorithm. To be specific, CMOMA outperforms FMOMA that is without the competition mechanism in all considered metrics, thereby verifying the effectiveness of the designed mechanism. Compared with NSGA-II, SPEA-2, and MOGWO, CMOMA also exhibits excellent performance. Moreover, as the size of the instances increases, the superiority of CMOMA over the comparison algorithms becomes evident.

This paper aims to introduce a novel DMSCO for Industrial Internet platforms, thereby providing fresh insights to stakeholders. Motivated by studies on optimal service selection, the developed DMSCO is comprehensible for relevant practitioners and easily compatible with existing processes. In parallel, the results of the experiments in this study reveal the proposed method’s potential for practical applications in scenarios with complex production processes and large production amounts, such as automobile, equipment, and household appliance manufacturing. A link in the production process can be assigned to multiple qualified manufacturing services, thus enhancing the overall service performance. However, it should be noted that the manufacturing service modeling in this paper focuses on physical processing resources, while Industrial Internet platforms also contain other types of resources, such as software and logistics resources. For future work, the proposed method will be validated and evaluated through real-world scenarios and experiments. Research on enhancing the expressive ability of manufacturing service modeling and the corresponding service collaboration model is recommended. In addition, this paper mainly focuses on collaboration solutions that satisfy customer requirements. Since the interests of service providers are also crucial, it is worth studying how to simultaneously improve the satisfaction of both sides.

《Acknowledgments》

Acknowledgments

This work is funded by the National Natural Science Foundation of China (51905396) and the China Scholarship Council.

《Compliance with ethics guidelines》

Compliance with ethics guidelines

Shibao Pang, Shunsheng Guo, Xi Vicent Wang, Lei Wang, and Lihui Wang declare that they have no conflict of interest or financial conflicts to disclose.