Time Factor in Operation Research Tasks for Smart Manufacturing Time Factor in Operation Research Tasks for Smart Manufacturing

The shift to the concepts Industry 4.0 and IIoT helps collect a vast amount of objective data about processes that take place in a production system, and thus, it creates back-ground for taking advantage of theoretical results in practice; it is a trend towards synchronizing production system processes and external (market) processes in practice. In order for the target to be achieved, we use the methods that formalize management tasks in the form of predictive models, consider the cases with the computational solution of management models and decision making in production system tasks which are set based on time factor and are solved by approximate methods. We also take a look at the problems of probabilistic nature of gained decisions and address the cases, when by computational solution of tasks we need to take into account restrictions and select time step in order to obtain the decision in a table form of the function of time. The problems that we investigate help obtain and solve management tasks of production systems with help of forecasting data for a group of indices that are involved in decision making – this all helps enhance the sufficiency and quality of management decisions.


Introduction
Presently, the information support of production systems management is mainly focused on the control and management of production systems (SCADA), the support of sales and production process (ERP, MRP, Just in Time), organizing production for known customers (CSRP), and product life cycle management (CALS). However, the aspects of tactical and strategical management get information support only on the stage of data preparation for decision making, yet on the stage when potential solutions are to be identified based on data we observe lack of information support. The integration of automatization and control systems, the trend towards Industry 4.0 and IIoT leads to an exponential growth of collected data. Hence, on the one hand, it can be expected that the use of big data might help obtain fundamentally new solutions due to their immense nature, but on the other hand, the issues of decision support automation become more acute as we face the trend of an ongoing automation expansion of production systems, and hereafter, can assume a concept of a virtual factory.
The use of the concepts Industry 4.0 outlines possibilities for the automation of production systems management taking into account the interaction of subsystems and the synchronization of their interaction with external factors. In the age of cutting edge innovation products we cannot talk about the stability of production processes since life cycle of such products is short, the number of modifications and parts is high, and power intensity and resources consumption is much higher. This proves the necessity of collecting reliable information with help of IIoT. The presence of such data helps build predictive models and use preventive control actions as production system is an inertial management object that is not able to adjust the ongoing processes instantaneously. Besides, the change of processes requires additional time resources, financial resources, labor competence, and organization resources.
The implementation of the concepts Industry 4.0 and Industrial Internet of Things [that deals with collecting information about each production unit and provides operation management over production processes in PS] [1] opens new possibilities for developing industrial engineering methods [2].
Taking into account long decades, when production systems were examined only on the basis of general data, data engineers had limited data to develop methods for decision making and took advantage of expert evaluations, i.e. the methods of utility theory (considering customer preferences as maximization of expected utility, probability models (see the works of O. Morgenstern), axiomatic theory of D. Savage that enables measure the utility and subjective probability simultaneously; decision tree approach that partitions the tasks into certain subtasks (look the works of H. Reif); multiple-criteria utility theory (developed in the works of R. Keeney); prospect theory methods, Electre methods (worked out by the French School on MCDA headed by B. Roy), hierarchy analysis method proposed by Saaty [3], heuristic methods (for instance, the method of the weighed sum of its evaluation ratings, compensation methods etc.), the models of bounded rationality by A. Rubinstein, the technic for order of preference by similarity to ideal solution (TOPSIS) [4].
The appearance of a big amount of statistical data encouraged the development of the methods of mathematical formalization used to solve tasks for the management of materials, parts, operations, and choice of suppliers [5] with the consideration of stochastic factors, probability approaches to measure risks taking into account different nature of examined events (joint, correlative, inconsistent and interdependent) used to solve planning tasks taking into account the dynamics of examined processes.
The consideration of random factors and the use of probability approaches help measure risks with help of models. There are planning risks (the risks related to decision making based on models [6] that depend on the current state of market (change in price, sales volumes etc.)) and production risks (the risks related to equipment mortality, failures in the delivery of necessary materials or parts etc.).
The use of probability models is based on the use of risk metrics [7], Bayes' Theorem [8] or Monte-Carlo Method [9].
2. Methodology aspect of management task setting in production systems Production system is regarded as management object that is placed in a state space. The coordinates on this n is the dimensional space are represented by the management parameters that are considered significant for achieving the targets, and their values describe the current state and remoteness from the selected targets.
If we mark target goal indexes by the vector P p , and the current state by the vector P a , we will receive a mathematically measurable metric P p ; P a À Á that shows how the current position deviates from the goal position that is deemed a sign of progress for project implementation (the end of implementation, P p ¼ P a ). However, to know the metrics P p ; P a À Á is not enough for management, we also need to know the vector of the parameters Y that greatly affect the state of project and consist of the values that describe project, production system and the environments in which project is implemented as well as dynamics of change and prognostic values of all these parameters. It should be noted, that the achievement of the goal values P p ¼ P a does not always mean the achievement of the vector values Y expected for this state.
In management tasks values and parameters can be classified in four groups [10]: parameters and values that describe a current state P i ð Þ p , values and parameters that describe the action (external factors and control action -Y ¼ A ∪ Θ, the A is the set of control actions, Θ is the set of environment values), values and parameters that describe a goal state P i ð Þ a , values and parameters that describe the output of system operation by shifting from the state P i ð Þ Therefore, management has to use an automaton where the consecutive state is defined by experts based on the current state and the state that was planned to be achieved on the previous stage and the time when it has to be done -P 0 In order for a new state to come, action A i ð Þ has to be defined. We can determine such action with help of the production system model that implements innovation projects w j ¼ U; S f g, where U is the vector of management parameter, S is the set of project resource needs, j is project number.
This approach helps work out hierarchically coordinated managerial decisions by taking into consideration system-interrelated external and internal factors that interact. Management process is considered then as a holistic undetermined process.
In general, the model can be presented in a form of a tuple: where w ¼ w 1 ; w 2 ; …; w n f gis the projects' vector, Y is the action vector on each step, R is the outcome vector on each step, P p is the vector of system states, P a is the vector of system goal states, T is the vector of decision points.
The use of the model (1) is described by an undetermined algorithm [11] see Figure 1.
As a result, management task becomes more transparent. However, it opens new sub-tasks, i.e. to determine decision points, to define the set of indexes and their values for each stage of project implementation, to build a model of production system by implementing the projects (w) in order to define the vector of control actions Y.
At the same time, the more formalized is the description of tuple parts (1) (less ambiguity), the higher is the quality of management [according to system properties]. Decision points can be defined in case if we know the set of controlled parameters [12], and have additional information that characterizes the production system that we manage (equipment maintenance periods, internal technological cycles etc.) [13].
The setting of management tasks taking into account time factor T i ð Þ leads to formalizing the The structure of the model sets formal interrelations between its parameters, and on each step the type of the model will depend on the managerial task that we consider (whether we forecast properties and behavior of the investigated management project; or when dealing with object management we select best actions by testing them on the model, investigate the object and look for the ways to improve management object).
The model itself can use both non-causal (component-oriented) and causal (block-oriented) modeling, and model components can set requirements to their development tool (for example, the possibility to 1) work with big data volumes set by time series 2) use the methods that are applied for incomplete data 3) solve tasks set in a form of mathematical programming 4) employ methods to work with probabilistic models etc.).
The specialization of models Y i ð Þ brings the problem of choosing approaches and ways for formalization based on the set of already known approaches, ways, methods and models [14] that will be collected as a composition (the compatibility of input and output areas).
For the implementation of each project in the considered production system, the model formation that is presented in a general form is as follows R; w; where P is the vector of external parameters that exert impact on the system, Π is the vector of system parameters, m i ð Þ j is the components or blocks of the model for time T i ð Þ ).
Despite the apparent simplicity of the approach, underlying this approach is a necessity to work out managerial decisions taking into account different levels (institutional, managerial, technical) and management types (finance management, production management, goods management, launch management, sales management, R&D management, institutional management), and subsystems of production systemall of that generates a whole group of managerial tasks that have to be solved together for each time period T i ð Þ ; the interrelation of the tasks is demonstrated in Figure 2.
Work with a model structure means that we need to consider several subtasks related to forecasting parameters of the considered project [15] and to formalizing an optimization task in a form of mathematical programming [16].
The examples of tasks that are considered in decision points can encompass the tasks of production and client analytics taking into account time factor, such as demand forecast and sales planning, volume planning, stock and procurement planning (including working life), equipment selection taking into account maintenance costs; these can be the tasks of optimizing stock work and minimizing the volumes of working assets, and obtaining optimal machine utilization and work force.
In this case, each of tasks can be described by a separate criterion; the use of a reflexive approach enables their joint solution as a set of optimization tasks that have common parameters and use forecast-based data.

Solving management tasks with help of predictive models
Let us now consider a general task of formalizing management processes for project implementation in PS. This task can be handled as a task of defining decision points and a cyclic solution of prognostic models that are represented by optimized formalizations based on forecast data and elaboration done on each step of processing data in order to make consecutive iterations with new data, and calculation results.
In order for the tasks to be formalized as tasks of optimal control, we have to input a set of indices, variables and parameters of management [9], for instance, like: i is the supplier's index; j is the index of production system/stock (PS); m is the part index or the demand in materials; n is the index of end item; k is the index of production operation; g is the index of machine or instrument; p is the index of operation; t is the time; x ijm is the number of parts m received from the supplier i for PS j; y jn is the number of parts n produced in PS j; r n is the number of returned items n for utilization; o m is the number of reused parts or materials m; d m is the number of items or materials m sent to utilization; ref jm is the number of reused items or materials m in PS j; bd n is the binary variable that possesses the value equal to 1 in case if it can be repeatedly used for the item n and 0 if not; Δt is the time step; sell n is the item's market price n; cost jn is the item's production cost n in PS j; price im is the price of the part m received from the supplier i; ship m=nij is the delivery cost of the part/ item m=n from the station i to the station j; inv j is the storage cost in PS j; setdis n is the preparation cost to get the parts out of the item n; disa m is the preparation cost to get the part m out for reuse; disp m is the utilization cost for the part m; refcost jm is the preparation/recovery cost of the part m for reuse in PS j; dem j ð Þn is the need/demand in the item n, if there is the index j the consumer get then j; req mn is the number of requested parts m required for the production of the item n; costeq pgj is the cost of the operation p on the equipment g in PS j; timeeq pgj is the time of operation performance p on the equipment g in PS j; part mpgj is the demand in parts/materials m in order to perform the operation p on the equipment g in PS j; eq pgj is the demand in the equipment g in order to perform the operation p in PS j; supmax im is the maximum size of the batch of the parts m that can be delivered from the supplier i; supmin im is the minimal size of the batch of the parts m that can be delivered from the supplier i; supmaxpart jm is the maximum potential number of parts and components m that can be delivered for production in PS j; supmaxeq jp is the maximum potential number of equipment units for the operation p in PS j; reuse m is the maximum percentage of the parts m that can be reused.
The approach described above helps state a set of optimization tasks that can be considered both, as joint and separate tasks. Let us give the examples of feasible task formalizations: • Profit maximization (production planning for demand), sell n t ð Þ À P j cost jn t ð Þ P j y jn t ð Þ ! max, ∀n; • Production cost minimization, • The minimization of costs for goods' storage, cost jn y jn t where y 1jn -the number of items stored in stock, y 2jn is the number of items sent to consumer; • The selection of suppliers taking into account that certain components can be reused, P j P n sell n À cost jn The tasks can be subject to different restrictions: • Production capacity restriction, P g eq pgj t ð Þ ≤ supmaxeq jp t ð Þ, ∀j, p, t; • The restriction related to delivery options of components and materials, P g part mpgj t ð Þ ≤ supmaxpart jm t ð Þ, ∀j, p, m, t; • Non-negativity restriction on the volumes of goods, orders etc., y jn t ð Þ, x ijm t ð Þ, r n t ð Þ, o m t ð Þ, • Demand volume restriction, P j y jn t ð Þ ≤ dem n t ð Þ, ∀n, t; • The description of technological process, P n req mn y jn t ð Þ ¼ • The restriction on the volume of orders, • etc.
The obtained tasks in their general form refer to a class of multi-parameter tasks with nonlinear restrictions. In such tasks a part of parameters is set by time functions. The outcome of the solution of such tasks will be the function of time as well (by numerical solution in a table form). Since today we lack analytical methods to solve such tasks, we will build then the solution of this task on multiple cyclic determination of numerical solutions of a multiparameter optimization task with the time period Δt ≤ min i¼1, n T iþ1 ð Þ À T i ð Þ that determines the accuracy of the description of the required function (see Figure 3).
Taking gradient calculation for finding solution was one of the first approaches to develop solution methods (gradient search method with the split of the step метод градиентного поиска с дроблением шага, steepest descent method, conjugate direction method, the Fletcher-Reeves method, the Davidon-Fletcher-Powell method). By these the goal function has to be differentiated two times and convex. The Newton's method and his modification the Newton-Raphson method is widespread. These methods also set the requirements to the goal function to be differentiated two times and be convex. Besides, these methods are sensitive to the selection of initial value. Moreover, in obtained optimization tasks we the cases can appear that are related with multiextremality, non-convex restrictions, multicoupling of the area of feasible solutions etc., and these methods cannot handle that appropriately. Modern methods can in general be split into three groups [17]: cluster methods, the methods of restrictions' distribution, metaheuristic methods. By choosing the solution method it is important to consider that the most significant feature of combinatoric optimization methods is their completeness and comprehension. A complete method ensures the finding of the task solution if it exists. However, the application of these methods can bring difficulties by a big dimension of search space, and we might not have sufficient amount of time that will be required for search in this case (for instance, due to time restrictions for decision making). If we use heuristic methods in task solutions, and heuristic elements complement combinatoric methods, it is getting more complicated to prove that the applied method is comprehensive. The methods of heuristic search are, in general, incomplete.
In practice hybrid algorithms are often used. Besides, the outcome of any algorithm work can be improved by building a joint solver. Due to the lack of specialized solution methods, for the obtained formalization we assume that we can use a developmental approachthe method of stochastic search. The drawback of developmental approaches is that in some cases the results and optimization time are dependent on the selection of initial approximation. This drawback can be eliminated by using as an initial approximation the solution that was worked out by experts. That is why, as a universal solution we suggest to use the method of stochastic search taking into consideration expert knowledge and indistinct preferences. However, in this case we need to direct attention to the fact that for some tasks we can obtain formalizations that already have methods of their solution. Hence, the decision about what method to apply should be taken dependent on the targets, i.e. how accurate the solution is expected to be and whether we have time restrictions for solution search (the methods of stochastic search can be limited in time required for solution search, which is crucial in integrated systems and IIoT that operate in real time).
In heuristic methods of random search we can distinguish two big groups: the methods of random search with learning and developmental programming [18]. In practical use the methods differ in convergence speed and the number of iterations required for search of a feasible solution (several methods, for example, genetic methods, ensure finding an extremal value, but not obligatorily an optimal one). The complexity of selection task is that the efficiency performance of certain methods of stochastic search (in particular, genetic algorithm) is determined by their parameters. As an example let us examine the application of the method of random search with inhibits (Pareto simulated annealing) [19]; along that, we take into account the set values, that were obtained by forecasting during the modification of task for work with restrictions. Before we start perform numerical calculation we need to determine the area for feasible solutions. The algorithm will consist in five steps and an additional sixth step; the latter step allows solve tasks with the restrictions set by functions and forecast values with the set accuracy and the criterion that can also use the values obtained by forecasting.
Let us now consider the search option of parameter values x i , i ¼ 1, N as points in space B i Let us assign Λ * * to the set of all points x i , that comply with the task restrictions: 1. Set N * * is the requested number of points from the set Λ * * (N * * is the parameter of algorithm). Depending on the certain task, the value N * * can alter.

2.
Find N * * points for each parameter x i ∈ Λ * * , scattered in the spaces B N ð Þ i randomly or by the use of expert knowledge, and use these points as an initial approximation.
3. For finding the solutions x i ∈ Λ Di (Λ Di is the set of feasible points) apply one of the heuristic methods of stochastic search. For this purpose, the point x i ∈ Λ * * is taken as a base point, and based on this point we build new points belonging to Λ * * where the criterion values are better than in a base point. Even if one such point is found, its base is used then for finding new values etc., and next search is done. All the points found this way x i ∈ Λ * * make the set Λ Di .

4.
All points x i ∈ Λ Di are studied for optimal factor, after that they are used to form an optimal set of solutions Λ P . The required sets are easily recovered from the labels of criteria in spaces.

The selection of the singular variant
Þ , from the Pareto-set is submitted to an expert, that has additional information that has not been formalized and neither taken into account in the model.

6.
For an operational reaction to altering external factors we should perform several iterations for task solution (by modeling the deviations of forecasting values within the confidence interval) and do that cyclically with the time period Δt.
As a result, we receive altering in time span (corridor) of potential solutions for each time period. At the same time, as several functions describe the parameters that are set by forecasts, where accuracy depends on the planning horizon, we can encounter the case, when the obtained values can fluctuate either towards the increase or the decrease. Such behavior will bring additional organization expenses for PS; however, it is possible to manage such behavior (smoothly adjust the altered values) by changing the dimension within the obtained corridors and the time step Δt (as a rule, such deviation is described by a stochastic variable that obeys normal distribution law).
In the result of the solution we can determine the diapasons and the values of the values that can be presented in a suitable way to the decision-maker (for instance, in a form of the Gantt chart that is so widespread in management) [20].

The generation of the area of feasible solutions by solving the tasks for optimal control of projects and production systems
By the implementation of management tasks as dynamic management tasks, where the solution is the function of time, it should be noted that the restrictions can also change in time. It happens as the characteristics of production system can alter in time, the changes can affect the schedule of supplies, the volume of resources allocated for the implementation of a certain project etc. The restrictions can be shown as follows: where M is the parameter or an expression with imposed restrictions, m 1 t ð Þ and m 2 t ð Þ is the restrictions set by the functions of time, m 3 t ð Þ is the area of feasible values can also alter in time.
The use of several criteria and a big number of restrictions often leads to the situation that we obtain an empty area of feasible values or the solution shows some deviations. In any case, in PS management tasks the final decision is taken by the expert. That is why, the restrictions can be presented by the functions F m 1 t ð Þ; m 2 t ð Þ; t ð Þ that can be represented in a form of additional criteria and used by performing the operation of criteria compression.
In case of discrete set values or if restrictions are set as an area of feasible values, the function Þbecomes piecewise-set. Hence, the values that belong to a feasible interval are maximum high by considering a maximization task, and the others become maximum low and vice versa by considering a minimization task. In general, for the consideration of all types of restrictions in one record the function can be written as F m 1 t ð Þ; ð m 2 t ð Þ; m 3 t ð Þ; tÞ: The membership with the area of feasible values can be validated then by calculating the value: where n is the number of restrictions.
If this value is equal to the sum of minimal or maximal values P n i¼1 min=maxF i m 1 t ð Þ; ð m 2 t ð Þ; m 3 t ð Þ; tÞ dependent on the type of the considered task (minimization or maximization), then it will belong to the area of feasible values. In practice, the restrictions can be considered not as stiff and we can determine the feasible deviation of values ( ∓ Δ).
Such approach helps add restrictions to a criterial function as additive components that allows get rid of restrictions and apply for solution the methods that do not work with restrictions. Since restrictions can be destroyed in this case, so the obtained functions are to be ranged with help of weight coefficients K. As a result, we receive a final setting of the task for extremum in the following form: where J is the criterial function.

The problems of obtaining solutions as functions of time
By solving tasks of optimal control taking into account time factor and some discrete time step Δt the solution will be a set function presented in a table form. In this case, the system interacts with the external environment and the found solution can be not achievable due to the changes of external or internal factors. According to Bayes' theorem [21] the probability of a successful transfer to another state (to a new solution) will depend on the previous state (the state that we are placed now). Hence, for selecting the path for project development it is useful to consider not just one solution, but a set of solutions that are Pareto optimal. So, the task solution will be a set of development paths that technically can be shown as a tree for each of the required parameter values (see Figure 4) that can be considered as Bayesian network.
The selection of a singular solution will be based on the choice of a path and on the potential of its implementation. The potential of each solution will be defined by chain rule [21]: Therefore, by the planning horizon in mΔt and n solutions on each step we will obtain Q m j¼1 n probabilities for leaf nodes in the built tree that should satisfy the following conditions P n i¼1 P X for each solution step.
If we assume that all X are unique, then the implementation potential for each solution will be equal. However, in practice solutions can repeat. It is connected with the fact that we use the method of random search for solving a task; more than that, for modeling deviations we need a multiple solution of a considered task. In this case, the probability of a transfer from the state X 0 ð Þ into the state X m ð Þ will be determined by the sum of probabilities of repeated values, and this value will determine the probability of a transfer from one decision point to another one.
This probability will not be a random value since multiple calculations are performed, as parameters that are obtained based on forecast data can have random walk described by the functions of probability density; the latter ones are necessary to be used for generating new forecast values by multiple calculations.
where σ 1 is the standard deviation, x 1 is the value obtained by forecasting. By a transfer to the consequent value the function will alter: in a new formula we add σ 2 2 is the Gaussian perturbation of constant dispersion that is calculated by the formula [22]: where D x ½ is the dispersion, M x 2 Â Ã is the mathematical expectation, x 1j is the possible values for x 1 (belonging to the interval σ in order to perform the validation for adequacy).
As a result, it is possible to define the probabilities of obtaining solutions and select the most probable ones.
The use of the probability density functions for modeling deviation helps measure the achievement probabilities of a series of consecutive states s 1 , s 2 , ⋯, s n . If the probability p 0 ð Þ 1 indicates that we are placed in the state s i and the state fully complies with the expected state (determined on the basis of previous stages), p ij shows the probability of the transfer from the state s i into the state s j , and p 1 ð Þ i indicates the probability that the state s i will be achieved.
and the management task adds up to the selection of a desired state from the set of possible states and the determination of a path (the set of delta states) to achieve this desired state. Therefore, it is possible to define the probabilities for obtaining decisions that will be taken into account for further selection of the most probabilistic ones based on the method of dynamic programming (Bellman method) (see Figure 5). Each state is determined by a risk metric (a value that is calculated on the base of the probability p ij depending on the path that we have taken to land at the examined state) and the dynamics of the change in the criterion value by the transfer from one special state into another one (see Figure 5).
By obtaining the solutions as the functions of time on each step of calculations the time step Δt becomes an important algorithm parameter. On the one hand, as a step we can choose the time between the decision points T iþ1 ð Þ À T i ð Þ , from the other hand, by such approach the sensitivity of the system to altering external factors is decreasing (it becomes inertial). That is why, the selection of time step will be a trade-off between sensitivity and persistence of system. At the same time, time step can be an altering dimension (Δt ¼ f t ð Þ) but it should be placed in the Underlying a new calculation is the output of values of a forecast parameter outside the bounder of the confidence interval AEσ. On the other hand, works related to changing production capacity, production and procurement scheduling etc., bring additional expenses for enterprise (in general, we encounter the situations, when production capacity is to be increased first and decreased afterwards, that in some cases can be balanced, particularly, by stocks. Therefore, we should consider this task as a separate management task and use the algorithm shown in Figure 6. The solution for the examined diapason T iþ1 ð Þ À T i ð Þ by, for instance, joint consideration of the tasks of volume planning and procurement management will be production plan, the value of the given criterion (with a potential deviation diapason of decisions), the value of risk metrics and the volumes of changes in required parts and components taking into account possible deviations from target production volumes (Figure 7).

Conclusion
The present chapter describes the approaches that thanks to the use of the concepts Industry 4.0 enable the formalization of the processes that are connected with the reasoning and Figure 7. The solution results of volume planning and procurement management tasks based on the collected data about production system for a discrete production: (a) an example of production output volume for one of the products by the use of different forecasting methods, (b) the values of risk metrics (solid line) and progressive risk metrics (dotted line) connected with the use of planning data, (c) adjusted criterion value by the use of best forecast results and the corridor of possible deviations by the use of normal distribution for their modeling and its correlation with the retrospective databased criterion value, (d) the need in one type of parts taking into account possible deviations in production plan.
preparation of managerial decisions which are based on real statistical data that take into consideration the interaction of subsystems in production system. Therefore, together with the use of predictive models IIoT helps not only enhance the level of automation and reduce a certain part of personnel production expenses but also consider such factors as increasing power intensity and resources consumption of productions, inertness of integration and management processes in production systems, and the situations that are connected with repair actions, equipment mortality, procurement failures, change in demand and prices etc.
We have investigated the question how to use and apply under existing conditions the approaches that search feasible and optimal solutions in the tasks of efficient management and planning (taking into account time factor). The changes that affect the setting and solution of tasks can be explained by the shift to automated and automatic enterprises, by the shift from mass production to single-part production. In this connection, the current situation requires operational rearrangement of ongoing production processes; we need to increase global economics mobility, i.e. the variability of external environment where production systems operate.
The approach that is described in the chapter is relevant as it tackles management tasks given as optimization tasks; besides, it helps deal with the phenomenon of NP is the completeness of obtained tasks.
The obtained results are sensitive to the quality of forecasts and lack time lags; more than that, we can observe a change in production volume that creates additional increased capacities for production system (related to the change in production schedule).
That is why, the shift to the concepts Industry 4.0 gives not only evident momentary advantages, but also outlines new areas for studies, i.e. the solution of tasks that take into consideration the inertness of production system and expenses that arise due to changes in production volume and risk metrics, that appear upon interaction with external systems (for example, delayed delivery, the delivery of faulty parts, return of goods etc.).
The development of mathematical formalization of these areas of studies can lead to additional effects in future and underlie the appearance of industrial concepts of next generations.