This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Mixed-model Sequencing with Stochastic Failures: A Case Study for Automobile Industry

I. Ozan Yilmazlar [email protected] Mary E. Kurz [email protected] Hamed Rahimian [email protected] Department of Industrial Engineering, Clemson University, Clemson SC, 29634
Abstract

In the automotive industry, the sequence of vehicles to be produced is determined ahead of the production day. However, there are some vehicles, failed vehicles, that cannot be produced due to some reasons such as material shortage or paint failure. These vehicles are pulled out of the sequence, and the vehicles in the succeeding positions are moved forward, potentially resulting in challenges for logistics or other scheduling concerns. This paper proposes a two-stage stochastic program for the mixed-model sequencing (MMS) problem with stochastic product failures, and provide improvements to the second-stage problem. To tackle the exponential number of scenarios, we employ the sample average approximation approach and two solution methodologies. On one hand, we develop an L-shaped decomposition-based algorithm, where the computational experiments show its superiority over solving the deterministic equivalent formulation with an off-the-shelf solver. Moreover, we provide a tabu search algorithm in addition to a greedy heuristic to tackle case study instances inspired by our car manufacturer partner. Numerical experiments show that the proposed solution methodologies generate high quality solutions by utilizing a sample of scenarios. Particularly, a robust sequence that is generated by considering car failures can decrease the expected work overload by more than 20% for both small- and large-sized instances.

keywords:
Stochastic programming , Mixed-model sequencing , Branch-and-Benders-cut , Heuristics , Tabu Search
journal: ArXivt1t1footnotetext: This research did not receive any specific grant from funding agencies in the public, commercial, or not-for-profit sectors.

1 Introduction

Mixed-model assembly lines (MMAL) are capable of producing several configurations, models, of a product. The number of models increases drastically as the complexity and customizability of the product expand. The number of theoretical configurations of vehicles from a German car manufacturer is up to 102410^{24} (Pil and Holweg, 2004). Different configurations require distinct tasks at each station which induces high variation in the processing times, though each station has a fixed maximum time available. In fact, station workload is distributed through line balancing such that each station’s average workload conforms to this maximum time. When a station has more work allocated to it for a particular model (work overload), interventions are needed to maintain the flow of products in the assembly line, thereby avoiding line stoppages. Interventions can be considered in advance, through sequencing decisions, or at the time of disruption, through utility workers. When these interventions fail, the line may stop production until the situation is resolved. Thus, it is essential to distribute the high work-load models along the planning horizon to avoid line stoppages.

The mixed-model sequencing (MMS) problem sequences products in an MMAL to minimize work overload at the stations. Data from our car manufacturing partner shows that the variation in processing times is high when customization appears on a main part, e.g., engine type: electric, diesel, gasoline, or hybrid. Car manufacturers have adapted their assembly lines for the mixed-model production of vehicles with diesel and gasoline engines. However, the assembly of electric vehicles (EV) in the same line has brought new challenges, while not eliminating the production of vehicles with diesel or gasoline engines. Unlike other vehicles, electric and hybrid vehicles have large batteries which causes a huge difference in tasks, e.g., at the station where the battery is loaded. As the proportion of electric and hybrid vehicles grows in a manufacturer’s mix, the impact of supply problems increases. Sometimes, a part is delayed from a supplier, so a designed sequence of vehicles will have a missing vehicle. Even if this vehicle has a gasoline or diesel engine, its absence may impact the battery-intensive stations. As a manufacturer’s mix of vehicles grows more specialized with more time-consuming content for a large subset, without alternative tasks for the vehicles without the specialized content, the impact of missing vehicles on a carefully designed sequence grows.

Some vehicles in a production sequence may not be ready for assembly on the production day for various reasons, such as the body not being ready, paint quality issues, or material shortage. Such vehicles, referred to as failed vehicles, need to be pulled out of the sequence. The resulting gap is closed by moving the succeeding vehicles forward. This process and the resulting additional work overload occurrence is illustrated in Figure 1 for a battery loading station. The processing time at this station is longer than the cycle time for EVs and shorter than the cycle time for non-EVs, and assume that back-to-back EVs cause work overload. We schedule five vehicles, two electric and three non-electric. One of the non-EVs (third in both scheduled sequences) has a high failure probability. The initial sequences with no failures, while different, both will lead to no work overload. Assuming the third vehicle fails, we have different consequences for the resultant sequence of vehicles. In the non-robust sequence, removing the failed non-EV results in two EVs in a row, which will cause a work overload. However, the robust sequence, which is composed of the same vehicles in a different order, can withstand the failure of the third vehicle without causing a work overload. We refer to this sequence as the “robust” sequence because no work overload occurs when the vehicle with high failure probability is pulled out of the sequence.

Refer to caption
(a) non-robust sequence
Refer to caption
(b) robust sequence
Figure 1: Illustration of a non-robust and robust sequence to stochastic failures

In this study, we generate robust sequences that consider the vehicles’ potential failures to reduce additional work overloads. We focus on the final assembly line, assuming that vehicles follow the same sequence as they arrive from the paint shop and resequencing is not an option; when a vehicle is removed from the sequence, the following vehicles close the gap. The contributions of this study are as follows:

  • 1.

    We provide a two-stage stochastic program for a MMS problem with stochastic product failures, and we provide improvements to the second-stage problem. To the best of our knowledge, this is the first study that considers stochastic failures of products in MMS.

  • 2.

    We adopt the sample average approximation (SAA) approach to tackle the exponential number of scenarios. The numerical experiments show that we can generate robust solutions with an optimality gap of less than 1% and 5% by utilizing a sample of scenarios, for the small-sized and industry-sized instances, respectively.

  • 3.

    We develop an L-shaped decomposition-based algorithm to solve small-sized instances. The numerical results show that the L-shaped algorithm outperforms an off-the-shelf solver, solving the deterministic equivalent formulation (DEF), in terms of both quality and computational time.

  • 4.

    To solve industry-sized instances, we propose a greedy heuristic and a tabu search (TS) algorithm that is accelerated in convergence with problem-specific tabu rules, and in objective reevaluation each time a new solution is visited.

  • 5.

    We conduct a case study with the data inspired by our car manufacturer industrial partner. The numerical experiments show that we can reduce the work overload by more than 20% by considering stochastic car failures and solving the corresponding problem with the proposed solution methodologies.

The remainder of this paper is structured as follows. MMS related literature is reviewed in Section 2. The tackled problem is defined, and the mathematical formulation of the proposed problem is presented in Section 3. Exact and heuristic solution approaches in addition to the SAA approach are presented in Section 4. In Section 5, we execute numerical experiments to analyze the performance of proposed solution methodologies and present the results. Finally, a summary of our findings and a discussion about future work are given in Section 6.

2 Related Work

Manufacturers use various design configurations of MMALs to maximize their revenue. The optimization process of the assembly line sequencing takes these design configurations into account. The first paper that articulates the MMS was presented by Kilbridge (1963). The researchers tackle the MMS with varied characteristics which required a systematic categorization of the components and the operating system of MMS problems. Dar-El (1978) categorizes the MMAL into four categories based on their main characteristics of assembly lines: product transfer system, product mobility on the conveyor, accessibility among adjacent stations, and the attribute of the launching period. An analytic framework for the categorization of Dar-El (1978) is given by Bard et al. (1992). Later, a survey is presented by Boysen et al. (2009), where they define tuple notation for the sequencing problems based on more detailed characteristics of assembly lines, including work overload management, processing time, concurrent work, line layout, and objective in addition to the main characteristics.

Several objectives are employed to evaluate the performance of the assembly line sequence. The most common objective in the literature, also adopted in this study, is minimizing the total work overload duration, proposed by Yano and Rachamadugu (1991). Tsai (1995) describes hiring utility workers to execute tasks so that production delays are avoided, which leads to the objective of minimizing the total utility work duration. Fattahi and Salehi (2009) minimize the total idle time in addition to utility work. Boysen et al. (2011) propose minimizing the number of utility workers instead of the total utility work duration in order to improve utility worker management.

A few exact solution methods are proposed in the literature to solve the deterministic MMS problem. Scholl et al. (1998) proposes a decomposition approach that uses patterns of different sequences, called pattern-based vocabulary building. They use a column generation method to solve the linear relaxation of the formulation and an informed tabu search is adapted to determine the pattern sequence. Bolat (2003) propose a job selection problem that is solved prior to the sequencing problem. They employ a due date-oriented cost function as an objective, and the work overload is restricted as a hard constraint. They develop a branch-and-bound (B&B) algorithm that is improved with some dominance criteria, a procedure to compare sequences based on the quality, which can select 50 jobs out of 100 in seconds. Kim and Jeong (2007) present a B&B algorithm to solve the MMS problem with sequence-dependent setup times. They calculate a lower bound on the work overload of the current sequence and the minimum possible work overload of the unconsidered configurations. The model can solve instances with up to five stations and 30 configurations. Boysen et al. (2011) integrate a skip policy for utility work into the MMS and formulate the new problem as an mixed-integer linear program (MILP). They propose a B&B algorithm with improved lower bound calculations and dominance rules.

There are several heuristic and meta-heuristic approaches related to the MMS. The most popular algorithm is the genetic algorithm (GA) which is adapted in several ways to solve MMS problems (Hyun et al., 1998; Kim et al., 1996; Ponnambalam et al., 2003; Leu et al., 1996; Celano et al., 1999; Akgündüz and Tunalı, 2010; Zhang et al., 2020). Akgündüz and Tunalı (2011) review GA-based MMS solution approaches. Other popular evolutionary algorithms that are used to solve MMS include ant colony optimization (Akpınar et al., 2013; Zhu and Zhang, 2011; Kucukkoc and Zhang, 2014), particle swarm (Rahimi-Vahed et al., 2007a; Wei-qi et al., 2011), scatter search algorithm (Rahimi-Vahed et al., 2007b; Cano-Belmán et al., 2010; Liu et al., 2014), and imperialist competitive algorithm (Zandieh and Moradi, 2019).

While the majority of the MMS literature focuses on models with deterministic parameters, there are a few studies that consider stochastic parameters on either processing times or demand. The seminal study with stochastic processing times is proposed by Zhao et al. (2007). They provide a Markov chain based approach that minimizes the expected work overload duration. This approximation is done by generating sub-intervals of possible positions of workers within the stations. The expected work overload is calculated based on the lower and upper bounds of the intervals. Mosadegh et al. (2017) propose a heuristic approach, inspired by Dijkstra’s algorithm, to tackle a single-station MMS with stochastic processing times. They formulate the problem as a shortest path problem. Mosadegh et al. (2020) formulate a multiple-station MMS with stochastic processing times as an MILP. They provide a Q-learning-based simulated annealing (SA) heuristic to solve industry-sized problems and show that the expected work overload is decreased compared to the deterministic problem. Brammer et al. (2022) propose a reinforcement learning approach to solve MMS by negatively rewarding work overload occurrences. They show that the proposed approach provides at least 7% better solutions than SA and GA. Moreover, stochastic parameters are considered in integrated mixed-model balancing and sequencing problems as well: in processing times (Agrawal and Tiwari, 2008; Özcan et al., 2011; Dong et al., 2014) and in demand (Sikora, 2021).

Although numerous studies have been conducted on the sequencing problems, only Hottenrott et al. (2021) consider the product failures in sequence planning, yet in the car sequencing structure. To the best of our knowledge, there is no research available that establishes robust sequences in the MMS structure that can withstand work overloads caused by product failures.

3 Problem Statement and Mathematical Formulation

In Section 3.1, we define the MMS with stochastic failures and illustrate the problem with an example. Then, in Section 3.2, we provide a two-stage stochastic program for our problem.

3.1 Problem Statement

In an MMAL, a set of workstations are connected by a conveyor belt. Products, launched with a fixed rate, move along the belt at a constant speed of one time unit (TU). The duration between two consecutive launches is called the cycle time cc, and we define the station length lkcl_{k}\geq c in TU as the total TU that the workpiece requires to cross the station kKk\in K. Operators work on the assigned tasks and must finish their job within the station length, otherwise, the line is stopped or a so-called utility worker takes over the remaining job. The excess work is called work overload. The sequence of products therefore has a great impact on the efficiency of the assembly line. MMS determines the sequence of a given set of products VV by assigning each product vVv\in V to one of the positions tTt\in T.

Formulating the MMS problem based on the vehicle configurations instead of vehicles is usual (Bolat and Yano, 1992; Bard et al., 1992; Scholl et al., 1998), however, automobile manufacturers offer billions of combinations of options (Pil and Holweg, 2004). When this high level of customization is combined with a short lead time promised to the customers, each vehicle produced in a planning horizon becomes unique. In this study, the vehicles are sequenced instead of configurations since the case study focuses on an automobile manufacturing facility with a high level of customization. In order to do so, we define a binary decision variable xvtx_{vt}, which takes value of 1 if vehicle vVv\in V is assigned to position tTt\in T. The processing time of vehicle vVv\in V at station kKk\in K is notated by pkvp_{kv}. The starting position and work overload of the vehicle at position tTt\in T for station kKk\in K are represented by zktz_{kt} and wktw_{kt}, respectively. Table 1 lists all the parameters and decision variables used in the proposed model. While second-stage decision variables are scenario-dependent, we drop such dependency for notation simplicity throughout the paper unless it is needed explicitly for clarity.

Sets and Index
V,vV,v Vehicles
K,kK,k Stations
T,tT,t Positions
Ω,ω\Omega,\omega Scenarios
Parameters
pkvp_{kv} The processing time of vehicle vVv\in V at station kKk\in K
lkl_{k} The length of station kKk\in K
cc The cycle time
fvf_{v} The failure probability of vehicle vVv\in V
evωe_{v\omega} 1 if vehicle vVv\in V exists at scenario ωΩ\omega\in\Omega, 0 otherwise
First-Stage Decision variables
xvtx_{vt} 1 if vehicle vVv\in V is assigned to position tTt\in T, 0 otherwise
Second-Stage Decision variables
wktw_{kt} The work overload at station kKk\in K at position tTt\in T
zktz_{kt} Starting position of operator at station kKk\in K at the beginning of position tTt\in T
bktb_{kt} The processing time at station kKk\in K at position tTt\in T
Table 1: List of parameters and decision variables used in the model

In this paper, we adopt the side-by-side policy as a work overload handling procedure. A utility worker is assumed to work with the regular worker side-by-side, enabling work to be completed within the station borders. The objective of MMS with the side-by-side policy is to minimize the total duration of work overloads, i.e., the total duration of the remaining tasks that cannot be completed within the station borders. The regular operator stops working on the piece at the station border so they can start working on the next workpiece at position lkcl_{k}-c in the same station.

We illustrate an MMAL with the side-by-side policy in Figure 2 which represents a station that processes five vehicles. The left and right vertical bold lines represent the left and right borders of the station. Assume that the cycle time cc is 7 and the station length lkl_{k} is 10 TU, i.e., it takes 10 TU for the conveyor to flow through the station. This specific station processes two different configurations of vehicles: configurations A and B. While configurations A requires option 1, configurations B does not, so the processing times of configurations A and B are 9 and 5, respectively. Figure 2 illustrates the first five vehicles in the sequence which is [A, B, B, A, A]. The diagonal straight lines represent the position of the vehicle in the station. The worker always starts working on the first vehicle at position zero, left border of the station. The second vehicle is already at position 2=9c2=9-c when the worker completes working on the first vehicle. Note that the next vehicle enters the station borders a cycle time after the current vehicle enters the station borders. The tasks of the second vehicle are completed when the third vehicle has just entered the station. The worker has 2 TU of idle time when the tasks of the third vehicle are completed. The worker starts working on the fifth vehicle on position 2 and the processing time of the fifth vehicle is 9 TU which causes a work overload of 1 TU, 2+9lk=12+9-l_{k}=1. The job of processing these five vehicles could not be completed within the station borders but with the help of a utility worker, we assume that the job is completed at the station border at the cost of 1 TU work overload. The worker will continue working on the sixth vehicle at position 3=lkc3=l_{k}-c, and this process continues.

Refer to caption

Figure 2: Illustration of mixed-model assembly line with five vehicles, cycle time c=7c=7, and station length lk=10l_{k}=10. From bottom to top, the diagonal lines correspond to vehicle configurations A, B, B, A, A

We note that the vehicles usually go through the body shop and paint shop in the scheduled sequence before the assembly process. Hence, the failed vehicles must be pulled out of the sequence, and its position cannot be compensated, i.e., resequencing is not an option. It is assumed that each vehicle vVv\in V (with its unique mix of configurations) has a failure probability fvf_{v}, and failures are independent of each other. The failures are related to the specifications of the vehicle, e.g., the increased production rate of EVs may induce higher failure rates, or painting a vehicle to a specific color may be more problematic. In our numerical experiments in Section 5, we estimate the failure probabilities from the historical data by doing feature analysis and using logistic regression.

3.2 Mathematical Model Under Uncertainty

In this section, first, we provide a two-stage stochastic program for our problem. Next, we discuss improvements of the proposed formulation.

Motivated by the dynamics of an MMAL, we formulate our problem as a two-stage stochastic program. The sequence of vehicles is decided in the first stage (here-and-now), before the car failures are realized. Once the car failures are realized, the work overload is minimized by determining the second-stage decisions (wait-and-see) given the sequence. First-stage decisions are determined by assigning each vehicle to a position such that the expected work overload in the second stage is minimized.

To formulate the problem, suppose that various realizations of the car failures are represented by a collection of finite scenarios Ω\Omega. As each vehicle either exists or fails at a scenario, we have a total of 2|V|2^{|V|} scenarios. We let Ω={ω1,,ω2|V|}\Omega=\{\omega_{1},\ldots,\omega_{2^{|V|}}\}, with ω\omega indicating a generic scenario. To denote a scenario ω\omega, let evω=1e_{v\omega}=1 if vehicle vv exists and evω=0e_{v\omega}=0 if vehicle vv fails at scenario ωΩ\omega\in\Omega. We can then calculate the probability of scenario ω\omega as ρω=v=1|V|fv1evω(1fv)evω\rho_{\omega}=\prod_{v=1}^{|V|}f_{v}^{1-e_{v\omega}}(1-f_{v})^{e_{v\omega}} such that ωΩρω=1\sum_{\omega\in\Omega}\rho_{\omega}=1, where fvf_{v} denotes the failure probability of vehicle vVv\in V.

A two-stage stochastic program for the full-information problem, where all possible realizations are considered, is as follows:

minx\displaystyle\min_{x}\quad ωΩρωQ(x,ω)\displaystyle\sum_{\omega\in\Omega}\rho_{\omega}Q(x,\omega) (1a)
s.t. vVxvt=1,tT\displaystyle\sum_{v\in V}x_{vt}=1,\qquad t\in T (1b)
tTxvt=1,vV\displaystyle\sum_{t\in T}x_{vt}=1,\qquad v\in V (1c)
xvt{0,1},tT,vV\displaystyle x_{vt}\in\{0,1\},\qquad t\in T,\enspace v\in V (1d)

where

Q(x,ω)=minz,w,b\displaystyle Q(x,\omega)=\min_{z,w,b}\quad kKtTωwkt\displaystyle\sum_{k\in K}\sum_{t\in T_{\omega}}w_{kt} (2a)
s.t. bkt=vVpkvxvt,kK,tTω\displaystyle b_{kt}=\sum_{v\in V}p_{kv}x_{vt},\qquad k\in K,\enspace t\in T_{\omega} (2b)
zktzk(t+1)wktcbkt,kK,tTω,\displaystyle z_{kt}-z_{k(t+1)}-w_{kt}\leq c-b_{kt},\qquad k\in K,\enspace t\in T_{\omega},\enspace (2c)
zktwktlkbkt,kK,tTω,\displaystyle z_{kt}-w_{kt}\leq l_{k}-b_{kt},\qquad k\in K,\enspace t\in T_{\omega}, (2d)
zk0=0,kK,\displaystyle z_{k0}=0,\qquad k\in K,\enspace (2e)
zk(|Tω|+1)=0,kK,\displaystyle z_{k(|T_{\omega}|+1)}=0,\qquad k\in K,\enspace (2f)
zkt,wkt0,kK,tTω,\displaystyle z_{kt},w_{kt}\geq 0,\qquad k\in K,\enspace t\in T_{\omega},\enspace (2g)

In the first-stage problem (1), the objective function represents the expected work overload, i.e., the cost associated with the second-stage problem. Constraint sets (1b) and (1c) ensure that exactly one vehicle is assigned to each position and each position has exactly one vehicle. respectively. Constraint set (1d) presents the domain of the binary first-stage variable. The second-stage problem (2) minimizes the total work overload throughout the planning horizon, given the sequence and scenario ωΩ\omega\in\Omega. Note that TωT_{\omega} denotes the set of positions of non-failed vehicles at scenario ωΩ\omega\in\Omega, which is obtained by removing failed vehicles. Constraint set (2b) determines the processing time bktb_{kt} at station kk at position tt. The starting position and workload of the vehicles at each station are determined by constraint sets (2c) and (2d), respectively. The constraint set (2e) ensures that the first position starts at the left border of the station. Constraint set (2f) builds regenerative production planning, in other words, the first position of the next planning horizon can start at the left border of the station. The constraint set (2g) defines the second-stage variables as continuous and nonnegative.

Several remarks are in order regarding the proposed two-stage stochastic program (1) and (2). First, the number of decision variables and the set of constraints in (2) are scenario-dependent, as the valid positions TωT_{\omega} are obtained based on the failure scenario ωΩ\omega\in\Omega. Second, the proposed two-stage stochastic program (1) and (2) has a simple recourse. That is, once the sequence is determined and the failures are realized, the work overloads are calculated from the sequence of the existing vehicles, without resequencing.

In the remainder of this section, we first provide two modified models for the second-stage problem so that the number of decision variables and the set of constraints are no longer scenario-dependent. Then, we provide two monolithic MILP formulations for the deterministic equivalent formulation (DEF) of the two-stage stochastic program of MMS with stochastic failures.

For each scenario, we modify the second-stage problem by updating the processing times of failed vehicles instead of removing the failed vehicles. In Figure 3, we demonstrate how the original model (2) and modified models represent car failures. To this end, we consider the example given in Figure 2 and assume that the vehicle at the second position fails. In the original model, the failed vehicles are removed from the sequence and the succeeding vehicles moved forward (Figure 3(a)). The proposed modified models, referred to as standard model and improved model, are explained below. In Section 5.2, we discuss the impact of these modified models on the computational time and solution quality.

Refer to caption
(a) Original model
Refer to caption
(b) (Modified) standard model
Refer to caption
(c) (Modified) improved model
Figure 3: Assembly line illustration of proposed models

Standard Model: In a preprocessing step, the processing time of vehicle vv is set to zero for all stations if the vehicle fails in scenario ω\omega. Accordingly, since this modification is conducted by adding uncertainty to the processing times, the scenario index ω\omega is added to the processing time. That is, evω=0pkvω=0kKe_{v\omega}=0\Rightarrow p_{kv\omega}=0\enspace k\in K. Based on this modification, the second-stage problem for a given scenario ω\omega can be presented as

Q(x,ω)=minw,z,b\displaystyle Q(x,\omega)=\min_{w,z,b}\quad kKtTwkt\displaystyle\sum_{k\in K}\sum_{t\in T}w_{kt} (3a)
s.t. bkt=vVpkvxvt,kK,tT,\displaystyle b_{kt}=\sum_{v\in V}p_{kv}x_{vt},\qquad k\in K,\enspace t\in T,\enspace (3b)
zkt+bktwktzk(t+1)c,kK,tT,\displaystyle z_{kt}+b_{kt}-w_{kt}-z_{k(t+1)}\leq c,\qquad k\in K,\enspace t\in T,\enspace (3c)
zkt+bktwktlk,kK,tT,\displaystyle z_{kt}+b_{kt}-w_{kt}\leq l_{k},\qquad k\in K,\enspace t\in T,\enspace (3d)
zk0=0,kK,\displaystyle z_{k0}=0,\qquad k\in K,\enspace (3e)
zk(T+1)=0,kK,\displaystyle z_{k(T+1)}=0,\qquad k\in K,\enspace (3f)
zktβkbktzk(t+1)0,kK,t={1,,T1}\displaystyle z_{kt}-\beta_{k}b_{kt}-z_{k(t+1)}\leq 0,\qquad k\in K,\enspace t=\{1,\ldots,T-1\} (3g)
zkTwkTβkbkT0,kK,\displaystyle z_{kT}-w_{kT}-\beta_{k}b_{kT}\leq 0,\qquad k\in K,\enspace (3h)
zkt,wkt,bkt0,kK,tT,\displaystyle z_{kt},w_{kt},b_{kt}\geq 0,\qquad k\in K,\enspace t\in T,\enspace (3i)

The objective function (3a) and the constraints (3b)–(3f) are the same as in formulation (2), except the set of positions and the length of the sequence are not scenario-dependent anymore. Constraint set (3g) guarantees that the starting position at station kk at position t+1t+1 equals the starting position of position tt when the vehicle assigned to position tt fails. Constraint set (3h) assures that the regenerative production planning is kept in case the vehicle at the end of the sequence fails. The parameter β\beta is calculated in a way that βkbktn>zktn\beta_{k}b_{ktn}>z_{ktn} which makes constraint sets (3g) and (3h) non-effective for the positions that have existing vehicles. Hence, βk\beta_{k} equals the maximum possible starting position divided by the minimum processing time for station kk, βk=lkcminvV{pkv}>0\beta_{k}=\frac{l_{k}-c}{\min_{v\in V}\{p_{kv}\}}>0. Note that the processing times in this calculation are the actual processing times before the preprocessing step. Also, βk\beta_{k} is well-defined as the the minimum processing time is strictly greater than zero. Figure 3(b) demonstrates that in the standard model, the processing time of the second vehicle is set to zero, so the operator starts working on the third vehicle at position two where the operator was going to start working on the second vehicle if it had not failed.

Improved Model: In order to reduce the size of the standard model, we modify this model as follows. During the preprocessing step, the processing time of vehicle vv is set to the cycle time for all stations if a vehicle fails at scenario ω\omega. Let us refer to the vehicles with processing time equal to cycle time for all stations as “neutral” because these vehicles do not have any impact on the schedule in terms of work overload (see Proposition 1 and its proof). In other words, we transform failed vehicles into neutral vehicles, i.e., evω=0pkvω=ckKe_{v\omega}=0\Rightarrow p_{kv\omega}=c\enspace k\in K.

Proposition 1.

A neutral vehicle has the same starting position as its succeeding vehicle at all stations. That is, bkt=czk(t+1)=zktb_{kt}=c\enspace\Rightarrow\enspace z_{k(t+1)}=z_{kt}.

Proof.

The operator’s starting position of the vehicle at t+1t+1 is zk(t+1)=zkt+bktcwktz_{k(t+1)}=z_{kt}+b_{kt}-c-w_{kt}. Assume that the vehicle at position tt is a neutral vehicle. We have zk(t+1)=zktwktz_{k(t+1)}=z_{kt}-w_{kt}. Hence, showing that the neutral vehicles never cause a work overload, wkt=0,w_{kt}=0, completes the proof. We know that the maximum starting position at a station is maxtT{zkt}=lkc\max_{t\in T}\{z_{kt}\}=l_{k}-c, which is a result of two extreme cases: an operator finishes working on a workpiece at the right border of a station or the operator cannot finish the work so we have a work overload. The starting position is less than lkcl_{k}-c for other cases. Therefore, a vehicle with a processing time less than or equal to cc at a station cannot cause any work overload. This completes the proof. ∎

As a result of Proposition 1, constraints (3g) and (3h) can be removed from the standard model. Hence, the problem size is reduced. Figure 3(c) contains an illustration for Proposition 1. The second vehicle becomes neutral when its processing time is set to cycle time so that the third vehicle starts at the same position as the second vehicle.

Using the standard or improved model, the DEF for MMS with stochastic failures can be obtained by adding the first-stage constraints (1b)–(1d) to the corresponding second-stage formulation, and by adding copies of all second-stage variables and constraints. We skip the details for brevity.

4 Solution Approaches

In Sections 4.1 and 4.2, we propose an L-shaped decomposition-based algorithm and a tabu search algorithm in addition to a greedy heuristic, respectively, to solve the models presented in Section 3.2. Then, in Section 4.3, the SAA approach is motivated and a solution quality assessment scheme is presented.

4.1 Exact Solution Approach

For the ease of exposition, we consider an abstract formulation of the two-stage stochastic program presented in Section 3.2 as follows:

z=minxX𝔼[Q(x,ξω)],\displaystyle z^{*}=\min_{x\in X}\mathbb{E}[Q(x,\xi_{\omega})], (4)

where xx denotes the first-stage decisions variables and X:={x{0,1}|V|×|T|:Ax=b}X:=\{x\in\{0,1\}^{|V|\times|T|}\colon Ax=b\} is the feasible region of decision variables xx, i.e., the set of points satisfying constraints (1b) - (1d). Moreover, we represent the second-stage problem for the standard or improved model, presented in Section 3.2, as

Q(x,ξω)=miny{qy|DyhωTωx,y0},Q(x,\xi_{\omega})=\min_{y}\{q^{\top}y|Dy\geq h_{\omega}-T_{\omega}x,\;y\geq 0\}, (5)

where yy represents the second-stage decision variables and ξω=(hω,Tω)\xi_{\omega}=(h_{\omega},T_{\omega}). The expectation of the recourse problem becomes 𝔼[Q(x,ξω)]=ωΩρωQ(x,ξω)\mathbb{E}[Q(x,\xi_{\omega})]=\sum_{\omega\in\Omega}\rho_{\omega}Q(x,\xi_{\omega}).

The L-shaped method is a procedure that has been successfully used to solve large-scale two-stage stochastic programs. Note that for any ωΩ\omega\in\Omega, function Q(x,ξω)Q(x,\xi_{\omega}), defined in (5), is convex in xx because xx appears on the right-hand side of constraints. Hence, we propose to iteratively construct its underestimator. To this end, for each ωΩ\omega\in\Omega and a given first-stage decision xXx\in X, we consider a subproblem that takes the form of (5). Moreover, we create a relaxed master problem, which contains a partial, but increasingly improving, representation of Q(x,ξω)Q(x,\xi_{\omega}), for each ωΩ\omega\in\Omega, through the so-called Benders’ cuts. Recall that our proposed two-stage stochastic programs have a relative complete recourse, that is, for any first-stage decision xx, there is a feasible second-stage variable yy. Thus, an underestimator of Q(x,ξω)Q(x,\xi_{\omega}) can be constructed by only the so-called Benders’ optimality cuts.

We now describe more details on our proposed L-shaped algorithm. We form the relaxed master problem for formulation (4) and (5) as follows:

minx,θ\displaystyle\min_{x,\theta}\quad ωΩρωθω\displaystyle\sum_{\omega\in\Omega}\rho_{\omega}\theta_{\omega} (6a)
s.t. xX\displaystyle x\in X (6b)
θωGωιx+gωι,ι{1,,l},ωΩ,\displaystyle\theta_{\omega}\geq G^{\iota}_{\omega}x+g^{\iota}_{\omega},\quad\iota\in\{1,\ldots,l\},\enspace\omega\in\Omega, (6c)

where the auxiliary variable θω\theta_{\omega} approximates the optimal value of the second-stage problem under scenario ωΩ\omega\in\Omega, i.e., Q(x,ξω)Q(x,\xi_{\omega}), through cuts θωGωιx+gωι\theta_{\omega}\geq G^{\iota}_{\omega}x+g^{\iota}_{\omega} formed up to iteration ll.

Let (x^ι,θ^ι)(\hat{x}^{\iota},\hat{\theta}^{\iota}) be an optimal solution to the relaxed master problem (6). For each scenario ωΩ\omega\in\Omega, we form a subproblem (5) at x^ι\hat{x}^{\iota}. Suppose that given x^ι\hat{x}^{\iota}, π^ωι\hat{\pi}_{\omega}^{\iota} denotes an optimal dual vector associated with the constraints in (5). That is, π^ωι\hat{\pi}_{\omega}^{\iota} is an optimal extreme point of the dual subproblem (DSP)

maxπ{πω(hωTωx^ι)|πωDq,πω0},\max_{\pi}\{\pi_{\omega}^{\top}(h_{\omega}-T_{\omega}\hat{x}^{\iota})|\pi_{\omega}^{\top}D\leq q^{\top},\;\pi_{\omega}\geq 0\}, (7)

where πω\pi_{\omega} is the associated dual vector. Then, using linear programming duality, we generate an optimality cut as

θωGωιx+gωι,\theta_{\omega}\geq G^{\iota}_{\omega}x+g^{\iota}_{\omega}, (8)

where Gωι=(π^ωι)TωG^{\iota}_{\omega}=-(\hat{\pi}_{\omega}^{\iota})^{\top}T_{\omega} and gωι=(π^ωι)hωg^{\iota}_{\omega}=(\hat{\pi}_{\omega}^{\iota})^{\top}h_{\omega}.

Our proposed L-shaped algorithm iterates between solving the relaxed master problem (6) and subproblems (5) (one for each ωΩ\omega\in\Omega) until a convergence criterion on the upper and lower bounds is satisfied. This algorithm results in an L-shaped method with multiple cuts.

In order to exploit the specific structure of the MMS problem and to provide improvements on the dual problem, let us define variables πsp\pi^{sp}, πwo\pi^{wo}, πfs\pi^{fs}, πch\pi^{ch}, πsf\pi^{sf}, and πcf\pi^{cf} corresponding to starting position constraints (3c), work overload constraints (3d), first station starting position constraints (3e), regenerative production planning constraints (3f), starting position of the vehicles following a failed vehicle (3f), and regenerative production planning with failed vehicles (3g), respectively. The DSP for scenario ωΩ\omega\in\Omega at a candidate solution x^ι\hat{x}^{\iota}, obtained by solving a relaxed master problem, can be formulated as follows:

maxπ\displaystyle\max_{\pi}\quad kKtTπktsp(vVpkvx^vtc)+πktwo(vVpkvx^vtlk)\displaystyle\sum_{k\in K}\sum_{t\in T}\pi^{sp}_{kt}(\sum_{v\in V}p_{kv}\hat{x}_{vt}-c)+\pi^{wo}_{kt}(\sum_{v\in V}p_{kv}\hat{x}_{vt}-l_{k}) (9a)
s.t. πk0sp+πk0wo+πkfs+πk0sf0,kK\displaystyle\pi^{sp}_{k0}+\pi^{wo}_{k0}+\pi^{fs}_{k}+\pi^{sf}_{k0}\leq 0,\qquad k\in K (9b)
πktspπk(t+1)spπk(t+1)wo+πktsfπk(t+1)sf0,kK,t{1,..,T1}\displaystyle\pi^{sp}_{kt}-\pi^{sp}_{k(t+1)}-\pi^{wo}_{k(t+1)}+\pi^{sf}_{kt}-\pi^{sf}_{k(t+1)}\leq 0,\qquad k\in K,\enspace t\in\{1,..,T-1\} (9c)
πkTspπkch0,kK\displaystyle\pi^{sp}_{kT}-\pi^{ch}_{k}\leq 0,\qquad k\in K (9d)
πktsp+πktwo1,kK,t{1,..,T1}\displaystyle\pi^{sp}_{kt}+\pi^{wo}_{kt}\leq 1,\qquad k\in K,\enspace t\in\{1,..,T-1\} (9e)
πkTsp+πkTwo+πkcf1,kK\displaystyle\pi^{sp}_{kT}+\pi^{wo}_{kT}+\pi^{cf}_{k}\leq 1,\qquad k\in K (9f)
πktsp,πktwo,πktsf,πkcf0,kK,tT\displaystyle\pi^{sp}_{kt},\pi^{wo}_{kt},\pi^{sf}_{kt},\pi^{cf}_{k}\geq 0,\qquad k\in K,\enspace t\in T (9g)
πkfs,πkchunrestricted,kK\displaystyle\pi^{fs}_{k},\pi^{ch}_{k}\text{unrestricted},\qquad k\in K (9h)

We provide improvements to the dual problem in several ways. The dual variables πfs\pi_{fs} and πcf\pi_{cf} are removed since the corresponding subproblem constraints (3f) and (3g) are eliminated in the improved model. The dual variables πfs\pi^{fs} and πch\pi^{ch} are not in the objective function and are unrestricted, which means that we can remove these variables and the constraints with those variables from the formulation without altering the optimal value of the problem. In our preliminary computational studies, we improved the dual subproblem by removing these variables. However, we observed that most of the DSPs have multiple optimal solutions, and as the number of vehicles and stations increase, it is more likely to have multiple optimal solutions. This naturally raises the question of what optimal dual vector provides the strongest cut, if we add only one cut per iteration per scenario. One can potentially add the cuts corresponding to all optimal dual extreme points, however, this results in an explosion in the size of the relaxed master problem after just a couple of iterations. While there is no reliable way to identify the weak cuts (Rahmaniani et al., 2017), we executed experiments in order to find a pattern for strong cuts. Our findings showed that adding the cut corresponding to the optimal dual extreme point with the most non-zero variables results in the fastest convergence. Thus, we added an 1\ell_{1} regularization term to the objective function of the DSP, hence, the new objective is encouraged to choose an optimal solution with the most non-zero variables. Accordingly, we propose an improved DSP formulation as follows:

maxπ\displaystyle\max_{\pi}\quad kKtTπktsp(vVpkvx^vtc+ϵ)+πktwo(vVpkvx^vtlk+ϵ)\displaystyle\sum_{k\in K}\sum_{t\in T}\pi^{sp}_{kt}(\sum_{v\in V}p_{kv}\hat{x}_{vt}-c+\epsilon)+\pi^{wo}_{kt}(\sum_{v\in V}p_{kv}\hat{x}_{vt}-l_{k}+\epsilon) (10a)
s.t. πktspπk(t+1)spπk(t+1)wo0,kK,t{1,..,T1}\displaystyle\pi^{sp}_{kt}-\pi^{sp}_{k(t+1)}-\pi^{wo}_{k(t+1)}\leq 0,\qquad k\in K,\enspace t\in\{1,..,T-1\} (10b)
πktsp+πktwo1,kK,tT\displaystyle\pi^{sp}_{kt}+\pi^{wo}_{kt}\leq 1,\qquad k\in K,\enspace t\in T (10c)
πktsp,πktwo0,kK,tT\displaystyle\pi^{sp}_{kt},\pi^{wo}_{kt}\geq 0,\qquad k\in K,\enspace t\in T (10d)

4.2 Heuristic Solution Approach

MMS is an NP-hard problem, and stochastic failures of products (cars) increases the computational burden of solving the problem drastically. Hence, it is essential to create efficient heuristic procedures in order to solve industry-sized problems. In this section, we provide a fast and easy-to-implement greedy heuristic to find a good initial feasible first-stage decision (i.e., a sequence of vehicles) and an efficient tabu search (TS) algorithm to improve the solution quality. Although all |V|!|V|! vehicle permutations are feasible, the proposed greedy heuristic aims to find a good initial feasible solution. To achieve this, a solution is generated for the deterministic counterpart of the proposed MMS problem, which excludes vehicle failures. We refer to this problem as the one-scenario problem, since the corresponding problem has a single scenario with no failed vehicles. Assuming that the failure probability of each vehicle is less than or equal to 0.5, the scenario with no failed vehicles has the highest probability. Once such a feasible sequence of vehicles is generated, the TS algorithm improves this solution in two parts: first, over the one-scenario problem, and then, over the full-information problem.

4.2.1 Greedy Heuristic

It is important for a local search heuristic algorithm to start with a good quality solution. A naive approach to generate an initial solution (sequence) is to always select the vehicle that causes the minimum new work overload for the next position. However, this approach is myopic since it only considers the current position. We remediate this issue by decreasing future work overloads which includes considering idle times and dynamic utilization rates. Accordingly, in order to generate a good initial solution, we propose utilizing an iterative greedy heuristic that follows a priority rule based on the work overload, idle time, and weighted sum of processing time, to be defined shortly.

Before explaining our proposed greedy heuristic, let us define some technical terms. The idle time refers to the duration that an operator waits for the next vehicle to enter the station borders. The weight of processing times is determined by using station utilization rates, which is inspired by car sequencing problem utilization rates (Solnon, 2000; Gottlieb et al., 2003). We describe the utilization rate of a station as the ratio between the average processing time on a station and the cycle time, so the utilization rate of station kk is vVpkv/(|V|c)\sum_{v\in V}p_{kv}/(|V|*c). At each iteration, after a new assignment of a vehicle, the dynamic utilization rates are calculated by considering only the unassigned vehicles. Accordingly, the weighted sum of the processing time of a vehicle vv is calculated using (11):

kKpkviV^pki|K||V^|c,\frac{\sum_{k\in K}p_{kv}\sum_{i\in\hat{V}}p_{ki}}{|K|*|\hat{V}|*c}, (11)

where V^\hat{V} denotes the set of unassigned vehicles. If the utilization rate of a station is greater than 1, then the average processing time is more than the cycle time, which induces an unavoidable work overload. On the other hand, a utilization rate close to 0 indicates that the average processing time is minimal compared to the station’s allocated time.

Our proposed greedy heuristic builds a sequence iteratively, one position at a time, starting from the first position and iterating over positions. We use tt to denote an iteration. At each iteration, t=1,,Tt={1,\ldots,T}, the unassigned vehicles that cause the minimum new work overload are determined, denoted by Vt,woV_{t,wo}. Ties are broken by selecting the vehicles from the set Vt,woV_{t,wo}, which causes the minimum new idle time, the new set of vehicles is denoted by Vt,idleV_{t,idle}. In the case of ties, the vehicle with the highest weighted sum of processing time from the set Vt,idleV_{t,idle} is assigned to the position tt of the sequence. Note that the first vehicle of the sequence is the vehicle with the highest weighted sum of processing time among the set V0,idleV_{0,idle} since there is no work overload initially.

Finally, we enhance the proposed greedy heuristic by considering the category of the vehicles. Motivated by our case study, we categorize the vehicles based on the engine type, electric or non-electric, because the engine type is the most restrictive feature due to the high EV ratio (number of EVs divided by the number of all vehicles). Moreover, the engine type leads to different processing times on a specific station. Hence, we modify our greedy heuristic to first decide whether an EV or a non-EV should be assigned to the next position at each iteration. Accordingly, first, the EV ratio is calculated, and an EV is assigned to the first position. The procedure always follows the EV ratio. For example, if the EV ratio is 1/31/3, an EV will be assigned to the positions 1+3t1+3t where t={1,,|T|31}t=\{1,\ldots,\frac{|T|}{3}-1\}. In case of a non-integer EV ratio, the position difference between any two consecutive EVs is the integer part of the ratio plus zero or one; randomly decided based on the decimal part of the ratio. Once the vehicle category is decided throughout the entire sequence, the specific vehicle to be assigned is selected based on the above-described procedure. We note that this enhancement in the greedy heuristic may be applied for any restrictive feature that causes large variations in processing times.

Vehicle Engine p1p_{1} p2p_{2}
A Electric 15 4
B Electric 16 3
C Gasoline 2 10
D Gasoline 3 8
E Gasoline 2 9
F Gasoline 4 7
Table 2: Illustration of greedy heuristic

To describe the greedy heuristic, consider an example with six vehicles and two stations. The processing times and engine types of vehicles are given in Table 2. The cycle time is 7 TU, and the length of the stations are 20 TU and 10 TU, respectively. The EV ratio is 1/3. We consider only EVs for the first position, vehicle A is designated to the first position since it causes less idle time than vehicle B. Next, none of the non-EVs causes work overload or idle time, so we assign the vehicle with the highest weighted sum of processing times to the second position, vehicle C. The procedure continues with another non-EV, and vehicle F is assigned to the third position because it is the only vehicle that does not cause any work overload. Consistent with the 1/31/3 EV ratio, an EV must be assigned to the fourth position, and vehicle B is assigned to this position as it is the only EV left. Vehicle E is assigned to the fifth position due to its higher weighted sum of processing times. Finally, vehicle D is assigned to the last position. The resulting sequence is A-C-F-B-E-D with a work overload of 3 TU, only at position 6 at station 2.

4.2.2 Tabu Search Algorithm

This section proposes a simulation-based local search algorithm on a very large neighborhood with tabu rules. The TS algorithm starts at the initial feasible solution (sequence) generated by the iterative greedy heuristic, and improves the initial solution via iterative improvements within the designed neighborhood. At each iteration of the TS, a transformation operator is randomly selected based on operator weights and applied to the incumbent solution to visit a random neighbor, respecting the tabu rules. The candidate solution is accepted if the objective function value is non-deteriorated, i.e., the candidate solution is rejected only if it has more total work overload. Then, another random operator is applied to the incumbent solution. This process repeats until the stopping criterion is met.

As aforementioned, the TS has two parts. The first part acts as the second step of the initial solution generation procedure since it improves the solution provided by the greedy heuristic for the one-scenario problem. In our preliminary numerical experiments we observed that this step can drastically improve the initial solution quality. Hence, we conduct this step for a duration τone\tau_{one}. Next, the algorithm makes a transition to the full-information problem and reevaluates the objective function value of the incumbent solution—the sequence generated by the first part of TS. In the second part of the TS algorithm, the objective function value corresponding to the sequence is evaluated for the full-information problem. To do this, we calculate the total work overload for all realizations ωΩ\omega\in\Omega, given the first-stage decision (sequence). That is, we calculate the objective function of (3) for each realization ωΩ\omega\in\Omega and take the weighted sum, each multiplied by the probability of the scenario. Observe from (3) that once the first-stage decision is fixed, the problem decomposes in scenarios and stations. Accordingly, the solution evaluation process is parallelized over scenarios and stations.

The TS algorithm continues evaluating the solution for the full-information problem for a duration τfull\tau_{full}. The allocated time for the second part, τfull\tau_{full}, is much larger than that of the first part, τone\tau_{one}, since iterating over one-scenario problem is much faster than that over a set of realizations. In the reminder of this section, we explain various components of the TS algorithm.

Objective Evaluation

The objective function of the problem for a given scenario is the same as the objective given in (3a), total work overload over all stations and positions. Evaluation of the objective, after each movement, is the bottleneck of our algorithm since the new total work overload needs to be determined. Note that the objective evaluation starts at the first position and is executed iteratively since there is a sequence dependency. Accordingly, we propose to reduce the computational burden in two ways.

First, reevaluating the whole sequence is unnecessary since transformation operators make local changes in the sequence, i.e., some parts of the sequence remain unaltered and do not require reevaluation. Hence, we apply partial reevaluation after each movement. To explain partial reevaluation, assume that the vehicles at positions t1t_{1} and t2t_{2} are swapped. We certainly know that the subsequence corresponding to positions [1,t11][1,t_{1}-1] is not impacted by the swap operation; hence, we do not reevaluate these positions. Additionally, we may not have to reevaluate all the positions in [t1,t21][t_{1},t_{2}-1] and the positions in [t2,|T|][t_{2},|T|]. In each of these subsequences, there could be a reset position which ensures that there is no change in the objective from that position until the end of the subsequence. Since the rest of the subsequence after the reset position is not changed, we can jump to the end of the subsequence. To highlight how a partial reevaluation may improve the objective reevaluation process, suppose that the vehicles at positions 350 and 380 are swapped. We certainly know the subsequence corresponding to positions [1, 349] is not impacted by the swap. Additionally, in the case that there is a reset point before position 380 (and |T||T|), we do not have to reevaluate all the positions between 350 and 380, and the positions between 380 and |T||T|.

Second, we calculate the objective function in an accelerated way. Traditionally work overload and starting position for position tt at station kk, respectively wktw_{kt} and zk(t+1)z_{k(t+1)}, are calculated as: wkt=zkt+bktlkw_{kt}=z_{kt}+b_{kt}-l_{k} and zk(t+1)=zkt+bktwktcz_{k(t+1)}=z_{kt}+b_{kt}-w_{kt}-c, where wkt,zkt0w_{kt},z_{kt}\geq 0. Instead of calculating work overload and starting position vectors separately, we propose using a single vector to extract these information, which in fact is a different representation of the starting position vector zz. If there is a work overload at position tt, then zk(t+1)=lkcz_{k(t+1)}=l_{k}-c. Otherwise, if there is not any work overload at position tt, then zk(t+1)=zkt+bktcz_{k(t+1)}=z_{kt}+b_{kt}-c, or we can equivalently write zk(t+1)=zk(t1)+bk(t1)c+bktcwk(t1)z_{k(t+1)}=z_{k(t-1)}+b_{k(t-1)}-c+b_{kt}-c-w_{k(t-1)}. Again, if there is a work overload at position t1t-1, then zk(t+1)=(lkc)+bktcz_{k(t+1)}=(l_{k}-c)+b_{kt}-c, otherwise, if there is not any work overload at t1t-1, then zk(t+1)=zk(t1)+(bk(t1)c)+(bktc)z_{k(t+1)}=z_{k(t-1)}+(b_{k(t-1)}-c)+(b_{kt}-c). Since we know that zk0=0z_{k0}=0, we can generalize it as zk(t+1)=h=1t(bkhc)z_{k(t+1)}=\sum_{h=1}^{t}(b_{kh}-c), which is the cumulative sum of vector ηk=bkc\eta_{k}=b_{k}-c up to and including position tt. However, this generalization has the assumption that there is not any work overload or idle time up to position tt. We note that there is an idle time at position t+1t+1 when zkt+bktc<0z_{kt}+b_{kt}-c<0. Accordingly, we can write a general formula as zk(t+1)=max(0,min(lkc,zkt+ηk(t+1)))z_{k(t+1)}=\max(0,\min(l_{k}-c,z_{kt}+\eta_{k(t+1)})), which is referred to as the conditional cumulative sum of ηk\eta_{k} up to position tt. Intuitively, the conditional cumulative sum is defined as follows: starting from position 0, the cumulative sum is calculated iteratively within the closed range [0,lkc][0,l_{k}-c]. Whenever the cumulative sum exceeds the lower bound zero or the upper bound lkcl_{k}-c, we set the cumulative sum to the corresponding bound’s value. If the cumulative sum is below the lower bound, the excess value is equal to the idle time. Otherwise, if the cumulative sum is above the upper bound, the excess value is equal to the work overload. For example, if the cumulative sum is -2 at a position, the cumulative sum is set to zero and there is a 2 TU of idle time at that position.

In light of the proposed improvements, the partial reevaluation process is executed in two subsequences, [t1,t2)[t_{1},t_{2}) and [t2,|T|][t_{2},|T|], assuming that t1t_{1} and t2t_{2} are the two selected positions by any transformation operator and t1<t2t_{1}<t_{2}. The process starts at the first position, called position one, of the corresponding subsequence. We set zk0=ηk1z_{k0}=\eta_{k1} and we calculate the starting position, work overload, and idle time for the positions in the subsequence iteratively as mentioned above. The reevaluation of the subsequence is completed when either a reset position is found or the whole subsequence is iterated. A reset position occurs at position tt differently in two cases as follows: 1) if zk,t+1=0z_{k,t+1}=0 when the processing time on the starting position t1t_{1} (or t2t_{2}) is decreased, 2) if the sum of idle time and work overload up to position tt in the current subsequence exceeds the total increased processing time at the corresponding starting position t1t_{1} (or t2t_{2}) when the processing time on the starting position t1t_{1} (or t2t_{2}) is increased.

Transformation Operators

In this section, we explain the details of the transformation operators. We employ swap, forward and backward insertions, and inversion operators. The swap operator interchanges the positions of two randomly selected cars. Insertion removes a car from position ii and inserts it at position jj. Insertion is applied in two different directions, backward and forward. When i>ji>j, the insertion is called a backward insertion, and all the vehicles between the positions jj and ii move one position to the right , i.e., scheduled later. On the contrary, forward insertion occurs when i<ji<j and all the vehicles between the positions ii and jj move one position to the left, i.e., scheduled earlier. Inversion takes two randomly selected positions in the sequence, and the subsequence between the selected positions is reversed. A repetitive application of these operators creates a very large neighborhood which helps the improvement procedure to escape local optima, especially when it is combined with a non-deteriorated solution acceptance procedure. The latter enables the algorithm to move on the plateaus that consist of the solutions with the same objective function value (see Section 5.3.2 for numerical experiments).

Tabu List

We design the tabu list in a non-traditional manner. This list includes the movements that induce undesired subsequences. Based on our observations, we define an undesired subsequence as back-to-back EVs because consecutive EVs cause a tremendous amount of work overload at the battery loading station. Accordingly, any movement that results in back-to-back EVs is a tabu. For the sake of clarity, we describe tabu movements in detail for each operator separately in A.

4.3 SAA Approach and Solution Quality Assessment

In (4), it is assumed that the probability of each scenario is known a priori, which may not hold in practice. In addition, the exponential growth of the number of scenarios causes an explosion in the size of the stochastic program. Hence, we utilize the SAA approach to tackle these issues. Consider the abstract formulation (4). The SAA method approximates the expected value function with an identically and independently distributed (i.i.d) random sample of NN realizations of the random vector ΩN:={ω1,,ωN}Ω\Omega_{N}:=\{\omega_{1},\ldots,\omega_{N}\}\subset\Omega as follows:

zN=minxX1NωΩNQ(x,ξω).\displaystyle z_{N}=\min_{x\in X}\frac{1}{N}\sum_{\omega\in\Omega_{N}}Q(x,\xi_{\omega}). (12)

The optimal value of (12), zNz_{N}, provides an estimate of the true optimal value (Kleywegt et al., 2002). Let x^N\hat{x}_{N} and xx^{*} denote an optimal solution to the SAA problem (12) and the true stochastic program (4), respectively. Note that 𝔼[Q(x^N,ξω)]𝔼[Q(x,ξω)]\mathbb{E}[Q(\hat{x}_{N},\xi_{\omega})]-\mathbb{E}[Q(x^{*},\xi_{\omega})] is the optimality gap of solution x^N\hat{x}_{N}, where 𝔼[Q(x^N,ξω)]\mathbb{E}[Q(\hat{x}_{N},\xi_{\omega})] is the (true) expected cost of solution x^N\hat{x}_{N} and 𝔼[Q(x,ξω)]\mathbb{E}[Q(x^{*},\xi_{\omega})] is the optimal value of the true problem (4). A high quality solution is implied by a small optimality gap. However, as xx^{*} (and hence, the optimal value of the true problem) may not be known, one may obtain a statistical estimate of the optimality gap to assess the quality of the candidate solution x^N\hat{x}_{N} (Homem-de Mello and Bayraksan, 2014). That is, given that 𝔼[zN]𝔼[Q(x,ξω)]\mathbb{E}[z_{N}]\leq\mathbb{E}[Q(x^{*},\xi_{\omega})], we can obtain an upper bound on the optimality gap as 𝔼[Q(x^N,ξω)]𝔼[zN]\mathbb{E}[Q(\hat{x}_{N},\xi_{\omega})]-\mathbb{E}[z_{N}]. We employ the multiple replication procedure of Mak et al. (1999) in order to assess the quality of a candidate solution by estimating an upper bound on its optimality gap. A pseudo-code for this procedure is given in Algorithm 1. We utilize MRP, in Section 5, to assess the quality of solutions generated by different approaches.

Algorithm 1 Multiple Replication Procedure MRPα(x^)\textrm{MRP}_{\alpha}(\hat{x})
Input: Candidate solution x^\hat{x}, replication size MM, and α(0,1)\alpha\in(0,1).
Output: A normalized %100(1α)\%100(1-\alpha) upper bound on the optimality gap of x^\hat{x}.
for m=1,2,,Mm=1,2,\ldots,Mdo
    Draw i.i.d. sample ΩNm\Omega^{m}_{N} of realizations ξωm\xi^{m}_{\omega}, ωΩNm\omega\in\Omega_{N}^{m}.
    Obtain zNm:=minxX1NωΩNmQ(x,ξωm)z^{m}_{N}:=\min\limits_{x\in X}\frac{1}{N}\sum\limits_{\omega\in\Omega_{N}^{m}}Q(x,\xi^{m}_{\omega}).
    Estimate the out-of-sample cost of x^\hat{x} as z^Nm:=1NωΩNmQ(x^,ξωm)\hat{z}^{m}_{N}:=\frac{1}{N}\sum\limits_{\omega\in\Omega_{N}^{m}}Q(\hat{x},\xi^{m}_{\omega}).
    Estimate the optimality gap of x^\hat{x} as GNm:=z^NmzNmG^{m}_{N}:=\hat{z}^{m}_{N}-z^{m}_{N}.
end for
 Calculate the sample mean and sample variance of the gap as
G¯N=1Mm=1MGNm\qquad\bar{G}_{N}=\frac{1}{M}\sum_{m=1}^{M}G_{N}^{m}\quad and sG2=1M1m=1M(GNmG¯N)2\quad s_{G}^{2}=\frac{1}{M-1}\sum_{m=1}^{M}(G_{N}^{m}-\bar{G}_{N})^{2}.
 Calculate a normalized %100(1α)\%100(1-\alpha) upper bound on the optimality gap as
1z¯N(G¯N+tα;M1sGM)\frac{1}{\bar{z}_{N}}\left(\bar{G}_{N}+t_{\alpha;M-1}\frac{s_{G}}{\sqrt{M}}\right), where z¯N=1Mm=1MzNm\bar{z}_{N}=\frac{1}{M}\sum_{m=1}^{M}z^{m}_{N}.

In addition, we propose an MRP integrated SAA approach for candidate solution generation and quality assessment, given in Algorithm 2. On the one hand, a candidate solution is generated by solving an SAA problem with a sample of NN realizations. Then, we use the MRP to estimate an upper bound on the optimality gap of the candidate solution. If the solution is ϵ\epsilon-optimal, i.e., estimated upper bound on its optimality gap is less than or equal to a ϵ\epsilon threshold, the algorithm stops. Otherwise, the sample size increases until a good quality solution is found. The algorithm returns a candidate solution and its optimality gap.

Algorithm 2 MRP integrated SAA
Input: List of sample sizes NlistN_{list} and ϵ,α(0,1)\epsilon,\alpha\in(0,1).
Output: Solution x^\hat{x} and OptGap.
for NN in NlistN_{list} do
    Obtain a candidate solution x^N\hat{x}_{N} by solving the SAA problem (12).
    Calculate a normalized %100(1α)\%100(1-\alpha) upper bound on the optimality gap as MRPα(x^N)\textrm{MRP}_{\alpha}(\hat{x}_{N}).
    if MRPα(x^N)ϵ\textrm{MRP}_{\alpha}(\hat{x}_{N})\leq\epsilon then
       x^x^N\hat{x}\leftarrow\hat{x}_{N} and OptGapMRPα(x^N)\textrm{OptGap}\leftarrow\textrm{MRP}_{\alpha}(\hat{x}_{N}).
       exit for loop
    end if
end for

We end this section by noting that each of the DEF, presented in Section 3.2, the L-shaped algorithm, presented in Section 4.1, and the heuristic algorithm, presented in Section 4.2, can be used to solve the SAA problem and obtain a candidate solution. However, the probability of scenarios ωΩ\omega\in\Omega, ρω\rho_{\omega}, must change in the formulations so that it reflect the scenarios in a sample ΩN\Omega_{N}. Let N^\hat{N} and nωn_{\omega} represent the set of unique scenarios in ΩN\Omega_{N} and the number of their occurrences. Thus, in the described DEF, L-shaped algorithm, and TS algorithm, ωΩρω()\sum_{\omega\in\Omega}\rho_{\omega}(\cdot) changes to 1NωΩN()\frac{1}{N}\sum_{\omega\in\Omega_{N}}(\cdot) or equivalently, 1NωN^nω()\frac{1}{N}\sum_{\omega\in\hat{N}}n_{\omega}(\cdot). Accordingly, in the L-shaped method, we generate one optimality cut for each unique scenario ωN^\omega\in\hat{N} by solving |N^||\hat{N}| number of subproblems at each iteration.

5 Numerical Experiments

In Section 5.1, we describe the experimental setup. Then, in Section 5.2 and 5.3, we assess solution quality and computational performance of the proposed L-shaped and heuristic algorithms applied to a SAA problem, respectively.

5.1 Experimental Setup

We generated real-world inspired instances from our automobile manufacturer partner’s assembly line and planning information. As given in Table 3, we generated three types of instances: (1) small-sized instances with 7-10 vehicles to assess the performance of L-shaped algorithm, (2) medium-sized instances with 40 vehicles to assess the performance of the TS algorithm for the one-scenario problem, (3) large-sized instances with 200, 300, and 400 vehicles to evaluate the performance of the TS algorithm. All instances have five stations, of which the first one is selected as the most restrictive station for EVs, the battery loading station. The rest are selected among other critical stations that conflict with the battery loading station.

Instance Type |V||V| |K||K| Number of Instances
Small 7, 8, 9, 10 5 30 ×\times 4
Medium 40 5 30 ×\times 1
Large 200, 300, 400 5 30 ×\times 3
Table 3: Data sets

The cycle time cc is 97 TU, and the station length ll is 120 TU for all but the battery loading station, which is two station lengths, 240 TU. The information about the distribution of the processing times is given in Table 4. It can be observed that the average and maximum processing times for each station are lower than the cycle time and the station length, respectively. Moreover, the ratio of the EVs is in the range of [0.25, 0.33] across all instances.

Time (s)
Station ID Min Mean Max
1 42.6 94.1 117.2
2 7.9 84.3 197.9
3 57.8 96.2 113.3
4 26.9 96.9 109.7
5 57.8 96.2 114.3
Table 4: Processing times distribution

We derived the failure rates from six months of historical data by performing predictive feature analysis on vehicles. Based on the analysis, two groups of vehicles are formed according to their failure probabilities, low-risk and high-risk vehicles, whose failure probabilities are in the range of [0.0, 0.01] and [0.2, 0.35], respectively. The failure probability is mostly higher for recently introduced features, e.g., the average failure probability of EVs is 50% higher than that of other vehicles. High-risk vehicles constitute [0.03, 0.05] of all vehicles. However, this percentage increases to [0.15, 0.25] for the small-sized instances in order to have a higher rate of failed vehicles. We note that the failures are not considered for the medium-sized instances since these instances are used for only the one-scenario problem, which does not involve failures by definition.

The number of failure scenarios, 2|V|2^{|V|}, increases exponentially in the number of vehicles. Thus, we generated an i.i.d random sample of NN realizations of the failure scenarios; hence, formed a SAA problem. For each failure scenario and vehicle, we first chose whether the vehicle was high risk or low risk (based on their prevalence). Then, depending on being a high-risk or low-risk vehicle, a failure probability was randomly selected from the respective range. Finally, it was determined whether the vehicle failed or not. In order to have a more representative sample of scenarios for large-sized instances, no low-risk vehicle was allowed to fail at any scenario.

For each parameter configuration, we generated 30 instances. The vehicles of each instance were randomly selected from a production day, respecting the ratios mentioned above. The algorithms were implemented in Python 3. For solving optimization problems we used Gurobi 9.0. The time limit is 600 seconds for all experiments unless otherwise stated. We run our experiments on computing nodes of the Clemson University supercomputer. The experiments with the exact solution approach were run on nodes with a single core and 15 GB of memory, and the experiments with the heuristic solution approach were on on nodes with 16 cores and 125 GB of memory.

5.2 Exact Solution Approach

In this section, we present results for the solution quality and computational performance of the L-shaped algorithm. We used MRP scheme, explained in Section 4.3, to assess the solution quality. We also compared the computational performance of the L-shaped algorithm with that of solving the DEF. We present the results for 120 small-sized instances consisting of 7 to 10 vehicles. We do not present the results for large-sized instances as our preliminary experiments showed that the number of instances that could be solved to optimality decreases drastically.

We also point that instead of solving a relaxed master problem to optimality at each iteration of the L-shaped algorithm, one can aim for just obtaining a feasible solution x^X\hat{x}\in X. This may result in saving a significant amount of computational time that would be otherwise spent on exploring solutions that are already eliminated in previous iterations. This kind of implementation, referred to as branch-and-Benders-cut (B&BC), is studied in the literature, see, e.g., (Hooker, 2011; Thorsteinsson, 2001; Codato and Fischetti, 2006). In our implementation of our proposed L-shaped algorithm, we used Gurobi’s Lazy constraint callback to generate cuts at a feasible integer solution found in the course of the branch-and-bound algorithm.

5.2.1 Solution Quality

Figure 4 shows the impact of sample size on the solution quality of the SAA problem. Observe that the improvement in the upper bound of the optimality gap, the MRP output, as the sample size increases from 100 to 1000 progressively. We set the number of replications MM to 30 and α=0.05\alpha=0.05 (95% confidence interval). While the mean of the optimality gap decreases gradually from 0.76% to 0.12%, a drastic enhancement is observed with the variance. We have 36 out of 120 solutions with an optimality gap of larger than 1% when the sample size is 100. However, all of the obtained solutions have less than a 1% optimality gap when the sample size is 1000. It can be seen in the figure that good solutions can be obtained with a sample size of 100, yet it is not assured due to the high variance of the approximation. Consequently, the results suggest that the sample size should be increased until the variance of objective estimation is small enough.

Refer to caption

Figure 4: Solution quality of the SAA problem based on sample sizes

Based on the results in Figure 4, we implemented the MRP integrated SAA scheme, presented in Section 4.3 and Algorithm 2, to balance the required computational effort and solution quality. We set α=0.05\alpha=0.05 (95% confidence interval) and ϵ=0.01\epsilon=0.01 in MRP. While it is ensured that we obtain solutions within a 1% optimality gap, most of the solutions are found with the least computational effort, e.g., at the first iteration with a sample size of 100. In Table 5, we provide key performance results to show the performance of the MRP integrated SAA scheme, where the number of replications MM is 30 and the MRP sample size NN is 5000. The average value for the optimal value, accepted candidate solution’s expected objective value, and optimality gap are presented in Table 5. The sample size of the accepted candidate solutions is 84, 20, 11, and 5 for the sample sizes Nlist={100,200,500,1000}N_{list}=\{100,200,500,1000\}, respectively. The average optimality gap is 0.2% which shows that SAA can produce high-quality solutions.

Statistical Lower
Bound 𝔼[zN]\mathbb{E}[z_{N}]
Estimated Objective
Value 𝔼[Q(x^N,ξω)]\mathbb{E}[Q(\hat{\textit{x}}_{N},\xi_{\omega})]
Optimality
Gap G¯N\bar{G}_{N}
55.80 55.91 0.11 (0.2%)
Table 5: Solution quality of the MRP integrated SAA

Additionally, we assess the solution quality of the one-scenario problem (i.e., the deterministic MMS problem without any car failure). Observe from Figure 5 that the average optimality gap is 23%, the maximum optimality gap is 274%, and the standard deviation is 39%. Comparing the performance of the SAA and the one-scenario problems shows that we can generate robust solutions by considering vehicle failures, which helps reduce work overloads by more than 20%.

Refer to caption

Figure 5: Solution quality of the one-scenario problem

5.2.2 Computational Performance

In this section, we conduct different experiments to compare the DEF and L-shaped algorithm. On the one hand, we assess the impact of using the improved model, described in Section 3.2, obtained by setting the processing time of failed vehicles to the cycle time, and compare the results with those obtained using the standard model. The DEF corresponding to the standard and improved models are denoted as DstdD_{std} and DimpD_{imp}, respectively. Similarly, the L-shaped algorithm corresponding to the standard and improved are denoted as LstdL_{std} and LimpL_{imp}. On the other hand, we assess the impact of our proposed cut selection strategy, described in Section 4.1, obtained using 1\ell_{1} norm regularization to find a cut with the least number of zero coefficients. We used the cut selection strategy for the improved model, and denote the corresponding L-shaped algorithm as LimpcsL_{imp-cs}.

In Table 6, we present the results on the impact of the improved model and cut selection strategy to compare the DEF and L-shaped algorithm for solving the SAA problem. We report the average and standard deviation of the computational time (in seconds), labelled as μt\mu_{t} and σt\sigma_{t}, respectively, and the optimality gap, labelled as Gap (in percentage). Additionally, the number of instances that could not be solved optimally within the time limit is given in parenthesis under the Gap columns. All time results are the average of instances (out of 30 instances) that could be solved optimally within the time limit, while the Gap results are the average of the instances that could not be solved optimally. Based on the results in Section 5.2.1, we conducted the computational experiments on the SAA problem with sample sizes 100, 200, 500, and 1000.

DstdD_{std} DimpD_{imp} LstdL_{std} LimpL_{imp} LimpcsL_{imp-cs}
|V||V| N μt\mu_{t} (s) σt\sigma_{t} (s) Gap (%) μt\mu_{t} (s) σt\sigma_{t} (s) Gap (%) μt\mu_{t} (s) σt\sigma_{t} (s) Gap (%) μt\mu_{t} (s) σt\sigma_{t} (s) Gap (%) μt\mu_{t} (s) σt\sigma_{t} (s) Gap (%)
7 100 6.3 3.8 - 1.4 1.1 - 4.9 3.2 - 1.2 0.5 - 1.1 0.5 -
200 10.2 6.4 - 2.0 1.1 - 7.9 5.7 - 1.9 0.9 - 1.6 0.6 -
500 15.0 8.8 - 3.2 1.8 - 10.2 6.5 - 2.4 1.0 - 2.1 0.8 -
1000 19.5 11.2 - 4.4 2.5 - 11.9 8.0 - 2.8 1.1 - 2.4 0.9 -
\hdashline8 100 29.3 19.7 - 11.4 10.5 - 51.7 59.4 - 9.4 6.6 - 5.2 4.4 -
200 50.5 31.5 - 19.1 16.5 - 83.2 83.4 - 16.7 22.9 - 8.1 6.9 -
500 88.9 53.3 - 32.3 24.9 - 145.0 146.2 - 24.9 26.4 - 13.4 10.9 -
1000 127.9 81.9 - 44.3 35.9 - 159.1 150.8 (1) 0.31 30.7 26.5 - 15.6 14.2 -
\hdashline9 100 170.3 157.6 (2) 0.33 35.3 27.3 - 187.5 130.3 (9) 0.54 43.1 39.0 - 25.0 19.1 -
200 238.9 165.1 (7) 0.34 68.4 53.1 - 315.6 170.2 (14) 0.54 80.0 96.4 - 41.1 49.4 -
500 263.6 140.3 (12) 0.38 122.6 120.3 (1) 0.20 357.0 170.0 (17) 0.55 105.0 79.9 (1) 0.16 58.6 49.1 -
1000 317.7 140.4 (16) 0.44 204.2 153.5 (1) 0.13 366.3 198.4 (22) 0.55 155.6 100.9 (1) 0.07 87.1 66.5 -
\hdashline10 100 279.1 159.7 (12) 0.28 91.8 61.7 - 486.5 253.8 (25) 0.61 169.3 137.1 (3) 0.16 133.7 152.2 -
200 258.5 161.2 (23) 0.34 160.9 117.9 - 344.2 361.7 (28) 0.61 238.5 179.6 (4) 0.18 158.8 138.3 (2) 0.16
500 565.2 58.4 (26) 0.48 245.3 151.4 (6) 0.13 283.9 0.0 (29) 0.69 264.4 193.9 (12) 0.18 223.5 170.9 (7) 0.16
1000 479.0 89.4 (28) 0.53 294.1 149.2 (13) 0.18 191.0 0.0 (29) 0.72 266.0 170.8 (18) 0.26 273.6 189.6 (8) 0.19
Table 6: Computational performance of the DEF and L-shaped algorithms for the SAA problem of small-sized instances

Observe from Table 6 that using the improved model instead of the standard model decreased the computational time and optimality gap of both the DEF and L-shaped algorithm drastically. In particular, the solution time decreased for all instances. On average (over different |V||V| and NN), we observe a 67% and 70% decrease for the DEF and L-shaped algorithm, respectively. Additionally, the decrease in the standard deviation is around 64% and 74% for the DEF and L-shaped algorithm, respectively, when non-optimal solutions are left out. Moreover, the number of instances that could not be solved optimally is reduced by using the improved model: on average, 83% and 78% of those instances are solved optimally with DimpD_{imp} and LimpL_{imp}, respectively. Additionally, the remaining non-optimal solutions are enhanced as a reduction in optimality gaps is achieved.

Another drastic improvement is provided by our cut selection strategy. Comparing LimpL_{imp} and LimpcsL_{imp-cs} in Table 6 shows that the mean and standard deviation of the computational time, and optimality gap are reduced by 35%, 33%, and 18%, on average. Furthermore, an optimal solution is found by LimpcsL_{imp-cs} for 56% of the instances that could not be solved optimally within the time limit by LimpL_{imp}.

Finally, we compare DimpD_{imp} and LimpcsL_{imp-cs}. Observe from Table 6 that LimpcsL_{imp-cs} resulted in lower mean computational time by 31%, 59%, 45%, and 1% for instances with 7, 8, 9, and 10 vehicles, respectively, and by 15%, 28%, 38%, and 45% for instances with 100, 200, 500, 1000 scenarios, respectively. In the same order, the variance decreased by 56%, 58%, 38%, and -52% for instances with 7,8,9, and 10 vehicles, respectively, and by -1%, 18%, 41%, and 42% for instances with 100, 200, 500, 1000 scenarios, respectively. We conclude that our L-shaped algorithm outperforms the DEF for instances with up to 9 vehicles, and they provide comparable results for instances with 10 vehicles. Additionally, the superiority of the L-shaped algorithm over the DEF escalates as the number of scenarios increases.

5.3 Heuristic Solution Approach

In this section, we present results for the solution quality and computational performance of the TS algorithm. The solution quality is evaluated by employing the MRP scheme, explained in Section 4.3. We also assess the computational performance of the TS algorithm in various aspects.

We set the operator selection probabilities (weights) based on our preliminary experiments. The weights of swap, forward insertion, backward insertion, and inversion are 0.45, 0.10, 0.15, 0.30, respectively. We set the time limit for the one-scenario problem to 10 seconds, i.e., τone=10\tau_{one}=10 seconds, which leaves τfull=590\tau_{full}=590 seconds. Finally, based on the results of the quality assessment, we set the sample size NN to 1000 for the computational performance assessments.

5.3.1 Solution Quality

We solved the SAA problem of large-sized instances using the TS algorithm. To assess the solution quality, we then used the proposed MRP integrated SAA scheme, given in Algorithm 2, with the replication M=30M=30, MRP sample size N=20000N=20000, α=0.05\alpha=0.05 (95% confidence interval), ϵ=0.05\epsilon=0.05, and the list of sample sizes Nlist={1000,2500,5000}N_{list}=\{1000,2500,5000\}. Table 7 reports the key performance result to show the performance of the MRP integrated SAA scheme. While the maximum optimality gap is 3.7%, the average optimality gap is 0.28% which indicates that solving the SAA problem with the proposed TS algorithm can produce high-quality solutions. Figure 6 further shows that the optimality gap for most of the solutions is less than 1% with only five outliers out of 90 instances.

Statistical Lower
Bound 𝔼[zN]\mathbb{E}[z_{N}]
Estimated Objective
Value 𝔼[Q(x^N,ξω)]\mathbb{E}[Q(\hat{\textit{x}}_{N},\xi_{\omega})]
Optimality
Gap G¯N\bar{G}_{N}
239.59 240.26 0.67 (0.28%)
Table 7: Solution quality of the MRP integrated SAA

Refer to caption

Figure 6: Solution quality of the SAA problem

Moreover, we assess the solution quality of the one-scenario problem over large-sized instances by using the same procedure in order to show its robustness. Observe from Figure 7 that the average optimality gap is 24.8%, the maximum optimality gap is 76.5%, and the standard deviation is 10.8%. Comparing the performance of the SAA and one-scenario problems demonstrates that we can generate robust solutions by considering vehicle failures. Accordingly, we reassure with the industry-sized instances that we can reduce the work overloads by more than 20% by considering stochastic car failures and solving the corresponding problem efficiently.

Refer to caption

Figure 7: Solution quality of one-scenario problem

5.3.2 Computational Performance

To assess the computational performance of the TS algorithm, we conducted different tests: 1) compared the solution found with the TS algorithm with the solution found by an off-the-shelf solver for the one-scenario problem of medium-sized instances, 2) compared the solution found with the TS algorithm with the optimal solution found by LimpcsL_{imp-cs} approach for the SAA problem of small-sized instances, 3) compared the solution found with the TS algorithm with that of a simulated annealing (SA) algorithm for the SAA problem of large-sized instances, 4) and analyzed the convergence of TS algorithm for the one-scenario and SAA problems of large-sized instance. We executed 30 runs for each of the instances and tests.

Table 8 reports the results for the first set of computational experiments for the one-scenario problem. We note that we generated 30 instances that could be solved within three hours time limit with Gurobi (solving the DEF). The minimum, average, maximum, and standard deviation of the computational time (in seconds) are shown for the TS algorithm and Gurobi. Additionally, the average number of movements for the TS algorithm is reported in order to provide some insight about the efficiency of the implementation. The optimal solutions, for all 30 instances and for all 30 runs, are found in under 10 seconds by the TS algorithm. The average computational times are 1140 and 0.33 seconds for the Gurobi solver and TS algorithm, respectively. These results show that the proposed TS algorithm can consistently provide optimal solutions to the one-scenario problem in a very short amount of time by avoiding any local optima.

Time (s) Move (#)
Method Min Mean Max Std. Dev. Mean
Gurobi 2.37 1140.91 9373.08 2613.95 -
TS 6e-4 0.33 9.44 0.81 16940
Table 8: Computational performance of Gurobi and TS for the one-scenario problem of medium-sized instances
Time (s) Optimality Gap (%)
Optimally
Solved (%)
Move
(#)
Method Min Mean Max Std. Dev. Mean Max Std. Dev. Mean
Gurobi 5.02 68.91 210.06 50.99 0 0 0 100 -
TS 1e-4 0.08 0.28 0.08 0.14 2.79 0.41 83 628
Table 9: Computational performance of Gurobi and TS for the SAA problem of smalls-sized instances

Recall that as demonstrated in Section 5.2.2, LimpcsL_{imp-cs} outperformed Gurobi in solving the SAA problem for small-sized instances. In the second set of experiments, we compared the computational effectiveness of solving the SAA problem of small-sized instances with the TS algorithm and LimpcsL_{imp-cs}. To this end, we chose a sample size of 1000 and solved 30 small-sized instances optimally by using LimpcsL_{imp-cs}. We then solved the same set of instances using the TS algorithm until either an optimal solution was obtained or the time limit was reached. Table 9 shows that TS algorithm was able to find optimal solutions in 83% of the experiments, 746 out of 900. However, the average optimality gap for these non-optimal solutions is 0.14%, indicating that the TS algorithm is reliable in terms of optimality. For the TS algorithm, we recorded the computational time as the time until the last improvement is done, so we can see that the TS algorithm is very efficient, with an average runtime of 0.08 seconds.

We observed that the TS algorithm terminated at a local optima often when the number of vehicles is very small. Our hypothesis is that there are plateaus with the same objective value when the number of vehicles is large. However, this is not the case when there are only 10 vehicles as there are a limited number of sequences and each sequence has a different objective function value. Hence, in the third set of experiments, we compared the TS algorithm and a SA algorithm for large-sized instances to analyze the impact of the tabu list and accepting worse solutions on escaping a local optima. We utilized the same local search algorithm for the SA with two differences: 1) disabled the tabu rules and 2) enabled accepting worse solutions based on the acceptance criterion. For the SA algorithm, we set the starting temperature Tinit=10T_{init}=10. We also adopted a geometric cooling with the cooling constant α=0.999\alpha=0.999. The acceptance ratio is calculated using the Boltzmann distribution P(Δf,T)=ef(x)f(x)TP(\Delta f,T)=e^{-\frac{f(x^{\prime})-f(x)}{T}}, where xx^{\prime} is a new solution, xx is the incumbent solution, and TT is the current temperature.

We note that the Kruskal-Wallis test shows no significant difference between the computational performance of the TS and SA algorithms at a 95% confidence level. However, as illustrated in Figure 8, the TS algorithm produces better results and converges faster than the SA algorithm (averaged over of 900 runs). This result shows that the proposed TS algorithm is capable of exploiting the search space while generally avoiding premature convergence to local optima. Accordingly, we conclude that there is no need to accept worse solutions in the local search.

Refer to caption

Figure 8: Convergence comparison of the TS and SA algorithms

Finally, in the last set of experiments, we conducted an analysis in order to provide insight into the reliability of the proposed TS algorithm’s convergence. In particular, in Figure 9, we presented the box plots of the standard deviation of the objective values for the one-scenario and SAA problems of large-sized instances. Each data point represents the standard deviation of 30 runs (for each of the 90 large-sized instances). The average standard deviations for the one-scenario and SAA problems are 0.19 and 0.93, while the means of the objective values are 212.77 and 239.69, respectively. Accordingly, the average coefficient of variations, the ratio between the average standard deviation and mean, are 9e-4 and 4e-3, which indicates that the proposed TS algorithm provides highly reliable results for both of the problems.

Refer to caption

Figure 9: Convergence of the objective value with TS algorithm for the one-scenario and SAA problems

6 Conclusion

This paper studied mixed-model sequencing (MMS) problem with stochastic failures. To the best of our knowledge, this is the first study that considers stochastic failures of products in MMS. The products (vehicles) fail according to various characteristics and are then removed from the sequence, moving succeeding vehicles forward to close the gap. Vehicle failure may cause extra work overloads that could be prevented by generating a robust sequence at the beginning. Accordingly, we formulated the problem as a two-stage stochastic program, and improvements were presented for the second-stage problem. We employed the sample average approximation approach to tackle the exponential number of scenarios. We developed L-shaped decomposition-based algorithms to solve small-sized instances. The numerical experiments showed that the L-shaped algorithm outperforms the deterministic equivalent formulation, solved with an off-the-shelf solver, in terms of both solution quality and computational time. To solve industry-sized instances efficiently, we developed a greedy heuristic and a tabu search algorithm that is accelerated with problem-specific tabu rules. Numerical results showed that we can provide good quality solutions, with less than a 5% statistical optimality gap, to industry-sized instances in under ten minutes. The numerical experiments also indicated that we can generate good quality robust solutions by utilizing a sample of scenarios. In particular, we can reduce the work overload by more than 20%, for both small- and large-sized instances, by considering possible car failures.

6.1 Managerial Insights

Car manufacturers are facing several challenges due to the increasing ratio of EVs in production. EVs have significant differences compared to non-EVs which require specific treatment when creating the sequence for the mixed-model assembly line. One of the main challenges is the battery installation process. Unlike traditional vehicles, EVs have large and heavy batteries that need to be installed in a specific order to avoid damaging the vehicle or the battery itself. Accordingly, a huge processing time variation at the battery loading station arises from this difference, which requires special treatment to ensure that the assembly line can continue to produce high-quality vehicles efficiently.

We have observed that consecutive EVs induce a significant amount of work overload, which generally requires line stoppage even with the help of utility workers. Planning the sequence by ruling out back-to-back EVs does not guarantee that there will not be any occurrence of consecutive EVs. The failure of vehicles disrupts the planned sequence, and the necessity of considering car failures during the planning process increases as the difference between product types expands.

In this study, we focused on generating robust schedules that take into account the possible deleterious effects resulting from the divergence between electric and non-electric vehicles. However, it is worth noting that our proposed solution methodologies are equally applicable to any feature that causes similar variations at specific work stations.

6.2 Future Research

One direction for future research is the reinsertion of the failed vehicles back into the sequence. Even though the reinsertion process is being conducted via a real-time decision, a robust approach may increase the efficiency of production. Another direction for future research is to include stochastic processing times in addition to stochastic product failures. This may potentially generate more robust schedules, particularly in case a connection between failures and processing times is observed. Finally, there are similarities between MMS and some variants of the traveling salesman problem (TSP). Since the TSP is one of the most studied combinatorial optimization problems, the state-of-art solution methodologies presented for TSP may be adapted to MMS.

Acknowledgements

The authors acknowledge Clemson University for generous allotment of compute time on Palmetto cluster.

Appendix A Tabu List for Local Search Algorithm

Assume that we have two positions is selected for any operator to be applied, t1,t2t_{1},t_{2} with t1<t2t_{1}<t_{2}. The tabu movements for each operator is described below under two cases: the vehicle at the position t1t_{1} is 1) an EV, 2) not an EV.
Swap
1) The position t2t_{2} cannot be a neighbor of an EV, e.g., the vehicles at the positions t21t_{2}-1 and t2+1t_{2}+1 cannot be an EV.
2) The vehicle at the position t2t_{2} cannot be an EV if the position t1t_{1} is a neighbor of an EV.
Forward Insertion
1) The vehicles at the positions t2t_{2} or t21t_{2}-1 cannot be an EV.
2) At most one of the vehicles at the positions t11t_{1}-1 and t1+1t_{1}+1 can be an EV.
Backward Insertion
1) The vehicles at the positions t1t_{1} or t11t_{1}-1 cannot be an EV.
2) At most one of the vehicles at the positions t21t_{2}-1 and t2+1t_{2}+1 can be an EV.
Inversion
1) The position t2t_{2} cannot be a left neighbor of an EV, e.g., the vehicle at the position t2+1t_{2}+1 cannot be an EV.
2) If the vehicle at the position t1t_{1} is a right neighbor of an EV, then the vehicle at the position t2t_{2} cannot be an EV.

References

  • Agrawal and Tiwari (2008) Agrawal, S., Tiwari, M., 2008. A collaborative ant colony algorithm to stochastic mixed-model u-shaped disassembly line balancing and sequencing problem. International Journal of Production Research 46, 1405–1429.
  • Akgündüz and Tunalı (2010) Akgündüz, O.S., Tunalı, S., 2010. An adaptive genetic algorithm approach for the mixed-model assembly line sequencing problem. International Journal of Production Research 48, 5157–5179.
  • Akgündüz and Tunalı (2011) Akgündüz, O.S., Tunalı, S., 2011. A review of the current applications of genetic algorithms in mixed-model assembly line sequencing. International Journal of Production Research 49, 4483–4503.
  • Akpınar et al. (2013) Akpınar, S., Bayhan, G.M., Baykasoglu, A., 2013. Hybridizing ant colony optimization via genetic algorithm for mixed-model assembly line balancing problem with sequence dependent setup times between tasks. Applied Soft Computing 13, 574–589.
  • Bard et al. (1992) Bard, J.F., Dar-Elj, E., Shtub, A., 1992. An analytic framework for sequencing mixed model assembly lines. The International Journal of Production Research 30, 35–48.
  • Bolat (2003) Bolat, A., 2003. A mathematical model for selecting mixed models with due dates. International Journal of Production Research 41, 897–918.
  • Bolat and Yano (1992) Bolat, A., Yano, C.A., 1992. Scheduling algorithms to minimize utility work at a single station on a paced assembly line. Production Planning & Control 3, 393–405.
  • Boysen et al. (2009) Boysen, N., Fliedner, M., Scholl, A., 2009. Sequencing mixed-model assembly lines: Survey, classification and model critique. European Journal of Operational Research 192, 349–373.
  • Boysen et al. (2011) Boysen, N., Kiel, M., Scholl, A., 2011. Sequencing mixed-model assembly lines to minimise the number of work overload situations. International Journal of Production Research 49, 4735–4760.
  • Brammer et al. (2022) Brammer, J., Lutz, B., Neumann, D., 2022. Stochastic mixed model sequencing with multiple stations using reinforcement learning and probability quantiles. OR Spectrum 44, 29–56.
  • Cano-Belmán et al. (2010) Cano-Belmán, J., Ríos-Mercado, R.Z., Bautista, J., 2010. A scatter search based hyper-heuristic for sequencing a mixed-model assembly line. Journal of Heuristics 16, 749–770.
  • Celano et al. (1999) Celano, G., Fichera, S., Grasso, V., La Commare, U., Perrone, G., 1999. An evolutionary approach to multi-objective scheduling of mixed model assembly lines. Computers & Industrial Engineering 37, 69–73.
  • Codato and Fischetti (2006) Codato, G., Fischetti, M., 2006. Combinatorial benders’ cuts for mixed-integer linear programming. Operations Research 54, 756–766.
  • Dar-El (1978) Dar-El, E.M., 1978. Mixed-model assembly line sequencing problems. Omega 6, 313–323.
  • Dong et al. (2014) Dong, J., Zhang, L., Xiao, T., Mao, H., 2014. Balancing and sequencing of stochastic mixed-model assembly u-lines to minimise the expectation of work overload time. International Journal of Production Research 52, 7529–7548.
  • Fattahi and Salehi (2009) Fattahi, P., Salehi, M., 2009. Sequencing the mixed-model assembly line to minimize the total utility and idle costs with variable launching interval. The International Journal of Advanced Manufacturing Technology 45, 987–998.
  • Gottlieb et al. (2003) Gottlieb, J., Puchta, M., Solnon, C., 2003. A study of greedy, local search, and ant colony optimization approaches for car sequencing problems, in: Workshops on Applications of Evolutionary Computation, Springer. pp. 246–257.
  • Hooker (2011) Hooker, J., 2011. Logic-based methods for optimization: combining optimization and constraint satisfaction. John Wiley & Sons.
  • Hottenrott et al. (2021) Hottenrott, A., Waidner, L., Grunow, M., 2021. Robust car sequencing for automotive assembly. European Journal of Operational Research 291, 983–994.
  • Hyun et al. (1998) Hyun, C.J., Kim, Y., Kim, Y.K., 1998. A genetic algorithm for multiple objective sequencing problems in mixed model assembly lines. Computers & Operations Research 25, 675–690.
  • Kilbridge (1963) Kilbridge, M., 1963. The assembly line model-mix sequencing problem, in: Proceedings of the 3rd International Conference on Operations Research, 1963.
  • Kim and Jeong (2007) Kim, S., Jeong, B., 2007. Product sequencing problem in mixed-model assembly line to minimize unfinished works. Computers & Industrial Engineering 53, 206–214.
  • Kim et al. (1996) Kim, Y.K., Hyun, C.J., Kim, Y., 1996. Sequencing in mixed model assembly lines: a genetic algorithm approach. Computers & Operations Research 23, 1131–1145.
  • Kleywegt et al. (2002) Kleywegt, A.J., Shapiro, A., Homem-de Mello, T., 2002. The sample average approximation method for stochastic discrete optimization. SIAM Journal on Optimization 12, 479–502.
  • Kucukkoc and Zhang (2014) Kucukkoc, I., Zhang, D.Z., 2014. Simultaneous balancing and sequencing of mixed-model parallel two-sided assembly lines. International Journal of Production Research 52, 3665–3687.
  • Leu et al. (1996) Leu, Y.Y., Matheson, L.A., Rees, L.P., 1996. Sequencing mixed-model assembly lines with genetic algorithms. Computers & Industrial Engineering 30, 1027–1036.
  • Liu et al. (2014) Liu, Q., Wang, W.x., Zhu, K.r., Zhang, C.y., Rao, Y.q., 2014. Advanced scatter search approach and its application in a sequencing problem of mixed-model assembly lines in a case company. Engineering Optimization 46, 1485–1500.
  • Mak et al. (1999) Mak, W.K., Morton, D.P., Wood, R.K., 1999. Monte Carlo bounding techniques for determining solution quality in stochastic programs. Operations Research Letters 24, 47–56.
  • Homem-de Mello and Bayraksan (2014) Homem-de Mello, T., Bayraksan, G., 2014. Monte carlo sampling-based methods for stochastic optimization. Surveys in Operations Research and Management Science 19, 56–85.
  • Mosadegh et al. (2017) Mosadegh, H., Fatemi Ghomi, S., Süer, G.A., 2017. Heuristic approaches for mixed-model sequencing problem with stochastic processing times. International Journal of Production Research 55, 2857–2880.
  • Mosadegh et al. (2020) Mosadegh, H., Ghomi, S.F., Süer, G.A., 2020. Stochastic mixed-model assembly line sequencing problem: Mathematical modeling and q-learning based simulated annealing hyper-heuristics. European Journal of Operational Research 282, 530–544.
  • Özcan et al. (2011) Özcan, U., Kellegöz, T., Toklu, B., 2011. A genetic algorithm for the stochastic mixed-model u-line balancing and sequencing problem. International Journal of Production Research 49, 1605–1626.
  • Pil and Holweg (2004) Pil, F.K., Holweg, M., 2004. Linking product variety to order-fulfillment strategies. Interfaces 34, 394–403.
  • Ponnambalam et al. (2003) Ponnambalam, S., Aravindan, P., Rao, M.S., 2003. Genetic algorithms for sequencing problems in mixed model assembly lines. Computers & Industrial Engineering 45, 669–690.
  • Rahimi-Vahed et al. (2007a) Rahimi-Vahed, A., Mirghorbani, S., Rabbani, M., 2007a. A new particle swarm algorithm for a multi-objective mixed-model assembly line sequencing problem. Soft Computing 11, 997–1012.
  • Rahimi-Vahed et al. (2007b) Rahimi-Vahed, A.R., Rabbani, M., Tavakkoli-Moghaddam, R., Torabi, S.A., Jolai, F., 2007b. A multi-objective scatter search for a mixed-model assembly line sequencing problem. Advanced Engineering Informatics 21, 85–99.
  • Rahmaniani et al. (2017) Rahmaniani, R., Crainic, T.G., Gendreau, M., Rei, W., 2017. The benders decomposition algorithm: A literature review. European Journal of Operational Research 259, 801–817.
  • Scholl et al. (1998) Scholl, A., Klein, R., Domschke, W., 1998. Pattern based vocabulary building for effectively sequencing mixed-model assembly lines. Journal of Heuristics 4, 359–381.
  • Sikora (2021) Sikora, C.G.S., 2021. Benders’ decomposition for the balancing of assembly lines with stochastic demand. European Journal of Operational Research 292, 108–124.
  • Solnon (2000) Solnon, C., 2000. Solving permutation constraint satisfaction problems with artificial ants, in: European Conference on Artificial Intelligence, Citeseer. pp. 118–122.
  • Thorsteinsson (2001) Thorsteinsson, E.S., 2001. Branch-and-check: A hybrid framework integrating mixed integer programming and constraint logic programming, in: International Conference on Principles and Practice of Constraint Programming, Springer. pp. 16–30.
  • Tsai (1995) Tsai, L.H., 1995. Mixed-model sequencing to minimize utility work and the risk of conveyor stoppage. Management Science 41, 485–495.
  • Wei-qi et al. (2011) Wei-qi, L., Qiong, L., Chao-yong, Z., Xin-yu, S., 2011. Hybrid particle swarm optimization for multi-objective sequencing problem in mixed model assembly lines. Computer Integrated Manufacturing System 17, 0.
  • Yano and Rachamadugu (1991) Yano, C.A., Rachamadugu, R., 1991. Sequencing to minimize work overload in assembly lines with product options. Management Science 37, 572–586.
  • Zandieh and Moradi (2019) Zandieh, M., Moradi, H., 2019. An imperialist competitive algorithm in mixed-model assembly line sequencing problem to minimise unfinished works. International Journal of Systems Science: Operations & Logistics 6, 179–192.
  • Zhang et al. (2020) Zhang, B., Xu, L., Zhang, J., 2020. A multi-objective cellular genetic algorithm for energy-oriented balancing and sequencing problem of mixed-model assembly line. Journal of Cleaner Production 244. Https://doi.org/10.1016/j.jclepro.2019.118845.
  • Zhao et al. (2007) Zhao, X., Liu, J., Ohno, K., Kotani, S., 2007. Modeling and analysis of a mixed-model assembly line with stochastic operation times. Naval Research Logistics 54, 681–691.
  • Zhu and Zhang (2011) Zhu, Q., Zhang, J., 2011. Ant colony optimisation with elitist ant for sequencing problem in a mixed model assembly line. International Journal of Production Research 49, 4605–4626.