Advertisement
Guest User

Purchased

a guest
Oct 22nd, 2019
255
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
Latex 317.77 KB | None | 0 0
  1. \documentclass[nonblindrev]{informs3}
  2. %\documentclass[nonblindrev]msom,
  3. %\documentclass[mnsc,blindrev]{informs3}
  4. %\documentclass[opre,nonblindrev]{informs3} % current default for manuscript submission
  5. %\documentclass[nonblindrev]{informs3}
  6.  
  7.  
  8. \OneAndAHalfSpacedXI
  9. %%\OneAndAHalfSpacedXII % Current default line spacing
  10. %%\DoubleSpacedXII
  11. %%\DoubleSpacedXI
  12.  
  13. % If hyperref is used, dvi-to-ps driver of choice must be declared as
  14. %   an additional option to the \documentclass. For example
  15. %\documentclass[dvips,mnsc]{informs3}      % if dvips is used
  16. %\documentclass[dvipsone,mnsc]{informs3}   % if dvipsone is used, etc.
  17.  
  18. % Private macros here (check that there is no clash with the style)
  19.  
  20.  
  21.  
  22.  
  23.  
  24.  
  25. \usepackage{boldline}
  26. \usepackage[table]{xcolor}
  27. \usepackage{tabularx,booktabs}
  28. \usepackage{array}
  29. \usepackage{endnotes}
  30. \usepackage{amsmath}
  31. \usepackage{longtable}
  32. \usepackage{adjustbox}
  33.  
  34. \usepackage{multirow}
  35. \usepackage{framed}
  36. \usepackage{makecell}
  37. \usepackage{threeparttable}
  38. \usepackage{comment}
  39. \usepackage{bm}
  40. \usepackage{subfloat}
  41. %\usepackage[flushleft]{threeparttable}
  42. \usepackage{subcaption}
  43. \usepackage{tabularx}
  44. \usepackage{float}
  45. \usepackage{rotating}
  46. \usepackage{caption}
  47.  
  48.  
  49.  
  50. \let\footnote=\endnote
  51. \let\enotesize=\normalsize
  52. \def\notesname{Endnotes}%
  53. \def\makeenmark{$^{\theenmark}$}
  54. \def\enoteformat{\rightskip0pt\leftskip0pt\parindent=1.75em
  55.  \leavevmode\llap{\theenmark.\enskip}}
  56. \newcommand{\mb}[1]{\mbox{\boldmath $#1$}}
  57. \newcommand{\mbt}[1]{\mbox{\boldmath $\tilde{#1}$}}
  58. \newcommand{\mbst}[1]{{\mbox{\boldmath \scriptsize{$\tilde{#1}$}}}}
  59. \newcommand {\E}{{\rm E}}
  60. %\newcommand{\proof}{{\bf Proof : }}
  61. %\newcommand{\qed}{\rule{7pt}{7pt}}
  62. %\newcommand{\remark}{{\bf Remark : }}
  63. \newcommand{\defi}{\stackrel{\Delta}{=}}
  64. %\newtheorem{theorem}{Theorem}
  65. %\newtheorem{lemma}{Lemma}
  66. %\newtheorem{prop}{Proposition}
  67. %\newtheorem{coro}{Corollary}
  68. %\newtheorem{claim}{Claim}
  69. \newtheorem{Defi}{Definition}
  70. %\usepackage{ntheorem}
  71. %\theoremstyple{break}
  72. \theoremstyle{break}
  73. \newtheorem{algorithm}{Algorithm}
  74.  
  75. \def\blot{\quad \mbox{$\vcenter{ \vbox{ \hrule height.4pt
  76.      \hbox{\vrule width.4pt height.9ex \kern.9ex \vrule width.4pt}
  77.           \hrule height.4pt}}$}}
  78. % Natbib setup for author-year style
  79. \usepackage{natbib}
  80. \bibpunct[, ]{(}{)}{,}{a}{}{,}%
  81.  \def\bibfont{\small}%
  82.  \def\bibsep{\smallskipamount}%
  83.  \def\bibhang{24pt}%
  84.  \def\newblock{\ }%
  85.  \def\BIBand{and}%
  86.  
  87. %\usepackage{threeparttable}
  88.  
  89. %% Setup of theorem styles. Outcomment only one.
  90. %% Preferred default is the first option.
  91. \TheoremsNumberedThrough     % Preferred (Theorem 1, Lemma 1, Theorem 2)
  92. %\TheoremsNumberedByChapter  % (Theorem 1.1, Lema 1.1, Theorem 1.2)
  93. \ECRepeatTheorems
  94.  
  95. %% Setup of the equation numbering system. Outcomment only one.
  96. %% Preferred default is the first option.
  97. \EquationsNumberedThrough    % Default: (1), (2), ...
  98. %\EquationsNumberedBySection % (1.1), (1.2), ...
  99.  
  100. % For new submissions, leave this number blank.
  101. % For revisions, input the manuscript number assigned by the on-line
  102. % system along with a suffix ".Rx" where x is the revision number.
  103. \MANUSCRIPTNO{}
  104. %(Please, provide the manuscript number!)
  105.  
  106.  
  107. \def \a{{\mbox{\boldmath $a$}}}
  108. \def \b{{\mbox{\boldmath $b$}}}
  109. \def \c{{\mbox{\boldmath $c$}}}
  110. \def \d{{\mbox{\boldmath $d$}}}
  111. \def \e{{\mbox{\boldmath $e$}}}
  112. \def \f{{\mbox{\boldmath $f$}}}
  113. \def \g{{\mbox{\boldmath $\mathbf{g}$}}}
  114. \def \h{{\mbox{\boldmath $h$}}}
  115. \def \q{{\mbox{\boldmath $q$}}}
  116. \def \p{{\mbox{\boldmath $p$}}}
  117. \def \u{{\mbox{\boldmath $u$}}}
  118. \def \v{{\mbox{\boldmath $v$}}}
  119. \def \w{{\mbox{\boldmath $w$}}}
  120. \def \r{{\mbox{\boldmath $r$}}}
  121. \def \s{{\mbox{\boldmath ${s}$}}}
  122. \def \x{\mathbf{x}}
  123. \def \b{\mathbf{b}}
  124. \def \y{\mathbf{y}}
  125. \def \z{\mathbf{z}}
  126. \def \q{\mathbf{q}}
  127. \def \p{\mathbf{p}}
  128. \def \g{\mathbf{g}}
  129. \def \s{\mathbf{s}}
  130. \def \t{\mathbf{t}}
  131. \def \d{\mathbf{d}}
  132. \def \u{\mathbf{u}}
  133. \def \v{\mathbf{v}}
  134. \def \h{\mathbf{h}}
  135. \def \r{\mathbf{r}}
  136. %\def \s{{\mbox{\boldmath $\mathbf{s}$}}}
  137. \def \t{{\mbox{\boldmath $t$}}}
  138. %\def \mathbf{x}{{\mbox{\boldmath $\mathbf{x}$}}}
  139. \def \lx{{\mbox{\footnotesize $\mathbf{x}$}}}
  140. \def \ly{{\mbox{\footnotesize $\y$}}}
  141. \def \y{{\mbox{\boldmath $\mathbf{y}$}}}
  142. %\def \z{{\mbox{\boldmath $\tilde{z}$}}}
  143. \def \z{{\mbox{\boldmath $\mathbf{z}$}}}
  144. \def \rz{{\mbox{\boldmath $\tilde{\mathbf{z}}$}}}
  145. \def \rs{{\mbox{\boldmath $\tilde{{s}}$}}}
  146. %\def \rs{{\mbox{\boldmath $\tilde{\mathbf{s}}$}}}
  147. \def \rw{{\mbox{\boldmath $\tilde{\mathbf{w}}$}}}
  148. \def \rrw{{\mbox{$\tilde{w}$}}}
  149.  
  150.  
  151.  
  152.  
  153.  
  154. \def \hxi{{\mbox{\boldmath $\xi$}}}
  155. \def \hnu{{\mbox{\boldmath $\nu$}}}
  156. \def \hpi{{\mbox{\boldmath $\pi$}}}
  157. \def \hiota{{\mbox{\boldmath $\iota$}}}
  158. \def \hpsi{{\mbox{\boldmath $\psi$}}}
  159. \def \hvarpi{{\mbox{\boldmath $\varpi$}}}
  160. \def \hmu{{\mbox{\boldmath $\mu$}}}
  161. \def \hchi{{\mbox{\boldmath $\chi$}}}
  162. \def \hlambda{{\mbox{\boldmath $\lambda$}}}
  163. \def \halpha{{\mbox{\boldmath $\alpha$}}}
  164. \def \hbeta{{\mbox{\boldmath $\beta$}}}
  165. \def \hrho{{\mbox{\boldmath $\rho$}}}
  166. \def \hvarrho{{\mbox{\boldmath $\varrho$}}}
  167. \def \hkappa{{\mbox{\boldmath $\kappa$}}}
  168. \def \bxi{{\mbox{\boldmath {\footnotesize $z$}}}}
  169. \def \heta{{\mbox{\boldmath $\eta$}}}
  170. \def \hzeta{{\mbox{\boldmath $\zeta$}}}
  171. \def \htheta{{\mbox{\boldmath $\theta$}}}
  172. \def \hvartheta{{\mbox{\boldmath $\vartheta$}}}
  173. \def \hvarphi{{\mbox{\boldmath $\varphi$}}}
  174. \def \hdelta{{\mbox{\boldmath $\delta$}}}
  175. \def \hdelta{{\mbox{\boldmath $\delta$}}}
  176. \def \htau{{\mbox{\boldmath $\tau$}}}
  177. \def \hphi{{\mbox{\boldmath $\phi$}}}
  178. \def \hepsilon{{\mbox{\boldmath $\epsilon$}}}
  179.  
  180.  
  181. \def \homega{{\mbox{\boldmath $\omega$}}}
  182. \def \hgamma{{\mbox{\boldmath $\gamma$}}}
  183. \def \hsigma{{\mbox{\boldmath $\sigma$}}}
  184. \def \hGamma{{\mbox{\boldmath $\Gamma$}}}
  185.  
  186. \def \hpi{{\mbox{\boldmath $\pi$}}}
  187. \def \hkappa{{\mbox{\boldmath $\kappa$}}}
  188. \def \htheta{{\mbox{\boldmath $\theta$}}}
  189. \def \hvartheta{{\mbox{\boldmath $\vartheta$}}}
  190. \def \hvarphi{{\mbox{\boldmath $\varphi$}}}
  191. \def \hphi{{\mbox{\boldmath $\phi$}}}
  192. \def \hvarsigma{{\mbox{\boldmath $\varsigma$}}}
  193. \def \hvarpi{{\mbox{\boldmath $\varpi$}}}
  194. \def \hUpsilon{{\mbox{\boldmath $\Upsilon$}}}
  195.  
  196. \def \D{{\mbox{\boldmath $D$}}}
  197.  
  198. \def \bB{{\mbox{\boldmath $B$}}}
  199. \def \C{{\mbox{\boldmath $C$}}}
  200. \def \H{{\mbox{\boldmath $\mathbf{H}$}}}
  201. \def \A{{\mbox{\boldmath $A$}}}
  202. \def \T{{\mbox{\boldmath $T$}}}
  203. \def \X{{\mbox{\boldmath $Y$}}}
  204. \def \Y{{\mbox{\boldmath $Y$}}}
  205. \def \bM{{\mbox{\boldmath $\mathbf{M}$}}}
  206. \def \bF{{\mbox{\boldmath $F$}}}
  207. \def \hL{{\mbox{\boldmath $\mathcal{L}$}}}
  208. \def \bV{{\mbox{$\boldsymbol{V}$}}}
  209. \def \bU{{\mbox{$\boldsymbol{U}$}}}
  210. \def \hT{{\mbox{\boldmath $\mathcal{T}$}}}
  211. \def \U{{\mbox{\boldmath $\mathcal{U}$}}}
  212. %\def \F{{\mbox{\boldmath $\mathcal{F}$}}}
  213. \def \hP{{\mbox{$\mathcal{P}$}}}
  214. \def \bI{{\mbox{\boldmath $I$}}}
  215. \def \T{{\mbox{\boldmath $T$}}}
  216. \def \W{{\mbox{\boldmath $W$}}}
  217. \def \bK{{\mbox{\boldmath $K$}}}
  218. \def \bQ{{\mbox{\boldmath $Q$}}}
  219. \def \bG{{\mbox{\boldmath $\mathbf{G}$}}}
  220. \def \bE{{\mbox{\boldmath $E$}}}
  221. \def \L{{\mbox{\boldmath $\mathcal{L}$}}}
  222. %\def \S{{\mbox{\boldmath $\mathbf{S}$}}}
  223. \def \S{{\mbox{\boldmath ${S}$}}}
  224. %\def \rS{{\mbox{\boldmath $\tilde{\mathbf{S}}$}}}
  225. \def \rS{{\mbox{\boldmath $\tilde{{S}}$}}}
  226. \def \I{{\mbox{\boldmath $\mathcal{I}$}}}
  227. \def \P{\mathbb{P}}
  228. \def \E{\mathbb{E}}
  229. \def \F{\mathcal{F}}
  230. \def \Q{\mathbb{Q}}
  231. \def \G{\mathcal{G}}
  232. \def \J{{\mbox{\boldmath $\mathcal{J}$}}}
  233. \def \B{{\mbox{\boldmath $\mathcal{B}$}}}
  234. \def \K{{\mbox{\boldmath $\mathcal{K}$}}}
  235. \def \K{{\mbox{$\mathcal{K}$}}}
  236. \def \O{{\mbox{\boldmath $\mathcal{O}$}}}
  237. \def \M{{\mbox{$\mathcal{M}$}}}
  238. \def \hW{{\mbox{\boldmath $\mathcal{W}$}}}
  239. \def \N{{\mbox{\boldmath $\mathcal{N}$}}}
  240.  
  241.  
  242. %-----------------------------------------------------------------------------
  243. % To set spacing:
  244. \def\spacingset#1{\renewcommand{\baselinestretch}%
  245.  {#1}\small\normalsize}
  246. \newcommand{\resetspacing}{\spacingset{1.70}}
  247. \newcommand{\unitspacing}{\spacingset{1.0}}
  248. \newcommand{\tightspacing}{\spacingset{1.25}}
  249. %-----------------------------------------------------------------------------
  250.  
  251.  
  252. \begin{document}
  253. \RUNAUTHOR{ }
  254.  
  255. \RUNTITLE{Redundancy Optimization with Side Information}
  256.  
  257. \TITLE{Redundancy Optimization with Side Information: State-Dependent Distributionally Robust Models}
  258. %\author{Shuming Wang}
  259. % use optional labels to link authors explicitly to addresses:
  260.  
  261. \ARTICLEAUTHORS{%
  262. }
  263.  
  264.  
  265. %\author{}
  266. %\affil[1]{School of Economics and Management, University of Chinese Academy of Sciences, China}
  267. %\affil[2]{University of Chinese Academy of Sciences, China}
  268. %% \address[label2]{}
  269. \ABSTRACT{{\color{red}
  270. In this paper, we present a state-dependent, distributionally robust framework for solving a series-parallel, multi-type component, mixed redundancy strategy system with uncertain component lifetimes. We assume that the distribution of component lifetimes are not exactly known, instead only partial distributional information (like mean and dispersion) can be extracted from the data set. We aim to product a system design that are reliable enough even under the worst possible distribution, given the partial information extracted. Moreover, We introduce clustering process in extracting information, which also includes techniques of dimension reduction and side information. We extract distributional information from each cluster, instead of from the entire data set. This helps us to incorporate more information into the ambiguity set, which results in smaller ambiguity set and therefore better result. Although the model itself is highly nonlinear, we utilize linearization technique to transform the model into a mixed integer linear program, without adding integer decision variables other than the original ones. This makes the solving the problem tractable, and easy to implement by off-the-shelves library. Finally, we present computational study to support our theoretical results.}
  271. }
  272. \KEYWORDS {Redundancy optimization, lifetime uncertainty, mixed redundancy strategies, robust optimization, linear programming, clustering}
  273.  
  274.  
  275. \maketitle
  276. %\end{frontmatter}
  277.  
  278.  
  279. \section{Introduction}
  280. {\color{blue}
  281. Redundancy optimization or redundancy allocation, roughly speaking, is  to determine the most cost-efficient allocations for redundancies while keeping system reliability above the predefined level (Kuo and Wan~2007)). It is well-known in multiple engineering domains, {\it e.g.}, railway engineering, nuclear engineering and aerospace engineering, to name a few. In a redundancy system, the cold-standby and the active-parallel are two typical and commonly used strategies in redundancy configuration ({\color{red} Ardakan~et~al.~2014}), and lifetime uncertainty of components has always been a crucial modeling issue for computing the system reliability and optimizing redundancy allocations, under different redundancy strategies.
  282.  
  283.  
  284.  
  285.  
  286. As a famous optimization problem under uncertainty, redundancy optimization has been extensively studied in the past decades. Different approaches have been proposed to deal with the modeling difficulty on the lifetime uncertainty and the resulting redundancy optimization problems. For instance, the optimization models with deterministic reliability levels of components({\it e.g.}, Coit~et~al.~1996, Kulturel-Konak~et~al.~2003, Ardakan~et~al.~2014) and with uncertain component reliability levels ({\it e.g.}, Feizollahi and Modarres~2012,  Feizollahi~et~al.~2014,~Feizollahi~et~al.~2015);  the stochastic programming approaches with known lifetime distributions of components ({\it e.g.}, Coit and Smith~1998,  Prasad et al.~2001, Zhao and Liu~2003); and most recently, the distributionally robust optimization (DRO) approach with uncertain lifetime distributions (Wang~et~al.~2019). A detailed literature review is provided in Section~\ref{sec:LR}. Several difficulties have been resolved, yet new challenges are constantly emerging.
  287.  
  288.  
  289. On the one hand, all the above mentioned approaches leverage solely on utilizing the lifetime information, which might not be sufficient for the increasingly complex systems nowadays. Technology advancements constantly drive the engineering systems to evolve toward ever higher complexities with larger number of components. For instance, the recent autonomous cars involve multiple sensors, control units, communication devices, safety barriers, artificial intelligence (AI) components, etc., which render them much more complex than the conventional electromechanical style of cars. Another example is the high-speed train, where the new technologies, such as automatic train operation (ATO) system, 5G telecommunication devices and auxiliary braking systems, are being integrated into the train system for more efficient operations. New technologies bring great upgrade to the existing system, which however also create the new sources of failures. The complexity of the system could largely complicate the uncertainty in the occurrence of the time-to-failures or lifetimes, and there is call for more related useful side information (in addition to lifetime data), {\it e.g.}, working condition (frequency and intensity), seasonal information (temperature and wether), quality information of the components, to be incorporated for estimating more effectively the component lifetime characteristics for redundancy optimization.
  290.  
  291.  
  292. On the other hand, the above mentioned side information is becoming more and more available. Thanks to
  293. the pervasive sensing capabilities of new generation of industrial IoT systems, it is often possible to have the senor records multiple types of data (even realtime data stream) on the working environment of components.
  294. For instance, there are thousands of sensors installed in one electric multiple unit (EMU) train, such as temperature sensor, humidity sensor, accelerator, {\it etc}. The sensors transmit their readings at a very high frequency, {\it i.e.}, 100Hz, to the onboard database system in the train, and the train will upload all stored data to the station once it completes the daily operation. Among all these data collected, some can be importantly relevant to the particular failure events. For example, the vibration signal has a strong correlation to the degradation condition or the abnormal condition of a running wheel of a train. Therefore,
  295. such side information can improve the description to the failures as well as to the lifetime distributions, and should be incorporated into the model together with the lifetime data itself to enhance the decision quality of the redundancy allocation under lifetime uncertainty.
  296.  
  297.  
  298. In this paper, we consider a redundancy optimization problem of a series-parallel system with  mixed redundancy strategy of active-parallels and cold-standbys, where the distributions of component lifetimes are assume to be uncertain. We develop a new optimization framework, based on distributionally robust optimization with conditional states, that is able to effectively incorporate the side information on component lifetimes, via the advanced machine learning techniques, {\it e.g.}, clustering, regression trees and PCA, to realize a more effective decision making of redundancy allocation under distributional uncertainty of lifetimes. Among all the above mentioned related studies, our paper is most relevant to Wang~et~al.~(2019), yet with key extensions in
  299. several directions. We briefly summarize our major contributions as follows:
  300. \begin{itemize}
  301. \item We develops a new distributionally robust redundancy optimization framework with multiple conditional states, which is able to flexibly incorporate the side information related lifetimes. To the best of our knowledge, our study is the first redundancy optimization approach with distributional uncertainty that is equipped with the machinery for harnessing the side information via the advanced machine learning techniques. In addition, the parameter selection of our framework can be readily realized by a cross validation approach developed.
  302. \item From the modeling perspective, our present study generalizes naturally Wang~et~al.~(2019) in two folds: (i) The distributionally robust redundancy model of Wang~et~al.~(2019) can be treated as a special case of our current model with a single state. (ii) Our current model considers a set of general conic constraints to capture the possible marginal and/or cross correlations of the lifetimes of different types of components, which also is the generation of the absolute dispersion constraints in the ambiguity set considered in Wang~et~al.~(2019).
  303. \item Methodologically,  our developed distributionally robust redundancy optimization model, to our best knowledge, is also the first distributionally robust chance-constrained optimization model over the state-dependent ambiguity set with general conic distributional constraints, which can also be readily extended to the general chance-constrained optimization problems.
  304. \item Computationally, we show that the worst-case system reliability level over the state-dependent ambiguity set given redundancy design, in several common cases, can be efficiently computed by solving a tractable conic program ({\it e.g.}, LP or SOCP), while the resulting distributionally robust redundancy optimization model solves a mixed integer conic program ({\it e.g.}, MILP or MISOCP). Furthermore, we develop a computationally viable supergradient-based decomposition algorithm to further enhance the scalability of the resulting MIP problems.
  305. \end{itemize}
  306.  
  307.  
  308.  
  309. The remaining of the paper is organized as follows. Section~\ref{sec:LR} reviews the related studies on redundancy optimization. Section~\ref{sec:base} introduces the base problem formulation of redundancy optimization that we are studying. In section~\ref{sec:DRRM}, we introduce our distributionally robust redundancy optimization model with side information, and discuss the tractable reformulations of the developed model as well as the hyperparameter selection via cross validation. In Section~\ref{sec:algo}, we discuss the development of the supergradient-based decomposition algorithm. Finally, we present numerical experiments and a case study in Section~\ref{sec:CS}, and conclude our study in~\ref{sec:conclusion}.
  310.  
  311.  
  312. }
  313.  
  314. \paragraph{\bf Notations:}
  315. We use the tilde to denote a random parameter, {\it e.g.,} $\bm{\tilde{z}} \in \mathbb{R}^n$; ``$\mathbb{P}$" to denote a specific probability distribution; and $\bm{\tilde{z}} \sim \P$ to denote the random variable $\bm{\tilde{z}}$ with  probability distribution $\P$.
  316. For a random variable $\bm{\tilde{z}} \in \mathbb{R}^n$ with distribution $\P$ and function $\g:\mathbb{R}^n \mapsto \mathbb{R}^m$, we denote $\E_{\P}(\g(\bm{\tilde{z}}))$ as the expectation of random variable $\g(\bm{\tilde{z}})$ with the probability measure $\P$. We use  $\mathcal{P}\left( \mathbb{R}^{n}\right)$ to represent the collection of all probability distributions of a random variable of dimension $n$, and ``$\mathbb{F}$" is a set of distributions for modeling the distributional ambiguity.
  317.  
  318. \section{Literature Review}\label{sec:LR}
  319. In this section we provide a brief review of approaches to RAP, including RAP with deterministic component reliability, RAP with uncertain component reliability and RAP with uncertain component lifetimes. In addition, we also briefly review the studies on robust optimization that are methodologically related to our work.
  320.  
  321. {\bf RAP with deterministic component reliability} Early studies on RAP focuses on formulations in which components have deterministic reliability. These problems are in general NP-hard(Chern.~1992), and efforts have been put into approximate algorithms to make them tractable. Coit~et~al.~(1996) uses a combination of neural network and genetic algorithm to search for the minimum cost design for a series-parallel system, given a minimum reliability constraint. Genetic algorithm is also applied to a system with cold-standby redundancy strategy (Ardakan~et~al.~2014). Liang~et~al.~(2004) applies ant colony optimization to a series-parallel system, in which the failure rates of components when not in use are the same as when in use (i.e., active redundancy). Kulturel-Konak~et~al.~(2003) introduces tabu search to solve a singled type component series-parallel system, with k-out-of-n subsystems and exponential component failure times.
  322.  
  323. It can be seen that RAP with deterministic component reliability suffers from its inherent theoretical intractability. Researchers has to develop various heuristic algorithms to mitigate this drawback, and there is currently no algorithm that is superior to others in all cases. Moreover, the assumption of known deterministic component reliability level is often impractical due to lack of data, as discussed in previous section. As the result, the focus recently shifted to RAP with uncertain component reliability.
  324.  
  325.  
  326. {\bf RAP with uncertain component reliability.} Most papers on RAP with uncertainty consider variations of the individual component reliability levels. Bhunia~et~al.~(2010) considered a reliability optimization problem for a series system as a stochastic chance constrained optimization problem with interval-valued reliability levels of individual components; the problem was transformed into an unconstrained integer programming problem with interval coefficients and solved by metaheuristics. Tekiner-Mogulkoc and Coit (2011) discussed an RAP that minimizes the coefficient of variation of the system reliability estimate with respect to a minimum system reliability constraint, in a series-parallel system, and an exact algorithm based on linear integer programming and a heuristic approach based on combined neighborhood search were proposed to solve the problems when component mixing is or is not allowed, respectively. Sun et al.~(2017) considered the uncertain component state performance and state probability in RAP, where the experts' epistemic estimations as the
  327. uncertainty parameters were modeled to be set-valued, and the resulting model was solved by a local-search-based metaheuristics approach.
  328.  
  329.  
  330.  
  331. Recently, some studies have addressed component reliability uncertainty in RAP using robust optimization (Bertsimas and Sim~2004), where component reliability levels are allowed to vary within an {\em uncertainty set}, and the resulting worst-case system reliability level is considered in the optimization of the RAP.  Feizollahi and Modarres (2012) and Feizollahi~et~al.~(2014) considered active redundancy in series-parallel systems and developed a robust RAP with an interval uncertainty set and a robust RAP with polyhedral budgeted uncertainty set, respectively. In both studies, the structures of the resulting robust RAP problems were well investigated, and problems could be transformed and solved iteratively with a series of MIP instances. Furthermore, Feizollahi~et~al.~(2015) developed a robust cold-standby redundancy allocation model for series-parallel systems with budgeted uncertainty; the problem could also be solved by an MIP-based iterative algorithm.
  332.  
  333. In addition, other related studies in this vein include multicriteria RAP models (Coit~et~al.~2004, Zaretalab~et~al.~2015, Govindan~et~al.~2017) and system reliability evaluation due to the lack of knowledge or the imprecision of human estimation (Li~et~al.~2014). More related papers can be found in an excellent survey (Kuo and Wan~2007).
  334.  
  335.  
  336.  
  337.  
  338.  
  339. {\bf RAP with stochastic lifetimes.} In the RAP literature, only a few studies have explicitly considered lifetime distributions. For example, Coit and Smith~(1998) and Prasad et al.~(2001) have proposed the maximization of a percentile life of the system subject to a budget constraint. Zhao and Liu~(2003) developed stochastic programming models for both parallel and standby redundancy optimization problems, which are solved by simulating the component lifetimes from any given known probability distributions and using metaheuristics. In addition, several studies have also performed theoretical analysis on the stochastic comparisons of redundancy allocation for a very limited number of components ({\it e.g.,} two or three) based on the given lifetime distributions (Li and Hu~2008, Zhao~et~al.~2011). All of these studies require the exactly known probability distributions of component lifetimes, which in practice is often difficult to specify or calibrate. In addition, Coit and Smith~(2002) considered uncertain Weibull component lifetime distributions with random-scale parameters in RAP; their model maximizes a lower percentile of the system time-to-failure distribution. The solution also leverages the approach of metaheuristics due to the complicated problem structure.
  340.  
  341.  
  342. Most of the resulting redundancy optimization models of the above RAP studies, except for the robust optimization models, are difficult to solve, and the proposed solution approaches ({\it e.g.,} metaheuristics), in general, are intractable. Among the related studies in the above two streams, our work is closest to that of Feizollahi and Modarres (2012), Feizollahi~et~al.~(2014), Feizollahi~et~al.~(2015), and Zhao and Liu~(2003). The research gap can be established in the following aspects:
  343. \begin{itemize}
  344.  \item The stochastic RAP model of Zhao and Liu~(2003) assumes the specific probability distributions for component lifetimes, which, as discussed previously, cannot be easily calibrated in the practice of reliability engineering, while our proposed RAP model allows for the lifetime distributions to be uncertain and vary over a set of possible distributions that are characterized by the available (limited) lifetime information. On the other hand, the robust RAP models of Feizollahi and Modarres (2012, Feizollahi~et~al.~(2014) and Feizollahi~et~al.~(2015) are based on regular robust optimization, which may sacrifice some critical distributional information of lifetimes, while our proposed RAP model is able to incorporate the distributional characteristics of component lifetimes ({\it e.g.,} mean, dispersion, and support information) into the redundancy optimization.
  345.  \item Our proposed RAP model considers a general setting with different redundancy strategies as well as multiple types of components for redundancy, while Zhao and Liu~(2003), Feizollahi and Modarres (2012) and Feizollahi~et~al.~(2014) only considered active redundancy and a single type of component. On the other hand, Feizollahi~et~al.~(2015) only considered cold-standby redundancy with a single type of component, while the multiple component types would violate the optimization structure utilized by the authors. Furthermore, our model is also able to incorporate starting failures for cold-standbys and common-cause failures for active redundancies, respectively, which have not been attempted in any of these related studies.
  346.  \item Finally, the resulting redundancy optimization problem of our proposed RAP model is equivalent to an MILP problem, which does not induce any binary variables in addition to the original redundancy allocation variables (binaries), while all the above four RAP models result in either bilinear MIP formulations (Feizollahi and Modarres~2012, Feizollahi~et~al.~2014), a linear MIP formulation with additional auxiliary integers (Feizollahi~et~al.~2015), or a highly intractable problem that was approached by metaheuristics (Zhao and Liu~2003).
  347. \end{itemize}
  348.  
  349.  
  350. {\bf Robust optimization.} Our work is based on distributionally robust optimization (Wiesemann et al.~2014), which is different from the abovementioned regular robust optimization (Bertsimas and Sim~2004, Ben-Tal~et~al.~2017): the latter focuses on the uncertainty of the actual values of the parameters, while the former concerns the uncertainty of the probability distributions of the parameters and is able to utilize the distributional information. In particular, our developed RAP model belongs to the class of distributionally robust chance-constrained programs (DR-CCPs), where the probability distributions of the uncertain parameters are allowed to vary within a distributional set or {\it Chebyshev ambiguity set}, and the worst-case chance level (reliability level in our context) is protected ({\it i.e.,} required to be above a threshold). In DR-CCPs, the case of joint chance constraint (our model belongs to this case) in general is much more difficult than the single chance constraint (Pr\'{e}kopa~1998), and several approximation approaches have been proposed, for instance, Bonferroni's inequality (Nemirovski and Shapiro~2006, Bertsimas~et~al.~2017), $\phi$-divergence (Yanikoglu and den Hertog~2013) and  conditional value-at-risk (CVaR) (Chen~et~al.~2010, Zymler~et~al.~2012). Most recently, some exact and tractable models have also been developed if the ambiguity set is carefully designed. Hanasusanto~et~al.~(2017) considered a joint constraint of affine functions with uncertainty and an ambiguity set that contains mean and support information, as well as an upper bound of dispersion for the DR-CCP, and they proved that the model is tractable whenever the support set and the dispersion function can be represented via polynomially numerous tractable (linear, conic quadratic and/or semidefinite) constraints.  In addition, Xie and Ahmed~(2018) considered a power flow optimization problem with two-sided chance constraints over the ambiguity set, with only mean and covariance information, and derived a second-order conic representable set for the feasible set of the distributionally robust two-sided chance constraints. Most recently, Wang, et al. (2019) proposed to use Chebyshev ambiguity set with distributional information of mean, support and dispersion to solve a mixed-redundancy strategy RAP. Due to the unique structure of the problem ({\it e.g.,} the mixed cold-standby and active parallel redundancy strategy), the resulting system lifetime function is a joint constraint of nonlinear functions with uncertainty, but Wang, et al. managed to reformulate the model into a linear MIP, where the binary variables are the same as the original redundancy allocation variables, thus the problem becomes tractable.
  351.  
  352. Our work build on the work of Wang, et al.. In particular, we introduce clustering process on data, and used it to construct conditional ambiguity set. In this way, more information and be incorporated into the ambiguity set, which can lead to more favorable result. We utilize cross validation to improve the clustering result. Our ambiguity set is also very general, and can incorporate any information of the data set that is defined by a convex function. Methodologically, we are also the first to attempt conditional ambiguity set in the distributionally robust chance constrained optimization, which enriches the modeling opportunities for the ambiguous CCPs.
  353.  
  354. As for our introduction of clustering algorithm to distributionally robust optimization, it is in some way similar to the work of Shang~et~al.~(2017), in which support vector clustering is applied as part of a data-driven robust optimization framework. The incorporation of side information into the clustering process is also a well known technique (Xing~et~al.~2003, Aggrawal~et~al.~2012, Liu~et~al.~2015).
  355.  
  356.  
  357.  
  358.  
  359. \section{\color{blue}Redundancy Allocation: The Base Problem}{\color{blue}
  360. In this section, we formally introduce the problem of redundancy allocation. We consider a system that consists of multiple subsystems indexed by $i \in \mathbf{N}$ that are connected in series. If any of the subsystems fails, the whole system fails. Furthermore, the subsystem $i$ consists of multiple types of components, indexed by $j \in \mathbf{M}_i$. In particular, every subsystem adopts a mixed redundancy strategy. That is, components of each type $j$ can be either in cold-standby fashion or active-parallel fashion, which makes up the two subsets of $\mathbf{M}_i=\mathbf{M}^{\rm c}_i \cup \mathbf{M}^{\rm a}_i$, where $\mathbf{M}^{\rm c}_i$ and $\mathbf{M}^{\rm a}_i$ indicate the index sets for cold-standbys and active-parallels, respectively. {\color{red} The cold-standby components begin to work (and thus begin its lifetime) only when its predecessor fails, while the active-parallel components begin their lifetime together.} Finally, each component of type $j$ in subsystem $i$ can have multiple redundant components, indexed by $t \in \mathbf{T}_{ij}$. The configuration of the series system with mixed redundancy strategy is illustrated in Figure~\ref{fig:mixed-redundancy}.
  361.  
  362.  
  363.  
  364. \begin{figure}[htp]
  365. \centering
  366. \includegraphics[scale=0.65]{AC-configuration0819.pdf}
  367. \caption{\color{red} A series-parallel system of mixed redundancy strategy}\label{fig:mixed-redundancy}
  368. \label{figure1}
  369. \end{figure}
  370.  
  371.  
  372. We denote by $\tilde{z}_{ijt}$ the lifetime random variable of the $t$-th redundant component of type $j$ in subsystem $i$, and use binary variables $x_{ijt}$ to denote that whether this component is used ($x_{ijt}=1$) or not ($x_{ijt}=0$), for $i \in \mathbf{N}, j \in \mathbf{M}_i, t \in \mathbf{T}_{ij}$. Thus, the system lifetime can be expressed as follows:
  373. $$
  374. \min\limits_{i \in \mathbf{N}}\left(\sum_{j \in \mathbf{M}^{\rm c}_i}\sum_{t\in \mathbf{T}_{ij}} \tilde{z}_{ijt}x_{ijt} + \max_{j \in \mathbf{M}^{\rm a}_i}  \max_{t \in \mathbf{T}_{ij}}\tilde{z}_{ijt}x_{ijt}\right).
  375. $$
  376. Now, we assume that the lifetime distribution
  377. $$\bm{\tilde{z}}=\left(\tilde{z}_{ijt}\right)_{i \in \mathbf{N}, j \in \mathbf{M}_i, t \in \mathbf{T}_{ij}} \sim \P$$ is known, and the redundancy allocation problem can be formulated as the following chance-constrained optimization problem:
  378. \begin{equation}\label{HP1-ambiguity-XY}
  379. \begin{array}{rcll}
  380. & \min\limits_{\x} &  \sum\limits_{i\in \mathbf{N}} \sum\limits_{j \in \mathbf{M}_i}\left[ \sum\limits_{t\in \mathbf{T}_{ij}} x_{ijt}\right]c_{ij} \\[0.3 cm]
  381. &{\rm s.t.} & \displaystyle \P\left[\min\limits_{i \in \mathbf{N}}\left(\sum_{j \in \mathbf{M}^{\rm c}_i}\sum_{t\in \mathbf{T}_{ij}} \tilde{z}_{ijt}x_{ijt} + \max_{j \in \mathbf{M}^{\rm a}_i}  \max_{t \in \mathbf{T}_{ij}}\tilde{z}_{ijt}x_{ijt}\right)> \mathcal{T}_R \right]\ge R_{0} & \\[0.3 cm]
  382. && L_{i} \le \sum\limits_{j\in \mathbf{M}_i}\sum\limits_{t\in \mathbf{T}_{ij}} x_{ijt}\le  U_{i}, \forall i \in \mathbf{N},\\
  383. && \x \in \{0,1\}^{H},
  384. \end{array}
  385. \end{equation}
  386. where $c_{ij}$ is the cost of each type $j$ redundant component in subsystem $i$, $L_i$ and $U_i$ are lower and upper limit of number of redundant component in each subsystem $i$, $\mathcal{T}_R$ is the required lifetime lower bound, and $R_0$ is the designed reliability level required to reach $\mathcal{T}_R$, and \begin{equation}
  387. H := \sum\limits_{i \in \mathbf{N}}\sum\limits_{j \in \mathbf{M}_i}|\mathbf{T}_{ij}|
  388. \end{equation}
  389. is the total number of possible components in the system.
  390. It is noted in the base model that we assume the distribution $\mathbb{P}$ is exactly known for the lifetimes $\bm{\tilde{z}}$ of all components. However, as mentioned in the Introduction, it is often hard in practice to calibrate such a high-dimensional lifetime distribution due to the scarce of the sufficient lifetime data in many practical reliability applications of redundancy systems. Furthermore, although the lifetime (or time-to-failure) observations might not be sufficient, we do have many observations (even realtime observations) on the working conditions related to component lifetimes. Therefore how to incorporate the possible {\it Side Information} to assist learning the lifetime patterns and achieve a more reliability and economic redundancy allocation is critical, especially in the insufficiency of lifetime data.  To this end, we propose in this work a {\it Distributionally Robust Optimization} model for the redundancy allocation that allows the lifetime distribution $\mathbb{P}$ to be ambiguous and is able to incorporate the possible valuable side information with learning. This will be discussed in detail in the forthcoming two sections.
  391. }
  392. \section{\color{blue}A Learning-based Distributionally Robust Redundancy Model}
  393. {\color{blue}In this section, we focus on the modeling of redundancy optimization problem with side information incorporated using the distributionally robust optimization technique. In particular, we first introduce the construction of the {\it Ambiguity Set}  with side information knowledge established from the machine learning methods, {\it e.g.}, clustering (Section~\ref{subsec:ambiguityset}). We then develop the resulting distributionally robust chance-constrained model for the redundancy allocation and discuss its tractable reformulation (Section~\ref{subsec:DROmodel}). Finally, we design a {\it Cross Validation} method for choosing the best learning parameter~(Section~\ref{subsec:CV}).
  394. }
  395. \subsection{Constructing ambiguity set with clustering using side information}\label{subsec:ambiguityset}
  396. {\color{blue}
  397. As mentioned in the Introduction that in many practical situations, the components of a given type (or even of different types) may fail due to several common causes which can usually reflected by the side information like  producer information, temperature, working intensity, maintenance, to name a few. In other words, these components, once with the side information, can be readily grouped or clustered together, and the resulting clusters or groups should provide valuable information for our redundancy allocation decision. This motivates us to use the {\it Ambiguity Set} with the conditional indicator variables ({\color{red}Chen~et~al.~2019}) to the distributional knowledge incorporated with the side information under the ambiguity.}
  398. {\color{blue}
  399. In particular, we assume that the true distribution $\P$ of the lifetime random variables $\bm{\tilde{z}}$ is not exactly known or {\em ambiguous}, and only partial information of the lifetime distribution is available. Furthermore, we also assume that some side information related to component lifetimes is also available and we let an indicator random variable $\tilde{k}$ to model the possible clustering knowledge learned from both side information and lifetime observations (which will be explained in detail later). Formally, we define the following distributional ambiguity set $\mathbb{F}$ by incorporating the above mentioned distributional information conditional on the indicator variable $\tilde{k}$, within which the distribution $\P$ is allowed to vary:}
  400. \begin{equation}\label{ambiguity-set}
  401. \mathbb{F}_{K}:=\left\{\P \in \mathcal{P}\left( \mathbb{R}_+^H\times[K]\right) \left |
  402. \begin{array}{ll}
  403. (\tilde{\bm{z} }, \tilde{k })\sim \P \\[0.3 cm]
  404. \mathbb{E}_{\P}\Big[\tilde{z}_{ijt}~\Big\vert~\tilde{k}=k\Big]\in \left[\underline{\mu}^{k}_{ij}, \overline{\mu}^{k}_{ij} \right], & \forall k \in [K], i \in \mathbf{N}, j\in \mathbf{M}_i, t \in \mathbf{T}_{ij}\\ [0.3 cm]
  405. \mathbb{E}_{\P}\Big[ g_{lk}(\bm{\tilde{z}})  ~\Big |~ \tilde{k}=k \Big]\le 0, & \forall k \in [K],  l \in \mathbf{L}_k\\[0.5 cm]
  406. \P\Big[\bm{\tilde{z}}\in \mathcal{Z}_k ~\Big |~ \tilde{k}=k \Big]=1, & \forall k \in [K]\\[0.3 cm]
  407. \P\Big[\tilde{k}=k\Big]=p_k, & \forall k \in [K]
  408. \end{array}\right. \right\}.
  409. \end{equation}
  410. {\color{blue}In the above ambiguity set, $K$ is the number of clusters of lifetime pattern learned (which will be explained later and the selection of the hyperparameter $K$ will discussed in Section~\ref{subsec:CV}), the second set of constraints capture the expected lifetime range $[\underline{\mu}^{k}_{ij}, \overline{\mu}^{k}_{ij}]$ of each type $j$ of components in subsystem $i$ in cluster $k$; The forth set of constraints captures the information of conditional support set $\mathcal{Z}_k$ given each cluster $k$, where unless specification we let
  411. $\mathcal{Z}_k=\mathcal{Z}, \forall k \in [K]$ and
  412. \begin{equation}\label{equ:Z}
  413. \mathcal{Z}:= \Big\{\z \in \mathbb{R}^{H}| z_{ijt} \in [\underline{z}_{ij},\overline{z}_{ij}], \forall i \in \mathbf{N}, j \in \mathbf{M}_i, t\in \mathbf{T}_{ij} \Big\}.
  414. \end{equation}
  415. The last set of constraints represent the probability $p_k$ of $\bm{z}$ falling within each cluster, which can be estimated upon the data points within each cluster. Finally, the third set of constraints
  416. \begin{equation}\label{cons:g}
  417. \mathbb{E}_{\P}\Big[ g_{lk}(\bm{\tilde{z}})~\Big |~ \tilde{k}=k \Big]\le 0, \quad \forall k \in [K],  l \in \mathbf{L}_k,
  418. \end{equation}
  419. where each $g_{lk}(\bm{{z}})$ is a convex function of $\bm{{z}}$, are utilized to incorporate, flexibly, more distributional information ({\it e.g.}, correlation) per necessary, for which we illustrate by the following examples.
  420. \begin{example}[Marginal Variance]\label{ex:ambiguityset-2}
  421. If we define  in \eqref{cons:g}
  422. \begin{equation}\label{equ:g-1}
  423. g_{ijtk}({\bm{z}}):=\left(z_{ijt}-\nu^{k}_{ij} \right)^2 - S^{k}_{ij}, \forall k \in [K], i \in \mathbf{N}, j \in {\mathbf{M}_{i}},
  424. \end{equation}
  425. where $\nu^{k}_{ij}, S^{k}_{ij}$ are sample means and standard deviation bounds estimated from the data of each cluster $k$ for each type $j$ in subsystem $i$. This leads to the conditional marginal variance constraints:
  426. $$
  427. \mathbb{E}_{\P}\left[\left. \left(\tilde{z}_{ijt}-\nu^{k}_{ij} \right)^2~\right|~ \tilde{k}=k \right]\le S^{k}_{ij}, \forall k \in [K], i \in \mathbf{N}, j\in \mathbf{M}_i.
  428. $$
  429. \end{example}
  430. \begin{example}[Conditional Average Dispersion]\label{ex:ambiguityset-1}
  431. If we define
  432. \begin{equation}\label{equ:g-2}
  433. g_{ijk}({\bm{z}}):=\sum\limits_{t \in \mathbf{T}_{ij}}\left|\frac{ z_{ijt}-\nu^{k}_{ij}}{|\mathbf{T}_{ij} |\sigma^{k}_{ij}} \right| - \epsilon^{k}_{ij}, \forall k \in [K], i \in \mathbf{N}, j \in {\mathbf{M}_{i}},
  434. \end{equation}
  435. where $\nu^{k}_{ij}, \sigma^{k}_{ij}$ and $\epsilon^{k}_{ij}$ are the sample mean values, standard deviations and dispersion upper bound levels estimated from the data of each cluster $k$. This leads to the conditional average dispersion constraints:
  436. $$
  437. \mathbb{E}_{\P}\left[\left. \frac{1}{|\mathbf{T}_{ij} |}\sum\limits_{t \in \mathbf{T}_{ij}}\left|\frac{ \tilde{z}_{ijt}-\nu^{k}_{ij}}{\sigma^{k}_{ij}} \right|  ~\right|~ \tilde{k}=k \right]\le \epsilon^{k}_{ij}, \forall k \in [K], i \in \mathbf{N}, j\in \mathbf{M}_i,\\
  438. $$
  439. which captures the modeling requirement that within each cluster $k$, the lifetimes of the same type of components in the same subsystem should be distributionally similar.  We also point out that when $K=1$ ({\it i.e.}, only 1 cluster considered), the model of \eqref{equ:g-2} with $g_{ijk}({\bm{z}})$ reduces to the regular average dispersion constraints employed in Wang~et~al.~(2019).
  440. \end{example}
  441. \begin{example}[Conditional Cross Variance]\label{ex:ambiguityset-1}
  442. If we define
  443. \begin{equation}\label{equ:g-3}
  444. g_{ik}({\bm{z}}):=\left(\sum\limits_{j \in {\mathbf{M}_{i}}}\sum\limits_{t \in \mathbf{T}_{ij}}\frac{z_{ijt}}{|\mathbf{T}_{ij}|}-\nu^{k}_{i} \right)^2 - S^k_{i}, \forall k \in [K], i \in \mathbf{N},
  445. \end{equation}
  446. where $\nu^{k}_{i}$ and $S^{k}_{i}$ are the estimated sample mean values and upper bound levels of variance of sum of component lifetimes in all types of each subsystem $i$ given cluster $k$. This leads to the conditional cross variance constraints:
  447. $$
  448. \mathbb{E}_{\P}\left[\left. \left(\sum\limits_{j \in {\mathbf{M}_{i}}}\sum\limits_{t \in \mathbf{T}_{ij}}\frac{z_{ijt}}{|\mathbf{T}_{ij}|}-\nu^{k}_{i} \right)^2  ~\right|~ \tilde{k}=k \right]\le S^k_{i}, \forall k \in [K], i \in \mathbf{N},
  449. $$
  450. which captures the possible correlations among the lifetimes of different types of components.
  451. \end{example}
  452. Finally, we elaborate on how the ambiguity set model $\mathbb{F}_K$ can incorporate, in a flexible fashion, the information of lifetime patterns extracted using machine learning approaches of clustering (Hastie~et~al.~2009). }
  453. {\color{blue}
  454. {\it Direct clustering on Lifetime Sample.} We first point out that even without side information, it is possible to identify more valuable component lifetime patterns by clustering directly over the lifetime sample and we lose nothing by just set $K=1$. For instance, we can use the {\it K-means}, which proves effective and is one of the most popular and unsupervised learning clustering algorithm. A modeling and computational advantage of using K-means for our DRO redundancy allocation model is that the resulting partition of the support $\mathcal{Z}$, which is also called {\it Voronoi Tesselation} (Hastie~et~al.~2009), forms naturally different polyhedral conditional support subsets:
  455. \begin{equation}\label{k-means-classifier}
  456. \mathcal{Z}_k := \Big\{\bm{z} \in \mathcal{Z}\:\big|\: 2(\hmu_i - \hmu_k)^{\top}\bm{z} \leq \hmu_{i}^{\top}\hmu_{i} - \hmu_{k}^{\top}\hmu_{k}, \forall i \in [K] \Big\}, \forall k \in [K],
  457. \end{equation}
  458. where $\hmu_k$ is the mean of cluster $\mathcal{Z}_k$ which is also the output of the K-means algorithm. This conditional support subsets $\mathcal{Z}_k$'s in \eqref{k-means-classifier} not only partitions the original support (therefore provides more effective conditional support sets), but also enjoy computationally viable geometry of polyhedra.  More details on K-means or clustering implementation issues, {\it e.g.}, using some tie-breaking rules to ensure each data point to belongs to only one cluster, can be found in {\color{red}the work of MacQueen (1967)}.
  459. }
  460. {\color{red}
  461. {\it Clustering with Dimension Reduction.} In some situations, it may be beneficial to incorporate dimension reduction into the clustering process. In complex systems with very large number of components, dimension reduction may speed up the clustering process. More importantly, if the lifetime ranges of different components vary to a great extent or if the lifetime of components correlate with each other, Euclidean distance on the raw data may not be a suitable distance metric to measure the similarity of data points. Dimension reduction is likely to reveal the underlying structure of the data, and provide a better distance metric, thus resulting in better clusters. There are many well-developed dimension reduction algorithms to be employed. Principle Component Analysis (PCA) is a popular linear dimension reduction algorithm that can execute very fast. t-Distributed Stochastic Neighbor Embedding (t-SNE) is a slower algorithm, but is particularly well-suited to reduce very high dimensional data into low dimensions like 2 or 3. Spectral Clustering is a technique that can combine a standard clustering method (like K-means) and dimension reduction seamlessly together (Ng et al. 2002). }
  462. {\color{blue}
  463. {\it Clustering based on Side Information.} As we mentioned in the Introduction, in many situations, the data collected on system redundancy design contains not only the samples of component lifetimes, but also the {\it covariates} of side information such as producer information, working conditions, time to last maintenance. These side information should be help for characterizing the lifetime patterns of the components, especially in the scarce of the historical lifetime data, see the following Example~\ref{ex:sideinformation}.
  464. \begin{example}[value of side information]\label{ex:sideinformation}
  465. \color{red} Consider an example of a simple railway system consists of two lines, North Line (case 0) and South Line (case 1). The climate alongside the lines is vastly different, which affects the expected lifetime of the braking system on the train. Consider the braking system as a single component, it has expected lifetime of 1 year operating on the North Line and 3 years on the South Line ( $\mathbb{E}_{\P}\Big[\tilde{z}~\Big\vert~\tilde{k}=0\Big] = 1, \mathbb{E}_{\P}\Big[\tilde{z}~\Big\vert~\tilde{k}=1\Big] = 3$). The two lines have same amount of traffic ($p_0 = p_1 = 0.5$). Without side information, we are able to construct the following ambiguity set, using only the mean information:
  466. \begin{equation}\label{ambiguity-set-foo}
  467. \mathbb{F}:=\left\{\P \in \mathcal{P}\left( \mathbb{R}_+\right) \left|
  468. \begin{array}{ll}
  469. \tilde{z}\sim \P \\[0.3 cm]
  470. \mathbb{E}_{\P}\Big[\tilde{z}]\in \left[1.6, 2.4\right]\\ [0.3 cm]
  471. \P\Big[{\tilde{z}}\in \mathcal{Z}\Big]=1
  472. \end{array}\right. \right\}.
  473. \end{equation}
  474. While using side information, the conditional ambiguity set is:
  475. \begin{equation}\label{ambiguity-set-bar}
  476. \mathbb{F}_{K}:=\left\{\P \in \mathcal{P}\left( \mathbb{R}_+\times[K]\right) \left |
  477. \begin{array}{ll}
  478. (\tilde{z}, \tilde{k })\sim \P \\[0.3 cm]
  479. \mathbb{E}_{\P}\Big[\tilde{z}~\Big\vert~\tilde{k}=0\Big]\in \left[0.8, 1.2\right] \\ [0.3 cm]
  480. \mathbb{E}_{\P}\Big[\tilde{z}~\Big\vert~\tilde{k}=1\Big]\in \left[2.4, 3.6\right] \\ [0.3 cm]
  481. \P\Big[\tilde{z}\in \mathcal{Z}_0 ~\Big |~ \tilde{k}=0 \Big]=1\\[0.3 cm]
  482. \P\Big[\tilde{z}\in \mathcal{Z}_1 ~\Big |~ \tilde{k}=1 \Big]=1\\[0.3 cm]
  483. \P\Big[\tilde{k}=0\Big]=0.5\\[0.3cm]
  484. \P\Big[\tilde{k}=1\Big]=0.5\\[0.3cm]
  485. \end{array}\right. \right\}.
  486. \end{equation}
  487. Note that the lower bound and upper bound of mean is constructed by shrinking or enlarging the mean by 20\%.
  488. It is clear that the conditional ambiguity set describes the data set better. While the lifetime distributions under different scenarios may not be so distinctive from each other in practice, with the introduction of side information we can incorporate much more information into the ambiguity set, and better describe the structure of the data set.
  489. \end{example}
  490. Therefore, we can extract the lifetime pattern information by clustering collectively both the lifetime data and the side information, using suitable machine learning approaches, such as K-means clustering, hierachical clustering ({\color{red}Sibson 1973}) and regression trees ({\color{red}Quinlan 1986}), per convenience. We can therefore form a number of clusters or scenarios $k\in [K]$ of the component lifetime patterns and achieve a more accurate description for the component lifetimes using the statistical information within each cluster (scenario).
  491. }
  492. {\color{blue} In the next section, we discuss the distributionally robust redundancy allocation model with ambiguity set $\mathbb{F}_K$ given the number $K$ of clusters, and derive its tractable reformulations.}
  493. \subsection{The model with tractable reformulation}\label{subsec:DROmodel}{\color{blue}
  494. With the ambiguity set $\mathbb{F}_K$ of lifetime distributions of all components, we now consider the worst-case probabilistic constraint for safeguarding the reliability level $R_0$, {\it i.e.},
  495. $$
  496. \inf\limits_{\P \in \mathbb{F}_{K}}\P\left[\min\limits_{i \in \mathbf{N}}\left(\sum_{j \in \mathbf{M}^{\rm c}_i}\sum_{t\in \mathbf{T}_{ij}} \tilde{z}_{ijt}x_{ijt} + \max_{j \in \mathbf{M}^{\rm a}_i}  \max_{t \in \mathbf{T}_{ij}}\tilde{z}_{ijt}x_{ijt}\right)> \mathcal{T}_S \right]\ge R_{0}
  497. $$
  498. in the original redundancy allocation problem~\eqref{HP1-ambiguity-XY}. This leads to the following {\em distributionally robust redundancy optimization (DRRO) model}:
  499. \begin{equation}\label{HP1-ambiguity-X}
  500. \begin{array}{rcll}
  501. & \min\limits_{\x} &  \sum\limits_{i\in \mathbf{N}} \sum\limits_{j \in \mathbf{M}_i}\left[ \sum\limits_{t\in \mathbf{T}_{ij}} x_{ijt}\right]c_{ij} \\[0.3 cm]
  502. &{\rm s.t.} & \displaystyle \inf\limits_{\P \in \mathbb{F}_{K}}\P\left[\min\limits_{i \in \mathbf{N}}\left(\sum_{j \in \mathbf{M}^{\rm c}_i}\sum_{t\in \mathbf{T}_{ij}} \tilde{z}_{ijt}x_{ijt} + \max_{j \in \mathbf{M}^{\rm a}_i}  \max_{t \in \mathbf{T}_{ij}}\tilde{z}_{ijt}x_{ijt}\right)> \mathcal{T}_S \right]\ge R_{0} & \\[0.3 cm]
  503. && L_{i} \le \sum\limits_{j\in \mathbf{M}_i}\sum\limits_{t\in \mathbf{T}_{ij}} x_{ijt}\le  U_{i}, \forall i \in \mathbf{N},\\
  504. && \x \in \{0,1\}^H.
  505. \end{array}
  506. \end{equation}
  507. In the above model, the distributionally robust chance constraint essentially ensures to achieve the required reliability level $R_{0}$ over all the qualified probability distributions $\P \in \mathbb{F}_K$.
  508. {\color{red}It can be shown that the conditional ambiguity set ($\mathbb{F}_{K}$) is tighter than the ambiguity set constructed without clustering (which can also be considered as the case when $K = 1$, denoted by $\mathbb{F}_{1}$). Therefore, by incorporating cluster information, we can achieve better result in solving the problem. We present a proof when using marginal variance in example 1. In the following proposition, parameters with subscript or superscript $k$ are parameters in $\mathbb{F}_K$, while the parameters without them are parameters in $\mathbb{F}_1$. $P_K$ is the program \eqref{HP1-ambiguity-X} where $\P \in \mathbb{F}_{K}$, with optimal solution $c^*_K$, and $P_1$ is the program \eqref{HP1-ambiguity-X} where $\P \in \mathbb{F}_{1}$, with optimal solution $c^*_1$.
  509. \begin{proposition}
  510. Given ambiguity set $\mathbb{F}_{K}$ and $\mathbb{F}_{1}$, assuming that $\underline{\hmu} = \sum\limits_{k \in [K]}p_k\underline{\hmu}_{k}$, $\overline{\hmu} = \sum\limits_{k \in [K]}p_k\overline{\hmu}_{k}$, $\hmu_k = \frac{\underline{\hmu}_{k} + \overline{\hmu}_{k}}{2}$, $S_{ij}^2 + \mu_{ij}^2 = \sum\limits_{k \in [K]}p_k({S_{ij}^k}^2 + {\mu_{ij}^k}^2), \forall i \in \mathbf{N}, j \in \mathbf{M}_i$, then $\mathbb{F}_{K} \subseteq \mathbb{F}_{1}$, and the optimal solution $c^*_K \leq c^*_1$.
  511. \end{proposition}
  512. {\bf Proof.} First, we prove that $\forall \P \in \mathbb{F}_{K}, \P \in \mathbb{F}_{1}$ too.
  513. %\begin{eqnarray}
  514. $$
  515. \begin{aligned}
  516. \mathbb{E}_{\P}\Big(\tilde{z}_{ijt}\Big) \geq \sum\limits_{k \in [K]}p_k\underline{\hmu}_{ij}^{k} &= \underline{\mu}_{ij} & \forall i \in \mathbf{N}, j \in \mathbf{M}_i, t \in \mathbf{T}_{ij}\\
  517. \mathbb{E}_{\P}\Big(\tilde{z}_{ijt}\Big) \leq \sum\limits_{k \in [K]}p_k\overline{\hmu}_{ij}^{k} &= \overline{\mu}_{ij} & \forall i \in \mathbf{N}, j \in \mathbf{M}_i, t \in \mathbf{T}_{ij}\\
  518. \mathbb{E}_{\P}\Big((\tilde{z}_{ijt} - \mu_{ij})^2 \Big) &= \sum\limits_{k \in [K]}p_k\mathbb{E}_{\P}\Big((\tilde{z}_{ijt} - \mu_{ij}^k)^2 | \tilde{k} = k) & \forall i \in \mathbf{N}, j \in \mathbf{M}_i, t \in \mathbf{T}_{ij}\\\\
  519. &= \sum\limits_{k \in [K]}p_k\mathbb{E}_{\P}\Big((\tilde{z}_{ijt})^2 | \tilde{k} = k\Big) - \mu_{ij}^2&\\
  520. &\leq \sum\limits_{k \in [K]}p_k\mathbb{E}_{\P}\Big(S_{ij}^2 + {\mu_{ij}^k}^2 | \tilde{k} = k \Big) - \mu_{ij}^2&\\
  521. &= S_{ij}^2 + {\mu_{ij}}^2 - \mu_{ij}^2&\\
  522. &= S_{ij}^2&
  523. \end{aligned}
  524. $$
  525. %\end{eqnarray}
  526.  
  527. Moreover, since $\mathcal{Z}_k \subseteq \mathcal{Z}$ by definition, $\bigcup\limits_{k \in [K]}\mathcal{Z}_k \subseteq \mathcal{Z}$. Therefore $\P[\tilde{z} \in \mathcal{Z}] = 1$. Hence, $\P \in \mathbb{F}_{1}$.
  528.  
  529. Since $\forall \P \in \mathbb{F}_{K}, \P \in \mathbb{F}_{1}$, $\mathbb{F}_{K} \subseteq \mathbb{F}_{1}$. Therefore any feasible solution $c_1$ to $P_1$ is also a feasible solution to $P_K$. Since $c^*_K$ and $c^*_1$ are optimal solution to a minimization problem, $c^*_K \leq c^*_1$. \blot
  530.  
  531. The conditions in the proposition are used to require that the the momemnt specifications should be consistent.
  532. }
  533.  
  534.  
  535.  
  536. Next, we derive a tractable formulation of the DRRO model \eqref{HP1-ambiguity-X}. Without loss of generality, we assume that $\mathbf{M}^{a}_{i} \neq \emptyset, \forall i \in \mathbf{N}$. The case of $\mathbf{M}^{a}_{i} = \emptyset, \exists i \in \mathbf{N}$, by our forthcoming discussions, can be treated similarly and is actually technically easier.  Also we illustrate our major results using
  537. $$
  538. g_{ijk}({\bm{z}}):=\sum\limits_{t \in \mathbf{T}_{ij}}\left|\frac{ z_{ijt}-\nu^{k}_{ij}}{|\mathbf{T}_{ij} |\sigma^{k}_{ij}} \right| - \epsilon^{k}_{ij}, \forall k \in [K], i \in \mathbf{N}, j \in {\mathbf{M}_{i}}
  539. $$
  540. in the ambiguity set $\mathbb{F}_K$, and consider the conditional support set with K-means
  541. \begin{equation}\label{equ:W}
  542. \mathcal{Z}_k = \Big\{\bm{z} \in \mathcal{Z}\:\big|\: 2(\hmu_i - \hmu_k)^{\top}\bm{z} \leq \hmu_{i}^{\top}\hmu_{i} - \hmu_{k}^{\top}\hmu_{k}, \forall i \in [K] \Big\}, \forall k \in [K]
  543. \end{equation}
  544. where $\mathcal{Z}$ is the original support set given by \eqref{equ:Z}.
  545. We begin by focusing on the computation of the following worst-case reliability function in the problem (\ref{HP1-ambiguity-X}) given system design $\x$:
  546. \begin{equation}\label{Prob-1}
  547. \displaystyle \inf\limits_{\P \in \mathbb{F}_K}\P\left[\min\limits_{i \in \mathbf{N}}\left(\sum_{j \in \mathbf{M}^{\rm c}_i}\sum_{t\in \mathbf{T}_{ij}} \tilde{z}_{ijt}x_{ijt} + \max_{j \in \mathbf{M}^{\rm a}_i}  \max_{t \in \mathbf{T}_{ij}}\tilde{z}_{ijt}x_{ijt}\right)> \mathcal{T}_R \right].
  548. \end{equation}
  549. For a better exposition of our approach, we denote by
  550. \begin{equation}\label{constraint-set}
  551. \mathcal{Z}_i(\x):=\left\{\bm{z} \in \mathbb{R}^H_+ ~\left|~ \sum_{j \in \mathbf{M}^{\rm c}_i}\sum_{t\in \mathbf{T}_{ij}} \tilde{z}_{ijt}x_{ijt} + \max_{j \in \mathbf{M}^{\rm a}_i}  \max_{t \in \mathbf{T}_{ij}}\tilde{z}_{ijt}x_{ijt} > \mathcal{T}_R \right.\right \}, \forall i \in \mathbf{N}.
  552. \end{equation}
  553. Since $\mathbf{M}^{a}_{i} \neq \emptyset, \forall i \in \mathbf{N}$, by the nature of maximal value, it is clear that
  554. \begin{equation}\label{constraint-set-2}
  555. \mathcal{Z}_i(\x) = \bigcup\limits_{{j} \in \mathbf{M}^{\rm a}_i}\bigcup\limits_{t \in \mathbf{T}_{i{j}}}\left\{\bm{z} \in \mathbb{R}^H_+ ~\left|~ \sum_{{j} \in \mathbf{M}^{\rm c}_i}\sum_{t\in \mathbf{T}_{i{j}}} \tilde{z}_{i{j}{t}}x_{i{j}{t}} + \tilde{z}_{i{j}{t}}x_{i{j}{t}} > \mathcal{T}_R \right.\right \}, \forall i \in \mathbf{N}.
  556. \end{equation}
  557. Then the complimentary set of $\mathcal{Z}_i$ for each $i \in \mathbf{N}$, denoted by $\overline{\mathcal{Z}}_i$,  is
  558. \begin{equation}\label{set-LT}
  559. \overline{\mathcal{Z}}_i(\x)=\bigcap\limits_{{j} \in \mathbf{M}^{\rm a}_i}\bigcap\limits_{t \in \mathbf{T}_{i{j}}}\left\{\bm{z} \in \mathbb{R}^H_+ ~\left|~ \sum_{{j} \in \mathbf{M}^{\rm c}_i}\sum_{t\in \mathbf{T}_{i{j}}} \tilde{z}_{i{j}{t}}x_{i{j}{t}} + \tilde{z}_{i{j}{t}}x_{i{j}{t}}\leq \mathcal{T}_R \right.\right \}, \forall i \in \mathbf{N}.
  560. \end{equation}
  561. The following result establishes an equivalent formulation of regular robust optimization for the above worst-case reliability function~\eqref{Prob-1}.
  562. \begin{lemma}\label{lem1-LT}
  563. Given system design $\x$, the worst-case reliability function (\ref{Prob-1}) is equivalent to the optimal value of the following  optimization problem:
  564. \begin{equation}\label{P1-ambiguity-sup-dual1}
  565. \begin{array}{rcll}
  566. &\!\!\!\!\!\!\!\!\!\!\!\! \max\limits_{\bm{\alpha}, \bm{\beta}, \bm{\lambda}, \bm{\tau}} &  1-\sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}}\sum\limits_{k\in[K]} \left(\alpha^{k}_{ijt}\underline{\mu}^{k}_{ij} + \beta^{k}_{ijt}\overline{\mu}^{k}_{ij}\right)-\sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i}\sum\limits_{k\in[K]}\epsilon^{k}_{ij}\lambda^{k}_{ij} - \sum\limits_{k\in[K]}\tau_{k} \\[0.3 cm]
  567. &\!\!\!\!\!\!\!\!\!\!\!\!{\rm s.t.} & \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}} \left[z_{ijt}\left(\alpha^{k}_{ijt}+ \beta^{k}_{ijt} \right) + \left|\frac{ z_{ijt}-\nu^{k}_{ij}}{|\mathbf{T}_{ij} |\sigma^{k}_{ij}} \right|\lambda^{k}_{ij}\right] + \tau_k \ge{p_k}, \forall \bm{z} \in  \mathcal{Z}_k\cap \overline{\mathcal{Z}}_i(\x), i \in \mathbf{N}, k \in [K]\\
  568. && \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}} \left[z_{ijt}\left(\alpha^{k}_{ijt}+ \beta^{k}_{ijt} \right) + \left|\frac{ z_{ij{t}}-\nu^{k}_{ij}}{|\mathbf{T}_{ij} |\sigma^{k}_{ij}} \right|\lambda^{k}_{ij}\right] + \tau_k \ge 0, \forall \bm{z} \in  \mathcal{Z}_k, k\in [K]\\
  569. && \halpha \le \mathbf{0}, \hbeta, \hlambda \ge \mathbf{0}, \bm{\tau} \in \mathbb{R}^K.
  570. \end{array}
  571. \end{equation}
  572.  
  573. \end{lemma}
  574.  
  575. %\begin{proof}
  576. {\bf Proof.}
  577. With the notation in (\ref{constraint-set}), the worst-case probabilistic chance function (\ref{Prob-1})  can be rewritten in terms of the probability of its complementary event:
  578. \begin{equation}\label{1minus}
  579. \inf\limits_{\P\in \mathbb{F}_{K}} \P\Big[\tilde{\bm{z} }\in \mathcal{Z}_i(\x),\forall i \in \mathbf{N} \Big]=1-\sup\limits_{\P \in \mathbb{F}_K } \P\Big[\cup_{i\in \mathbf{N} }\{\bm{\tilde{z}} \in \overline{\mathcal{Z}}_i(\x)\} \Big].
  580. \end{equation}
  581.  
  582. Given the probability distribution of  $\tilde{s}$ as
  583. $$
  584. \P\Big[\tilde{k}=k\Big]=p_k, \forall k \in [K].
  585. $$
  586. We now define $\P_k$ as the conditional distribution of $\bm{\tilde{z}}$ given $\tilde{k}=k$ for $k \in [K]$, we then can decompose any distribution $\P \in \mathbb{F}_K$ using $\{\P_k, k\in [K]\}$ and rewrite the worst-case chance
  587. $$
  588. \sup\limits_{\P \in \mathbb{F}_K } \P\Big[\cup_{i\in \mathbf{N} }\{\bm{\tilde{z}} \in \overline{\mathcal{Z}}_i(\x)\} \Big]
  589. $$
  590. using total probability law as following formulation:
  591. \begin{eqnarray}
  592. \sup\limits_{\P \in \mathbb{F}_K } \P\Big[\cup_{i\in \mathbf{N} }\{\bm{\tilde{z}} \in \overline{\mathcal{Z}}_i(\x)\} \Big]
  593. &=&\sup\limits_{\P_k, \forall k\in[K]}\sum\limits_{k\in[K]}p_k\P_k\Big[\cup_{i\in \mathbf{N} }\{\bm{\tilde{z}} \in \overline{\mathcal{Z}}_i(\x)\} \Big]\label{P1-ambiguity-sup} \\[0.2 cm]
  594. &=&\sup\limits_{\P_k}\sum\limits_{k\in[K]}\displaystyle \int_{\cup_{i\in \mathbf{N} }\left\{\bm{\tilde{z}} \in \overline{\mathcal{Z}}_i(\x)\right\}}p_k {\rm d}\P_k(\bm{\tilde{z}}) \\[0.2 cm]
  595. &&{\rm s.t.}\displaystyle \int_{\bm{\tilde{z}} \in \mathcal{Z}_k} \tilde{z}_{ijt} {\rm d}\P_k(\bm{\tilde{z}}) \ge \underline{\mu}^{k}_{ij},    \forall i \in \mathbf{N}, j\in \mathbf{M}_i, t\in \mathbf{T}_{ij}, k\in [K]\\ [0.2 cm]
  596. &&~\quad \displaystyle \int_{\bm{\tilde{z}} \in \mathcal{Z}_k} \tilde{z}_{ijt} {\rm d}\P_k(\bm{\tilde{z}}) \le \overline{\mu}^{k}_{ij},    \forall i \in \mathbf{N}, j\in \mathbf{M}_i, t\in \mathbf{T}_{ij},k\in [K]\\ [0.2 cm]
  597. &&~\quad  \displaystyle \int_{\bm{\tilde{z}} \in \mathcal{Z}_k} \sum\limits_{t\in \mathbf{T}_{ij}}\left|\frac{\tilde{z}_{ijt}-\nu^{k}_{ij}}{|\mathbf{T}_{ij} |\sigma^{k}_{ij}} \right| {\rm d}\P_k(\bm{\tilde{z}}) \le \epsilon^{k}_{ij},   \forall i \in \mathbf{N}, j\in \mathbf{M}_i,k\in [K]\\[0.2 cm]
  598. &&~\quad \displaystyle \int_{\bm{\tilde{z}} \in \mathcal{Z}_k} {\rm d}\P_k(\bm{\tilde{z}}) =1, \forall k\in[K], \label{P1-ambiguity-sup-2}
  599. \end{eqnarray}
  600. where the support $\mathcal{Z}_k$ is given by (\ref{equ:W}). The Lagrange dual of above moment problem \eqref{P1-ambiguity-sup}-\eqref{P1-ambiguity-sup-2} has the following formulation (Wiesemann~et al.~2014):
  601. \begin{equation}\label{P1-ambiguity-sup-dual0}
  602. \begin{array}{rcl}
  603. &\!\!\!\!\!\!\!\!\!\!\!\! \min\limits_{\bm{\alpha}, \bm{\beta}, \bm{\lambda}, \bm{\tau}} &  \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}}\sum\limits_{k\in[K]} \left(\alpha^{k}_{ijt}\underline{\mu}^{k}_{ij}+ \beta^{k}_{ijt}\overline{\mu}^{k}_{ij}\right)+\sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i}\sum\limits_{k\in[K]}\epsilon^{k}_{ij}\lambda^{k}_{ij} + \sum\limits_{k\in[K]}\tau_{k} \\[0.3 cm]
  604. &\!\!\!\!\!\!\!\!\!\!\!\!{\rm s.t.} & \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}}  \left[z_{ijt}\left(\alpha^{k}_{ijt}+ \beta^{k}_{ijt} \right) + \left|\frac{ z_{ijt}-\nu^{k}_{ij}}{|\mathbf{T}_{ij} |\sigma^{k}_{ij}} \right|\lambda^{k}_{ij}\right] + \tau_k \ge p_k \mathbb{I}\Big({\bigcup\limits_{i\in \mathbf{N} }\left\{ \bm{z} \in \overline{\mathcal{Z}}_i\right\}} \Big), \forall \bm{z} \in \mathcal{Z}_k, k \in [K]\\
  605. && \bm{\alpha} \le \mathbf{0}, \bm{\beta}, \bm{\lambda} \ge \mathbf{0}, \bm{\tau} \in \mathbb{R}^K,
  606. \end{array}
  607. \end{equation}
  608. where $\mathbb{I}(\mathcal{A})$ is the indicator function with respect to set $\mathcal{A}$, and  $(\halpha, \hbeta, \hlambda, \htau)$ are the dual variables associated with the constraints of the primal problem \eqref{P1-ambiguity-sup}-\eqref{P1-ambiguity-sup-2}.
  609. Furthermore, we show the strong duality holds. Since ${\mu^{k}_{ij}}$ is the expectation of $\bm{\tilde{z}}_{ijt}$, we can always find a Dirac probability distribution $\P^{\dag}_{\bm{\mu}}$ with $\underline{\hmu}<\hmu<\overline{\hmu}$ which is relative interior point of the feasible set of problem \eqref{P1-ambiguity-sup}-\eqref{P1-ambiguity-sup-2}. Therefore, the Slater condition holds, and then the optimal value of (\ref{P1-ambiguity-sup-dual0}) is equivalent to that of problem \eqref{P1-ambiguity-sup}-\eqref{P1-ambiguity-sup-2}.
  610. Next, expanding the indication function $\mathbb{I}\left({\cup_{i\in \mathbf{N} }\left\{\bm{\tilde{z}} \in \overline{\mathcal{Z}}_i(\x)\right\}} \right)$ for different cases of $\bm{z}$, the above problem (\ref{P1-ambiguity-sup-dual0}) is also equivalent to the following formulation:
  611. \begin{equation}\label{P1-ambiguity-sup-dual00}
  612. \begin{array}{rcll}
  613. &\!\!\!\!\!\!\!\!\!\!\!\! \min\limits_{\bm{\alpha}, \bm{\beta}, \bm{\lambda}, \bm{\tau}} &  \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}}\sum\limits_{k\in[K]} \left(\alpha^{k}_{ijt}\underline{\mu}^{k}_{ij}+ \beta^{k}_{ijt}\overline{\mu}^{k}_{ij}\right)+\sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i}\sum\limits_{k\in[K]}\epsilon^{k}_{ij}\lambda^{k}_{ij} + \sum\limits_{k\in[K]}\tau_{k} \\[0.3 cm]
  614. &\!\!\!\!\!\!\!\!\!\!\!\!{\rm s.t.} & \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}}  \left[z_{ijt}\left(\alpha^{k}_{ijt}+ \beta^{k}_{ijt} \right) + \left|\frac{ z_{ijt}-\nu^{k}_{ij}}{|\mathbf{T}_{ij} |\sigma^{k}_{ij}} \right|\lambda^{k}_{ij}\right] + \tau_k \ge p_k, \forall \bm{z} \in \mathcal{Z}_k\cap\overline{\mathcal{Z}}_i, i \in \mathbf{N}, k \in [K]\\
  615. && \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}}  \left[z_{ijt}\left(\alpha^{k}_{ijt}+ \beta^{k}_{ijt} \right) + \left|\frac{ z_{ijt}-\nu^{k}_{ij}}{|\mathbf{T}_{ij} |\sigma^{k}_{ij}} \right|\lambda^{k}_{ij}\right] + \tau_k \ge0, \forall \bm{z} \in \mathcal{Z}_k, k \in [K]\\\nonumber
  616. && \halpha \le \mathbf{0}, \hbeta, \hlambda \ge \mathbf{0}, \bm{\tau} \in \mathbb{R}^K.
  617. \end{array}
  618. \end{equation}
  619. Finally, plugging this formulation  into the equation (\ref{1minus}), we arrive at the the formulation of (\ref{P1-ambiguity-sup-dual1}) whose optimal objective value is exactly the worst-case value of probabilistic chance function (\ref{Prob-1}).  The proof is completed. \blot
  620. %\end{proof}
  621.  
  622. It is noted that the derived optimization problem (\ref{P1-ambiguity-sup-dual1}) in current version still belongs to semi-infinitely dimensional optimization problems which are not directly computable.  In the following, we show that by duality argument the problem can be further transformed into a computationally tractable formulation of linear program.
  623.  
  624. \begin{proposition}\label{P-proposition1b}
  625. Given a system design $\x$, The worst-case reliability function (\ref{Prob-1}) solves the following linear program (LP):
  626. \begin{eqnarray}
  627. &\!\!\!\!\!\! \max &  1-\sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}}\sum\limits_{k\in[K]} \left(\alpha^{k}_{ijt}\underline{\mu}^{k}_{ij} + \beta^{k}_{ijt}\overline{\mu}^{k}_{ij}\right) - \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i}\sum\limits_{k\in[K]}\epsilon^{k}_{ij}\lambda^{k}_{ij} - \sum\limits_{k\in[K]}\tau_{k} \label{HP1-ambiguity-LP-FL} \\
  628. &\!\!\!\!\!\!{\rm s.t.} & \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}} \Big[\phi^{lk}_{ijt}\underline{z}_{ij}+\varphi^{lk}_{ijt}\overline{z}_{ij}+{\nu^{k}_{ij}\left(\pi^{lk}_{ijt}-\varpi^{lk}_{ijt} \right)}  \nonumber\\
  629.  &&+  \sum\limits_{n \in [K]}\psi^{lk}_{n}({(\mu^{n}_{ijt})}^2 - {(\mu^{k}_{ijt})}^2)\Big] + \sum\limits_{j\in \mathbf{M}^{\rm a}_{l}}\sum\limits_{t\in \mathbf{T}_{lj}} q^{lk}_{jt}\mathcal{T}_R+ \tau_k \geq p_k, \forall l \in \mathbf{N}, k \in [K]  \label{HP1-ambiguity-LP-FL1}\\
  630.  && \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}} \Big[\rho^{k}_{ijt}\underline{z}_{ij}+\varrho^{k}_{ijt}\overline{z}_{ij}+{\nu^{k}_{ij}\left(\gamma^{k}_{ijt}-\theta^{k}_{ijt} \right)} \nonumber\\ && + \sum\limits_{n \in [K]}\varsigma^{k}_{n}({(\mu^{n}_{ijt})}^2 - {(\mu^{k}_{ijt})}^2) \Big] + \tau_k \geq0, \forall k \in [K]\label{HP1-ambiguity-LP-FL1-2}\\
  631. &&  {q^{lk}_{jt}}x_{l jt}  +\sum\limits_{n \in [K]}2(\mu^{n}_{ljt} - \mu^{k}_{ljt})\psi^{lk}_{n}+ \phi^{lk}_{ljt}+\varphi^{lk}_{ljt}+\pi^{lk}_{ljt}-\varpi^{lk}_{ljt} \nonumber \\ &&  = \alpha^{k}_{ljt}+\beta^{k}_{ljt}, \forall l \in \mathbf{N}, j\in \mathbf{M}^{\rm a}_l, t \in \mathbf{T}_{l j}, k \in [K] \\[0.3 cm]
  632. &&  x_{l jt}\sum\limits_{j \in \mathbf{M}^{\rm a}_l}\sum\limits_{t \in \mathbf{T}_{l j}}{q^{lk}_{jt}}  +\sum\limits_{n \in [K]}2(\mu^{n}_{ljt} - \mu^{k}_{ljt})\psi^{lk}_{n}+ \phi^{lk}_{ljt}+\varphi^{lk}_{ljt}+\pi^{lk}_{ljt}-\varpi^{lk}_{ljt} \nonumber \\ &&  = \alpha^{k}_{ljt}+\beta^{k}_{ljt}, \forall l \in \mathbf{N}, j\in \mathbf{M}^{\rm c}_l, t \in \mathbf{T}_{l j}, k \in [K] \\[0.3 cm]
  633. &&   \sum\limits_{n \in [K]}2(\mu^{n}_{ijt} - \mu^{k}_{ijt})\psi^{lk}_{n}+  \phi^{lk}_{ijt}+\varphi^{lk}_{ijt}+\pi^{lk}_{ijt}-\varpi^{lk}_{ijt} = \alpha^{k}_{ijt}+\beta^{k}_{ijt},\nonumber\\ && ~ \forall l \in \mathbf{N}, i \in \mathbf{N}\setminus\{l \}, j \in \mathbf{M}_i, t\in \mathbf{T}_{ij}, k \in [K] \label{HP2-ambiguity-LP-FL2} \\
  634. &&{|\mathbf{T}_{ij} |\sigma^{k}_{ij}}(\pi^{lk}_{ijt}+\varpi^{lk}_{ijt}) =\lambda^{k}_{ij},  ~ \forall l  \in \mathbf{N}, i \in \mathbf{N}, j\in \mathbf{M}_i, t\in \mathbf{T}_{ij}, k \in [K] \\
  635. && \sum\limits_{n \in [K]}2(\mu^{n}_{ijt} - \mu^{k}_{ijt})\varsigma^{k}_{n}+ \rho^{k}_{ijt}+\varrho^{k}_{ijt}+\gamma^{k}_{ijt}-\theta^{k}_{ijt} = \alpha^{k}_{ijt}+\beta^{k}_{ijt},\nonumber\\ && ~ \forall i \in \mathbf{N}, j \in \mathbf{M}_i, t\in \mathbf{T}_{ij}, k \in [K]  \\
  636. &&  {|\mathbf{T}_{ij} |\sigma^{k}_{ij}}(\gamma^{k}_{ijt}+\theta^{k}_{ijt}) =\lambda^{k}_{ij},  ~ \forall i \in \mathbf{N}, j\in \mathbf{M}_i, t\in \mathbf{T}_{ij}, k \in [K] \\
  637. %&&  q_{l jk}\le y_{{l jk}}, ~\forall l  \in \mathbf{N}, j \in \mathbf{M}_{l}, k\in \mathcal{N}(l,j) \\[0.3 cm]
  638. %&& y_{{l jk}} \ge M x_{l jk}, ~\forall l  \in \mathbf{N}, j \in \mathbf{M}_{l}, k\in \mathcal{N}(l,j) \\[0.3 cm]
  639. %&& y_{{l jk}} \le  q_{l jk }+(x_{l jk}-1)M, ~\forall l  \in \mathbf{N}, j \in \mathcal{J}({l}), k \in \mathcal{N}(l,j)\\[0.3 cm]
  640. %&&  \sum\limits_{k\in \mathbf{K}_{ij}} x^{k}_{ij}\ge  L_{ij}, ~  \forall  i \in \mathbf{N}, j\in \mathbf{M}_i \\
  641. && \halpha,\q, \hpsi, \hvarphi, \hvarsigma, \hvarrho \le \mathbf{0}, \htau \in \mathbb{R}^K,  \\
  642. && \hbeta, \hlambda, \hphi, \hrho, \hpi,\hvarpi, \hgamma, \htheta \ge \mathbf{0}, \label{HP2-ambiguity-LP-FL}
  643. \end{eqnarray}
  644. where $\halpha, \hbeta, \hlambda, \htau, \q, \hpsi, \hphi, \hvarphi, \hpi, \hvarpi, \hrho, \hvarrho, \hvarsigma, \hgamma, \htheta$ are auxiliary variables.
  645. \end{proposition}
  646.  
  647. %\begin{proof}
  648. {\bf Proof. }
  649. First of all, for a given $l  \in \mathbf{N}$ we deal with the infinitely dimensional constraints
  650. $$
  651. \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}}  \left[z_{ijt}\left(\alpha^{k}_{ijt}+ \beta^{k}_{ijt} \right) + \left|\frac{ z_{ijt}-\nu^{k}_{ij}}{|\mathbf{T}_{ij} |\sigma^{k}_{ij}} \right|\lambda^{k}_{ij}\right] + \tau_k \ge p_k, \forall \bm{z} \in \mathcal{Z}_k\cap\overline{\mathcal{Z}}_l, k \in [K].
  652. $$
  653. Recall that
  654. \begin{equation}
  655. \overline{\mathcal{Z}}_l(\x)=\bigcap\limits_{{j} \in \mathbf{M}^{\rm a}_l}\bigcap\limits_{{t} \in \mathbf{T}_{l{j}}}\left\{\bm{z} \in \mathbb{R}^H_+ ~\left|~ \sum_{{j} \in \mathbf{M}^{\rm c}_l}\sum_{{t}\in \mathbf{T}_{l{j}}} \tilde{z}_{l{j}{t}}x_{l{j}{t}} + \tilde{z}_{l{j}{t}}x_{l{j}{t}}\leq \mathcal{T}_R \right.\right \}
  656. \end{equation}
  657. \begin{equation}
  658. =\left\{\bm{z} \in \mathbb{R}^H_+ ~\left|~ \sum_{j \in \mathbf{M}^{\rm c}_l}\sum_{{t}\in \mathbf{T}_{l{j}}} \tilde{z}_{l{j}{t}}x_{l{j}{t}} + \tilde{z}_{l{j}{t}}x_{l{j}{t}}\leq \mathcal{T}_R, \forall j \in \mathbf{M}^{\rm a}_l, {t}\in \mathbf{T}_{l{j}}  \right.\right \},
  659. \end{equation}
  660. and
  661. $$
  662. \mathcal{Z}_k = \left\{\bm{z} \in \mathbb{R}^{H} \left| \begin{array}{l}
  663.                                                        2(\hmu_i - \hmu_k)^{\top}\bm{z} \leq \hmu_{i}^{\top}\hmu_{i} - \hmu_{k}^{\top}\hmu_{k}, \forall i \in [K]  \\[0.25 cm]
  664.                                                        z_{ijt} \in [\underline{z}_{ij},\overline{z}_{ij}], \forall i \in \mathbf{N}, j \in \mathbf{M}_i, t\in \mathbf{T}_{ij}
  665.                                                       \end{array}
  666. \right.\right\}
  667. $$
  668. for any $k \in [K]$.
  669. First of all, we claim that for any $k \in [K]$
  670. \begin{equation}\label{Lifting-1}
  671. \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}}  \left[z_{ijt}\left(\alpha^{k}_{ijt}+ \beta^{k}_{ijt} \right) + \left|\frac{ z_{ijt}-\nu^{k}_{ij}}{|\mathbf{T}_{ij} |\sigma^{k}_{ij}} \right|\lambda^{k}_{ij}\right] + \tau_k \ge p_k, \forall \bm{z} \in \mathcal{Z}_k\cap\overline{\mathcal{Z}}_l
  672. \end{equation}
  673. is equivalent to
  674. \begin{equation}\label{Lifting-2}
  675. \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}}  \left[z_{ijt}\left(\alpha^{k}_{ijt}+ \beta^{k}_{ijt} \right) + u^{k}_{ijt}\lambda^{k}_{ij}\right] + \tau_k \ge p_k, \forall (\bm{z},\u) \in \mathcal{W}_k,
  676. \end{equation}
  677. {\color{red} where
  678. $$
  679. \mathcal{W}_k := \left\{(\z, \u) \middle| \: \z \in \mathcal{Z}_k\cap\overline{\mathcal{Z}}_l, \left|\frac{ z_{ijt}-\nu^{k}_{ij}}{|\mathbf{T}_{ij} |\sigma^{k}_{ij}} \right| \leq u^{k}_{ijt}, \forall i \in \mathbf{N}, j \in \mathbf{M}_i, t \in \mathbf{T}_{ij} \right\}, \forall k \in [K].
  680. $$
  681. }
  682. In fact, on the one hand, if \eqref{Lifting-1} holds, since $\hlambda \geq \mathbf{0}$, we have for any $(\bm{z}, \u) \in \mathcal{W}_k$
  683. $$
  684. u^{k}_{ijt}\lambda^{k}_{ij} \geq \left|\frac{ z_{ijt}-\nu^{k}_{ij}}{|\mathbf{T}_{ij} |\sigma^{k}_{ij}} \right|\lambda^{k}_{ij}, \forall i \in \mathbf{N}, j \in \mathbf{M}_i, t \in \mathbf{T}_{ij}.
  685. $$
  686. Therefore \eqref{Lifting-2} holds. On the other hand, assume \eqref{Lifting-2} holds, then for any $\hat{\bm{z}} \in \mathcal{Z}_k\cap\overline{\mathcal{Z}}_l$, we take
  687. $$
  688. u^{k}_{ijt} = \left|\frac{ \hat{z}_{ijt}-\nu^{k}_{ij}}{|\mathbf{T}_{ij} |\sigma^{k}_{ij}} \right|, \forall i \in \mathbf{N}, j \in \mathbf{M}_i, t \in \mathbf{T}_{ij}.
  689. $$
  690. Then we have
  691. $$
  692. \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}}  \left[\hat{z}_{ijt}\left(\alpha^{k}_{ijt}+ \beta^{k}_{ijt} \right) + \left|\frac{ \hat{z}_{ijt}-\nu^{k}_{ij}}{|\mathbf{T}_{ij} |\sigma^{k}_{ij}} \right|\lambda^{k}_{ij}\right] + \tau_k
  693. =\sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}}  \left[\hat{z}_{ijt}\left(\alpha^{k}_{ijt}+ \beta^{k}_{ijt} \right) + u^{k}_{ijt}\lambda^{k}_{ij}\right] + \tau_k \geq p_k.
  694. $$
  695. Therefore \eqref{Lifting-1} holds, and \eqref{Lifting-1} and \eqref{Lifting-2} are equivalent.
  696. Then, by introducing auxiliary variables $u^{k}_{ijt}, \forall i \in \mathbf{N}, j \in \mathbf{M}_i, t\in \mathbf{T}_{ij}$, we can equivalently lift the above constraints into the following optimization-based formulation:
  697. \begin{equation}\label{H-system1}
  698. \left.\begin{array}{rcll}
  699.  p_k-\tau_k\le & \min\limits_{\bm{z}, \u} & \displaystyle \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}} \left[z_{ijt}\left(\alpha^{k}_{ijt}+ \beta^{k}_{ijt} \right) + u^{k}_{ijt} \lambda^{k}_{ij}\right]  \\[0.3 cm]
  700. &{\rm s.t.} & \displaystyle   {\sum_{j \in \mathbf{M}^{\rm c}_l}\sum_{{t}\in \mathbf{T}_{l{j}}} \tilde{z}_{l{j}{t}}x_{l{j}{t}} + \tilde{z}_{l{j}{t}}x_{l{j}{t}}\leq \mathcal{T}_R} & \forall {j} \in \mathbf{M}^{\rm a}_l, {t}\in \mathbf{T}_{l{j}}  \\[0.3 cm]
  701. &&  {|\mathbf{T}_{ij} |\sigma^{k}_{ij}}u^{k}_{ijt}- { z_{ijt}} \ge { -\nu^{k}_{ij}}, & \forall i \in \mathbf{N}, j\in \mathbf{M}_i, t\in \mathbf{T}_{ij} \\[0.3 cm]
  702. &&  {|\mathbf{T}_{ij} |\sigma^{k}_{ij}}u^{k}_{ijt}+  { z_{ijt}} \ge  {\nu^{k}_{ij} }, & \forall i \in \mathbf{N}, j\in \mathbf{M}_i, t\in \mathbf{T}_{ij} \\[0.3 cm]
  703. &&  \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}} 2(\mu^{n}_{ijt} - \mu^{k}_{ijt})z_{ijt}\\[0.3 cm]
  704. && \leq \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}} {(\mu^{n}_{ijt})}^2 - {(\mu^{k}_{ijt})}^2, & \forall n \in [K]\\[0.3 cm]
  705. && \underline{z}_{ij} \leq z_{ijt} \leq \overline{z}_{ij} & \forall i \in \mathbf{N}, j \in \mathbf{M}_i, t\in \mathbf{T}_{ij}\\[0.3cm]
  706. && u^{k}_{ijt} \in \mathbb{R}, & \forall i \in \mathbf{N}, j\in \mathbf{M}_i, t\in \mathbf{T}_{ij}.
  707. \end{array}\right\}, \forall k \in [K]
  708. \end{equation}
  709. By the strong duality of linear programming, the above constraint is also equivalent to the following system: for all $k$ in $[K]$,
  710. \begin{equation*}
  711. \left\{  \begin{array}{rl}
  712. & p_k-\tau_k\le \sum\limits_{j\in \mathbf{M}^{\rm a}_{l}}\sum\limits_{t\in \mathbf{T}_{lj}} q^{lk}_{jt}\mathcal{T}_R\\[0.3 cm]
  713. & +\sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}} \left[\phi^{lk}_{ijt}\underline{z}_{ij}+\varphi^{lk}_{ijt}\overline{z}_{ij}+{\nu^{k}_{ij}\left(\pi^{lk}_{ijt}-\varpi^{lk}_{ijt} \right)}  + \sum\limits_{n \in [K]}\psi^{lk}_{n}({(\mu^{n}_{ijt})}^2 - {(\mu^{k}_{ijt})}^2)\right]  \\[0.3 cm]
  714. &  {q^{lk}_{jt}}x_{l jt}  +\sum\limits_{n \in [K]}2(\mu^{n}_{ljt} - \mu^{k}_{ljt})\psi^{lk}_{n}+ \phi^{lk}_{ljt}+\varphi^{lk}_{ljt}+\pi^{lk}_{ljt}-\varpi^{lk}_{ljt} \nonumber \\ &  = \alpha^{k}_{ljt}+\beta^{k}_{ljt}, \forall j\in \mathbf{M}^{\rm a}_l, t \in \mathbf{T}_{l j} \\[0.3 cm]
  715. &  x_{ljt}\sum\limits_{j \in \mathbf{M}^{\rm a}_l}\sum\limits_{t \in \mathbf{T}_{l j}}{q^{lk}_{jt}}  +\sum\limits_{n \in [K]}2(\mu^{n}_{ljt} - \mu^{k}_{ljt})\psi^{lk}_{k}+ \phi^{lk}_{ljt}+\varphi^{lk}_{ljt}+\pi^{lk}_{ljt}-\varpi^{lk}_{ljt} \nonumber \\ &  = \alpha^{k}_{ljt}+\beta^{k}_{ljt}, \forall j\in \mathbf{M}^{\rm c}_l, t \in \mathbf{T}_{l j} \\[0.3 cm]
  716. &   \sum\limits_{n \in [K]}2(\mu^{n}_{ijt} - \mu^{k}_{ijt})\psi^{lk}_{n}+  \phi^{lk}_{ijt}+\varphi^{lk}_{ijt}+\pi^{lk}_{ijt}-\varpi^{lk}_{ijt} = \alpha^{k}_{ijt}+\beta^{k}_{ijt}, ~ \forall i \in \mathbf{N}\setminus\{l\}, j \in \mathbf{M}_i, t\in \mathbf{T}_{ij}  \\[0.3 cm]
  717. &{|\mathbf{T}_{ij} |\sigma^{k}_{ij}}(\pi^{lk}_{ijt}+\varpi^{lk}_{ijt}) =\lambda^{k}_{ij},  ~ \forall i \in \mathbf{N}, j\in \mathbf{M}_i, t\in \mathbf{T}_{ij} \\[0.3 cm]
  718. &   q^{lk}_{jt}\le 0, ~\forall j\in \mathbf{M}^{\rm a}_l, t\in \mathbf{T}_{lj}\\& \psi^{lk}_{n} \le 0, \phi^{lk}_{ijt} \geq 0, \varphi^{lk}_{ijt} \leq 0, \pi^{lk}_{ijt}\ge 0,\varpi^{lk}_{ijt}\ge 0, ~\forall n \in [K], i \in \mathbf{N}, j\in \mathbf{M}_i, t\in \mathbf{T}_{ij}.
  719. \end{array}
  720. \right\}
  721. \end{equation*}
  722. Likewise, the constraints
  723. $$
  724. \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}}  \left[z_{ijt}\left(\alpha^{k}_{ijt}+ \beta^{k}_{ijt} \right) + \left|\frac{ z_{ijt}-\nu^{k}_{ij}}{|\mathbf{T}_{ij} |\sigma^{k}_{ij}} \right|\lambda^{k}_{ij}\right] + \tau_k \ge0, \forall \bm{z} \in \mathcal{Z}_k, k \in [K]\\
  725. $$
  726. can also be dualized similarly.  Leveraging on the derived formulation (\ref{P1-ambiguity-sup-dual1}) in Lemma~\ref{lem1-LT}, we can arrive at the formulation of the linear program (\ref{HP1-ambiguity-LP-FL})--(\ref{HP2-ambiguity-LP-FL}). We are done. \blot
  727. %\end{proof}
  728.  
  729.  
  730. Leveraging on the derived reformulation (\ref{HP1-ambiguity-LP-FL})-(\ref{HP2-ambiguity-LP-FL}) of linear program, we can readily linearize the
  731. bilinear terms $q^{lk}_{jt}x_{l jt}$ in the overall DRRO model~\eqref{HP1-ambiguity-X} using standard MIP techniques. This results in the following mixed integer linear program (MILP) reformulation for the DRRO model~\eqref{HP1-ambiguity-X}. In particular, the advantage of the resulting MIP reformulation is that it does increase the number of integers.
  732.  
  733.  
  734. \begin{proposition}\label{proposition1b}
  735. The problem (\ref{HP1-ambiguity-X}) is equivalent to the following mixed integer linear program:
  736. \begin{eqnarray}
  737. & \min\limits_{\x} &  \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \left[\sum\limits_{t\in \mathbf{T}_{ij}} x_{ijt}\right]c_{ij} \label{HP1-ambiguity-MILP-FL1}\\
  738. &{\rm s.t.} & 1-\sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}}\sum\limits_{k\in[K]} \left(\alpha^{k}_{ijt}\underline{\mu}^{k}_{ij}+ \beta^{k}_{ijt}\overline{\mu}^{k}_{ij}\right)\nonumber\\
  739. && -\sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i}\sum\limits_{k\in[K]}\epsilon^{k}_{ij}\lambda^{k}_{ij} - \sum\limits_{k\in[K]}\tau_{k}\ge R_{0}   \\
  740. &&  L_{i} \le \sum\limits_{j\in \mathbf{M}_i}\sum\limits_{t\in \mathbf{T}_{ij}} x_{ijt}\le  U_{i}, ~  \forall  i \in \mathbf{N} \\
  741. %&& \sum\limits_{j\in \mathbf{M}_{l}} \sum\limits_{k\in \mathbf{K}_{l j}}q_{l jk}\mathcal{T}_R\nonumber\\
  742. % &&+\sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{k\in \mathbf{K}_{ij}} \left[ \phi^{l }^{k}_{ij}\underline{z}_{ij}+ \varphi^{\varsigma}^{k}_{ij}\overline{z}_{ij}  + {\nu_{ij}\left(\pi^{l }^{k}_{ij}-\varpi^{l }^{k}_{ij} \right)} \right]+\tau \ge 1,~\forall {l  \in \mathbf{N}}  \\
  743. && y^{{\rm a}lk}_{jt}  +\sum\limits_{n \in [K]}2(\mu^{n}_{ljt} - \mu^{k}_{ljt})\psi^{lk}_{n}+ \phi^{lk}_{ljt}+\varphi^{lk}_{ljt}+{ \pi^{lk}_{ljt}-\varpi^{lk}_{ljt} } = \alpha^{k}_{ljt}+\beta^{k}_{ljt},\nonumber\\ &&  \forall l \in \mathbf{N}, j\in \mathbf{M}^{\rm a}_l, t \in \mathbf{T}_{l j}, k \in [K]  \label{con:32}  \\
  744. && y^{{\rm c}lk}_{jt}  +\sum\limits_{n \in [K]}2(\mu^{n}_{ljt} - \mu^{k}_{ljt})\psi^{lk}_{n}+ \phi^{lk}_{ljt}+\varphi^{lk}_{ljt}+\pi^{lk}_{ljt}-\varpi^{lk}_{ljt}  = \alpha^{k}_{ljt}+\beta^{k}_{ljt},\nonumber\\ &&  \forall l \in \mathbf{N}, j\in \mathbf{M}^{\rm c}_l, t \in \mathbf{T}_{l j}, k \in [K]    \\
  745. && (\ref{HP1-ambiguity-LP-FL1}-\ref{HP1-ambiguity-LP-FL1-2}); (\ref{HP2-ambiguity-LP-FL2})-(\ref{HP2-ambiguity-LP-FL})\\
  746. &&  q^{lk}_{jt}\le y^{{\rm a}lk}_{jt}, ~\forall l  \in \mathbf{N}, j \in \mathbf{M}^{\rm a}_{l}, t\in \mathbf{T}_{l j}, k \in [K] \\
  747. && y^{{\rm a}lk}_{jt} \ge M x_{l jt}, ~\forall l  \in \mathbf{N}, j \in \mathbf{M}^{\rm a}_{l}, t\in \mathbf{T}_{l j}, k \in [K] \\
  748. && y^{{\rm a}lk}_{jt} \le  q^{lk}_{jt}+(x_{l jt}-1)M, ~\forall l  \in \mathbf{N}, j \in \mathbf{M}^{\rm a}_{l}, t \in \mathbf{T}_{l j}, k \in [K]\\
  749. &&  \sum\limits_{j \in \mathbf{M}^{\rm a}_{l}}\sum\limits_{t\in \mathbf{T}_{l j}}q^{lk}_{jt}\le y^{{\rm c}lk}_{jt}, ~\forall l  \in \mathbf{N},j \in \mathbf{M}^{\rm c}_{l}, t\in \mathbf{T}_{l j}, k \in [K] \\
  750. && y^{{\rm c}lk}_{jt} \ge M x_{l jt}, ~\forall l  \in \mathbf{N}, j \in \mathbf{M}^{\rm c}_{l}, t\in \mathbf{T}_{l j}, k \in [K] \\
  751. && y^{{\rm c}lk}_{jt} \le  \sum\limits_{j \in \mathbf{M}^{\rm a}_{l}}\sum\limits_{t\in \mathbf{T}_{l j}}q^{lk}_{jt}+(x_{l jt}-1)M, ~\forall l  \in \mathbf{N}, j \in \mathbf{M}^{\rm c}_{l}, t \in \mathbf{T}_{l j}, k \in [K]\\
  752. && \y^{\rm a}, \y^{\rm c} \le \mathbf{0}, \x\in \{0,1\}^{H},  \label{HP1-ambiguity-MILP-FL2}
  753. \end{eqnarray}
  754. where $\halpha, \hbeta, \hlambda, \htau, \q, \hpsi, \hphi, \hvarphi, \hpi, \hvarpi, \hrho, \hvarrho, \hvarsigma, \hgamma, \htheta, \y^{\rm a}$ and $\y^{\rm c}$ are auxiliary variables and $M$ is a sufficiently small negative number.
  755. \end{proposition}
  756. %\begin{proof}
  757. {\bf Proof. }
  758. In the proof of Proposition~\ref{P-proposition1b}, the feasible set $\mathcal{Z}_k\cap \overline{\mathcal{Z}}_{l }(\x)$ of the minimization problem
  759. $$
  760. \min\limits_{\bm{z}} \displaystyle \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}} \left[z_{ijt}\left(\alpha^{k}_{ijt}+ \beta^{k}_{ijt} \right) + \left|\frac{ z_{ijt}-\nu^{k}_{ij}}{|\mathbf{T}_{ij} |\sigma^{k}_{ij}} \right| \lambda^{k}_{ij}\right]
  761. $$
  762. is bounded. Assuming that it is nonempty, then its lifted equivalent form of the inner minimization problem in (\ref{H-system1}) is also bounded and nonempty. Therefore, the dual variables $q^{lk}_{jt}$ are also bounded. Therefore, we can linearize the bilinear terms by introducing new variables $y^{{\rm a}lk}_{jt}$ and $y^{{\rm c}lk}_{jt}$, such that
  763. \begin{eqnarray}
  764. &&  q^{lk}_{jt}\le y^{{\rm a}lk}_{jt}, ~\forall l  \in \mathbf{N}, j \in \mathbf{M}^{\rm a}_{l}, t\in \mathbf{T}_{l j}, k \in [K] \nonumber\\
  765. && y^{{\rm a}lk}_{jt} \ge M x_{l jt}, ~\forall l  \in \mathbf{N}, j \in \mathbf{M}^{\rm a}_{l}, t\in \mathbf{T}_{l j}, k \in [K] \nonumber\\
  766. && y^{{\rm a}lk}_{jt} \le  q^{lk}_{jt}+(x_{l jt}-1)M, ~\forall l  \in \mathbf{N}, j \in \mathbf{M}^{\rm a}_{l}, t \in \mathbf{T}_{l j}, k \in [K]\nonumber\\
  767. &&  \sum\limits_{j \in \mathbf{M}^{\rm a}_{l}}\sum\limits_{t\in \mathbf{T}_{l j}}q^{lk}_{jt}\le y^{{\rm c}lk}_{jt}, ~\forall l  \in \mathbf{N}, , j \in \mathbf{M}^{\rm c}_{l}, t\in \mathbf{T}_{l j}, k \in [K] \nonumber\\
  768. && y^{{\rm c}lk}_{jt} \ge M x_{l jt}, ~\forall l  \in \mathbf{N}, j \in \mathbf{M}^{\rm c}_{l}, t\in \mathbf{T}_{l j}, k \in [K] \nonumber\\
  769. && y^{{\rm c}lk}_{jt} \le  \sum\limits_{j \in \mathbf{M}^{\rm a}_{l}}\sum\limits_{t\in \mathbf{T}_{l j}}q^{lk}_{jt}+(x_{l jt}-1)M, ~\forall l  \in \mathbf{N}, j \in \mathbf{M}^{\rm c}_{l}, t \in \mathbf{T}_{l j}, k \in [K],\nonumber
  770. \end{eqnarray}
  771. where $M$ is a sufficiently small negative number (in numerical computation, M can be set to a negative number with very large absolute value).  Using this linearization technique, we can arrive at the MILP  reformulation \eqref{HP1-ambiguity-MILP-FL1}-\eqref{HP1-ambiguity-MILP-FL2} for the DRRO model (\ref{HP1-ambiguity-X}). \blot
  772. }
  773. \subsection{Learning the clustering parameter $K$ by Cross Validation}\label{subsec:CV}{\color{blue}
  774. It is known that the clustering approaches ({e.g.}, K-means) in general belong to the {\it unsupervised learning} methods (Hastie~et~al.~2009), for which there is no labeled loss function for validating the parameter $K$\footnotemark \footnotetext{One of the popular methods to choose $K$ is the elbow method ({\color{red}Thorndike 1953, Ketchen et al. 1996}), in which the ratio of in-class variance and total variance is plotted, and optimal $K$ is chosen by observing the plot, which apparently depends partially on subjective understanding of the data.}.
  775. Interestingly, our distributionally robust redundancy optimization framework with lifetime pattern clustering (of hyper-parameter $K$) can essentially be treated as a {\it supervised learning} from statistical learning perspective ({\color{red} Friedman et al. 2001, James et al. 2013}). In particular, the proposed optimization framework provides two labels of (i) redundancy cost and (ii) feasibility of reliability requirement, in the DRRO model~\eqref{HP1-ambiguity-X}. This important feature of our framework enables us to design {\it Cross Validation} to choose the {\it best} parameter $K$. Specifically, considering the problem structure of chance constrained optimization where the objective of  redundancy cost and feasibility of constraint (reliability) needs to be balanced, where the balance actually reflects the user's design preference in between the two. This motivates us introduce a preference level $\lambda$ to design the Cross Validation. Specifically, for each choice of $K$, we consider the number of constraint violation in cross validation, as well as its average design cost. We normalize both into the range of $[0, 1]$ separately, and combine them linearly by the coefficient $\lambda$. Low $\lambda$ means that the designer cares more about low cost than robustness (reliability); high $\lambda$ indicates the contrary. By applying this metric, cross validation can help choosing $K$ to find a balance of cost and robustness as per necessary. The procedures of cross validation are summarized in the following Algorithm 1.
  776. \noindent\rule{\textwidth}{0.1em}\vspace{-5pt}\\
  777. \noindent {\bf Algorithm 1. Cross Validation for Selecting $K$} \vspace{-10pt}\\
  778. \noindent\rule{\textwidth}{0.05em} \\
  779. {\bf Input:} Data sample $\mathcal{D}$, preference level $\lambda$ and the set of possible candidates for $K$: $\{K_l, l \in [L]\}$.
  780. \begin{enumerate}
  781. \item  Split $\mathcal{D}$ into $M$ disjoint subsets $\mathcal{D}_m, m \in [M]$, each of them have equal size.
  782. \item  For each $K_l, l \in [L]$:
  783. \begin{enumerate}
  784.  \item Training: For each subset $\mathcal{D}_m, m \in [M]$, compute the optimal solution $\x^{*}(\mathcal{D}_m, K_l)$ and associated optimal cost $c(\mathcal{D}_m, K_l)$ by solving the model, using $\mathcal{D} \setminus \mathcal{D}_m$ as input data and split it into $K_l$ clusters.
  785.  \item Validation: For each $m \in [M]$: For $\mathcal{D}_m$, use it as validation set to compute $R(\x^{*}(\mathcal{D}_m, K_l))$, the reliability level under design $\x^{*}(\mathcal{D}_m, K_l)$. If $R(\x^{*}(\mathcal{D}_m, K_l)) < R_0$, count it as one instance of violation. Sum up the total number of violations across all $m \in [M]$, that is, $$v(K_l) := \sum\limits_{m \in [M]}\mathbb{I}\Big[ R(\x^{*}(\mathcal{D}_m, K_l)) < R_0 \Big],$$
  786.      where $\mathbb{I}[\cdot]$ is the indicator function. Compute average cost as $$\overline{c}(K_l) := \frac{1}{n}\sum\limits_{m \in M}c(\mathcal{D}_m, K_l)$$
  787.  \item Compute the combined metric $\lambda{v( K_l)} + (1-\lambda)\overline{c}(K_l)$. Let $$K_{\rm opt} := \underset{K_l}{\mathrm{arg\,min}}\,\lambda{v(K_l)} + (1-\lambda)\overline{c}(K_l).$$
  788. \end{enumerate}
  789. \end{enumerate}
  790. \noindent{\bf Output:} The optimal value $K_{\rm opt}$ for parameter $K$.  
  791. \vspace{-5pt}
  792. \noindent\rule{\textwidth}{0.1em}\vspace{-17pt}\\
  793. }
  794. \section{A Supergradient-based Decomposition Algorithm}{\color{blue}
  795. Although the linearized MIP formulation is tractable, when $K$ is large, it may take long time to solve it due to the large number of constraints. In this subsection we provide an iterative algorithm that exploits the structure of the problem, that decomposed that problem into subproblems that can be parallelized.
  796. We first introduce $\y^a$ and $\y^c$ to separate $\q$ and $\x$ in \eqref{HP1-ambiguity-LP-FL}, by the same approach in the proof of proposition 3. Then, computing its dual,
  797. we obtain the following formulation:
  798. \begin{eqnarray} \label{dual}
  799. &\!\!\!\!\!\! \min &
  800. \sum\limits_{l\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}^{a}_l} \sum\limits_{t\in \mathbf{T}_{lj}}\sum\limits_{k\in[K]}(\gamma^{lk}_{jt}+\delta^{lk}_{jt})Mx_{ljt} +
  801. \sum\limits_{l\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}^{c}_l} \sum\limits_{t\in \mathbf{T}_{lj}}\sum\limits_{k\in[K]}(\lambda^{lk}_{jt}+\tau^{lk}_{jt})Mx_{ljt} \nonumber\\
  802. && + \sum\limits_{l\in \mathbf{N}}\sum\limits_{k\in[K]}{p_k}{\alpha^{l}_k} -
  803. \sum\limits_{l\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}^{a}_l} \sum\limits_{t\in \mathbf{T}_{lj}}\sum\limits_{k\in[K]}\delta^{lk}_{jt}M -
  804. \sum\limits_{l\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}^{c}_l} \sum\limits_{t\in \mathbf{T}_{lj}}\sum\limits_{k\in[K]}\tau^{lk}_{jt}M \\
  805. &\!\!\!\!\!\!{\rm s.t.} & \rho^{ik}_{jt}+\sum\limits_{l \in \mathbf{N} \setminus \{i\}}\phi^{lk}_{ijt} + \varphi^{k}_{ijt} \ge \underline{\mu}^{k}_{ij}, \forall i \in \mathbf{N}, j\in \mathbf{M}^{a}_i, t\in \mathbf{T}_{ij}, k\in[K]\label{model-c-1}\\
  806. && \varrho^{ik}_{jt}+\sum\limits_{l \in \mathbf{N} \setminus \{i\}}\phi^{lk}_{ijt} + \varphi^{k}_{ijt} \ge \underline{\mu}^{k}_{ij}, \forall i \in \mathbf{N}, j\in \mathbf{M}^{c}_i, t\in \mathbf{T}_{ij}, k\in[K]\\
  807. && \rho^{ik}_{jt}+\sum\limits_{l \in \mathbf{N} \setminus \{i\}}\phi^{lk}_{ijt} + \varphi^{k}_{ijt} \le \overline{\mu}^{k}_{ij}, \forall i \in \mathbf{N}, j\in \mathbf{M}^{a}_i, t\in \mathbf{T}_{ij}, k\in[K]\\
  808. && \varrho^{ik}_{jt}+\sum\limits_{l \in \mathbf{N} \setminus \{i\}}\phi^{lk}_{ijt} + \varphi^{k}_{ijt} \le \overline{\mu}^{k}_{ij}, \forall i \in \mathbf{N}, j\in \mathbf{M}^{c}_i, t\in \mathbf{T}_{ij}, k\in[K]\\
  809. && \sum\limits_{l \in \mathbf{N}}\sum\limits_{t\in \mathbf{T}_{ij}}\pi^{lk}_{ijt}+\sum\limits_{t\in \mathbf{T}_{ij}}\varpi^{k}_{ijt} \le \epsilon^{k}_{ij}, \forall i \in \mathbf{N}, j\in \mathbf{M}_i, k\in[K]\\
  810. && \sum\limits_{i \in \mathbf{N}}\alpha_{ik} + \beta_{k} = -1, \forall k \in [K]\\
  811. && \rho^{ik}_{jt}-\theta^{ik}_{jt}+\gamma^{ik}_{jt}+\delta^{ik}_{jt}\le 0, \forall i \in \mathbf{N}, j\in \mathbf{M}^{a}_i, t\in \mathbf{T}_{ij}, k\in[K]\\
  812. && \varrho^{ik}_{jt}-\vartheta^{ik}_{jt}+\lambda^{ik}_{jt}+\tau^{ik}_{jt}\le 0, \forall i \in \mathbf{N}, j\in \mathbf{M}^{c}_i, t\in \mathbf{T}_{ij}, k\in[K]\\
  813. && \mathcal{T}_R\alpha_{ik} + \theta^{ik}_{jt} - \delta^{ik}_{jt} + \sum\limits_{j \in \mathbf{M}^c_i}\sum\limits_{t \in \mathbf{T}_{ij}}(\vartheta^{ik}_{jt} - \tau^{ik}_{jt}) \le 0, \forall i \in \mathbf{N}, j\in \mathbf{M}^{a}_i, t\in \mathbf{T}_{ij}, k\in[K]\\
  814. && \underline{z}_{lj}\alpha_{lk} + \rho^{lk}_{jt} \ge 0, \forall l \in \mathbf{N}, j\in \mathbf{M}^a_l, t\in \mathbf{T}_{lj}, k\in[K]\\
  815. && \underline{z}_{lj}\alpha_{lk} + \varrho^{lk}_{jt} \ge 0, \forall l \in \mathbf{N}, j\in \mathbf{M}^c_l, t\in \mathbf{T}_{lj}, k\in[K]\\
  816. && \underline{z}_{ij}\alpha_{lk} + \phi^{lk}_{ijt} \ge 0, \forall l \in \mathbf{N}, i \in \mathbf{N} \setminus \{l\}, j\in \mathbf{M}_i, t\in \mathbf{T}_{ij}, k\in[K]\\
  817. && \overline{z}_{lj}\alpha_{lk} + \rho^{lk}_{jt} \le 0, \forall l \in \mathbf{N}, j\in \mathbf{M}^a_l, t\in \mathbf{T}_{lj}, k\in[K]\\
  818. && \overline{z}_{lj}\alpha_{lk} + \varrho^{lk}_{jt} \le 0, \forall l \in \mathbf{N}, j\in \mathbf{M}^c_l, t\in \mathbf{T}_{lj}, k\in[K]\\
  819. && \overline{z}_{ij}\alpha_{lk} + \phi^{lk}_{ijt} \le 0, \forall l \in \mathbf{N}, i \in \mathbf{N} \setminus \{l\}, j\in \mathbf{M}_i, t\in \mathbf{T}_{ij}, k\in[K]\\
  820. && \sum\limits_{i \in \mathbf{N}}\sum\limits_{j \in \mathbf{M}_i}\sum\limits_{t \in \mathbf{T}_{ij}}\Big[\Big((\mu^n_{ijt})^2-(\mu^k_{ijt})^2\Big)\alpha_{lk}\Big] + \sum\limits_{j \in \mathbf{M}^a_l}\sum\limits_{t \in \mathbf{T}_{lj}}2(\mu^n_{ljt}-\mu^k_{ljt})\rho^{lk}_{jt} + \sum\limits_{j \in \mathbf{M}^c_l}\sum\limits_{t \in \mathbf{T}_{lj}}2(\mu^n_{ljt}-\mu^k_{ljt})\varrho^{lk}_{jt}\nonumber \\&& + \sum\limits_{i \in \mathbf{N} \setminus \{l\}}\sum\limits_{j \in \mathbf{M}_i}\sum\limits_{t \in \mathbf{T}_{ij}}2(\mu^n_{ijt}-\mu^k_{ijt})\phi^{lk}_{ijt} \le 0 ,\forall l \in \mathbf{N}, n \in [K], k \in [K]\\
  821. && \nu^{k}_{lj}\alpha_{lk} + \rho^{lk}_{jt} + \pi^{lk}_{ljt} \ge 0, \forall l \in \mathbf{N}, j\in \mathbf{M}^a_l, t\in \mathbf{T}_{lj}, k\in[K]\\
  822. && \nu^{k}_{lj}\alpha_{lk} + \varrho^{lk}_{jt} + \pi^{lk}_{ljt} \ge 0, \forall l \in \mathbf{N}, j\in \mathbf{M}^c_l, t\in \mathbf{T}_{lj}, k\in[K]\\
  823. && \nu^{k}_{ij}\alpha_{lk} + \phi^{lk}_{ijt} + \pi^{lk}_{ijt} \ge 0, \forall l \in \mathbf{N}, i \in \mathbf{N} \setminus \{l\}, j\in \mathbf{M}_i, t\in \mathbf{T}_{ij}, k\in[K]\\
  824. && \nu^{k}_{lj}\alpha_{lk} + \rho^{lk}_{jt} - \pi^{lk}_{ljt} \le 0, \forall l \in \mathbf{N}, j\in \mathbf{M}^a_l, t\in \mathbf{T}_{lj}, k\in[K]\\
  825. && \nu^{k}_{lj}\alpha_{lk} + \varrho^{lk}_{jt} - \pi^{lk}_{ljt} \le 0, \forall l \in \mathbf{N}, j\in \mathbf{M}^c_l, t\in \mathbf{T}_{lj}, k\in[K]\\
  826. && \nu^{k}_{ij}\alpha_{lk} + \phi^{lk}_{ijt} - \pi^{lk}_{ijt} \le 0, \forall l \in \mathbf{N}, i \in \mathbf{N} \setminus \{l\}, j\in \mathbf{M}_i, t\in \mathbf{T}_{ij}, k\in[K]\\
  827. && \underline{z}_{ij}\beta_k + \varphi^{k}_{ijt} \ge 0, \forall i \in \mathbf{N}, j\in \mathbf{M}_i, t\in \mathbf{T}_{ij}, k\in[K]\\
  828. && \overline{z}_{ij}\beta_k + \varphi^{k}_{ijt} \le 0, \forall i \in \mathbf{N}, j\in \mathbf{M}_i, t\in \mathbf{T}_{ij}, k\in[K]\\
  829. && \sum\limits_{i \in \mathbf{N}}\sum\limits_{j \in \mathbf{M}_i}\sum\limits_{t \in \mathbf{T}_{ij}}\Big[\Big((\mu^n_{ijt})^2-(\mu^k_{ijt})^2\Big)\beta_k + 2(\mu^n_{ijt}-\mu^k_{ijt})\varphi^{k}_{ijt}\Big] \le 0 ,\forall n \in [K], k \in [K]\\
  830. && \nu^{k}_{ij}\beta_k + \varphi^{k}_{ijt} + \varpi^{k}_{ijt} \ge 0, \forall i \in \mathbf{N}, j\in \mathbf{M}_i, t\in \mathbf{T}_{ij}, k\in[K]\\
  831. && \nu^{k}_{ij}\beta_k + \varphi^{k}_{ijt} - \varpi^{k}_{ijt} \le 0, \forall i \in \mathbf{N}, j\in \mathbf{M}_i, t\in \mathbf{T}_{ij}, k\in[K]\\
  832. && \halpha,\hbeta, \hgamma, \hlambda \le \mathbf{0},  \\
  833. && \htheta, \hdelta, \hvartheta, \htau \ge \mathbf{0}, \label{model:dual-S}
  834. \end{eqnarray}
  835. where $({\halpha}, {\hbeta}, {\hphi}, {\hpi}, {\hvarphi}, {\hvarpi}, {\hrho}, {\hvarrho}, {\htheta}, {\hgamma}, {\hdelta}, {\hvartheta}, {\hlambda}, {\htau})$ are auxiliary variables.
  836. Let $D(\x)$ be the optimal solution of the above dual program \eqref{dual}-\eqref{model:dual-S} given $\x$, it is noted that computing $D(\x)$ solves a linear program. Furthermore, This linear program problem $D(\x)$ can be decomposed into $K$ subproblems by splitting the constraints and objective based on $k \in [K]$.  Therefore, the computation of $D(\x)$ can be efficiently even if $K$ is very large.
  837.  
  838. %\begin{eqnarray}
  839. %&\!\!\!\!\!\! D_k(\x) = \min &  \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}} \left(\alpha^{k}_{ijt}\underline{\mu}^{k}_{ij} + \beta^{k}_{ijt}\overline{\mu}^{k}_{ij}\right) + \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i}\epsilon^{k}_{ij}\lambda^{k}_{ij} - \tau_{k} \\
  840. % &\!\!\!\!\!\!{\rm s.t.} & \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}} \Big[\phi^{lk}_{ijt}\underline{z}_{ij}+\varphi^{lk}_{ijt}\overline{z}_{ij}+{\nu^{k}_{ij}\left(\pi^{lk}_{ijt}-\varpi^{lk}_{ijt} \right)}  \nonumber\\
  841. %  &&+  \sum\limits_{n \in [K]}\psi^{lk}_{n}({(\mu^{n}_{ijt})}^2 - {(\mu^{k}_{ijt})}^2)\Big] + \sum\limits_{j\in \mathbf{M}^{\rm a}_{l}}\sum\limits_{t\in \mathbf{T}_{lj}} q^{lk}_{jt}\mathcal{T}_R+ \tau_k \geq p_k, \forall l \in \mathbf{N}\\
  842. %  && \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}} \Big[\rho^{k}_{ijt}\underline{z}_{ij}+\varrho^{k}_{ijt}\overline{z}_{ij}+{\nu^{k}_{ij}\left(\gamma^{k}_{ijt}-\theta^{k}_{ijt} \right)} \nonumber\\ && + \sum\limits_{n \in [K]}\varsigma^{k}_{n}({(\mu^{n}_{ijt})}^2 - {(\mu^{k}_{ijt})}^2) \Big] + \tau_k \geq0\\
  843. %&&  {q^{lk}_{jt}}x_{l jt}  +\sum\limits_{n \in [K]}2(\mu^{n}_{ljt} - \mu^{k}_{ljt})\psi^{lk}_{n}+ \phi^{lk}_{ljt}+\varphi^{lk}_{ljt}+\pi^{lk}_{ljt}-\varpi^{lk}_{ljt} \nonumber \\ &&  = \alpha^{k}_{ljt}+\beta^{k}_{ljt}, \forall l \in \mathbf{N}, j\in \mathbf{M}^{\rm a}_l, t \in \mathbf{T}_{l j} \\[0.3 cm]
  844. %&&  x_{l jt}\sum\limits_{j \in \mathbf{M}^{\rm a}_l}\sum\limits_{t \in \mathbf{T}_{l j}}{q^{lk}_{jt}}  +\sum\limits_{n \in [K]}2(\mu^{n}_{ljt} - \mu^{k}_{ljt})\psi^{lk}_{n}+ \phi^{lk}_{ljt}+\varphi^{lk}_{ljt}+\pi^{lk}_{ljt}-\varpi^{lk}_{ljt} \nonumber \\ &&  = \alpha^{k}_{ljt}+\beta^{k}_{ljt}, \forall l \in \mathbf{N}, j\in \mathbf{M}^{\rm c}_l, t \in \mathbf{T}_{l j} \\[0.3 cm]
  845. %&&   \sum\limits_{n \in [K]}2(\mu^{n}_{ijt} - \mu^{k}_{ijt})\psi^{lk}_{n}+  \phi^{lk}_{ijt}+\varphi^{lk}_{ijt}+\pi^{lk}_{ijt}-\varpi^{lk}_{ijt} = \alpha^{k}_{ijt}+\beta^{k}_{ijt},\nonumber\\ && ~ \forall l \in \mathbf{N}, i \in \mathbf{N}\setminus\{l \}, j \in \mathbf{M}_i, t\in \mathbf{T}_{ij} \\
  846. %&&{|\mathbf{T}_{ij} |\sigma^{k}_{ij}}(\pi^{lk}_{ijt}+\varpi^{lk}_{ijt}) =\lambda^{k}_{ij},  ~ \forall l  \in \mathbf{N}, i \in \mathbf{N}, j\in \mathbf{M}_i, t\in \mathbf{T}_{ij} \\
  847. %&& \sum\limits_{n \in [K]}2(\mu^{n}_{ijt} - \mu^{k}_{ijt})\varsigma^{k}_{n}+ \rho^{k}_{ijt}+\varrho^{k}_{ijt}+\gamma^{k}_{ijt}-\theta^{k}_{ijt} = \alpha^{k}_{ijt}+\beta^{k}_{ijt},\nonumber\\ && ~ \forall i \in \mathbf{N}, j \in \mathbf{M}_i, t\in \mathbf{T}_{ij}  \\
  848. %&&  {|\mathbf{T}_{ij} |\sigma^{k}_{ij}}(\gamma^{k}_{ijt}+\theta^{k}_{ijt}) =\lambda^{k}_{ij},  ~ \forall i \in \mathbf{N}, j\in \mathbf{M}_i, t\in \mathbf{T}_{ij} \\
  849. %%&&  q_{l jk}\le y_{{l jk}}, ~\forall l  \in \mathbf{N}, j \in \mathbf{M}_{l}, k\in \mathcal{N}(l,j) \\[0.3 cm]
  850. %%&& y_{{l jk}} \ge M x_{l jk}, ~\forall l  \in \mathbf{N}, j \in \mathbf{M}_{l}, k\in \mathcal{N}(l,j) \\[0.3 cm]
  851. %%&& y_{{l jk}} \le  q_{l jk }+(x_{l jk}-1)M, ~\forall l  \in \mathbf{N}, j \in \mathcal{J}({l}), k \in \mathcal{N}(l,j)\\[0.3 cm]
  852. %%&&  \sum\limits_{k\in \mathbf{K}_{ij}} x^{k}_{ij}\ge  L_{ij}, ~  \forall  i \in \mathbf{N}, j\in \mathbf{M}_i \\
  853. %&& \halpha,\q, \hpsi, \hvarphi, \hvarsigma, \hvarrho \le \mathbf{0}, \htau_{k} \in \mathbb{R}  \\
  854. %&& \hbeta, \hlambda, \hphi, \hrho, \hpi,\hvarpi, \hgamma, \htheta \ge \mathbf{0}
  855. %\end{eqnarray}. \fi
  856.  
  857. Now the problem \eqref{HP1-ambiguity-X} can be rewritten as
  858. \begin{equation}\label{HP1-ambiguity-XI}
  859. \begin{array}{rcll}
  860. & \min\limits_{\x} &  C(\x)=\sum\limits_{i\in \mathbf{N}} \sum\limits_{j \in \mathbf{M}_i}\left[ \sum\limits_{t\in \mathbf{T}_{ij}} x_{ijt}\right]c_{ij} \\[0.5 cm]
  861. &{\rm s.t.} & D(\x) \ge R_{0} & \\[0.3 cm]
  862. && L_{i} \le \sum\limits_{j\in \mathbf{M}_i}\sum\limits_{t\in \mathbf{T}_{ij}} x_{ijt}\le  U_{i}, \forall i \in \mathbf{N}\\[0.3 cm]
  863. && \x \in \{0,1\}^H
  864. \end{array}
  865. \end{equation}
  866. which is termed the {\it master} problem.
  867.  
  868.  
  869. We now approach this master problem \eqref{HP1-ambiguity-XI} by solving a sequence of {\it supergradient} based relaxed problems iteratively to approximate the optimality. If its optimal solution is feasible in the original problem, then the solution is also optimal in the original problem. Otherwise, we add more constraints into the problem. Repeat this process until either the problem becomes infeasible, or an optimal solution is obtained.
  870.  
  871.  
  872. It is noted that $D(\cdot)$ is a concave function which always has a {\it supergradient}. In particular, we denote by $\s(\x)$ a {\it supergradient} of $D(\x)$ at $\x$, which by definition means that $D(\x) \le D(\y) + s(\y)^{\top}(\x-\y), \forall \y \in S$. Note that when $\x = \y$, the two sides are equal. With supergradient $\s(\x)$, we can formulate the {\it relaxed master problem} for \eqref{HP1-ambiguity-XI} as
  873. \begin{equation}\label{HP1-ambiguity-XIII}
  874. \begin{array}{rcll}
  875. & \min\limits_{\x} &  C(\x)=\sum\limits_{i\in \mathbf{N}} \sum\limits_{j \in \mathbf{M}_i}\left[ \sum\limits_{t\in \mathbf{T}_{ij}} x_{ijt}\right]c_{ij} \\[0.5 cm]
  876. &{\rm s.t.} & D(\y) + s(\y)^{\top}(\x-\y) \ge R_{0}, & \forall \y \in \mathcal{Y} \subset \{0,1\}^H \\[0.3 cm]
  877. && L_{i} \le \sum\limits_{j\in \mathbf{M}_i}\sum\limits_{t\in \mathbf{T}_{ij}} x_{ijt}\le  U_{i}, & \forall i \in \mathbf{N}\\[0.3 cm]
  878. && \x \in \{0,1\}^H.
  879. \end{array}
  880. \end{equation}
  881.  
  882.  
  883. The following lemma computes the supergradient $\s(\x)$ of $D(\x)$.
  884. \begin{lemma}
  885. Given any feasible $\x$ and $(\overline{\halpha}, \overline{\hbeta}, \overline{\hphi}, \overline{\hpi}, \overline{\hvarphi}, \overline{\hvarpi}, \overline{\hrho}, \overline{\hvarrho}, \overline{\htheta}, \overline{\hgamma}, \overline{\hdelta}, \overline{\hvartheta}, \overline{\hlambda}, \overline{\htau})$ is the optimal solution of the dual problem \eqref{dual}-\eqref{model:dual-S} given $\x$. Let
  886. $$\s(\x)=\Big(s_{ijt}(\x)\Big)_{i\in \mathbf{N}, j \in \mathbf{M}_i, t \in \mathbf{T}_{ij}}$$
  887. with
  888. \begin{equation}
  889. s_{ijt}(\x) := \left\{
  890.                \begin{array}{ll}
  891.                \sum\limits_{k\in[K]}\left(\overline{\gamma}^{lk}_{jt}+\overline{\delta}^{lk}_{jt}\right)M, \forall l \in \mathbf{N}, j \in \mathbf{M}^a_l, t \in \mathbf{T}_{lj}\\
  892.                \sum\limits_{k\in[K]}\left(\overline{\lambda}^{lk}_{jt}+\overline{\tau}^{lk}_{jt}\right)M, \forall l \in \mathbf{N}, j \in \mathbf{M}^c_l, t \in \mathbf{T}_{lj}\end{array}\right.,
  893. \end{equation}
  894. where  and define $\overline{\hgamma}^{l}_j = \overline{\hdelta}^{l}_j = \mathbf{0}, \forall j \in \mathbf{M}_{l}^{c}$ and $\overline{\hlambda}^{l}_j = \overline{\htau}^{l}_j = \mathbf{0}, \forall j \in \mathbf{M}_{l}^{a}$. Then $\s(\x)$ is a supergradient of $D(\x)$ at $\x$.
  895. \end{lemma}
  896. {\bf Proof. }
  897. Since $D(\x)$ is concave, it is equivalent to prove the following inequality:
  898. \begin{equation}\label{subgrad}
  899. D(\hat{\x}) \le D(\x) + s(\x)^{\top}(\hat{\x}-\x), \forall \hat{\x} \in \mathcal{X},
  900. \end{equation}
  901. where $$\mathcal{X}:= \left\{\x \in \{0,1\}^H:  L_{i} \le \sum\limits_{j\in \mathbf{M}_i}\sum\limits_{t\in \mathbf{T}_{ij}} x_{ijt}\le  U_{i}, \forall i \in \mathbf{N}\right\}.$$
  902. After obtaining $(\overline{\halpha}, \overline{\hbeta}, \overline{\hphi}, \overline{\hpi}, \overline{\hvarphi}, \overline{\hvarpi}, \overline{\hrho}, \overline{\hvarrho}, \overline{\htheta}, \overline{\hgamma}, \overline{\hdelta}, \overline{\hvartheta}, \overline{\hlambda}, \overline{\htau})$ as the optimal solution for given $\x$, we can rewrite the right side of \eqref{subgrad} as:
  903. \begin{eqnarray}
  904. && \sum\limits_{l\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}^{a}_l} \sum\limits_{t\in \mathbf{T}_{lj}}\sum\limits_{k\in[K]}(\overline{\gamma}^{lk}_{jt}+\overline{\delta}^{lk}_{jt})M\hat{x}_{ljt} +
  905. \sum\limits_{l\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}^{c}_l} \sum\limits_{t\in \mathbf{T}_{lj}}\sum\limits_{k\in[K]}(\overline{\lambda}^{lk}_{jt}+\overline{\tau}^{lk}_{jt})M\hat{x}_{ljt} \nonumber\\
  906. & +& \sum\limits_{l\in \mathbf{N}}\sum\limits_{k\in[K]}{p_k}{\overline{\alpha}^{l}_k} -
  907. \sum\limits_{l\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}^{a}_l} \sum\limits_{t\in \mathbf{T}_{lj}}\sum\limits_{k\in[K]}\overline{\delta}^{lk}_{jt}M -
  908. \sum\limits_{l\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}^{c}_l} \sum\limits_{t\in \mathbf{T}_{lj}}\sum\limits_{k\in[K]}\overline{\tau}^{lk}_{jt}M,
  909. \end{eqnarray}
  910. and the left side as
  911. \begin{eqnarray*}
  912. &\!\!\!\!\!\! \min &
  913. \sum\limits_{l\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}^{a}_l} \sum\limits_{t\in \mathbf{T}_{lj}}\sum\limits_{k\in[K]}(\gamma^{lk}_{jt}+\delta^{lk}_{jt})M\hat{x}_{ljt} +
  914. \sum\limits_{l\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}^{c}_l} \sum\limits_{t\in \mathbf{T}_{lj}}\sum\limits_{k\in[K]}(\lambda^{lk}_{jt}+\tau^{lk}_{jt})M\hat{x}_{ljt} \nonumber\\
  915. && + \sum\limits_{l\in \mathbf{N}}\sum\limits_{k\in[K]}{p_k}{\alpha^{l}_k} -
  916. \sum\limits_{l\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}^{a}_l} \sum\limits_{t\in \mathbf{T}_{lj}}\sum\limits_{k\in[K]}\delta^{lk}_{jt}M -
  917. \sum\limits_{l\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}^{c}_l} \sum\limits_{t\in \mathbf{T}_{lj}}\sum\limits_{k\in[K]}\tau^{lk}_{jt}M\\
  918. &\!\!\!\!\!\!{\rm s.t.} & \eqref{model-c-1}-\eqref{model:dual-S}.
  919. \end{eqnarray*}
  920. Since the constraints are independent of $\hat{\x}$, $(\overline{\halpha}, \overline{\hbeta}, \overline{\hphi}, \overline{\hpi}, \overline{\hvarphi}, \overline{\hvarpi}, \overline{\hrho}, \overline{\hvarrho}, \overline{\htheta}, \overline{\hgamma}, \overline{\hdelta}, \overline{\hvartheta}, \overline{\hlambda}, \overline{\htau})$ is a feasible solution of $D(\hat{\x})$. Since $D(\hat{\x})$ is the solution to a minimization problem,
  921. \begin{eqnarray*}
  922. D(\hat{\x})&\le& \sum\limits_{l\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}^{a}_l} \sum\limits_{t\in \mathbf{T}_{lj}}\sum\limits_{k\in[K]}(\overline{\gamma}^{lk}_{jt}+\overline{\delta}^{lk}_{jt})M\hat{x}_{ljt} +
  923. \sum\limits_{l\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}^{c}_l} \sum\limits_{t\in \mathbf{T}_{lj}}\sum\limits_{k\in[K]}(\overline{\lambda}^{lk}_{jt}+\overline{\tau}^{lk}_{jt})M\hat{x}_{ljt} \nonumber\\
  924. &+& \sum\limits_{l\in \mathbf{N}}\sum\limits_{k\in[K]}{p_k}{\overline{\alpha}^{l}_k} -
  925. \sum\limits_{l\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}^{a}_l} \sum\limits_{t\in \mathbf{T}_{lj}}\sum\limits_{k\in[K]}\overline{\delta}^{lk}_{jt}M -
  926. \sum\limits_{l\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}^{c}_l} \sum\limits_{t\in \mathbf{T}_{lj}}\sum\limits_{k\in[K]}\overline{\tau}^{lk}_{jt}M, \forall \hat{\x} \in \mathcal{X}.
  927. \end{eqnarray*}
  928. It is equivalent to \eqref{subgrad}, so we have proved that $\s(\x)$ is a supergradient of $D(\x)$ at $\x$. \blot\\
  929.  
  930. %Furthermore, if we compute $D(\y^m)$ and the supergradient $\s(\y^m)$ with $M$ points $\y^m \in S, \forall m \in [M](M \le |S|)$, we form the following relaxed problem:
  931. %\begin{equation}\label{HP1-ambiguity-XIII}
  932. %\begin{array}{rcll}
  933. %& \min\limits_{\x} &  C(\x) \\[0.5 cm]
  934. %&{\rm s.t.} & D(\y^m) + s(\y^m)^{\top}(\x-\y^m) \ge R_{0}, & \forall m \in [M]\\[0.3 cm]
  935. %&& \x \in S.
  936. %\end{array}
  937. %\end{equation}
  938.  
  939. It is noted that if the optimal solution to the relaxed problem \eqref{HP1-ambiguity-XIII} is feasible in the original problem, it is also the optimal solution of the original problem. We present the overall algorithm below:
  940.  
  941. \noindent\rule{\textwidth}{0.1em}\vspace{-5pt}\\
  942. \noindent {\bf Algorithm 2.} Supergradient-based iterative optimization framework for problem \eqref{HP1-ambiguity-XI}\vspace{-10pt}\\
  943. \noindent\rule{\textwidth}{0.05em} \\
  944. {\bf Input:} $y^0 \in S, m=M=0, LB=-\infty, UB=\infty$. \\
  945. {\bf Output:} Design $\x^{*}$ and associated cost.
  946.  
  947. \begin{enumerate}
  948. \item  Compute $D(\y^m)$;
  949. \item  Compute supergradient $\s(\y^m)$ at $\y^m$, and solve \eqref{HP1-ambiguity-XIII}: if it is infeasible, STOP and declare the overall problem infeasible; otherwise, obtain $\x^{*}$ as the optimal solution;
  950. \item  If $D(\x^*) \ge R_0$, STOP; else, let $m = m + 1$, $M = M + 1$, $\y^m = \x^{*}$ and go to STEP 1;
  951. \item  {\bf return} Design $\x^{*}$ and associated cost $C(\x^{*})$.
  952. \end{enumerate}
  953.  
  954. \vspace{-5pt}
  955. \noindent\rule{\textwidth}{0.1em}\vspace{-17pt}\\
  956. }
  957. \iffalse
  958. Note that the algorithm above produce exact solution to the problem. By relaxing the condition $UB \le LB$ in STEP 3 to $UB - LB \le \epsilon$ for some $\epsilon > 0$, we can also get an algorithm that produces approximate solution to the problem, potentially with much less time.
  959. \fi
  960.  
  961.  
  962.  
  963.  
  964.  
  965. %\section{Computational Study}
  966. %In this section we present numerical experiments of our model, as well as a case study. The computational study consists of six parts: (i) visualizing the result of dimension reduction and clustering; (ii) testing how design changes when parameters vary, including $K$ and other hyperparameters; (iii) choosing best parameter $K$ by cross validation; (iv) comparing our design with a baseline probabilistic model;  (v) experimenting on the value of side information and (vi) a real-life case. The distribution used in experiment (i)-(v) is generated by a real data set.
  967. %All computational result were done on a PC with Intel(R) Core(TM) i7-7500U CPU at 2.7 GHz, coded in Python. The MIP models were solved by library called Gurobi, version 8.1.1.
  968. %
  969. %\subsection{Visualizing clusters}
  970. %In this section we present a visualization of dimension reduction and clustering. We first apply tSNE algorithm, a state-of-art algorithm for dimension reduction and visualization, \iffalse (refer to Maaten L, Hinton G. Visualizing data using t-SNE[J]. Journal of machine learning research, 2008, 9(Nov): 2579-2605.)\fi to reduce the lifetime data from 28 dimensions to 2 dimensions, and then perform K-means clustering. We present figures of different choice of number of clusters, $K=2$ and $K=5$, respectively. As we can see in the figure, the original multi-dimensional data can be well clustered after dimension reduction. The clustering results, including results from other choices of $K$, are used in %following experiments.
  971. %
  972. %\begin{figure}[H]
  973. %\centering
  974. %%\includegraphics{TSNE_1.png}
  975. %\includegraphics[width=\columnwidth]{2D_tsne.jpg}
  976. %\caption{\footnotesize Visualization of dimension reduction by tSNE algorithm and clustering by K-means algorithms, into 2 clusters on the left and 5 clusters on the right. Note that the two figures are produced by different data.}
  977. %\label{figure_TSNE_1}
  978. %\end{figure}
  979. %
  980. %\subsection{System design variation with different parameters}
  981. %
  982. %\begin{figure}[H]
  983. %\centering
  984. %\includegraphics[scale=0.8]{KVARIATION11.pdf}
  985. %\caption{ The multi-type series-parallel system we are experimenting in this section. It consists of 3 subsystems, with 1 type of active parallel component and 2 types of cold-standby component in each subsystem. Each type of component can have up to 3 redundancies.}
  986. %\label{figure1}
  987. %\end{figure}
  988. %
  989. %In this subsection we experiment on adjusting parameters of the model and observe how system design $x$ changed accordingly.  we consider the following setting: a series-parallel system with  subsystems ($|\mathbf{N}|=3$), with each subsystem containing 3 types of components ($|\mathbf{M}_i|= 3, \forall i \in \mathbf{N}$), among which 1 types are active-parallel and 2 types are cold-standby, and each type is of 3 redundant components ($|\mathbf{T}_{ij}|=3, \forall i \in \mathbf{N}, j\in \mathbf{M}_i$). The lifetime requirement $\mathcal{T}_R = 29$.  The parameters we are adjusting includes $K \in [1, 10], R_0 \in \{0.95, 0.97, 0.99\}$, $\epsilon_{ij} \in \{0.05, 0.075, 0.1\},\forall i \in [3], j \in [3]$, and $\Delta$ to adjust $\underline{\hmu}$ and $\overline{\hmu}$:.
  990. %$$
  991. %\underline{\mu}^{k}_{ij}={\nu}^{k}_{ij}-\Delta,
  992. %\quad \overline{\mu}^{k}_{ij}={\nu}^{k}_{ij}+\Delta,
  993. %$$
  994. %where $\Delta \in \{0.025, 0.05, 0.075, 0.1\}$.  Therefore, by changing the values of $\Delta$, we can have different sets of parameters $\underline{\hmu}$ and $\overline{\hmu}$. Specifically, large values of $\Delta$ correspond to the large gaps between $\underline{\hmu}$ and $\overline{\hmu}$.
  995. %
  996. %\begin{table}[!htbp]\label{d-table}\scriptsize%\footnotesize%\small%
  997. %\caption{ \footnotesize  The design table for different $K$ under $\mathcal{T}_R = 29$, $\epsilon = 0.05$. AP, CS-I and CS-II represents the active parallel type, cold-standby type 1 and cold-standby type 2, respectively.}
  998. %\begin{center}
  999. %\begin{adjustbox}{angle=270}
  1000. %\scalebox{1}{
  1001. %\begin{tabular}{|c|c|c| c|| ccccc|| ccccc|| ccccc||}\hline
  1002. % \multirow{2}{*}{$\epsilon$} & \multirow{2}{*}{$\Delta$} & \multirow{2}{*}{Subsystem} & \multirow{2}{*}{Type} & \multicolumn{5}{c||}{$R_0 = 0.95, K$} & \multicolumn{5}{c||}{$R_0 = 0.97, K$} & \multicolumn{5}{c||}{$R_0 = 0.99, K$}\\
  1003. % \cline{5-19}
  1004. %      &&&&1 & 3 & 5 &  8 &  10 &1 & 3 & 5 &  8 &  10 & 1 & 3 & 5 &  8 &  10  \\
  1005. %     \hline
  1006. %                    &&& AP & 1 & 1 & 1 &1 & 1 & 0 &0 & 0 & 0 & 1 & 1 & 1 & 1 &1 & 1\\
  1007. %        &&1         & CS-l & 1 & 1 & 1 &1 & 1 & 2 & 2 & 2 & 2& 1 & 2 & 2 & 2 &2 & 2\\
  1008. %                    &&& CS-ll & 0 & 0 & 0 &0 & 0 & 0 & 0 & 0 &0 & 0& 0 & 0 & 0 &0 & 0\\
  1009. %     \cline{3-19}
  1010. %                    &&& AP & 1 & 1 & 1 &1 & 1 & 1 & 1 & 1 &1 & 1& 1 & 1 & 1 &1 & 1\\
  1011. %        &0.025&2    & CS-l & 0& 0 & 0 & 0 & 1 & 0 & 0 & 0 &0& 0& 0 & 0 & 0 &0 & 0\\
  1012. %                    &&& CS-ll & 2 & 2 & 2 &2 & 2 & 2 & 2 & 2 &2 & 2& 2 & 2 & 2 &2 & 2\\
  1013. %     \cline{3-19}
  1014. %                    &&& AP &0 & 0 & 1& 1 & 1 & 0 &1 & 1 &1 & 1 & 1 &0 &0 &0 & 0\\
  1015. %        &&3         & CS-l & 1 & 1& 0& 0 & 0 & 1 & 0 & 0& 0 & 1 & 1 & 1 & 1 &1 & 1\\
  1016. %                    &&& CS-ll & 3 & 3 & 3 &3 & 3&  3  & 3 & 3& 3 & 3& 3 & 3 & 3 &3 & 3\\
  1017. %\cline{2-19}
  1018. %\multicolumn{1}{|c|}{} & \multicolumn{3}{c||}{Design cost (k\$)} & 39.5& 39.5&\!\! 39    &39& 39 & 44.5&44&44&\!\!44 & 41 &55 & 54.5 &\!\! 54.5 &54.5& 54.5 \\
  1019. %     \cline{2-19}
  1020. %                    &&& AP& 1 & 1 & 1 &1 & 1 & 0 & 0 & 0 &0 & 0 & 1& 1 & 1 & 1 &1\\
  1021. %        &&1         & CS-l& 1 & 1 & 1 &1 & 1 & 2 & 2 & 2 &2 & 2& 2 & 2 & 2 &2 & 2 \\
  1022. %                    &&& CS-ll & 0 & 0 & 0 &0 & 0 & 0 & 0 & 0 &0 & 0& 0 & 0 & 0 &0& 0\\
  1023. %     \cline{3-19}
  1024. %                    &&& AP & 1 & 1 & 1 &1 & 1 & 1 & 1 & 1 &1 & 1 & 1 & 1 & 1 &1 & 1 \\
  1025. %        &0.05&2     & CS-l &1 & 0 & 0 &0 & 0 & 0 & 0 & 0 &0 & 0& 0 & 0 & 0 &0 & 0\\
  1026. %                    &&& CS-ll & 2 & 2 & 2 &2 & 2 & 2 & 2 &2 &2 &2& 2 & 2 & 2 &2 &2\\
  1027. %     \cline{3-19}
  1028. %                    &&& AP & 1 & 1 & 1 &1 & 0 & 1& 1 & 0& 0 & 0 & 1 &1 & 1 &1 & 1\\
  1029. %        &&3         & CS-l & 3 & 3 & 3 &3 & 1 & 3 & 3 & 1& 1 & 1 &  3 & 3 & 3 & 3 & 3\\
  1030. %                    &&& CS-ll & 1 & 1 & 1 &1 & 3 & 1 & 1& 3& 3 & 3 & 1 & 1 & 1 &1 & 1\\
  1031. %\cline{2-19}
  1032. %\multicolumn{1}{|c|}{0.05} & \multicolumn{3}{c||}{Design cost (k\$)} & 40&40 &\!\! 40  &40& 39.5 & 45&45&\!\! 44.5 & 44.5 & 44.5 & 55&55&\!\! 55  & 55 & 55\\
  1033. %     \cline{2-19}
  1034. %                     &&& AP& 1& 1 & 1 & 1 &1 & 0 & 0 & 0 &0 & 0 & 1& 1 & 1 & 1 &1\\
  1035. %        &&1           &CS-l2 & 1& 1 & 1 & 1 &1 & 2 & 2 & 2 &2 & 2& 2 & 2 & 2 &2 & 2\\
  1036. %                    &&& CS-ll & 0 & 0 & 0 &0 & 0 & 0 & 0 & 0 &0 & 0& 0 & 0 & 0 &0 & 0 \\
  1037. %     \cline{3-19}
  1038. %                    &&& AP & 1& 1 & 1 & 1 &1 & 1 &1 & 1 & 1 &1 & 1& 1 & 1 & 1 &1  \\
  1039. %        &0.075&2    & CS-l & 0 & 0 & 0 &0 & 0 & 0 & 0 & 0 &0 & 0& 0 & 0 & 0 &0 & 0 \\
  1040. %                    &&& CS-ll & 2 & 2 & 2 &2 & 2 & 2 & 2 & 2 &2 & 2& 2 & 2 & 2 &2 & 2 \\
  1041. %     \cline{3-19}
  1042. %                    &&& AP & 1& 1 & 1 & 1 &1  & 1 &1 & 1 & 1 &1 & 1& 1 & 1 & 1 &1\\
  1043. %        &&3           & CS-l & 3 & 3 & 3 & 3 & 3 & 3 & 3 & 3 &3 & 3& 3 & 3 & 3 &3 & 3\\
  1044. %                    &&& CS-ll & 1& 1 & 1 & 1 &1  & 1 &1 & 1 & 1 &1 & 1& 1 & 1 & 1 &1\\
  1045. %\cline{2-19}
  1046. %\multicolumn{1}{|c|}{} & \multicolumn{3}{c||}{Design cost (k\$)} & 40&40&\!\! 40 & 40 & 40 & 45&45&\!\! 45& 45&45 & 55&55 &\!\! 55    & 55& 55\\
  1047. %
  1048. %     \cline{2-19}
  1049. %                    &&& AP& 1& 1 & 1 & 1 &1 & 0 & 0 & 0 &0 & 0 & 1& 1 & 1 & 1 &1\\
  1050. %
  1051. %        &&1         & CS-l & 1& 1 & 1 & 1 &1 & 2 & 2 & 2 &2 & 2& 2 & 2 & 2 &2 & 2\\
  1052. %                    &&&CS-ll3 & 0 & 0 & 0 &0 & 0 & 0 & 0 & 0 &0 & 0& 0 & 0 & 0 &0 & 0\\
  1053. %     \cline{3-19}
  1054. %                    &&& AP & 1& 1 & 1 & 1 &1  &  1 &1 & 1 & 1 &1 & 1& 1 & 1 & 1 &1 \\
  1055. %        &0.1&2          & CS-l & 0 & 0 & 0 &0 & 0 & 0 & 0 & 0 &0 & 0& 0 & 0 & 0 &0 & 0 \\
  1056. %                    &&& CS-ll & 2 & 2 & 2 &2 & 2 & 2 & 2 & 2 &2 & 2& 2 & 2 & 2 &2 & 2\\
  1057. %     \cline{3-19}
  1058. %                    &&& AP &1 & 1& 1 & 1 & 1 & 1  & 1 & 1 & 1 &1 & 1& 1 & 1 & 1 &1 \\
  1059. %        &&3         & CS-l & 3 & 3 & 3 &3 & 3 & 3 & 3 & 3 &3 & 3& 3 & 3 & 3 &3 & 3\\
  1060. %                    &&& CS-ll& 1& 1 & 1 & 1 &1  & 1 & 1 & 1 & 1 &1 & 1& 1 & 1 & 1 & 1 \\
  1061. %\cline{2-19}
  1062. %\multicolumn{1}{|c|}{} & \multicolumn{3}{c||}{Design cost (k\$)} & 40&\!\! 40   & 40&40 & 40 & 45&45&\!\! 45 &45& 45 & 55&55&\!\! 55 & 55 & 55 \\
  1063. %\hline
  1064. %\end{tabular}}
  1065. %\end{adjustbox}
  1066. %\end{center}
  1067. %\end{table}
  1068. %
  1069. %The resulting system design in the case of $\epsilon_{ij} = 0.05$ are shown in the table. We leave the rest of experiment results in the Electronic Companion. Note that when $K = 1$, the model reduces to the case of robust model without clustering by Wang, et. al.
  1070. %
  1071. %The observations of the experimental results are the following:
  1072. %(i) the cost increases when the variation range $[\underline{\hmu}, \overline{\hmu}]$ of expected lifetimes increases, or the dispersion parameter $\epsilon$ of the lifetimes of components enlarges. Such increased cost is due to the enlarged ambiguity set $\mathbb{F}_K$ resulting from the change of above distributional parameters $([\underline{\hmu}, \overline{\hmu}]$. (ii) Intuitively, cost also increases if the required reliability level $R_0$ is increased.
  1073. %(iii) the cost decreases as number of clusters $K$ increases. This is due to to the fact that since the mean and dispersion information of each cluster are included in the ambiguity set, more clusters implies more information and therefore smaller ambiguity set.
  1074. %
  1075. %We then perform out-of-sample tests on the designs $K = 1, 4, 8$. To test the robustness of the design, we generate testing data with smaller lifetime mean and much larger standard deviation. In particular, we let $\hmu_{test} = (1 - \Delta_{m})\hmu_{train}$, and $\hsigma_{test} = \Delta_{s}\hsigma_{train}$. $\Delta_m$ is set to 10\%, and $\Delta_s$ is set to 2000\%, 3000\% and 4000\%, respectively.
  1076. %
  1077. %\begin{table}[htp]\footnotesize%\small%
  1078. %\caption{\label{tab-compare2} \footnotesize The out of sample reliability of designs generated by $K = 1, 4, 8$ models under different $\Delta_{s}$. $R_0$ is set to $0.9$. In the design columns, the 3 columns are corresponding to different subsystems. The three numbers in each column indicates the number of redundancies used that is active parallel, cold-standby type 1, or cold-standby type 2, respectively. }
  1079. %\begin{center}
  1080. %\begin{tabular}{|c||c|c|| c|c|c| c |  c |c   |}\hline
  1081. %\multirow{2}{*}{$\mathcal{T}_R$}  & \multirow{2}{*}{$(\Delta_{m}, \Delta_{s})$} & \multirow{2}{*}{Model} & \multicolumn{3}{c|}{Design} & \multirow{2}{*}{cost} & \multirow{2}{*}{Mean of out-of-sample reliability level} & \multirow{2}{*}{StD} \\
  1082. %\cline{4-6}
  1083. %&&& 1 & 2 & 3 &&&\\
  1084. %\hline
  1085. %\multirow{9}{*}{25} & \multirow{3}{*}{(10\%, 2000\%)} &  K=1 Model & (1,1,0) & (1,2,0) & (1,0,3)& 38.5 &0.999 & 0.035   \\
  1086. %     & & K=4 Model & (1,1,0) & (1,2,0) & (0,3,1)& 38.0 &0.966 & 0.182 \\
  1087. %     & & K=8 Model & (1,1,0) & (1,2,0) & (0,0,3)& 37.0 &0.607 & 0.489  \\
  1088. %    \cline{2-9}
  1089. %     &\multirow{3}{*}{(10\%, 3000\%)}  &  K=1 Model & (1,1,0) & (1,2,0) & (1,0,3)& 38.5 &  0.989 & 0.103   \\
  1090. %     & & K=4 Model & (1,1,0) & (1,2,0) & (0,3,1)& 38.0 &0.925 & 0.263 \\
  1091. %     & & K=8 Model & (1,1,0) & (1,2,0) & (0,0,3)& 37.0 &0.576 & 0.494  \\
  1092. %    \cline{2-9}
  1093. %     &\multirow{3}{*}{(10\%, 4000\%)}  &  K=1 Model & (1,1,0) & (1,2,0) & (1,0,3)& 38.5 &  0.975 & 0.157   \\
  1094. %     & & K=4 Model & (1,1,0) & (1,2,0) & (0,3,1)& 38.0 &0.894 & 0.308 \\
  1095. %     & & K=8 Model & (1,1,0) & (1,2,0) & (0,0,3)& 37.0 &0.551 & 0.497  \\
  1096. %\hline
  1097. %
  1098. %\end{tabular}
  1099. %\end{center}
  1100. %\end{table}
  1101. %
  1102. %\begin{figure}[H]
  1103. %\centering
  1104. %\includegraphics[width=\columnwidth]{Out_of_sample_K.png}
  1105. %\caption{\footnotesize Figure (a)(b)(c) represents the out of sample reliability of designs generated by $K = 1, 4, 8$ models under different $\Delta_{s}$, respectively. The vertical beam represents $\mathcal{T}_R$. The fraction of the lifetime histogram on the right side of beam represents the out-of-sample reliability level.}
  1106. %\label{figureK}
  1107. %\end{figure}
  1108. %
  1109. %From the results, we can observe that with moderate number of clusters ($K = 4$), we can obtain designs with less costs than designs generated by robust model with no clustering ($K = 1$), while retaining robustness even when there is a significant shrink in mean lifetime and much larger standard deviation. This shows that our framework can produce designs that are robust enough and cheaper, by the incorporation of clustering. However, if the number of clusters becomes too large ($K = 8$), the resulting model lost robustness in these extreme tests. Thus, it is crucial to choose the optimal $K$. We present an experiment of choosing $K$ by cross validation in the next subsection.
  1110. %
  1111. %\subsection{Choosing $K$ by cross validation}
  1112. %In this subsection, we present a cross validation experiment by applying Algorithm 2. In particular, we choose $m = 10$ and do a 10-fold cross validation. The number of instances of constraints violation, as well as the cost of designs, are plotted in the figure below on the left.  We also offers a combined metric for cross validation. First, the cost and number of validation are both normalized to the range $[0, 1]$. Then, compute $(1-\lambda)cost(K) + \lambda{{\#}violation(K)}$, where $\lambda \in [0,1]$. By assigning different $\lambda$, we can adapt to scenarios in of different cost-violation tradeoffs. In particular, high $\lambda$ means that robustness of the design is more valued than the cost; low $\lambda$ indicates the contrary. The combined metric under different $\lambda$ are plotted in the figure below on the right. Observe that when $\lambda$ is low, large $K$ such as 7 and 9 are preferred; when $\lambda$ is high, moderate $K$ like 5 are better. Since large $K$ generally correspond to less cost, this result matches the intuition that people are willing to pay more cost if robustness is more valued.
  1113. % %least constraint violation occurs, so $K = 5$ is the ideal parameter to cluster this data set. $K = 5$ will be used in the following subsections.
  1114. %\iffalse
  1115. %\begin{figure}[H]
  1116. %\begin{subfigure}{0.5\textwidth}
  1117. %\includegraphics[scale=0.65]{cross_validation.png}
  1118. %\caption{\footnotesize}
  1119. %\label{figure4-1}
  1120. %\end{subfigure}
  1121. %\begin{subfigure}{0.5\textwidth}
  1122. %\centering
  1123. %\includegraphics[scale=0.37]{cross_validation_lambda.png}
  1124. %\caption{\footnotesize }
  1125. %\label{figure4-2}
  1126. %\end{subfigure}
  1127. %\caption{\footnotesize (a) The number of violations and costs with different $K$. (b) Costs penalized by $\lambda$ with different $K$, with each line associated with a different $\lambda$.}
  1128. %\end{figure}
  1129. %\fi
  1130. %
  1131. %\begin{figure}[H]
  1132. %\centering
  1133. %\includegraphics[width=\columnwidth]{cv.jpg}
  1134. %\caption{\footnotesize (a) The number of violations and costs with different $K$. (b) Costs penalized by $\lambda$ with different $K$, with each line associated with a different $\lambda$.}
  1135. %\label{figure4-2}
  1136. %\end{figure}
  1137. %
  1138. %
  1139. %
  1140. %
  1141. %\iffalse
  1142. %\begin{table}[h*]\footnotesize%\small%
  1143. %\caption{\label{tab-compare2} \footnotesize  Out-of-sample reliability level comparison ($R_0=0.85$), where `Design' specifies the number of components allocated in each of 3 subsystems, and `P-Model' and `C-DRO-Model' refer to the probabilistic model and robust model, respectively. }
  1144. %\begin{center}
  1145. %\begin{tabular}{|c||c|c|| c| c |c |c   |}\hline
  1146. %    $\mathcal{T}_R$ & Out of sample $\sigma$ &Model & Design &   Designed reliability level &Mean of out-of-sample reliability level & StD \\
  1147. %     \hline
  1148. %    \multirow{6}{*}{7.625} & \multirow{2}{*}{2}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.85$}& 0.834 & 0.372   \\
  1149. %     && C-DRO-Model & (1,1,3) &  &0.997 & 0.053  \\
  1150. %        \cline{2-7}
  1151. %     &\multirow{2}{*}{6}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.85$}& 0.724 & 0.447   \\
  1152. %     && C-DRO-Model & (1,1,3) &  &0.982 & 0.132  \\
  1153. %        \cline{2-7}
  1154. %     &\multirow{2}{*}{10}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.85$}& 0.673 & 0.469   \\
  1155. %     && C-DRO-Model & (1,1,3) &  &0.971 & 0.168  \\
  1156. %\hline
  1157. %    \multirow{6}{*}{7.75} & \multirow{2}{*}{2}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.85$}& 0.781 & 0.414   \\
  1158. %     && C-DRO-Model & (1,1,4) &  &0.999 & 0.028  \\
  1159. %                 \cline{2-7}
  1160. %     &\multirow{2}{*}{6}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.85$}& 0.690 & 0.462   \\
  1161. %     && C-DRO-Model & (1,1,4) &  &0.991 & 0.095  \\
  1162. %                  \cline{2-7}
  1163. %     &\multirow{2}{*}{10}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.85$}& 0.637 & 0.481   \\
  1164. %     && C-DRO-Model & (1,1,4) &  &0.989 & 0.105  \\
  1165. %\hline
  1166. %    \multirow{6}{*}{7.875} & \multirow{2}{*}{2}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.85$}& 0.716 & 0.451   \\
  1167. %     && C-DRO-Model & (1,1,7) &  &1.000 & 0.000  \\
  1168. %               \cline{2-7}
  1169. %     &\multirow{2}{*}{6}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.85$}& 0.640 & 0.480   \\
  1170. %     && C-DRO-Model & (1,1,7) &  &0.998 & 0.040  \\
  1171. %               \cline{2-7}
  1172. %     &\multirow{2}{*}{10}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.85$}& 0.602 & 0.490   \\
  1173. %     && C-DRO-Model & (1,1,7) &  &0.999 & 0.035  \\
  1174. %\hline
  1175. %\end{tabular}
  1176. %\end{center}
  1177. %\end{table}
  1178. %
  1179. %\begin{table}[h*]\footnotesize%\small%
  1180. %\caption{\label{tab-compare2} \footnotesize  Out-of-sample reliability level comparison ($R_0=0.90$), where `Design' specifies the number of components allocated in each of 3 subsystems, and `P-Model' and `C-DRO-Model' refer to the probabilistic model and robust model, respectively. }
  1181. %\begin{center}
  1182. %\begin{tabular}{|c||c|c|| c| c |c |c   |}\hline
  1183. %    $\mathcal{T}_R$ & Out of sample $\sigma$ &Model & Design &   Designed reliability level &Mean of out-of-sample reliability level & StD \\
  1184. %     \hline
  1185. %    \multirow{6}{*}{7.625} & \multirow{2}{*}{2}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.618 & 0.486   \\
  1186. %     && C-DRO-Model & (1,1,3) &  &0.998 & 0.047  \\
  1187. %        \cline{2-7}
  1188. %     &\multirow{2}{*}{6}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.576 & 0.494   \\
  1189. %     && C-DRO-Model & (1,1,3) &  &0.989 & 0.103   \\
  1190. %        \cline{2-7}
  1191. %     &\multirow{2}{*}{10}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.560 & 0.496   \\
  1192. %     && C-DRO-Model & (1,1,3) &  &0.988 & 0.111   \\
  1193. %\hline
  1194. %    \multirow{6}{*}{7.75} & \multirow{2}{*}{2}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.528 & 0.499   \\
  1195. %     && C-DRO-Model & (1,1,4) &  &0.999 & 0.037   \\
  1196. %                 \cline{2-7}
  1197. %     &\multirow{2}{*}{6}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.538 & 0.499   \\
  1198. %     && C-DRO-Model & (1,1,4) &  &0.996 & 0.063   \\
  1199. %                  \cline{2-7}
  1200. %     &\multirow{2}{*}{10}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.530 & 0.499   \\
  1201. %     && C-DRO-Model & (1,1,4) &  &0.994 & 0.080   \\
  1202. %\hline
  1203. %    \multirow{6}{*}{7.875} & \multirow{2}{*}{2}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.458 & 0.498   \\
  1204. %     && C-DRO-Model & (1,1,7) &  &0.996 & 0.060  \\
  1205. %               \cline{2-7}
  1206. %     &\multirow{2}{*}{6}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.481 & 0.500   \\
  1207. %     && C-DRO-Model & (1,1,7) &  &0.996 & 0.060   \\
  1208. %               \cline{2-7}
  1209. %     &\multirow{2}{*}{10}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.496 & 0.500   \\
  1210. %     && C-DRO-Model & (1,1,7) &  &0.995 & 0.069  \\
  1211. %\hline
  1212. %\end{tabular}
  1213. %\end{center}
  1214. %\end{table}
  1215. %
  1216. %\begin{table}[h*]\footnotesize%\small%
  1217. %\caption{\label{tab-compare2} \footnotesize  Out-of-sample reliability level comparison ($R_0=0.90$), where `Design' specifies the number of components allocated in each of 3 subsystems, and `P-Model' and `C-DRO-Model' refer to the probabilistic model and robust model, respectively. }
  1218. %\begin{center}
  1219. %\begin{tabular}{|c||c|c|| c| c |c |c   |}\hline
  1220. %    $\mathcal{T}_R$ & Out of sample $\sigma$ &Model & Design &   Designed reliability level &Mean of out-of-sample reliability level & StD \\
  1221. %     \hline
  1222. %    \multirow{6}{*}{7.625} & \multirow{2}{*}{2}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.143 & 0.350   \\
  1223. %     && C-DRO-Model & (1,1,3) &  &0.653 & 0.476  \\
  1224. %        \cline{2-7}
  1225. %     &\multirow{2}{*}{6}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.576 & 0.494   \\
  1226. %     && C-DRO-Model & (1,1,3) &  &0.989 & 0.103   \\
  1227. %        \cline{2-7}
  1228. %     &\multirow{2}{*}{10}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.560 & 0.496   \\
  1229. %     && C-DRO-Model & (1,1,3) &  &0.988 & 0.111   \\
  1230. %\hline
  1231. %    \multirow{6}{*}{7.75} & \multirow{2}{*}{2}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.528 & 0.499   \\
  1232. %     && C-DRO-Model & (1,1,4) &  &0.999 & 0.037   \\
  1233. %                 \cline{2-7}
  1234. %     &\multirow{2}{*}{6}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.538 & 0.499   \\
  1235. %     && C-DRO-Model & (1,1,4) &  &0.996 & 0.063   \\
  1236. %                  \cline{2-7}
  1237. %     &\multirow{2}{*}{10}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.530 & 0.499   \\
  1238. %     && C-DRO-Model & (1,1,4) &  &0.994 & 0.080   \\
  1239. %\hline
  1240. %    \multirow{6}{*}{7.875} & \multirow{2}{*}{2}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.458 & 0.498   \\
  1241. %     && C-DRO-Model & (1,1,7) &  &0.996 & 0.060  \\
  1242. %               \cline{2-7}
  1243. %     &\multirow{2}{*}{6}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.481 & 0.500   \\
  1244. %     && C-DRO-Model & (1,1,7) &  &0.996 & 0.060   \\
  1245. %               \cline{2-7}
  1246. %     &\multirow{2}{*}{10}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.496 & 0.500   \\
  1247. %     && C-DRO-Model & (1,1,7) &  &0.995 & 0.069  \\
  1248. %\hline
  1249. %\end{tabular}
  1250. %\end{center}
  1251. %\end{table}
  1252. %
  1253. %\begin{table}[h*]\footnotesize%\small%
  1254. %\caption{\label{tab-compare2} \footnotesize  Out-of-sample reliability level comparison ($R_0=0.90$), where `Design' specifies the number of components allocated in each of 3 subsystems, and `P-Model' and `C-DRO-Model' refer to the probabilistic model and robust model, respectively. }
  1255. %\begin{center}
  1256. %\begin{tabular}{|c||c|c|| c| c |c |c   |}\hline
  1257. %    $\mathcal{T}_R$ & Out of sample $\sigma$ &Model & Design &   Designed reliability level &Mean of out-of-sample reliability level & StD \\
  1258. %     \hline
  1259. %    \multirow{6}{*}{7.625} & \multirow{2}{*}{2}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.348 & 0.476   \\
  1260. %     && C-DRO-Model & (1,1,3) &  &0.930 & 0.256  \\
  1261. %        \cline{2-7}
  1262. %     & \multirow{2}{*}{6}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.427 & 0.495   \\
  1263. %     && C-DRO-Model & (1,1,3) &  &0.948 & 0.221  \\
  1264. %        \cline{2-7}
  1265. %     &\multirow{2}{*}{10}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.445 & 0.497   \\
  1266. %     && C-DRO-Model & (1,1,3) &  &0.957 & 0.203   \\
  1267. %\hline
  1268. %    \multirow{6}{*}{7.75} & \multirow{2}{*}{2}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.278 & 0.448   \\
  1269. %     && C-DRO-Model & (1,1,4) &  &0.933 & 0.250   \\
  1270. %                 \cline{2-7}
  1271. %     &\multirow{2}{*}{6}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.381 & 0.486   \\
  1272. %     && C-DRO-Model & (1,1,4) &  &0.974 & 0.159   \\
  1273. %                 \cline{2-7}
  1274. %     &\multirow{2}{*}{10}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.409 & 0.492   \\
  1275. %     && C-DRO-Model & (1,1,4) &  &0.979 & 0.143   \\
  1276. %\hline
  1277. %    \multirow{6}{*}{7.875} & \multirow{2}{*}{2}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.214 & 0.410   \\
  1278. %     && C-DRO-Model & (1,1,7) &  &0.909 & 0.287  \\
  1279. %               \cline{2-7}
  1280. %     & \multirow{2}{*}{6}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.345 & 0.475   \\
  1281. %     && C-DRO-Model & (1,1,7) &  &0.973 & 0.163  \\
  1282. %               \cline{2-7}
  1283. %     &\multirow{2}{*}{10}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.383 & 0.486   \\
  1284. %     && C-DRO-Model & (1,1,7) &  &0.985 & 0.122   \\
  1285. %\hline
  1286. %\end{tabular}
  1287. %\end{center}
  1288. %\end{table}
  1289. %\fi
  1290. %
  1291. %\iffalse
  1292. %From the experiment result, we can observe that despite a smaller ambiguity set, our design ($x^{(2)}$) can achieve robustness level that is comparable to the design without clustering $x^{(1)}$, and are far better than the baseline probabilistic model ($x^{(3)}$).
  1293. %\fi
  1294. %
  1295. %\subsection{Comparison with a baseline probabilistic model}
  1296. %
  1297. %\begin{figure}[H]
  1298. %\centering
  1299. %\includegraphics[scale=0.8]{KVARIATION55.pdf}
  1300. %\caption{\footnotesize The series-parallel system we are studying in this section, consists of a single type of component, with active parallel strategy only. This simplicity is due to the limitation of baseline probabilistic model.}
  1301. %\label{figure1}
  1302. %\end{figure}
  1303. %
  1304. %To illustrate the performance of our robust reliability model, we compare the design ($x^{(1)}$) obtained from the proposed robust redundancy optimization model with the design ($x^{(2)}$) obtained from a probabilistic redundancy optimization model. We choose $K = 5$, correspond to $\lambda = 0.8$ in the previous subsection. As mentioned in the Introduction and Literature Review, when the situation involves multiple types ({\it i.e.,} $|\mathbf{M}_i|>1$), or both the cold-standby and active parallel redundant subsystems are considered, the probabilistic model generally becomes intractable. Therefore, for a fair comparison, we consider a series-parallel system with $|\mathbf{N}| = 3$ and $|\mathbf{M}_i|=1, \forall i \in [3]$, which preserves a linear MIP formulation for the probabilistic model. For a coherent exposition of the experimental study, we place the details of the probabilistic redundancy model as well as its MIP %transformation in the Electronic Companion.
  1305. %
  1306. %\iffalse
  1307. %In particular, we first randomly generate lifetime samples (size=2500) and then compute the probability levels $\P[\tilde{z}_{i}\le \mathcal{T}_R ], \forall i \in [3]$ and the parameters $(\hnu, \underline{\hmu}, \overline{\hmu}, \bm{\hsigma}, \p)$ from the generated lifetime samples for parameter inputs of probabilistic and robust models, respectively, where $R_0=0.95$ and $\mathcal{T}_R=7.5$. We obtain the designs by solving the perspective redundancy models. \fi We perform out-of-sample experiments in the similar way as the out-of-sample-test performed in Section 5.2. First, we keep the lifetime mean as the same ($\Delta_m = 0$), but increases the out-of-sample standard deviation ($\Delta_s = 500\%, 1000\%, 1500\%$, respectively). This corresponds to the first row of the figure. Then, we shrink the lifetime mean ($\Delta_m = 10\%, 12.5\%, 15\%$) as the same time as $\Delta_s$ enlarges. This corresponds to the second row of the figure. The out-of-sample system lifetimes are compared and plotted in figure, and the comparison of out-of-sample reliability levels is provided in table.
  1308. %\begin{table}[H]\footnotesize%\small%
  1309. %\caption{\label{tab-compare2} \footnotesize  Out-of-sample reliability level comparison ($R_0=0.95$), where `Design' specifies the number of components allocated in each of 3 subsystems, and `P-Model' and `C-DRO-Model' refer to the probabilistic model and robust model, respectively. }
  1310. %\begin{center}
  1311. %\begin{tabular}{|c||c|c|| c| c |c |c   |}\hline
  1312. %    $\mathcal{T}_R$ & $(\Delta_{m}, \Delta_{s})$ &Model & Design &   Designed reliability level &Mean of out-of-sample reliability level & StD \\
  1313. %     \hline
  1314. %    \multirow{12}{*}{7.5} & \multirow{2}{*}{(0, 500\%)}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.95$}& 0.952 & 0.214   \\
  1315. %     && C-DRO-Model & (1,2,8) &  &1.000 & 0.000  \\
  1316. %        \cline{2-7}
  1317. %     &\multirow{2}{*}{(0, 1000\%)}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.95$}& 0.886 & 0.318   \\
  1318. %     && C-DRO-Model & (1,2,8) &  &1.000 & 0.000  \\
  1319. %        \cline{2-7}
  1320. %     &\multirow{2}{*}{(0, 1500\%)}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.95$}& 0.800 & 0.400   \\
  1321. %     && C-DRO-Model & (1,2,8) &  &1.000 & 0.000  \\
  1322. %     \cline{2-7}
  1323. %     & \multirow{2}{*}{(10\%, 500\%)}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.95$}& 0.408 & 0.492   \\
  1324. %     && C-DRO-Model & (1,2,8) &  &0.995 & 0.069  \\
  1325. %        \cline{2-7}
  1326. %     &\multirow{2}{*}{(12.5\%, 1000\%)}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.95$}& 0.328 & 0.469   \\
  1327. %     && C-DRO-Model & (1,2,8) &  &0.970 & 0.172  \\
  1328. %        \cline{2-7}
  1329. %     &\multirow{2}{*}{(15\%, 1500\%)}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.95$}& 0.291 & 0.454   \\
  1330. %     && C-DRO-Model & (1,2,8) &  &0.959 & 0.198  \\
  1331. %\hline
  1332. %\end{tabular}
  1333. %\end{center}
  1334. %\end{table}
  1335. %
  1336. %\begin{figure}[H]
  1337. %\centering
  1338. %\includegraphics[width=\columnwidth]{Out_of_sample_def.png}
  1339. %\caption{\footnotesize The out-of-sample system lifetime scenarios of robust model with clustering and probabilistic model under different $\Delta_{m}$ and $\Delta_{s}$. Figure (a), (b), (c) represents the scenario that the mean of out-of-sample data is kept the same, with different $\Delta_{s}$ that is much larger than the training set; figure (d), (e), (f) represents the scenario that the mean of out-of-sample data shrinks ($\Delta_{m} > 0$), in additional to much larger $\Delta_{s}$. The vertical beam represents $\mathcal{T}_R$. The fraction of the lifetime histogram on the right side of beam represents the out-of-sample reliability level.}
  1340. %\label{figure_out_of_sample_test_def}
  1341. %\end{figure}
  1342. %
  1343. %We can observe that when the mean lifetime are kept the same, the robust model produces very robust designs, while the baseline probabilistic model produces design that works fine for moderate $\Delta_s$, but becomes unsatisfiable under larger $\Delta_s$; when the lifetime mean shrinks, however, the out-of-sample reliability of the design from baseline model becomes extremely awful, while the design from robust model can still keep the reliability level intact. This illustrates that our model outperforms the baseline model in that it is significantly more robust, especially under extreme circumstances. In addition, as mentioned before, our model is tractable for multi-type mixed strategy systems, while the probabilistic model will become intractable. Thus our model is superior in both robustness and computability.
  1344. %
  1345. %
  1346. %\subsection{Value of side information}
  1347. %In this subsection we experiment on clustering according to side information. The system we are studying is the same as the one in Section 5.2. We choose $K = 5$, corresponding to $\lambda = 0.8$ in the cross validation section. When generating samples from distributions, we also obtain the side information of which of the 5 distributions the sample is drawn from. We then cluster the data set by $K = 5$, based solely on the side information, and compute parameters $(\hnu, [\underline{\hmu}, \overline{\hmu}], \bm{\hsigma}, \p)$ of the model from it. We obtain design ($x^{(2)}$) from the model, and compare it with design ($x^{(1)}$) obtained from the model in which $K$ is also 5, but is directly clustered based on lifetime information instead of the side information.
  1348. %
  1349. %\begin{table}[H]\label{d-table}\scriptsize%\footnotesize%\small%
  1350. %\caption{ \footnotesize  The design table for $K = 5$ model with and without side information (S.I). AP, CS-I and CS-II represents the active parallel type, cold-standby type 1 and cold-standby type 2, respectively.}
  1351. %\begin{center}
  1352. %\begin{tabular}{|c|c| c|| ccccc || ccccc || ccccc |}\hline
  1353. %\multirow{2}{*}{S.I} & \multirow{2}{*}{Subsystem} & \multirow{2}{*}{Type} & \multicolumn{5}{c||}{$R_0=0.95$,~$\mathcal{T}_R$ (yrs) }& \multicolumn{5}{c||}{$R_0=0.97$,~$\mathcal{T}_R$ (yrs)}& \multicolumn{5}{c|}{$R_0=0.99$,~$\mathcal{T}_R$ (yrs)} \\
  1354. % \cline{4-18}
  1355. %      &&&28 & 28.5 & 29 &  29.5 & 30 &28 & 28.5 & 29 &  29.5 & 30 &28 & 28.5 & 29 &  29.5 & 30  \\
  1356. %     \cline{1-18}
  1357. %                    && AP & 1 & 1 &1 & 0 & 0  & 0 & 0 &0 & 0 & 0    & 1 & 1 & 1 & 1 & 1 \\                       &
  1358. %        1            & CS-l & 1 & 1 &1 & 2 & 2  & 2 & 2 &2 & 2 & 2    & 2 & 2 & 2 & 2 & 2 \\
  1359. %                    &&CS-ll & 0 & 0 &0 & 0 & 0  & 0 & 0 &0 & 0 & 0    & 0 & 0 & 0 & 0 & 0 \\
  1360. %     \cline{2-18}
  1361. %                    && AP  & 1 & 1 &1 & 1 & 1  & 1 & 1 &1 & 1 & 1    & 1 & 1 & 1 & 1 & 1 \\ With  &
  1362. %        2            & CS-l & 1 & 0 &0 & 0 & 0  & 1 & 0 &0 & 0 & 3    & 1 & 0 & 0 & 0 & 3 \\ S.I &
  1363. %                     & CS-ll & 1 & 2 &2 & 2 & 2  & 1 & 2 &2 & 2 & 0    & 1 & 2 & 2 & 2 & 0 \\
  1364. %     \cline{2-18}
  1365. %                    && AP & 1 & 0 &1 & 1 & 0  & 1 & 1 &0 & 1 & 1    & 0 & 0 & 0 & 1 & 1 \\                       &
  1366. %        3            & CS-l & 0 & 1 &3 & 0 & 1  & 0 & 0 &1 & 3 & 2    & 1 & 1 & 1 & 3 & 2 \\
  1367. %                    && CS-ll & 3 & 3 &1 & 3 & 3  & 3 & 3 &3 & 1 & 2    & 3 & 3 & 3 & 1 & 2 \\
  1368. %\hline
  1369. %\multicolumn{3}{|c||}{Design cost (k\$)} & 38.5& 39.5 &40&\!\! 44 &\!\! 44.5    & 43.5 & 44 &44.5& 45 & 46    & 54 & \!\! 54.5\!\! &54.5\!\!& 55\!\! &56 \\
  1370. %\hline
  1371. %\multicolumn{3}{|c||}{Cost saved (k\$)} & 1 & 4.5 &4&\!\! 0 &\!\! 1    & 0.5 & 1 &0.5& 1 & 1.5    & 0 & \!\! 0.5\!\! &0.5\!\!& 0\!\! &0 \\
  1372. %\cline{1-18}
  1373. %     \cline{2-18}
  1374. %                    && AP & 1 & 0 &0 & 0 & 0  & 0 & 0 &0 & 0 & 0    & 1 & 1 & 1 & 1 & 1 \\                          &
  1375. %        1            & CS-l & 1 & 2 &2 & 2 & 2  & 2 & 2 &2 & 2 & 2    & 2 & 2 & 2 & 2 & 2 \\
  1376. %                    && CS-ll & 0 & 0 &0 & 0 & 0  & 0 & 0 &0 & 0 & 0    & 0 & 0 & 0 & 0 & 0 \\
  1377. %     \cline{2-18}
  1378. %                    && AP & 1 & 1 &1 & 1 & 1  & 1 & 1 &1 & 1 & 1    & 1 & 1 & 1 & 1 & 1 \\ Without &
  1379. %        2            & CS-l & 1 & 2 &0 & 0 & 3  & 1 & 0 &0 & 2 & 2    & 1 & 0 & 0 & 0 & 3 \\ S.I &
  1380. %                     & CS-ll & 1 & 0 &2 & 2 & 0  & 1 & 2 &2 & 1 & 1    & 1 & 2 & 2 & 2 & 0 \\
  1381. %     \cline{2-18}
  1382. %                    && AP  & 1 & 1 &1 & 1 & 1  & 0 & 1 &1 & 1 & 0    & 0 & 1 & 1 & 1 & 1 \\                          &
  1383. %       3             & CS-l & 3 & 3 &0 & 0 & 3  & 1 & 3 &3 & 3 & 2    & 1 & 3 & 3 & 3 & 2 \\
  1384. %                    && CS-ll& 1 & 1 &3 & 3 & 1  & 3 & 1 &1 & 1 & 3    & 3 & 1 & 1 & 1 & 2 \\
  1385. %\hline
  1386. %\multicolumn{3}{|c||}{Design cost (k\$)} & 39.5& 44 &44&\!\! 44 &\!\! 45.5    & 44 & 45 &45& 46 & 47.5    & 54 & \!\! 55\!\! &55\!\!& 55\!\! &56 \\
  1387. %\hline
  1388. %\end{tabular}
  1389. %\end{center}
  1390. %\end{table}
  1391. %
  1392. %The result shows that when side information is incorporated, we can achieve a design with much lower cost. We then choose $\mathcal{T}_R = 28.5$, which correspond to the largest cost saved, and perform out-of-sample tests, in the similar way as previous experiments. We can observe that even with this significant cost saved, the design obtained by clustering by side information still performs well enough under mean shrink and large standard deviation.
  1393. %
  1394. %\begin{table}[htp]\footnotesize%\small%
  1395. %\caption{\label{tab-compare3} \footnotesize The out-of-sample result of designs obtained at $\mathcal{T}_R = 28.5$. $R_0$ is set to $0.9$. In the design columns, the 3 columns are corresponding to different subsystems. The three numbers in each column indicates the number of redundancies used that is active parallel, cold-standby type I, or cold-standby type II, respectively. }
  1396. %\begin{center}
  1397. %\begin{tabular}{|c||c|c|| c|c|c| c |  c |c  |}\hline
  1398. %\multirow{2}{*}{$\mathcal{T}_R$} & \multirow{2}{*}{($\Delta_{m}, \Delta_{s}$)} &\multirow{2}{*}{Model} & \multicolumn{3}{c|}{Design} & \multirow{2}{*}{cost}  & \multirow{2}{*}{Mean of out-of-sample reliability level} & \multirow{2}{*}{StD} \\
  1399. %\cline{4-6}
  1400. %&&& 1 & 2 & 3 &&&\\
  1401. %\hline
  1402. %\multirow{6}{*}{28.5} & \multirow{2}{*}{(5\%, 800\%)} &  With S.I & (1,1,0) & (1,0,2) & (0,1,3)& 39.5 &0.985 & 0.122   \\
  1403. %     & & Without S.I & (0,2,0) & (1,2,0) & (1,3,1)& 44  &1.000 & 0.000 \\
  1404. %    \cline{2-9}
  1405. %     &\multirow{2}{*}{(5\%, 1000\%)}  &  With S.I & (1,1,0) & (1,0,2) & (0,1,3)& 39.5 & 0.970 & 0.172   \\
  1406. %     & & Without S.I & (0,2,0) & (1,2,0) & (1,3,1)& 44  &0.999 & 0.028 \\
  1407. %    \cline{2-9}
  1408. %     &\multirow{2}{*}{(5\%, 1200\%)}  &  With S.I & (1,1,0) & (1,0,2) &(0,1,3)& 39.5 & 0.968 & 0.175   \\
  1409. %     & & Without S.I & (0,2,0) & (1,2,0) & (1,3,1)& 44  &0.996 & 0.060 \\
  1410. %\hline
  1411. %\end{tabular}
  1412. %\end{center}
  1413. %\end{table}
  1414. %
  1415. %\begin{figure}[H]
  1416. %\centering
  1417. %\includegraphics[width=\columnwidth]{Out_of_sample_side.png}
  1418. %\caption{\footnotesize The out-of-sample test result of designs obtained at $\mathcal{T}_R = 28.5$, with and without side information (S.I).  The vertical beam represents $\mathcal{T}_R$. The fraction of the lifetime histogram on the right side of beam represents the out-of-sample reliability level.}
  1419. %\label{figure4}
  1420. %\end{figure}
  1421. %
  1422. %
  1423. %
  1424. %\iffalse
  1425. %\begin{table}[H]\footnotesize%\small%
  1426. %\caption{\label{tab-compare2} \footnotesize  Out-of-sample reliability level comparison ($R_0=0.95$), where `Design' specifies the number of components allocated in each of 3 subsystems, and `P-Model' and `C-DRO-Model' refer to the probabilistic model and robust model, respectively. }
  1427. %\begin{center}
  1428. %\begin{tabular}{|c||c|c|| c| c |c |c   |}\hline
  1429. %    $\mathcal{T}_R$ & $\Delta_{m}, \Delta_{s}$ &Model & Design &   Designed reliability level &Mean of out-of-sample reliability level & StD \\
  1430. %     \hline
  1431. %    \multirow{12}{*}{7.5} & \multirow{2}{*}{1, 5}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.95$}& 0.952 & 0.214   \\
  1432. %     && C-DRO-Model & (1,2,8) &  &1.000 & 0.000  \\
  1433. %        \cline{2-7}
  1434. %     &\multirow{2}{*}{1, 10}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.95$}& 0.886 & 0.318   \\
  1435. %     && C-DRO-Model & (1,2,8) &  &1.000 & 0.000  \\
  1436. %        \cline{2-7}
  1437. %     &\multirow{2}{*}{1, 20}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.95$}& 0.800 & 0.400   \\
  1438. %     && C-DRO-Model & (1,2,8) &  &1.000 & 0.000  \\
  1439. %\hline
  1440. %\end{tabular}
  1441. %\end{center}
  1442. %\end{table}
  1443. %
  1444. %\begin{table}[!htbp]\label{d-table}\scriptsize%\footnotesize%\small%
  1445. %\caption{ \footnotesize  The design table for $K = 10$ model with side information}
  1446. %\begin{center}
  1447. %\begin{tabular}{|c| c|| ccccc || ccccc || ccccc |}\hline
  1448. %\multirow{2}{*}{Subsystem} & \multirow{2}{*}{Type} & \multicolumn{5}{c||}{$R_0=0.95$,~$\mathcal{T}_R$ (yrs) }& \multicolumn{5}{c||}{$R_0=0.97$,~$\mathcal{T}_R$ (yrs)}& \multicolumn{5}{c|}{$R_0=0.99$,~$\mathcal{T}_R$ (yrs)} \\
  1449. % \cline{3-17}
  1450. %      &&28 & 28.5 & 29 &  29.5 & 30 &28 & 28.5 & 29 &  29.5 & 30 &28 & 28.5 & 29 &  29.5 & 30  \\
  1451. %     \hline
  1452. %                    & 1 & 1 & 1 &1 & 0 & 0  & 0 & 0 &0 & 0 & 0    & 1 & 1 & 1 & 1 & 1 \\
  1453. %        1           & 2 & 1 & 1 &1 & 2 & 2  & 2 & 2 &2 & 2 & 2    & 2 & 2 & 2 & 2 & 2 \\
  1454. %                    & 3 & 0 & 0 &0 & 0 & 0  & 0 & 0 &0 & 0 & 0    & 0 & 0 & 0 & 0 & 0 \\
  1455. %     \hline
  1456. %                    & 1 & 1 & 1 &1 & 1 & 1  & 1 & 1 &1 & 1 & 1    & 1 & 1 & 1 & 1 & 1 \\
  1457. %        2           & 2 & 1 & 0 &0 & 0 & 0  & 1 & 0 &0 & 0 & 3    & 1 & 0 & 0 & 0 & 3 \\
  1458. %                    & 3 & 1 & 2 &2 & 2 & 2  & 1 & 2 &2 & 2 & 0    & 1 & 2 & 2 & 2 & 0 \\
  1459. %     \hline
  1460. %                    & 1 & 1 & 0 &1 & 1 & 0  & 1 & 1 &0 & 1 & 1    & 0 & 0 & 0 & 1 & 1 \\
  1461. %       3            & 2 & 0 & 1 &3 & 0 & 1  & 0 & 0 &1 & 3 & 2    & 1 & 1 & 1 & 3 & 2 \\
  1462. %                    & 3 & 3 & 3 &1 & 3 & 3  & 3 & 3 &3 & 1 & 2    & 3 & 3 & 3 & 1 & 2 \\
  1463. %\hline
  1464. %\multicolumn{2}{|c||}{Design cost (k\$)} & 38.5& 39.5 &40&\!\! 44 &\!\! 44.5    & 43.5 & 44 &44.5& 45 & 46    & 54 & \!\! 54.5\!\! &54.5\!\!& 55\!\! &56\\
  1465. %\hline
  1466. %\end{tabular}
  1467. %\end{center}
  1468. %\end{table}
  1469. %
  1470. %
  1471. %\begin{table}[!htbp]\label{d-table}\scriptsize%\footnotesize%\small%
  1472. %\caption{ \footnotesize  The design table for $K = 10$ model without side information}
  1473. %\begin{center}
  1474. %\begin{tabular}{|c| c|| ccccc || ccccc || ccccc |}\hline
  1475. %\multirow{2}{*}{Subsystem} & \multirow{2}{*}{Type} & \multicolumn{5}{c||}{$R_0=0.95$,~$\mathcal{T}_R$ (yrs) }& \multicolumn{5}{c||}{$R_0=0.97$,~$\mathcal{T}_R$ (yrs)}& \multicolumn{5}{c|}{$R_0=0.99$,~$\mathcal{T}_R$ (yrs)} \\
  1476. % \cline{3-17}
  1477. %      &&28 & 28.5 & 29 &  29.5 & 30 &28 & 28.5 & 29 &  29.5 & 30 &28 & 28.5 & 29 &  29.5 & 30  \\
  1478. %     \hline
  1479. %                    & 1 & 1 & 0 &0 & 0 & 0  & 0 & 0 &0 & 0 & 0    & 1 & 1 & 1 & 1 & 1 \\
  1480. %        1           & 2 & 1 & 2 &2 & 2 & 2  & 2 & 2 &2 & 2 & 2    & 2 & 2 & 2 & 2 & 2 \\
  1481. %                    & 3 & 0 & 0 &0 & 0 & 0  & 0 & 0 &0 & 0 & 0    & 0 & 0 & 0 & 0 & 0 \\
  1482. %     \hline
  1483. %                    & 1 & 1 & 1 &1 & 1 & 1  & 1 & 1 &1 & 1 & 1    & 1 & 1 & 1 & 1 & 1 \\
  1484. %        2           & 2 & 1 & 2 &0 & 0 & 3  & 1 & 0 &0 & 2 & 2    & 1 & 0 & 0 & 0 & 3 \\
  1485. %                    & 3 & 1 & 0 &2 & 2 & 0  & 1 & 2 &2 & 1 & 1    & 1 & 2 & 2 & 2 & 0 \\
  1486. %     \hline
  1487. %                    & 1 & 1 & 1 &1 & 1 & 1  & 0 & 1 &1 & 1 & 0    & 0 & 1 & 1 & 1 & 1 \\
  1488. %       3            & 2 & 3 & 3 &0 & 0 & 3  & 1 & 3 &3 & 3 & 2    & 1 & 3 & 3 & 3 & 2 \\
  1489. %                    & 3 & 1 & 1 &3 & 3 & 1  & 3 & 1 &1 & 1 & 3    & 3 & 1 & 1 & 1 & 2 \\
  1490. %\hline
  1491. %\multicolumn{2}{|c||}{Design cost (k\$)} & 39.5& 44 &44&\!\! 44 &\!\! 45.5    & 44 & 45 &45& 46 & 47.5    & 54 & \!\! 55\!\! &55\!\!& 55\!\! &56\\
  1492. %\hline
  1493. %\end{tabular}
  1494. %\end{center}
  1495. %\end{table}
  1496. %\fi
  1497. %
  1498. %\iffalse
  1499. %Data creating process finised
  1500. %Academic license - for non-commercial use only
  1501. %cost is 38.5003857478587 , \mathcal{T}_R is 26, R_0 is 0.95
  1502. %
  1503. %(((1.0, -0.0, -0.0), (1.0, -0.0, -0.0), (7.563683503866358e-06, 7.563683503866358e-06, 7.563683503866358e-06)), ((-0.0, -0.0, 1.0), (-0.0, 1.0, -0.0), (1.0, -0.0, -0.0)), ((-0.0, -0.0, -0.0), (1.0, -0.0, 1.0), (1.0, 1.0, -0.0)))
  1504. %cost is 43.0 , \mathcal{T}_R is 26, R_0 is 0.97
  1505. %
  1506. %(((-0.0, -0.0, -0.0), (-0.0, 1.0, 1.0), (-0.0, -0.0, -0.0)), ((-0.0, -0.0, 1.0), (-0.0, 1.0, 1.0), (-0.0, -0.0, -0.0)), ((-0.0, -0.0, -0.0), (1.0, 1.0, -0.0), (1.0, -0.0, 1.0)))
  1507. %cost is 53.0 , \mathcal{T}_R is 26, R_0 is 0.99
  1508. %
  1509. %(((-0.0, 1.0, -0.0), (-0.0, 1.0, 1.0), (-0.0, -0.0, -0.0)), ((-0.0, 1.0, -0.0), (-0.0, 1.0, 1.0), (-0.0, -0.0, -0.0)), ((-0.0, -0.0, -0.0), (1.0, 1.0, -0.0), (-0.0, 1.0, 1.0)))
  1510. %cost is 44.0 , \mathcal{T}_R is 27, R_0 is 0.95
  1511. %
  1512. %(((-0.0, -0.0, -0.0), (1.0, -0.0, 1.0), (0.0, -0.0, -0.0)), ((-0.0, 1.0, -0.0), (1.0, -0.0, -0.0), (1.0, -0.0, -0.0)), ((-0.0, -0.0, -0.0), (-0.0, -0.0, 1.0), (1.0, 1.0, 1.0)))
  1513. %cost is 44.0 , \mathcal{T}_R is 27, R_0 is 0.97
  1514. %
  1515. %(((-0.0, -0.0, -0.0), (1.0, 1.0, -0.0), (0.0, -0.0, -0.0)), ((-0.0, -0.0, 1.0), (-0.0, 1.0, -0.0), (0.0, 1.0, -0.0)), ((-0.0, -0.0, -0.0), (-0.0, 1.0, -0.0), (1.0, 1.0, 1.0)))
  1516. %cost is 54.0 , \mathcal{T}_R is 27, R_0 is 0.99
  1517. %
  1518. %(((-0.0, 1.0, -0.0), (-0.0, 1.0, 1.0), (0.0, -0.0, -0.0)), ((-0.0, 1.0, -0.0), (1.0, -0.0, -0.0), (-0.0, -0.0, 1.0)), ((-0.0, -0.0, -0.0), (-0.0, 1.0, -0.0), (1.0, 1.0, 1.0)))
  1519. %cost is 44.0 , \mathcal{T}_R is 28, R_0 is 0.95
  1520. %
  1521. %(((-0.0, -0.0, -0.0), (-0.0, 1.0, 1.0), (-0.0, -0.0, -0.0)), ((-0.0, 1.0, -0.0), (-0.0, 1.0, -0.0), (-0.0, 1.0, -0.0)), ((-0.0, -0.0, -0.0), (-0.0, -0.0, 1.0), (1.0, 1.0, 1.0)))
  1522. %cost is 45.00017123897305 , \mathcal{T}_R is 28, R_0 is 0.97
  1523. %
  1524. %(((5.707965768311441e-06, 5.707965768311441e-06, 5.707965768311441e-06), (1.0, 1.0, -0.0), (0.0, 0.0, -0.0)), ((0.0, 1.0, -0.0), (1.0, 1.0, 1.0), (-0.0, -0.0, -0.0)), ((-0.0, -0.0, 0.0), (-0.0, 1.0, -0.0), (1.0, 1.0, 1.0)))
  1525. %cost is 54.0 , \mathcal{T}_R is 28, R_0 is 0.99
  1526. %
  1527. %(((-0.0, -0.0, 1.0), (-0.0, 1.0, 1.0), (-0.0, -0.0, -0.0)), ((-0.0, 1.0, -0.0), (-0.0, 1.0, -0.0), (-0.0, 1.0, -0.0)), ((-0.0, -0.0, -0.0), (-0.0, -0.0, 1.0), (1.0, 1.0, 1.0)))
  1528. %cost is 45.0 , \mathcal{T}_R is 29, R_0 is 0.95
  1529. %
  1530. %(((-0.0, -0.0, -0.0), (1.0, -0.0, 1.0), (-0.0, -0.0, -0.0)), ((-0.0, -0.0, 1.0), (-0.0, -0.0, -0.0), (1.0, -0.0, 1.0)), ((-0.0, 1.0, -0.0), (1.0, 1.0, 1.0), (-0.0, -0.0, 1.0)))
  1531. %cost is 54.5 , \mathcal{T}_R is 29, R_0 is 0.97
  1532. %
  1533. %(((-0.0, -0.0, 1.0), (-0.0, 1.0, 1.0), (-0.0, -0.0, -0.0)), ((1.0, -0.0, -0.0), (-0.0, -0.0, -0.0), (1.0, 1.0, -0.0)), ((-0.0, -0.0, -0.0), (1.0, -0.0, -0.0), (1.0, 1.0, 1.0)))
  1534. %cost is 55.0 , \mathcal{T}_R is 29, R_0 is 0.99
  1535. %
  1536. %(((-0.0, -0.0, 1.0), (1.0, 1.0, -0.0), (-0.0, -0.0, -0.0)), ((-0.0, -0.0, 1.0), (-0.0, -0.0, -0.0), (1.0, -0.0, 1.0)), ((-0.0, 1.0, -0.0), (1.0, 1.0, 1.0), (-0.0, 1.0, -0.0)))
  1537. %cost is 55.0 , \mathcal{T}_R is 30, R_0 is 0.95
  1538. %
  1539. %(((-0.0, -0.0, 1.0), (1.0, 1.0, -0.0), (-0.0, -0.0, -0.0)), ((-0.0, -0.0, 1.0), (1.0, 1.0, 1.0), (-0.0, -0.0, -0.0)), ((-0.0, -0.0, -0.0), (1.0, -0.0, -0.0), (1.0, 1.0, 1.0)))
  1540. %cost is 55.5 , \mathcal{T}_R is 30, R_0 is 0.97
  1541. %
  1542. %(((-0.0, -0.0, 1.0), (-0.0, 1.0, 1.0), (0.0, -0.0, -0.0)), ((-0.0, -0.0, 1.0), (1.0, 1.0, 1.0), (-0.0, -0.0, -0.0)), ((-0.0, -0.0, 1.0), (1.0, 1.0, 1.0), (-0.0, 1.0, -0.0)))
  1543. %cost is 56.0 , \mathcal{T}_R is 30, R_0 is 0.99
  1544. %
  1545. %(((1.0, -0.0, -0.0), (1.0, 1.0, -0.0), (-0.0, -0.0, -0.0)), ((-0.0, -0.0, 1.0), (1.0, 1.0, 1.0), (-0.0, -0.0, -0.0)), ((1.0, 0.0, -0.0), (1.0, 1.0, -0.0), (-0.0, 1.0, 1.0)))
  1546. %\fi
  1547. %
  1548. %\subsection{Case study}
  1549. %
  1550. %
  1551. %\section{Conclusion}
  1552. %
  1553. %\begin{thebibliography}{}
  1554. %\bibitem{Aggarwal2012}Aggarwal, C. C., Y. Zhao, P. S. Yu. 2012. On Text Clustering with Side Information. {\em 2012 IEEE 28th International Conference on Data Engineering}, 894--904.
  1555. %
  1556. %\bibitem{Ardakan2014} Ardakan, M. A., A. Z. Hamadani. 2014. Reliability-redundancy allocation problem with cold-standbyredundancy strategy. {\em Simulation Modelling Practice and Theory}~{\bf 42}:107--118.
  1557. %
  1558. %\bibitem{Bertsimas2011}Bertsimas, D., M. Sim. 2004. The price of robustness. {\em Operations Research}~{\bf 52}(1):35--53.
  1559. %
  1560. %\bibitem{Bhunia2010} Bhunia, A. K., L. Sahoo,  D. Roy. 2010. Reliability stochastic optimization
  1561. %for a series system with interval component reliability via genetic
  1562. %algorithm.  {\em Appl.Math. Computat.}~{\bf 216}~(3): 929--939, 2010.
  1563. %
  1564. %
  1565. %\bibitem{Chern1992}Chern, M.S. 1992. On the computational complexity of reliability redundancy allocation in a series system. {\em Operations research letters}~{\bf 11}~(5):309--315.
  1566. %
  1567. %\bibitem{Cheng2009}Cheng, Z., X. Wang, C. Tian, F. Wang. 2009. Mission reliability simulation of High-speed EMU service braking system. {\em Proceedings of the 8th International Conference on Reliability, Maintainability and Safety} (ICRMS 2009), 253--256.
  1568. %
  1569. %\bibitem{Coit1996}Coit, D. W., A.E. Smith. 1996. Solving the redundancy allocation problem using a combined neural network/genetic algorithm approach. {\em Computers \& Operations Research}~{\bf 23}~(6):515--526.
  1570. %
  1571. %\bibitem{Coit1998}Coit, D. W., A.E. Smith. 1998. Redundancy allocation to maximize a lower percentile of the system time-to-failure distribution. {\em IEEE Transactions on Reliability}~ {\bf 47}~(1):79--87.
  1572. %
  1573. %\bibitem{Coit2003}Coit, D. W. 2003. Maximization of system reliability with a choice of redundancy strategies. {\em IIE Transactions}~{\bf 35}~(6):535-543.
  1574. %
  1575. %\bibitem{Coit2004} Coit, D.W.,  T. Jin,  N. Wattanapongsakorn. 2004. System optimization
  1576. %with component reliability estimation uncertainty: A multi-criteria approach. {\em IEEE Trans. Rel.}~ {\bf 53}~(3) :  369--380, 2004.
  1577. %
  1578. %
  1579. %
  1580. %
  1581. %\bibitem{Elsayed2012}Elsayed, E A. 2012. {\em Reliability Engineering}. 2nd Edition. Wiley.
  1582. %
  1583. %
  1584. %\bibitem{Govindan2017}Govindan, K., A. Jafarian, M.E. Azbari, T.M. Choi. 2017. Optimal bi-objective redundancy allocation for systems reliability and risk management. {\em IEEE Transactions on Cybernetics}~{\bf 46}~(8):1735--1748.
  1585. %
  1586. %
  1587. %
  1588. %%\bibitem{Lam2012}Lam SW, T.S. Ng, and M. Sim. (2012). Multiple objectives satisficing under uncertainty. To appear in Operations Research, 2012.
  1589. %
  1590. %%\bibitem{Lin2011}Lin J, Muthuraman K, Lawley M (2011) Optimal and approximate algorithms for sequential clinical scheduling with no-shows. {\it IIE Transactions on Healthcare Systems Engineering} 1:20--36.
  1591. %
  1592. % %\bibitem{McCarthy2000} McCarthy K, McGee HM, O'Boyle CA. 2000. Outpatient clinic waiting times and non-attendance as indicators of quality. {\it Psychology, Health and Medicine} 5: 287--293.
  1593. %
  1594. %\bibitem{Grani2017} Hanasusanto, G. A., V. Roitch, D. Kuhn, W. Wiesemann. 2017. Ambiguous joint chance constraints under mean and dispersion information. {\em Operations Research}~{\bf 65}~(3):715--767.
  1595. %
  1596. %
  1597. %
  1598. %
  1599. %
  1600. %\bibitem{Elegbede2003}Elegbede, A.C., C. Chu, K.H. Adjallah, F. Yalaoui. 2003. Reliability allocation through cost minimization. {\em IEEE Transactions on reliability}~{\bf 52}~(1):106--111.
  1601. %
  1602. %\bibitem{Feizollahi2012} Feizollahi, M.J., M. Modarres. 2012. The robust deviation redundancy allocation problem with interval component reliabilities. {\em IEEE Transactions on reliability}~{\bf 61}~(4):957--965.
  1603. %
  1604. %
  1605. %
  1606. %
  1607. %
  1608. %\bibitem{Feizollahi2014}Feizollahi, M.J., S. Ahmed, M. Modarres. 2014. The robust redundancy allocation problem in series-parallel systems with budgeted uncertainty.  {\em IEEE Transactions on reliability}~{\bf 63}~(1):239--250.
  1609. %
  1610. %\bibitem{Feizollahi2015} Feizollahi, M.J., R. Soltan, H. Feyzollahi. 2015. The robust cold standby redundancy allocation in series-parallel systems with budgeted uncertainty. {\em IEEE Transactions on reliability}~{\bf 64}~(2):799--806.
  1611. %
  1612. %\bibitem{Hasegawa1999}Hasegawa, I., Uchida, S. 1999. Braking systems. {\em Japan Railway and Transport Review}~{\bf 20}:52-59.
  1613. %
  1614. %\bibitem{Kulturel-Konak2003}Kulturel-Konak, S., A. Smith., D., Coit. 2003. Efficiently Solving the Redundancy Allocation Problem Using Tabu Search. {\em IIE Transactions.}~{\bf 35}:515--526.
  1615. %
  1616. %\bibitem{Kuo2001}Kuo, W., V.R. Prasad, F.A. Tillman, C.L. Hwang. 2001. {\em Optimal Reliability Design: Fundamentals and Applications.} Cambridge university press. Cambridge.
  1617. %
  1618. %\bibitem{Kuo2007}Kuo, W., R. Wan. 2007. Recent advances in optimal reliability allocation. {\em IEEE Transactions on Systems, Man, and Cybernetics-Part A: Systems and Humans}, {\bf 37}~(2):143-156.
  1619. %
  1620. %\bibitem{Li2014}Li, Y.F., Y. Ding, E. Zio. 2014. Random fuzzy extension of the universal generating function approach for the
  1621. %reliability assessment of multi-state systems under aleatory and epistemic uncertainties. {\em IEEE Transactions on Reliability}~{\bf 63}~(1):13--25.
  1622. %
  1623. %\bibitem{Li2011}Li, C.Y., X. Chen, X.S. Yi, J.Y. Tao. 2011. Interval-valued reliability analysis of multi-state systems. {\em IEEE Transactions on Reliability}~{\bf 60}~(1):323--330.
  1624. %
  1625. %\bibitem{Li2008} Li, X.,  X. Hu. 2008. Some new stochastic comparisons for redundancy
  1626. %allocations in series and parallel systems.~{\em  Statist. Probabil. Lett.}~{\bf 78}~(18): 3388--3394.
  1627. %
  1628. %\bibitem{Liang2004} Liang, Y., A. E. Smith. 2004 An ant colony optimization algorithm for the redundancy allocation problem (RAP). {\em IEEE Transactions on Reliability}~{\bf 53}~(3):417--423.
  1629. %
  1630. %\bibitem{Liao2014}Liao, L., F. K\"{o}ttig. 2014. Review of hybrid prognostics approaches for remaining useful life prediction of engineered systems, and an application to battery life prediction. {\em IEEE Transactions on Reliability}. {\bf 63}~(1):191--207.
  1631. %
  1632. %\bibitem{Liu2015 }Liu, H., Y. Fu. 2015. Clustering with Partition Level Side Information. {\em 2015 IEEE International Conference on Data Mining}, 877--882.
  1633. %
  1634. %\bibitem{Military1992}Military, U.S. 1992. Reliability prediction of electronic equipment. MIL-HDBK-217F Notice 1.
  1635. %
  1636. %\bibitem{Marseguerra2005} Marseguerra, M., E. Zio, L. Podofillini, D. W. Coit. 2005. Optimal design
  1637. %of reliable network systems in presence of uncertainty. {\em IEEE
  1638. %Trans. Rel.}~{\bf 54}~(2): 243--253.
  1639. %
  1640. %
  1641. %\bibitem{Ng2014} Ng, S. Y., Xing Y., K. L. Tsui. 2014. A naive Bayes model for robust remaining useful life prediction of lithium-ion battery. {\em Applied Energy}~{\bf 118}: 114-123.
  1642. %
  1643. %
  1644. %
  1645. %
  1646. %\bibitem{Prasad2001}Prasad, V. R., W. Kuo, K. O. Kim. 2001. Maximization of a percentile life of a series system through component redundancy allocation. {\em IIE Transactions}~{\bf 33}~(12):1071--1079.
  1647. %
  1648. %\bibitem{Pecht2008} Pecht, M. 2008. {\em Prognostics and Health Management of Electronics.} John Wiley \& Sons, Ltd.
  1649. %
  1650. %\bibitem{Shang2017} Shang, C., X. Huang, F. You. 2017. Data-driven robust optimization based on kernel learning. {\em Computers \& Chemical Engineering}~{\bf 106}:464--479.
  1651. %
  1652. %\bibitem{Shapiro2001}Shapiro~A.~2001.~On duality theory of conic linear problems. In {\em Semi-Infinite Programming}, chapter 7, 135--165, Kluwer Academic Publishers, 2001.
  1653. %
  1654. %\bibitem{Soltani2015}Soltani R., J. Safari, S.J. Sadjadi. 2015. Robust counterpart optimization for the redundancy allocation problem in series-parallel systems with component mixing under uncertainty. {\em Applied Mathematics \& Computation}~{\bf 271}~(C): 80--88.
  1655. %
  1656. %
  1657. %
  1658. %
  1659. %\bibitem{Sun2017} Sun, M. X., Y. F. Li, E. Zio. 2017. On the optimal redundancy allocation for multi-state series-parallel systems under epistemic uncertainty. {\em Reliability Engineering \& System Safety}. Accepted.
  1660. %
  1661. %\bibitem{Tang2014}Tang, S., C. Yu, X. Wang, X. Guo, X. Si. 2014. Remaining useful life prediction of lithium-ion batteries based on the wiener process with measurement error. {\em Energies}~{\bf 7}~(2):520--547.
  1662. %
  1663. %\bibitem{Tekiner-Mogulkoc2011}Tekiner-Mogulkoc, H., D. W. Coit. 2011. System reliability optimization
  1664. %considering uncertainty: Minimization of the coefficient of variation
  1665. %for series-parallel systems.~{\em  IEEE Trans. Rel.}~{\bf 60}~(30): 667--674, 2011.
  1666. %
  1667. %
  1668. %\bibitem{Wang2012} Wang, Y., L. Li, S. Huang, Q. Chang. 2012. Reliability and covariance estimation of weighted k-out-of-n multi-state Systems. {\em European Journal of Operational Research}~{\bf 221}:~138--147.
  1669. %
  1670. %\bibitem{Wang2019} Wang, S., Y. Li, T. Jia.  2019. Distributionally Robust Design for Redundancy Allocation. {\em INFORMS Journal on Computing}, in press.
  1671. %
  1672. %\bibitem{Wisemann2014} Wiesemann, W.,  D. Kuhn, M. Sim. 2014. Distributionally robust convex optimization. {\em Operations Research}~ {\bf 62} ~(6)~ 1358--1376.
  1673. %
  1674. %\bibitem{Xie2017} Xie, W., Ahmed, S. 2017. Distributionally robust chance constrained optimal power flow with renewables: A conic reformulation. {\em IEEE Transactions on Power Systems.} Accepted.
  1675. %
  1676. %\bibitem{Xing2002} Xing, E.P., A.Y, Ng, M.I., Jordan, S., Russell. 2002. Distance metric learning, with application to clustering with side-information. {\em Proceedings of the 15th International Conference on Neural Information Processing Systems}, ~521--528
  1677. %
  1678. %
  1679. %\bibitem{Yalaoui2005}Yalaoui, A., E.  Chatelet, C. Chu. 2005. A new dynamic programming method for reliability redundancy allocation in a parallel-series system. {\em IEEE transactions on reliability}.~{\bf 54}~(2):254--261.
  1680. %
  1681. %
  1682. %\bibitem{Zaretalab2015}Zaretalab, A., V. Hajipour, M. Sharifi, M. R. Shahriari. 2015. A knowledge-based archive multi-objective simulated annealing algorithm to optimize series-parallel system with choice of redundancy strategies. {\em Computers \& Industrial Engineering}~{\bf 80}:33-44.
  1683. %
  1684. %\bibitem{Zhao2003} Zhao, R., B. Liu. 2003. Stochastic programming models for general redundancy-optimization problems.~{\em IEEE Trans. Rel.}~{\bf 52}~(2): 181--191, 2003.
  1685. %
  1686. %\bibitem{Zhao2011} Zhao, P., P.S. Chan, H.K.T. Ng. 2011. Optimal allocation of redundancies in series systems. {\em European Journal of Operational Research}~{\bf 220}~(3):673--683.
  1687. %
  1688. %\bibitem{Kuhn2013}Zymler, S., D. Kuhn, B. Rustem. 2013. Distributionally robust joint chance constraints with second-order moment information, {\em Mathematical Programming}~{\bf 137}~(1-2):167--198.
  1689. %\end{thebibliography}
  1690. %\newpage
  1691. %\ECSwitch
  1692. %\ECHead{Electronic Companion}\small
  1693. %\iffalse
  1694. %\section{Implementation Of K-means}
  1695. %The detail of K-means algorithm:
  1696. %
  1697. %1. Initialization: Randomly pick $K$ data points in $\mathcal{Z}$ to be the initial means $\hmu_1, \hmu_2, ... \hmu_K$.
  1698. %
  1699. %2. Assignment step: Assign each data point to one cluster, according to the rule
  1700. %\begin{equation}\label{k-means-assignment}
  1701. %\bm{z} \in \mathcal{Z}_{\argmin\limits_{k \in [K]}\|\bm{z} - \hmu_k\|^2}, \forall \bm{z} \in \mathcal{Z}
  1702. %\end{equation} and some tie-breaking rule.
  1703. %
  1704. %3. Update step: Compute the mean of each cluster:
  1705. %\begin{equation}\label{k-means-update}
  1706. %\hmu_k = \frac{1}{|\mathcal{Z}_k|}\sum\limits_{\bm{z} \in \mathcal{Z}_k}\bm{z}, \forall k \in [K]
  1707. %\end{equation}
  1708. %
  1709. %4. Check for convergence: if the new means computed in step 3 is the same as the before, then the algorithm converges and stops. Otherwise, go back to step 2.
  1710. %\fi
  1711. %
  1712. %\section{Discussion Of The Parallel Case}
  1713. %In this section, we discuss the case in which the cold-standby and active-parallel parts of a subsystem is activated together. First, consider the following worst-case probabilistic chance function in the problem given system design $\x$:
  1714. %\begin{equation}\label{Prob-2}
  1715. %\displaystyle \inf\limits_{\P \in \mathbb{F}_K}\P\left[\min\limits_{i \in \mathbf{N}}\left(\sum_{j \in \mathbf{M}^{\rm c}_i}\sum_{t\in \mathbf{T}_{ij}} \tilde{z}_{ijt}x_{ijt} \bigvee \max_{j \in \mathbf{M}^{\rm a}_i}  \max_{t \in \mathbf{T}_{ij}}\tilde{z}_{ijt}x_{ijt}\right)> \mathcal{T}_R \right].
  1716. % \end{equation}
  1717. %For a better exposition of our approach, we denote by
  1718. %\begin{equation}\label{constraint-set-2}
  1719. %\mathcal{Z}_i(\x):=\left\{\bm{z} \in \mathbb{R}^H_+ ~\left|~ \sum_{j \in \mathbf{M}^{\rm c}_i}\sum_{t\in \mathbf{T}_{ij}} \tilde{z}_{ijt}x_{ijt} \bigvee \max_{j \in \mathbf{M}^{\rm a}_i}  \max_{t \in \mathbf{T}_{ij}}\tilde{z}_{ijt}x_{ijt} > \mathcal{T}_R \right.\right \}, \forall i \in \mathbf{N},
  1720. %\end{equation}
  1721. %while the complimentary set of $\mathcal{Z}_i$ for each $i \in \mathbf{N}$, denoted by $\overline{\mathcal{Z}}_i$,  is
  1722. %\begin{equation}\label{set-LT-2}
  1723. %\overline{\mathcal{Z}}_i(\x)=\overline{\mathcal{Z}}^{\rm c}_i(\x) \cap \overline{\mathcal{Z}}^{\rm a}_i(\x),
  1724. %\end{equation}
  1725. %where
  1726. %$$
  1727. %\overline{\mathcal{Z}}^{\rm c}_i(\x):= \left\{\bm{z} \in \mathbb{R}^H_+ ~\left|~ \sum_{j\in \mathbf{M}^{\rm c}_i}\sum_{t\in \mathbf{T}_{ij}} z_{ijt} x_{ijt}\le \mathcal{T}_R \right.\right\}
  1728. %$$
  1729. %and
  1730. %$$
  1731. %\overline{\mathcal{Z}}^{\rm a}_i(\x):=\bigcap\limits_{j\in \mathbf{M}^{\rm a}_i}\bigcap\limits_{t\in \mathbf{T}_{ij}} \Big\{\bm{z} \in \mathbb{R}^H_+ ~\left|~z_{ijt} x_{ijt}  \le \mathcal{T}_R \right.\Big\}.
  1732. %$$
  1733. %
  1734. %The following result establishes an equivalent formulation of regular robust optimization for the above worst-case probabilistic chance  function~\eqref{Prob-2}.
  1735. %
  1736. %\begin{lemma}\label{lem1-LT-2}
  1737. %Given system design $\x$, the worst-case probabilistic chance function (\ref{Prob-2}) is equivalent to the optimal value of the following  optimization problem:
  1738. %\begin{equation}\label{P2-ambiguity-sup-dual1-2}
  1739. %\begin{array}{rcll}
  1740. %&\!\!\!\!\!\!\!\!\!\!\!\! \max\limits_{\bm{\alpha}, \bm{\beta}, \bm{\lambda}, \bm{\tau}} &  1-\sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}}\sum\limits_{k\in[K]} \left(\alpha^{k}_{ijt}\underline{\mu}^{k}_{ij} + \beta^{k}_{ijt}\overline{\mu}^{k}_{ij}\right) - \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i}\sum\limits_{k\in[K]}\epsilon^{k}_{ij}\lambda^{k}_{ij} - \sum\limits_{k\in[K]}\tau_{k} \\[0.3 cm]
  1741. %&\!\!\!\!\!\!\!\!\!\!\!\!{\rm s.t.} & \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}} \left[z_{ijt}\left(\alpha^{k}_{ijt}+ \beta^{k}_{ijt} \right) + \left|\frac{ z_{ijt}-\nu^{k}_{ij}}{|\mathbf{T}_{ij} |\sigma^{k}_{ij}} \right|\lambda^{k}_{ij}\right] + \tau_k \ge{p_k}, & \!\!\!\!\!\!\!\!\!\!\!\!\!\!\!\!\!\!\!\!\!\!\!\!\!\!\!\forall \bm{z} \in  \mathcal{Z}_k\cap \overline{\mathcal{Z}}_i(\x), i \in \mathbf{N}, k \in [K]\\
  1742. %&& \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}} \left[z_{ijt}\left(\alpha^{k}_{ijt}+ \beta^{k}_{ijt} \right) + \left|\frac{ z_{ij{t}}-\nu^{k}_{ij}}{|\mathbf{T}_{ij} |\sigma^{k}_{ij}} \right|\lambda^{k}_{ij}\right] + \tau_k \ge 0, & \!\!\!\!\!\!\!\!\!\!\!\!\!\!\!\!\!\!\!\!\!\!\!\!\!\!\!\forall \bm{z} \in  \mathcal{Z}_k, k\in [K]\\
  1743. %&& \halpha \le \mathbf{0}, \hbeta, \hlambda \ge \mathbf{0}, \bm{\tau} \in \mathbb{R}^K,
  1744. %\end{array}
  1745. %\end{equation}
  1746. %where
  1747. %\begin{equation}\label{equ:W-2}
  1748. %\mathcal{Z}_k = \big\{\bm{z} \in \mathcal{Z}\:\big|\: 2(\hmu_i - \hmu_k)^{\top}\bm{z} \leq \hmu_{i}^{\top}\hmu_{i} - \hmu_{k}^{\top}\hmu_{k}, \forall i \in [K] \big\}, \forall k \in [K].
  1749. %\end{equation}
  1750. %\end{lemma}
  1751. %\begin{proof}
  1752. %With the notation in (\ref{constraint-set-2}), the worst-case probabilistic chance function (\ref{Prob-2})  can be rewritten in terms of the probability of its complementary event:
  1753. %\begin{equation}\label{1minus-2}
  1754. %\inf\limits_{\P\in \mathbb{F}_{K}} \P\Big[\tilde{\bm{z} }\in \mathcal{Z}_i(\x),\forall i \in \mathbf{N} \Big]=1-\sup\limits_{\P \in \mathbb{F}_K } \P\Big[\cup_{i\in \mathbf{N} }\{\bm{\tilde{z}} \in \overline{\mathcal{Z}}_i(\x)\} \Big].
  1755. %\end{equation}
  1756. %
  1757. %Given the probability distribution of  $\tilde{s}$ as
  1758. %$$
  1759. %\P\Big[\tilde{k}=k\Big]=p_k, \forall k \in [K].
  1760. %$$
  1761. %We now define $\P_k$ as the conditional distribution of $\bm{\tilde{z}}$ given $\tilde{k}=k$ for $k \in [K]$, we then can decompose any distribution $\P \in \mathbb{F}_K$ using $\{\P_k, k\in [K]\}$ and rewrite the worst-case chance
  1762. %$$
  1763. %\sup\limits_{\P \in \mathbb{F}_K } \P\Big[\cup_{i\in \mathbf{N} }\{\bm{\tilde{z}} \in \overline{\mathcal{Z}}_i(\x)\} \Big]
  1764. %$$
  1765. %using total probability law as following formulation:
  1766. %\begin{eqnarray}\label{P2-ambiguity-sup-2}
  1767. %&&\sup\limits_{\P \in \mathbb{F}_K } \P\Big[\cup_{i\in \mathbf{N} }\{\bm{\tilde{z}} \in \overline{\mathcal{Z}}_i(\x)\} \Big]\\[0.35 cm]
  1768. %&=&\sup\limits_{\P_k, \forall k\in[K]}\sum\limits_{k\in[K]}p_k\P_k\Big[\cup_{i\in \mathbf{N} }\{\bm{\tilde{z}} \in \overline{\mathcal{Z}}_i(\x)\} \Big]\\[0.25 cm]
  1769. %&=&\sup\limits_{\P_k}\sum\limits_{k\in[K]}\displaystyle \int_{\cup_{i\in \mathbf{N} }\left\{\bm{\tilde{z}} \in \overline{\mathcal{Z}}_i(\x)\right\}}p_k {\rm d}\P_k(\bm{\tilde{z}}) \\[0.25 cm]
  1770. %&{\rm s.t.} &\displaystyle \int_{\bm{\tilde{z}} \in \mathcal{Z}_k} \tilde{z}_{ijt} {\rm d}\P_k(\bm{\tilde{z}}) \ge \underline{\mu}^{k}_{ij},    \forall i \in \mathbf{N}, j\in \mathbf{M}_i, t\in \mathbf{T}_{ij}, k\in [K]\\ [0.35 cm]
  1771. %&& \displaystyle \int_{\bm{\tilde{z}} \in \mathcal{Z}_k} \tilde{z}_{ijt} {\rm d}\P_k(\bm{\tilde{z}}) \le \overline{\mu}^{k}_{ij},    \forall i \in \mathbf{N}, j\in \mathbf{M}_i, t\in \mathbf{T}_{ij},k\in [K]\\ [0.35 cm]
  1772. %&&  \displaystyle \int_{\bm{\tilde{z}} \in \mathcal{Z}_k} \sum\limits_{t\in \mathbf{T}_{ij}}\left|\frac{\tilde{z}_{ijt}-\nu^{k}_{ij}}{|\mathbf{T}_{ij} |\sigma^{k}_{ij}} \right| {\rm d}\P_k(\bm{\tilde{z}}) \le \epsilon^{k}_{ij},   \forall i \in \mathbf{N}, j\in \mathbf{M}_i,k\in [K]\\[0.35 cm]
  1773. %&& \displaystyle \int_{\bm{\tilde{z}} \in \mathcal{Z}_k} {\rm d}\P_k(\bm{\tilde{z}}) =1, \forall k\in[K],
  1774. %\end{eqnarray}
  1775. %where the support $\mathcal{Z}_k$ is given by (\ref{equ:W-2}). The Lagrange dual of above moment problem (\ref{P2-ambiguity-sup-2}) has the following formulation (Wiesemann~et al.~2014):
  1776. %\begin{equation}\label{P1-ambiguity-sup-dual0-2}
  1777. %\begin{array}{rcl}
  1778. %&\!\!\!\!\!\!\!\!\!\!\!\! \min\limits_{\bm{\alpha}, \bm{\beta}, \bm{\lambda}, \bm{\tau}} &  \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}}\sum\limits_{k\in[K]} \left(\alpha^{k}_{ijt}\underline{\mu}^{k}_{ij}+ \beta^{k}_{ijt}\overline{\mu}^{k}_{ij}\right)+\sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i}\sum\limits_{k\in[K]}\epsilon^{k}_{ij}\lambda^{k}_{ij} + \sum\limits_{k\in[K]}\tau_{k} \\[0.3 cm]
  1779. %&\!\!\!\!\!\!\!\!\!\!\!\!{\rm s.t.} & \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}}  \left[z_{ijt}\left(\alpha^{k}_{ijt}+ \beta^{k}_{ijt} \right) + \left|\frac{ z_{ijt}-\nu^{k}_{ij}}{|\mathbf{T}_{ij} |\sigma^{k}_{ij}} \right|\lambda^{k}_{ij}\right] + \tau_k\\
  1780. %&& \ge p_k\mathbb{I}\Big({\cup_{i\in \mathbf{N} }\left\{ \bm{z} \in \overline{\mathcal{Z}}_i\right\}} \Big), \forall \bm{z} \in \mathcal{Z}_k, k \in [K]\\
  1781. %&& \halpha \le \mathbf{0}, \hbeta, \hlambda \ge \mathbf{0}, \bm{\tau} \in \mathbb{R}^K,
  1782. %\end{array}
  1783. %\end{equation}
  1784. %where $\mathbb{I}(\{\cdot\})$ is the indicator function with respect to set $\{\cdot\}$, and  $(\halpha, \hbeta, \hlambda, \htau)$ are the dual variables associated with the constraints of the primal problem (\ref{P2-ambiguity-sup-2}).
  1785. %
  1786. %
  1787. %
  1788. %Furthermore, we show the strong duality holds. Since ${\mu^{k}_{ij}}$ is the expectation of $\bm{\tilde{z}}_{ijt}$, we can always find a Dirac probability distribution $\P^{\dag}_{\bm{\mu}}$ with $\underline{\hmu}<\hmu<\overline{\hmu}$ which is relative interior point of the feasible set of problem (\ref{P2-ambiguity-sup-2}). Therefore, the Slater condition holds, and then the optimal value of (\ref{P1-ambiguity-sup-dual0-2}) is equivalent to that of problem (\ref{P2-ambiguity-sup-2}).
  1789. %
  1790. %
  1791. %
  1792. %
  1793. %
  1794. %Next, expanding the indication function $\mathbb{I}\left({\cup_{i\in \mathbf{N} }\left\{\bm{\tilde{z}} \in \overline{\mathcal{Z}}_i(\x)\right\}} \right)$ for different cases of $\bm{z}$, the above problem (\ref{P1-ambiguity-sup-dual0-2}) is also equivalent to the following formulation:
  1795. %\begin{equation}\label{P1-ambiguity-sup-dual00-2}
  1796. %\begin{array}{rcll}
  1797. %&\!\!\!\!\!\!\!\!\!\!\!\! \min\limits_{\bm{\alpha}, \bm{\beta}, \bm{\lambda}, \bm{\tau}} &  \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}}\sum\limits_{k\in[K]} \left(\alpha^{k}_{ijt}\underline{\mu}^{k}_{ij}+ \beta^{k}_{ijt}\overline{\mu}^{k}_{ij}\right)+\sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i}\sum\limits_{k\in[K]}\epsilon^{k}_{ij}\lambda^{k}_{ij} + \sum\limits_{k\in[K]}\tau_{k} \\[0.3 cm]
  1798. %&\!\!\!\!\!\!\!\!\!\!\!\!{\rm s.t.} & \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}}  \left[z_{ijt}\left(\alpha^{k}_{ijt}+ \beta^{k}_{ijt} \right) + \left|\frac{ z_{ijt}-\nu^{k}_{ij}}{|\mathbf{T}_{ij} |\sigma^{k}_{ij}} \right|\lambda^{k}_{ij}\right] + \tau_k \ge p_k, \forall \bm{z} \in \mathcal{Z}_k\cap\overline{\mathcal{Z}}_i, i \in \mathbf{N}, k \in [K]\\
  1799. %&& \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}}  \left[z_{ijt}\left(\alpha^{k}_{ijt}+ \beta^{k}_{ijt} \right) + \left|\frac{ z_{ijt}-\nu^{k}_{ij}}{|\mathbf{T}_{ij} |\sigma^{k}_{ij}} \right|\lambda^{k}_{ij}\right] + \tau_k \ge0, \forall \bm{z} \in \mathcal{Z}_k, k \in [K]\\\nonumber
  1800. %&& \halpha \le \mathbf{0}, \hbeta, \hlambda \ge \mathbf{0}, \bm{\tau} \in \mathbb{R}^K.
  1801. %\end{array}
  1802. %\end{equation}
  1803. %Finally, plugging this formulation into the equation (\ref{1minus-2}), we arrive at the the formulation of (\ref{P2-ambiguity-sup-dual1-2}) whose optimal objective value is exactly the worst-case value of probabilistic chance function (\ref{Prob-2}) for all the $\mathbb{P}_k \in \mathbb{F}_k$, given system design $\x$.  The proof is completed.
  1804. %\end{proof}
  1805. %
  1806. %It is noted that the derived optimization problem (\ref{P2-ambiguity-sup-dual1-2}) in current version still belongs to semi-infinitely dimensional optimization problems which are not directly computable.  In the following, we show that by duality argument the problem can be further transformed into a computationally tractable formulation of linear program.
  1807. %
  1808. %\begin{proposition}\label{P-proposition1b-2}
  1809. %Given a system design $\x$, the worst-case probabilistic chance function (\ref{Prob-2}) solves the following linear program (LP):
  1810. %\begin{eqnarray}
  1811. %&\!\!\!\!\!\! \max &  1-\sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}}\sum\limits_{k\in[K]} \left(\alpha^{k}_{ijt}\underline{\mu}^{k}_{ij} + \beta^{k}_{ijt}\overline{\mu}^{k}_{ij}\right) - \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i}\sum\limits_{k\in[K]}\epsilon^{k}_{ij}\lambda^{k}_{ij} - \sum\limits_{k\in[K]}\tau_{k} \label{HP1-ambiguity-LP-FL-2} \\
  1812. % &\!\!\!\!\!\!{\rm s.t.} & \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}} \Big[\phi^{lk}_{ijt}\underline{z}_{ij}+\varphi^{lk}_{ijt}\overline{z}_{ij}+{\nu^{k}_{ij}\left(\pi^{lk}_{ijt}-\varpi^{lk}_{ijt} \right)}  \nonumber\\
  1813. %  &&+  \sum\limits_{n \in [K]}\psi^{lk}_{n}({(\mu^{n}_{ijt})}^2 - {(\mu^{k}_{ijt})}^2)\Big] + \sum\limits_{j\in \mathbf{M}^{\rm a}_{l}}\sum\limits_{t\in \mathbf{T}_{lj}} q^{lk}_{jt}\mathcal{T}_R+ s^{lk}\mathcal{T}_R + \tau_k \geq p_k, \forall l \in \mathbf{N}, k \in [K]  \label{HP1-ambiguity-LP-FL1-3}\\
  1814. %  && \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}} \Big[\rho^{k}_{ijt}\underline{z}_{ij}+\varrho^{k}_{ijt}\overline{z}_{ij}+{\nu^{k}_{ij}\left(\gamma^{k}_{ijt}-\theta^{k}_{ijt} \right)} \nonumber\\ && + \sum\limits_{n \in [K]}\varsigma^{k}_{n}({(\mu^{n}_{ijt})}^2 - {(\mu^{k}_{ijt})}^2)\Big] + \tau_k \geq0, \forall k \in [K]\label{HP1-ambiguity-LP-FL1-2-2}\\
  1815. % &&  q^{lk}_{jt}x_{l jt}  +\sum\limits_{n \in [K]}2(\mu^{n}_{ljt} - \mu^{k}_{ljt})\psi^{lk}_{n}+ \phi^{lk}_{ljt}+\varphi^{lk}_{ljt}+{ \pi^{lk}_{ljt}-\varpi^{lk}_{ljt} } = \alpha^{k}_{ljt}+\beta^{k}_{ljt},\nonumber\\ &&  \forall l \in \mathbf{N}, j\in \mathbf{M}^{\rm a}_l, t \in \mathbf{T}_{l j}, k \in [K] \\
  1816. %&&  s^{lk}x_{l jt}  +\sum\limits_{n \in [K]}2(\mu^{n}_{ljt} - \mu^{k}_{ljt})\psi^{lk}_{n}+ \phi^{lk}_{ljt}+\varphi^{lk}_{ljt}+\pi^{lk}_{ljt}-\varpi^{lk}_{ljt}  = \alpha^{k}_{ljt}+\beta^{k}_{ljt},\nonumber\\ &&  \forall l \in \mathbf{N}, j\in \mathbf{M}^{\rm c}_l, t \in \mathbf{T}_{l j}, k \in [K] \\
  1817. %&&   \sum\limits_{n \in [K]}2(\mu^{n}_{ijt} - \mu^{k}_{ijt})\psi^{lk}_{n}+  \phi^{lk}_{ijt}+\varphi^{lk}_{ijt}+\pi^{lk}_{ijt}-\varpi^{lk}_{ijt} = \alpha^{k}_{ijt}+\beta^{k}_{ijt},\nonumber\\ && ~ \forall l \in \mathbf{N}, i \in \mathbf{N}\setminus\{l \}, j \in \mathbf{M}_i, t\in \mathbf{T}_{ij}, k \in [K], \label{HP2-ambiguity-LP-FL2-2} \\
  1818. %&&{|\mathbf{T}_{ij} |\sigma^{k}_{ij}}(\pi^{lk}_{ijt}+\varpi^{lk}_{ijt}) =\lambda^{k}_{ij},  ~ \forall l  \in \mathbf{N}, i \in \mathbf{N}, j\in \mathbf{M}_i, t\in \mathbf{T}_{ij}, k \in [K] \\
  1819. %&& \sum\limits_{n \in [K]}2(\mu^{n}_{ijt} - \mu^{k}_{ijt})\varsigma^{k}_{n}+ \rho^{k}_{ijt}+\varrho^{k}_{ijt}+\gamma^{k}_{ijt}-\theta^{k}_{ijt} = \alpha^{k}_{ijt}+\beta^{k}_{ijt},\nonumber\\ && ~ \forall i \in \mathbf{N}, j \in \mathbf{M}_i, t\in \mathbf{T}_{ij}, k \in [K]  \\
  1820. %&&  {|\mathbf{T}_{ij} |\sigma^{k}_{ij}}(\gamma^{k}_{ijt}+\theta^{k}_{ijt}) =\lambda^{k}_{ij},  ~ \forall i \in \mathbf{N}, j\in \mathbf{M}_i, t\in \mathbf{T}_{ij}, k \in [K] \\
  1821. %%&&  q_{l jk}\le y_{{l jk}}, ~\forall l  \in \mathbf{N}, j \in \mathbf{M}_{l}, k\in \mathcal{N}(l,j) \\[0.3 cm]
  1822. %%&& y_{{l jk}} \ge M x_{l jk}, ~\forall l  \in \mathbf{N}, j \in \mathbf{M}_{l}, k\in \mathcal{N}(l,j) \\[0.3 cm]
  1823. %%&& y_{{l jk}} \le  q_{l jk }+(x_{l jk}-1)M, ~\forall l  \in \mathbf{N}, j \in \mathcal{J}({l}), k \in \mathcal{N}(l,j)\\[0.3 cm]
  1824. %%&&  \sum\limits_{k\in \mathbf{K}_{ij}} x^{k}_{ij}\ge  L_{ij}, ~  \forall  i \in \mathbf{N}, j\in \mathbf{M}_i \\
  1825. %&& \halpha,\q, \s, \hpsi, \hvarphi, \hvarsigma, \hvarrho \le \mathbf{0}, \htau \in \mathbb{R}^K,  \\
  1826. %&& \hbeta, \hlambda, \hphi, \hrho, \hpi,\hvarpi, \hgamma, \htheta \ge \mathbf{0}, \label{HP2-ambiguity-LP-FL-2}
  1827. %\end{eqnarray}
  1828. %where $\halpha, \hbeta, \hlambda, \htau, \q, \s, \hpsi, \hphi, \hvarphi, \hpi, \hvarpi, \hrho, \hvarrho, \hvarsigma, \hgamma, \htheta$ are auxiliary variables.
  1829. %\end{proposition}
  1830. %
  1831. %{\bf Proof.}
  1832. %First of all, for a given $l  \in \mathbf{N}$ we deal with the infinitely dimensional constraints
  1833. %$$
  1834. %\sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}}  \left[z_{ijt}\left(\alpha^{k}_{ijt}+ \beta^{k}_{ijt} \right) + \left|\frac{ z_{ijt}-\nu^{k}_{ij}}{|\mathbf{T}_{ij} |\sigma^{k}_{ij}} \right|\lambda^{k}_{ij}\right] + \tau_k \ge p_k, \forall \bm{z} \in \mathcal{Z}_k\cap\overline{\mathcal{Z}}_l, k \in [K]\\
  1835. %$$
  1836. %Recall that
  1837. %$$
  1838. %\overline{\mathcal{Z}}_l(\x)=\overline{\mathcal{Z}}^{\rm c}_l(\x) \cap \overline{\mathcal{Z}}^{\rm a}_l(\x)= \left\{\bm{z} \in \mathbb{R}^{H}_+ ~\left|~\begin{array}{l}
  1839. %\displaystyle \sum_{j\in \mathbf{M}^{\rm c}_{l}}\sum_{t\in \mathbf{T}_{lj}} z_{ljt} x_{ljt}\le \mathcal{T}_R \\
  1840. % z_{ljt} x_{ljt}  \le \mathcal{T}_R,  j\in \mathbf{M}^a_{l}, t \in \mathbf{T}_{lj}
  1841. %                                                                                                                                             \end{array}
  1842. %\right.\right\},
  1843. %$$
  1844. %and
  1845. %$$
  1846. %\mathcal{Z}_k = \big\{\bm{z} \in \mathcal{Z}\:\big|\: 2(\hmu_i - \hmu_k)^{\top}\bm{z} \leq \hmu_{i}^{\top}\hmu_{i} - \hmu_{k}^{\top}\hmu_{k}, \forall i \in [K] \big\}, \forall k \in [K],
  1847. %$$
  1848. %where
  1849. %$$
  1850. %\mathcal{Z} := \big\{\bm{z} \in \mathbb{R}^{H}| z_{ijt} \in [\underline{z}_{ij},\overline{z}_{ij}], \forall i \in \mathbf{N}, j \in \mathbf{M}_i, t\in \mathbf{T}_{ij} \big\}.
  1851. %$$
  1852. %
  1853. %First of all, we claim that for any $k \in [K]$
  1854. %\begin{equation}\label{Lifting-3}
  1855. %\sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}}  \left[z_{ijt}\left(\alpha^{k}_{ijt}+ \beta^{k}_{ijt} \right) + \left|\frac{ z_{ijt}-\nu^{k}_{ij}}{|\mathbf{T}_{ij} |\sigma^{k}_{ij}} \right|\lambda^{k}_{ij}\right] + \tau_k \ge p_k, \forall \bm{z} \in \mathcal{Z}_k\cap\overline{\mathcal{Z}}_l
  1856. %\end{equation}
  1857. %is equivalent to
  1858. %\begin{equation}\label{Lifting-4}
  1859. %\sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}}  \left[z_{ijt}\left(\alpha^{k}_{ijt}+ \beta^{k}_{ijt} \right) + u^{k}_{ijt}\lambda^{k}_{ij}\right] + \tau_k \ge p_k, \forall (\bm{z},\u) \in \mathcal{W}_k,
  1860. %\end{equation}
  1861. %where
  1862. %$$
  1863. %\mathcal{W}_k := \left\{(\bm{z}, \u) \middle| \: \bm{z} \in \mathcal{Z}_k\cap\overline{\mathcal{Z}}_l, \left|\frac{ z_{ijt}-\nu^{k}_{ij}}{|\mathbf{T}_{ij} |\sigma^{k}_{ij}} \right| \leq u^{k}_{ijt}, \forall i \in \mathbf{N}, j \in \mathbf{M}_i, t \in \mathbf{T}_{ij} \right\}, \forall k \in [K].
  1864. %$$
  1865. %
  1866. %Assume \eqref{Lifting-3} holds, since $\hlambda \geq \mathbf{0}$,
  1867. %$$
  1868. %u^{k}_{ijt}\lambda^{k}_{ij} \geq \left|\frac{ z_{ijt}-\nu^{k}_{ij}}{|\mathbf{T}_{ij} |\sigma^{k}_{ij}} \right|\lambda^{k}_{ij}, \forall i \in \mathbf{N}, j \in \mathbf{M}_i, t \in \mathbf{T}_{ij}, (\bm{z}, \u) \in \mathcal{W}_k.
  1869. %$$
  1870. %Therefore \eqref{Lifting-4} holds.
  1871. %
  1872. %Assume \eqref{Lifting-4} holds, then for any $\hat{\bm{z}} \in \mathcal{Z}_k\cap\overline{\mathcal{Z}}_l$,
  1873. %$$
  1874. %\min\limits_{(\hat{\bm{z}}, \u) \in \mathcal{W}_k}u^{k}_{ijt} = \left|\frac{ \hat{z}_{ijt}-\nu^{k}_{ij}}{|\mathbf{T}_{ij} |\sigma^{k}_{ij}} \right|, \forall i \in \mathbf{N}, j \in \mathbf{M}_i, t \in \mathbf{T}_{ij}.
  1875. %$$
  1876. %Since $\hlambda \geq \mathbf{0}$,
  1877. %$$
  1878. %\sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}}  \left[\hat{z}_{ijt}\left(\alpha^{k}_{ijt}+ \beta^{k}_{ijt} \right) + \left|\frac{ \hat{z}_{ijt}-\nu^{k}_{ij}}{|\mathbf{T}_{ij} |\sigma^{k}_{ij}} \right|\lambda^{k}_{ij}\right] + \tau_k
  1879. %$$
  1880. %$$
  1881. %=\min\limits_{(\hat{\bm{z}}, \u) \in \mathcal{W}_k}\sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}}  \left[\hat{z}_{ijt}\left(\alpha^{k}_{ijt}+ \beta^{k}_{ijt} \right) + u^{k}_{ijt}\lambda^{k}_{ij}\right] + \tau_k \geq p_k.
  1882. %$$
  1883. %Therefore \eqref{Lifting-3} holds.
  1884. %
  1885. %Therefore \eqref{Lifting-3} and \eqref{Lifting-4} are equivalent.
  1886. %
  1887. %Then, by introducing auxiliary variables $u^{k}_{ijt}, \forall i \in \mathbf{N}, j \in \mathbf{M}_i, t\in \mathbf{T}_{ij}$, we can equivalently lift the above constraints into the following optimization-based formulation:
  1888. %\begin{equation}\label{H-system2}
  1889. %\left.\begin{array}{rcll}
  1890. %  p_k-\tau_k\le & \min\limits_{\bm{z}, \u} & \displaystyle \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}} \left[z_{ijt}\left(\alpha^{k}_{ijt}+ \beta^{k}_{ijt} \right) + u^{k}_{ijt} \lambda^{k}_{ij}\right]  \\[0.3 cm]
  1891. %&{\rm s.t.} &    z_{l jt}x_{l jt}\le \mathcal{T}_R, &\forall j\in \mathbf{M}^{\rm a}_{l}, t\in \mathbf{T}_{l j}  \\[0.3 cm]
  1892. %&& \displaystyle \sum_{j\in \mathbf{M}^{\rm c}_{l}}\sum_{t\in \mathbf{T}_{lj}} z_{ljt} x_{ljt}\le \mathcal{T}_R, &   \\[0.3 cm]
  1893. %&&  {|\mathbf{T}_{ij} |\sigma^{k}_{ij}}u^{k}_{ijt}- { z_{ijt}} \ge { -\nu^{k}_{ij}}, & \forall i \in \mathbf{N}, j\in \mathbf{M}_i, t\in \mathbf{T}_{ij} \\[0.3 cm]
  1894. %&&  {|\mathbf{T}_{ij} |\sigma^{k}_{ij}}u^{k}_{ijt}+  { z_{ijt}} \ge  {\nu^{k}_{ij} }, & \forall i \in \mathbf{N}, j\in \mathbf{M}_i, t\in \mathbf{T}_{ij} \\[0.3 cm]
  1895. %&&  \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}} 2(\mu^{n}_{ijt} - \mu^{k}_{ijt})z_{ijt}\\[0.3 cm]
  1896. %&& \leq \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}} {(\mu^{n}_{ijt})}^2 - {(\mu^{k}_{ijt})}^2, & \forall n \in [K]\\[0.3 cm]
  1897. %&& \underline{z}_{ij} \leq z_{ijt} \leq \overline{z}_{ij} & \forall i \in \mathbf{N}, j \in \mathbf{M}_i, t\in \mathbf{T}_{ij}\\[0.3cm]
  1898. %&& u^{k}_{ijt} \in \mathbb{R}, & \forall i \in \mathbf{N}, j\in \mathbf{M}_i, t\in \mathbf{T}_{ij}.
  1899. %\end{array}\right\}, \forall k \in [K]
  1900. %\end{equation}
  1901. %
  1902. %By the strong duality of linear programming, the above constraint is also equivalent to the following system: for all $k$ in $[K]$,
  1903. %\begin{equation*}
  1904. %\left\{  \begin{array}{rl}
  1905. %& p_k-\tau_k\le \sum\limits_{j\in \mathbf{M}^{\rm a}_{l}}\sum\limits_{t\in \mathbf{T}_{lj}} q^{lk}_{jt}\mathcal{T}_R+ s^{lk}\mathcal{T}_R\\[0.3 cm]
  1906. %& +\sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}} \Big[\phi^{lk}_{ijt}\underline{z}_{ij}+\varphi^{lk}_{ijt}\overline{z}_{ij}+{\nu^{k}_{ij}\left(\pi^{lk}_{ijt}-\varpi^{lk}_{ijt} \right)}  + \sum\limits_{n \in [K]}\psi^{lk}_{n}({(\mu^{n}_{ijt})}^2 - {(\mu^{k}_{ijt})}^2)\Big]  \\[0.3 cm]
  1907. %&  q^{lk}_{jt}x_{l jt}  +\sum\limits_{n \in [K]}2(\mu^{n}_{ljt} - \mu^{k}_{ljt})\psi^{lk}_{n}+ \phi^{lk}_{ljt}+\varphi^{lk}_{ljt}+{ \pi^{lk}_{ljt}-\varpi^{lk}_{ljt} } = \alpha^{k}_{ljt}+\beta^{k}_{ljt}, \forall j\in \mathbf{M}^{\rm a}_l, t \in \mathbf{T}_{l j} \\[0.3 cm]
  1908. %&  s^{lk}x_{l jt}  +\sum\limits_{n \in [K]}2(\mu^{n}_{ljt} - \mu^{k}_{ljt})\psi^{lk}_{n}+ \phi^{lk}_{ljt}+\varphi^{lk}_{ljt}+\pi^{lk}_{ljt}-\varpi^{lk}_{ljt}  = \alpha^{k}_{ljt}+\beta^{k}_{ljt}, \forall j\in \mathbf{M}^{\rm c}_l, t \in \mathbf{T}_{l j} \\[0.3 cm]
  1909. %&   \sum\limits_{n \in [K]}2(\mu^{n}_{ijt} - \mu^{k}_{ijt})\psi^{lk}_{n}+  \phi^{lk}_{ijt}+\varphi^{lk}_{ijt}+\pi^{lk}_{ijt}-\varpi^{lk}_{ijt} = \alpha^{k}_{ijt}+\beta^{k}_{ijt}, ~ \forall i \in \mathbf{N}\setminus\{l \}, j \in \mathbf{M}_i, t\in \mathbf{T}_{ij}  \\[0.3 cm]
  1910. %&{|\mathbf{T}_{ij} |\sigma^{k}_{ij}}(\pi^{lk}_{ijt}+\varpi^{lk}_{ijt}) =\lambda^{k}_{ij},  ~ \forall i \in \mathbf{N}, j\in \mathbf{M}_i, t\in \mathbf{T}_{ij} \\[0.3 cm]
  1911. %&   q^{lk}_{jt}\le 0, s^{lk}\le 0, \psi^{lk}_{n} \le 0, \phi^{lk}_{ijt} \geq 0, \varphi^{lk}_{ijt} \leq 0, \pi^{lk}_{ijt}\ge 0,\varpi^{lk}_{ijt}\ge 0, ~\forall n \in [K], i \in \mathbf{N}, j\in \mathbf{M}_i, t\in \mathbf{T}_{ij}.
  1912. %\end{array}
  1913. %\right\}
  1914. %\end{equation*}
  1915. %
  1916. %
  1917. %
  1918. %
  1919. %Likewise, the constraints
  1920. %$$
  1921. %\sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}}  \left[z_{ijt}\left(\alpha^{k}_{ijt}+ \beta^{k}_{ijt} \right) + \left|\frac{ z_{ijt}-\nu^{k}_{ij}}{|\mathbf{T}_{ij} |\sigma^{k}_{ij}} \right|\lambda^{k}_{ij}\right] + \tau_k \ge0, \forall \bm{z} \in \mathcal{Z}_k, k \in [K]\\
  1922. %$$
  1923. %can also be dualized similarly.  Leveraging on the derived formulation (\ref{P2-ambiguity-sup-dual1-2}) in Lemma~\ref{lem1-LT-2}, we can arrive at the formulation of the linear program (\ref{HP1-ambiguity-LP-FL-2})--(\ref{HP2-ambiguity-LP-FL-2}). We are done.
  1924. %\blot
  1925. %
  1926. %
  1927. %Although the derived formulation (\ref{HP1-ambiguity-LP-FL-2})-(\ref{HP2-ambiguity-LP-FL-2}) can be considered as a linear program given a system design $\mathbf{x}$, in the original problem $\mathbf{x}$ is the decision variable, so there are bilinear terms $q^{lk}_{jt}x_{l jt}$ and $s^{lk}x_{l jt}$ in the overall formulation after the derived formulation is plugged back into the original problem. We can linearize these terms.
  1928. %
  1929. %
  1930. %\begin{proposition}\label{proposition1b-2-2}
  1931. %The overall problem  is equivalent to the following mixed integer linear program:
  1932. %\begin{eqnarray}
  1933. % & \min\limits_{\x} &  \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \left[\sum\limits_{t\in \mathbf{T}_{ij}} x_{ijt}\right]c_{ij} \label{HP1-ambiguity-MILP-FL1-4-2}\\
  1934. % &{\rm s.t.} & 1-\sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}}\sum\limits_{k\in[K]} \left(\alpha^{k}_{ijt}\underline{\mu}^{k}_{ij} + \beta^{k}_{ijt}\overline{\mu}^{k}_{ij}\right)\nonumber\\
  1935. % && - \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i}\sum\limits_{k\in[K]}\epsilon^{k}_{ij}\lambda^{k}_{ij} - \sum\limits_{k\in[K]}\tau_{k}\ge R_{0}   \\
  1936. % &&  L_{i} \le \sum\limits_{j\in \mathbf{M}_i}\sum\limits_{t\in \mathbf{T}_{ij}} x_{ijt}\le  U_{i}, ~  \forall  i \in \mathbf{N} \\
  1937. %%&& \sum\limits_{j\in \mathbf{M}_{l}} \sum\limits_{k\in \mathbf{K}_{l j}}q_{l jk}\mathcal{T}_R\nonumber\\
  1938. %% &&+\sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{k\in \mathbf{K}_{ij}} \left[ \phi^{l }^{k}_{ij}\underline{z}_{ij}+ \varphi^{\varsigma}^{k}_{ij}\overline{z}_{ij}  + {\nu_{ij}\left(\pi^{l }^{k}_{ij}-\varpi^{l }^{k}_{ij} \right)} \right]+\tau \ge 1,~\forall {l  \in \mathbf{N}}  \\
  1939. %&& y^{\rm a}_{l jk}  +\sum\limits_{n \in [K]}2(\mu^{n}_{ljt} - \mu^{k}_{ljt})\psi^{lk}_{n}+ \phi^{lk}_{ljt}+\varphi^{lk}_{ljt}+{ \pi^{lk}_{ljt}-\varpi^{lk}_{ljt} } = \alpha^{k}_{ljt}+\beta^{k}_{ljt},\nonumber\\ &&  \forall l \in \mathbf{N}, j\in \mathbf{M}^{\rm a}_l, t \in \mathbf{T}_{l j}, k \in [K]  \label{con:32}  \\
  1940. %&& y^{\rm c}_{l jk}  +\sum\limits_{n \in [K]}2(\mu^{n}_{ljt} - \mu^{k}_{ljt})\psi^{lk}_{n}+ \phi^{lk}_{ljt}+\varphi^{lk}_{ljt}+\pi^{lk}_{ljt}-\varpi^{lk}_{ljt}  = \alpha^{k}_{ljt}+\beta^{k}_{ljt},\nonumber\\ &&  \forall l \in \mathbf{N}, j\in \mathbf{M}^{\rm c}_l, t \in \mathbf{T}_{l j}, k \in [K]    \\
  1941. % && (\ref{HP1-ambiguity-LP-FL1-3}-\ref{HP1-ambiguity-LP-FL1-2-2}); (\ref{HP2-ambiguity-LP-FL2-2})-(\ref{HP2-ambiguity-LP-FL-2})\\
  1942. %%&&\phi^{l }^{k}_{ij}+\varphi^{l }^{k}_{ij}+ { \pi^{l }^{k}_{ij}-\varpi^{l }^{k}_{ij} }= \alpha^{k}_{ij}+\beta^{k}_{ij}, ~ \forall {l  \in \mathbf{N}}, i \in \mathbf{N}\setminus\{l \}, j \in \mathbf{M}_i, k\in \mathbf{K}_{ij}  \\
  1943. %%&&{|\mathbf{K}_{ij} |\sigma_{ij}} (\pi^{l }^{k}_{ij}+\varpi^{l }^{k}_{ij}) =\lambda_{ij},  ~ \forall {l  \in \mathbf{N}}, i \in \mathbf{N}, j\in \mathbf{M}_i, k\in \mathbf{K}_{ij} \\
  1944. %%&& \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{k\in \mathbf{K}_{ij}} \left[\left(\varsigma^{k}_{ij}\underline{z}_{ij}+ \vartheta^{k}_{ij}\overline{z}_{ij} \right) +  {\nu_{ij}\left(\gamma^{k}_{ij}-\theta^{k}_{ij} \right)}\right]+\tau \ge 0\\
  1945. %%&& \varsigma^{k}_{ij}+ \vartheta^{k}_{ij} + { \gamma^{k}_{ij}-\theta^{k}_{ij} }= \alpha^{k}_{ij}+\beta^{k}_{ij}, ~ \forall  i \in \mathbf{N}, j\in \mathbf{M}_i, k\in \mathbf{K}_{ij} \\
  1946. %%&&  {|\mathbf{K}_{ij} |\sigma_{ij}}(\gamma^{k}_{ij}+\theta^{k}_{ij}) = \lambda_{{ij}}, ~  \forall  i \in \mathbf{N}, j\in \mathbf{M}_i, k\in \mathbf{K}_{ij} \\
  1947. %&&  q^{lk}_{jt}\le y^{{\rm a}lk}_{jt}, ~\forall l  \in \mathbf{N}, j \in \mathbf{M}^{\rm a}_{l}, t\in \mathbf{T}_{l j}, k \in [K] \\
  1948. %&& y^{{\rm a}lk}_{jt} \ge M x_{l jt}, ~\forall l  \in \mathbf{N}, j \in \mathbf{M}^{\rm a}_{l}, t\in \mathbf{T}_{l j}, k \in [K] \\
  1949. %&& y^{{\rm a}lk}_{jt} \le  q^{lk}_{jt}+(x_{l jt}-1)M, ~\forall l  \in \mathbf{N}, j \in \mathbf{M}^{\rm a}_{l}, t \in \mathbf{T}_{l j}, k \in [K]\\
  1950. %&&  s^{lk}\le y^{{\rm c}lk}_{jt}, ~\forall l  \in \mathbf{N}, j \in \mathbf{M}^{\rm c}_{l}, t\in \mathbf{T}_{l j}, k \in [K] \\
  1951. %&& y^{{\rm c}lk}_{jt} \ge M x_{l jt}, ~\forall l  \in \mathbf{N}, j \in \mathbf{M}^{\rm c}_{l}, t\in \mathbf{T}_{l j}, k \in [K] \\
  1952. %&& y^{{\rm c}lk}_{jt} \le  s^{lk}+(x_{l jt}-1)M, ~\forall l  \in \mathbf{N}, j \in \mathbf{M}^{\rm c}_{l}, t \in \mathbf{T}_{l j}, k \in [K]\\
  1953. %%&& \alpha^{k}_{ij}\le 0, \beta^{k}_{ij}\ge 0, \lambda_{{ij}} \ge 0, \tau \in \Re, ~\forall i \in \mathbf{N}, j\in \mathbf{M}_i, k\in \mathbf{K}_{ij}\\
  1954. %%&&q_{l jk}\le 0, y_{l jk} \le 0, ~\forall {l  \in \mathbf{N}}, j\in \mathbf{M}_{l}, k\in \mathbf{K}_{l j} \\
  1955. %%&& \phi^{l }^{k}_{ij} \ge 0,  \varphi^{l }^{k}_{ij} \le 0,  \pi^{l }^{k}_{ij}\ge 0,\varpi^{l }^{k}_{ij}\ge 0, ~\forall {l  \in \mathbf{N}},  i \in \mathbf{N}, j\in \mathbf{M}_i, k\in \mathbf{K}_{ij}\\
  1956. %%&& \theta^{k}_{ij}\ge 0, \gamma^{k}_{ij}\ge 0, \varsigma^{k}_{ij}\ge 0, \vartheta^{k}_{ij} \le 0, ~\forall  i \in \mathbf{N}, j\in \mathbf{M}_i, k\in \mathbf{K}_{ij}\\
  1957. %&& \y^{\rm a}, \y^{\rm c} \le \mathbf{0}, \x\in \{0,1\}^{H},  \label{HP1-ambiguity-MILP-FL2-4-2}
  1958. %\end{eqnarray}
  1959. %where $\halpha, \hbeta, \hlambda, \htau, \q, \s, \hpsi, \hphi, \hvarphi, \hpi, \hvarpi, \hrho, \hvarrho, \hvarsigma, \hgamma, \htheta, \y^{\rm a}$ and $\y^{\rm c}$ are auxiliary variables and $M$ is a sufficiently small negative number.
  1960. %\end{proposition}
  1961. %{\bf Proof. }
  1962. %In the proof of Proposition~\ref{P-proposition1b-2}, the feasible set $\mathcal{Z}_k\cap \overline{\mathcal{Z}}_{l }(\x)$ of the minimization problem
  1963. %$$
  1964. %\min\limits_{\bm{z}} \displaystyle \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{t\in \mathbf{T}_{ij}} \left[z_{ijt}\left(\alpha^{k}_{ijt}+ \beta^{k}_{ijt} \right) + \left|\frac{ z_{ijt}-\nu^{k}_{ij}}{|\mathbf{T}_{ij} |\sigma^{k}_{ij}} \right| \lambda^{k}_{ij}\right]
  1965. %$$
  1966. %is bounded. Assuming that it is nonempty, then its lifted equivalent form of the inner minimization problem in (\ref{H-system2}) is also bounded and nonempty. Therefore, the dual variables $q^{lk}_{jt}$ and $s^{lk}$ are also bounded. Therefore, we can linearize the bilinear terms $q^{lk}_{jt}x_{l jt}$ and $s^{lk}x_{l jt}$ by introducing new variables $y^{{\rm a}lk}_{jt}$ and $y^{{\rm c}lk}_{jt}$, such that
  1967. %\begin{equation}\label{equ:linear1}
  1968. %q^{lk}_{jt}\le y^{{\rm a}lk}_{jt}\le 0,~y^{{\rm a}lk}_{jt} \ge M x_{l jt},~y^{{\rm a}lk}_{jt} \le  q^{lk}_{jt}+(x_{l jt}-1)M,
  1969. %\end{equation}
  1970. %and
  1971. %\begin{equation}\label{equ:linear2}
  1972. %s^{lk}\le y^{{\rm c}lk}_{jt}\le 0,~y^{{\rm c}lk}_{jt} \ge M x_{l jt},~y^{{\rm c}lk}_{jt} \le  s^{lk}+(x_{l jt}-1)M,
  1973. %\end{equation}
  1974. %respectively, where $M$ is a sufficiently small negative number (in numerical computation, M can be set to a negative number with very large absolute value).  Using this linearization technique, we can arrive at the following formulation of mixed integer linear program (MILP) for the distributionally robust redundancy allocation problem. \blot\\
  1975. % Note that in the linearized MIP formulation~\eqref{HP1-ambiguity-MILP-FL1-4-2}-\eqref{HP1-ambiguity-MILP-FL2-4-2}, the integer variables are still $x_{ijt}$, the original redundancy allocation decision variables, therefore the problem remains tractable.
  1976. %
  1977. %\section{A baseline probabilistic reliability model}\label{EC4}
  1978. %For comparison, we consider the following probabilistic reliability model, which is an MIP with binaries (Feizollahi and Modarres~2012, Wang et.al. 2019):
  1979. %\begin{eqnarray}
  1980. %\begin{array}{rcll}
  1981. %& \min\limits_{\x} & \displaystyle \sum\limits_{i\in \mathbf{N}} \left[L_{i}+\sum_{k=0}^{U_{i}-L_{i}}kx_{ik}\right]c_{i} \\[0.3 cm]
  1982. %& {\rm s.t.} &  \displaystyle \sum_{i \in \mathbf{N}} \sum_{k=0}^{U_{i}-L_{i}}x_{ik}\ln\left[1-r_{i}^{L_{i}+k} \right]\ge \ln R_0  & \\[0.3 cm]
  1983. %&& \displaystyle \sum_{k=0}^{U_{i}-L_{i}}x_{ik}=1, & \forall i \in \mathbf{N}\\
  1984. %&& x_{ik} \in \{0,1\},  & \forall i \in \mathbf{N}, k \in [0; U_{i}-L_{i}],
  1985. %\end{array}
  1986. %\end{eqnarray}
  1987. %which can be solved by off-the-shelf MIP solvers. Note that the above linear MIP transformation holds only for the regular series-parallel redundant systems with a single type of component; as for the situation that involves multiple types ({\it i.e.,} $|\mathbf{M}_i|>1$ for some $i \in \mathbf{N}$) or the cold-standby subsystems are considered, the probabilistic model, in general, becomes intractable.
  1988. %
  1989. %\iffalse
  1990. %
  1991. %Furthermore, to enhance the scalability, we can harness both the structures of formulation (\ref{HP1-ambiguity-LP-FL})-(\ref{HP2-ambiguity-LP-FL}) and linearization \eqref{equ:linear1}-\eqref{equ:linear2} and design a Benders decomposition styled algorithm to solve the overall problem \eqref{HP1-ambiguity-X} iteratively. We briefly describe the procedures as follows.
  1992. %
  1993. %
  1994. %To ease the exposition, we denote by
  1995. %$$
  1996. %c(\x):=\sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \left[\sum\limits_{k\in \mathbf{K}_{ij}} x^{k}_{ij}\right]c_{ij},
  1997. %$$
  1998. %and
  1999. %$$
  2000. %R(\x):=\inf\limits_{\P \in \mathbb{F}}\P\left[\min\limits_{i \in \mathbf{N}}\left(\sum_{j \in \mathbf{M}^{\rm c}_i}\sum_{k\in \mathbf{K}_{ij}} \tilde{z}^{k}_{ij}x^{k}_{ij} \bigvee \max_{j \in \mathbf{M}^{\rm a}_i}  \max_{k \in \mathbf{K}_{ij}}\tilde{z}^{k}_{ij}x^{k}_{ij}\right)> \mathcal{T}_R \right].
  2001. %$$
  2002. %If we applying the linearizations \eqref{equ:linear1}-\eqref{equ:linear2} to replace the bilinear terms $x_{l jk}q_{l jk}$ and $x_{l jk}p_{l}$ in the formulation (\ref{HP1-ambiguity-LP-FL})-(\ref{HP2-ambiguity-LP-FL})  of the worst-case probabilistic chance function (\ref{Prob-1}), we can re-express the chance function (\ref{Prob-1}) given $\x$ as the following compact form:
  2003. %$$
  2004. %R(\x)=\max_{\mathbf{A}\mathbf{d}=\mathbf{B}\mathbf{x}+\mathbf{r},\mathbf{d}\ge \mathbf{0}}\mathbf{d}'\mathbf{b},
  2005. %$$
  2006. %where $\mathbf{A}\mathbf{d}=\mathbf{B}\mathbf{x}+\mathbf{r},\mathbf{d}\ge \mathbf{0}$ represents the constraints in the problem (\ref{HP1-ambiguity-LP-FL})-(\ref{HP2-ambiguity-LP-FL}) after the linearization in a standardized form of linear program with $\mathbf{d}$ being the standardized decision variables, and $\mathbf{A},\mathbf{B}$ and $\b,\mathbf{r}$ being the matrix and vector coefficients or inputs with comfortable dimensions. Then by the duality argument, we can rewrite $R(\x)$ as
  2007. %\begin{equation}\label{algo-1}
  2008. %R(\x)=\min_{\mathbf{A}'\mathbf{g}\ge \mathbf{b}}\x'(\mathbf{B}'\mathbf{g})+ \mathbf{r}'\mathbf{g},
  2009. %\end{equation}
  2010. %where $\mathbf{g}$ are the dual variables. Hence the {\em relaxed master problem} of overall problem \eqref{HP1-ambiguity-X} can be expressed as
  2011. %\begin{equation}\label{algo-master}
  2012. %\begin{array}{rcll}
  2013. %& \min\limits_{\x} &  c(\x) \\[0.3 cm]
  2014. %& {\rm s.t.} & \x'(\mathbf{B}'\mathbf{g})+ \mathbf{r}'\mathbf{g} \ge R_0,& \forall \mathbf{g} \in \mathcal{G}\\
  2015. %&& L_{i} \le \sum\limits_{j\in \mathbf{M}_i}\sum\limits_{k\in \mathbf{K}_{ij}} x^{k}_{ij}\le  U_{i}, &\forall  i \in \mathbf{N} \\
  2016. %&& \x\in \{0,1\}^K,
  2017. %\end{array}
  2018. %\end{equation}
  2019. %where $\mathcal{G}\subset \mathcal{V}$ with $\mathcal{V}$ being the set of vertices of polyhedron $\{\g: \mathbf{A}'\mathbf{g}\ge \mathbf{b}\}$.  In particular, we have the following optimality condition:
  2020. %\begin{proposition}\label{optimality}
  2021. %Let $\x$ is a solution of the relaxed master problem \eqref{algo-master}, if $R(\x) \ge R_0$ then it is the optimal solution of \eqref{HP1-ambiguity-X}.
  2022. %\end{proposition}
  2023. %\begin{proof}
  2024. %Let $c^*$ be the optimal cost of overall problem \eqref{HP1-ambiguity-X}. Since $\x$ is a solution of relaxed master problem \eqref{algo-master}, we have $c(\x) \le c^*$. Now if $R(\x) \ge R_0$, $\x$ is also a feasible solution of \eqref{HP1-ambiguity-X}, it is then optimal.
  2025. %\end{proof}
  2026. %
  2027. %In our algorithm, for a given design $\x$, the linear program \eqref{algo-1} is used to establish the separation problem and optimality condition, where its solution $\mathbf{g}^*$ is used to generate cuts. Since $|\mathcal{V}|$ is finite, the algorithm can achieve the optimality in finite steps. The procedure of the algorithm is summarized as follows.
  2028. %
  2029. %\noindent\rule{\textwidth}{0.1em}\vspace{-5pt}\\
  2030. %\noindent {\bf A Benders Decomposition Algorithm for \eqref{HP1-ambiguity-X}.} \vspace{-10pt}\\
  2031. %\noindent\rule{\textwidth}{0.05em} \\
  2032. %{\bf Initialization:} $\mathcal{G}=\emptyset$.
  2033. %
  2034. %\begin{enumerate}
  2035. %\item  Solve the relaxed master problem \eqref{algo-master}, and obtain solution $\x$.
  2036. %\item  Compute $R(\x)$ with linear program \eqref{algo-1}:
  2037. %\begin{enumerate}
  2038. %  \item If $R(\x) \ge R_0$, then STOP and let $\x^*=\x$.
  2039. %  \item Otherwise, obtain solution $\mathbf{g}^*$ and update $\mathcal{G}=\mathcal{G} \cup \{\mathbf{g}^*\}$. Go to STEP 1.
  2040. %\end{enumerate}
  2041. %\end{enumerate}
  2042. %\noindent{\bf Output:} The optimal redundancy design $\x^*$.  
  2043. %
  2044. %\vspace{-5pt}
  2045. %\noindent\rule{\textwidth}{0.1em}\vspace{-17pt}\\
  2046. %%decision vector $\mathbf{d}$ represents the variables $\halpha, \hbeta, \hlambda, \tau, \q, \s, \y^{\rm a}, \y^{\rm c}, \hphi, \hvarphi, \hpi, \hvarpi, \htheta, \hvartheta, \hvarsigma$ and $\hvartheta$ of the,  $\r$ is the input vector with entries of $1, 0$ and $M$.
  2047. %
  2048. %
  2049. %%\begin{remark}\color{blue}
  2050. %%If we impose the independence of the lifetimes, then the problem becomes
  2051. %%\begin{eqnarray*}
  2052. %%&&\max_{\x}\P\left[\min\limits_{i \in \mathbf{N}} \left(\max\limits_{j \in \mathbf{M}_i}  \max\limits_{k \in \mathbf{K}_{ij}}   \tilde{z}^{k}_{ij} x^{k}_{ij}\right)> \mathcal{T}_R \right]\\
  2053. %%&=&\max_{\x}\prod_{i \in \mathbf{N}}\left[1-\prod_{j \in \mathbf{M}_i}\prod_{k \in \mathbf{K}_{ij}} \P\Big[\tilde{z}^{k}_{ij} x^{k}_{ij}\le \mathcal{T}_R \Big]\right]
  2054. %%\end{eqnarray*}
  2055. %%which is non-convex with respect to $\P\Big[\tilde{z}^{k}_{ij} x^{k}_{ij}\le \mathcal{T}_R \Big]$.
  2056. %%\end{remark}
  2057. %
  2058. %
  2059. %
  2060. %
  2061. %
  2062. %
  2063. %
  2064. %
  2065. %
  2066. %
  2067. %
  2068. %
  2069. %
  2070. %\section{Extreme Lifetime Distributions}\label{sec:ELD}
  2071. %In this section, we develop another model that can recover efficiently the extreme joint probability distribution of component lifetimes $\bm{\tilde{z}}$ that achieves the worst-case probability level (\ref{Prob-1}) for a given system design $\x$.
  2072. %
  2073. %\begin{proposition}\label{extreme}
  2074. %Given a system design $\x$, the extreme probability distribution of the component lifetimes  $\bm{\tilde{z}}$ can be constructed by the following discrete distribution with $|\mathbf{N}|$+1 scenarios:
  2075. %\begin{equation}\label{equ:extreme1-1H}
  2076. %\P^*\!\!\left[\tilde{z}^{k}_{ij}=\frac{b^*^{k}_{ij}}{p^*_i}, \forall j \in \mathbf{M}_i, k\in \mathbf{K}_{ij}; \tilde{z}_{ljk}=\frac{d^*_{iljk}}{p^*_i},\forall l\in \mathbf{N}\backslash \{i\}, j \in \mathbf{M}_l, k\in \mathbf{K}_{lj} \right]=p^*_i, \forall i \in \mathbf{N},
  2077. %\end{equation}
  2078. %and
  2079. %\begin{equation}\label{equ:extreme1-2H}
  2080. %\P^*\!\!\left[\tilde{z}^{k}_{ij}=\frac{s^*^{k}_{ij}}{p^*_{|\mathbf{N}|+1}}, \forall i \in \mathbf{N}, j \in \mathbf{M}_i, k\in \mathbf{K}_{ij}\right]=p^*_{|\mathbf{N}|+1},
  2081. %\end{equation}
  2082. %where $\b^*, \d^*, \s^*$ and $\p^*$ can be obtained by solving the following linear program:
  2083. %\begin{equation}\label{equ:extreme2H}\left.
  2084. % \begin{array}{rcll}
  2085. %& \max\limits_{\b,\d,\s,\p}  & \displaystyle \sum\limits_{i\in \mathbf{N}}p_i\\[0.2 cm]
  2086. %&{\rm s.t.} & \displaystyle \sum\limits_{i\in \mathbf{N}}p_i + p_{|\mathbf{N}|+1}=1 & \\[0.2 cm]
  2087. %&&\displaystyle b^{k}_{ij} x^{k}_{ij} \le p_i \mathcal{T}_R, \forall i \in \mathbf{N}, j\in \mathbf{M}^{\rm a}_{i}, k\in \mathbf{K}_{i j}   \\[0.2 cm]
  2088. %&&\displaystyle \sum_{j\in \mathbf{M}^{\rm c}_{i}}\sum_{k\in \mathbf{K}_{i j}}b^{k}_{ij} x^{k}_{ij} \le p_i \mathcal{T}_R,  \forall i \in \mathbf{N} \\[0.2 cm]
  2089. %&&\displaystyle \underline{\mu}_{ij}\le b^{k}_{ij}+\sum_{l\in \mathbf{N}\backslash \{i\}}d_{lijk} + s^{k}_{ij} \le \overline{\mu}_{ij}, \forall i \in \mathbf{N}, j\in \mathbf{M}_{i}, k\in \mathbf{K}_{i j}   \\[0.2 cm]
  2090. %&&\displaystyle \underline{z}_{ij}p_i\le b^{k}_{ij}\le  \overline{{z}}_{ij}p_i, \forall i \in \mathbf{N}, j\in \mathbf{M}_i, k\in \mathbf{K}_{ij} \\[0.3 cm]
  2091. %&&\displaystyle \underline{z}_{lj}p_i\le d_{iljk}\le  \overline{{z}}_{lj}p_i,  \forall i \in \mathbf{N}, l\in \mathbf{N}\backslash \{i\}, j\in \mathbf{M}_l, k\in \mathbf{K}_{lj}\\[0.2 cm]
  2092. %&&\displaystyle  \underline{{z}}_{{ij}}p_{|\mathbf{N}|+1} \le s^{k}_{ij}\le \overline{{z}}_{{ij}}p_{|\mathbf{N}|+1},  \forall i \in \mathbf{N}, j\in \mathbf{M}_i, k\in \mathbf{K}_{ij}  \\[0.2 cm]
  2093. %%&&\displaystyle  \sum_{k \in \mathbf{K}_{ij}}\frac{|b^{k}_{ij}-\nu_{ij}p_i|}{|\mathbf{K}_{ij}||\sigma_{ij}|}+\\
  2094. %%&&\displaystyle\sum_{l\in \mathbf{N}\backslash \{i\}}\sum_{k \in \mathbf{K}_{ij}}\frac{|d_{lijk}-\nu_{ij}p_l|}{|\mathbf{K}_{ij}||\sigma_{ij}|}+&\\
  2095. %%&&\displaystyle\sum_{k \in \mathbf{K}_{ij}}\frac{|s^{k}_{ij}-\nu_{ij}p_{|\mathbf{N}|+1}|}{|\mathbf{K}_{ij}||\sigma_{ij}|}\le \epsilon_{ij}, & \forall i \in \mathbf{N}, j\in \mathbf{M}_i\\
  2096. %&&\displaystyle  \sum_{k \in \mathbf{K}_{ij}} \xi^{k}_{ij}+\sum_{l\in \mathbf{N}\backslash \{i\}}\sum_{k \in \mathbf{K}_{ij}} \eta_{lijk}+ \sum_{k \in \mathbf{K}_{ij}} \bm{z}eta^{k}_{ij}\le \epsilon_{ij}, \forall i \in \mathbf{N}, j\in \mathbf{M}_{i}\\[0.2 cm]
  2097. %&& \displaystyle b^{k}_{ij}-\nu_{ij}p_i  -\xi^{k}_{ij}|\mathbf{K}_{ij}||\sigma_{ij}|\le 0,  \forall i \in \mathbf{N}, j\in \mathbf{M}_{i}, k\in \mathbf{K}_{ij} \\[0.2 cm]
  2098. %&& \displaystyle \nu_{ij}p_i- b^{k}_{ij} -\xi^{k}_{ij}|\mathbf{K}_{ij}||\sigma_{ij}|\le 0,  \forall i \in \mathbf{N}, j\in \mathbf{M}_{i}, k\in \mathbf{K}_{ij} \\[0.2 cm]
  2099. %&& \displaystyle  d_{lijk}-\nu_{ij}p_l -\eta_{lijk}|\mathbf{K}_{ij}||\sigma_{ij}|\le 0, \forall i \in \mathbf{N}, l\in \mathbf{N}\backslash \{i\},  j\in \mathbf{M}_{l}, k\in \mathbf{K}_{lj}\\[0.2 cm]
  2100. %&& \displaystyle \nu_{ij}p_l -d_{lijk} -\eta_{lijk}|\mathbf{K}_{ij}||\sigma_{ij}|\le 0, \forall i \in \mathbf{N}, l\in \mathbf{N}\backslash \{i\}, j\in \mathbf{M}_{l}, k\in \mathbf{K}_{lj}\\[0.2 cm]
  2101. %&& \displaystyle  s^{k}_{ij}-\nu_{ij}p_{|\mathbf{N}|+1}- \bm{z}eta^{k}_{ij}|\mathbf{K}_{ij}||\sigma_{ij}|\le 0, \forall i \in \mathbf{N}, j\in \mathbf{M}_{i}, k\in \mathbf{K}_{ij}  \\[0.2 cm]
  2102. %&& \displaystyle  \nu_{ij}p_{|\mathbf{N}|+1}-s^{k}_{ij}-\bm{z}eta^{k}_{ij}|\mathbf{K}_{ij}||\sigma_{ij}| \le 0, \forall i \in \mathbf{N}, j\in \mathbf{M}_{i}, k\in \mathbf{K}_{ij} \\[0.2 cm]
  2103. %&& \p\in \mathbb{R}^{|\mathbf{N}|+1}_+,  \b \in \mathbb{R}^{K}, \s \in \mathbb{R}^{K}, \d \in  \mathbb{R}^{K(|\mathbf{N}|+1)}, \hxi \in \mathbb{R}_+^{K}, \heta \in \mathbb{R}_+^{K(|\mathbf{N}|+1)}, \hzeta \in \mathbb{R}_+^{K},
  2104. %%&& \p\in \mathbb{R}^{|\mathbf{N}|+1}_+,  \b \in \mathbb{R}^{K}, \s \in \mathbb{R}^{K}, \d \in  \mathbb{R}^{K(|\mathbf{N}|+1)}.
  2105. %\end{array}
  2106. %\right\}
  2107. %\end{equation}
  2108. %where $\hxi,\heta$ and $\hzeta$ are auxiliary variables.
  2109. %\end{proposition}
  2110. %\begin{proof}
  2111. %For the given system design $\x$, recall that
  2112. %\begin{eqnarray*}
  2113. %\inf\limits_{\P \in \mathbb{F}}\P\Big[\tilde{\bm{z} }\in \mathcal{Z}_i(\x),\forall i \in \mathbf{N} \Big]=\P^*\!\Big[\tilde{\bm{z} }\in \mathcal{Z}_i(\x),\forall i \in \mathbf{N} \Big]
  2114. %&=&1-\P^*\!\Big[\cup_{i\in \mathbf{N} }\{\bm{\tilde{z}} \in \overline{\mathcal{Z}}_i(\x)\} \Big]\\
  2115. %&=&1-\sup\limits_{\P \in \mathbb{F}}\P\Big[\cup_{i\in \mathbf{N} }\{\bm{\tilde{z}} \in \overline{\mathcal{Z}}_i(\x)\} \Big].
  2116. %\end{eqnarray*}
  2117. %It suffices to prove that the probability distribution $\P^*$ obtained in (\ref{equ:extreme1-1H})-(\ref{equ:extreme1-2H}) is the extreme distribution for
  2118. %\begin{equation}\label{equ:sup-1H}
  2119. %\sup\limits_{\P \in \mathbb{F}}\P\Big[\cup_{i\in \mathbf{N} }\{\bm{\tilde{z}} \in \overline{\mathcal{Z}}_i(\x)\} \Big].
  2120. %\end{equation}
  2121. %In fact, by proposition~\eqref{P-proposition1b}, the above supremum (\ref{equ:sup-1H}) is equivalent to the optimal objective of
  2122. %$$
  2123. %\min\left\{\sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{k\in \mathbf{K}_{ij}} \left(\alpha^{k}_{ij}\underline{\mu}_{{ij}}+ \beta^{k}_{ij}\overline{\mu}_{{ij}}\right)+\sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i}\epsilon_{ij}\lambda_{ij} + \tau: (\ref{HP1-ambiguity-LP-FL1})-(\ref{HP2-ambiguity-LP-FL}) \right\},
  2124. %$$
  2125. %which by duality argument is also equivalent the optimal objective of (\ref{equ:extreme2H}). Now we are left to show the probability distribution (\ref{equ:extreme1-1H})-(\ref{equ:extreme1-2H}) formed by the solution $(\b^*, \d^*, \s^*,\p^*)$ of (\ref{equ:extreme2H}) is a qualified distribution in the ambiguity set (\ref{ambiguity-set}). This can be readily seen from the constraints in (\ref{equ:extreme2H}), by which the probability distribution $\mathbb{P}^*$ identified in (\ref{equ:extreme1-1H})-(\ref{equ:extreme1-2H}) satisfies the following constraints:
  2126. %$$
  2127. %\left\{ \begin{array}{rcll}
  2128. %&& \displaystyle \sum\limits_{i\in \mathbf{N}}p^*_i + p^*_{|\mathbf{N}|+1}=1 & \\[0.3 cm]
  2129. %%&&\displaystyle \left(\frac{b_{igjk}}{p_i}\right)x^g^{k}_{ij} \le \mathcal{T}_R, \forall i \in \mathbf{N}, g \in \mathbf{G}_i, j\in \mathbf{M}^{\rm a}_{ig}, k\in \mathbf{K}^g_{i j}   \\[0.3 cm]
  2130. %%&&\displaystyle \sum_{j\in \mathbf{M}^{\rm c}_{ig}}\sum_{k\in \mathbf{K}^g_{i j}}\left(\frac{b_{igjk}}{p_i }\right) x^g^{k}_{ij} \le \mathcal{T}_R,  \forall i \in \mathbf{N}, g \in \mathbf{G}_i \\[0.3 cm]
  2131. %&&\displaystyle \underline{\mu}_{ij}\le p^*_i\left[\frac{b^*^{k}_{ij}}{p^*_i}\right]+\sum_{l\in \mathbf{N}\backslash \{i\}} p^*_l\left[\frac{d^*_{lijk}}{p^*_l}\right] + p^*_{|\mathbf{N}|+1}\left[\frac{s^*^{k}_{ij}}{p^*_{|\mathbf{N}|+1}}\right] \le \overline{\mu}_{ij}, \forall i \in \mathbf{N},  j\in \mathbf{M}_{i}, k\in \mathbf{K}_{i j}   \\[0.3 cm]
  2132. %&&\displaystyle \underline{z}_{ij}\le \frac{b^*^{k}_{ij}}{p^*_i}\le  \overline{{z}}_{ij}, \forall i \in \mathbf{N}, j\in \mathbf{M}_{i}, k\in \mathbf{K}_{ij} \\[0.3 cm]
  2133. %&&\displaystyle \underline{z}_{lj} \le \frac{d^*_{iljk}}{p^*_i}\le  \overline{{z}}_{lj},  \forall i \in \mathbf{N}, l\in \mathbf{N}\backslash \{i\}, j\in \mathbf{M}_{l}, k\in \mathbf{K}_{lj}\\[0.3 cm]
  2134. %&&\displaystyle  \underline{{z}}_{{ij}} \le \frac{s^*^{k}_{ij}}{p_{|\mathbf{N}|+1}}\le \overline{{z}}_{{ij}},  \forall i \in \mathbf{N}, j\in \mathbf{M}_{i}, k\in \mathbf{K}_{ij}  \\[0.3 cm]
  2135. %&&\displaystyle  p^*_i\left[\sum_{k \in \mathbf{K}_{ij}}\frac{|b^*^{k}_{ij}/p^*_i-\nu_{ij}|}{|\mathbf{K}_{ij}||\sigma_{ij}|}\right]+\sum_{l\in \mathbf{N}\backslash \{i\}}p^*_l\left[\sum_{k \in \mathbf{K}_{ij}}\frac{|d^*_{lijk}/p^*_l-\nu_{ij}|}{|\mathbf{K}_{ij}||\sigma_{ij}|}\right]+&\\
  2136. %&&\displaystyle p^*_{|\mathbf{N}|+1}\left[\sum_{k \in \mathbf{K}_{ij}}\frac{|s^*^{k}_{ij}/p^*_{|\mathbf{N}|+1}-\nu_{ij}|}{|\mathbf{K}_{ij}||\sigma_{ij}|}\right]\le \epsilon_{ij}, \forall i \in \mathbf{N},  j\in \mathbf{M}_{i},
  2137. %\end{array}\right\}
  2138. %$$
  2139. %which is exactly a qualified distribution in the ambiguity set $\mathbb{F}$ in (\ref{ambiguity-set}).
  2140. %% of $\P^*\!\Big[\tilde{\bm{z} }\in \mathcal{Z}_i(\x),\forall i \in \mathbf{N} \Big]$, by the duality argument, we can have the extreme probability level of the complementary event
  2141. %%$$
  2142. %%\P^*\!\Big[\cup_{i\in \mathbf{N} }\{\bm{\tilde{z}} \in \overline{\mathcal{Z}}_i(\x)\} \Big]=\P^*\!\left[\bigcup_{i\in \mathbf{N} }\Big\{z^{k}_{ij}x^{k}_{ij}\le \mathcal{T}_R, \forall j \in \mathbf{M}_i, k\in \mathbf{K}_{ij}\Big\} \right]
  2143. %%$$
  2144. %%solves the optimization problem (\ref{equ:extreme2}). It can be seen that the problem (\ref{equ:extreme2}) solves exactly for the discrete probability distribution of $\bm{\tilde{z}}$ in (\ref{equ:extreme1-1})-(\ref{equ:extreme1-2}) with $|\mathbf{N}|+1$ scenarios that achieves the highest (extreme) probability level of event
  2145. %%$$
  2146. %%\bigcup_{i\in \mathbf{N} }\Big\{z^{k}_{ij}x^{k}_{ij}\le \mathcal{T}_R, \forall j \in \mathbf{M}_i, k\in \mathbf{K}_{ij}\Big\}
  2147. %%$$
  2148. %%and satisfies the distributional requirement in the ambiguity set (\ref{ambiguity-set}).
  2149. %\end{proof}
  2150. %
  2151. %
  2152. %
  2153. %\section{Computational Study}\label{CS3}
  2154. %
  2155. %In this section, we demonstrate the performance of the developed models by performing sufficient numerical experiments and a case study. The computational study consists of three parts: (i) Evaluating the worst-case probability level, i.e., model (\ref{HP1-ambiguity-LP-FL})-(\ref{HP2-ambiguity-LP-FL}), for a given system design under different  parameter settings. (ii) Robust reliability optimization using model (\ref{HP1-ambiguity-MILP-FL1})-(\ref{HP1-ambiguity-MILP-FL2}). (iii) Performance comparison with probabilistic RAP model using out-of-sample testing. (iv) A real-life case of hybrid cold-standby and parallel high speed train system.
  2156. %
  2157. %
  2158. %
  2159. %All experiments were run on a PC with an Intel Core(TM) i7-4510U processor at 2.60 Ghz. All  MIP models were coded in Python and solved using the callback library of Mosek version 8.1.0.34.
  2160. %
  2161. %
  2162. %
  2163. %
  2164. %
  2165. %
  2166. %
  2167. %
  2168. %\subsection{Computing the reliability level}\label{sec:comp1}
  2169. %In the parts (i) and (ii) of numerical experiments (Sections~\ref{sec:comp1} and~\ref{sec:comp2}), we consider the following setting: a series-parallel system with 5 subsystems ($|\mathbf{N}|=5$), and each subsystem contains 3 types of components ($|\mathbf{M}_i|\equiv|\mathbf{M}|= 3, \forall i \in \mathbf{N}$) where each type is of 3 redundant components ($|\mathbf{K}_{ij}|\equiv 3, \forall i \in \mathbf{N}, j\in \mathbf{M}_i$). Furthermore, we set the distributional parameters of component lifetimes $([\underline{\hmu}, \overline{\hmu}], [\underline{\bm{z}}, \overline{\bm{z}}], \bm{\hsigma}, \hepsilon)$ in the ambiguity set $\mathbb{F}$  as follows.
  2170. %
  2171. %We utilize parameters $\chi \ge 0$ and $t \ge 0$ to adjust the gap level of $[\underline{\mu}_{ij}, \overline{\mu}_{ij}]$ and $[\underline{z}_{ij}, \overline{z}_{ij}]$, respectively, as follows
  2172. %$$
  2173. %\underline{\mu}_{ij}=\underline{\mu}^o_{ij}-\chi(\overline{\mu}^o_{ij}-\underline{\mu}^o_{ij}),\quad \overline{\mu}_{ij}=\overline{\mu}^o_{ij}+\chi(\overline{\mu}^o_{ij}-\underline{\mu}^o_{ij})
  2174. %$$
  2175. %where $\underline{\mu}^o_{ij} \in \mathbf{U}[10,13], \overline{\mu}^o_{ij} \in \mathbf{U}[15,18], \forall i \in [5], j \in [3]$, and
  2176. %$$
  2177. %\underline{z}_{ij}=\underline{z}^o_{ij}-t(\overline{z}^o_{ij}-\underline{z}^o_{ij}),\quad \overline{z}_{ij}=\overline{z}^o_{ij}+t(\overline{z}^o_{ij}-\underline{z}^o_{ij}),
  2178. %$$
  2179. %where $\underline{z}^o_{ij} \in \mathbf{U}[5,10], \overline{z}^o_{ij} \in \mathbf{U}[30,50], \forall i \in [5], j \in [3]$. Also we let
  2180. %$$\sigma_{ij}=\kappa \sigma^o_{ij}, \quad \epsilon_{ij}=\theta\epsilon^o_{ij},$$
  2181. %where $\sigma^o_{ij} \in \mathbf{U}[4,5], \epsilon^o_{ij} \in \mathbf{U}[0, 0.5], \forall i\in [5], j\in [3]$ and $\kappa, \theta \ge 0$ are the adjustable parameters. Therefore, by changing the values of $(\chi,t,\kappa,\theta)$, we can have different sets of parameters $([\underline{\hmu}, \overline{\hmu}], [\underline{\bm{z}}, \overline{\bm{z}}], \bm{\hsigma}, \hepsilon)$. Specifically, large values of $(\chi,t)$ correspond to the large gaps of $[\underline{\mu}_{ij}, \overline{\mu}_{ij}]$ and $[\underline{z}_{ij}, \overline{z}_{ij}], \forall i \in [5], j \in [3]$, while large values of $(\kappa,\theta)$ correspond to large values of $\sigma_{ij}$ and $\epsilon_{ij}, \forall i \in [5], j \in [3].$ Finally, we set the system lifetime requirement $\mathcal{T}_R \in [16, 19]$.
  2182. %
  2183. %
  2184. %In this first part of numerical experiments, we fix a system design $\x$, such that
  2185. %$$
  2186. %\left[\sum_{k \in \mathbf{K}_{ij}}x^{k}_{ij}\right]_{|\mathbf{N}||\mathbf{M}|}=\left[
  2187. %\begin{matrix}
  2188. %1 & 1 & 1  \\
  2189. %1 & 1 & 1 \\
  2190. %2 & 1 & 1 \\
  2191. %1 & 1 & 1  \\
  2192. %1 & 2 & 2  \\
  2193. %\end{matrix}
  2194. %\right],
  2195. %$$
  2196. %where each subsystem (out of five) is assigned with one or two components for each type (out of three). Using model (\ref{HP1-ambiguity-LP-FL})-(\ref{HP2-ambiguity-LP-FL}), we compute the worst-case probability level (\ref{Prob-1}) with different settings of  parameters $([\underline{\hmu}, \overline{\hmu}], [\underline{\bm{z}}, \overline{\bm{z}}], \bm{\hsigma}, \hepsilon)$ in the ambiguity set $\mathbb{F}$ and the parameter $\mathcal{T}_R$ of system lifetime requirement. The worst-case probability levels under different testing purposes are provided in the Figure~\ref{fig:exp-1}.
  2197. %
  2198. %
  2199. %
  2200. %
  2201. %\begin{figure}[h!]
  2202. %   \centering
  2203. %   \includegraphics[width=1\linewidth]{exp-set-1}
  2204. %   \caption{\footnotesize The worst-case probability (reliability) level of a given system design with different settings of distributional parameters $([\underline{\hmu}, \overline{\hmu}], [\underline{\bm{z}}, \overline{\bm{z}}], \bm{\hsigma}, \hepsilon)$ and system lifetime requirement level $\mathcal{T}_R$. (a) The probability level vs. gap of $[\underline{\hmu}, \overline{\hmu}]$ under different $\mathcal{T}_R$ levels, where the gap enlarges as the parameter $\chi$ (`chi') increases. (b) The probability level vs. gap of $[\underline{\bm{z}}, \overline{\bm{z}}]$ under different $\mathcal{T}_R$ levels, where the gap enlarges as the parameter $t$ ('psi') increases. (c)The probability level vs. values of $\hsigma$ under different $\mathcal{T}_R$ levels, where the values of $\hsigma$ increase as the parameter $\kappa$ ('kappa') increases. (d) The probability level vs. values of $\hepsilon$ under different $\mathcal{T}_R$ levels, where the values of $\hepsilon$ increase as the parameter $\theta$ ('theta') increases.}
  2205. %   \label{fig:exp-1}
  2206. %\end{figure}
  2207. %
  2208. %The observations of the experimental results are the following:
  2209. %(i) the worst-case probability (reliability) level decreases when the variation range $[\underline{\hmu}, \overline{\hmu}]$ of expected lifetimes increases (Figure~\ref{fig:exp-1}-(a)), the variation range $[\underline{\bm{z}}, \overline{\bm{z}}]$ of the lifetimes increases (Figure~\ref{fig:exp-1}-(b)), or the difference in average among the lifetimes of the same type of components enlarges (Figure~\ref{fig:exp-1}-(c) and (d)). Such decreased probability is due to the enlarged ambiguity set $\mathbb{F}$ resulted by the changed of above distributional parameters $([\underline{\hmu}, \overline{\hmu}], [\underline{\bm{z}}, \overline{\bm{z}}], \bm{\hsigma}, \hepsilon)$. (ii) Intuitively, the probability level also decreases if the system lifetime requirement $\mathcal{T}_R$ is increased.
  2210. %(iii) The results in Figure~\ref{fig:exp-1} also provide the variation limits for the distributional parameters for an aspirational probability level with the given system lifetime requirement $\mathcal{T}_R$. For instance, if the decision maker needs the (worst-case) probability higher than 0.8 for system lifetime above $\mathcal{T}_R=16.2$, then the components lifetime range $[\underline{\bm{z}}, \overline{\bm{z}}]$ cannot be larger than the case such that parameter $t \ge 0.01$.
  2211.  
  2212.  
  2213. %\subsubsection{Experiment on $\underline{z}$ and $\overline{z}$}
  2214. %In this subsection, wo do experiments with different dispersion levels between the parameters $\bm{\underline{z}}$ and $\bm{\overline{z}}$. For simplicity, we also introduce an auxiliary variable $t\in[1,+\infty)$, and control the values of specified parameters as follow:
  2215. %\begin{eqnarray}
  2216. %\widetilde{\underline{z}}^{k}_{ij}' = \frac{1}{t} * \widetilde{\underline{z}}^{k}_{ij}&&\forall i\in I, j_i\in J_i, k_{ij}\in N_{ij}
  2217. %\end{eqnarray}
  2218. %\begin{eqnarray}
  2219. %\widetilde{\overline{z}}^{k}_{ij}' = t * \widetilde{\overline{z}}^{k}_{ij}&&\forall i\in I, j_i\in J_i, k_{ij}\in N_{ij}
  2220. %\end{eqnarray}
  2221. %where $\widetilde{\overline{z}}^{k}_{ij}'$ and $\widetilde{\overline{z}}^{k}_{ij}'$ are the experiment observation values of parameters $\underline{z}_{ij}$ and $\overline{z}_{ij}$, so the greater value of $t$, the greater of dispersion level between $\underline{z}_{ij}$ and $\overline{z}_{ij}$. The comparsion is shown in Figure~\ref{fig:xi}.
  2222. %
  2223. %\begin{figure}[h!]
  2224. %   \centering
  2225. %   \includegraphics[width=0.5\linewidth]{xi}
  2226. %   \caption[Obj value under $t$]{The objective values under different value of $t\in[1,2]$ of five $\mathcal{T}_R$ value cases ($\mathcal{T}_R\in\{15, 20, 25, 30, 35\}$), in each case, the result shows that the greater dispersion level between $\bm{\underline{z}}$ and $\bm{\overline{z}}$, the smaller obj value is.}
  2227. %   \label{fig:xi}
  2228. %\end{figure}
  2229. %
  2230. %\subsubsection{Experiment on $\mathcal{T}_R$}
  2231. %In this subsection, we do experiments with different $\mathcal{T}_R$ in candidate set $\{0, 5, 10, ..., 80, 85, 90\}$, the result is shown in Figure~\ref{fig:dl}.
  2232. %
  2233. %\begin{figure}[h!]
  2234. %   \centering
  2235. %   \includegraphics[width=0.5\linewidth]{Dl}
  2236. %   \caption[Obj value under $\mathcal{T}_R$]{The objective values under different value of parameter $\mathcal{T}_R$, with the value of $\mathcal{T}_R$ increase from 0 to 90, the objective value decrease from 1 to 0.}
  2237. %   \label{fig:dl}
  2238. %\end{figure}
  2239. %
  2240. %\subsubsection{Experiment on $\hsigma$}
  2241. %In this subsection, we do experiments with different value of $\bm{\sigma}$. For simplicity of experiments, we introduce an
  2242. %auxiliary variable $\kappa\in(0,+\infty)$ to control the value of $\sigma_{ij}$, $\forall i\in I$, $j_i\in J_i$:
  2243. %\begin{eqnarray}
  2244. %\widetilde{\sigma}_{ij}' = \kappa * \widetilde{\sigma}_{ij}&&\forall i\in I, j_i\in J_i, k_{ij}\in N_{ij}
  2245. %\end{eqnarray}
  2246. %where $\widetilde{\sigma}_{ij}'$ is the experiment observation values of parameters $\sigma_{ij}$ and $\widetilde{\sigma}_{ij}$ is the initial setting value of $\sigma_{ij}$.
  2247. %The result is shown in Figure~\ref{fig:sigma}.
  2248. %\begin{figure}[h!]
  2249. %   \centering
  2250. %   \includegraphics[width=0.5\linewidth]{sigma}
  2251. %   \caption[Obj value under $\bm{\sigma}$]{The objective values under different value of $\kappa\in[0,2.5]$ of nine $\mathcal{T}_R$ value cases ($\mathcal{T}_R\in\{10, 20, 30, 40, 50, 60, 70, 80, 90\}$), in each case, the result shows that the greater value of $\bm{\sigma}$ corresponds to the smaller objective value. All objective values will be 1 if $\mathcal{T}_R<10$.}
  2252. %   \label{fig:sigma}
  2253. %\end{figure}
  2254. %
  2255. %\subsubsection{Experiment on $\epsilon$}
  2256. %In this section, we do experiments with different values of parameter $\bm{\epsilon}$. Here we set
  2257. %$\epsilon_{ij}\in\{0, 0.05, 0.1, 0.15, 0.2, 0.25, 0.3, 0.35, 0.4, 0.45, 0.5\}$, $\forall i\in I$, $j_i\in J_i$ for simplicity and clear. The results is shown in Figure~\ref{fig:epsilon}.
  2258. %\begin{figure}[h!]
  2259. %   \centering
  2260. %   \includegraphics[width=0.5\linewidth]{epsilon}
  2261. %   \caption[Obj value under $\bm{\epsilon}$]{The objective values under different value of $\epsilon_{ij}$, $\forall i\in I$, $j_i\in J_i$ of nine $\mathcal{T}_R$ value cases ($\mathcal{T}_R\in\{10, 20, 30, 40, 50, 60, 70, 80, 90\}$), in each case, the result shows that the greater value of $\epsilon_{ij}$ corresponds to the smaller objective value.}
  2262. %   \label{fig:epsilon}
  2263. %\end{figure}
  2264. %%
  2265. %
  2266. %
  2267.  
  2268.  
  2269. %\subsection{Reliability Optimization}\label{sec:comp2}
  2270. %Next we perform the robust redundancy optimization using  model (\ref{HP1-ambiguity-MILP-FL1})-(\ref{HP1-ambiguity-MILP-FL2}) to find the optimal system design $\x^*$ with different combinations $(\mathcal{T}_R,R_0)$ of system lifetime requirement level and reliability requirement level. Specifically, we set the minimum number of components of each type as $L_{ij}=1, \forall i \in [5], j \in [3]$, and allow the $\mathcal{T}_R$ to vary from 16.5 to 19.0, and $R_0$ from 0.1 to 0.9.  The  optimal (minimal) cost is plotted in Figure~\ref{fig:exp-2}.
  2271. %\begin{figure}[h!]
  2272. %   \centering
  2273. %   \includegraphics[width=0.6\linewidth]{Figure_R0}
  2274. %   \caption{\footnotesize  Optimal cost with different system lifetime requirement $\mathcal{T}_R$ and reliability requirement $R_0$, where the vertical line implies the case of infeasibility.}
  2275. %   \label{fig:exp-2}
  2276. %\end{figure}
  2277. %
  2278. %From the results, the optimal cost and the number of components increase as the $\mathcal{T}_R$ or $R_0$ increases. In particular, Figure~\ref{fig:exp-2} implies, for each system lifetime requirement $\mathcal{T}_R$, there are several threshold levels of system reliability requirement $R_0$ beyond which no system design would be feasible. For instance, for $\mathcal{T}_R=18$, there is no feasible system design for the reliability level $R_0>0.7$, while when $\mathcal{T}_R$ is raised to $18.5$, there would be no feasible system design can be found even for the reliability level $R_0>0.2$. In other words, Figure~\ref{fig:exp-2} provides the physical limits of both the lifetime and reliability requirements for the system to meet, under the lifetime distributional ambiguity of components.
  2279. %
  2280. %%On the other hand, the Table~\ref{table-design} provides with the system design information (and also the cost information) with different combinations of $(\mathcal{T}_R, R_0)$. The designer can then choose the most suitable system designs with these information.
  2281. %%\begin{table}[h]\scriptsize%\tiny%\footnotesize
  2282. %%\begin{center}
  2283. %%  \caption{\footnotesize The optimal system redundancy design $\x^*$: the total number of components (of all types) in each  subsystem along with different system lifetime requirement $\mathcal{T}_R$ and reliability requirement $R_0$, where `$-$' indicates a no-solution case due to the overly high $\mathcal{T}_R$ and/or $R_0$. }\label{table-design}
  2284. %%  \begin{tabular}{|c|c||cccccccc c|}
  2285. %%      \hline
  2286. %%  %   \rule{0pt}{12pt} & && & & & $R_0$ & & & & \\[0pt]       \hline
  2287. %%      \multirow{2}{*}{\scriptsize  System Lifetime $\mathcal{T}_R$} & \multirow{2}{*}{\scriptsize  Subsystem $i$} & \multicolumn{9}{c|}{$R_0$}\\
  2288. %%\cline{3-11}
  2289. %%        && 0.1 & 0.2 & 0.3 & 0.4 & 0.5 & 0.6 & 0.7 & 0.8 & 0.9 \\[0pt]       
  2290. %%\hline
  2291. %%      % DL = 16.5
  2292. %%                  &1& 3 & 3 & 3 & 3 & 3 & 3 & 3 & 3 & 4 \\[0pt]
  2293. %%                &2& 3 & 3 & 3 & 3 & 3 & 3 & 3 & 3 & 3 \\[0pt]
  2294. %%        $16.5$  &3& 3 & 3 & 3 & 3 & 3 & 3 & 3 & 3 & 3 \\[0pt]
  2295. %%                &4& 3 & 3 & 3 & 3 & 3 & 3 & 3 & 3 & 3 \\[0pt]
  2296. %%                &5& 3 & 3 & 3 & 3 & 3 & 3 & 3 & 3 & 3 \\
  2297. %%      \hline
  2298. %%        % \multicolumn{2}{|c|}{Cost} &&&&&&&&--&-- \\ \hline
  2299. %%      % DL = 17.0
  2300. %%                &1 & 3 & 3 & 4 & 3 & 4 & 5 & 4 & \multirow{5}{*}{--} & \multirow{5}{*}{--}  \\[0pt]
  2301. %%                &2 & 3 & 3 & 3 & 4 & 4 & 4 & 5 &   &   \\[0pt]
  2302. %%       $17.0$   &3 & 3 & 3 & 3 & 3 & 3 & 3 & 3 &   &   \\[0pt]
  2303. %%                &4 & 3 & 3 & 3 & 3 & 3 & 3 & 3 &   &   \\[0pt]
  2304. %%                &5 & 3 & 3 & 3 & 3 & 3 & 3 & 4 &   &   \\
  2305. %%      \hline
  2306. %%      %\multicolumn{2}{|c|}{Cost} &&&&&&&&--&-- \\ \hline
  2307. %%                &1& 3 & 3 & 4 & 4 & 4 & 4 & 5 & \multirow{5}{*}{--} & \multirow{5}{*}{--} \\[0pt]
  2308. %%                &2& 3 & 3 & 3 & 3 & 4 & 4 & 5 &   &   \\[0pt]
  2309. %%        $17.5$  &3& 3 & 3 & 3 & 3 & 3 & 3 & 3 &   &   \\[0pt]
  2310. %%                &4& 3 & 3 & 3 & 3 & 3 & 3 & 3 &   &   \\[0pt]
  2311. %%                &5& 3 & 3 & 3 & 3 & 3 & 4 & 4 &   &   \\
  2312. %%      \hline
  2313. %%      %\multicolumn{2}{|c|}{Cost} &&&&&&&&--&-- \\ \hline
  2314. %%                &1& 3 & 4 & 3 & 4 & 5 & 5 & 5 & \multirow{5}{*}{--}  & \multirow{5}{*}{--} \\[0pt]
  2315. %%                &2& 3 & 3 & 4 & 4 & 4 & 4 & 5 &   &   \\[0pt]
  2316. %%       $18.0$   &3& 3 & 3 & 3 & 3 & 3 & 3 & 3 &   &   \\[0pt]
  2317. %%                &4& 3 & 3 & 3 & 3 & 3 & 3 & 3 &   &   \\[0pt]
  2318. %%                &5& 3 & 3 & 3 & 3 & 3 & 4 & 5 &   &   \\
  2319. %%      \hline
  2320. %%      %\multicolumn{2}{|c|}{Cost} &&&&&&&&--&-- \\ \hline
  2321. %%                &1& 4 & 5 & \multirow{5}{*}{--} & \multirow{5}{*}{--} & \multirow{5}{*}{--} & \multirow{5}{*}{--} & \multirow{5}{*}{--} & \multirow{5}{*}{--} & \multirow{5}{*}{--} \\[0pt]
  2322. %%                &2& 5 & 5 &   &   &   &   &   &   &   \\[0pt]
  2323. %%       $18.5$   &3& 5 & 5 &   &   &   &   &   &   &   \\[0pt]
  2324. %%                &4& 4 & 5 &   &   &   &   &   &   &   \\[0pt]
  2325. %%                &5& 4 & 5 &   &   &   &   &   &   &   \\
  2326. %%      \hline
  2327. %%%\multicolumn{2}{|c|}{Cost} &&&&&&&&--&-- \\ \hline
  2328. %%                &1& 5 & \multirow{5}{*}{--} & \multirow{5}{*}{--} & \multirow{5}{*}{--} & \multirow{5}{*}{--} & \multirow{5}{*}{--} & \multirow{5}{*}{--} & \multirow{5}{*}{--} & \multirow{5}{*}{--} \\[0pt]
  2329. %%                &2& 5 &   &   &   &   &   &   &   &   \\[0pt]
  2330. %%        $19.0$  &3& 5 &   &   &   &   &   &   &   &   \\[0pt]
  2331. %%                &4& 5 &   &   &   &   &   &   &   &   \\[0pt]
  2332. %%                &5& 5 &   &   &   &   &   &   &   &   \\
  2333. %%      \hline
  2334. %%%\multicolumn{2}{|c|}{Cost} &&&&&&&&--&-- \\ \hline
  2335. %%  \end{tabular}
  2336. %%\end{center}
  2337. %%\end{table}
  2338. %
  2339. %%\subsubsection{Performance comparison with probabilistic reliability model}
  2340. %%\begin{equation}\label{RO-Model}
  2341. %%\begin{array}{rcll}
  2342. %%& \min\limits_{\x} &  \sum\limits_{i\in \mathbf{N}} \sum\limits_{j \in \mathbf{M}_i}\left[ \sum\limits_{k\in \mathbf{K}_{ij}} x^{k}_{ij}\right]c_{ij} \\[0.3 cm]
  2343. %%&{\rm s.t.} &  \min\limits_{i \in \mathbf{N}} \left(\max\limits_{j \in \mathbf{M}_i}  \max\limits_{k \in \mathbf{K}_{ij}}   z^{k}_{ij} x^{k}_{ij}\right)\ge \mathcal{T}_R   & \forall \bm{z} \in \mathcal{U}(\Gamma) \\[0.3 cm]
  2344. %%&&  \sum\limits_{k\in \mathbf{K}_{ij}} x^{k}_{ij}\ge  L_{ij}, & \forall i \in \mathbf{N}, j\in \mathbf{M}_i\\
  2345. %%&& x^{k}_{ij} \in \{0,1\},  & \forall i \in \mathbf{N}, j\in \mathbf{M}_i, k \in \mathbf{K}_{ij}.
  2346. %%\end{array}
  2347. %%\end{equation}
  2348. %%The above budgeted robust convex (nonlinear) optimization model is in general intractable even when uncertainty set $\mathcal{U}(\Gamma)$ is a general polytope, by noting that the robust counterpart of convex (piecewise linear) constraints
  2349. %%$$
  2350. %%\max\limits_{j \in \mathbf{M}_i}  \max\limits_{k \in \mathbf{K}_{ij}}   z^{k}_{ij} x^{k}_{ij} \ge \mathcal{T}_R, \forall i \in \mathbf{N}
  2351. %%$$
  2352. %%cannot be tractably dualized.  Nevertheless, under the independence condition of component lifetimes $\tilde{z}^{k}_{ij}, \forall i \in \mathbf{N}, j\in \mathbf{M}_i, k \in \mathbf{K}_{ij}$, if we define
  2353. %%$$
  2354. %%\mathcal{U}(\gamma):=\left\{\tilde{z}^{k}_{ij}\in \Big[\underline{{z}}_{ij}(\gamma), \overline{{z}}_{ij}(\gamma) \Big], \forall i \in \mathbf{N}, j\in \mathbf{M}_i, k \in \mathbf{K}_{ij} \right\},
  2355. %%$$
  2356. %%then the above regular robust reliability model can be equivalently expressed as the following MIP problem:
  2357. %%\begin{equation}\label{RO-Model}
  2358. %%\begin{array}{rcll}
  2359. %%& \min\limits_{\x} &  \sum\limits_{i\in \mathbf{N}} \sum\limits_{j \in \mathbf{M}_i}\left[ \sum\limits_{k\in \mathbf{K}_{ij}} x^{k}_{ij}\right]c_{ij} \\[0.3 cm]
  2360. %%&{\rm s.t.} &   \max\limits_{j \in \mathbf{M}_i}  \max\limits_{k \in \mathbf{K}_{ij}}  \underline{{z}}_{ij}(\gamma) x^{k}_{ij} \ge \mathcal{T}_R,   & \forall i \in \mathbf{N} \\[0.3 cm]
  2361. %%&&  \sum\limits_{k\in \mathbf{K}_{ij}} x^{k}_{ij}\ge  L_{ij}, & \forall i \in \mathbf{N}, j\in \mathbf{M}_i\\
  2362. %%&& x^{k}_{ij} \in \{0,1\},  & \forall i \in \mathbf{N}, j\in \mathbf{M}_i, k \in \mathbf{K}_{ij}.
  2363. %%\end{array}
  2364. %%\end{equation}
  2365. %\subsection{Out-of-sample performance comparison with the probabilistic model}\label{sec:test-1}
  2366. %To further illustrate the performance of our robust reliability model, we compare the design obtained from the proposed robust redundancy optimization  model (\ref{HP1-ambiguity-MILP-FL1})-(\ref{HP1-ambiguity-MILP-FL2}), termed {\em robust design} with the design of the probabilistic redundancy optimization model, termed {\em probabilistic design}. As mentioned in the Introduction, when the situation  involves multiple types (i.e. $|\mathbf{M}_i|>1$), or both the cold-standby and active parallel redundant subsystems are considered, the probabilistic model in general becomes intractable (as mentioned in the sections of Introduction and Literature Review). Therefore, for a fair comparison, we consider a series-parallel system with $|\mathbf{M}_i|=1, \forall i \in [5]$, which preserves a linear MIP formulation. For a coherent exposition of the experimental study, we place the details of the probabilistic redundancy model as well as its MIP transformation in the Appendix.
  2367. %
  2368. %In particular, we first randomly generate lifetime samples (size=100), and then  compute the probability levels $\P[\tilde{z}_{i}\le \mathcal{T}_R ], \forall i \in [5]$ and the parameters $([\underline{\hmu}, \overline{\hmu}], [\underline{\bm{z}}, \overline{\bm{z}}], \bm{\hsigma}, \hepsilon)$ from the generated lifetime samples  for parameter inputs of probabilistic and robust models, respectively, where $R_0=0.9$ and $\mathcal{T}_R=18,19$ and $20$. We obtain probabilistic and robust designs by solving the perspective redundancy models. Then we perform 1000 sets of out-of-sample experiments where in each set, 1000 lifetime data points are generated using the sample mean and StD of the original lifetime samples. The out-of-sample system lifetimes as well as its mean values of both probabilistic and robust designs are compared and plotted in Figure~\ref{Fig:lifetime-compare1}, and the comparison of out-of-sample reliability levels is provided in Table~\ref{tab-compare2}.
  2369. %
  2370. %
  2371. %
  2372. %
  2373. %
  2374. %
  2375. %\begin{figure}[h*]
  2376. %  \centering
  2377. %  \subfigure[]{\includegraphics[scale=0.35]{lifetime-1}}\!\!\!\!\!\!\!\!
  2378. %  \subfigure[]{\includegraphics[scale=0.35]{lifetime-2}}\!\!\!\!\!\!\!\!
  2379. %  \subfigure[]{\includegraphics[scale=0.35]{lifetime-3}}\!\!\!\!\!\!\!\!
  2380. %  \subfigure[]{\includegraphics[scale=0.35]{mean-lifetime-1}}\!\!\!\!\!\!\!\!
  2381. %  \subfigure[]{\includegraphics[scale=0.35]{mean-lifetime-2}}\!\!\!\!\!\!\!\!
  2382. %  \subfigure[]{\includegraphics[scale=0.35]{mean-lifetime-3}}\!\!\!\!\!\!\!\!
  2383. %\caption{\footnotesize (a)-(c) The out-of-sample system lifetime scenarios (in 1 set of experiments with 1000 lifetime data) of probabilistic and robust design under $\mathcal{T}_R=18,19$ and $20$, respectively.
  2384. %(d)-(f) The (sample) mean values of out-of-sample system lifetimes (in 1000 sets of experiments) of probabilistic and robust design under $\mathcal{T}_R=18,19$ and $20$, respectively.}
  2385. %\label{Fig:lifetime-compare1}
  2386. %\end{figure}
  2387. %
  2388. %
  2389. %
  2390. %\begin{table}[h*]\footnotesize%\small%
  2391. %\caption{\label{tab-compare2} \footnotesize  Out-of-sample reliability levels comparison ($R_0=0.9$), where the `Design' specifies the number of redundant components allocated in each of 5 subsystems, `P-Model' and `C-DRO-Model' refer to the probabilistic model and conditionally robust model, respectively. }
  2392. %\begin{center}
  2393. %\begin{tabular}{|c|c|| c| c |c |c   |}\hline
  2394. %    $\mathcal{T}_R$ &Model & Design &   Designed reliability level &Mean of out-of-sample reliability level & StD \\
  2395. %     \hline
  2396. %    \multirow{2}{*}{18}& P-Model & (1,1,1,1,1) & \multirow{2}{*}{$R_0=0.9$}& 0.617 & 0.015   \\
  2397. %     & C-DRO-Model & (2,1,2,2,2) &  &0.873 & 0.010  \\
  2398. %\hline
  2399. %    \multirow{2}{*}{19}& P-Model & (2,1,1,1,1)&\multirow{2}{*}{$R_0=0.9$} & 0.564 & 0.015   \\
  2400. %     & C-DRO-Model & (2,1,2,2,2) && 0.884 & 0.009  \\
  2401. %\hline
  2402. %    \multirow{2}{*}{20}& P-Model & (2,1,1,3,2) &\multirow{2}{*}{$R_0=0.9$}& 0.554 & 0.016   \\
  2403. %     & C-DRO-Model & (2,2,3,3,2) && 0.909 & 0.009  \\
  2404. %\hline
  2405. %\end{tabular}
  2406. %\end{center}
  2407. %\end{table}
  2408. %
  2409. %
  2410. %From the comparative results in Figure~\ref{Fig:lifetime-compare1} and Table~\ref{tab-compare2}, the robust designs in all the different system lifetime requirements ($\mathcal{T}_R=18,19$ and $20$)  provide more reliable performance profiles than the probabilistic designs, in protecting the reliability level $R_0$. Specifically, in each set of out-of-sample tests, compared with the probabilistic design, the robust design pulls the sample system lifetimes (Figure~\ref{Fig:lifetime-compare1}:(a)-(c)) and sample mean of the system lifetimes (Figure~\ref{Fig:lifetime-compare1}:(d)-(f)) further away from the lifetime requirement $\mathcal{T}_R$ the to the right side, forming a wider safety gap which is able to absorb more lifetime uncertainty. Furthermore, from Table~\ref{tab-compare2}, the sample mean of reliability levels achieved by robust design is $0.873, 0.884$ and $0.909$ under different $\mathcal{T}_R$ levels which is very close or exceeds the designed level $R_0=0.9$ of reliability requirement. In contrast, the probabilistic design can only achieve $0.617, 0.564$ and $0.554$ in perspective $\mathcal{T}_R$ levels, which fall to meet the designed reliability level. Also, reliability levels achieved by robust design have a smaller variance than that of probabilistic design.
  2411. %On the other hand, although it is noted from the specifications of the `Design' column in Table~\ref{tab-compare2} that the higher performance achieved by the robust design is at the cost of using more components, our distributional robust design is not {\em ad-hoc}, but is calibrated by the distributional information of component lifetimes as given in the ambiguity set. In other words, it provides a more informed decision of redundancy allocation that can ensure the designed system reliability level given the component lifetime distributions being consistent to the characteristics of the ambiguity set.
  2412. %
  2413. %
  2414. %
  2415. %The implications of the comparison is that when the component lifetimes are highly uncertain, it could be risky to implement the probabilistic design (from the redundancy model that assumes known component reliability levels), since the actual reliability level achieved could be far below the designed level. In this situation, the robust design, albeit with more conservative redundancy allocation, is able to protect (or at least be very close to) the designed reliability level. Hence, it effectively reduces the risk of the system being `unreliable'.
  2416. %%\begin{eqnarray*}
  2417. %%\begin{array}{rcll}
  2418. %%& \min\limits_{\y} &  \sum\limits_{i\in \mathbf{N}} \sum\limits_{j \in \mathbf{M}_i}\left[y_{ij}\right]c_{ij} \\[0.3 cm]
  2419. %%& {\rm s.t.} &  \displaystyle \prod_{i \in \mathbf{N}}\left[1-\prod_{j \in \mathbf{M}_i}(r_{ij})^{y_{ij}}\right]\ge R_0  & \\[0.3 cm]
  2420. %%&& L_{ij}\le y_{ij}\le U_{ij}, & \forall i \in \mathbf{N}, j\in \mathbf{M}_i\\
  2421. %%&& y_{ij} \in \mathbb{Z}_+,  & \forall i \in \mathbf{N}, j\in \mathbf{M}_i,
  2422. %%\end{array}
  2423. %%\end{eqnarray*}
  2424. %%where
  2425. %%$$
  2426. %%r_{ij}=\P\Big[\tilde{z}_{ij}\le \mathcal{T}_R \Big],
  2427. %%$$
  2428. %%which can be estimated from the data. By transforming the integer variable $y_{ij}$ with binaries $x^{k}_{ij}$:
  2429. %%$$
  2430. %%y_{ij}=L_{ij}+\sum_{k=0}^{U_{ij}-L_{ij}}kx^{k}_{ij},~\mbox{with}~\sum_{k=0}^{U_{ij}-L_{ij}}x^{k}_{ij}=1
  2431. %%$$
  2432. %%The above model can be readily linearized as the following MIP with binaries:
  2433. %%\begin{eqnarray}
  2434. %%\begin{array}{rcll}
  2435. %%& \min\limits_{\x} & \displaystyle \sum\limits_{i\in \mathbf{N}} \sum\limits_{j \in \mathbf{M}_i}\left[L_{ij}+\sum_{k=0}^{U_{ij}-L_{ij}}kx^{k}_{ij}\right]c_{ij} \\[0.3 cm]
  2436. %%& {\rm s.t.} &  \displaystyle \sum_{i \in \mathbf{N}} \sum_{k=0}^{U_{ij}-L_{ij}}x^{k}_{ij}\ln\left[1-\prod_{j \in \mathbf{M}_i} (r_{ij})^{L_{ij}+k} \right]\ge \ln R_0  & \\[0.3 cm]
  2437. %%&& \displaystyle \sum_{k=0}^{U_{ij}-L_{ij}}x^{k}_{ij}=1, & \forall i \in \mathbf{N}, j\in \mathbf{M}_i\\
  2438. %%&& x^{k}_{ij} \in \{0,1\},  & \forall i \in \mathbf{N}, j\in \mathbf{M}_i, k \in \mathbf{K}_{ij}
  2439. %%\end{array}
  2440. %%\end{eqnarray}
  2441. %
  2442. %
  2443. %
  2444. %
  2445. %
  2446. %\subsection{A case of air brake system in high speed train}\label{sec:rlcs}
  2447. %A high speed train typically consists of several electric multiple units (EMUs) which are self-propelled carriages using electricity as the motive power. The air brake system is one critical subsystem of the EMU \cite{Hasegawa1999}. As shown in Figure \ref{figure-cs}, it mainly consists of one air compressor which generates compressed air, one check valve which ensures single direction of the airflow, one main air reservoir which stores the compressed air for braking operation, one relay valve which controls the air pressure according to the received signal and one boosting cylinder which transfers the air pressure to the amplified oil pressure to active the basic braking mechanisms to slow down or stop the unit \cite{Hasegawa1999,Cheng2009}.
  2448. %
  2449. %\begin{figure}[h]
  2450. %\centering
  2451. %\includegraphics[scale=0.5]{figure-case-study.pdf}
  2452. %\caption{\small Diagram of the air brake system of a EMU}
  2453. %\label{figure-cs}
  2454. %\end{figure}
  2455. %
  2456. %The high speed train is highly reliability-demanding, since one small accident would result to large number of passenger life losses given a high operational speed. To enhance the reliability of the air brake system, different types of redundancies, e.g. cold standby or hot standby, need to be introduced. For the two valves cold standbies are used, whereas for the rest components hot standbies are needed. The parameters of the various components are presented in Table \ref{tab-cs}, which are modified based the real data from a railway company in Europe, due to confidential agreement. Due to the space limit, the maximal number of components connected in parallel can not excess 4 and there must be at least one component in each subsystem. This system is required to operate without failure for 15-20 years at probability of 0.95-0.99.
  2457. %
  2458. %\begin{table}[t*]\scriptsize%\footnotesize%\small%
  2459. %\caption{\label{tab-cs} \footnotesize  Lifetime parameters of the available components for air brake system.}
  2460. %\begin{center}
  2461. %\begin{tabular}{|c| c | c || c c c  c |}\hline
  2462. %     Subsystem & Redundancy Strategy & Component Type & $\underline{z}_{ij}$ (yrs) & $\overline{z}_{ij}$ (yrs) &$\mu_{ij}$ (yrs) &  $c_{ij}$ (k\$) \\
  2463. %     \hline
  2464. %                    &\multirow{5}{*}{Active redundant} & 1 & 10 & 23 &15 & 10.0  \\
  2465. %                    & &2 & 12 & 21 &16& 15.0 \\
  2466. %     Air compressor & &3 & 16 & 24 &20& 17.0   \\
  2467. %                    & &4 & 17 & 25 &22& 25.0  \\
  2468. %                    & &5 & 20 & 26 &23& 32.0 \\
  2469. %     \hline
  2470. %                    &\multirow{5}{*}{Cold Standby} &1 & 4 & 6 &10& 1.0   \\
  2471. %                    & &2 & 6 & 10.5 &8& 1.5 \\
  2472. %      Check valve   & &3 & 7.5 & 11 &9& 2.0   \\
  2473. %                    & &4 & 8.5 & 12 &10& 2.5 \\
  2474. %                    & &5 & 15 & 17.5 &16& 3.0   \\
  2475. %     \hline
  2476. %                    &\multirow{6}{*}{Cold Standby} &1 & 4.5 & 8 &6& 1.5  \\
  2477. %                    & &2 & 6 & 9 &7& 2.0   \\
  2478. %   Control valve    & &3 & 7.5 & 10 &9& 2.5  \\
  2479. %                    & &4 & 8 & 11 &10& 3.0  \\
  2480. %                    & &5 & 10 & 12 &10.5& 3.5  \\
  2481. %                    & &6 & 10.5 & 12.5 &12& 4.0  \\
  2482. %\hline
  2483. %                    &\multirow{5}{*}{Active redundant} &1  & 15  & 21 &18& 15.0 \\
  2484. %                    & &2  & 17  & 22 &20& 20.5 \\
  2485. % Main air reservoir && 3  & 18  & 23 &21& 23.5 \\
  2486. %                    & &4  & 19  & 26 &23& 25.5 \\
  2487. %                    & &5  & 20  & 27 &25& 30.0 \\
  2488. % \hline
  2489. %                    &\multirow{5}{*}{Active redundant} &1  & 14  & 20 &16& 20.0 \\
  2490. %                    & &2  & 16  & 21 &18& 22.5 \\
  2491. %  Boosting cylinder & &3  & 18  & 24 &22& 25.5 \\
  2492. %                    & &4  & 20  & 28 &24& 30.5 \\
  2493. %                    & &5  & 25  & 31 &28& 34.0 \\
  2494. %\hline
  2495. %\end{tabular}
  2496. %\end{center}
  2497. %\end{table}
  2498. %
  2499. %Implementing the proposed robust redundancy allocation model, we can obtain the design table (Table~\ref{d-table}) which provides with the system design information (and also the cost information) with different combinations of $(\mathcal{T}_R, R_0)$. The designer can then choose the most suitable system designs with these information. From Table~\ref{d-table}, we see that as the system lifetime requirement $\mathcal{T}_R$ or system reliability level $R_0$ increases, the design allocates more components (e.g. the number of components allocated to `Air compressor' is increased from 1 to 2 as $\mathcal{T}_R$ increases from 19 yrs to 20 yrs, under $R_0=0.95$) and/or shifts the components to more reliable ones (with longer expected lifetimes, e.g. the 1 component allocated to `Main air reservoir' is shifted from  Type 2 to Type 5, as $\mathcal{T}_R$ increases from 16 yrs to 19 yrs, under $R_0=0.99$).
  2500. %
  2501. %\begin{sidewaystable}[!htbp]\scriptsize%\footnotesize%\small%
  2502. %\caption{\label{tab-cs} \footnotesize  Design table for the air brake system: The number of components of each allocated in each subsystem for $R_0=0.95, 0.97$ and $0.99$, and $\mathcal{T}_R=15-20$ yrs, where no feasible design can be found for $R_0=0.99, \mathcal{T}_R=20$.}\label{d-table}
  2503. %\begin{center}
  2504. %\begin{tabular}{|c| c|| cccccc || cccccc || cccccc |}\hline
  2505. %\multirow{2}{*}{Subsystem} & \multirow{2}{*}{Type} & \multicolumn{6}{c||}{$R_0=0.95$,~$\mathcal{T}_R$ (yrs) }& \multicolumn{6}{c||}{$R_0=0.97$,~$\mathcal{T}_R$ (yrs)}& \multicolumn{6}{c|}{$R_0=0.99$,~$\mathcal{T}_R$ (yrs)} \\
  2506. % \cline{3-20}
  2507. %      &&15 & 16 & $17$ &  $18$& $19$& $20$&$15$ & $16$ & $17$ &  $18$& $19$& $20$&$15$ & $16$ & $17$ &  $18$& $19$ & 20 \\
  2508. %     \hline
  2509. %                    & 1 & 0 & 0 &0 & 0 & 0& 0   & 0 & 0 &0 & 0& 0 & 0   & 0 & 0 &0 & 0& 0 &\multirow{5}{*}{--} \\
  2510. %      Air           & 2 & 0 & 0 &0 & 0 & 0& 0   & 0 & 0 &0 & 0&0  & 0   & 0 & 0 &0& 0 &0 &\\
  2511. %     compressor     & 3 & 1 & 1 &1 & 1 & 0& 0   & 1 & 1 &0 & 0&0  & 0   & 1 & 0 &0& 0 &0 & \\
  2512. %                    & 4 & 0 & 0 &0 & 0 & 1& 2   & 0 & 0 &1 & 1&0  & 0   & 0 & 1 &0& 0 &0 &\\
  2513. %                    & 5 & 0 & 0 &0 & 0 & 0& 0   & 0 & 0 &0 & 0&1  & 2   & 0 & 0 &1& 1 &1 &\\
  2514. %     \hline
  2515. %                    & 1 & 1 & 1 &1& 1 & 0 & 0   & 1 & 1 &1& 1 & 0 & 0   & 1 & 1 &1& 1 & 0 &\multirow{5}{*}{--} \\
  2516. %       Check        & 2 & 0 & 2 &0& 0 & 1 & 1   & 0 & 0 &0& 0 & 1& 1    & 0 & 0 &0& 0 & 1 &\\
  2517. %      valve         & 3 & 0 & 0 &0& 0 & 0 & 0   & 0 & 0 &0& 0 & 0 & 1   & 0 & 0 &0& 0 & 0 & \\
  2518. %                    & 4 & 0 & 0 &0& 0 & 0 & 0   & 0 & 0 &0& 0 & 0& 0    & 0 & 0 &0& 0 & 0 &\\
  2519. %                    & 5 & 1 & 0 &1& 1 & 1 & 1   & 1 & 1 &1& 1 & 1 & 1   & 1 & 1 &1& 1 & 1 &\\
  2520. %     \hline
  2521. %                    & 1 & 0 & 0 &1& 1 & 1& 1   & 0 & 3 &0& 0 & 1& 1    & 0 & 1 &0& 1 & 1 &\multirow{5}{*}{--}\\
  2522. %       Control      & 2 & 0 & 0 &1& 0 & 0 & 1   & 0 & 0 &0& 2 & 0& 1    & 0 & 2 &0& 1 & 0 &\\
  2523. %   valve            & 3 & 1 & 1 &1& 2 & 2& 0    & 1 & 1 &1& 1 & 2& 0    & 1 & 0 &1& 0 & 2 &\\
  2524. %                    & 4 & 1 & 0 &0& 0 & 0& 0    & 1 & 0 &0& 0 & 0& 0    & 1 & 0 &0& 1 & 0 &\\
  2525. %                    & 5 & 0 & 0 &0& 0 & 0& 1    & 0 & 0 &1& 0 & 0& 1    & 0 & 0 &1& 0 & 0 &\\
  2526. %                    & 6 & 0 & 0 &0& 0 & 0& 0    & 0 & 0 &0& 0 & 0& 0    & 0 & 0 &0& 0 & 0 &\\
  2527. %\hline
  2528. %                    & 1  & 1  & 0 &0& 0 & 0& 0  & 0 & 0 &0& 0 & 0& 0    & 0 & 0 &0& 0 & 0&\multirow{5}{*}{--}\\
  2529. % Main               & 2  & 0  & 1 &0& 0 & 0& 0  & 1 & 1 &0& 0 & 0& 0    & 1 & 1 &0& 0 & 0&\\
  2530. % air reservoir      & 3  & 0  & 0 &1& 0 & 0& 0  & 0 & 0 &1& 0 & 0& 0    & 0 & 0 &1& 0 & 0&\\
  2531. %                    & 4  & 0  & 0 &0& 1 & 0& 0  & 0 & 0 &0& 1 & 1& 0    & 0 & 0 &0& 1 & 0&\\
  2532. %                    & 5  & 0  & 0 &0& 0 & 1& 1  & 0 & 0 &0& 0 & 0& 2    & 0 & 0 &0& 0 & 1&\\
  2533. % \hline
  2534. %                    & 1  & 0  & 0 &0& 0 & 0& 0     & 0 & 0 &0& 0 & 0& 0    & 0 & 0 &0& 0 & 0&\multirow{5}{*}{--}\\
  2535. % Boosting           & 2  & 1  & 0 &0& 0 & 0& 0    & 1 & 0 &0& 0 & 0& 0    & 1 & 0 &0& 0 & 0&\\
  2536. % cylinder           & 3  & 0  & 1 &1& 0 & 0& 0    & 0 & 1 &1& 0 & 0& 0    & 0 & 1 &1& 0 & 0&\\
  2537. %                    & 4  & 0  & 0 &0& 1 & 1& 0    & 0 & 0 &0& 1 & 1& 0    & 0 & 0 &0& 1 & 1&\\
  2538. %                    & 5  & 0  & 0 &0& 0 & 0& 1    & 0 & 0 &0& 0 & 0& 1    & 0 & 0 &0& 0 & 0&\\
  2539. %\hline
  2540. %\multicolumn{2}{|c||}{Design cost (k\$)} & 64& 72.5 &76&\!\! 83.5 &\!\! 96.5&\!\! 125.5    & 69.5 & 72.5 &84& 91.5 & 99& 169.5    & 69.5& \!\! 80.5\!\! &91\!\!& 98.5\!\! & 103.5\!\!&--\\
  2541. %\hline
  2542. %\end{tabular}
  2543. %\end{center}
  2544. %\end{sidewaystable}
  2545.  
  2546. \section{Computational Study}
  2547. In this section we present numerical experiments of our model, as well as a case study. The computational study consists of six parts: (i) visualizing the result of dimension reduction and clustering; (ii) testing how design changes when parameters vary, including $K$ and other hyperparameters; (iii) choosing best parameter $K$ by cross validation; (iv) experimenting on the value of side information ;  (v)comparing our design with a baseline probabilistic model  and (vi) a real-life case of maintenance in high speed train industry. The distribution used in experiment (i)-(v) is generated by an accountable real data set from (Wang~et~al.~2019).
  2548. All computational result were done on a PC with Intel(R) Core(TM) i7-7500U CPU at 2.7 GHz, coded in Python. The MIP models were solved by a solver library called Gurobi, version 8.1.1.
  2549.  
  2550. \subsection{Visualizing clusters}
  2551. In this section we present a visualization of dimension reduction and clustering. We first apply tSNE algorithm, a state-of-art algorithm for dimension reduction and visualization, \iffalse (refer to Maaten L, Hinton G. Visualizing data using t-SNE[J]. Journal of machine learning research, 2008, 9(Nov): 2579-2605.)\fi to reduce the lifetime data from 28 dimensions to 2 dimensions, and then perform K-means clustering. We present figures of different choice of number of clusters, $K=2$ and $K=5$, respectively. As we can see in the figure, the original multi-dimensional data can be well clustered after dimension reduction. The clustering results, including results from other choices of $K$, are used in following experiments.
  2552.  
  2553. \begin{figure}[H]
  2554. \centering
  2555. %\includegraphics{TSNE_1.png}
  2556. \includegraphics[width=\columnwidth]{2D_tsne.jpg}
  2557. \caption{\footnotesize Visualization of dimension reduction by tSNE algorithm and clustering by K-means algorithms, into 2 clusters on the left and 5 clusters on the right. Note that the two figures are produced by different data.}
  2558. \label{figure_TSNE_1}
  2559. \end{figure}
  2560.  
  2561. \subsection{System design variation with different parameters}
  2562.  
  2563. \begin{figure}[H]
  2564. \centering
  2565. \includegraphics[scale=0.8]{KVARIATION11.pdf}
  2566. \caption{ The multi-type series-parallel system we are experimenting in this section. It consists of 3 subsystems, with 1 type of active parallel component and 2 types of cold-standby component in each subsystem. Each type of component can have up to 3 redundancies.}
  2567. \label{figure1}
  2568. \end{figure}
  2569.  
  2570. In this subsection we experiment on adjusting parameters of the model and observe how system design $x$ changed accordingly.  we consider the following setting: a series-parallel system with  subsystems ($|\mathbf{N}|=3$), with each subsystem containing 3 types of components ($|\mathbf{M}_i|= 3, \forall i \in \mathbf{N}$), among which 1 types are active-parallel and 2 types are cold-standby, and each type is of 3 redundant components ($|\mathbf{T}_{ij}|=3, \forall i \in \mathbf{N}, j\in \mathbf{M}_i$). The lifetime requirement $\mathcal{T}_S = 29$. $\epsilon_{ij} = 0.05,\forall i \in [3], j \in [3]$. The parameters we are adjusting includes $K \in [1,5, 10], R_0 \in \{0.95, 0.97, 0.99\}$, and $\Omega$ to adjust $\underline{\hmu}$ and $\overline{\hmu}$:.
  2571. $$
  2572. \underline{\mu}^{k}_{ij}={\nu}^{k}_{ij}-\mathcal{R},
  2573. \quad \overline{\mu}^{k}_{ij}={\nu}^{k}_{ij}+\mathcal{R} ,
  2574. $$
  2575. where $\mathcal{R}\in \{0.025, 0.05, 0.075, 0.1\}$.  Therefore, by changing the values of $\mathcal{R}$, we can have different sets of parameters $\underline{\hmu}$ and $\overline{\hmu}$. Specifically, large values of $\mathcal{R}$ correspond to the large gaps between $\underline{\hmu}$ and $\overline{\hmu}$.
  2576.  
  2577. \begin{table}[!htbp]\label{d-table}\scriptsize%\footnotesize%\small%\begin{table}[!htbp]\label{d-table}\scriptsize%\footnotesize%\small%
  2578. \caption{ \footnotesize  The design table for different $K$ under $\mathcal{T}_S = 29$}
  2579. \begin{center}
  2580. \begin{adjustbox}{angle=270}
  2581. \scalebox{1}{
  2582. \begin{tabular}{|c|c|c| c|| ccccc|| ccccc|| ccccc||}\hline
  2583. \multirow{3}{*}{$\epsilon$} & \multirow{3}{*}{$\mathcal{R}$} & \multirow{3}{*}{Subsystem} & \multirow{3}{*}{Type} & \multicolumn{5}{c||}{$R_0 = 0.95$} & \multicolumn{5}{c||}{$R_0 = 0.97$} & \multicolumn{5}{c||}{$R_0 = 0.99$}\\
  2584. \cline{5-19}
  2585.  &&  &  & \multicolumn{5}{c||}{$ K$} & \multicolumn{5}{c||}{$K$} & \multicolumn{5}{c||}{$ K$}\\
  2586.      &&&&1 & 3 & 5 &  8 &  10 &1 & 3 & 5 &  8 &  10 & 1 & 3 & 5 &  8 &  10  \\
  2587.     \hline
  2588.                    &&& AP & 1 & 1 & 1 &1 & 1 & 0 &0 & 0 & 0 & 1 & 1 & 1 & 1 &1 & 1\\
  2589.        &&1         & CS-I & 1 & 1 & 1 &1 & 1 & 2 & 2 & 2 & 2& 1 & 2 & 2 & 2 &2 & 2\\
  2590.                    &&& CS-II & 0 & 0 & 0 &0 & 0 & 0 & 0 & 0 &0 & 0& 0 & 0 & 0 &0 & 0\\
  2591.     \cline{3-19}
  2592.                    &&& AP & 1 & 1 & 1 &1 & 1 & 1 & 1 & 1 &1 & 1& 1 & 1 & 1 &1 & 1\\
  2593.        &0.025&2    & CS-I & 0& 0 & 0 & 0 & 1 & 0 & 0 & 0 &0& 0& 0 & 0 & 0 &0 & 0\\
  2594.                    &&& CS-II & 2 & 2 & 2 &2 & 2 & 2 & 2 & 2 &2 & 2& 2 & 2 & 2 &2 & 2\\
  2595.     \cline{3-19}
  2596.                    &&& AP &0 & 0 & 1& 1 & 1 & 0 &1 & 1 &1 & 1 & 1 &0 &0 &0 & 0\\
  2597.        &&3         & CS-I & 1 & 1& 0& 0 & 0 & 1 & 0 & 0& 0 & 1 & 1 & 1 & 1 &1 & 1\\
  2598.                    &&& CS-II & 3 & 3 & 3 &3 & 3&  3  & 3 & 3& 3 & 3& 3 & 3 & 3 &3 & 3\\
  2599. \cline{2-19}
  2600. \multicolumn{1}{|c|}{} & \multicolumn{3}{c||}{Design cost (k\$)} & 39.5& 39.5&\!\! 39    &39& 39 & 44.5&44&44&\!\!44 & 41 &55 & 54.5 &\!\! 54.5 &54.5& 54.5 \\
  2601.     \cline{2-19}
  2602.                    &&& AP& 1 & 1 & 1 &1 & 1 & 0 & 0 & 0 &0 & 0 & 1& 1 & 1 & 1 &1\\
  2603.        &&1         & CS-I& 1 & 1 & 1 &1 & 1 & 2 & 2 & 2 &2 & 2& 2 & 2 & 2 &2 & 2 \\
  2604.                    &&& CS-II & 0 & 0 & 0 &0 & 0 & 0 & 0 & 0 &0 & 0& 0 & 0 & 0 &0& 0\\
  2605.     \cline{3-19}
  2606.                    &&& AP & 1 & 1 & 1 &1 & 1 & 1 & 1 & 1 &1 & 1 & 1 & 1 & 1 &1 & 1 \\
  2607.        &0.05&2     & CS-I &1 & 0 & 0 &0 & 0 & 0 & 0 & 0 &0 & 0& 0 & 0 & 0 &0 & 0\\
  2608.                    &&& CS-II & 2 & 2 & 2 &2 & 2 & 2 & 2 &2 &2 &2& 2 & 2 & 2 &2 &2\\
  2609.     \cline{3-19}
  2610.                    &&& AP & 1 & 1 & 1 &1 & 0 & 1& 1 & 0& 0 & 0 & 1 &1 & 1 &1 & 1\\
  2611.        &&3         & CS-I & 3 & 3 & 3 &3 & 1 & 3 & 3 & 1& 1 & 1 &  3 & 3 & 3 & 3 & 3\\
  2612.                    &&& CS-II & 1 & 1 & 1 &1 & 3 & 1 & 1& 3& 3 & 3 & 1 & 1 & 1 &1 & 1\\
  2613. \cline{2-19}
  2614. \multicolumn{1}{|c|}{0.05} & \multicolumn{3}{c||}{Design cost (k\$)} & 40&40 &\!\! 40  &40& 39.5 & 45&45&\!\! 44.5 & 44.5 & 44.5 & 55&55&\!\! 55  & 55 & 55\\
  2615.     \cline{2-19}
  2616.                     &&& AP& 1& 1 & 1 & 1 &1 & 0 & 0 & 0 &0 & 0 & 1& 1 & 1 & 1 &1\\
  2617.        &&1           &CS-I2 & 1& 1 & 1 & 1 &1 & 2 & 2 & 2 &2 & 2& 2 & 2 & 2 &2 & 2\\
  2618.                    &&& CS-II & 0 & 0 & 0 &0 & 0 & 0 & 0 & 0 &0 & 0& 0 & 0 & 0 &0 & 0 \\
  2619.     \cline{3-19}
  2620.                    &&& AP & 1& 1 & 1 & 1 &1 & 1 &1 & 1 & 1 &1 & 1& 1 & 1 & 1 &1  \\
  2621.        &0.075&2    & CS-I & 0 & 0 & 0 &0 & 0 & 0 & 0 & 0 &0 & 0& 0 & 0 & 0 &0 & 0 \\
  2622.                    &&& CS-II & 2 & 2 & 2 &2 & 2 & 2 & 2 & 2 &2 & 2& 2 & 2 & 2 &2 & 2 \\
  2623.     \cline{3-19}
  2624.                    &&& AP & 1& 1 & 1 & 1 &1  & 1 &1 & 1 & 1 &1 & 1& 1 & 1 & 1 &1\\
  2625.        &&3           & CS-I & 3 & 3 & 3 & 3 & 3 & 3 & 3 & 3 &3 & 3& 3 & 3 & 3 &3 & 3\\
  2626.                    &&& CS-II & 1& 1 & 1 & 1 &1  & 1 &1 & 1 & 1 &1 & 1& 1 & 1 & 1 &1\\
  2627. \cline{2-19}
  2628. \multicolumn{1}{|c|}{} & \multicolumn{3}{c||}{Design cost (k\$)} & 40&40&\!\! 40 & 40 & 40 & 45&45&\!\! 45& 45&45 & 55&55 &\!\! 55    & 55& 55\\
  2629.     \cline{2-19}
  2630.                    &&& AP& 1& 1 & 1 & 1 &1 & 0 & 0 & 0 &0 & 0 & 1& 1 & 1 & 1 &1\\
  2631.        &&1         & CS-I & 1& 1 & 1 & 1 &1 & 2 & 2 & 2 &2 & 2& 2 & 2 & 2 &2 & 2\\
  2632.                    &&&CS-II3 & 0 & 0 & 0 &0 & 0 & 0 & 0 & 0 &0 & 0& 0 & 0 & 0 &0 & 0\\
  2633.     \cline{3-19}
  2634.                    &&& AP & 1& 1 & 1 & 1 &1  &  1 &1 & 1 & 1 &1 & 1& 1 & 1 & 1 &1 \\
  2635.        &0.1&2          & CS-I & 0 & 0 & 0 &0 & 0 & 0 & 0 & 0 &0 & 0& 0 & 0 & 0 &0 & 0 \\
  2636.                    &&& CS-II & 2 & 2 & 2 &2 & 2 & 2 & 2 & 2 &2 & 2& 2 & 2 & 2 &2 & 2\\
  2637.     \cline{3-19}
  2638.                    &&& AP &1 & 1& 1 & 1 & 1 & 1  & 1 & 1 & 1 &1 & 1& 1 & 1 & 1 &1 \\
  2639.        &&3         & CS-I & 3 & 3 & 3 &3 & 3 & 3 & 3 & 3 &3 & 3& 3 & 3 & 3 &3 & 3\\
  2640.                    &&& CS-II& 1& 1 & 1 & 1 &1  & 1 & 1 & 1 & 1 &1 & 1& 1 & 1 & 1 & 1 \\
  2641. \cline{2-19}
  2642. \multicolumn{1}{|c|}{} & \multicolumn{3}{c||}{Design cost (k\$)} & 40&\!\! 40   & 40&40 & 40 & 45&45&\!\! 45 &45& 45 & 55&55&\!\! 55 & 55 & 55 \\
  2643. \hline
  2644. \end{tabular}}
  2645. \end{adjustbox}
  2646. \end{center}
  2647. \end{table}
  2648.  
  2649. %The resulting system design in the case of $\epsilon_{ij} = 0.05$ are shown in the table. We leave the rest of experiment results in the Electronic Companion.
  2650.  
  2651. Note that when $K = 1$, the model reduces to the case of robust model without clustering by Wang, et. al.
  2652.  
  2653. The observations of the experimental results are the following:
  2654. (i) the cost increases when the variation range $[\underline{\hmu}, \overline{\hmu}]$ of expected lifetimes increases, or the dispersion parameter $\epsilon$ of the lifetimes of components enlarges. Such increased cost is due to the enlarged ambiguity set $\mathbb{F}_K$ resulting from the change of above distributional parameters $([\underline{\hmu}, \overline{\hmu}]$. (ii) Intuitively, cost also increases if the required reliability level $R_0$ is increased.
  2655. (iii) The cost decreases as number of clusters $K$ increases. This is due to to the fact that since the mean and dispersion information of each cluster are included in the ambiguity set, more clusters implies more information and therefore smaller ambiguity set.
  2656.  
  2657. We then perform out-of-sample tests on the designs $K = 1, 5, 10$. To test the robustness of the design, we generate testing data with smaller lifetime mean and larger standard deviation. In particular, we let $\hmu_{test} = (1 - \Delta_{M})\hmu_{train}$, and $\hsigma_{test} = \Delta_{S}\hsigma_{train}$. $\Delta_M$ is set to 8\%, and $\Delta_S$ is set to 10\%, 20\% and 30\%, respectively.
  2658.  
  2659. \begin{table}[htp]\footnotesize%\small%
  2660. \caption{\label{tab-compare2} \footnotesize The out of sample reliability of designs generated by $K = 1, 5,10$ models under different $\Delta_{s}$. $R_0$ is set to $0.95$. In the design columns, the 3 columns are corresponding to different subsystems. The three numbers in each column indicates the number of redundancies used that is active parallel, cold-standby type I, or cold-standby type II, respectively. }
  2661. \begin{center}
  2662. \begin{tabular}{|c||c|c|| c|c|c| c |  c |c   |}\hline
  2663. \multirow{2}{*}{$\mathcal{T}_S$}  & \multirow{2}{*}{$(\Delta_{M}, \Delta_{S})$} & \multirow{2}{*}{Model} & \multicolumn{3}{c|}{Design} & \multirow{2}{*}{cost} & \multirow{2}{*}{Mean of out-of-sample reliability level} & \multirow{2}{*}{StD} \\
  2664. \cline{4-6}
  2665. &&& AP & CS-I & CS-II &&&\\
  2666. \hline
  2667. \multirow{9}{*}{29} & \multirow{3}{*}{(8\%, 10\%)} &  K=1 Model & (0,2,0) & (1,3,0) & (0,1,3)& 45.5 &0.976 & 0.152   \\
  2668.     & & K=5 Model & (0,2,0) & (1,2,0) & (1,2,2)& 45.0 &0.959 & 0.198 \\
  2669.     & & K=10 Model & (1,1,0) & (1,3,0) & (1,1,3)& 42.0 &0.607 & 0.489  \\
  2670.    \cline{2-9}
  2671.     &\multirow{3}{*}{(8\%, 20\%)}  &  K=1 Model & (0,2,0) & (1,3,0) & (0,1,3)& 45.5  &  0.972 & 0.167   \\
  2672.     & & K=5 Model &  (0,2,0) & (1,2,0) & (1,2,2)& 45.0&0.953 & 0.211 \\
  2673.     & & K=10 Model &  (1,1,0) & (1,3,0) & (1,1,3)& 42.0  &0.620 & 0.485  \\
  2674.    \cline{2-9}
  2675.     &\multirow{3}{*}{(8\%, 30\%)}  &  K=1 Model & (0,2,0) & (1,3,0) & (0,1,3)& 45.5 &  0.964 & 0.186   \\
  2676.     & & K=5 Model & (0,2,0) & (1,2,0) & (1,2,2)& 45.0 &0.947 & 0.223 \\
  2677.     & & K=10 Model &  (1,1,0) & (1,3,0) & (1,1,3)& 42.0 &0.614 & 0.487  \\
  2678. \hline
  2679.  
  2680. \end{tabular}
  2681. \end{center}
  2682. \end{table}
  2683.  
  2684. \begin{figure}[H]
  2685. \centering
  2686. \includegraphics[width=\columnwidth]{Out_of_sample_K.png}
  2687. \caption{\footnotesize Figure (a)(b)(c) represents the out of sample reliability of designs generated by $K = 1, 5, 10$ models under different $\Delta_{S}$, respectively. The vertical beam represents $\mathcal{T}_S $. The fraction of the lifetime histogram on the right side of beam represents the out-of-sample reliability level.}
  2688. \label{figureK}
  2689. \end{figure}
  2690.  
  2691. From the results, we can observe that with moderate number of clusters ($K = 5$), we can obtain designs with less costs than designs generated by robust model with no clustering ($K = 1$), while retaining robustness even when there is a significant shrink in mean lifetime and much larger standard deviation. This shows that our framework can produce designs that are robust enough and cheaper, by the incorporation of clustering. However, if the number of clusters becomes too large ($K = 10$), the out-of-sample reliability drops significantly. The possible reason is that when $K$ is too high, the model has to split natural clusters into smaller ones, which means that it learns unnecessary information and overfits. Thus, it is crucial to choose the optimal $K$. We present an experiment of choosing $K$ by cross validation in the next subsection.
  2692.  
  2693. \subsection{Choosing $K$ by cross validation}
  2694. In this subsection, we present a cross validation experiment by applying Algorithm 2. In particular, we choose $m = 10$ and do a 10-fold cross validation. The number of instances of constraints violation, as well as the cost of designs, are plotted in the figure below on the left.  \iffalse We also offers a combined metric for cross validation. First, the cost and number of validation are both normalized to the range $[0, 1]$. Then, compute $(1-\lambda)cost(K) + \lambda{{\#}violation(K)}$, where $\lambda \in [0,1]$. By assigning different $\lambda$, we can adapt to scenarios in of different cost-violation tradeoffs. In particular, high $\lambda$ means that robustness of the design is more valued than the cost; low $\lambda$ indicates the contrary. \fi The combined metric under different $\lambda$ are plotted in the figure below on the right. Observe that when $\lambda$ is low, large $K$ such as 7 and 9 are preferred; when $\lambda$ is high, moderate $K$ like 5 are better. Since large $K$ generally correspond to less cost, this result matches the intuition that people are willing to pay more cost if robustness is more valued.
  2695.  %least constraint violation occurs, so $K = 5$ is the ideal parameter to cluster this data set. $K = 5$ will be used in the following subsections.
  2696. \iffalse
  2697. \begin{figure}[H]
  2698. \begin{subfigure}{0.5\textwidth}
  2699. \includegraphics[scale=0.65]{cross_validation.png}
  2700. \caption{\footnotesize}
  2701. \label{figure4-1}
  2702. \end{subfigure}
  2703. \begin{subfigure}{0.5\textwidth}
  2704. \centering
  2705. \includegraphics[scale=0.37]{cross_validation_lambda.png}
  2706. \caption{\footnotesize }
  2707. \label{figure4-2}
  2708. \end{subfigure}
  2709. \caption{\footnotesize (a) The number of violations and costs with different $K$. (b) Costs penalized by $\lambda$ with different $K$, with each line associated with a different $\lambda$.}
  2710. \end{figure}
  2711. \fi
  2712.  
  2713. \begin{figure}[H]
  2714. \centering
  2715. \includegraphics[width=\columnwidth]{cv.jpg}
  2716. \caption{\footnotesize (a) The number of violations and costs with different $K$. (b) Costs penalized by $\lambda$ with different $K$, with each line associated with a different $\lambda$.}
  2717. \label{figure4-2}
  2718. \end{figure}
  2719.  
  2720.  
  2721.  
  2722.  
  2723. \iffalse
  2724. \begin{table}[h*]\footnotesize%\small%
  2725. \caption{\label{tab-compare2} \footnotesize  Out-of-sample reliability level comparison ($R_0=0.85$), where `Design' specifies the number of components allocated in each of 3 subsystems, and `P-Model' and `C-DRO-Model' refer to the probabilistic model and robust model, respectively. }
  2726. \begin{center}
  2727. \begin{tabular}{|c||c|c|| c| c |c |c   |}\hline
  2728.    $L_S$ & Out of sample $\sigma$ &Model & Design &   Designed reliability level &Mean of out-of-sample reliability level & StD \\
  2729.     \hline
  2730.    \multirow{6}{*}{7.625} & \multirow{2}{*}{2}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.85$}& 0.834 & 0.372   \\
  2731.     && C-DRO-Model & (1,1,3) &  &0.997 & 0.053  \\
  2732.        \cline{2-7}
  2733.     &\multirow{2}{*}{6}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.85$}& 0.724 & 0.447   \\
  2734.     && C-DRO-Model & (1,1,3) &  &0.982 & 0.132  \\
  2735.        \cline{2-7}
  2736.     &\multirow{2}{*}{10}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.85$}& 0.673 & 0.469   \\
  2737.     && C-DRO-Model & (1,1,3) &  &0.971 & 0.168  \\
  2738. \hline
  2739.    \multirow{6}{*}{7.75} & \multirow{2}{*}{2}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.85$}& 0.781 & 0.414   \\
  2740.     && C-DRO-Model & (1,1,4) &  &0.999 & 0.028  \\
  2741.                 \cline{2-7}
  2742.     &\multirow{2}{*}{6}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.85$}& 0.690 & 0.462   \\
  2743.     && C-DRO-Model & (1,1,4) &  &0.991 & 0.095  \\
  2744.                  \cline{2-7}
  2745.     &\multirow{2}{*}{10}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.85$}& 0.637 & 0.481   \\
  2746.     && C-DRO-Model & (1,1,4) &  &0.989 & 0.105  \\
  2747. \hline
  2748.    \multirow{6}{*}{7.875} & \multirow{2}{*}{2}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.85$}& 0.716 & 0.451   \\
  2749.     && C-DRO-Model & (1,1,7) &  &1.000 & 0.000  \\
  2750.               \cline{2-7}
  2751.     &\multirow{2}{*}{6}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.85$}& 0.640 & 0.480   \\
  2752.     && C-DRO-Model & (1,1,7) &  &0.998 & 0.040  \\
  2753.               \cline{2-7}
  2754.     &\multirow{2}{*}{10}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.85$}& 0.602 & 0.490   \\
  2755.     && C-DRO-Model & (1,1,7) &  &0.999 & 0.035  \\
  2756. \hline
  2757. \end{tabular}
  2758. \end{center}
  2759. \end{table}
  2760.  
  2761. \begin{table}[h*]\footnotesize%\small%
  2762. \caption{\label{tab-compare2} \footnotesize  Out-of-sample reliability level comparison ($R_0=0.90$), where `Design' specifies the number of components allocated in each of 3 subsystems, and `P-Model' and `C-DRO-Model' refer to the probabilistic model and robust model, respectively. }
  2763. \begin{center}
  2764. \begin{tabular}{|c||c|c|| c| c |c |c   |}\hline
  2765.    $L_S$ & Out of sample $\sigma$ &Model & Design &   Designed reliability level &Mean of out-of-sample reliability level & StD \\
  2766.     \hline
  2767.    \multirow{6}{*}{7.625} & \multirow{2}{*}{2}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.618 & 0.486   \\
  2768.     && C-DRO-Model & (1,1,3) &  &0.998 & 0.047  \\
  2769.        \cline{2-7}
  2770.     &\multirow{2}{*}{6}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.576 & 0.494   \\
  2771.     && C-DRO-Model & (1,1,3) &  &0.989 & 0.103   \\
  2772.        \cline{2-7}
  2773.     &\multirow{2}{*}{10}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.560 & 0.496   \\
  2774.     && C-DRO-Model & (1,1,3) &  &0.988 & 0.111   \\
  2775. \hline
  2776.    \multirow{6}{*}{7.75} & \multirow{2}{*}{2}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.528 & 0.499   \\
  2777.     && C-DRO-Model & (1,1,4) &  &0.999 & 0.037   \\
  2778.                 \cline{2-7}
  2779.     &\multirow{2}{*}{6}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.538 & 0.499   \\
  2780.     && C-DRO-Model & (1,1,4) &  &0.996 & 0.063   \\
  2781.                  \cline{2-7}
  2782.     &\multirow{2}{*}{10}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.530 & 0.499   \\
  2783.     && C-DRO-Model & (1,1,4) &  &0.994 & 0.080   \\
  2784. \hline
  2785.    \multirow{6}{*}{7.875} & \multirow{2}{*}{2}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.458 & 0.498   \\
  2786.     && C-DRO-Model & (1,1,7) &  &0.996 & 0.060  \\
  2787.               \cline{2-7}
  2788.     &\multirow{2}{*}{6}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.481 & 0.500   \\
  2789.     && C-DRO-Model & (1,1,7) &  &0.996 & 0.060   \\
  2790.               \cline{2-7}
  2791.     &\multirow{2}{*}{10}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.496 & 0.500   \\
  2792.     && C-DRO-Model & (1,1,7) &  &0.995 & 0.069  \\
  2793. \hline
  2794. \end{tabular}
  2795. \end{center}
  2796. \end{table}
  2797.  
  2798. \begin{table}[h*]\footnotesize%\small%
  2799. \caption{\label{tab-compare2} \footnotesize  Out-of-sample reliability level comparison ($R_0=0.90$), where `Design' specifies the number of components allocated in each of 3 subsystems, and `P-Model' and `C-DRO-Model' refer to the probabilistic model and robust model, respectively. }
  2800. \begin{center}
  2801. \begin{tabular}{|c||c|c|| c| c |c |c   |}\hline
  2802.    $L_S$ & Out of sample $\sigma$ &Model & Design &   Designed reliability level &Mean of out-of-sample reliability level & StD \\
  2803.     \hline
  2804.    \multirow{6}{*}{7.625} & \multirow{2}{*}{2}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.143 & 0.350   \\
  2805.     && C-DRO-Model & (1,1,3) &  &0.653 & 0.476  \\
  2806.        \cline{2-7}
  2807.     &\multirow{2}{*}{6}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.576 & 0.494   \\
  2808.     && C-DRO-Model & (1,1,3) &  &0.989 & 0.103   \\
  2809.        \cline{2-7}
  2810.     &\multirow{2}{*}{10}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.560 & 0.496   \\
  2811.     && C-DRO-Model & (1,1,3) &  &0.988 & 0.111   \\
  2812. \hline
  2813.    \multirow{6}{*}{7.75} & \multirow{2}{*}{2}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.528 & 0.499   \\
  2814.     && C-DRO-Model & (1,1,4) &  &0.999 & 0.037   \\
  2815.                 \cline{2-7}
  2816.     &\multirow{2}{*}{6}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.538 & 0.499   \\
  2817.     && C-DRO-Model & (1,1,4) &  &0.996 & 0.063   \\
  2818.                  \cline{2-7}
  2819.     &\multirow{2}{*}{10}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.530 & 0.499   \\
  2820.     && C-DRO-Model & (1,1,4) &  &0.994 & 0.080   \\
  2821. \hline
  2822.    \multirow{6}{*}{7.875} & \multirow{2}{*}{2}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.458 & 0.498   \\
  2823.     && C-DRO-Model & (1,1,7) &  &0.996 & 0.060  \\
  2824.               \cline{2-7}
  2825.     &\multirow{2}{*}{6}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.481 & 0.500   \\
  2826.     && C-DRO-Model & (1,1,7) &  &0.996 & 0.060   \\
  2827.               \cline{2-7}
  2828.     &\multirow{2}{*}{10}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.496 & 0.500   \\
  2829.     && C-DRO-Model & (1,1,7) &  &0.995 & 0.069  \\
  2830. \hline
  2831. \end{tabular}
  2832. \end{center}
  2833. \end{table}
  2834.  
  2835. \begin{table}[h*]\footnotesize%\small%
  2836. \caption{\label{tab-compare2} \footnotesize  Out-of-sample reliability level comparison ($R_0=0.90$), where `Design' specifies the number of components allocated in each of 3 subsystems, and `P-Model' and `C-DRO-Model' refer to the probabilistic model and robust model, respectively. }
  2837. \begin{center}
  2838. \begin{tabular}{|c||c|c|| c| c |c |c   |}\hline
  2839.    $L_S$ & Out of sample $\sigma$ &Model & Design &   Designed reliability level &Mean of out-of-sample reliability level & StD \\
  2840.     \hline
  2841.    \multirow{6}{*}{7.625} & \multirow{2}{*}{2}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.348 & 0.476   \\
  2842.     && C-DRO-Model & (1,1,3) &  &0.930 & 0.256  \\
  2843.        \cline{2-7}
  2844.     & \multirow{2}{*}{6}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.427 & 0.495   \\
  2845.     && C-DRO-Model & (1,1,3) &  &0.948 & 0.221  \\
  2846.        \cline{2-7}
  2847.     &\multirow{2}{*}{10}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.445 & 0.497   \\
  2848.     && C-DRO-Model & (1,1,3) &  &0.957 & 0.203   \\
  2849. \hline
  2850.    \multirow{6}{*}{7.75} & \multirow{2}{*}{2}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.278 & 0.448   \\
  2851.     && C-DRO-Model & (1,1,4) &  &0.933 & 0.250   \\
  2852.                 \cline{2-7}
  2853.     &\multirow{2}{*}{6}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.381 & 0.486   \\
  2854.     && C-DRO-Model & (1,1,4) &  &0.974 & 0.159   \\
  2855.                 \cline{2-7}
  2856.     &\multirow{2}{*}{10}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.409 & 0.492   \\
  2857.     && C-DRO-Model & (1,1,4) &  &0.979 & 0.143   \\
  2858. \hline
  2859.    \multirow{6}{*}{7.875} & \multirow{2}{*}{2}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.214 & 0.410   \\
  2860.     && C-DRO-Model & (1,1,7) &  &0.909 & 0.287  \\
  2861.               \cline{2-7}
  2862.     & \multirow{2}{*}{6}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.345 & 0.475   \\
  2863.     && C-DRO-Model & (1,1,7) &  &0.973 & 0.163  \\
  2864.               \cline{2-7}
  2865.     &\multirow{2}{*}{10}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.90$}& 0.383 & 0.486   \\
  2866.     && C-DRO-Model & (1,1,7) &  &0.985 & 0.122   \\
  2867. \hline
  2868. \end{tabular}
  2869. \end{center}
  2870. \end{table}
  2871. \fi
  2872.  
  2873. \iffalse
  2874. From the experiment result, we can observe that despite a smaller ambiguity set, our design ($x^{(2)}$) can achieve robustness level that is comparable to the design without clustering $x^{(1)}$, and are far better than the baseline probabilistic model ($x^{(3)}$).
  2875. \fi
  2876.  
  2877.  
  2878. \subsection{Value of side information}
  2879. In this subsection we experiment on clustering according to side information. The system we are studying is the same as the one in Section 5.2. We choose $K = 5$, corresponding to $\lambda = 0.8$ in the cross validation section. When generating samples from distributions, we also obtain the side information of which of the 5 distributions the sample is drawn from. We then cluster the data set by $K = 5$, based solely on the side information, and compute parameters $(\hnu, [\underline{\hmu}, \overline{\hmu}], \bm{\hsigma}, \p)$ of the model from it. We obtain design ($x^{(2)}$) from the model, and compare it with design ($x^{(1)}$) obtained from the model in which $K$ is also 5, but is directly clustered based on lifetime information instead of the side information.
  2880.  
  2881. \begin{table}[H]\label{d-table}\scriptsize%\footnotesize%\small%
  2882. \caption{ \footnotesize  The design table for $K = 5$ model with and without side information (S.I)}
  2883. \begin{center}
  2884. \begin{tabular}{|c|c| c|| ccccc || ccccc || ccccc |}\hline
  2885. \multirow{3}{*}{S.I} & \multirow{3}{*}{Subsystem} & \multirow{3}{*}{Type} & \multicolumn{5}{c||}{$R_0=0.95$ }& \multicolumn{5}{c||}{$R_0=0.97$}& \multicolumn{5}{c|}{$R_0=0.99$} \\
  2886.  \cline{4-18}
  2887. &&&\multicolumn{5}{c||}{$\mathcal{T}_S $ (yrs) }& \multicolumn{5}{c||}{$\mathcal{T}_S $ (yrs)}& \multicolumn{5}{c||}{$\mathcal{T}_S $ (yrs)}\\
  2888.  
  2889.      &&&28 & 28.5 & 29 &  29.5 & 30 &28 & 28.5 & 29 &  29.5 & 30 &28 & 28.5 & 29 &  29.5 & 30  \\
  2890.     \cline{1-18}
  2891.                    && AP & 0 & 0 &0 & 0 & 0  & 0 & 0 &1 & 1 & 1    & 1 & 1 & 0 & 1 & 1 \\                       &
  2892.        1            & CS-l & 2 & 2 &2 & 2 & 2  & 2 & 2 &2 & 2 & 2    & 3 & 3 & 3 & 3 & 3 \\
  2893.                    &&CS-ll & 3 & 0 &0 & 0 & 0  & 0 & 0 &0 & 0 & 0    & 0 & 0 & 0 & 0 & 0 \\
  2894.     \cline{2-18}
  2895.                    && AP  & 2 & 2 &2 & 1 & 3  & 2& 1 &2 & 2 & 2    & 1 & 1 & 1 & 1 & 1 \\ With  &
  2896.        2            & CS-l & 3 & 3 &0 & 3 & 0  &0 & 2 &3 & 0 & 0   & 3 & 3 & 3& 3& 3 \\ S.I &
  2897.                     & CS-ll & 0 & 0 &2 & 0 & 3  & 3 & 3 &0 & 3 & 3    & 3 & 3& 3 & 3 & 3 \\
  2898.     \cline{2-18}
  2899.                    && AP & 1 & 1 &1 & 1 & 1    & 1 & 0 &1 & 1 & 1    & 1 & 1 & 1 & 1 & 1 \\                       &
  2900.        3            & CS-l & 0 & 0 &1 & 1 & 1   & 2 & 3 &1 & 1 & 1    & 3 & 3 & 3 & 3 & 3 \\
  2901.                    && CS-ll & 3 & 3 &3 & 3 & 3   & 3 & 3 &3 & 3 & 3    & 3 & 3 & 3 & 3 & 3 \\
  2902. \hline
  2903. \multicolumn{3}{|c||}{Design cost (k\$)} & 45.5& 45.5 &47&\!\! 46.5 &\!\! 50    & 51 & 54.5 &57.5& 59 & 59   & 81.5 & \!\! 81.5\!\! &81.5\!\!& 81.5\!\! &81.5 \\
  2904. \hline
  2905. \multicolumn{3}{|c||}{Cost saved (k\$)} & 1 & 1 &1.5&\!\! 4 &\!\! 4.5    & 3.5 & 4.5 &2& 0.5 & 2    & 0 & \!\! 0\!\! &25\!\!& 15\!\! &15 \\
  2906. \cline{1-18}
  2907.     \cline{2-18}
  2908.                    && AP & 0 & 0 &0 & 0 & 0  & 0 & 1&1 & 1 & 1    & 1 & 1 & 1 & 1 & 1 \\                          &
  2909.        1            & CS-l & 2 & 2 &2 & 2 & 2  & 2 & 2 &2 & 2 & 2    & 3 & 3 & 3 & 3 & 3 \\
  2910.                    && CS-ll & 0 & 0 &0 & 0 & 0  & 0 & 0 &0 & 0 & 0    & 0 & 0 & 1 & 1 & 1 \\
  2911.     \cline{2-18}
  2912.                    && AP & 1 & 1 &3 & 3 & 1 & 1 & 2 &2 & 2 & 2    & 1 & 1 & 1 & 1 & 1 \\ Without &
  2913.        2            & CS-l & 3 & 3 &3 & 3 & 3  & 3 & 0 &3 & 3 & 0    & 3 & 3& 3 & 3 & 3 \\ S.I &
  2914.                     & CS-ll & 0 & 0 &0 & 0 & 3  & 3 & 3 &0 & 0 & 3    & 3 & 3 & 3 & 3 & 3 \\
  2915.     \cline{2-18}
  2916.                    && AP  & 1 & 1 &1 & 1 & 1  & 1 & 1 &1  & 1    & 1  & 1& 1 & 1 & 1 & 1 \\                          &
  2917.       3             & CS-l & 1 & 1 &1 & 1 & 2  & 2 & 1 &2  & 2    & 2  & 3& 3 & 2 & 2 & 2 \\
  2918.                    && CS-ll& 3 & 3 &3 & 3   &3  & 3 & 3 &3 & 3    & 3  & 3& 3 & 3 & 3 & 3 \\
  2919. \hline
  2920. \multicolumn{3}{|c||}{Design cost (k\$)} & 46.5& 46.5 &48.5&\!\!50.5 &\!\! 54.5 &54.5     & 59 & 59.5 &59.5& 61 & 81.5    & 81.5\!\! & \!\! 96.5\!\! &96.5\!\!& 96.5\\
  2921. \hline
  2922. \end{tabular}
  2923. \end{center}
  2924. \end{table}
  2925.  
  2926.  
  2927.  
  2928. \begin{table}[htp]\footnotesize%\small%
  2929. \caption{\label{tab-compare3} \footnotesize The out-of-sample result of designs obtained at $\mathcal{T}_S  = 29$. $R_0$ is set to $0.95$. In the design columns, the 3 columns are corresponding to different subsystems. The three numbers in each column indicates the number of redundancies used that is active parallel, cold-standby type I, or cold-standby type II, respectively. }
  2930. \begin{center}
  2931. \begin{tabular}{|c|c|c||c|c|c|| c|c|c|}\hline
  2932. \multirow{2}{*}{($\Delta_{M}, \Delta_{S}$)} &\multirow{2}{*}{Model}  &  Reliability & \multirow{2}{*}{($\Delta_{M}, \Delta_{S}$)} &\multirow{2}{*}{Model}  &  Reliability& \multirow{2}{*}{($\Delta_{M}, \Delta_{S}$)} &\multirow{2}{*}{Model}  &  Reliability \\
  2933. && level &&& level &&& level \\
  2934. \hline
  2935.  
  2936.       \multirow{2}{*}{(5\%, 10\%)}&  With S.I   &0.9999& \multirow{2}{*}{(8\%, 10\%)}&  With S.I& 0.994& \multirow{2}{*}{(10\%, 10\%)} &  With S.I &  0.9626   \\
  2937.      & Without S.I &1.0& & Without S.I &  0.9989&& Without S.I &0.9877  \\
  2938.    \hline
  2939.     \multirow{2}{*}{(5\%, 15\%)}  & With S.I& 0.9999 &\multirow{2}{*}{(8\%, 15\%)}  &  With S.I  & 0.9939&\multirow{2}{*}{(10\%, 15\%)}  &  With S.I & 0.9557  \\
  2940.     & Without S.I & 1.0& & Without S.I & 0.9985  & & Without S.I & 0.9883\\
  2941.    \hline
  2942.      \multirow{2}{*}{(5\%, 20\%)}& With S.I  & 1.0 &\multirow{2}{*}{(8\%, 20\%)}  &  With S.I & 0.994  &\multirow{2}{*}{(10\%, 20\%)}  &  With S.I & 0.9588 \\
  2943.      & Without S.I &1.0 & & Without S.I &  0.9986& & Without S.I & 0.9853\\
  2944.    \hline
  2945.      \multirow{2}{*}{(5\%, 30\%)}  &  With S.I& 0.9996 &\multirow{2}{*}{(8\%, 30\%)}  &  With S.I & 0.9919  &\multirow{2}{*}{(10\%, 30\%)}  &  With S.I &  0.9519 \\
  2946.     & Without S.I  &1.0  & & Without S.I  &0.9972 & & Without S.I & 0.9814 \\
  2947.  
  2948.  
  2949. \hline
  2950. \end{tabular}
  2951.  
  2952. \end{center}
  2953. \end{table}
  2954.  
  2955.  
  2956.  
  2957.  
  2958.  
  2959.  
  2960. \iffalse
  2961. \begin{tabular}{|c|c||c|c|| c|c||c| c||}\hline
  2962. \multirow{2}{*}{$L_S$} & \multirow{2}{*}{($\Delta_{M}, \Delta_{S}$)} &\multirow{2}{*}{Model} & \multicolumn{3}{c|}{Design} & \multirow{2}{*}{cost}  & \multirow{2}{*}{Mean of out-of-sample reliability level} & \multirow{2}{*}{StD} \\
  2963. \cline{4-6}
  2964. &&& AP & CS-I & CS-II &&&\\
  2965. \hline
  2966. \multirow{25}{*}{29} & \multirow{2}{*}{(5\%, 10\%)} &  With S.I & (0,2,0) & (2,3,0) & (0,2,3)& 49 &0.9999 & 0.01   \\
  2967.     & & Without S.I & (0,2,0) & (2,3,0) & (0,3,3)& 51  &1.0 & 0.0 \\
  2968.    \cline{2-9}
  2969.     &\multirow{2}{*}{(5\%, 15\%)}  &  With S.I & (0,2,0) & (2,3,0) & (0,2,3)& 49 & 0.9999 & 0.01   \\
  2970.     & & Without S.I & (0,2,0) & (2,3,0) & (0,3,3)& 51    &1.0 & 0.0 \\
  2971.    \cline{2-9}
  2972.     &\multirow{2}{*}{(5\%, 20\%)}  & With S.I &(0,2,0) & (2,3,0) & (0,2,3)& 49 & 1.0 & 0.0   \\
  2973.     & & Without S.I & (0,2,0) & (2,3,0) & (0,3,3)& 51    &1.0 & 0.0 \\
  2974.    \cline{2-9}
  2975.     &\multirow{2}{*}{(5\%, 30\%)}  &  With S.I & (0,2,0) & (2,3,0) & (0,2,3)& 49 & 0.9996 & 0.02   \\
  2976.     & & Without S.I & (0,2,0) & (2,3,0) & (0,3,3)& 51   &1.0 & 0.0 \\
  2977.    \clineB{2-9}{2}
  2978. & \multirow{2}{*}{(8\%, 10\%)} &  With S.I & (0,2,0) & (2,3,0) & (0,2,3)& 49  &0.994 & 0.0772   \\
  2979.     & & Without S.I & (0,2,0) & (2,3,0) & (0,3,3)& 51    &0.9997 & 0.0173 \\
  2980.    \cline{2-9}
  2981.     &\multirow{2}{*}{(8\%, 15\%)}  &  With S.I & (0,2,0) & (2,3,0) & (0,2,3)& 49  & 0.9939 & 0.0779   \\
  2982.     & & Without S.I & (0,2,0) & (2,3,0) & (0,3,3)& 51  &0.9998 & 0.0141 \\
  2983.    \cline{2-9}
  2984.     &\multirow{2}{*}{(8\%, 20\%)}  &  With S.I &(0,2,0) & (2,3,0) & (0,2,3)& 49  & 0.994 & 0.0772   \\
  2985.     & & Without S.I & (0,2,0) & (2,3,0) & (0,3,3)& 51   &0.9998 & 0.0264 \\
  2986.    \cline{2-9}
  2987.     &\multirow{2}{*}{(8\%, 30\%)}  &  With S.I & (0,2,0) & (2,3,0) & (0,2,3)& 49 & 0.9919 & 0.0869   \\
  2988.     & & Without S.I &(0,2,0) & (2,3,0) & (0,3,3)& 51   &0.9988 & 0.0346 \\
  2989.     \cline{2-9}
  2990.     \clineB{2-9}{2}
  2991. & \multirow{2}{*}{(10\%, 10\%)} &  With S.I & (0,2,0) & (2,3,0) & (0,2,3)& 49  &0.9633 & 0.1880   \\
  2992.     & & Without S.I & (0,2,0) & (2,3,0) & (0,3,3)& 51    &0.9953 & 0.0683 \\
  2993.    \cline{2-9}
  2994.     &\multirow{2}{*}{(10\%, 15\%)}  &  With S.I & (0,2,0) & (2,3,0) & (0,2,3)& 49  & 0.9563 & 0.2044   \\
  2995.     & & Without S.I & (0,2,0) & (2,3,0) & (0,3,3)& 51    &0.9963& 0.0607 \\
  2996.    \cline{2-9}
  2997.     &\multirow{2}{*}{(10\%, 20\%)}  &  With S.I & (0,2,0) & (2,3,0) & (0,2,3)& 49  & 0.9589 & 0.1985   \\
  2998.     & & Without S.I & (0,2,0) & (2,3,0) & (0,3,3)& 51   &0.9942 & 0.0759 \\
  2999.    \cline{2-9}
  3000.     &\multirow{2}{*}{(10\%, 30\%)}  &  With S.I & (0,2,0) & (2,3,0) & (0,2,3)& 49  & 0.9523 & 0.2131   \\
  3001.     & & Without S.I & (0,2,0) & (2,3,0) & (0,3,3)& 51    &0.991 & 0.0944 \\
  3002. \hline
  3003. \end{tabular}
  3004.  
  3005. \end{center}
  3006. \end{table}
  3007. \fi
  3008. The result in table 3 shows that when side information is incorporated, we can achieve a design with much lower cost. We then choose $\mathcal{T}_S  = 29$, and perform out-of-sample tests, in the similar way as previous experiments. We can observe that even with this significant cost saved, the design obtained by clustering by side information still performs well enough under significant $\Delta_M$ and $\Delta_S$
  3009. \begin{figure}[H]
  3010. \centering
  3011. \includegraphics[scale=0.2]{Out_of_sample_side.png}
  3012. \caption{\footnotesize The out-of-sample test result of designs obtained at $\mathcal{T}_S = 29$, with and without S.I.  The vertical beam represents $\mathcal{T}_S $. The fraction of the lifetime histogram on the right side of beam represents the out-of-sample reliability level.}
  3013. \label{figure4}
  3014. \end{figure}
  3015.  
  3016.  
  3017.  
  3018.  
  3019. \iffalse
  3020. \begin{table}[H]\footnotesize%\small%
  3021. \caption{\label{tab-compare2} \footnotesize  Out-of-sample reliability level comparison ($R_0=0.95$), where `Design' specifies the number of components allocated in each of 3 subsystems, and `P-Model' and `C-DRO-Model' refer to the probabilistic model and robust model, respectively. }
  3022. \begin{center}
  3023. \begin{tabular}{|c||c|c|| c| c |c |c   |}\hline
  3024.    $L_S$ & $\Delta_{m}, \Delta_{s}$ &Model & Design &   Designed reliability level &Mean of out-of-sample reliability level & StD \\
  3025.     \hline
  3026.    \multirow{12}{*}{7.5} & \multirow{2}{*}{1, 5}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.95$}& 0.952 & 0.214   \\
  3027.     && C-DRO-Model & (1,2,8) &  &1.000 & 0.000  \\
  3028.        \cline{2-7}
  3029.     &\multirow{2}{*}{1, 10}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.95$}& 0.886 & 0.318   \\
  3030.     && C-DRO-Model & (1,2,8) &  &1.000 & 0.000  \\
  3031.        \cline{2-7}
  3032.     &\multirow{2}{*}{1, 20}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.95$}& 0.800 & 0.400   \\
  3033.     && C-DRO-Model & (1,2,8) &  &1.000 & 0.000  \\
  3034. \hline
  3035. \end{tabular}
  3036. \end{center}
  3037. \end{table}
  3038.  
  3039. \begin{table}[!htbp]\label{d-table}\scriptsize%\footnotesize%\small%
  3040. \caption{ \footnotesize  The design table for $K = 10$ model with side information}
  3041. \begin{center}
  3042. \begin{tabular}{|c| c|| ccccc || ccccc || ccccc |}\hline
  3043. \multirow{2}{*}{Subsystem} & \multirow{2}{*}{Type} & \multicolumn{5}{c||}{$R_0=0.95$,~$L_S$ (yrs) }& \multicolumn{5}{c||}{$R_0=0.97$,~$L_S$ (yrs)}& \multicolumn{5}{c|}{$R_0=0.99$,~$L_S$ (yrs)} \\
  3044. \cline{3-17}
  3045.      &&28 & 28.5 & 29 &  29.5 & 30 &28 & 28.5 & 29 &  29.5 & 30 &28 & 28.5 & 29 &  29.5 & 30  \\
  3046.     \hline
  3047.                    & 1 & 1 & 1 &1 & 0 & 0  & 0 & 0 &0 & 0 & 0    & 1 & 1 & 1 & 1 & 1 \\
  3048.        1           & 2 & 1 & 1 &1 & 2 & 2  & 2 & 2 &2 & 2 & 2    & 2 & 2 & 2 & 2 & 2 \\
  3049.                    & 3 & 0 & 0 &0 & 0 & 0  & 0 & 0 &0 & 0 & 0    & 0 & 0 & 0 & 0 & 0 \\
  3050.     \hline
  3051.                    & 1 & 1 & 1 &1 & 1 & 1  & 1 & 1 &1 & 1 & 1    & 1 & 1 & 1 & 1 & 1 \\
  3052.        2           & 2 & 1 & 0 &0 & 0 & 0  & 1 & 0 &0 & 0 & 3    & 1 & 0 & 0 & 0 & 3 \\
  3053.                    & 3 & 1 & 2 &2 & 2 & 2  & 1 & 2 &2 & 2 & 0    & 1 & 2 & 2 & 2 & 0 \\
  3054.     \hline
  3055.                    & 1 & 1 & 0 &1 & 1 & 0  & 1 & 1 &0 & 1 & 1    & 0 & 0 & 0 & 1 & 1 \\
  3056.       3            & 2 & 0 & 1 &3 & 0 & 1  & 0 & 0 &1 & 3 & 2    & 1 & 1 & 1 & 3 & 2 \\
  3057.                    & 3 & 3 & 3 &1 & 3 & 3  & 3 & 3 &3 & 1 & 2    & 3 & 3 & 3 & 1 & 2 \\
  3058. \hline
  3059. \multicolumn{2}{|c||}{Design cost (k\$)} & 38.5& 39.5 &40&\!\! 44 &\!\! 44.5    & 43.5 & 44 &44.5& 45 & 46    & 54 & \!\! 54.5\!\! &54.5\!\!& 55\!\! &56\\
  3060. \hline
  3061. \end{tabular}
  3062. \end{center}
  3063. \end{table}
  3064.  
  3065.  
  3066. \begin{table}[!htbp]\label{d-table}\scriptsize%\footnotesize%\small%
  3067. \caption{ \footnotesize  The design table for $K = 10$ model without side information}
  3068. \begin{center}
  3069. \begin{tabular}{|c| c|| ccccc || ccccc || ccccc |}\hline
  3070. \multirow{2}{*}{Subsystem} & \multirow{2}{*}{Type} & \multicolumn{5}{c||}{$R_0=0.95$,~$L_S$ (yrs) }& \multicolumn{5}{c||}{$R_0=0.97$,~$L_S$ (yrs)}& \multicolumn{5}{c|}{$R_0=0.99$,~$L_S$ (yrs)} \\
  3071. \cline{4-18}
  3072.      &&28 & 28.5 & 29 &  29.5 & 30 &28 & 28.5 & 29 &  29.5 & 30 &28 & 28.5 & 29 &  29.5 & 30  \\
  3073.     \hline
  3074.                    & 1 & 1 & 0 &0 & 0 & 0  & 0 & 0 &0 & 0 & 0    & 1 & 1 & 1 & 1 & 1 \\
  3075.        1           & 2 & 1 & 2 &2 & 2 & 2  & 2 & 2 &2 & 2 & 2    & 2 & 2 & 2 & 2 & 2 \\
  3076.                    & 3 & 0 & 0 &0 & 0 & 0  & 0 & 0 &0 & 0 & 0    & 0 & 0 & 0 & 0 & 0 \\
  3077.     \hline
  3078.                    & 1 & 1 & 1 &1 & 1 & 1  & 1 & 1 &1 & 1 & 1    & 1 & 1 & 1 & 1 & 1 \\
  3079.        2           & 2 & 1 & 2 &0 & 0 & 3  & 1 & 0 &0 & 2 & 2    & 1 & 0 & 0 & 0 & 3 \\
  3080.                    & 3 & 1 & 0 &2 & 2 & 0  & 1 & 2 &2 & 1 & 1    & 1 & 2 & 2 & 2 & 0 \\
  3081.     \hline
  3082.                    & 1 & 1 & 1 &1 & 1 & 1  & 0 & 1 &1 & 1 & 0    & 0 & 1 & 1 & 1 & 1 \\
  3083.       3            & 2 & 3 & 3 &0 & 0 & 3  & 1 & 3 &3 & 3 & 2    & 1 & 3 & 3 & 3 & 2 \\
  3084.                    & 3 & 1 & 1 &3 & 3 & 1  & 3 & 1 &1 & 1 & 3    & 3 & 1 & 1 & 1 & 2 \\
  3085. \hline
  3086. \multicolumn{2}{|c||}{Design cost (k\$)} & 39.5& 44 &44&\!\! 44 &\!\! 45.5    & 44 & 45 &45& 46 & 47.5    & 54 & \!\! 55\!\! &55\!\!& 55\!\! &56\\
  3087. \hline
  3088. \end{tabular}
  3089. \end{center}
  3090. \end{table}
  3091. \fi
  3092. \iffalse
  3093. Data creating process finised
  3094. Academic license - for non-commercial use only
  3095. cost is 38.5003857478587 , L_S is 26, R_0 is 0.95
  3096. (((1.0, -0.0, -0.0), (1.0, -0.0, -0.0), (7.563683503866358e-06, 7.563683503866358e-06, 7.563683503866358e-06)), ((-0.0, -0.0, 1.0), (-0.0, 1.0, -0.0), (1.0, -0.0, -0.0)), ((-0.0, -0.0, -0.0), (1.0, -0.0, 1.0), (1.0, 1.0, -0.0)))
  3097. cost is 43.0 , L_S is 26, R_0 is 0.97
  3098. (((-0.0, -0.0, -0.0), (-0.0, 1.0, 1.0), (-0.0, -0.0, -0.0)), ((-0.0, -0.0, 1.0), (-0.0, 1.0, 1.0), (-0.0, -0.0, -0.0)), ((-0.0, -0.0, -0.0), (1.0, 1.0, -0.0), (1.0, -0.0, 1.0)))
  3099. cost is 53.0 , L_S is 26, R_0 is 0.99
  3100. (((-0.0, 1.0, -0.0), (-0.0, 1.0, 1.0), (-0.0, -0.0, -0.0)), ((-0.0, 1.0, -0.0), (-0.0, 1.0, 1.0), (-0.0, -0.0, -0.0)), ((-0.0, -0.0, -0.0), (1.0, 1.0, -0.0), (-0.0, 1.0, 1.0)))
  3101. cost is 44.0 , L_S is 27, R_0 is 0.95
  3102. (((-0.0, -0.0, -0.0), (1.0, -0.0, 1.0), (0.0, -0.0, -0.0)), ((-0.0, 1.0, -0.0), (1.0, -0.0, -0.0), (1.0, -0.0, -0.0)), ((-0.0, -0.0, -0.0), (-0.0, -0.0, 1.0), (1.0, 1.0, 1.0)))
  3103. cost is 44.0 , L_S is 27, R_0 is 0.97
  3104. (((-0.0, -0.0, -0.0), (1.0, 1.0, -0.0), (0.0, -0.0, -0.0)), ((-0.0, -0.0, 1.0), (-0.0, 1.0, -0.0), (0.0, 1.0, -0.0)), ((-0.0, -0.0, -0.0), (-0.0, 1.0, -0.0), (1.0, 1.0, 1.0)))
  3105. cost is 54.0 , L_S is 27, R_0 is 0.99
  3106. (((-0.0, 1.0, -0.0), (-0.0, 1.0, 1.0), (0.0, -0.0, -0.0)), ((-0.0, 1.0, -0.0), (1.0, -0.0, -0.0), (-0.0, -0.0, 1.0)), ((-0.0, -0.0, -0.0), (-0.0, 1.0, -0.0), (1.0, 1.0, 1.0)))
  3107. cost is 44.0 , L_S is 28, R_0 is 0.95
  3108. (((-0.0, -0.0, -0.0), (-0.0, 1.0, 1.0), (-0.0, -0.0, -0.0)), ((-0.0, 1.0, -0.0), (-0.0, 1.0, -0.0), (-0.0, 1.0, -0.0)), ((-0.0, -0.0, -0.0), (-0.0, -0.0, 1.0), (1.0, 1.0, 1.0)))
  3109. cost is 45.00017123897305 , L_S is 28, R_0 is 0.97
  3110. (((5.707965768311441e-06, 5.707965768311441e-06, 5.707965768311441e-06), (1.0, 1.0, -0.0), (0.0, 0.0, -0.0)), ((0.0, 1.0, -0.0), (1.0, 1.0, 1.0), (-0.0, -0.0, -0.0)), ((-0.0, -0.0, 0.0), (-0.0, 1.0, -0.0), (1.0, 1.0, 1.0)))
  3111. cost is 54.0 , L_S is 28, R_0 is 0.99
  3112. (((-0.0, -0.0, 1.0), (-0.0, 1.0, 1.0), (-0.0, -0.0, -0.0)), ((-0.0, 1.0, -0.0), (-0.0, 1.0, -0.0), (-0.0, 1.0, -0.0)), ((-0.0, -0.0, -0.0), (-0.0, -0.0, 1.0), (1.0, 1.0, 1.0)))
  3113. cost is 45.0 , L_S is 29, R_0 is 0.95
  3114. (((-0.0, -0.0, -0.0), (1.0, -0.0, 1.0), (-0.0, -0.0, -0.0)), ((-0.0, -0.0, 1.0), (-0.0, -0.0, -0.0), (1.0, -0.0, 1.0)), ((-0.0, 1.0, -0.0), (1.0, 1.0, 1.0), (-0.0, -0.0, 1.0)))
  3115. cost is 54.5 , L_S is 29, R_0 is 0.97
  3116. (((-0.0, -0.0, 1.0), (-0.0, 1.0, 1.0), (-0.0, -0.0, -0.0)), ((1.0, -0.0, -0.0), (-0.0, -0.0, -0.0), (1.0, 1.0, -0.0)), ((-0.0, -0.0, -0.0), (1.0, -0.0, -0.0), (1.0, 1.0, 1.0)))
  3117. cost is 55.0 , L_S is 29, R_0 is 0.99
  3118. (((-0.0, -0.0, 1.0), (1.0, 1.0, -0.0), (-0.0, -0.0, -0.0)), ((-0.0, -0.0, 1.0), (-0.0, -0.0, -0.0), (1.0, -0.0, 1.0)), ((-0.0, 1.0, -0.0), (1.0, 1.0, 1.0), (-0.0, 1.0, -0.0)))
  3119. cost is 55.0 , L_S is 30, R_0 is 0.95
  3120. (((-0.0, -0.0, 1.0), (1.0, 1.0, -0.0), (-0.0, -0.0, -0.0)), ((-0.0, -0.0, 1.0), (1.0, 1.0, 1.0), (-0.0, -0.0, -0.0)), ((-0.0, -0.0, -0.0), (1.0, -0.0, -0.0), (1.0, 1.0, 1.0)))
  3121. cost is 55.5 , L_S is 30, R_0 is 0.97
  3122. (((-0.0, -0.0, 1.0), (-0.0, 1.0, 1.0), (0.0, -0.0, -0.0)), ((-0.0, -0.0, 1.0), (1.0, 1.0, 1.0), (-0.0, -0.0, -0.0)), ((-0.0, -0.0, 1.0), (1.0, 1.0, 1.0), (-0.0, 1.0, -0.0)))
  3123. cost is 56.0 , L_S is 30, R_0 is 0.99
  3124. (((1.0, -0.0, -0.0), (1.0, 1.0, -0.0), (-0.0, -0.0, -0.0)), ((-0.0, -0.0, 1.0), (1.0, 1.0, 1.0), (-0.0, -0.0, -0.0)), ((1.0, 0.0, -0.0), (1.0, 1.0, -0.0), (-0.0, 1.0, 1.0)))
  3125. \fi
  3126. \subsection{Comparison with a baseline probabilistic model}
  3127. \begin{figure}[H]
  3128. \centering
  3129. \includegraphics[scale=0.8]{KVARIATION55.pdf}
  3130. \caption{\footnotesize The series-parallel system we are studying in this section, consists of a single type of component, with active parallel strategy only. This simplicity is due to the limitation of baseline probabilistic model.}
  3131. \label{figure1}
  3132. \end{figure}
  3133. To illustrate the performance of our robust reliability model, we compare the design ($x^{(1)}$) obtained from the proposed robust redundancy optimization model with the design ($x^{(2)}$) obtained from a probabilistic redundancy optimization model. We choose $K = 5$, correspond to $\lambda = 0.8$ in the previous subsection. As mentioned in the Introduction and Literature Review, when the situation involves multiple types ({\it i.e.,} $|\mathbf{M}_i|>1$), or both the cold-standby and active parallel redundant subsystems are considered, the probabilistic model generally becomes intractable. Therefore, for a fair comparison, we consider a series-parallel system with $|\mathbf{N}| = 3$ and $|\mathbf{M}_i|=1, \forall i \in [3]$, which preserves a linear MIP formulation for the probabilistic model. For a coherent exposition of the experimental study, we place the details of the probabilistic redundancy model as well as its MIP transformation in the Electronic Companion.
  3134. \iffalse
  3135. In particular, we first randomly generate lifetime samples (size=2500) and then compute the probability levels $\P[\tilde{z}_{i}\le L_S ], \forall i \in [3]$ and the parameters $(\hnu, \underline{\hmu}, \overline{\hmu}, \bm{\hsigma}, \p)$ from the generated lifetime samples for parameter inputs of probabilistic and robust models, respectively, where $R_0=0.95$ and $L_S=7.5$. We obtain the designs by solving the perspective redundancy models. \fi We perform out-of-sample experiments in the similar way as the out-of-sample-test performed in Section 5.2. We increases the out-of-sample standard deviation ($\Delta_S = 10\%, 15\%, 20\%$, respectively),and shrink the lifetime mean ($\Delta_M = 5\%, 8\%, 10\%$) at the same time. The out-of-sample system lifetimes are compared and plotted in figure, and the comparison of out-of-sample reliability levels is provided in table.
  3136. \begin{table}[H]\footnotesize%\small%
  3137. \caption{\label{tab-compare2} \footnotesize  Out-of-sample reliability level comparison ($R_0=0.95$), where `Design' specifies the number of components allocated in each of 3 subsystems, and `P-Model' and `C-DRO-Model' refer to the probabilistic model and robust model, respectively. }
  3138. \begin{center}
  3139. \begin{tabular}{|c||c|c|| c| c |c |c   |}\hline
  3140.    $\mathcal{T}_S $ & $(\Delta_{M}, \Delta_{S})$ &Model & Design &   Designed reliability level &Mean of out-of-sample reliability level & StD \\
  3141.     \hline
  3142.    \multirow{19}{*}{7} & \multirow{2}{*}{(5\%, 10\%)}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.95$}& 0.914 & 0.280   \\
  3143.     && C-DRO-Model & (1,1,2) &  &0.993 & 0.084  \\
  3144.        \cline{2-7}
  3145.     &\multirow{2}{*}{(5\%, 15\%)}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.95$}& 0.913 & 0.282   \\
  3146.     && C-DRO-Model & (1,1,2) &  &0.990 & 0.097  \\
  3147.        \cline{2-7}
  3148.     &\multirow{2}{*}{(5\%, 20\%)}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.95$}& 0.908 & 0.289   \\
  3149.     && C-DRO-Model & (1,1,2) &  &0.994 & 0.075  \\
  3150.  
  3151.  
  3152.  
  3153.     \clineB{2-7}{2}
  3154.  
  3155.     & \multirow{2}{*}{(8\%, 10\%)}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.95$}& 0.835 & 0.371   \\
  3156.     && C-DRO-Model & (1,1,2) &  &0.971 & 0.167  \\
  3157.        \cline{2-7}
  3158.     &\multirow{2}{*}{(8\%, 15\%)}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.95$}& 0.830 & 0.37   \\
  3159.     && C-DRO-Model & (1,1,2) &  &0.968 & 0.177  \\
  3160.        \cline{2-7}
  3161.     &\multirow{2}{*}{(8\%, 20\%)}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.95$}& 0.825 & 0.380   \\
  3162.     && C-DRO-Model & (1,1,2) &  &0.968 & 0.177  \\
  3163.  
  3164.     \clineB{2-7}{2}
  3165.     & \multirow{2}{*}{(10\%, 10\%)}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.95$}& 0.745 & 0.436   \\
  3166.     && C-DRO-Model & (1,1,2) &  &0.932 & 0.252  \\
  3167.        \cline{2-7}
  3168.     &\multirow{2}{*}{(10\%, 15\%)}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.95$}& 0.754 & 0.431   \\
  3169.     && C-DRO-Model & (1,1,2) &  &0.930 & 0.256  \\
  3170.        \cline{2-7}
  3171.     &\multirow{2}{*}{(10\%, 20\%)}& P-Model & (1,1,1) & \multirow{2}{*}{$R_0=0.95$}& 0.742 & 0.438   \\
  3172.     && C-DRO-Model & (1,1,2) &  &0.946 & 0.227  \\
  3173. \hline
  3174. \end{tabular}
  3175. \end{center}
  3176. \end{table}
  3177.  
  3178. \begin{figure}[H]
  3179. \centering
  3180. \includegraphics[width=\columnwidth]{Out_of_sample_def.png}
  3181. \caption{\footnotesize The out-of-sample system lifetime scenarios of robust model with clustering and probabilistic model under different $\Delta_{M}$ and $\Delta_{S}$. The vertical beam represents $\mathcal{T}_S $. The fraction of the lifetime histogram on the right side of beam represents the out-of-sample reliability level.}
  3182. \label{figure_out_of_sample_test_def}
  3183. \end{figure}
  3184.  
  3185. We can observe that the out-of-sample reliability of the design from baseline model becomes unsatisfiable, while the design from robust model can still keep the reliability level intact. Robust model outperforms baseline model more as $\Delta_M$ and $\Delta_S$ increases. This illustrates that our model outperforms the baseline model in that it is significantly more robust, especially under extreme circumstances. In addition, as mentioned before, our model is tractable for multi-type mixed strategy systems, while the probabilistic model will become intractable. Thus our model is superior in both robustness and computability.
  3186.  
  3187. \subsection{Case study}
  3188.  
  3189. In this section we present a case study of the braking system on high speed train, to provide an example of applying our model in practice. The braking system is a critical subsystem on the train, as the failure of the system can cost hundreds of life. Therefore, it is a common practice to install multiple redundant braking mechanisms on the train, either cold-standby or active-parallel. In this study we model the problem as a system consists of a single subsystem, with 3 types of braking components:  MTB, pneumatically brake stop cock, and emergency brake valve. MTB components works in active-parallel, while the other twos are cold-standby redundancies. Up to 2 MTBs, 8 pneumatically brake stop cocks, and 2 emergency brake valves may be installed.
  3190.  
  3191. \begin{figure}[H]
  3192. \centering
  3193. \includegraphics[scale=0.6]{case-description-edited.pdf}
  3194. \caption{ The braking system in this case. It consists of 3 types of component: pneumatically brake stop cock, MTB, and emergency brake valve. MTB components works in active-parallel, while the other twos are cold-standby redundancies. Only two redundant components are shown in figure for each type for simplicity.}
  3195. \label{figure1}
  3196. \end{figure}
  3197.  
  3198. A subset of the dataset is presented below.
  3199.  
  3200. \begin{table}[H]\label{without side information-table}%\scriptsize%\footnotesize%\small%
  3201. \caption{ \footnotesize  A sample of dataset.}
  3202. \begin{center}
  3203. \begin{tabular}{|c|c| c|c|}\hline
  3204. Type & {Speed (km/s)} & {Emergency brake valve status} & {Mileage at failure(kms)} \\
  3205. \hline
  3206. \multirow{5}{*} {Pneumatically brake stop cock} & 0 & 1 & 1205489.233 \\
  3207. &3.191489362 &0&1210654.084\\
  3208. &68.08510638&0&1212826.491\\
  3209. &100&1&1214834.413\\
  3210. &\vdots&\vdots&\vdots\\
  3211. \hline
  3212. \multirow{5}{*} {MTB} & 0 & 1 & 1209897.965\\
  3213. & 25.53191489 & 0 & 1213133.429\\
  3214. & 3.191489362   &1 & 1213997.540\\
  3215. & 58.5106383 & 0 & 1215082.013\\
  3216. &\vdots&\vdots&\vdots\\
  3217.  
  3218. \hline
  3219. \multirow{5}{*} {Emergency brake valve} & 3.191489362 & 0 & 1210654.084\\
  3220. & 4.255319149 & 1 & 1212022.818\\
  3221. & 68.08510638   &0 & 1212826.491\\
  3222. & 35.10638298 & 0 & 1217498.174\\
  3223. &\vdots&\vdots&\vdots\\
  3224. \hline
  3225.  
  3226. \end{tabular}
  3227. \end{center}
  3228. \end{table}
  3229.  
  3230. In this case, lifetime is measured by the distance the train has run when the component is installed, instead of time, since failure rate is more closely related to the former. We can compute the lifetime by subtracting the mileage when the component is installed by the mileage when the component fails.
  3231.  
  3232. The speed and the binary internal status of emergency brake valve at the point of failure are suspect to be correlated to failure rate. High speed may overload the braking system that already accumulates much attrition. The internal status, although does not directly tells us whether the component fails, might contain useful information of the braking system as a whole at the point of failure, like the flight recorder ("black box") on aeroplanes. Therefore, we incorporated them as side information in the clustering process.
  3233.  
  3234. First, we applies the cross validation algorithm to search for ideal $K$. We choose $\lambda = 0.8$, since robustness is critical in high speed train. The result is shown in the figure below.
  3235.  
  3236. \begin{figure}[H]
  3237. \centering
  3238. \includegraphics[width=\columnwidth]{real_case_cv.jpg}
  3239. \caption{\footnotesize (a) The number of violations and costs with different $K$. (b) Costs penalized by $\lambda$ with different $K$, with each line associated with a different $\lambda$.}
  3240. \label{figure4-2}
  3241. \end{figure}
  3242.  
  3243.  
  3244. The results shows that $K = 3, 4, 5$ is ideal. We subsequently obtain the design in the table below using $K=5$, using side information in the clustering process. In particular, we divide the speed into 3 cases: stationary (under 20 km/s), low speed (20-50 km/s) and high speed (over 50 km/s). The valve openness is a boolean value and is naturally divided into 2 cases. Therefore, we can divide the dataset to 6 clusters based on speed and valve status. However, in our dataset there is no instance in the class of low speed and valve status 1. This leaves us with 5 clusters ($K = 5$). We then collect distributional information from each of the clusters, and obtain the design.
  3245.  
  3246. \iffalse
  3247. \begin{table}[H]\label{without side information-table}\scriptsize%\footnotesize%\small%
  3248. \caption{ \footnotesize  The design table for optimal clusters $K = 5$ of real case model. Cock, MTB and Emergency represents the pneumatically brake stop cocks, MTB and emergency brake valves, respectively.}
  3249. \begin{center}
  3250. \begin{tabular}{|c|c| c|| ccc || ccc || ccc || ccc |}\hline
  3251. \multirow{2}{*}{$\epsilon$} & \multirow{2}{*}{Subsystem} & \multirow{2}{*}{Type} & \multicolumn{3}{c||}{$R_0=0.95$,~$L_S$ (kms) }& \multicolumn{3}{c||}{$R_0=0.97$,~$L_S$ (kms)}& \multicolumn{3}{c||}{$R_0=0.98$,~$L_S$ (kms)}& \multicolumn{3}{c|}{$R_0=0.99$,~$L_S$ (kms)} \\
  3252. \cline{4-15}
  3253.      &&&2000 & 2250 & 2500 &  2000 & 2250 &2500 & 2000 & 2250 &  2500 & 2000 & 2250 &  2500  \\
  3254.     \cline{1-15}
  3255.                  && Cock & 5 & 6 &7 & 6 & 6  & 7 & 6 &7& 7 & 6 &7& 8 \\
  3256. $\epsilon=0.5$      & 1 & MTB & 1 & 1 &1 & 0 & 1  & 1 & 0 &0 & 1& 1 &1 & 1 \\
  3257.                 &&Emergency  & 2 & 2 &2 & 2 & 2 &2  & 2 & 2 &2  & 1 & 1 &1  \\
  3258.  
  3259. \hline
  3260. \multicolumn{3}{|c||}{Design cost (k\$)} & 27& 31 &35&\!\! 28 &\!\! 31    &35 & 28 &32& 35 & 29    & 33 & \!\! 37 \\
  3261. \hline
  3262. \end{tabular}
  3263. \end{center}
  3264. \end{table}
  3265. \fi
  3266.  
  3267.  
  3268. \begin{table}[H]\label{new-table}\scriptsize%\footnotesize%\small%
  3269. \caption{ \footnotesize  The design table for $K = 5$ real case model with side information (S.I)}
  3270. \begin{center}
  3271. \begin{tabular}{|c|c| c|| ccc || ccc || ccc || ccc |}\hline
  3272. \multirow{3}{*}{With S.I,$\epsilon$} & \multirow{3}{*}{Subsystem} & \multirow{3}{*}{Type} &\multicolumn{3}{c||}{$R_0=0.95$ }& \multicolumn{3}{c||}{$R_0=0.97$}& \multicolumn{3}{c||}{$R_0=0.98$}& \multicolumn{3}{c|}{$R_0=0.99$} \\
  3273. \cline{4-15}
  3274. &&&\multicolumn{3}{c||}{~$\mathcal{T}_S $ (kms) }& \multicolumn{3}{c||}{~$\mathcal{T}_S $ (kms)}& \multicolumn{3}{c||}{~$\mathcal{T}_S $ (kms)}& \multicolumn{3}{c|}{~$\mathcal{T}_S $ (kms)}\\
  3275.  
  3276.      &&&2000 & 2250 & 2500 &  2000 & 2250 &2500 & 2000 & 2250 &  2500 & 2000 & 2250 &  2500  \\
  3277.     \cline{1-15}
  3278.                  && Cock & 5 & 6 &7 & 5 & 6  & 7 & 6 &6& 7 & 6 &6& 7 \\
  3279.        $\epsilon=0.025$& 1           & MTB& 1 & 1 &1 & 1 & 1  & 1 & 0 &1 & 1& 1 &1 & 1  \\
  3280.                  &&Emergency & 2 & 2 &2 & 2 & 2 &2  & 2 & 2 &2  & 1 & 2 &2  \\
  3281.  
  3282. \hline
  3283. \multicolumn{3}{|c||}{Design cost (k\$)} & 27& 31 &35&\!\! 27 &\!\! 31    &35 & 28 &31& 35 & 29    & 31 & \!\! 35 \\
  3284.     \cline{1-15}
  3285.                 && Cock& 5 & 6 &7 & 6 & 6  & 7 & 6 &7& 7 & 6 &7& 8  \\
  3286.         $\epsilon=0.05$& 1           & MTB & 1 & 1 &1 & 0 & 1  & 1 & 0 &1 & 1& 1 &1 & 0 \\
  3287.                 &&Emergency & 2 & 2 &2 & 2 & 2 &2  & 2 & 2 &2  & 1 & 2 &2  \\
  3288. \hline
  3289. \multicolumn{3}{|c||}{Design cost (k\$)} & 27& 31 &35&\!\! 28 &\!\! 31    &35 & 28 &32& 35 & 29    & 33 & \!\! 36 \\
  3290.  
  3291.     \cline{1-15}
  3292.                  && Cock& 6 & 6 &7 & 6 & 7 & 7 & 6 &7& 8 & 7 &7& 8 \\
  3293.        $\epsilon=0.075$& 1           & MTB  & 0 & 1 &1 & 1 & 0  & 1 & 1 &1 & 0& 0 &1 & 1 \\
  3294.                 &&Emergency & 2 & 2 &2 & 1 & 2 &2  & 1 & 1 &2  & 1 & 1 &1  \\
  3295.  
  3296. \hline
  3297. \multicolumn{3}{|c||}{Design cost (k\$)} & 28& 31 &35&\!\! 29 &\!\! 32    &35 & 29 &33& 36 & 30    & 33 & \!\! 37 \\
  3298. \hline
  3299. \end{tabular}
  3300. \end{center}
  3301. \end{table}
  3302. We can observe that as lifetime requirement $\mathcal{T}_S $, level of robustness $R_0$, or dispersion parameter $\epsilon$ increases, the cost of design generally increases. This is due to either extra components installed (for example, when $\mathcal{T}_S $ increases from 2000 to 2250 under $R_0 = 0.95$, $\epsilon = 0.05$, an extra cock is installed), or a cheaper but less robust component is replaced by a more expensive but more robust one (for example, when $R_0$ increases from 0.95 to 0.97 under $\mathcal{T}_S = 2000$,$\epsilon = 0.05$, the single MTB is replaced by a cock).
  3303.  
  3304. \section{Conclusions}{\color{red}
  3305. Redundancy allocation problem (RAP) is a critical problem in the field of reliability engineering. Distributionally robust optimization (DRO) models has been employed to high robustness in RAP. We present a framework of combining clustering and dimension reduction into DRO models, to achieve a smaller ambiguity set, which helps alleviate the problem that DRO models can sometimes be over conservative. This framework can incorporate side information to improve the ambiguity set, or, in absence of side information, use clustering to discover and utilize underlying structures of the data. We develop a mixed integer linear program (MILP)reformulation of the model without adding additional integer variables, thus keeping the model tractable. We also devise the supergradient based algorithm to decompose the MILP formulation, so that the model can be solved by multiple computers in parallel, which makes the model practically viable even with large number of clusters.
  3306. }
  3307.  
  3308. \begin{thebibliography}{}
  3309.  
  3310. \bibitem{Ardakan2014}Ardakan M. A., A. Z. Hamadani, M. Alinaghian. Optimizing bi-objective redundancy allocation problem with a mixed redundancy strategy. {\em ISA transactions}~{\bf 55} :116--128.
  3311.  
  3312. \bibitem{Bertsimas2011}Bertsimas D., M. Sim. 2004. The price of robustness. {\em Operations Research}~{\bf 52}~(1):35--53.
  3313.  
  3314. \bibitem{Bhunia2010} Bhunia, A. K., L. Sahoo,  D. Roy. 2010. Reliability stochastic optimization
  3315. for a series system with interval component reliability via genetic
  3316. algorithm.  {\em Appl.Math. Computat.}~{\bf 216}~(3): 929--939, 2010.
  3317.  
  3318.  
  3319. \bibitem{Chern1992}Chern, M.S. 1992. On the computational complexity of reliability redundancy allocation in a series system. {\em Operations research letters}~{\bf 11}~(5):309--315.
  3320.  
  3321. \bibitem{Cheng2009}Cheng, Z., X. Wang, C. Tian, F. Wang. 2009. Mission reliability simulation of High-speed EMU service braking system. {\em Proceedings of the 8th International Conference on Reliability}, Maintainability and Safety (ICRMS 2009), 253--256.
  3322.  
  3323. \bibitem{Coit1998}Coit, D. W., A.E. Smith. 1998. Redundancy allocation to maximize a lower percentile of the system time-to-failure distribution. {\em IEEE Trans. Rel.}~ {\bf 47}~(1):79--87.
  3324.  
  3325. \bibitem{Coit2003}Coit, D. W. 2003. Maximization of system reliability with a choice of redundancy strategies. {\em IIE Transactions}~{\bf 35}~(6):535-543.
  3326.  
  3327. \bibitem{Coit2004} Coit, D.W.,  T. Jin,  N. Wattanapongsakorn. 2004. System optimization
  3328. with component reliability estimation uncertainty: A multi-criteria approach. {\em IEEE Trans. Rel.}~ {\bf 53}~(3) :  369--380, 2004.
  3329.  
  3330.  
  3331.  
  3332.  
  3333. \bibitem{Elsayed2012}Elsayed, E A. 2012. {\em Reliability Engineering}. 2nd Edition. Wiley.
  3334.  
  3335.  
  3336. \bibitem{Govindan2017}Govindan, K., A. Jafarian, M.E. Azbari, T.M. Choi. 2017. Optimal bi-objective redundancy allocation for systems reliability and risk management. {\em IEEE Transactions on Cybernetics}~{\bf 46}~(8):1735--1748.
  3337.  
  3338.  
  3339.  
  3340. %\bibitem{Lam2012}Lam SW, T.S. Ng, and M. Sim. (2012). Multiple objectives satisficing under uncertainty. To appear in Operations Research, 2012.
  3341.  
  3342. %\bibitem{Lin2011}Lin J, Muthuraman K, Lawley M (2011) Optimal and approximate algorithms for sequential clinical scheduling with no-shows. {\it IIE Transactions on Healthcare Systems Engineering} 1:20--36.
  3343.  
  3344.  %\bibitem{McCarthy2000} McCarthy K, McGee HM, O'Boyle CA. 2000. Outpatient clinic waiting times and non-attendance as indicators of quality. {\it Psychology, Health and Medicine} 5: 287--293.
  3345.  
  3346. \bibitem{Grani2017} Hanasusanto, G. A., V. Roitch, D. Kuhn, W. Wiesemann. 2017. Ambiguous joint chance constraints under mean and dispersion information. {\it Operations Research}~{\bf 65}~(3):715--767.
  3347.  
  3348.  
  3349.  
  3350.  
  3351.  
  3352. \bibitem{Elegbede2003}Elegbede, A.C., C. Chu, K.H. Adjallah, F. Yalaoui. 2003. Reliability allocation through cost minimization. {\em IEEE Transactions on reliability}~{\bf 52}~(1):106--111.
  3353.  
  3354. \bibitem{Feizollahi2012} Feizollahi, M.J., M. Modarres. 2012. The robust deviation redundancy allocation problem with interval component reliabilities. {\em IEEE Transactions on reliability}~{\bf 61}~(4):957--965.
  3355.  
  3356.  
  3357.  
  3358.  
  3359.  
  3360. \bibitem{Feizollahi2014}Feizollahi, M.J., S. Ahmed, M. Modarres. 2014. The robust redundancy allocation problem in series-parallel systems with budgeted uncertainty.  {\em IEEE Transactions on reliability}~{\bf 63}~(1):239--250.
  3361.  
  3362. \bibitem{Feizollahi2015} Feizollahi, M.J., R. Soltan, H. Feyzollahi. 2015. The robust cold standby redundancy allocation in series-parallel systems with budgeted uncertainty. {\em IEEE Transactions on reliability}~{\bf 64}~(2):799--806.
  3363.  
  3364. \bibitem{Friedman2001} Friedman, J., T. Hastie, R. Tibshirani. 2001. {\em The elements of statistical learning.} Springer series in statistics. New York.
  3365.  
  3366. \bibitem{Hasegawa1999}Hasegawa, I., Uchida, S. 1999. Braking systems. {\em Japan Railway and Transport Review}~{\bf 20}:52-59.
  3367.  
  3368. \bibitem{James2013} James, G., D. Witten, T. Hastie, R. Tibshirani. 2013. {\em An introduction to statistical learning.} Springer series in statistics. New York.
  3369.  
  3370. \bibitem{Ketchen1996}Ketchen, Jr. D.J., C.L. Shook, 1996. The application of cluster analysis in Strategic Management Research: An analysis and critique. {\em Strategic Management Journal}~{\bf 17}~(6):441-458.
  3371.  
  3372. \bibitem{Kuo2001}Kuo, W., V.R. Prasad, F.A. Tillman, C.L. Hwang. 2001. {\em Optimal Reliability Design: Fundamentals and Applications.} Cambridge university press. Cambridge.
  3373.  
  3374. \bibitem{Kuo2007}Kuo, W., R. Wan. 2007. Recent advances in optimal reliability allocation. {\em IEEE Transactions on Systems, Man, and Cybernetics-Part A: Systems and Humans}~{\bf 37}~(2):143-156.
  3375.  
  3376. \bibitem{Li2014}Li, Y.F., Y. Ding, E. Zio. 2014. Random fuzzy extension of the universal generating function approach for the
  3377. reliability assessment of multi-state systems under aleatory and epistemic uncertainties. {\em IEEE Transactions on Reliability}~{\bf 63}~(1):13--25.
  3378.  
  3379. \bibitem{Li2011}Li, C.Y., X. Chen, X.S. Yi, J.Y. Tao. 2011. Interval-valued reliability analysis of multi-state systems. {\em IEEE Transactions on Reliability}~{\bf 60}~(1):323--330.
  3380.  
  3381. \bibitem{Li2008} Li, X.,  X. Hu. 2008. Some new stochastic comparisons for redundancy
  3382. allocations in series and parallel systems.~{\em  Statist. Probabil. Lett.}~{\bf 78}~(18): 3388--3394.
  3383.  
  3384.  
  3385.  
  3386. \bibitem{Liao2014}Liao, L., F. K\"{o}ttig. 2014. Review of hybrid prognostics approaches for remaining useful life prediction of engineered systems, and an application to battery life prediction. {\em IEEE Transactions on Reliability}. {\bf 63}~(1):191--207.
  3387.  
  3388.  
  3389.  
  3390.  
  3391. \bibitem{MacQueen1967} MacQueen, J. 1967. Some methods for classification and analysis of multivariate observations. {\em Proceedings of the Fifth Berkeley Symposium on Mathematical Statistics and Probability} {\bf 1}~(Statistics):281--297.
  3392.  
  3393.  
  3394.  
  3395. \bibitem{ Marseguerra2005} Marseguerra, M., E. Zio, L. Podofillini, D. W. Coit. 2005. Optimal design of reliable network systems in presence of uncertainty. {\em IEEE Trans Rel.}~{\bf 54}~(2):243--253.
  3396.  
  3397. \bibitem{Military1992}Military, U.S. 1992. Reliability prediction of electronic equipment. MIL-HDBK-217F Notice 1.
  3398.  
  3399. \bibitem{Ng2002} Ng, A. Y, M. I. Jordan, Y. Weiss. 2002. On spectral clustering: Analysis and an algorithm. {\em Advances in neural information processing systems}:849-856.
  3400.  
  3401. \bibitem{Ng2014} Ng, S. Y., Y. Xing, K. L. Tsui. 2014. A naive Bayes model for robust remaining useful life prediction of lithium-ion battery. {\em Applied Energy}~{\bf 118}: 114-123.
  3402.  
  3403.  
  3404.  
  3405.  
  3406. \bibitem{Prasad2001}Prasad, V. R., W. Kuo, K. O. Kim. 2001. Maximization of a percentile life of a series system through component redundancy allocation. {\em IIE Transactions}~{\bf 33}~(12):1071--1079.
  3407.  
  3408. \bibitem{Pecht2008} Pecht, M. 2008. {\em Prognostics and Health Management of Electronics.} John Wiley \& Sons, Ltd.
  3409.  
  3410. \bibitem{Quinlan1986}Quinlan, J. R. 1986. Induction of decision trees. {\em Machine Learning}~{\bf 1}: 81--106.
  3411.  
  3412.  
  3413. \bibitem{Shapiro2001}Shapiro~A.~2001.~On duality theory of conic linear problems. In {\em Semi-Infinite Programming}, chapter 7, 135--165, Kluwer Academic Publishers, 2001.
  3414.  
  3415. \bibitem{Sibson1973}Sibson R. 1973. SLINK: an optimally efficient algorithm for the single-link cluster method. {\em The Computer Journal. British Computer Society}~{\bf 16}~(1):30-34.
  3416.  
  3417. \bibitem{Soltani2015}Soltani R., J. Safari, S.J. Sadjadi. 2015. Robust counterpart optimization for the redundancy allocation problem in series-parallel systems with component mixing under uncertainty. {\em Applied Mathematics \& Computation}~{\bf 271}~(C): 80--88.
  3418.  
  3419.  
  3420.  
  3421.  
  3422. \bibitem{Sun2017} Sun, M. X., Y. F. Li, E. Zio. 2017. On the optimal redundancy allocation for multi-state series-parallel systems under epistemic uncertainty. {\em Reliability Engineering \& System Safety}. Accepted.
  3423.  
  3424. \bibitem{Tang2014}Tang, S., C. Yu, X. Wang, X. Guo, X. Si. 2014. Remaining useful life prediction of lithium-ion batteries based on the wiener process with measurement error. {\em Energies}~{\bf 7}~(2):520--547.
  3425.  
  3426. \bibitem{Tekiner-Mogulkoc2011}Tekiner-Mogulkoc, H., D. W. Coit. 2011. System reliability optimization
  3427. considering uncertainty: Minimization of the coefficient of variation
  3428. for series-parallel systems.~{\em  IEEE Trans. Rel.}~{\bf 60}~(30): 667--674, 2011.
  3429.  
  3430. \bibitem{Thorndike1953}Thorndike R. L. 1953. Who Belongs in the Family?. {\em Psychometrika}~{\bf 18}~(4):267-276.
  3431.  
  3432. \bibitem{Wang2012} Wang, Y., L. Li, S. Huang, Q. Chang. 2012. Reliability and covariance estimation of weighted k-out-of-n multi-state Systems. {\em European Journal of Operational Research}~{\bf 221}:~138--147.
  3433.    
  3434. {\color{red} \bibitem{Wang2019} Wang, S.,...}
  3435.  
  3436. \bibitem{Wisemann2014} Wiesemann, W.,  D. Kuhn, M. Sim. 2014. Distributionally robust convex optimization. {\it Operations Research}~{\bf 62} ~(6)~ 1358--1376.
  3437.  
  3438. \bibitem{xie2017} Xie, W., Ahmed, S. 2017. Distributionally robust chance constrained optimal power flow with renewables: A conic reformulation. {\em IEEE Transactions on Power Systems.} Accepted.
  3439.  
  3440.  
  3441. \bibitem{Yalaoui2005}Yalaoui, A., E.  Chatelet, C. Chu. 2005. A new dynamic programming method for reliability redundancy allocation in a parallel-series system. {\em IEEE transactions on reliability}.~{\bf 54}~(2):254--261.
  3442.  
  3443.  
  3444. \bibitem{Zaretalab2015}Zaretalab, A., V. Hajipour, M. Sharifi, M. R. Shahriari. 2015. A knowledge-based archive multi-objective simulated annealing algorithm to optimize series-parallel system with choice of redundancy strategies. {\em Computers \& Industrial Engineering}~{\bf 80}:33-44.
  3445.  
  3446. \bibitem{Zhao2003} Zhao, R., B. Liu. 2003. Stochastic programming models for general redundancy-optimization problems.~{\em IEEE Trans. Rel.}~{\bf 52}~(2): 181--191, 2003.
  3447.  
  3448. \bibitem{Zhao2011} Zhao, P., P.S. Chan, H.K.T. Ng. 2011. Optimal allocation of redundancies in series systems. {\em European Journal of Operational Research}~{\bf 220}~(3):673--683.
  3449.  
  3450. \bibitem{Kuhn2013}Zymler, S., D. Kuhn, B. Rustem. 2013. Distributionally robust joint chance constraints with second-order moment information, {\em Mathematical Programming}~{\bf 137}~(1-2):167--198.
  3451. \end{thebibliography}
  3452. \newpage
  3453. \section*{Appendix: A benchmark probabilistic reliability model}
  3454. As a benchmark for fair comparison, we consider the following probabilistic reliability model
  3455. \begin{eqnarray*}
  3456. \begin{array}{rcll}
  3457. & \min\limits_{\y} &  \sum\limits_{i\in \mathbf{N}}y_{i}c_{i} \\[0.3 cm]
  3458. & {\rm s.t.} &  \displaystyle \prod_{i \in \mathbf{N}}\left(1-r_{i}^{y_{i}}\right)\ge R_0  & \\[0.3 cm]
  3459. && L_{i}\le y_{i}\le U_{i}, & \forall i \in \mathbf{N}\\
  3460. && y_{i} \in \mathbb{Z}_+,  & \forall i \in \mathbf{N},
  3461. \end{array}
  3462. \end{eqnarray*}
  3463. where each subsystem is equipped with one type of components (i.e., $|\mathbf{M}_i|\equiv 1, \forall i \in \mathbf{N}$ ), and
  3464. $$
  3465. r_{i}=\P\Big[\tilde{z}_{i}\le \mathcal{T}_R \Big],
  3466. $$
  3467. which can be estimated from the data. By transforming the integer variable $y_{i}$ with binaries $x_{ik}$:
  3468. $$
  3469. y_{i}=L_{i}+\sum_{k=0}^{U_{i}-L_{i}}kx_{ik},~\mbox{with}~\sum_{k=0}^{U_{i}-L_{i}}x_{ik}=1,
  3470. $$
  3471. the above model can be linearized as the following MIP with binaries (Feizollahi and Modarres~2012):
  3472. \begin{eqnarray}
  3473. \begin{array}{rcll}
  3474. & \min\limits_{\x} & \displaystyle \sum\limits_{i\in \mathbf{N}} \left[L_{i}+\sum_{k=0}^{U_{i}-L_{i}}kx_{ik}\right]c_{i} \\[0.3 cm]
  3475. & {\rm s.t.} &  \displaystyle \sum_{i \in \mathbf{N}} \sum_{k=0}^{U_{i}-L_{i}}x_{ik}\ln\left[1-r_{i}^{L_{i}+k} \right]\ge \ln R_0  & \\[0.3 cm]
  3476. && \displaystyle \sum_{k=0}^{U_{i}-L_{i}}x_{ik}=1, & \forall i \in \mathbf{N}\\
  3477. && x_{ik} \in \{0,1\},  & \forall i \in \mathbf{N}, k \in [0; U_{i}-L_{i}],
  3478. \end{array}
  3479. \end{eqnarray}
  3480. which can be solved by off-the-shelf MIP solvers. Nevertheless, the above linear MIP transformation holds only for the regular series-parallel redundant systems with single type of components,  as for the situation that involves multiple types (i.e. $|\mathbf{M}_i|>1$ for some $i \in \mathbf{N}$) or the cold-standby subsystems are considered, the probabilistic model in general becomes intractable.
  3481.  
  3482. %\section*{Appendix II}
  3483. %\begin{proposition}\label{P-proposition1b}
  3484. %Given a system design $\x$, the worst-case probabilistic chance function (\ref{Prob-1}) solves the following linear program (LP):
  3485. %\begin{eqnarray}
  3486. %&\!\!\!\!\!\! \max  &  1-  \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{k\in \mathbf{K}_{ij}} \left(\alpha^{k}_{ij}\underline{\mu}_{{ij}}+ \beta^{k}_{ij}\overline{\mu}_{{ij}}\right)-\sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i}\epsilon_{ij}\lambda_{ij} - \tau \label{HP1-ambiguity-LP-FLaa} \\
  3487. % &\!\!\!\!\!\!{\rm s.t.} & \sum\limits_{j\in \mathbf{M}^{\rm p}_{l}}\sum\limits_{k\in \mathbf{K}_{lj}} q_{ljk}\mathcal{T}_R\nonumber\\
  3488. % &&+\sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{k\in \mathbf{K}_{ij}} \left[ \phi^{l }^{k}_{ij}\underline{z}_{ij}+ \varphi^{l}^{k}_{ij}\overline{z}_{ij}  + {\nu_{ij}\left(\pi^{l }^{k}_{ij}-\varpi^{l }^{k}_{ij} \right)} \right]+\tau \ge 1,~\forall {l  \in [3;5]}  \label{HP1-ambiguity-LP-FL1aa}\\
  3489. % &&p_{l}\mathcal{T}_R+\sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{k\in \mathbf{K}_{ij}} \left[ \phi^{l }^{k}_{ij}\underline{z}_{ij}+ \varphi^{l}^{k}_{ij}\overline{z}_{ij}  + {\nu_{ij}\left(\pi^{l }^{k}_{ij}-\varpi^{l }^{k}_{ij} \right)} \right]+\tau \ge 1,~\forall {l  \in [1;2]}\\
  3490. %  && \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{k\in \mathbf{K}_{ij}} \left[ \varsigma^{k}_{ij}\underline{z}_{ij}+ \vartheta^{k}_{ij}\overline{z}_{ij}  +  {\nu_{ij}\left(\gamma^{k}_{ij}-\theta^{k}_{ij} \right)}\right]+\tau \ge 0\label{HP1-ambiguity-LP-FL1-2aa}\\
  3491. % && q_{l jk}x_{l jk} +\phi^{l }_{l jk}+\varphi^{l }_{l jk}+ { \pi^{l }_{l jk}-\varpi^{l }_{l jk} }  = \alpha_{l jk}+\beta_{l jk},~\forall {l \in [3;5]}, j \in \mathbf{M}^{\rm p}_l, k\in \mathbf{K}_{l j}   \\
  3492. %&& p_{l}x_{l jk} +\phi^{l }_{l jk}+\varphi^{l }_{l jk}+ { \pi^{l }_{l jk}-\varpi^{l }_{l jk} }  = \alpha_{l jk}+\beta_{l jk},~\forall {l \in [1;2]}, j \in \mathbf{M}^{\rm c}_l, k\in \mathbf{K}_{l j}   \\
  3493. %&&\phi^{l }^{k}_{ij}+\varphi^{l }^{k}_{ij}+ { \pi^{l }^{k}_{ij}-\varpi^{l }^{k}_{ij} }= \alpha^{k}_{ij}+\beta^{k}_{ij}, ~ \forall {l  \in \mathbf{N}}, i \in \mathbf{N}\setminus\{l \}, j \in \mathbf{M}_i, k\in \mathbf{K}_{ij}  \label{HP1-ambiguity-LP-FL2aa}\\
  3494. %&&{|\mathbf{K}_{ij} |\sigma_{ij}} (\pi^{l }^{k}_{ij}+\varpi^{l }^{k}_{ij}) =\lambda_{ij},  ~ \forall {l  \in \mathbf{N}}, i \in \mathbf{N}, j\in \mathbf{M}_i, k\in \mathbf{K}_{ij} \\
  3495. %&& \varsigma^{k}_{ij}+ \vartheta^{k}_{ij} + { \gamma^{k}_{ij}-\theta^{k}_{ij} }= \alpha^{k}_{ij}+\beta^{k}_{ij}, ~ \forall  i \in \mathbf{N}, j\in \mathbf{M}_i, k\in \mathbf{K}_{ij} \\
  3496. %&&  {|\mathbf{K}_{ij} |\sigma_{ij}}(\gamma^{k}_{ij}+\theta^{k}_{ij}) = \lambda_{{ij}}, ~  \forall  i \in \mathbf{N}, j\in \mathbf{M}_i, k\in \mathbf{K}_{ij} \\
  3497. %%&&  q_{l jk}\le y_{{l jk}}, ~\forall l  \in \mathbf{N}, j \in \mathbf{M}_{l}, k\in \mathcal{N}(l,j) \\[0.3 cm]
  3498. %%&& y_{{l jk}} \ge M x_{l jk}, ~\forall l  \in \mathbf{N}, j \in \mathbf{M}_{l}, k\in \mathcal{N}(l,j) \\[0.3 cm]
  3499. %%&& y_{{l jk}} \le  q_{l jk }+(x_{l jk}-1)M, ~\forall l  \in \mathbf{N}, j \in \mathcal{J}({l}), k \in \mathcal{N}(l,j)\\[0.3 cm]
  3500. %%&&  \sum\limits_{k\in \mathbf{K}_{ij}} x^{k}_{ij}\ge  L_{ij}, ~  \forall  i \in \mathbf{N}, j\in \mathbf{M}_i \\
  3501. %&& \alpha^{k}_{ij}\le 0, \beta^{k}_{ij}\ge 0, \lambda_{{ij}} \ge 0, \tau \in \mathbb{R}, ~\forall i \in \mathbf{N}, j\in \mathbf{M}_i, k\in \mathbf{K}_{ij}\\
  3502. %&&q_{l jk}\le 0, ~\forall {l  \in [3;5]}, j\in \mathbf{M}^{\rm p}_{l}, k \in \mathbf{K}_{lj}\\
  3503. %&&p_{l}\le 0, ~\forall {l  \in [1;2]}\\
  3504. %&& \phi^{l }^{k}_{ij} \ge 0,  \varphi^{l }^{k}_{ij} \le 0,  \pi^{l }^{k}_{ij}\ge 0,\varpi^{l }^{k}_{ij}\ge 0, ~\forall {l  \in \mathbf{N}},  i \in \mathbf{N}, j\in \mathbf{M}_i, k\in \mathbf{K}_{ij}\\
  3505. %&& \theta^{k}_{ij}\ge 0, \gamma^{k}_{ij}\ge 0, \varsigma^{k}_{ij}\ge 0, \vartheta^{k}_{ij} \le 0, ~\forall  i \in \mathbf{N}, j\in \mathbf{M}_i, k\in \mathbf{K}_{ij}, \label{HP2-ambiguity-LP-FLaa}
  3506. %\end{eqnarray}
  3507. %where $\halpha, \hbeta, \hlambda, \tau, \q, \s, \hphi, \hvarphi, \hpi, \hvarpi, \htheta, \hgamma, \hvarsigma$ and $\hvartheta$ are auxiliary variables.
  3508. %\end{proposition}
  3509. %
  3510. %\begin{proposition}\label{proposition1b}
  3511. %The robust system reliability redundancy allocation problem (\ref{HP1-ambiguity-X}) can be cast into the following mixed integer linear program (MILP):
  3512. %\begin{eqnarray}
  3513. % & \min\limits_{\x} &  \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \left[\sum\limits_{k\in \mathbf{K}_{ij}} x^{k}_{ij}\right]c_{ij} \label{HP1-ambiguity-MILP-FL1aa}\\
  3514. % &{\rm s.t.} & 1-  \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{k\in \mathbf{K}_{ij}} \left(\alpha^{k}_{ij}\underline{\mu}_{{ij}}+ \beta^{k}_{ij}\overline{\mu}_{{ij}}\right)-\sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i}\epsilon_{ij}\lambda_{ij} - \tau\ge R_{0}   \\
  3515. % &&  \sum_{j\in \mathbf{M}_i}\sum\limits_{k\in \mathbf{K}_{ij}} x^{k}_{ij}\ge  L_{i}, ~  \forall  i \in \mathbf{N} \\
  3516. %%&& \sum\limits_{j\in \mathbf{M}_{l}} \sum\limits_{k\in \mathbf{K}_{l j}}q_{l jk}\mathcal{T}_R\nonumber\\
  3517. %% &&+\sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{k\in \mathbf{K}_{ij}} \left[ \phi^{l }^{k}_{ij}\underline{z}_{ij}+ \varphi^{\varsigma}^{k}_{ij}\overline{z}_{ij}  + {\nu_{ij}\left(\pi^{l }^{k}_{ij}-\varpi^{l }^{k}_{ij} \right)} \right]+\tau \ge 1,~\forall {l  \in \mathbf{N}}  \\
  3518. %&& y^{\rm p}_{l jk} +\phi^{l }_{l jk}+\varphi^{l }_{l jk}+ { \pi^{l }_{l jk}-\varpi^{l }_{l jk} }  = \alpha_{l jk}+\beta_{l jk},~\forall {l \in [3;5]}, j \in \mathbf{M}^{\rm p}_{l}, k\in \mathbf{K}_{l j}    \\
  3519. %&& y^{\rm c}_{l jk} +\phi^{l }_{l jk}+\varphi^{l }_{l jk}+ { \pi^{l }_{l jk}-\varpi^{l }_{l jk} }  = \alpha_{l jk}+\beta_{l jk},~\forall {l \in [1;2]}, j \in \mathbf{M}^{\rm c}_{l}, k\in \mathbf{K}_{l j}    \\
  3520. % && (\ref{HP1-ambiguity-LP-FL1aa}-\ref{HP1-ambiguity-LP-FL1-2aa}); (\ref{HP1-ambiguity-LP-FL2aa})-(\ref{HP2-ambiguity-LP-FLaa})\\
  3521. %%&&\phi^{l }^{k}_{ij}+\varphi^{l }^{k}_{ij}+ { \pi^{l }^{k}_{ij}-\varpi^{l }^{k}_{ij} }= \alpha^{k}_{ij}+\beta^{k}_{ij}, ~ \forall {l  \in \mathbf{N}}, i \in \mathbf{N}\setminus\{l \}, j \in \mathbf{M}_i, k\in \mathbf{K}_{ij}  \\
  3522. %%&&{|\mathbf{K}_{ij} |\sigma_{ij}} (\pi^{l }^{k}_{ij}+\varpi^{l }^{k}_{ij}) =\lambda_{ij},  ~ \forall {l  \in \mathbf{N}}, i \in \mathbf{N}, j\in \mathbf{M}_i, k\in \mathbf{K}_{ij} \\
  3523. %%&& \sum\limits_{i\in \mathbf{N}}\sum\limits_{j\in \mathbf{M}_i} \sum\limits_{k\in \mathbf{K}_{ij}} \left[\left(\varsigma^{k}_{ij}\underline{z}_{ij}+ \vartheta^{k}_{ij}\overline{z}_{ij} \right) +  {\nu_{ij}\left(\gamma^{k}_{ij}-\theta^{k}_{ij} \right)}\right]+\tau \ge 0\\
  3524. %%&& \varsigma^{k}_{ij}+ \vartheta^{k}_{ij} + { \gamma^{k}_{ij}-\theta^{k}_{ij} }= \alpha^{k}_{ij}+\beta^{k}_{ij}, ~ \forall  i \in \mathbf{N}, j\in \mathbf{M}_i, k\in \mathbf{K}_{ij} \\
  3525. %%&&  {|\mathbf{K}_{ij} |\sigma_{ij}}(\gamma^{k}_{ij}+\theta^{k}_{ij}) = \lambda_{{ij}}, ~  \forall  i \in \mathbf{N}, j\in \mathbf{M}_i, k\in \mathbf{K}_{ij} \\
  3526. %&&  q_{l jk}\le y^{\rm p}_{{l jk}}, ~\forall l  \in [3;5], j \in \mathbf{M}^{\rm p}_{l}, k\in \mathbf{K}_{l j} \\
  3527. %&& y^{\rm p}_{{l jk}} \ge M_1 x_{l jk}, ~\forall l  \in [3;5], j \in \mathbf{M}^{\rm p}_{l}, k\in \mathbf{K}_{l j} \\
  3528. %&& y^{\rm p}_{{l jk}} \le  q_{l jk}+(x_{l jk}-1)M_1, ~\forall l  \in [3;5], j \in \mathbf{M}^{\rm p}_{l}, k \in \mathbf{K}_{l j}\\
  3529. %&&  p_{l}\le y^{\rm c}_{{l jk}}, ~\forall l  \in [1;2], j \in \mathbf{M}^{\rm c}_{l}, k\in \mathbf{K}_{l j} \\
  3530. %&& y^{\rm c}_{{l jk}} \ge M_2 x_{l jk}, ~\forall l  \in [1;2], j \in \mathbf{M}^{\rm c}_{l}, k\in \mathbf{K}_{l j} \\
  3531. %&& y^{\rm c}_{{l jk}} \le  p_{l }+(x_{l jk}-1)M_2, ~\forall l  \in [1;2], j \in \mathbf{M}^{\rm c}_{l}, k \in \mathbf{K}_{l j}\\
  3532. %%&& \alpha^{k}_{ij}\le 0, \beta^{k}_{ij}\ge 0, \lambda_{{ij}} \ge 0, \tau \in \Re, ~\forall i \in \mathbf{N}, j\in \mathbf{M}_i, k\in \mathbf{K}_{ij}\\
  3533. %%&&q_{l jk}\le 0, y_{l jk} \le 0, ~\forall {l  \in \mathbf{N}}, j\in \mathbf{M}_{l}, k\in \mathbf{K}_{l j} \\
  3534. %%&& \phi^{l }^{k}_{ij} \ge 0,  \varphi^{l }^{k}_{ij} \le 0,  \pi^{l }^{k}_{ij}\ge 0,\varpi^{l }^{k}_{ij}\ge 0, ~\forall {l  \in \mathbf{N}},  i \in \mathbf{N}, j\in \mathbf{M}_i, k\in \mathbf{K}_{ij}\\
  3535. %%&& \theta^{k}_{ij}\ge 0, \gamma^{k}_{ij}\ge 0, \varsigma^{k}_{ij}\ge 0, \vartheta^{k}_{ij} \le 0, ~\forall  i \in \mathbf{N}, j\in \mathbf{M}_i, k\in \mathbf{K}_{ij}\\
  3536. %&& y^{\rm p}_{{l jk}}\le 0,  ~ \forall l \in [3;5], j\in \mathbf{M}^{\rm a}_i, k \in \mathbf{K}_{l j}\\
  3537. %&& y^{\rm c}_{{l jk}}\le 0,  ~ \forall l \in [1;2], j\in \mathbf{M}^{\rm c}_i, k \in \mathbf{K}_{l j}\\
  3538. %&&  x^{k}_{ij} \in \{0,1\},  ~ \forall i \in \mathbf{N}, j\in \mathbf{M}_i, k \in \mathbf{K}_{ij}, \label{HP1-ambiguity-MILP-FL2aa}
  3539. %\end{eqnarray}
  3540. %where $\halpha, \hbeta, \hlambda, \tau, \q, \s, \y^{\rm p}, \y^{\rm c}, \hphi, \hvarphi, \hpi, \hvarpi, \htheta, \hvartheta, \hvarsigma$ and $\hvartheta$ are auxilary variables and $M$ is a sufficiently small negative number.
  3541. %\end{proposition}
  3542.  
  3543. \end{document}
  3544. %%
  3545. %% E    nd of file `elsarticle-template-num.tex'.
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement