首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 529 毫秒
1.
Consider a k-out-of-n system with independent repairable components. Assume that the repair and failure distributions are exponential with parameters {μ1, ?,μn} and {λ1, ?,λn}, respectively. In this article we show that if λi – μi = Δ for all i then the life distribution of the system is increasing failure rate (IFR).  相似文献   

2.
Consider a system consisting of n separately maintained independent components where the components alternate between intervals in which they are “up” and in which they are “down”. When the ith component goes up [down] then, independent of the past, it remains up [down] for a random length of time, having distribution Fi[Gi], and then goes down [up]. We say that component i is failed at time t if it has been “down” at all time points s ?[t-A.t]: otherwise it is said to be working. Thus, a component is failed if it is down and has been down for the previous A time units. Assuming that all components initially start “up,” let T denote the first time they are all failed, at which point we say the system is failed. We obtain the moment-generating function of T when n = l, for general F and G, thus generalizing previous results which assumed that at least one of these distributions be exponential. In addition, we present a condition under which T is an NBU (new better than used) random variable. Finally we assume that all the up and down distributions Fi and Gi i = l,….n, are exponential, and we obtain an exact expression for E(T) for general n; in addition we obtain bounds for all higher moments of T by showing that T is NBU.  相似文献   

3.
The signature of a system with independent and identically distributed (i.i.d.) component lifetimes is a vector whose ith element is the probability that the ith component failure is fatal to the system. System signatures have been found to be quite useful tools in the study and comparison of engineered systems. In this article, the theory of system signatures is extended to versions of signatures applicable in dynamic reliability settings. It is shown that, when a working used system is inspected at time t and it is noted that precisely k failures have occurred, the vector s [0,1]nk whose jth element is the probability that the (k + j)th component failure is fatal to the system, for j = 1,2,2026;,nk, is a distribution‐free measure of the design of the residual system. Next, known representation and preservation theorems for system signatures are generalized to dynamic versions. Two additional applications of dynamic signatures are studied in detail. The well‐known “new better than used” (NBU) property of aging systems is extended to a uniform (UNBU) version, which compares systems when new and when used, conditional on the known number of failures. Sufficient conditions are given for a system to have the UNBU property. The application of dynamic signatures to the engineering practice of “burn‐in” is also treated. Specifically, we consider the comparison of new systems with working used systems burned‐in to a given ordered component failure time. In a reliability economics framework, we illustrate how one might compare a new system to one successfully burned‐in to the kth component failure, and we identify circumstances in which burn‐in is inferior (or is superior) to the fielding of a new system. © 2009 Wiley Periodicals, Inc. Naval Research Logistics, 2009  相似文献   

4.
This paper presents a branch and bound algorithm for computing optimal replacement policies in a discrete‐time, infinite‐horizon, dynamic programming model of a binary coherent system with n statistically independent components, and then specializes the algorithm to consecutive k‐out‐of‐n systems. The objective is to minimize the long‐run expected average undiscounted cost per period. (Costs arise when the system fails and when failed components are replaced.) An earlier paper established the optimality of following a critical component policy (CCP), i.e., a policy specified by a critical component set and the rule: Replace a component if and only if it is failed and in the critical component set. Computing an optimal CCP is a optimization problem with n binary variables and a nonlinear objective function. Our branch and bound algorithm for solving this problem has memory storage requirement O(n) for consecutive k‐out‐of‐n systems. Extensive computational experiments on such systems involving over 350,000 test problems with n ranging from 10 to 150 find this algorithm to be effective when n ≤ 40 or k is near n. © 2002 Wiley Periodicals, Inc. Naval Research Logistics 49: 288–302, 2002; Published online in Wiley InterScience (www.interscience.wiley.com). DOI 10.1002/nav.10017  相似文献   

5.
A system reliability is often evaluated by individual tests of components that constitute the system. These component test plans have advantages over complete system based tests in terms of time and cost. In this paper, we consider the series system with n components, where the lifetime of the i‐th component follows exponential distribution with parameter λi. Assuming test costs for the components are different, we develop an efficient algorithm to design a two‐stage component test plan that satisfies the usual probability requirements on the system reliability and in addition minimizes the maximum expected cost. For the case of prior information in the form of upper bounds on λi's, we use the genetic algorithm to solve the associated optimization problems which are otherwise difficult to solve using mathematical programming techniques. The two‐stage component test plans are cost effective compared to single‐stage plans developed by Rajgopal and Mazumdar. We demonstrate through several numerical examples that our approach has the potential to reduce the overall testing costs significantly. © 2002 John Wiley & Sons, Inc. Naval Research Logistics, 49: 95–116, 2002; DOI 10.1002/nav.1051  相似文献   

6.
T identical exponential lifetime components out of which G are initially functioning (and B are not) are to be allocated to N subsystems, which are connected either in parallel or in series. Subsystem i, i = 1,…, N, functions when at least Ki of its components function and the whole system is maintained by a single repairman. Component repair times are identical independent exponentials and repaired components are as good as new. The problem of the determination of the assembly plan that will maximize the system reliability at any (arbitrary) time instant t is solved when the component failure rate is sufficiently small. For the parallel configuration, the optimal assembly plan allocates as many components as possible to the subsystem with the smallest Ki and allocates functioning components to subsystems in increasing order of the Ki's. For the series configuration, the optimal assembly plan allocates both the surplus and the functioning components equally to all subsystems whenever possible, and when not possible it favors subsystems in decreasing order of the Ki's. The solution is interpreted in the context of the optimal allocation of processors and an initial number of jobs in a problem of routing time consuming jobs to parallel multiprocessor queues. © John Wiley & Sons, Inc. Naval Research Logistics 48: 732–746, 2001  相似文献   

7.
Suppose that observations from populations π1, …, πk (k ≥ 1) are normally distributed with unknown means μ1., μk, respectively, and a common known variance σ2. Let μ[1] μ … ≤ μ[k] denote the ranked means. We take n independent observations from each population, denote the sample mean of the n observation from π1 by X i (i = 1, …, k), and define the ranked sample means X [1] ≤ … ≤ X [k]. The problem of confidence interval estimation of μ(1), …,μ[k] is stated and related to previous work (Section 1). The following results are obtained (Section 2). For i = 1, …, k and any γ(0 < γ < 1) an upper confidence interval for μ[i] with minimal probability of coverage γ is (? ∞, X [i]+ h) with h = (σ/n1/2) Φ?11/k-i+1), where Φ(·) is the standard normal cdf. A lower confidence interval for μ[i] with minimal probability of coverage γ is (X i[i]g, + ∞) with g = (σ/n1/2) Φ?11/i). For the upper confidence interval on μ[i] the maximal probability of coverage is 1– [1 – γ1/k-i+1]i, while for the lower confidence interval on μ[i] the maximal probability of coverage is 1–[1– γ1/i] k-i+1. Thus the maximal overprotection can always be calculated. The overprotection is tabled for k = 2, 3. These results extend to certain translation parameter families. It is proven that, under a bounded completeness condition, a monotone upper confidence interval h(X 1, …, X k) for μ[i] with probability of coverage γ(0 < γ < 1) for all μ = (μ[1], …,μ[k]), does not exist.  相似文献   

8.
The opportunistic maintenance of a k‐out‐of‐n:G system with imperfect preventive maintenance (PM) is studied in this paper, where partial failure is allowed. In many applications, the optimal maintenance actions for one component often depend on the states of the other components and system reliability requirements. Two new (τ, T) opportunistic maintenance models with the consideration of reliability requirements are proposed. In these two models, only minimal repairs are performed on failed components before time τ and the corrective maintenance (CM) of all failed components are combined with PM of all functioning but deteriorated components after τ; if the system survives to time T without perfect maintenance, it will be subject to PM at time T. Considering maintenance time, asymptotic system cost rate and availability are derived. The results obtained generalize and unify some previous research in this area. Application to aircraft engine maintenance is presented. © 2000 John Wiley & Sons;, Inc. Naval Research Logistics 47: 223–239, 2000  相似文献   

9.
We consider a class of production scheduling models with m identical machines in parallel and k different product types. It takes a time pi to produce one unit of product type i on any one of the machines. There is a demand stream for product type i consisting of ni units with each unit having a given due date. Before a machine starts with the production of a batch of products of type i a setup cost c is incurred. We consider several different objective functions. Each one of the objective functions has three components, namely a total setup cost, a total earliness cost, and a total tardiness cost. In our class of problems we find a relatively large number of problems that can be solved either in polynomial time or in pseudo‐polynomial time. The polynomiality or pseudo‐polynomiality is achieved under certain special conditions that may be of practical interest; for example, a regularity pattern in the string of due dates combined with earliness and tardiness costs that are similar for different types of products. The class of models we consider includes as special cases discrete counterparts of a number of inventory models that have been considered in the literature before, e.g., Wagner and Whitin (Manage Sci 5 (1958), 89–96) and Zangwill (Oper Res 14 (1966), 486–507; Manage Sci 15 (1969), 506–527). © 2008 Wiley Periodicals, Inc. Naval Research Logistics, 2008  相似文献   

10.
There are given k (? 2) univariate cumulative distribution functions (c.d.f.'s) G(x; θi) indexed by a real-valued parameter θi, i=1,…, k. Assume that G(x; θi) is stochastically increasing in θi. In this paper interval estimation on the ith smallest of the θ's and related topics are studied. Applications are considered for location parameter, normal variance, binomial parameter, and Poisson parameter.  相似文献   

11.
This article proposes a modified preventive maintenance (PM) policy which may be done only at scheduled times nT (n = 1,2, …): The PM is done at the next such time if and only if the total number of failures exceeds a specified number k. The optimal number k* to minimize the expected cost rate is discussed. Further, four alternative similar PM models are considered, when the system fails due to a certain number of faults, uses, shocks, and unit failures.  相似文献   

12.
We study a class of new scheduling problems which involve types of teamwork tasks. Each teamwork task consists of several components, and requires a team of processors to complete, with each team member to process a particular component of the task. Once the processor completes its work on the task, it will be available immediately to work on the next task regardless of whether the other components of the last task have been completed or not. Thus, the processors in a team neither have to start, nor have to finish, at the same time as they process a task. A task is completed only when all of its components have been processed. The problem is to find an optimal schedule to process all tasks, under a given objective measure. We consider both deterministic and stochastic models. For the deterministic model, we find that the optimal schedule exhibits the pattern that all processors must adopt the same sequence to process the tasks, even under a general objective function GC = F(f1(C1), f2(C2), … , fn(Cn)), where fi(Ci) is a general, nondecreasing function of the completion time Ci of task i. We show that the optimal sequence to minimize the maximum cost MC = max fi(Ci) can be derived by a simple rule if there exists an order f1(t) ≤ … ≤ fn(t) for all t between the functions {fi(t)}. We further show that the optimal sequence to minimize the total cost TC = ∑ fi(Ci) can be constructed by a dynamic programming algorithm. For the stochastic model, we study three optimization criteria: (A) almost sure minimization; (B) stochastic ordering; and (C) expected cost minimization. For criterion (A), we show that the results for the corresponding deterministic model can be easily generalized. However, stochastic problems with criteria (B) and (C) become quite difficult. Conditions under which the optimal solutions can be found for these two criteria are derived. © 2004 Wiley Periodicals, Inc. Naval Research Logistics, 2004  相似文献   

13.
This paper develops a methodology for optimizing operation of a multipurpose reservoir with a finite capacity V. The input of water into the reservoir is a Wiener process with positive drift. There are n purposes for which water is demanded. Water may be released from the reservoir at any rate, and the release rate can be increased or decreased instantaneously with zero cost. In addition to the reservoir, a supplementary source of water can supply an unlimited amount of water demanded during any period of time. There is a cost of Ci dollars per unit of demand supplied by the supplementary source to the ith purpose (i = 1, 2, …, n). At any time, the demand rate Ri associated with the ith purpose (i = 1, 2, …, n) must be supplied. A controller must continually decide the amount of water to be supplied by the reservoir for each purpose, while the remaining demand will be supplied through the supplementary source with the appropriate costs. We consider the problem of specifying an output policy which minimizes the long run average cost per unit time.  相似文献   

14.
To location Li we are to allocate a “generator” and ni “machines” for i = 1, …,k, where n1n1 ≧ … ≧ nk. Although the generators and machines function independently of one another, a machine is operable only if it and the generator at its location are functioning. The problem we consider is that of finding the arrangement or allocation optimizing the number of operable machines. We show that if the objective is to maximize the expected number of operable machines at some future time, then it is best to allocate the best generator and the n1 best machines to location L1, the second-best generator and the n2-next-best machines to location L2, etc. However, this arrangement is not always stochastically optimal. For the case of two generators we give a necessary and sufficient condition that this arrangement is stochastically best, and illustrate the result with several examples.  相似文献   

15.
Various methods and criteria for comparing coherent systems are discussed. Theoretical results are derived for comparing systems of a given order when components are assumed to have independent and identically distributed lifetimes. All comparisons rely on the representation of a system's lifetime distribution as a function of the system's “signature,” that is, as a function of the vector p= (p1, … , pn), where pi is the probability that the system fails upon the occurrence of the ith component failure. Sufficient conditions are provided for the lifetime of one system to be larger than that of another system in three different senses: stochastic ordering, hazard rate ordering, and likelihood ratio ordering. Further, a new preservation theorem for hazard rate ordering is established. In the final section, the notion of system signature is used to examine a recently published conjecture regarding componentwise and systemwise redundancy. © 1999 John Wiley & Sons, Inc. Naval Research Logistics 46: 507–523, 1999  相似文献   

16.
Consider a network G(N. A) with n nodes, where node 1 designates its source node and node n designates its sink node. The cuts (Zi, =), i= 1…, n - 1 are called one-node cuts if 1 ? Zi,. n q Zi, Z1-? {1}, Zi ? Zi+1 and Zi and Zi+l differ by only one node. It is shown that these one-node cuts decompose G into 1 m n/2 subnetworks with known minimal cuts. Under certain circumstances, the proposed one-node decomposition can produce a minimal cut for G in 0(n2 ) machine operations. It is also shown that, under certain conditions, one-node cuts produce no decomposition. An alternative procedure is also introduced to overcome this situation. It is shown that this alternative procedure has the computational complexity of 0(n3).  相似文献   

17.
The authors study a discrete-time, infinite-horizon, dynamic programming model for the replacement of components in a binary k-out-of-n failure system. (The system fails when k or more of its n components fail.) Costs are incurred when the system fails and when failed components are replaced. The objective is to minimize the long-run expected average undiscounted cost per period. A companion article develops a branch-and-bound algorithm for computing optimal policies. Extensive computational experiments find it effective for k to be small or near n; however, difficulties are encountered when n ≥ 30 and 10 ≤ kn − 4. This article presents a simple, intuitive heuristic rule for determining a replacement policy whose memory storage and computation time requirements are O(n − k) and O(n(n − k) + k), respectively. This heuristic is based on a plausible formula for ranking components in order of their usefulness. The authors provide sufficient conditions for it to be optimal and undertake computational experiments that suggest that it handles parallel systems (k = n) effectively and, further, that its effectiveness increases as k moves away from n. In our test problems, the mean relative errors are under 5% when n ≤ 100 and under 2% when kn − 3 and n ≤ 50. © 1997 John Wiley & Sons, Inc. Naval Research Logistics 44, 273–286, 1997.  相似文献   

18.
Following a review of the basic ideas in structural reliability, including signature‐based representation and preservation theorems for systems whose components have independent and identically distributed (i.i.d.) lifetimes, extensions that apply to the comparison of coherent systems of different sizes, and stochastic mixtures of them, are obtained. It is then shown that these results may be extended to vectors of exchangeable random lifetimes. In particular, for arbitrary systems of sizes m < n with exchangeable component lifetimes, it is shown that the distribution of an m‐component system's lifetime can be written as a mixture of the distributions of k‐out‐of‐n systems. When the system has n components, the vector of coefficients in this mixture representation is precisely the signature of the system defined in Samaniego, IEEE Trans Reliabil R–34 (1985) 69–72. These mixture representations are then used to obtain new stochastic ordering properties for coherent or mixed systems of different sizes. © 2008 Wiley Periodicals, Inc. Naval Research Logistics, 2008  相似文献   

19.
It is often assumed in the facility location literature that functions of the type øi(xi, y) = βi[(xi-x)2+(yi-y)2]K/2 are twice differentiable. Here we point out that this is true only for certain values of K. Convexity proofs that are independent of the value of K are given.  相似文献   

20.
基于高温超导材料结构的准二维性,建立了高温超导的层状理论,得到了高温超导材料的临界温度Tc与Cu-O面数l间的关系,指出高物质密度N_i~0和高电子密度N_e~0的超导材料可能有更高的Tc。  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号