首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
In progressive censoring, items are removed at certain times during the life test. Commonly, it is assumed that the removed items are used for further testing. In order to take into account information about these additional testing in inferential procedures, we propose a two‐step model of stage life testing with one fixed stage‐change time which incorporates information about both the removed items (further tested under different conditions) and those remaining in the current life test. We show that some marginal distributions in our model correspond either to progressive censoring with a fixed censoring time or to a simple‐step stress model. Furthermore, assuming a cumulative exposure model, we establish exact inferential results for the distribution parameters when the lifetimes are exponentially distributed. An extension to Weibull distributed lifetimes is also discussed.  相似文献   

2.
A new connection between the distribution of component failure times of a coherent system and (adaptive) progressively Type‐II censored order statistics is established. Utilizing this property, we develop inferential procedures when the data is given by all component failures until system failure in two scenarios: In the case of complete information, we assume that the failed component is also observed whereas in the case of incomplete information, we have only information about the failure times but not about the components which have failed. In the first setting, we show that inferential methods for adaptive progressively Type‐II censored data can directly be applied to the problem. For incomplete information, we face the problem that the corresponding censoring plan is not observed and that the available inferential procedures depend on the knowledge of the used censoring plan. To get estimates for distributional parameters, we propose maximum likelihood estimators which can be obtained by solving the likelihood equations directly or via an Expectation‐Maximization‐algorithm type procedure. For an exponential distribution, we discuss also a linear estimator to estimate the mean. Moreover, we establish exact distributions for some estimators in the exponential case which can be used, for example, to construct exact confidence intervals. The results are illustrated by a five component bridge system. © 2015 Wiley Periodicals, Inc. Naval Research Logistics 62: 512–530, 2015  相似文献   

3.
Mixed censoring is useful extension of Type I and Type II censoring and combines some advantages of both types of censoring. This paper proposes a general Bayesian framework for designing a variable acceptance sampling scheme with mixed censoring. A general loss function which includes the sampling cost, the time‐consuming cost, the salvage value, and the decision loss is employed to determine the Bayes risk and the corresponding optimal sampling plan. An explicit expression of the Bayes risk is derived. The new model can easily be adapted to create life testing models for different distributions. Specifically, two commonly used distributions including the exponential distribution and the Weibull distribution are considered with a special decision loss function. We demonstrate that the proposed model is superior to models with Type I or Type II censoring. Numerical examples are reported to illustrate the effectiveness of the method proposed. © 2004 Wiley Periodicals, Inc. Naval Research Logistics, 2004  相似文献   

4.
Burn‐in is a widely used method to improve the quality of products or systems after they have been produced. In this paper, we consider the problem of determining bounds to the optimal burn‐in time and optimal replacement policy maximizing the steady state availability of a repairable system. It is assumed that two types of system failures may occur: One is Type I failure (minor failure), which can be removed by a minimal repair, and the other is Type II failure (catastrophic failure), which can be removed only by a complete repair. Assuming that the underlying lifetime distribution of the system has a bathtub‐shaped failure rate function, upper and lower bounds for the optimal burn‐in time are provided. Furthermore, some other applications of optimal burn‐in are also considered. © 2004 Wiley Periodicals, Inc. Naval Research Logistics, 2004  相似文献   

5.
Chen and Bhattacharyya [Exact confidence bounds for an exponential parameter under hybrid censoring, Commun Statist Theory Methods 17 (1988), 1857–1870] considered a hybrid censoring scheme and obtained the exact distribution of the maximum likelihood estimator of the mean of an exponential distribution along with an exact lower confidence bound. Childs et al. [Exact likelihood inference based on Type‐I and Type‐II hybrid censored samples from the exponential distribution, Ann Inst Statist Math 55 (2003), 319–330] recently derived an alternative simpler expression for the distribution of the MLE. These authors also proposed a new hybrid censoring scheme and derived similar results for the exponential model. In this paper, we propose two generalized hybrid censoring schemes which have some advantages over the hybrid censoring schemes already discussed in the literature. We then derive the exact distribution of the maximum likelihood estimator as well as exact confidence intervals for the mean of the exponential distribution under these generalized hybrid censoring schemes. © 2004 Wiley Periodicals, Inc. Naval Research Logistics, 2004  相似文献   

6.
As a generalization of k‐out‐of‐n:F and consecutive k‐out‐of‐n:F systems, the consecutive k‐within‐m‐out‐of‐n:F system consists of n linearly ordered components such that the system fails iff there are m consecutive components which include among them at least k failed components. In this article, the reliability properties of consecutive k‐within‐m‐out‐of‐n:F systems with exchangeable components are studied. The bounds and approximations for the survival function are provided. A Monte Carlo estimator of system signature is obtained and used to approximate survival function. The results are illustrated and numerics are provided for an exchangeable multivariate Pareto distribution. © 2009 Wiley Periodicals, Inc. Naval Research Logistics 2009  相似文献   

7.
There has been much research on the general failure model recently. In the general failure model, when the unit fails at its age t, Type I failure (minor failure) occurs with probability 1 ? p(t) and Type II failure (catastrophic failure) occurs with probability p(t). In the previous research, some specific shapes (constant, non‐decreasing, or bathtub‐shape) on the probability function p(t) are assumed. In this article, general results on some probability functions are obtained and applied to study the shapes of p(t). The results are also applied to determining the optimal inspection and allocation policies in maintenance problems. © 2006 Wiley Periodicals, Inc. Naval Research Logistics, 2007  相似文献   

8.
By running life tests at higher stress levels than normal operating conditions, accelerated life testing (ALT) quickly yields information on the lifetime distribution of a test unit. The lifetime at the design stress is then estimated through extrapolation using a regression model. In constant‐stress testing, a unit is tested at a fixed stress level until failure or the termination time point of test, whereas step‐stress testing allows the experimenter to gradually increase the stress levels at some prefixed time points during the test. In this work, the optimal k‐level constant‐stress and step‐stress ALTs are compared for the exponential failure data under complete sampling and Type‐I censoring. The objective is to quantify the advantage of using the step‐stress testing relative to the constant‐stress one. Assuming a log‐linear life–stress relationship with the cumulative exposure model for the effect of changing stress in step‐stress testing, the optimal design points are determined under C/D/A‐optimality criteria. The efficiency of step‐stress testing to constant‐stress one is then discussed in terms of the ratio of optimal objective functions based on the information matrix. © 2013 Wiley Periodicals, Inc. Naval Research Logistics 00: 000–000, 2013  相似文献   

9.
In this paper we deal with the d‐dimensional vector packing problem, which is a generalization of the classical bin packing problem in which each item has d distinct weights and each bin has d corresponding capacities. We address the case in which the vectors of weights associated with the items are totally ordered, i.e., given any two weight vectors ai, aj, either ai is componentwise not smaller than aj or aj is componentwise not smaller than ai. An asymptotic polynomial‐time approximation scheme is constructed for this case. As a corollary, we also obtain such a scheme for the bin packing problem with cardinality constraint, whose existence was an open question to the best of our knowledge. We also extend the result to instances with constant Dilworth number, i.e., instances where the set of items can be partitioned into a constant number of totally ordered subsets. We use ideas from classical and recent approximation schemes for related problems, as well as a nontrivial procedure to round an LP solution associated with the packing of the small items. © 2002 Wiley Periodicals, Inc. Naval Research Logistics, 2003  相似文献   

10.
We study a knapsack problem with an additional minimum filling constraint, such that the total weight of selected items cannot be less than a given threshold. The problem has several applications in shipping, e‐commerce, and transportation service procurement. When the threshold equals the knapsack capacity, even finding a feasible solution to the problem is NP‐hard. Therefore, we consider the case when the ratio α of threshold to capacity is less than 1. For this case, we develop an approximation scheme that returns a feasible solution with a total profit not less than (1 ‐ ε) times the total profit of an optimal solution for any ε > 0, and with a running time polynomial in the number of items, 1/ε, and 1/(1‐α). © 2012 Wiley Periodicals, Inc. Naval Research Logistics, 2013  相似文献   

11.
The present study is concerned with the determination of a few observations from a sufficiently large complete or censored sample from the extreme value distribution with location and scale parameters μ and σ, respectively, such that the asymptotically best linear unbiased estimators (ABLUE) of the parameters in Ref. [24] yield high efficiencies among other choices of the same number of observations. (All efficiencies considered are relative to the Cramér-Rao lower bounds for regular unbiased estimators.) The study is on the asymptotic theory and under Type II censoring scheme. For the estimation of μ when σ is known, it has been proved that there exists a unique optimum spacing whether the sample is complete, right censored, left censored, or doubly censored. Several tables are prepared to aid in the numerical computation of the estimates as well as to furnish their efficiencies. For the estimation of σ when μ is known, it has been observed that there does not exist a unique optimum spacing. Accordingly we have obtained a spacing based on a complete sample which yields high efficiency. A similar table as above is prepared. When both μ and σ are unknown, we have considered four different spacings based on a complete sample and chosen the one yielding highest efficiency. A table of the efficiencies is also prepared. Finally we apply the above results for the estimation of the scale and/or shape parameters of the Weibull distribution.  相似文献   

12.
The signature of a system with independent and identically distributed (i.i.d.) component lifetimes is a vector whose ith element is the probability that the ith component failure is fatal to the system. System signatures have been found to be quite useful tools in the study and comparison of engineered systems. In this article, the theory of system signatures is extended to versions of signatures applicable in dynamic reliability settings. It is shown that, when a working used system is inspected at time t and it is noted that precisely k failures have occurred, the vector s [0,1]nk whose jth element is the probability that the (k + j)th component failure is fatal to the system, for j = 1,2,2026;,nk, is a distribution‐free measure of the design of the residual system. Next, known representation and preservation theorems for system signatures are generalized to dynamic versions. Two additional applications of dynamic signatures are studied in detail. The well‐known “new better than used” (NBU) property of aging systems is extended to a uniform (UNBU) version, which compares systems when new and when used, conditional on the known number of failures. Sufficient conditions are given for a system to have the UNBU property. The application of dynamic signatures to the engineering practice of “burn‐in” is also treated. Specifically, we consider the comparison of new systems with working used systems burned‐in to a given ordered component failure time. In a reliability economics framework, we illustrate how one might compare a new system to one successfully burned‐in to the kth component failure, and we identify circumstances in which burn‐in is inferior (or is superior) to the fielding of a new system. © 2009 Wiley Periodicals, Inc. Naval Research Logistics, 2009  相似文献   

13.
Accelerated life testing (ALT) is widely used to determine the failure time distribution of a product and the associated life‐stress relationship in order to predict the product's reliability under normal operating conditions. Many types of stress loadings such as constant‐stress, step‐stress and cyclic‐stress can be utilized when conducting ALT. Extensive research has been conducted on the analysis of ALT data obtained under a specified stress loading. However, the equivalency of ALT experiments involving different stress loadings has not been investigated. In this article, a log‐location‐scale distribution under Type I censoring is considered in planning ALT. An idea is provided for the equivalency of various ALT plans involving different stress loadings. Based on this idea, general equivalent ALT plans and some special types of equivalent ALT plans are explored. For demonstration, a constant‐stress ALT and a ramp‐stress ALT for miniature lamps are presented and their equivalency is investigated. © 2010 Wiley Periodicals, Inc. Naval Research Logistics, 2010  相似文献   

14.
In this paper, we study a m‐parallel machine scheduling problem with a non‐crossing constraint motivated by crane scheduling in ports. We decompose the problem to allow time allocations to be determined once crane assignments are known and construct a backtracking search scheme that manipulates domain reduction and pruning strategies. Simple approximation heuristics are developed, one of which guarantees solutions to be at most two times the optimum. For large‐scale problems, a simulated annealing heuristic that uses random neighborhood generation is provided. Computational experiments are conducted to test the algorithms. © 2006 Wiley Periodicals, Inc. Naval Research Logistics, 2007.  相似文献   

15.
Under a free-replacement warranty of duration W, the customer is provided, for an initial cost of C, as many replacement items as needed to provide service for a period W. Payments of C are not made at fixed intervals of length W, but in random cycles of length Y = W + γ(W), where γ(W) is the (random) remaining life-time of the item in service W time units after the beginning of a cycle. The expected number of payments over the life cycle, L, of the item is given by MY(L), the renewal function for the random variable Y. We investigate this renewal function analytically and numerically and compare the latter with known asymptotic results. The distribution of Y, and hence the renewal function, depends on the underlying failure distribution of the items. Several choices for this distribution, including the exponential, uniform, gamma and Weibull, are considered.  相似文献   

16.
One important thrust in the reliability literature is the development of statistical procedures under various “restricted family” model assumptions such as the increasing failure rate (IFR) and decreasing failure rate (DFR) distributions. However, relatively little work has been done on the problem of testing fit to such families as a null hypothesis. Barlow and Campo proposed graphical methods for assessing goodness of fit to the IFR model in single-sample problems. For the same problem with complete data, Tenga and Santner studied several analytic tests of the null hypothesis that the common underlying distribution is IFR versus the alternative that it is not IFR for complete data. This article considers the same problem for the case of four types of censored data: (i) Type I (time) censoring, (ii) Type I1 (order statistic) censoring, (iii) a hybrid of Type I and Type I1 censoring, and (iv) random censorship. The least favorable distributions of several intuitive test statistics are derived for each of the four types of censoring so that valid small-sample-size α tests can be constructed from them. Properties of these tests are investigated.  相似文献   

17.
18.
We consider a processing network in which jobs arrive at a fork‐node according to a renewal process. Each job requires the completion of m tasks, which are instantaneously assigned by the fork‐node to m task‐processing nodes that operate like G/M/1 queueing stations. The job is completed when all of its m tasks are finished. The sojourn time (or response time) of a job in this G/M/1 fork‐join network is the total time it takes to complete the m tasks. Our main result is a closed‐form approximation of the sojourn‐time distribution of a job that arrives in equilibrium. This is obtained by the use of bounds, properties of D/M/1 and M/M/1 fork‐join networks, and exploratory simulations. Statistical tests show that our approximation distributions are good fits for the sojourn‐time distributions obtained from simulations. © 2008 Wiley Periodicals, Inc. Naval Research Logistics, 2008  相似文献   

19.
The problem of minimum makespan on an m machine jobshop with unit execution time (UET) jobs (m ≥ 3) is known to be strongly NP‐hard even with no setup times. We focus in this article on the two‐machine case. We assume UET jobs and consider batching with batch availability and machine‐dependent setup times. We introduce an efficient \begin{align*}(O(\sqrt{n}))\end{align*} algorithm, where n is the number of jobs. We then introduce a heuristic for the multimachine case and demonstrate its efficiency for two interesting instances. © 2011 Wiley Periodicals, Inc. Naval Research Logistics, 2011  相似文献   

20.
Most papers in the scheduling field assume that a job can be processed by only one machine at a time. Namely, they use a one‐job‐on‐one‐machine model. In many industry settings, this may not be an adequate model. Motivated by human resource planning, diagnosable microprocessor systems, berth allocation, and manufacturing systems that may require several resources simultaneously to process a job, we study the problem with a one‐job‐on‐multiple‐machine model. In our model, there are several alternatives that can be used to process a job. In each alternative, several machines need to process simultaneously the job assigned. Our purpose is to select an alternative for each job and then to schedule jobs to minimize the completion time of all jobs. In this paper, we provide a pseudopolynomial algorithm to solve optimally the two‐machine problem, and a combination of a fully polynomial scheme and a heuristic to solve the three‐machine problem. We then extend the results to a general m‐machine problem. Our algorithms also provide an effective lower bounding scheme which lays the foundation for solving optimally the general m‐machine problem. Furthermore, our algorithms can also be applied to solve a special case of the three‐machine problem in pseudopolynomial time. Both pseudopolynomial algorithms (for two‐machine and three‐machine problems) are much more efficient than those in the literature. © 1999 John Wiley & Sons, Inc. Naval Research Logistics 46: 57–74, 1999  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号