首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
Let us assume that observations are obtained at random and sequentially from a population with density function In this paper we consider a sequential rule for estimating μ when σ is unknown corresponding to the following class of cost functions In this paper we consider a sequential rule for estimating μ when σ is unknown corresponding to the following class of cost functions Where δ(XI,…,XN) is a suitable estimator of μ based on the random sample (X1,…, XN), N is a stopping variable, and A and p are given constants. To study the performance of the rule it is compared with corresponding “optimum fixed sample procedures” with known σ by comparing expected sample sizes and expected costs. It is shown that the rule is “asymptotically efficient” when absolute loss (p=-1) is used whereas the one based on squared error (p = 2) is not. A table is provided to show that in small samples similar conclusions are also true.  相似文献   

2.
In this paper we consider a simple three-order-statistic asymptotically unbiased estimator of the Weibull shape parameter c for the case in which all three parameters are unknown. Optimal quantiles that minimize the asymptotic variance of this estimator, c? are determined and shown to depend only on the true (unknown) shape parameter value c and in a rather insensitive way. Monte Carlo studies further verified that, in practice where the true shape parameter c is unknown, using always c? with the optimal quantities that correspond to c = 2.0 produces estimates, c?, remarkably close to the theoretical optimal. A second stage estimation procedure, namely recalculating c? based on the optimal quantiles corresponding to c?, was not worth the additional effort. Benchmark simulation comparisons were also made with the best percentile estimator of Zanakis [20] and with a new estimator of Wyckoff, Bain and Engelhardt [18], one that appears to be the best of proposed closed-form estimators but uses all sample observations. The proposed estimator, c?, should be of interest to practitioners having limited resources and to researchers as a starting point for more accurate iterative estimation procedures. Its form is independent of all three Weibull parameters and, for not too large sample sizes, it requires the first, last and only one other (early) ordered observation. Practical guidelines are provided for choosing the best anticipated estimator of shape for a three-parameter Weibull distribution under different circumstances.  相似文献   

3.
An approximation for P(X2 + Y2 ≤ K2σ21) based on an unpublished result of Kleinecke is derived, where X and Y are independent normal variables having zero means and variances σ21 and σ22 and σ1 ≥ σ2. Also, we provide asymptotic expressions for the probabilities for large values of β = K2(1 - c2)/4c2 where c = σ21. These are illustrated by comparing with values tabulated by Harter [6]. Solution of K for specified P and c is also considered. The main point of this note is that simple and easily calculable approximations for P and K can be developed and there is no need for numerical evaluation of integrals.  相似文献   

4.
For each n, X1(n),…, Xn(n) are independent and identically distributed random variables, each with cumulative distribution function F(x) which is known to be absolutely continuous but is otherwise unknown. The problem is to test the hypothesis that \documentclass{article}\pagestyle{empty}\begin{document}$ F(x) = G\left( {{\textstyle{{x - \theta _1 } \over {\theta _2 }}}} \right) $\end{document}, where the cumulative distribution function Gx is completely specified and satisfies certain regularity conditions, and the parameters θ1, θ2 are unknown and unspecified, except that the scale parameter θ2, is positive. Y1 (n) ≦ Y2 (n) ≦ … ≦ Yn (n)are the ordered values of X1(n),…, Xn(n). A test based on a certain subset of {Yi(n)} is proposed, is shown to have asymptotically a normal distribution when the hypothesis is true, and is shown to be consistent against all alternatives satisfying a mild regularity condition.  相似文献   

5.
A system reliability is often evaluated by individual tests of components that constitute the system. These component test plans have advantages over complete system based tests in terms of time and cost. In this paper, we consider the series system with n components, where the lifetime of the i‐th component follows exponential distribution with parameter λi. Assuming test costs for the components are different, we develop an efficient algorithm to design a two‐stage component test plan that satisfies the usual probability requirements on the system reliability and in addition minimizes the maximum expected cost. For the case of prior information in the form of upper bounds on λi's, we use the genetic algorithm to solve the associated optimization problems which are otherwise difficult to solve using mathematical programming techniques. The two‐stage component test plans are cost effective compared to single‐stage plans developed by Rajgopal and Mazumdar. We demonstrate through several numerical examples that our approach has the potential to reduce the overall testing costs significantly. © 2002 John Wiley & Sons, Inc. Naval Research Logistics, 49: 95–116, 2002; DOI 10.1002/nav.1051  相似文献   

6.
Suppose that observations from populations π1, …, πk (k ≥ 1) are normally distributed with unknown means μ1., μk, respectively, and a common known variance σ2. Let μ[1] μ … ≤ μ[k] denote the ranked means. We take n independent observations from each population, denote the sample mean of the n observation from π1 by X i (i = 1, …, k), and define the ranked sample means X [1] ≤ … ≤ X [k]. The problem of confidence interval estimation of μ(1), …,μ[k] is stated and related to previous work (Section 1). The following results are obtained (Section 2). For i = 1, …, k and any γ(0 < γ < 1) an upper confidence interval for μ[i] with minimal probability of coverage γ is (? ∞, X [i]+ h) with h = (σ/n1/2) Φ?11/k-i+1), where Φ(·) is the standard normal cdf. A lower confidence interval for μ[i] with minimal probability of coverage γ is (X i[i]g, + ∞) with g = (σ/n1/2) Φ?11/i). For the upper confidence interval on μ[i] the maximal probability of coverage is 1– [1 – γ1/k-i+1]i, while for the lower confidence interval on μ[i] the maximal probability of coverage is 1–[1– γ1/i] k-i+1. Thus the maximal overprotection can always be calculated. The overprotection is tabled for k = 2, 3. These results extend to certain translation parameter families. It is proven that, under a bounded completeness condition, a monotone upper confidence interval h(X 1, …, X k) for μ[i] with probability of coverage γ(0 < γ < 1) for all μ = (μ[1], …,μ[k]), does not exist.  相似文献   

7.
A statistic is determined for testing the hypothesis of equality for scale parameters from two populations, each of which has the first asymptotic distribution of smallest (extreme) values. The probability distribution is derived for this statistic, and critical values are determined and given in tabular form for a one-sided or two-sided alternative, for censored samples of size n1 and n2, n1 = 2, 3, …. 6, n2 = 2, 3, …. 6. The power function of the test for certain alternatives is also calculated and listed in each case considered.  相似文献   

8.
Let Xt, t = 1,2, ?, be a stationary Gaussian Markov process with E(Xt) = μ and Cov(Xt, Xt+k) = σ2ρk. We derive a prediction interval for X2n+1 based on the preceding 2n observations X1,X2, ?,X2n.  相似文献   

9.
In this note some extensions are made to previous work by a number of authors on the development of tests for exponentiality. The most recent example is due to Fercho and Ringer in which they compare the small sample powers of a few well-known test statistics for the hypothesis of a constant failure rate. It is the primary intent of this current work to extend Gnedenko's F test to situations with hypercensoring and to provide guidance for its use, particularly when a log-normal distribution is the alternative.  相似文献   

10.
Consider an experiment in which only record-breaking values (e.g., values smaller than all previous ones) are observed. The data available may be represented as X1,K1,X2,K2, …, where X1,X2, … are successive minima and K1,K2, … are the numbers of trials needed to obtain new records. We treat the problem of estimating the mean of an underlying exponential distribution, and we consider both fixed sample size problems and inverse sampling schemes. Under inverse sampling, we demonstrate certain global optimality properties of an estimator based on the “total time on test” statistic. Under random sampling, it is shown than an analogous estimator is consistent, but can be improved for any fixed sample size.  相似文献   

11.
This paper investigates the problem of choosing between two simple hypothesis, H0 and H1, in terms of independent, identically distributed random variables, when observations can be taken in groups. At any stage in the decision process it must be decided whether to stop and take action now or to continue, in which case the size of the next group of observations must be decided upon. The problem is to find an optimal procedure incorporating a stopping, group size (batch) and terminal action rule. It is proven, in general, that the optimal stopping and terminal action rule is of the sequential probability ratio type (SPRT). Fixed stopping rules of the SPRT type are studied and an iterative procedure of the policy improvement type, both with and without a value determination step, is developed. It is shown, for the general situation, that both the average risk and scheduling rule converge to the optima. Also, six suboptimal scheduling rules are considered with respect to the average risks they achieve. Numerical results are presented to illustrate the effectiveness of the procedures.  相似文献   

12.
There are given k (? 2) univariate cumulative distribution functions (c.d.f.'s) G(x; θi) indexed by a real-valued parameter θi, i=1,…, k. Assume that G(x; θi) is stochastically increasing in θi. In this paper interval estimation on the ith smallest of the θ's and related topics are studied. Applications are considered for location parameter, normal variance, binomial parameter, and Poisson parameter.  相似文献   

13.
A new approach is presented for analyzing multiple-attribute decision problems in which the set of actions is finite and the utility function is additive. The problem can be resolved if the decision makers (or group of decision makers) specifies a set of nonnegative weights for the various attributes or criteria, but we here assume that the decision maker(s) cannot provide a numerical value for each such weight. Ordinal information about these weights is therefore obtained from the decision maker(s), and this information is translated into a set of linear constraints which restrict the values of the weights. These constraints are then used to construct a polytope W of feasible weight vectors, and the subsets Hi (polytopes) of W over which each action ai has the greatest utility are determined. With the Comparative Hypervolume Criterion we calculate for each action the ratio of the hypervolume of Hi to the hypervolume of W and suggest the choice of an action with the largest such ratio. Justification of this choice criterion is given, and a computational method for accurately approximating the hypervolume ratios is described. A simple example is provided to evaluate the efficiency of a computer code developed to implement the method.  相似文献   

14.
A stochastic production-maximizing problem with transportation constraints is considered where the production rates, Rij, of man i — job j combinations are random variables rather than constants. It is shown that for the family of Weibull distributions (of which the Exponential is a special case) with scale parameters λij and shape parameter β, the plan that maximizes the expected rate of the entire line is obtained by solving a deterministic fixed charge transportation problem with no linear costs and with “set-up” cost matrix ‖λij‖.  相似文献   

15.
From an original motivation in quantitative inventory modeling, we develop methods for testing the hypothesis that the service times of an M/G/1 queue are exponentially distributed, given a sequence of observations of customer line and/or system waits. The approaches are mostly extensions of the well-known exponential goodness-of-fit test popularized by Gnedenko, which results from the observation that the sum of a random exponential sample is Erlang distributed and thus that the quotient of two independent exponential sample means is F distributed.  相似文献   

16.
Given a target T in Euclidean n-space Rn and a point bomb whose point of impact in Rn is governed by a probability distribution about the aim point a, what choice of a maximizes the probability of a hit va(T)? Of course, only in special cases is an exact solution of this problem obtainable. This paper treats targets T which are symmetric about the origin o and demonstrates conditions on the extent of T and the impact density f, a density with respect to Lebesgue measure, sufficient for va(T) to be monotone in the distance from a to o and maximized at a = o. The results are applied to various tactical situations.  相似文献   

17.
One branch of the reliability literature is concerned with devising statistical procedures with various nonparametric “restricted family” model assumptions because of the potential improved operating characteristics of such procedures over totally nonparametric ones. In the single-sample problem with unknown increasing failure rate (IFR) distribution F, (1) maximum-likelihood estimators of F have been calculated, (2) upper or lower tolerance limits for F have been determined, and (3) tests of the null hypothesis that F is exponential have been constructed. Barlow and Campo proposed graphical methods for assessing goodness of fit to the IFR model when the validity of this assumption is unknown. This article proposes several analytic tests of the IFR null hypothesis based on the maximum distance and area between the cumulative hazard function and its greatest convex minorant (GCM), and the maximum distance and area between the total time on test statistic and its GCM. A table of critical points is provided to implement a specific test having good overall power properties.  相似文献   

18.
Consider an auction in which increasing bids are made in sequence on an object whose value θ is known to each bidder. Suppose n bids are received, and the distribution of each bid is conditionally uniform. More specifically, suppose the first bid X1 is uniformly distributed on [0, θ], and the ith bid is uniformly distributed on [Xi?1, θ] for i = 2, …?, n. A scenario in which this auction model is appropriate is described. We assume that the value θ is un known to the statistician and must be esimated from the sample X1, X2, …?, Xn. The best linear unbiased estimate of θ is derived. The invariance of the estimation problem under scale transformations in noted, and the best invariant estimation problem under scale transformations is noted, and the best invariant estimate of θ under loss L(θ, a) = [(a/θ) ? 1]2 is derived. It is shown that this best invariant estimate has uniformly smaller mean-squared error than the best linear unbiased estimate, and the ratio of the mean-squared errors is estimated from simulation experiments. A Bayesian formulation of the estimation problem is also considered, and a class of Bayes estimates is explicitly derived.  相似文献   

19.
A new piece of equipment has been purchased in a lot of size m. Some of the items can be used in destructive testing before the item is put into use. Testing uncovers faults which can be removed from the remaining pieces of equipment in the lot. If t < m pieces of equipment are tested, then those that remain, m1 = mt, have reduced fault incidence and are more reliable than initially, but m1 may be too small to be useful, or than is desirable. In this paper models are studied to address this question: given the lot size m, how to optimize by choice of t the effectiveness of the pieces of equipment remaining after the test. The models used are simplistic and illustrative; they can be straightforwardly improved. © 1997 John Wiley & Sons, Inc. Naval Research Logistics 44: 623–637, 1997  相似文献   

20.
The problem treated here involves a mixed fleet of vehicles comprising two types of vehicles: K1 tanker-type vehicles capable of refueling themselves and other vehicles, and K2 nontanker vehicles incapable of refueling. The two groups of vehicles have different fuel capacities as well as different fuel consumption rates. The problem is to find the tanker refueling sequence that maximizes the range attainable for the K2 nontankers. A tanker refueling sequence is a partition of the tankers into m subsets (2 ≤ mK1). A given sequence of the partition provides a realization of the number of tankers participating in each successive refueling operation. The problem is first formulated as a nonlinear mixed-integer program and reduced to a linear program for a fixed sequence which may be solved by a simple recursive procedure. It is proven that a “unit refueling sequence” composed of one tanker refueling at each of K1 refueling operations is optimal. In addition, the problem of designing the “minimum fleet” (minimum number of tankers) required for a given set of K2 nontankers to attain maximal range is resolved. Also studied are extensions to the problem with a constraint on the number of refueling operations, different nontanker recovery base geometry, and refueling on the return trip.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号