首页 | 本学科首页   官方微博 | 高级检索  
文章检索
  按 检索   检索词:      
出版年份:   被引次数:   他引次数: 提示:输入*表示无穷大
  收费全文   403篇
  免费   10篇
  2021年   11篇
  2019年   10篇
  2017年   10篇
  2016年   13篇
  2015年   13篇
  2014年   8篇
  2013年   83篇
  2011年   4篇
  2010年   2篇
  2008年   7篇
  2007年   5篇
  2006年   7篇
  2005年   7篇
  2004年   4篇
  2003年   2篇
  2002年   7篇
  2001年   7篇
  2000年   5篇
  1999年   4篇
  1998年   8篇
  1997年   9篇
  1996年   5篇
  1995年   6篇
  1994年   8篇
  1993年   7篇
  1992年   7篇
  1991年   8篇
  1990年   2篇
  1988年   3篇
  1986年   16篇
  1985年   9篇
  1984年   6篇
  1983年   6篇
  1982年   4篇
  1981年   8篇
  1980年   6篇
  1979年   11篇
  1978年   7篇
  1977年   8篇
  1976年   2篇
  1975年   9篇
  1974年   7篇
  1973年   5篇
  1972年   7篇
  1971年   4篇
  1970年   4篇
  1969年   6篇
  1968年   3篇
  1967年   2篇
  1966年   3篇
排序方式: 共有413条查询结果,搜索用时 15 毫秒
381.
We undertake inference for a stochastic form of the Lanchester combat model. In particular, given battle data, we assess the type of battle that occurred and whether or not it makes any difference to the number of casualties if an army is attacking or defending. Our approach is Bayesian and we use modern computational techniques to fit the model. We illustrate our method using data from the Ardennes campaign. We compare our results with previous analyses of these data by Bracken and Fricker. Our conclusions are somewhat different to those of Bracken. Where he suggests that a linear law is appropriate, we show that the logarithmic or linear‐logarithmic laws fit better. We note however that the basic Lanchester modeling assumptions do not hold for the Ardennes data. Using Fricker's modified data, we show that although his “super‐logarithmic” law fits best, the linear, linear‐logarithmic, and logarithmic laws cannot be ruled out. We suggest that Bayesian methods can be used to make inference for battles in progress. We point out a number of advantages: Prior information from experts or previous battles can be incorporated; predictions of future casualties are easily made; more complex models can be analysed using stochastic simulation techniques. © 2000 John Wiley & Sons, Inc. Naval Research Logistics 47: 541–558, 2000  相似文献   
382.
Acceptance sampling plans are used to assess the quality of an ongoing production process, in addition to the lot acceptance. In this paper, we consider sampling inspection plans for monitoring the Markov‐dependent production process. We construct sequential plans that satisfy the usual probability requirements at acceptable quality level and rejectable quality level and, in addition, possess the minimum average sample number under semicurtailed inspection. As these plans result in large sample sizes, especially when the serial correlation is high, we suggest new plans called “systematic sampling plans.” The minimum average sample number systematic plans that satisfy the probability requirements are constructed. Our algorithm uses some simple recurrence relations to compute the required acceptance probabilities. The optimal systematic plans require much smaller sample sizes and acceptance numbers, compared to the sequential plans. However, they need larger production runs to make a decision. Tables for choosing appropriate sequential and systematic plans are provided. The problem of selecting the best systematic sampling plan is also addressed. The operating characteristic curves of some of the sequential and the systematic plans are compared, and are observed to be almost identical. © 2001 John Wiley & Sons, Inc. Naval Research Logistics 48: 451–467, 2001  相似文献   
383.
The existing literature on economic design of process control charts generally assumes perfect process adjustment, such that the process mean is returned to an exactly centered “in control” state following any real or false alarm control chart signal. This paper presents a model which demonstrates the effects of imperfect process adjustment on the economically designed control chart parameters. The model demonstrates that the optimal control limit width depends fundamentally on the precision with which the process can be adjusted. The greater the process adjustment error, all else constant, the wider will be the optimal control limits, in order to alleviate the potential for process overcontrol and tampering effects. By endogenously modeling these effects, the new model helps to rectify the problem of poor statistical properties for which the economic design approach has been criticized. © 1999 John Wiley & Sons, Inc. Naval Research Logistics 46: 597–612, 1999  相似文献   
384.
A number of results pertaining to preservation of aging properties (IFR, IFRA etc.) under various shock models are available in the literature. Our aim in this paper is to examine in the same spirit, the preservation of unimodality under various shock models. For example, it is proved that in a non-homogeneous Poisson shock model if {pk}K≥0, the sequence of probabilities with which the device fails on the kth shock, is unimodal then under some suitable conditions on the mean value function Λ (t), the corresponding survival function is also unimodal. The other shock models under which the preservation of unimodality is considered in this paper are pure birth shock model and a more general shock model in which shocks occur according to a general counting process. © 1999 John Wiley & Sons, Inc. Naval Research Logistics 46: 952–957, 1999  相似文献   
385.
We describe the development of a heuristic algorithm for determining efficient 2-dimensional packings in cargo aircraft where cargo placement constraints are critically important in determining the feasibility of packing locations. We review the performance of a new algorithm versus some traditional ones for aircraft loading. The algorithm is also tested in a more generalized setting where there exist no additional constraints on items, to suggest applicability in other environments. The new algorithm has been used worldwide in the Automated Air Load Planning System (AALPS) for cargo aircraft loading, with much success. © 1998 John Wiley & Sons, Inc. Naval Research Logistics 45: 751–768, 1998  相似文献   
386.
In this paper, we consider a general covering problem in which k subsets are to be selected such that their union covers as large a weight of objects from a universal set of elements as possible. Each subset selected must satisfy some structural constraints. We analyze the quality of a k-stage covering algorithm that relies, at each stage, on greedily selecting a subset that gives maximum improvement in terms of overall coverage. We show that such greedily constructed solutions are guaranteed to be within a factor of 1 − 1/e of the optimal solution. In some cases, selecting a best solution at each stage may itself be difficult; we show that if a β-approximate best solution is chosen at each stage, then the overall solution constructed is guaranteed to be within a factor of 1 − 1/eβ of the optimal. Our results also yield a simple proof that the number of subsets used by the greedy approach to achieve entire coverage of the universal set is within a logarithmic factor of the optimal number of subsets. Examples of problems that fall into the family of general covering problems considered, and for which the algorithmic results apply, are discussed. © 1998 John Wiley & Sons, Inc. Naval Research Logistics 45: 615–627, 1998  相似文献   
387.
We discuss the problem of scheduling several jobs on a single machine with the objective of minimizing the weighted mean absolute deviation of flow times around the weighted mean flow time. We first show that the optimal schedule is W-shaped. For the unweighted case, we show that all optimal schedules are V-shaped. This characterization enables us to show that the problem is NP-hard. We then provide a pseudopolynomial algorithm for the unweighted problem. Finally, we consider three heuristic algorithms for the unweighted problem and report computational experience with these algorithms. © 1998 John Wiley & Sons, Inc. Naval Research Logistics 45: 297–311, 1998  相似文献   
388.
Acceptance sampling is often used to monitor the quality of raw materials and components when product testing is destructive, time-consuming, or expensive. In this paper we consider the effect of a buyer-imposed acceptance sampling policy on the optimal batch size and optimal quality level delivered by an expected cost minimizing supplier. We define quality as the supplier's process capability, i.e., the probability that a unit conforms to all product specifications, and we assume that unit cost is an increasing function of the quality level. We also assume that the supplier faces a known and constant “pass-through” cost, i.e., a fixed cost per defective unit passed on to the buyer. We show that the acceptance sampling plan has a significant impact on the supplier's optimal quality level, and we derive the conditions under which zero defects (100% conformance) is the policy that minimizes the supplier's expected annual cost. © 1997 John Wiley & Sons, Inc. Naval Research Logistics 44: 515–530, 1997  相似文献   
389.
An economic sequential screening procedure is considered for limited failure populations in which defective items fail soon after they are put in operation and nondefective ones never fail during the technical life of the items. A cost model is constructed that involves screening test cost and external failure cost. A sequential scheme that minimizes the expected cost is derived from the solution of a dynamic programming formulation and the optimal decision at each stage is obtained in a closed form. © 1994 John Wiley & Sons, Inc.  相似文献   
390.
This article examines the applicability of acceptance sampling and the effectiveness of Deming's kp rule in relation to the degree of process stability achieved through statistical process control techniques. A discrete-event simulation model is used to characterize the correlation between the number of defective units in a randomly drawn sample versus in the remainder of a lot, in response to a number of system and control chart parameters. The model reveals that such correlation is typically present when special causes of variation affect the production process from time to time, even though the process is tightly monitored through statistical process control. Comparison of these results to an analogous mixed binomial scenario reveals that the mixed binomial model overstates the correlation in question if the state of the process is not necessarily constant during lot production. A generalization of the kp analysis is presented that incorporates the possibility of dependence between a sample and the unsampled portion of the lot. This analysis demonstrates that acceptance sampling is generally ineffective for lots generated by a process subject to statistical process control, despite the fact that the number of defectives in the sample and in the remainder of the lot are not strictly independent. © 1994 John Wiley & Sons, Inc.  相似文献   
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号