全文获取类型
收费全文 | 233篇 |
免费 | 44篇 |
出版年
2021年 | 2篇 |
2019年 | 8篇 |
2018年 | 2篇 |
2017年 | 8篇 |
2016年 | 17篇 |
2015年 | 17篇 |
2014年 | 8篇 |
2013年 | 77篇 |
2012年 | 10篇 |
2011年 | 7篇 |
2010年 | 4篇 |
2009年 | 8篇 |
2008年 | 11篇 |
2007年 | 12篇 |
2006年 | 9篇 |
2005年 | 13篇 |
2004年 | 8篇 |
2003年 | 11篇 |
2002年 | 11篇 |
2001年 | 7篇 |
2000年 | 6篇 |
1999年 | 7篇 |
1997年 | 2篇 |
1993年 | 1篇 |
1991年 | 1篇 |
1988年 | 3篇 |
1987年 | 1篇 |
1986年 | 1篇 |
1983年 | 1篇 |
1982年 | 1篇 |
1980年 | 1篇 |
1977年 | 1篇 |
1971年 | 1篇 |
排序方式: 共有277条查询结果,搜索用时 125 毫秒
81.
Degradation experiments are widely used to assess the reliability of highly reliable products which are not likely to fail under the traditional life tests. In order to conduct a degradation experiment efficiently, several factors, such as the inspection frequency, the sample size, and the termination time, need to be considered carefully. These factors not only affect the experimental cost, but also affect the precision of the estimate of a product's lifetime. In this paper, we deal with the optimal design of a degradation experiment. Under the constraint that the total experimental cost does not exceed a predetermined budget, the optimal decision variables are solved by minimizing the variance of the estimated 100pth percentile of the lifetime distribution of the product. An example is provided to illustrate the proposed method. Finally, a simulation study is conducted to investigate the robustness of this proposed method. © 1999 John Wiley & Sons, Inc. Naval Research Logistics 46: 689–706, 1999 相似文献
82.
Multiple-facility loading (MFL) involves the allocation of products among a set of finite-capacity facilities. Applications of MFL arise naturally in a variety of production scheduling environments. MFL models typically assume that capacity is consumed as a linear function of products assigned to a facility. Product similarities and differences, however, result in capacity-based economies or diseconomies of scope, and thus the effective capacity of the facility is often a (nonlinear) function of the set of tasks assigned to the facility. This article addresses the multiple-facility loading problem under capacity-based economies (and diseconomies) of scope (MFLS). We formulate MFLS as a nonlinear 0–1 mixed-integer programming problem, and we discuss some useful properties. MFLS generalizes many well-known combinatorial optimization problems, such as the capacitated facility location problem and the generalized assignment problem. We also define a tabu-search heuristic and a branch-and-bound algorithm for MFLS. The tabu-search heuristic alternates between two search phases, a regional search and a diversification search, and offers a novel approach to solution diversification. We also report computational experience with the procedures. In addition to demonstrating MFLS problem tractability, the computational results indicate that the heuristic is an effective tool for obtaining high-quality solutions to MFLS. © 1997 John Wiley & Sons, Inc. Naval Research Logistics 44: 229–256, 1997 相似文献
83.
This paper introduces a general or “distribution‐free” model to analyze the lifetime of components under accelerated life testing. Unlike the accelerated failure time (AFT) models, the proposed model shares the advantage of being “distribution‐free” with the proportional hazard (PH) model and overcomes the deficiency of the PH model not allowing survival curves corresponding to different values of a covariate to cross. In this research, we extend and modify the extended hazard regression (EHR) model using the partial likelihood function to analyze failure data with time‐dependent covariates. The new model can be easily adopted to create an accelerated life testing model with different types of stress loading. For example, stress loading in accelerated life testing can be a step function, cyclic, or linear function with time. These types of stress loadings reduce the testing time and increase the number of failures of components under test. The proposed EHR model with time‐dependent covariates which incorporates multiple stress loadings requires further verification. Therefore, we conduct an accelerated life test in the laboratory by subjecting components to time‐dependent stresses, and we compare the reliability estimation based on the developed model with that obtained from experimental results. The combination of the theoretical development of the accelerated life testing model verified by laboratory experiments offers a unique perspective to reliability model building and verification. © 1999 John Wiley & Sons, Inc. Naval Research Logistics 46: 303–321, 1999 相似文献
84.
Reliability data obtained from life tests and degradation tests have been extensively used for purposes such as estimating product reliability and predicting warranty costs. When there is more than one candidate model, an important task is to discriminate between the models. In the literature, the model discrimination was often treated as a hypothesis test and a pairwise model discrimination procedure was carried out. Because the null distribution of the test statistic is unavailable in most cases, the large sample approximation and the bootstrap were frequently used to find the acceptance region of the test. Although these two methods are asymptotically accurate, their performance in terms of size and power is not satisfactory in small sample size. To enhance the small‐sample performance, we propose a new method to approximate the null distribution, which builds on the idea of generalized pivots. Conventionally, the generalized pivots were often used for interval estimation of a certain parameter or function of parameters in presence of nuisance parameters. In this study, we further extend the idea of generalized pivots to find the acceptance region of the model discrimination test. Through extensive simulations, we show that the proposed method performs better than the existing methods in discriminating between two lifetime distributions or two degradation models over a wide range of sample sizes. Two real examples are used to illustrate the proposed methods. 相似文献
85.
Christian Leuprecht Joseph Szeman David B. Skillicorn 《Contemporary Security Policy》2019,40(3):382-407
Cyberspace is a new domain of operation, with its own characteristics. Cyber weapons differ qualitatively from kinetic ones: They generate effects by non-kinetic means through information, technology, and networks. Their properties, opportunities, and constraints are comparable to the qualitative difference between conventional and nuclear weapons. New weapons and their target sets in a new domain raise a series of unresolved policy challenges at the domestic, bilateral, and international levels about deterrence, attribution, and response. They also introduce new risks: uncertainty about unintended consequences, expectations of efficacy, and uncertainty about both the target’s and the international community’s response. Cyber operations offer considerable benefits for states to achieve strategic objectives both covertly and overtly. However, without a strategic framework to contain and possibly deter their use, make state and non-state behavior more predictable in the absence of reciprocal norms, and limit their impact, an environment where states face persistent attacks that nonetheless fall below the threshold of armed conflict presents a policy dilemma that reinforces collective insecurity. 相似文献
86.
In this paper, we present an O(nm log(U/n)) time maximum flow algorithm. If U = O(n) then this algorithm runs in O(nm) time for all values of m and n. This gives the best available running time to solve maximum flow problems satisfying U = O(n). Furthermore, for unit capacity networks the algorithm runs in O(n2/3m) time. It is a two‐phase capacity scaling algorithm that is easy to implement and does not use complex data structures. © 2000 John Wiley & Sons, Inc. Naval Research Logistics 47: 511–520, 2000 相似文献
87.
We deal with the problem of minimizing makespan on a single batch processing machine. In this problem, each job has both processing time and size (capacity requirement). The batch processing machine can process a number of jobs simultaneously as long as the total size of these jobs being processed does not exceed the machine capacity. The processing time of a batch is just the processing time of the longest job in the batch. An approximation algorithm with worst‐case ratio 3/2 is given for the version where the processing times of large jobs (with sizes greater than 1/2) are not less than those of small jobs (with sizes not greater than 1/2). This result is the best possible unless P = NP. For the general case, we propose an approximation algorithm with worst‐case ratio 7/4. A number of heuristics by Uzosy are also analyzed and compared. © 2001 John Wiley & Sons, Inc. Naval Research Logistics 48: 226–240, 2001 相似文献
88.
Ebru K. Bish Thin‐Yin Leong Chung‐Lun Li Jonathan W. C. Ng David Simchi‐Levi 《海军后勤学研究》2001,48(5):363-385
We consider a container terminal discharging containers from a ship and locating them in the terminal yard. Each container has a number of potential locations in the yard where it can be stored. Containers are moved from the ship to the yard using a fleet of vehicles, each of which can carry one container at a time. The problem is to assign each container to a yard location and dispatch vehicles to the containers so as to minimize the time it takes to download all the containers from the ship. We show that the problem is NP‐hard and develop a heuristic algorithm based on formulating the problem as an assignment problem. The effectiveness of the heuristic is analyzed from both worst‐case and computational points of view. © 2001 John Wiley & Sons, Inc. Naval Research Logistics 48: 363–385, 2001 相似文献
89.
We consider the problem of finding the system with the best primary performance measure among a finite number of simulated systems in the presence of a stochastic constraint on a single real‐valued secondary performance measure. Solving this problem requires the identification and removal from consideration of infeasible systems (Phase I) and of systems whose primary performance measure is dominated by that of other feasible systems (Phase II). We use indifference zones in both phases and consider two approaches, namely, carrying out Phases I and II sequentially and carrying out Phases I and II simultaneously, and we provide specific example procedures of each type. We present theoretical results guaranteeing that our approaches (general and specific, sequential and simultaneous) yield the best system with at least a prespecified probability, and we provide a portion of an extensive numerical study aimed at evaluating and comparing the performance of our approaches. The experimental results show that both new procedures are useful for constrained ranking and selection, with neither procedure showing uniform superiority over the other.© 2010 Wiley Periodicals, Inc. Naval Research Logistics, 2010 相似文献
90.
MacGregor and Harris (J Quality Technol 25 (1993) 106–118) proposed the exponentially weighted mean squared deviation (EWMS) and the exponentially weighted moving variance (EWMV) charts as ways of monitoring process variability. These two charts are particularly useful for individual observations where no estimate of variability is available from replicates. However, the control charts derived by using the approximate distributions of the EWMS and EWMV statistics are difficult to interpret in terms of the average run length (ARL). Furthermore, both control charting schemes are biased procedures. In this article, we propose two new control charts by applying a normal approximation to the distributions of the logarithms of the weighted sum of chi squared random variables, which are respectively functions of the EWMS and EWMV statistics. These new control charts are easy to interpret in terms of the ARL. On the basis of the simulation studies, we demonstrate that the proposed charts are superior to the EWMS and EWMV charts and they both are nearly unbiased for the commonly used smoothing constants. We also compare the performance of the proposed charts with that of the change point (CP) CUSUM chart of Acosta‐Mejia (1995). The design of the proposed control charts is discussed. An example is also given to illustrate the applicability of the proposed control charts. © 2009 Wiley Periodicals, Inc. Naval Research Logistics, 2009 相似文献