首页 | 本学科首页   官方微博 | 高级检索  
文章检索
  按 检索   检索词:      
出版年份:   被引次数:   他引次数: 提示:输入*表示无穷大
  收费全文   466篇
  免费   53篇
  国内免费   1篇
  2021年   6篇
  2019年   12篇
  2018年   6篇
  2017年   10篇
  2016年   15篇
  2015年   18篇
  2014年   14篇
  2013年   111篇
  2012年   13篇
  2011年   10篇
  2010年   9篇
  2009年   10篇
  2008年   12篇
  2007年   14篇
  2006年   12篇
  2005年   13篇
  2004年   9篇
  2003年   14篇
  2002年   12篇
  2001年   8篇
  2000年   8篇
  1999年   8篇
  1997年   5篇
  1996年   7篇
  1994年   5篇
  1993年   10篇
  1992年   6篇
  1991年   7篇
  1989年   7篇
  1988年   5篇
  1987年   4篇
  1986年   6篇
  1985年   8篇
  1984年   5篇
  1983年   4篇
  1982年   5篇
  1981年   4篇
  1980年   8篇
  1979年   7篇
  1978年   8篇
  1976年   7篇
  1975年   10篇
  1974年   6篇
  1973年   8篇
  1972年   5篇
  1971年   4篇
  1970年   5篇
  1969年   4篇
  1968年   5篇
  1967年   3篇
排序方式: 共有520条查询结果,搜索用时 15 毫秒
161.
Within a reasonable life‐testing time, how to improve the reliability of highly reliable products is one of the great challenges to today's manufacturers. By using a resolution III experiment together with degradation test, Tseng, Hamada, and Chiao (1995) presented an interesting case study of improving the reliability of fluorescent lamps. However, in conducting such an experiment, they did not address the problem of how to choose the optimal settings of variables, such as sample size, inspection frequency, and termination time for each run, which are influential to the correct identification of significant factors and the experimental cost. Assuming that the product's degradation paths satisfy Wiener processes, this paper proposes a systematic approach to the aforementioned problem. First, an intuitively appealing identification rule is proposed. Next, under the constraints of a minimum probability of correct decision and a maximum probability of incorrect decision of the proposed identification rule, the optimum test plan (including the determinations of inspection frequency, sample size, and termination time for each run) can be obtained by minimizing the total experimental cost. An example is provided to illustrate the proposed method. © 2002 Wiley Periodicals, Inc. Naval Research Logistics 49: 514–526, 2002; Published online in Wiley InterScience (www.interscience.wiley.com). DOI 10.1002/nav.10024  相似文献   
162.
In this paper, a single‐machine scheduling problem with weighted earliness and tardiness penalties is considered. Idle time between two adjacent jobs is permitted and due dates of jobs could be unequal. The dominance rules are utilized to develop a relationship matrix, which allows a branch‐and‐bound algorithm to eliminate a high percentage of infeasible solutions. After combining this matrix with a branching strategy, a procedure to solve the problem is proposed. © 2002 Wiley Periodicals, Inc. Naval Research Logistics 49: 760–780, 2002; Published online in Wiley InterScience (www.interscience.wiley.com). DOI 10.1002/nav.10039  相似文献   
163.
This paper studies a scheduling problem arising in a beef distribution system where pallets of various types of beef products in the warehouse are first depalletized and then individual cases are loaded via conveyors to the trucks which deliver beef products to various customers. Given each customer's demand for each type of beef, the problem is to find a depalletizing and truck loading schedule that fills all the demands at a minimum total cost. We first show that the general problem where there are multiple trucks and each truck covers multiple customers is strongly NP‐hard. Then we propose polynomial‐time algorithms for the case where there are multiple trucks, each covering only one customer, and the case where there is only one truck covering multiple customers. We also develop an optimal dynamic programming algorithm and a heuristic for solving the general problem. By comparing to the optimal solutions generated by the dynamic programming algorithm, the heuristic is shown to be capable of generating near optimal solutions quickly. © 2003 Wiley Periodicals, Inc. Naval Research Logistics, 2003  相似文献   
164.
Weighted voting classifiers considered in this paper consist of N units each providing individual classification decisions. The entire system output is based on tallying the weighted votes for each decision and choosing the one which has total support weight exceeding a certain threshold. Each individual unit may abstain from voting. The entire system may also abstain from voting if no decision support weight exceeds the threshold. Existing methods of evaluating the reliability of weighted voting systems can be applied to limited special cases of these systems and impose some restrictions on their parameters. In this paper a universal generating function method is suggested which allows the reliability of weighted voting classifiers to be exactly evaluated without imposing constraints on unit weights. Based on this method, the classifier reliability is determined as a function of a threshold factor, and a procedure is suggested for finding the threshold which minimizes the cost of damage caused by classifier failures (misclassification and abstention may have different price.) Dynamic and static threshold voting rules are considered and compared. A method of analyzing the influence of units' availability on the entire classifier reliability is suggested, and illustrative examples are presented. © 2003 Wiley Periodicals, Inc. Naval Research Logistics 50: 322–344, 2003.  相似文献   
165.
In this article we present and test two heuristics for the economic lot scheduling problem. The first heuristic was developed by one of us (P.C. Geng) during Ph.D. research, while the other is a convergent implementation of an algorithm due to Doll and Whybark. We study the performance of these heuristics on a large set of test problems constructed using a new form of problem generation that yields random problems within an experimental design.  相似文献   
166.
MacGregor and Harris (J Quality Technol 25 (1993) 106–118) proposed the exponentially weighted mean squared deviation (EWMS) and the exponentially weighted moving variance (EWMV) charts as ways of monitoring process variability. These two charts are particularly useful for individual observations where no estimate of variability is available from replicates. However, the control charts derived by using the approximate distributions of the EWMS and EWMV statistics are difficult to interpret in terms of the average run length (ARL). Furthermore, both control charting schemes are biased procedures. In this article, we propose two new control charts by applying a normal approximation to the distributions of the logarithms of the weighted sum of chi squared random variables, which are respectively functions of the EWMS and EWMV statistics. These new control charts are easy to interpret in terms of the ARL. On the basis of the simulation studies, we demonstrate that the proposed charts are superior to the EWMS and EWMV charts and they both are nearly unbiased for the commonly used smoothing constants. We also compare the performance of the proposed charts with that of the change point (CP) CUSUM chart of Acosta‐Mejia (1995). The design of the proposed control charts is discussed. An example is also given to illustrate the applicability of the proposed control charts. © 2009 Wiley Periodicals, Inc. Naval Research Logistics, 2009  相似文献   
167.
We consider a processing network in which jobs arrive at a fork‐node according to a renewal process. Each job requires the completion of m tasks, which are instantaneously assigned by the fork‐node to m task‐processing nodes that operate like G/M/1 queueing stations. The job is completed when all of its m tasks are finished. The sojourn time (or response time) of a job in this G/M/1 fork‐join network is the total time it takes to complete the m tasks. Our main result is a closed‐form approximation of the sojourn‐time distribution of a job that arrives in equilibrium. This is obtained by the use of bounds, properties of D/M/1 and M/M/1 fork‐join networks, and exploratory simulations. Statistical tests show that our approximation distributions are good fits for the sojourn‐time distributions obtained from simulations. © 2008 Wiley Periodicals, Inc. Naval Research Logistics, 2008  相似文献   
168.
Since a system and its components usually deteriorate with age, preventive maintenance (PM) is often performed to restore or keep the function of a system in a good state. Furthermore, PM is capable of improving the health condition of the system and thus prolongs its effective age. There has been a vast amount of research to find optimal PM policies for deteriorating repairable systems. However, such decisions involve numerous uncertainties and the analyses are typically difficult to perform because of the scarcity of data. It is therefore important to make use of all information in an efficient way. In this article, a Bayesian decision model is developed to determine the optimal number of PM actions for systems which are maintained according to a periodic PM policy. A non‐homogeneous Poisson process with a power law failure intensity is used to describe the deteriorating behavior of the repairable system. It is assumed that the status of the system after a PM is somewhere between as good as new for a perfect repair and as good as old for a minimal repair, and for failures between two preventive maintenances, the system undergoes minimal repairs. Finally, a numerical example is given and the results of the proposed approach are discussed after performing sensitivity analysis. © 2007 Wiley Periodicals, Inc. Naval Research Logistics, 2008  相似文献   
169.
This article proposes new location models for emergency medical service stations. The models are generated by incorporating a survival function into existing covering models. A survival function is a monotonically decreasing function of the response time of an emergency medical service (EMS) vehicle to a patient that returns the probability of survival for the patient. The survival function allows for the calculation of tangible outcome measures—the expected number of survivors in case of cardiac arrests. The survival‐maximizing location models are better suited for EMS location than the covering models which do not adequately differentiate between consequences of different response times. We demonstrate empirically the superiority of the survival‐maximizing models using data from the Edmonton EMS system. © 2007 Wiley Periodicals, Inc. Naval Research Logistics, 2008  相似文献   
170.
We consider the problem of efficiently scheduling deliveries by an uncapacitated courier from a central location under online arrivals. We consider both adversary‐controlled and Poisson arrival processes. In the adversarial setting we provide a randomized (3βΔ/2δ ? 1) ‐competitive algorithm, where β is the approximation ratio of the traveling salesman problem, δ is the minimum distance between the central location and any customer, and Δ is the length of the optimal traveling salesman tour overall customer locations and the central location. We provide instances showing that this analysis is tight. We also prove a 1 + 0.271Δ/δ lower‐bound on the competitive ratio of any algorithm in this setting. In the Poisson setting, we relax our assumption of deterministic travel times by assuming that travel times are distributed with a mean equal to the excursion length. We prove that optimal policies in this setting follow a threshold structure and describe this structure. For the half‐line metric space we bound the performance of the randomized algorithm in the Poisson setting, and show through numerical experiments that the performance of the algorithm is often much better than this bound.  相似文献   
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号