首页 | 本学科首页   官方微博 | 高级检索  
文章检索
  按 检索   检索词:      
出版年份:   被引次数:   他引次数: 提示:输入*表示无穷大
  收费全文   233篇
  免费   40篇
  2021年   3篇
  2019年   7篇
  2018年   3篇
  2017年   6篇
  2016年   13篇
  2015年   14篇
  2014年   9篇
  2013年   76篇
  2012年   9篇
  2011年   7篇
  2010年   4篇
  2009年   7篇
  2008年   9篇
  2007年   12篇
  2006年   7篇
  2005年   11篇
  2004年   7篇
  2003年   11篇
  2002年   10篇
  2001年   6篇
  2000年   5篇
  1999年   7篇
  1998年   1篇
  1995年   2篇
  1994年   2篇
  1993年   1篇
  1992年   2篇
  1990年   1篇
  1989年   2篇
  1988年   2篇
  1987年   1篇
  1986年   2篇
  1985年   2篇
  1982年   1篇
  1980年   1篇
  1979年   1篇
  1978年   1篇
  1977年   1篇
  1973年   2篇
  1972年   1篇
  1971年   1篇
  1970年   2篇
  1968年   1篇
排序方式: 共有273条查询结果,搜索用时 296 毫秒
51.
Instead of measuring a Wiener degradation or performance process at predetermined time points to track degradation or performance of a product for estimating its lifetime, we propose to obtain the first‐passage times of the process over certain nonfailure thresholds. Based on only these intermediate data, we obtain the uniformly minimum variance unbiased estimator and uniformly most accurate confidence interval for the mean lifetime. For estimating the lifetime distribution function, we propose a modified maximum likelihood estimator and a new estimator and prove that, by increasing the sample size of the intermediate data, these estimators and the above‐mentioned estimator of the mean lifetime can achieve the same levels of accuracy as the estimators assuming one has failure times. Thus, our method of using only intermediate data is useful for highly reliable products when their failure times are difficult to obtain. Furthermore, we show that the proposed new estimator of the lifetime distribution function is more accurate than the standard and modified maximum likelihood estimators. We also obtain approximate confidence intervals for the lifetime distribution function and its percentiles. Finally, we use light‐emitting diodes as an example to illustrate our method and demonstrate how to validate the Wiener assumption during the testing. © 2008 Wiley Periodicals, Inc. Naval Research Logistics, 2008  相似文献   
52.
In this paper, we present an optimization model for coordinating inventory and transportation decisions at an outbound distribution warehouse that serves a group of customers located in a given market area. For the practical problems which motivated this paper, the warehouse is operated by a third party logistics provider. However, the models developed here may be applicable in a more general context where outbound distribution is managed by another supply chain member, e.g., a manufacturer. We consider the case where the aggregate demand of the market area is constant and known per period (e.g., per day). Under an immediate delivery policy, an outbound shipment is released each time a demand is realized (e.g., on a daily basis). On the other hand, if these shipments are consolidated over time, then larger (hence more economical) outbound freight quantities can be dispatched. In this case, the physical inventory requirements at the third party warehouse (TPW) are determined by the consolidated freight quantities. Thus, stock replenishment and outbound shipment release policies should be coordinated. By optimizing inventory and freight consolidation decisions simultaneously, we compute the parameters of an integrated inventory/outbound transportation policy. These parameters determine: (i) how often to dispatch a truck so that transportation scale economies are realized and timely delivery requirements are met, and (ii) how often, and in what quantities, the stock should be replenished at the TPW. We prove that the optimal shipment release timing policy is nonstationary, and we present algorithms for computing the policy parameters for both the uncapacitated and finite cargo capacity problems. The model presented in this study is considerably different from the existing inventory/transportation models in the literature. The classical inventory literature assumes that demands should be satisfied as they arrive so that outbound shipment costs are sunk costs, or else these costs are covered by the customer. Hence, the classical literature does not model outbound transportation costs. However, if a freight consolidation policy is in place then the outbound transportation costs can no longer be ignored in optimization. Relying on this observation, this paper models outbound transportation costs, freight consolidation decisions, and cargo capacity constraints explicitly. © 2002 Wiley Periodicals, Inc. Naval Research Logistics 49: 531–556, 2002; Published online in Wiley InterScience (www.interscience.wiley.com). DOI 10.1002/nav.10030  相似文献   
53.
For computing an optimal (Q, R) or kindred inventory policy, the current literature provides mixed signals on whether or when it is safe to approximate a nonnormal lead‐time‐demand (“LTD”) distribution by a normal distribution. The first part of this paper examines this literature critically to justify why the issue warrants further investigations, while the second part presents reliable evidence showing that the system‐cost penalty for using the normal approximation can be quite serious even when the LTD‐distribution's coefficient of variation is quite low—contrary to the prevalent view of the literature. We also identify situations that will most likely lead to large system‐cost penalty. Our results indicate that, given today's technology, it is worthwhile to estimate an LTD‐distribution's shape more accurately and to compute optimal inventory policies using statistical distributions that more accurately reflect the LTD‐distributions' actual shapes. © 2003 Wiley Periodicals, Inc. Naval Research Logistics, 2003  相似文献   
54.
Today, many products are designed and manufactured to function for a long period of time before they fail. Determining product reliability is a great challenge to manufacturers of highly reliable products with only a relatively short period of time available for internal life testing. In particular, it may be difficult to determine optimal burn‐in parameters and characterize the residual life distribution. A promising alternative is to use data on a quality characteristic (QC) whose degradation over time can be related to product failure. Typically, product failure corresponds to the first passage time of the degradation path beyond a critical value. If degradation paths can be modeled properly, one can predict failure time and determine the life distribution without actually observing failures. In this paper, we first use a Wiener process to describe the continuous degradation path of the quality characteristic of the product. A Wiener process allows nonconstant variance and nonzero correlation among data collected at different time points. We propose a decision rule for classifying a unit as normal or weak, and give an economic model for determining the optimal termination time and other parameters of a burn‐in test. Next, we propose a method for assessing the product's lifetime distribution of the passed units. The proposed methodologies are all based only on the product's initial observed degradation data. Finally, an example of an electronic product, namely contact image scanner (CIS), is used to illustrate the proposed procedure. © 2002 Wiley Periodicals, Inc. Naval Research Logistics, 2003  相似文献   
55.
56.
In many practical manufacturing environments, jobs to be processed can be divided into different families such that a setup is required whenever there is a switch from processing a job of one family to another job of a different family. The time for setup could be sequence independent or sequence dependent. We consider two particular scheduling problems relevant to such situations. In both problems, we are given a set of jobs to be processed on a set of identical parallel machines. The objective of the first problem is to minimize total weighted completion time of jobs, and that of the second problem is to minimize weighted number of tardy jobs. We propose column generation based branch and bound exact solution algorithms for the problems. Computational experiments show that the algorithms are capable of solving both problems of medium size to optimality within reasonable computational time. © 2003 Wiley Periodicals, Inc. Naval Research Logistics 50: 823–840, 2003.  相似文献   
57.
In this paper we consider a major assembly composed of two or more subassemblies. The failure of any subassembly causes the major assembly to not function. Every failed subassembly is repaired or replaced. A total investment in stocks of spare components is to be distributed among the various subassemblies and the major assembly so as to provide the best possible customer service. This is a complicated problem: relevant factors are the failure rates, unit costs, and repair times of the various components. For the case of Poisson failures, a heuristic solution is developed which is a compromise between theoretical optimality and practical usefulness.  相似文献   
58.
The paper presents the formulation and several solutions of a model for allocating a fixed number of aircraft to carriers and to missions. The amount of damage that can be inflicted is maximized. A nonseparable concave nonlinear objective function expresses diminishing marginal damage. Linear constraints on aireraft, carrier space, and aircraft availability for missions are included. The model is solved using the sequential unconstrained minimization technique (SUMT). The model is presented in terms of a scenario. Several different exponential damage functions are treated, and S-shaped damage functions are discussed.  相似文献   
59.
A heuristic for 0–1 integer programming is proposed that features a specific rule for breaking ties that occur when attempting to determine a variable to set to 1 during a given iteration. It is tested on a large number of small- to moderate-sized randomly generated generalized set-packing models. Solutions are compared to those obtained using an existing well-regarded heuristic and to solutions to the linear programming relaxations. Results indicate that the proposed heuristic outperforms the existing heuristic except for models in which the number of constraints is large relative to the number of variables. In this case, it performs on par with the existing heuristic. Results also indicate that use of a specific rule for tie breaking can be very effective, especially for low-density models in which the number of variables is large relative to the number of constraints.  相似文献   
60.
We consider a two‐echelon inventory system with a manufacturer operating from a warehouse supplying multiple distribution centers (DCs) that satisfy the demand originating from multiple sources. The manufacturer has a finite production capacity and production times are stochastic. Demand from each source follows an independent Poisson process. We assume that the transportation times between the warehouse and DCs may be positive which may require keeping inventory at both the warehouse and DCs. Inventory in both echelons is managed using the base‐stock policy. Each demand source can procure the product from one or more DCs, each incurring a different fulfilment cost. The objective is to determine the optimal base‐stock levels at the warehouse and DCs as well as the assignment of the demand sources to the DCs so that the sum of inventory holding, backlog, and transportation costs is minimized. We obtain a simple equation for finding the optimal base‐stock level at each DC and an upper bound for the optimal base‐stock level at the warehouse. We demonstrate several managerial insights including that the demand from each source is optimally fulfilled entirely from a single distribution center, and as the system's utilization approaches 1, the optimal base‐stock level increases in the transportation time at a rate equal to the demand rate arriving at the DC. © 2011 Wiley Periodicals, Inc. Naval Research Logistics, 2011  相似文献   
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号