首页 | 本学科首页   官方微博 | 高级检索  
文章检索
  按 检索   检索词:      
出版年份:   被引次数:   他引次数: 提示:输入*表示无穷大
  收费全文   238篇
  免费   55篇
  国内免费   1篇
  2021年   1篇
  2020年   1篇
  2019年   7篇
  2018年   5篇
  2017年   6篇
  2016年   12篇
  2015年   14篇
  2014年   10篇
  2013年   70篇
  2012年   9篇
  2011年   8篇
  2010年   7篇
  2009年   7篇
  2008年   11篇
  2007年   14篇
  2006年   8篇
  2005年   13篇
  2004年   8篇
  2003年   14篇
  2002年   13篇
  2001年   6篇
  2000年   8篇
  1999年   12篇
  1998年   3篇
  1997年   5篇
  1996年   3篇
  1995年   2篇
  1994年   4篇
  1993年   2篇
  1992年   4篇
  1990年   3篇
  1989年   2篇
  1988年   1篇
  1983年   1篇
排序方式: 共有294条查询结果,搜索用时 15 毫秒
21.
Many manufacturers sell their products through retailers and share the revenue with those retailers. Given this phenomenon, we build a stylized model to investigate the role of revenue sharing schemes in supply chain coordination and product variety decisions. In our model, a monopolistic manufacturer serves two segments of consumers, which are distinguished by their willingness to pay for quality. In the scenario with exogenous revenue sharing ratios, when the potential gain from serving the low segment is substantial (e.g., the low‐segment consumers' willingness to pay is high enough or the low segment takes a large enough proportion of the market), the retailer is better off abandoning the revenue sharing scheme. Moreover, when the potential gain from serving the low (high) segment is substantial enough, the manufacturer finds it profitable to offer a single product. Furthermore, when revenue sharing ratios are endogenous, we divide our analysis into two cases, depending on the methods of cooperation. When revenue sharing ratios are negotiated at the very beginning, the decentralized supply chain causes further distortion. This suggests that the central premise of revenue sharing—the coordination of supply chains—may be undermined if supply chain parties meticulously bargain over it.  相似文献   
22.
23.
We seek dynamic server assignment policies in finite‐capacity queueing systems with flexible and collaborative servers, which involve an assembly and/or a disassembly operation. The objective is to maximize the steady‐state throughput. We completely characterize the optimal policy for a Markovian system with two servers, two feeder stations, and instantaneous assembly and disassembly operations. This optimal policy allocates one server per station unless one of the stations is blocked, in which case both servers work at the unblocked station. For Markovian systems with three stations and instantaneous assembly and/or disassembly operations, we consider similar policies that move a server away from his/her “primary” station only when that station is blocked or starving. We determine the optimal assignment of each server whose primary station is blocked or starving in systems with three stations and zero buffers, by formulating the problem as a Markov decision process. Using this optimal assignment, we develop heuristic policies for systems with three or more stations and positive buffers, and show by means of a numerical study that these policies provide near‐optimal throughput. Furthermore, our numerical study shows that these policies developed for assembly‐type systems also work well in tandem systems. © 2008 Wiley Periodicals, Inc. Naval Research Logistics, 2008  相似文献   
24.
In this article, the Building Evacuation Problem with Shared Information (BEPSI) is formulated as a mixed integer linear program, where the objective is to determine the set of routes along which to send evacuees (supply) from multiple locations throughout a building (sources) to the exits (sinks) such that the total time until all evacuees reach the exits is minimized. The formulation explicitly incorporates the constraints of shared information in providing online instructions to evacuees, ensuring that evacuees departing from an intermediate or source location at a mutual point in time receive common instructions. Arc travel time and capacity, as well as supply at the nodes, are permitted to vary with time and capacity is assumed to be recaptured over time. The BEPSI is shown to be NP‐hard. An exact technique based on Benders decomposition is proposed for its solution. Computational results from numerical experiments on a real‐world network representing a four‐story building are given. Results of experiments employing Benders cuts generated in solving a given problem instance as initial cuts in addressing an updated problem instance are also provided. © 2008 Wiley Periodicals, Inc. Naval Research Logistics, 2008  相似文献   
25.
In this paper, we derive new families of facet‐defining inequalities for the finite group problem and extreme inequalities for the infinite group problem using approximate lifting. The new valid inequalities for the finite group problem include two‐ and three‐slope facet‐defining inequalities as well as the first family of four‐slope facet‐defining inequalities. The new valid inequalities for the infinite group problem include families of two‐ and three‐slope extreme inequalities. These new inequalities not only illustrate the diversity of strong inequalities for the finite and infinite group problems, but also provide a large variety of new cutting planes for solving integer and mixed‐integer programming problems. © 2008 Wiley Periodicals, Inc. Naval Research Logistics, 2008  相似文献   
26.
Instead of measuring a Wiener degradation or performance process at predetermined time points to track degradation or performance of a product for estimating its lifetime, we propose to obtain the first‐passage times of the process over certain nonfailure thresholds. Based on only these intermediate data, we obtain the uniformly minimum variance unbiased estimator and uniformly most accurate confidence interval for the mean lifetime. For estimating the lifetime distribution function, we propose a modified maximum likelihood estimator and a new estimator and prove that, by increasing the sample size of the intermediate data, these estimators and the above‐mentioned estimator of the mean lifetime can achieve the same levels of accuracy as the estimators assuming one has failure times. Thus, our method of using only intermediate data is useful for highly reliable products when their failure times are difficult to obtain. Furthermore, we show that the proposed new estimator of the lifetime distribution function is more accurate than the standard and modified maximum likelihood estimators. We also obtain approximate confidence intervals for the lifetime distribution function and its percentiles. Finally, we use light‐emitting diodes as an example to illustrate our method and demonstrate how to validate the Wiener assumption during the testing. © 2008 Wiley Periodicals, Inc. Naval Research Logistics, 2008  相似文献   
27.
We consider the problem of sequencing n jobs on a single machine, with each job having a processing time and a common due date. The common due date is assumed to be so large that all jobs can complete by the due date. It is known that there is an O(n log n)‐time algorithm for finding a schedule with minimum total earliness and tardiness. In this article, we consider finding a schedule with dual criteria. The primary goal is to minimize the total earliness and tardiness. The secondary goals are to minimize: (1) the maximum earliness and tardiness; (2) the sum of the maximum of the squares of earliness and tardiness; (3) the sum of the squares of earliness and tardiness. For the first two criteria, we show that the problems are NP‐hard and we give a fully polynomial time approximation scheme for both of them. For the last two criteria, we show that the ratio of the worst schedule versus the best schedule is no more than . © 2002 Wiley Periodicals, Inc. Naval Research Logistics 49: 422–431, 2002; Published online in Wiley InterScience (www.interscience.wiley.com). DOI 10.1002/nav.10020  相似文献   
28.
Analytical resolution of search theory problems, as formalized by B.O. Koopman, may be applied with some model extension to various resource management issues. However, a fundamental prerequisite is the knowledge of the prior target density. Though this assumption has the definite advantage of simplicity, its drawback is clearly that target reactivity is not taken into account. As a preliminary step towards reactive target study stands the problem of resource planning under a min–max game context. This paper is related to Nakai's work about the game planning of resources for the detection of a stationary target. However, this initial problem is extended by adding new and more general constraints, allowing a more realistic modeling of the target and searcher behaviors. © 2007 Wiley Periodicals, Inc. Naval Research Logistics, 2007  相似文献   
29.
This study combines inspection and lot‐sizing decisions. The issue is whether to INSPECT another unit or PRODUCE a new lot. A unit produced is either conforming or defective. Demand need to be satisfied in full, by conforming units only. The production process may switch from a “good” state to a “bad” state, at constant rate. The proportion of conforming units in the good state is higher than in the bad state. The true state is unobservable and can only be inferred from the quality of units inspected. We thus update, after each inspection, the probability that the unit, next candidate for inspection, was produced while the production process was in the good state. That “good‐state‐probability” is the basis for our decision to INSPECT or PRODUCE. We prove that the optimal policy has a simple form: INSPECT only if the good‐state‐probability exceeds a control limit. We provide a methodology to calculate the optimal lot size and the expected costs associated with INSPECT and PRODUCE. Surprisingly, we find that the control limit, as a function of the demand (and other problem parameters) is not necessarily monotone. Also, counter to intuition, it is possible that the optimal action is PRODUCE, after revealing a conforming unit. © 2007 Wiley Periodicals, Inc. Naval Research Logistics, 2007  相似文献   
30.
We study a parallel machine scheduling problem, where a job j can only be processed on a specific subset of machines Mj, and the Mj subsets of the n jobs are nested. We develop a two‐phase heuristic for minimizing the total weighted tardiness subject to the machine eligibility constraints. In the first phase, we compute the factors and statistics that characterize a problem instance. In the second phase, we propose a new composite dispatching rule, the Apparent Tardiness Cost with Flexibility considerations (ATCF) rule, which is governed by several scaling parameters of which the values are determined by the factors obtained in the first phase. The ATCF rule is a generalization of the well‐known ATC rule which is very widely used in practice. We further discuss how to improve the dispatching rule using some simple but powerful properties without requiring additional computation time, and the improvement is quite satisfactory. We apply the Sequential Uniform Design Method to design our experiments and conduct an extensive computational study, and we perform tests on the performance of the ATCF rule using a real data set from a large hospital in China. We further compare its performance with that of the classical ATC rule. We also compare the schedules improved by the ATCF rule with what we believe are Near Optimal schedules generated by a general search procedure. The computational results show that especially with a low due date tightness, the ATCF rule performs significantly better than the well‐known ATC rule generating much improved schedules that are close to the Near Optimal schedules. © 2017 Wiley Periodicals, Inc. Naval Research Logistics 64: 249–267, 2017  相似文献   
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号