首页 | 本学科首页   官方微博 | 高级检索  
文章检索
  按 检索   检索词:      
出版年份:   被引次数:   他引次数: 提示:输入*表示无穷大
  收费全文   473篇
  免费   0篇
  473篇
  2021年   8篇
  2019年   14篇
  2018年   5篇
  2017年   9篇
  2016年   10篇
  2015年   11篇
  2014年   8篇
  2013年   80篇
  2012年   5篇
  2011年   8篇
  2010年   6篇
  2009年   5篇
  2008年   5篇
  2007年   8篇
  2006年   8篇
  2005年   5篇
  2004年   6篇
  2003年   6篇
  2002年   5篇
  2001年   5篇
  1999年   9篇
  1997年   7篇
  1996年   8篇
  1994年   8篇
  1993年   13篇
  1992年   11篇
  1991年   11篇
  1990年   5篇
  1989年   9篇
  1988年   8篇
  1987年   6篇
  1986年   8篇
  1985年   9篇
  1984年   7篇
  1982年   5篇
  1981年   7篇
  1980年   8篇
  1979年   9篇
  1978年   9篇
  1977年   4篇
  1976年   8篇
  1975年   11篇
  1974年   9篇
  1973年   8篇
  1972年   6篇
  1971年   7篇
  1970年   5篇
  1969年   6篇
  1968年   7篇
  1967年   4篇
排序方式: 共有473条查询结果,搜索用时 15 毫秒
451.
    
This paper considers a two-agent scheduling problem with linear resource-dependent processing times, in which each agent has a set of jobs that compete with that of the other agent for the use of a common processing machine, and each agent aims to minimize the weighted number of its tardy jobs. To meet the due date requirements of the jobs of the two agents, additional amounts of a common resource, which may be in discrete or continuous quantities, can be allocated to the processing of the jobs to compress their processing durations. The actual processing time of a job is a linear function of the amount of the resource allocated to it. The objective is to determine the optimal job sequence and resource allocation strategy so as to minimize the weighted number of tardy jobs of one agent, while keeping the weighted number of tardy jobs of the other agent, and the total resource consumption cost within their respective predetermined limits. It is shown that the problem is -hard in the ordinary sense, and there does not exist a polynomial-time approximation algorithm with performance ratio unless ; however it admits a relaxed fully polynomial time approximation scheme. A proximal bundle algorithm based on Lagrangian relaxation is also presented to solve the problem approximately. To speed up convergence and produce sharp bounds, enhancement strategies including the design of a Tabu search algorithm and integration of a Lagrangian recovery heuristic into the algorithm are devised. Extensive numerical studies are conducted to assess the effectiveness and efficiency of the proposed algorithms.  相似文献   
452.
    
For the single‐machine scheduling problem with the objective of simultaneously minimizing total flow time and number of tardy jobs, a lower bound on the number of efficient sequences is known. However, the proof thereof, which makes use of a modified version of Smith's algorithm, is unduly lengthy and sophisticated. Adopting a totally new point of view, we present in this short article a much simpler proof based on the naive idea of pairwise interchange. © 2016 Wiley Periodicals, Inc. Naval Research Logistics 63: 346–348, 2016  相似文献   
453.
    
“Evergreening” is a strategy wherein an innovative pharmaceutical firm introduces an upgrade of its current product when the patent on this product expires. The upgrade is introduced with a new patent and is designed to counter competition from generic manufacturers that seek to imitate the firm's existing product. However, this process is fraught with uncertainty because the upgrade is subject to stringent guidelines and faces approval risk. Thus, an incumbent firm has to make an upfront production capacity investment without clarity on whether the upgrade will reach the market. This uncertainty may also affect the capacity investment of a competing manufacturer who introduces a generic version of the incumbent's existing product but whose market demand depends on the success or failure of the upgrade. We analyze a game where capacity investment occurs before uncertainty resolution and firms compete on prices thereafter. Capacity considerations that arise due to demand uncertainty introduce new factors into the evergreening decision. Equilibrium analysis reveals that the upgrade's estimated approval probability needs to exceed a threshold for the incumbent to invest in evergreening. This threshold for evergreening increases as the intensity of competition in the generic market increases. If evergreening is optimal, the incumbent's capacity investment is either decreasing or nonmonotonic with respect to low end market competition depending on whether the level of product improvement in the upgrade is low or high. If the entrant faces a capacity constraint, then the probability threshold for evergreening is higher than the case where the entrant is not capacity constrained. Finally, by incorporating the risk‐return trade‐off that the incumbent faces in terms of the level of product improvement versus the upgrade success probability, we can characterize policy for a regulator. We show that the introduction of capacity considerations may maximize market coverage and/or social surplus at incremental levels of product improvement in the upgrade. This is contrary to the prevalent view of regulators who seek to curtail evergreening involving incremental product improvement. © 2016 Wiley Periodicals, Inc. Naval Research Logistics 63: 71–89, 2016  相似文献   
454.
    
We consider the problem of scheduling n independent and simultaneously available jobs without preemption on a single machine, where the machine has a fixed maintenance activity. The objective is to find the optimal job sequence to minimize the total amount of late work, where the late work of a job is the amount of processing of the job that is performed after its due date. We first discuss the approximability of the problem. We then develop two pseudo‐polynomial dynamic programming algorithms and a fully polynomial‐time approximation scheme for the problem. Finally, we conduct extensive numerical studies to evaluate the performance of the proposed algorithms. © 2016 Wiley Periodicals, Inc. Naval Research Logistics 63: 172–183, 2016  相似文献   
455.
    
We investigate a single‐machine scheduling problem for which both the job processing times and due windows are decision variables to be determined by the decision maker. The job processing times are controllable as a linear or convex function of the amount of a common continuously divisible resource allocated to the jobs, where the resource allocated to the jobs can be used in discrete or continuous quantities. We use the common flow allowances due window assignment method to assign due windows to the jobs. We consider two performance criteria: (i) the total weighted number of early and tardy jobs plus the weighted due window assignment cost, and (ii) the resource consumption cost. For each resource consumption function, the objective is to minimize the first criterion, while keeping the value of the second criterion no greater than a given limit. We analyze the computational complexity, devise pseudo‐polynomial dynamic programming solution algorithms, and provide fully polynomial‐time approximation schemes and an enhanced volume algorithm to find high‐quality solutions quickly for the considered problems. We conduct extensive numerical studies to assess the performance of the algorithms. The computational results show that the proposed algorithms are very efficient in finding optimal or near‐optimal solutions. © 2017 Wiley Periodicals, Inc. Naval Research Logistics, 64: 41–63, 2017  相似文献   
456.
    
In this article, we study generalizations of some of the inventory models with nonlinear costs considered by Rosling in (Oper. Res. 50 (2002) 797–809). In particular, we extend the study of both the periodic review and the compound renewal demand processes from a constant lead time to a random lead time. We find that the quasiconvexity properties of the cost function (and therefore the existence of optimal (s, S) policies), holds true when the lead time has suitable log‐concavity properties. The results are derived by structural properties of renewal delayed processes stopped at an independent random time and by the study of log‐concavity properties of compound distributions. © 2015 Wiley Periodicals, Inc. Naval Research Logistics 62: 345–356, 2015  相似文献   
457.
We investigate the strategy of transshipments in a dynamic deterministic demand environment over a finite planning horizon. This is the first time that transshipments are examined in a dynamic or deterministic setting. We consider a system of two locations which replenish their stock from a single supplier, and where transshipments between the locations are possible. Our model includes fixed (possibly joint) and variable replenishment costs, fixed and variable transshipment costs, as well as holding costs for each location and transshipment costs between locations. The problem is to determine how much to replenish and how much to transship each period; thus this work can be viewed as a synthesis of transshipment problems in a static stochastic setting and multilocation dynamic deterministic lot sizing problems. We provide interesting structural properties of optimal policies which enhance our understanding of the important issues which motivate transshipments and allow us to develop an efficient polynomial time algorithm for obtaining the optimal strategy. By exploring the reasons for using transshipments, we enable practitioners to envision the sources of savings from using this strategy and therefore motivate them to incorporate it into their replenishment strategies. © 2001 John Wiley & Sons, Inc. Naval Research Logistics 48:386–408, 2001  相似文献   
458.
    
Populations of many types of component are heterogeneous and often consist of a small number of different subpopulations. This is called a mixture and it arises in a number of situations. For example, a majority of products in industrial populations are mixtures of defective items with shorter lifetimes and standard items with longer lifetimes. It is a well‐known result that distributions with decreasing failure rates are closed under mixture. However, mixtures of distributions with increasing failure rates are not easily classifiable. If the subpopulations involved in the mixture have increasing failure rates, there might be some upward movement in the mixture and later a general downward pull towards the strongest component. Little work has been done in describing the shape of mixture failure rates when all subpopulations do not have decreasing failure rate. In this paper, we present general results that describe the shape and behavior of a failure rate of a mixture obtained from two Weibull subpopulations with strictly increasing failure rates. © 2004 Wiley Periodicals, Inc. Naval Research Logistics, 2004  相似文献   
459.
    
We develop a simple, approximately optimal solution to a model with Erlang lead time and deterministic demand. The method is robust to misspecification of the lead time and has good accuracy. We compare our approximate solution to the optimal for the case where we have prior information on the lead‐time distribution, and another where we have no information, except for computer‐generated sample data. It turns out that our solution is as easy as the EOQ's, with an accuracy rate of 99.41% when prior information on the lead‐time distribution is available and 97.54–99.09% when only computer‐generated sample information is available. Apart from supplying the inventory practitioner with an easy heuristic, we gain insights into the efficacy of stochastic lead time models and how these could be used to find the cost and a near‐optimal policy for the general model, where both demand rate and lead time are stochastic. © 2004 Wiley Periodicals, Inc. Naval Research Logistics, 2004  相似文献   
460.
    
In recent years military procurement agencies have used “progress payments” in contracting. Although, overall, the adoption of such a policy may well have contributed to a decrease in total costs of procurement, we identify a class of inefficiencies that are likely to have been generated by it.  相似文献   
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号