首页 | 本学科首页   官方微博 | 高级检索  
文章检索
  按 检索   检索词:      
出版年份:   被引次数:   他引次数: 提示:输入*表示无穷大
  收费全文   472篇
  免费   109篇
  国内免费   13篇
  594篇
  2024年   1篇
  2022年   3篇
  2021年   5篇
  2020年   5篇
  2019年   19篇
  2018年   6篇
  2017年   22篇
  2016年   27篇
  2015年   20篇
  2014年   18篇
  2013年   81篇
  2012年   34篇
  2011年   35篇
  2010年   31篇
  2009年   36篇
  2008年   33篇
  2007年   42篇
  2006年   38篇
  2005年   26篇
  2004年   27篇
  2003年   17篇
  2002年   18篇
  2001年   17篇
  2000年   13篇
  1999年   7篇
  1998年   1篇
  1997年   1篇
  1995年   3篇
  1993年   2篇
  1992年   1篇
  1991年   1篇
  1990年   1篇
  1989年   2篇
  1988年   1篇
排序方式: 共有594条查询结果,搜索用时 0 毫秒
31.
Under quasi‐hyperbolic discounting, the valuation of a payoff falls relatively rapidly for earlier delay periods, but then falls more slowly for longer delay periods. When the salespersons with quasi‐hyperbolic discounting consider the product sale problem, they would exert less effort than their early plan, thus resulting in losses of future profit. We propose a winner‐takes‐all competition to alleviate the above time inconsistent behaviors of the salespersons, and allow the company to maximize its revenue by choosing an optimal bonus. To evaluate the effects of the competition scheme, we define the group time inconsistency degree of the salespersons, which measures the consequence of time inconsistent behaviors, and two welfare measures, the group welfare of the salespersons and the company revenue. We show that the competition always improves the group welfare and the company revenue as long as the company chooses to run the competition in the first place. However, the effect on group time inconsistency degree is mixed. When the optimal bonus is moderate (extreme high), the competition motivates (over‐motivates) the salesperson to work hard, thus alleviates (worsens) the time inconsistent behaviors. © 2017 Wiley Periodicals, Inc. Naval Research Logistics 64: 357–372, 2017  相似文献   
32.
We study a setting with a single type of resource and with several players, each associated with a single resource (of this type). Unavailability of these resources comes unexpectedly and with player‐specific costs. Players can cooperate by reallocating the available resources to the ones that need the resources most and let those who suffer the least absorb all the costs. We address the cost savings allocation problem with concepts of cooperative game theory. In particular, we formulate a probabilistic resource pooling game and study them on various properties. We show that these games are not necessarily convex, do have non‐empty cores, and are totally balanced. The latter two are shown via an interesting relationship with Böhm‐Bawerk horse market games. Next, we present an intuitive class of allocation rules for which the resulting allocations are core members and study an allocation rule within this class of allocation rules with an appealing fairness property. Finally, we show that our results can be applied to a spare parts pooling situation.  相似文献   
33.
Today, many products are designed and manufactured to function for a long period of time before they fail. Determining product reliability is a great challenge to manufacturers of highly reliable products with only a relatively short period of time available for internal life testing. In particular, it may be difficult to determine optimal burn‐in parameters and characterize the residual life distribution. A promising alternative is to use data on a quality characteristic (QC) whose degradation over time can be related to product failure. Typically, product failure corresponds to the first passage time of the degradation path beyond a critical value. If degradation paths can be modeled properly, one can predict failure time and determine the life distribution without actually observing failures. In this paper, we first use a Wiener process to describe the continuous degradation path of the quality characteristic of the product. A Wiener process allows nonconstant variance and nonzero correlation among data collected at different time points. We propose a decision rule for classifying a unit as normal or weak, and give an economic model for determining the optimal termination time and other parameters of a burn‐in test. Next, we propose a method for assessing the product's lifetime distribution of the passed units. The proposed methodologies are all based only on the product's initial observed degradation data. Finally, an example of an electronic product, namely contact image scanner (CIS), is used to illustrate the proposed procedure. © 2002 Wiley Periodicals, Inc. Naval Research Logistics, 2003  相似文献   
34.
Wildfire managers use initial attack (IA) to control wildfires before they grow large and become difficult to suppress. Although the majority of wildfire incidents are contained by IA, the small percentage of fires that escape IA causes most of the damage. Therefore, planning a successful IA is very important. In this article, we study the vulnerability of IA in wildfire suppression using an attacker‐defender Stackelberg model. The attacker's objective is to coordinate the simultaneous ignition of fires at various points in a landscape to maximize the number of fires that cannot be contained by IA. The defender's objective is to optimally dispatch suppression resources from multiple fire stations located across the landscape to minimize the number of wildfires not contained by IA. We use a decomposition algorithm to solve the model and apply the model on a test case landscape. We also investigate the impact of delay in the response, the fire growth rate, the amount of suppression resources, and the locations of fire stations on the success of IA.  相似文献   
35.
It is well‐known that the efficient set of a multiobjective linear programming (MOLP) problem can be represented as a union of the maximal efficient faces of the feasible region. In this paper, we propose a method for finding all maximal efficient faces for an MOLP. The new method is based on a condition that all efficient vertices (short for the efficient extreme points and rays) for the MOLP have been found and it relies on the adjacency, affine independence and convexity results of efficient sets. The method uses a local top‐down search strategy to determine maximal efficient faces incident to every efficient vertex for finding maximal efficient faces of an MOLP problem. To our knowledge, the proposed method is the first top‐down search method that uses the adjacency property of the efficient set to find all maximal efficient faces. We discuss this and other advantages and disadvantages of the algorithm. We also discuss some computational experience we have had with our computer code for implementing the algorithm. This computational experience involved solving several MOLP problems with the code.  相似文献   
36.
为提高信号到达角(AOA)的测量精度,采用修正的窗函数加权平均算法和基于机动加速度的非零均值时间相关模型的自适应卡尔曼滤波算法,将信号到达角的变化轨迹看作是一种做加速度机动的运动,实现对连续变化信号到达角的参数估计.计算机仿真验证了算法的有效性和可行性,证明该方法有一定的应用价值.  相似文献   
37.
In scheduling problems with two competing agents, each one of the agents has his own set of jobs to be processed and his own objective function, and both share a common processor. In the single‐machine problem studied in this article, the goal is to find a joint schedule that minimizes the total deviation of the job completion times of the first agent from a common due‐date, subject to an upper bound on the maximum deviation of job completion times of the second agent. The problem is shown to be NP‐hard even for a nonrestrictive due‐date, and a pseudopolynomial dynamic program is introduced and tested numerically. For the case of a restrictive due‐date (a sufficiently small due‐date that may restrict the number of early jobs), a faster pseudopolynomial dynamic program is presented. We also study the multiagent case, which is proved to be strongly NP‐hard. A simple heuristic for this case is introduced, which is tested numerically against a lower bound, obtained by extending the dynamic programming algorithm. © 2013 Wiley Periodicals, Inc. Naval Research Logistics 61: 1–16, 2014  相似文献   
38.
Service systems such as call centers and hospital emergency rooms typically have strongly time‐varying arrival rates. Thus, a nonhomogeneous Poisson process (NHPP) is a natural model for the arrival process in a queueing model for performance analysis. Nevertheless, it is important to perform statistical tests with service system data to confirm that an NHPP is actually appropriate, as emphasized by Brown et al. [8]. They suggested a specific statistical test based on the Kolmogorov–Smirnov (KS) statistic after exploiting the conditional‐uniform (CU) property to transform the NHPP into a sequence of i.i.d. random variables uniformly distributed on [0,1] and then performing a logarithmic transformation of the data. We investigate why it is important to perform the final data transformation and consider what form it should take. We conduct extensive simulation experiments to study the power of these alternative statistical tests. We conclude that the general approach of Brown et al. [8] is excellent, but that an alternative data transformation proposed by Lewis [22], drawing upon Durbin [10], produces a test of an NHPP test with consistently greater power. We also conclude that the KS test after the CU transformation, without any additional data transformation, tends to be best to test against alternative hypotheses that primarily differ from an NHPP only through stochastic and time dependence. © 2014 Wiley Periodicals, Inc. Naval Research Logistics 61: 66–90, 2014  相似文献   
39.
In this article, we study aging properties of parallel and series systems with a random number of components. We show that the decreasing likelihood ratio property is closed under the formation of random minima. We also show, by counterexamples, that other aging properties are not closed under the formation of random minima or maxima. Some mistakes in the literature are corrected. © 2014 Wiley Periodicals, Inc. Naval Research Logistics 61: 238–243, 2014  相似文献   
40.
We develop the first approximation algorithm with worst‐case performance guarantee for capacitated stochastic periodic‐review inventory systems with setup costs. The structure of the optimal control policy for such systems is extremely complicated, and indeed, only some partial characterization is available. Thus, finding provably near‐optimal control policies has been an open challenge. In this article, we construct computationally efficient approximate optimal policies for these systems whose demands can be nonstationary and/or correlated over time, and show that these policies have a worst‐case performance guarantee of 4. We demonstrate through extensive numerical studies that the policies empirically perform well, and they are significantly better than the theoretical worst‐case guarantees. We also extend the analyses and results to the case with batch ordering constraints, where the order size has to be an integer multiple of a base load. © 2014 Wiley Periodicals, Inc. Naval Research Logistics 61: 304–319, 2014  相似文献   
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号