全文获取类型
收费全文 | 472篇 |
免费 | 109篇 |
国内免费 | 13篇 |
专业分类
594篇 |
出版年
2024年 | 1篇 |
2022年 | 3篇 |
2021年 | 5篇 |
2020年 | 5篇 |
2019年 | 19篇 |
2018年 | 6篇 |
2017年 | 22篇 |
2016年 | 27篇 |
2015年 | 20篇 |
2014年 | 18篇 |
2013年 | 81篇 |
2012年 | 34篇 |
2011年 | 35篇 |
2010年 | 31篇 |
2009年 | 36篇 |
2008年 | 33篇 |
2007年 | 42篇 |
2006年 | 38篇 |
2005年 | 26篇 |
2004年 | 27篇 |
2003年 | 17篇 |
2002年 | 18篇 |
2001年 | 17篇 |
2000年 | 13篇 |
1999年 | 7篇 |
1998年 | 1篇 |
1997年 | 1篇 |
1995年 | 3篇 |
1993年 | 2篇 |
1992年 | 1篇 |
1991年 | 1篇 |
1990年 | 1篇 |
1989年 | 2篇 |
1988年 | 1篇 |
排序方式: 共有594条查询结果,搜索用时 0 毫秒
31.
Under quasi‐hyperbolic discounting, the valuation of a payoff falls relatively rapidly for earlier delay periods, but then falls more slowly for longer delay periods. When the salespersons with quasi‐hyperbolic discounting consider the product sale problem, they would exert less effort than their early plan, thus resulting in losses of future profit. We propose a winner‐takes‐all competition to alleviate the above time inconsistent behaviors of the salespersons, and allow the company to maximize its revenue by choosing an optimal bonus. To evaluate the effects of the competition scheme, we define the group time inconsistency degree of the salespersons, which measures the consequence of time inconsistent behaviors, and two welfare measures, the group welfare of the salespersons and the company revenue. We show that the competition always improves the group welfare and the company revenue as long as the company chooses to run the competition in the first place. However, the effect on group time inconsistency degree is mixed. When the optimal bonus is moderate (extreme high), the competition motivates (over‐motivates) the salesperson to work hard, thus alleviates (worsens) the time inconsistent behaviors. © 2017 Wiley Periodicals, Inc. Naval Research Logistics 64: 357–372, 2017 相似文献
32.
We study a setting with a single type of resource and with several players, each associated with a single resource (of this type). Unavailability of these resources comes unexpectedly and with player‐specific costs. Players can cooperate by reallocating the available resources to the ones that need the resources most and let those who suffer the least absorb all the costs. We address the cost savings allocation problem with concepts of cooperative game theory. In particular, we formulate a probabilistic resource pooling game and study them on various properties. We show that these games are not necessarily convex, do have non‐empty cores, and are totally balanced. The latter two are shown via an interesting relationship with Böhm‐Bawerk horse market games. Next, we present an intuitive class of allocation rules for which the resulting allocations are core members and study an allocation rule within this class of allocation rules with an appealing fairness property. Finally, we show that our results can be applied to a spare parts pooling situation. 相似文献
33.
Today, many products are designed and manufactured to function for a long period of time before they fail. Determining product reliability is a great challenge to manufacturers of highly reliable products with only a relatively short period of time available for internal life testing. In particular, it may be difficult to determine optimal burn‐in parameters and characterize the residual life distribution. A promising alternative is to use data on a quality characteristic (QC) whose degradation over time can be related to product failure. Typically, product failure corresponds to the first passage time of the degradation path beyond a critical value. If degradation paths can be modeled properly, one can predict failure time and determine the life distribution without actually observing failures. In this paper, we first use a Wiener process to describe the continuous degradation path of the quality characteristic of the product. A Wiener process allows nonconstant variance and nonzero correlation among data collected at different time points. We propose a decision rule for classifying a unit as normal or weak, and give an economic model for determining the optimal termination time and other parameters of a burn‐in test. Next, we propose a method for assessing the product's lifetime distribution of the passed units. The proposed methodologies are all based only on the product's initial observed degradation data. Finally, an example of an electronic product, namely contact image scanner (CIS), is used to illustrate the proposed procedure. © 2002 Wiley Periodicals, Inc. Naval Research Logistics, 2003 相似文献
34.
An attacker‐defender model for analyzing the vulnerability of initial attack in wildfire suppression 下载免费PDF全文
Wildfire managers use initial attack (IA) to control wildfires before they grow large and become difficult to suppress. Although the majority of wildfire incidents are contained by IA, the small percentage of fires that escape IA causes most of the damage. Therefore, planning a successful IA is very important. In this article, we study the vulnerability of IA in wildfire suppression using an attacker‐defender Stackelberg model. The attacker's objective is to coordinate the simultaneous ignition of fires at various points in a landscape to maximize the number of fires that cannot be contained by IA. The defender's objective is to optimally dispatch suppression resources from multiple fire stations located across the landscape to minimize the number of wildfires not contained by IA. We use a decomposition algorithm to solve the model and apply the model on a test case landscape. We also investigate the impact of delay in the response, the fire growth rate, the amount of suppression resources, and the locations of fire stations on the success of IA. 相似文献
35.
Adjacency‐based local top‐down search method for finding maximal efficient faces in multiple objective linear programming 下载免费PDF全文
It is well‐known that the efficient set of a multiobjective linear programming (MOLP) problem can be represented as a union of the maximal efficient faces of the feasible region. In this paper, we propose a method for finding all maximal efficient faces for an MOLP. The new method is based on a condition that all efficient vertices (short for the efficient extreme points and rays) for the MOLP have been found and it relies on the adjacency, affine independence and convexity results of efficient sets. The method uses a local top‐down search strategy to determine maximal efficient faces incident to every efficient vertex for finding maximal efficient faces of an MOLP problem. To our knowledge, the proposed method is the first top‐down search method that uses the adjacency property of the efficient set to find all maximal efficient faces. We discuss this and other advantages and disadvantages of the algorithm. We also discuss some computational experience we have had with our computer code for implementing the algorithm. This computational experience involved solving several MOLP problems with the code. 相似文献
36.
37.
In scheduling problems with two competing agents, each one of the agents has his own set of jobs to be processed and his own objective function, and both share a common processor. In the single‐machine problem studied in this article, the goal is to find a joint schedule that minimizes the total deviation of the job completion times of the first agent from a common due‐date, subject to an upper bound on the maximum deviation of job completion times of the second agent. The problem is shown to be NP‐hard even for a nonrestrictive due‐date, and a pseudopolynomial dynamic program is introduced and tested numerically. For the case of a restrictive due‐date (a sufficiently small due‐date that may restrict the number of early jobs), a faster pseudopolynomial dynamic program is presented. We also study the multiagent case, which is proved to be strongly NP‐hard. A simple heuristic for this case is introduced, which is tested numerically against a lower bound, obtained by extending the dynamic programming algorithm. © 2013 Wiley Periodicals, Inc. Naval Research Logistics 61: 1–16, 2014 相似文献
38.
Choosing arrival process models for service systems: Tests of a nonhomogeneous Poisson process 下载免费PDF全文
Service systems such as call centers and hospital emergency rooms typically have strongly time‐varying arrival rates. Thus, a nonhomogeneous Poisson process (NHPP) is a natural model for the arrival process in a queueing model for performance analysis. Nevertheless, it is important to perform statistical tests with service system data to confirm that an NHPP is actually appropriate, as emphasized by Brown et al. [8]. They suggested a specific statistical test based on the Kolmogorov–Smirnov (KS) statistic after exploiting the conditional‐uniform (CU) property to transform the NHPP into a sequence of i.i.d. random variables uniformly distributed on [0,1] and then performing a logarithmic transformation of the data. We investigate why it is important to perform the final data transformation and consider what form it should take. We conduct extensive simulation experiments to study the power of these alternative statistical tests. We conclude that the general approach of Brown et al. [8] is excellent, but that an alternative data transformation proposed by Lewis [22], drawing upon Durbin [10], produces a test of an NHPP test with consistently greater power. We also conclude that the KS test after the CU transformation, without any additional data transformation, tends to be best to test against alternative hypotheses that primarily differ from an NHPP only through stochastic and time dependence. © 2014 Wiley Periodicals, Inc. Naval Research Logistics 61: 66–90, 2014 相似文献
39.
In this article, we study aging properties of parallel and series systems with a random number of components. We show that the decreasing likelihood ratio property is closed under the formation of random minima. We also show, by counterexamples, that other aging properties are not closed under the formation of random minima or maxima. Some mistakes in the literature are corrected. © 2014 Wiley Periodicals, Inc. Naval Research Logistics 61: 238–243, 2014 相似文献
40.
We develop the first approximation algorithm with worst‐case performance guarantee for capacitated stochastic periodic‐review inventory systems with setup costs. The structure of the optimal control policy for such systems is extremely complicated, and indeed, only some partial characterization is available. Thus, finding provably near‐optimal control policies has been an open challenge. In this article, we construct computationally efficient approximate optimal policies for these systems whose demands can be nonstationary and/or correlated over time, and show that these policies have a worst‐case performance guarantee of 4. We demonstrate through extensive numerical studies that the policies empirically perform well, and they are significantly better than the theoretical worst‐case guarantees. We also extend the analyses and results to the case with batch ordering constraints, where the order size has to be an integer multiple of a base load. © 2014 Wiley Periodicals, Inc. Naval Research Logistics 61: 304–319, 2014 相似文献