首页 | 本学科首页   官方微博 | 高级检索  
文章检索
  按 检索   检索词:      
出版年份:   被引次数:   他引次数: 提示:输入*表示无穷大
  收费全文   320篇
  免费   71篇
  国内免费   1篇
  392篇
  2021年   2篇
  2019年   11篇
  2018年   4篇
  2017年   15篇
  2016年   21篇
  2015年   15篇
  2014年   15篇
  2013年   67篇
  2012年   16篇
  2011年   23篇
  2010年   21篇
  2009年   18篇
  2008年   22篇
  2007年   31篇
  2006年   20篇
  2005年   15篇
  2004年   21篇
  2003年   15篇
  2002年   16篇
  2001年   10篇
  2000年   11篇
  1999年   2篇
  1997年   1篇
排序方式: 共有392条查询结果,搜索用时 0 毫秒
1.
Wildfire managers use initial attack (IA) to control wildfires before they grow large and become difficult to suppress. Although the majority of wildfire incidents are contained by IA, the small percentage of fires that escape IA causes most of the damage. Therefore, planning a successful IA is very important. In this article, we study the vulnerability of IA in wildfire suppression using an attacker‐defender Stackelberg model. The attacker's objective is to coordinate the simultaneous ignition of fires at various points in a landscape to maximize the number of fires that cannot be contained by IA. The defender's objective is to optimally dispatch suppression resources from multiple fire stations located across the landscape to minimize the number of wildfires not contained by IA. We use a decomposition algorithm to solve the model and apply the model on a test case landscape. We also investigate the impact of delay in the response, the fire growth rate, the amount of suppression resources, and the locations of fire stations on the success of IA.  相似文献   
2.
It is well‐known that the efficient set of a multiobjective linear programming (MOLP) problem can be represented as a union of the maximal efficient faces of the feasible region. In this paper, we propose a method for finding all maximal efficient faces for an MOLP. The new method is based on a condition that all efficient vertices (short for the efficient extreme points and rays) for the MOLP have been found and it relies on the adjacency, affine independence and convexity results of efficient sets. The method uses a local top‐down search strategy to determine maximal efficient faces incident to every efficient vertex for finding maximal efficient faces of an MOLP problem. To our knowledge, the proposed method is the first top‐down search method that uses the adjacency property of the efficient set to find all maximal efficient faces. We discuss this and other advantages and disadvantages of the algorithm. We also discuss some computational experience we have had with our computer code for implementing the algorithm. This computational experience involved solving several MOLP problems with the code.  相似文献   
3.
Finding all nondominated vectors for multi‐objective combinatorial optimization (MOCO) problems is computationally very hard in general. We approximate the nondominated frontiers of MOCO problems by fitting smooth hypersurfaces. For a given problem, we fit the hypersurface using a single nondominated reference vector. We experiment with different types of MOCO problems and demonstrate that in all cases the fitted hypersurfaces approximate all nondominated vectors well. We discuss that such an approximation is useful to find the neighborhood of preferred regions of the nondominated vectors with very little computational effort. Further computational effort can then be spent in the identified region to find the actual nondominated vectors the decision maker will prefer. © 2009 Wiley Periodicals, Inc. Naval Research Logistics, 2009  相似文献   
4.
Todas information and communication network requires a design that is secure to tampering. Traditional performance measures of reliability and throughput must be supplemented with measures of security. Recognition of an adversary who can inflict damage leads toward a game‐theoretic model. Through such a formulation, guidelines for network designs and improvements are derived. We opt for a design that is most robust to withstand both natural degradation and adversarial attacks. Extensive computational experience with such a model suggests that a Nash‐equilibrium design exists that can withstand the worst possible damage. Most important, the equilibrium is value‐free in that it is stable irrespective of the unit costs associated with reliability vs. capacity improvement and how one wishes to trade between throughput and reliability. This finding helps to pinpoint the most critical components in network design. From a policy standpoint, the model also allows the monetary value of information‐security to be imputed. © 2009 Wiley Periodicals, Inc. Naval Research Logistics, 2009  相似文献   
5.
We consider the problem of scheduling customer orders in a flow shop with the objective of minimizing the sum of tardiness, earliness (finished goods inventory holding), and intermediate (work‐in‐process) inventory holding costs. We formulate this problem as an integer program, and based on approximate solutions to two different, but closely related, Dantzig‐Wolfe reformulations, we develop heuristics to minimize the total cost. We exploit the duality between Dantzig‐Wolfe reformulation and Lagrangian relaxation to enhance our heuristics. This combined approach enables us to develop two different lower bounds on the optimal integer solution, together with intuitive approaches for obtaining near‐optimal feasible integer solutions. To the best of our knowledge, this is the first paper that applies column generation to a scheduling problem with different types of strongly ????‐hard pricing problems which are solved heuristically. The computational study demonstrates that our algorithms have a significant speed advantage over alternate methods, yield good lower bounds, and generate near‐optimal feasible integer solutions for problem instances with many machines and a realistically large number of jobs. © 2004 Wiley Periodicals, Inc. Naval Research Logistics, 2004.  相似文献   
6.
Failure rate and mean residual life are two important characteristics for studying reliability of products. In literature, some work studied the shape of failure rate function based on the knowledge of the associated probability density function; some other work investigated the shape of mean residual life function based on the shape of the associated failure rate function separately for continuous case and discrete case. In this article, a general approach is developed which can be applied to the aforementioned studies. © 2004 Wiley Periodicals, Inc. Naval Research Logistics, 2004  相似文献   
7.
In this paper we study higher‐order Markov chain models for analyzing categorical data sequences. We propose an efficient estimation method for the model parameters. Data sequences such as DNA and sales demand are used to illustrate the predicting power of our proposed models. In particular, we apply the developed higher‐order Markov chain model to the server logs data. The objective here is to model the users' behavior in accessing information and to predict their behavior in the future. Our tests are based on a realistic web log and our model shows an improvement in prediction. © 2004 Wiley Periodicals, Inc. Naval Research Logistics, 2004  相似文献   
8.
We consider the problem of scheduling a set of jobs on a single machine subject to random breakdowns. We focus on the preemptive‐repeat model, which addresses the situation where, if a machine breaks down during the processing of a job, the work done on the job prior to the breakdown is lost and the job will have to be started from the beginning again when the machine resumes its work. We allow that (i) the uptimes and downtimes of the machine follow general probability distributions, (ii) the breakdown process of the machine depends upon the job being processed, (iii) the processing times of the jobs are random variables following arbitrary distributions, and (iv) after a breakdown, the processing time of a job may either remain a same but unknown amount, or be resampled according to its probability distribution. We first derive the optimal policy for a class of problems under the criterion to maximize the expected discounted reward earned from completing all jobs. The result is then applied to further obtain the optimal policies for other due date‐related criteria. We also discuss a method to compute the moments and probability distributions of job completion times by using their Laplace transforms, which can convert a general stochastic scheduling problem to its deterministic equivalent. The weighted squared flowtime problem and the maintenance checkup and repair problem are analyzed as applications. © 2004 Wiley Periodicals, Inc. Naval Research Logistics, 2004  相似文献   
9.
Extending Sastry's result on the uncapacitated two‐commodity network design problem, we completely characterize the optimal solution of the uncapacitated K‐commodity network design problem with zero flow costs for the case when K = 3. By solving a set of shortest‐path problems on related graphs, we show that the optimal solutions can be found in O(n3) time when K = 3, where n is the number of nodes in the network. The algorithm depends on identifying a list of “basic patterns”; the number of basic patterns grows exponentially with K. We also show that the uncapacitated K‐commodity network design problem can be solved in O(n3) time for general K if K is fixed; otherwise, the time for solving the problem is exponential. © 2004 Wiley Periodicals, Inc. Naval Research Logistics, 2004  相似文献   
10.
This paper reports on a study using the available oil monitoring information, such as the data obtained using the Spectrometric Oil Analysis Programme (SOAP), to predict the residual life of a set of aircraft engines. The relationship between oil monitoring information and the residual life is established using the concept of the proportional residual, which states that the predicted residual life may be proportional to the wear increment measured by the oil analysis programmes. Assuming such a relationship between wear and the residual life exists, we formulated a recursive prediction model for the item's residual life given measured oil monitoring information to date. A set of censored life data of 30 aircraft engines (right censored due to preventive overhaul) along with the history of their monitored metal concentration information are available to us. The metal concentration information includes many variables, such as Fe, Cu, Al, etc.; not all of them are useful, and some of them may be correlated. The principal component analysis (PCA) has been adopted to reduce the dimension of the original data set and to produce a new set of uncorrelated variables, which we shall use in the prediction model. The procedure associated with estimating model parameters is discussed. The model is fitted to the actual SOAP data from the aircraft engines, and the goodness‐of‐fit test has been carried out. © 2005 Wiley Periodicals, Inc. Naval Research Logistics, 2005.  相似文献   
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号