首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
Computer simulation has many advantages. However, one major disadvantage is that, in all too many cases, the attempt to use computer simulation to find an optimum solution to a problem rapidly degenerates into a trial-and-error process. Techniques for overcoming this disadvantage, i. e., for making optimization and computer simulation more compatible, are applicable at two points in the development of the overall computer simulation. Techniques which are used within actual construction of the mathematical models comprising the simulation will be labeled as internal methods, while those which are used after the simulation has been completely developed will be termed external methods Because external methods appear to offer the largest potential payoff, discussion is restricted to these methods, which are essentially search techniques. In addition, the development of an “Optimizer” computer program based on these techniques is suggested Although drawbacks to the use of search techniques in the computer simulation framework exist, these techniques do offer potential for “optimization.” The modification of these techniques to satisfy the requirements of an “Optimizer” is discussed.  相似文献   

2.
We investigate the relative effectiveness of top‐down versus bottom‐up strategies for forecasting the demand of an item that belongs to a product family. The demand for each item in the family is assumed to follow a first‐order univariate autoregressive process. Under the top‐down strategy, the aggregate demand is forecasted by using the historical data of the family demand. The demand forecast for the items is then derived by proportional allocation of the aggregate forecast. Under the bottom‐up strategy, the demand forecast for each item is directly obtained by using the historical demand data of the particular item. In both strategies, the forecasting technique used is exponential smoothing. We analytically evaluate the condition under which one forecasting strategy is preferred over the other when the lag‐1 autocorrelation of the demand time series for all the items is identical. We show that when the lag‐1 autocorrelation is smaller than or equal to 1/3, the maximum difference in the performance of the two forecasting strategies is only 1%. However, if the lag‐1 autocorrelation of the demand for at least one of the items is greater than 1/3, then the bottom‐up strategy consistently outperforms the top‐down strategy, irrespective of the items' proportion in the family and the coefficient of correlation between the item demands. A simulation study reveals that the analytical findings hold even when the lag‐1 autocorrelation of the demand processes is not identical. © 2006 Wiley Periodicals, Inc. Naval Research Logistics, 2007.  相似文献   

3.
In this study, we illustrate a real‐time approximate dynamic programming (RTADP) method for solving multistage capacity decision problems in a stochastic manufacturing environment, by using an exemplary three‐stage manufacturing system with recycle. The system is a moderate size queuing network, which experiences stochastic variations in demand and product yield. The dynamic capacity decision problem is formulated as a Markov decision process (MDP). The proposed RTADP method starts with a set of heuristics and learns a superior quality solution by interacting with the stochastic system via simulation. The curse‐of‐dimensionality associated with DP methods is alleviated by the adoption of several notions including “evolving set of relevant states,” for which the value function table is built and updated, “adaptive action set” for keeping track of attractive action candidates, and “nonparametric k nearest neighbor averager” for value function approximation. The performance of the learned solution is evaluated against (1) an “ideal” solution derived using a mixed integer programming (MIP) formulation, which assumes full knowledge of future realized values of the stochastic variables (2) a myopic heuristic solution, and (3) a sample path based rolling horizon MIP solution. The policy learned through the RTADP method turned out to be superior to polices of 2 and 3. © 2010 Wiley Periodicals, Inc. Naval Research Logistics 2010  相似文献   

4.
Decades ago, simulation was famously characterized as a “method of last resort,” to which analysts should turn only “when all else fails.” In those intervening decades, the technologies supporting simulation—computing hardware, simulation‐modeling paradigms, simulation software, design‐and‐analysis methods—have all advanced dramatically. We offer an updated view that simulation is now a very appealing option for modeling and analysis. When applied properly, simulation can provide fully as much insight, with as much precision as desired, as can exact analytical methods that are based on more restrictive assumptions. The fundamental advantage of simulation is that it can tolerate far less restrictive modeling assumptions, leading to an underlying model that is more reflective of reality and thus more valid, leading to better decisions. Published 2015 Wiley Periodicals, Inc. Naval Research Logistics 62: 293–303, 2015  相似文献   

5.
We study a selling practice that we refer to as locational tying (LT), which seems to be gaining wide popularity among retailers. Under this strategy, a retailer “locationally ties” two complementary items that we denote by “primary” and “secondary.” The retailer sells the primary item in an appropriate “department” of his or her store. To stimulate demand, the secondary item is offered in the primary item's department, where it is displayed in very close proximity to the primary item. We consider two variations of LT: In the multilocation tying strategy (LT‐M), the secondary item is offered in its appropriate department in addition to the primary item's department, whereas in the single‐location tying strategy (LT‐S), it is offered only in the primary item's location. We compare these LT strategies to the traditional independent components (IC) strategy, in which the two items are sold independently (each in its own department), but the pricing/inventory decisions can be centralized (IC‐C) or decentralized (IC‐D). Assuming ample inventory, we compare and provide a ranking of the optimal prices of the four strategies. The main insight from this comparison is that relative to IC‐D, LT decreases the price of the primary item and adjusts the price of the secondary item up or down depending on its popularity in the primary item's department. We also perform a comparative statics analysis on the effect of demand and cost parameters on the optimal prices of various strategies, and identify the conditions that favor one strategy over others in terms of profitability. Then we study inventory decisions in LT under exogenous pricing by developing a model that accounts for the effect of the primary item's stock‐outs on the secondary item's demand. We find that, relative to IC‐D, LT increases the inventory level of the primary item. We also link the profitability of different strategies to the trade‐off between the increase in demand volume of the secondary item as a result of LT and the potential increase in inventory costs due to decentralizing the inventory of the secondary item. © 2009 Wiley Periodicals, Inc. Naval Research Logistics 2009  相似文献   

6.
Ranking is a common task for selecting and evaluating alternatives. In the past few decades, combining rankings results from various sources into a consensus ranking has become an increasingly active research topic. In this study, we focus on the evaluation of rank aggregation methods. We first develop an experimental data generation method, which can provide ground truth ranking for alternatives based on their “inherent ability.” This experimental data generation method can generate the required individual synthetic rankings with adjustable accuracy and length. We propose characterizing the effectiveness of rank aggregation methods by calculating the Kendall tau distance between the aggregated ranking and the ground truth ranking. We then compare four classical rank aggregation methods and present some useful findings on the relative performances of the four methods. The results reveal that both the accuracy and length of individual rankings have a remarkable effect on the comparison results between rank aggregation methods. Our methods and results may be helpful to both researchers and decision‐makers.  相似文献   

7.
We consider a make‐to‐order manufacturer facing random demand from two classes of customers. We develop an integrated model for reserving capacity in anticipation of future order arrivals from high priority customers and setting due dates for incoming orders. Our research exhibits two distinct features: (1) we explicitly model the manufacturer's uncertainty about the customers' due date preferences for future orders; and (2) we utilize a service level measure for reserving capacity rather than estimating short and long term implications of due date quoting with a penalty cost function. We identify an interesting effect (“t‐pooling”) that arises when the (partial) knowledge of customer due date preferences is utilized in making capacity reservation and order allocation decisions. We characterize the relationship between the customer due date preferences and the required reservation quantities and show that not considering the t‐pooling effect (as done in traditional capacity and inventory rationing literature) leads to excessive capacity reservations. Numerical analyses are conducted to investigate the behavior and performance of our capacity reservation and due date quoting approach in a dynamic setting with multiple planning horizons and roll‐overs. One interesting and seemingly counterintuitive finding of our analyses is that under certain conditions reserving capacity for high priority customers not only improves high priority fulfillment, but also increases the overall system fill rate. © 2008 Wiley Periodicals, Inc. Naval Research Logistics, 2008  相似文献   

8.
Research and development activities in a business firm or government laboratory are portrayed as a multi-stage information generation and conversion process. A “basic research” phase generates opportunities, in the form of findings in a set of scientific disciplines, which are available for subsequent exploitation. It is assumed that increments to information in a subject area are stochastic, proportional to the amount of knowledge which already exists in the area, and have values which are randomly distributed. An “exploratory development” phase is viewed as a process of selecting a subset of alternative research opportunities, improving each opportunity in the direction of its applications, estimating the value of the improved opportunity and using these estimates to choose the exploratory development results to be implemented in engineering development. The “engineering development” phase makes the value of exploratory results realizable without changing value or risk. Engineering development costs are assumed to increase as value increases. If exploratory development is not successful, additional costs in engineering development must be incurred to bring the design up to a minimum desirable level. The model is intended as a step toward formulating and analyzing problems in management planning and control of the several interrelated stages of the research and development process.  相似文献   

9.
This article is a sequel to a recent article that appeared in this journal, “An extensible modeling framework for dynamic reassignment and rerouting in cooperative airborne operations” [ 17 ], in which an integer programming formulation to the problem of rescheduling in‐flight assets due to changes in battlespace conditions was presented. The purpose of this article is to present an improved branch‐and‐bound procedure to solve the dynamic resource management problem in a timely fashion, as in‐flight assets must be quickly re‐tasked to respond to the changing environment. To facilitate the rapid generation of attractive updated mission plans, this procedure uses a technique for reducing the solution space, supports branching on multiple decision variables simultaneously, incorporates additional valid cuts to strengthen the minimal network constraints of the original mathematical model, and includes improved objective function bounds. An extensive numerical analysis indicates that the proposed approach significantly outperforms traditional branch‐and‐bound methodologies and is capable of providing improved feasible solutions in a limited time. Although inspired by the dynamic resource management problem in particular, this approach promises to be an effective tool for solving other general types of vehicle routing problems. © 2013 Wiley Periodicals, Inc. Naval Research Logistics, 2013  相似文献   

10.
A simple formula is found to be just as accurate as a complicated one for estimating the probability of detection achievable by an ingenious searcher patrolling a channel or barrier. The difference between “detection” and “closure” is emphasized in an extension.  相似文献   

11.
The use of commercial business management techniques is widespread in all government departments, including the Ministry of Defence. This article examines the use of popular management techniques in the Armed Forces and argues that their application is misplaced. It looks at what the “effs” – “efficiency” and “effectiveness” – mean in the business world and to the Armed Forces. It compares the definitions both in business and the Armed Forces and finds that there are few, if any, situations where the same measurements can be applied. Whilst many management techniques are suited for business, the function of the Armed Forces and its output cannot be measured in the same way, complicated by the different metrics of “efficiency” in peace and in war. This difference may not be clearly understood by some politicians, or indeed by some senior military personnel. Using examples from some of the most popular management techniques such as “Lean” and “Agile” it is possible to see that their use might actually diminish the capabilities of the Armed Forces when it comes to performing their principal role – the use of force to achieve political objectives.  相似文献   

12.
This paper obtains the uniformly minimum variance unbiased estimates of two indices of performance of a system which alternates between two states “up” or “down” in accordance with a Markov process. The two indices are (1) operational readiness, which measures the probability that the system will be up when needed; and (2) operational reliability, which measures the probability that the system will be up during the entire time of need. For the purpose of obtaining these estimates, two types of observations are considered: (a) those which reveal only the state of system at isolated time-points, and (b) those which continuously record the duration of the “up” and “down” times of the system.  相似文献   

13.
This article provides formulas for estimating the parameters to be used in the basic EOQ lot-size model. The analysis assumes that the true values of these parameters are unknown over known ranges and perhaps nonstationary over time. Two measures of estimator “goodness” are derived from EOQ sensitivity analysis. Formulas are given for computing the minimax choice and the minimum expected value choice for the parameter estimates using both measures of estimator “goodness”. A numerical example is included.  相似文献   

14.
This study combines inspection and lot‐sizing decisions. The issue is whether to INSPECT another unit or PRODUCE a new lot. A unit produced is either conforming or defective. Demand need to be satisfied in full, by conforming units only. The production process may switch from a “good” state to a “bad” state, at constant rate. The proportion of conforming units in the good state is higher than in the bad state. The true state is unobservable and can only be inferred from the quality of units inspected. We thus update, after each inspection, the probability that the unit, next candidate for inspection, was produced while the production process was in the good state. That “good‐state‐probability” is the basis for our decision to INSPECT or PRODUCE. We prove that the optimal policy has a simple form: INSPECT only if the good‐state‐probability exceeds a control limit. We provide a methodology to calculate the optimal lot size and the expected costs associated with INSPECT and PRODUCE. Surprisingly, we find that the control limit, as a function of the demand (and other problem parameters) is not necessarily monotone. Also, counter to intuition, it is possible that the optimal action is PRODUCE, after revealing a conforming unit. © 2007 Wiley Periodicals, Inc. Naval Research Logistics, 2007  相似文献   

15.
This article examines a problem faced by a firm procuring a material input or good from a set of suppliers. The cost to procure the material from any given supplier is concave in the amount ordered from the supplier, up to a supplier‐specific capacity limit. This NP‐hard problem is further complicated by the observation that capacities are often uncertain in practice, due for instance to production shortages at the suppliers, or competition from other firms. We accommodate this uncertainty in a worst‐case (robust) fashion by modeling an adversarial entity (which we call the “follower”) with a limited procurement budget. The follower reduces supplier capacity to maximize the minimum cost required for our firm to procure its required goods. To guard against uncertainty, the firm can “protect” any supplier at a cost (e.g., by signing a contract with the supplier that guarantees supply availability, or investing in machine upgrades that guarantee the supplier's ability to produce goods at a desired level), ensuring that the anticipated capacity of that supplier will indeed be available. The problem we consider is thus a three‐stage game in which the firm first chooses which suppliers' capacities to protect, the follower acts next to reduce capacity from unprotected suppliers, and the firm then satisfies its demand using the remaining capacity. We formulate a three‐stage mixed‐integer program that is well‐suited to decomposition techniques and develop an effective cutting‐plane algorithm for its solution. The corresponding algorithmic approach solves a sequence of scaled and relaxed problem instances, which enables solving problems having much larger data values when compared to standard techniques. © 2013 Wiley Periodicals, Inc. Naval Research Logistics, 2013  相似文献   

16.
Clustering problems are often difficult to solve due to nonlinear cost functions and complicating constraints. Set partitioning formulations can help overcome these challenges, but at the cost of a very large number of variables. Therefore, techniques such as delayed column generation must be used to solve these large integer programs. The underlying pricing problem can suffer from the same challenges (non‐linear cost, complicating constraints) as the original problem, however, making a mathematical programming approach intractable. Motivated by a real‐world problem in printed circuit board (PCB) manufacturing, we develop a search‐based algorithm (Rank‐Cluster‐and‐Prune) as an alternative, present computational results for the PCB problem to demonstrate the tractability of our approach, and identify a broader class of clustering problems for which this approach can be used. © 2009 Wiley Periodicals, Inc. Naval Research Logistics 2009  相似文献   

17.
We study an infinite‐horizon, N‐stage, serial production/inventory system with two transportation modes between stages: regular shipping and expedited shipping. The optimal inventory policy for this system is a top–down echelon base‐stock policy, which can be computed through minimizing 2N nested convex functions recursively (Lawson and Porteus, Oper Res 48 (2000), 878–893). In this article, we first present some structural properties and comparative statics for the parameters of the optimal inventory policies, we then derive simple, newsvendor‐type lower and upper bounds for the optimal control parameters. These results are used to develop near optimal heuristic solutions for the echelon base‐stock policies. Numerical studies show that the heuristic performs well. © 2009 Wiley Periodicals, Inc. Naval Research Logistics, 2010  相似文献   

18.
It is well‐known that the efficient set of a multiobjective linear programming (MOLP) problem can be represented as a union of the maximal efficient faces of the feasible region. In this paper, we propose a method for finding all maximal efficient faces for an MOLP. The new method is based on a condition that all efficient vertices (short for the efficient extreme points and rays) for the MOLP have been found and it relies on the adjacency, affine independence and convexity results of efficient sets. The method uses a local top‐down search strategy to determine maximal efficient faces incident to every efficient vertex for finding maximal efficient faces of an MOLP problem. To our knowledge, the proposed method is the first top‐down search method that uses the adjacency property of the efficient set to find all maximal efficient faces. We discuss this and other advantages and disadvantages of the algorithm. We also discuss some computational experience we have had with our computer code for implementing the algorithm. This computational experience involved solving several MOLP problems with the code.  相似文献   

19.
Consider a system consisting of n separately maintained independent components where the components alternate between intervals in which they are “up” and in which they are “down”. When the ith component goes up [down] then, independent of the past, it remains up [down] for a random length of time, having distribution Fi[Gi], and then goes down [up]. We say that component i is failed at time t if it has been “down” at all time points s ?[t-A.t]: otherwise it is said to be working. Thus, a component is failed if it is down and has been down for the previous A time units. Assuming that all components initially start “up,” let T denote the first time they are all failed, at which point we say the system is failed. We obtain the moment-generating function of T when n = l, for general F and G, thus generalizing previous results which assumed that at least one of these distributions be exponential. In addition, we present a condition under which T is an NBU (new better than used) random variable. Finally we assume that all the up and down distributions Fi and Gi i = l,….n, are exponential, and we obtain an exact expression for E(T) for general n; in addition we obtain bounds for all higher moments of T by showing that T is NBU.  相似文献   

20.
In this article, we address a stochastic generalized assignment machine scheduling problem in which the processing times of jobs are assumed to be random variables. We develop a branch‐and‐price (B&P) approach for solving this problem wherein the pricing problem is separable with respect to each machine, and has the structure of a multidimensional knapsack problem. In addition, we explore two other extensions of this method—one that utilizes a dual‐stabilization technique and another that incorporates an advanced‐start procedure to obtain an initial feasible solution. We compare the performance of these methods with that of the branch‐and‐cut (B&C) method within CPLEX. Our results show that all B&P‐based approaches perform better than the B&C method, with the best performance obtained for the B&P procedure that includes both the extensions aforementioned. We also utilize a Monte Carlo method within the B&P scheme, which affords the use of a small subset of scenarios at a time to estimate the “true” optimal objective function value. Our experimental investigation reveals that this approach readily yields solutions lying within 5% of optimality, while providing more than a 10‐fold savings in CPU times in comparison with the best of the other proposed B&P procedures. © 2014 Wiley Periodicals, Inc. Naval Research Logistics 61: 131–143, 2014  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号