全文获取类型
收费全文 | 796篇 |
免费 | 304篇 |
国内免费 | 75篇 |
出版年
2024年 | 1篇 |
2023年 | 9篇 |
2022年 | 8篇 |
2021年 | 23篇 |
2020年 | 25篇 |
2019年 | 19篇 |
2018年 | 15篇 |
2017年 | 36篇 |
2016年 | 54篇 |
2015年 | 31篇 |
2014年 | 57篇 |
2013年 | 66篇 |
2012年 | 64篇 |
2011年 | 67篇 |
2010年 | 34篇 |
2009年 | 72篇 |
2008年 | 49篇 |
2007年 | 53篇 |
2006年 | 74篇 |
2005年 | 62篇 |
2004年 | 56篇 |
2003年 | 46篇 |
2002年 | 37篇 |
2001年 | 35篇 |
2000年 | 33篇 |
1999年 | 22篇 |
1998年 | 27篇 |
1997年 | 13篇 |
1996年 | 17篇 |
1995年 | 11篇 |
1994年 | 16篇 |
1993年 | 5篇 |
1992年 | 18篇 |
1991年 | 6篇 |
1990年 | 9篇 |
1989年 | 3篇 |
1988年 | 2篇 |
排序方式: 共有1175条查询结果,搜索用时 15 毫秒
901.
We consider a problem of scheduling jobs on m parallel machines. The machines are dedicated, i.e., for each job the processing machine is known in advance. We mainly concentrate on the model in which at any time there is one unit of an additional resource. Any job may be assigned the resource and this reduces its processing time. A job that is given the resource uses it at each time of its processing. No two jobs are allowed to use the resource simultaneously. The objective is to minimize the makespan. We prove that the two‐machine problem is NP‐hard in the ordinary sense, describe a pseudopolynomial dynamic programming algorithm and convert it into an FPTAS. For the problem with an arbitrary number of machines we present an algorithm with a worst‐case ratio close to 3/2, and close to 3, if a job can be given several units of the resource. For the problem with a fixed number of machines we give a PTAS. Virtually all algorithms rely on a certain variant of the linear knapsack problem (maximization, minimization, multiple‐choice, bicriteria). © 2008 Wiley Periodicals, Inc. Naval Research Logistics, 2008 相似文献
902.
We consider a manufacturer, served by a single supplier, who has to quote due dates to arriving customers in a make‐to‐order production environment. The manufacturer is penalized for long lead times and for missing due dates. To meet due dates, the manufacturer has to obtain components from a supplier. We model this manufacturer and supplier as a two‐machine flow shop, consider several variations of this problem, and design effective due‐date quotation and scheduling algorithms for centralized and decentralized versions of the model. We perform extensive computational testing to assess the effectiveness of our algorithms and to compare the centralized and decentralized models to quantify the value of centralized control in a make‐to‐order supply chain. Since complete information exchange and centralized control is not always practical or cost‐effective, we explore the value of partial information exchange for this system. © 2008 Wiley Periodicals, Inc. Naval Research Logistics, 2008 相似文献
903.
Data envelopment analysis (DEA) is a method for measuring the efficiency of peer decision making units (DMUs). This tool has been utilized by a number of authors to examine two‐stage processes, where all the outputs from the first stage are the only inputs to the second stage. The current article examines and extends these models using game theory concepts. The resulting models are linear, and imply an efficiency decomposition where the overall efficiency of the two‐stage process is a product of the efficiencies of the two individual stages. When there is only one intermediate measure connecting the two stages, both the noncooperative and centralized models yield the same results as applying the standard DEA model to the two stages separately. As a result, the efficiency decomposition is unique. While the noncooperative approach yields a unique efficiency decomposition under multiple intermediate measures, the centralized approach is likely to yield multiple decompositions. Models are developed to test whether the efficiency decomposition arising from the centralized approach is unique. The relations among the noncooperative, centralized, and standard DEA approaches are investigated. Two real world data sets and a randomly generated data set are used to demonstrate the models and verify our findings. © 2008 Wiley Periodicals, Inc. Naval Research Logistics, 2008 相似文献
904.
为了满足JPEG2000遥感图像实时压缩设备的实时性要求,提出了一种基于并行机制的“零时间”数据搬移策略。通过对EBC算法结构进行拆分,并采用EDMA方式进行数据搬移,使得算法运算与数据搬移过程完全并行,搬移时间等效为零。实验结果表明,采用新的数据搬移策略后,编码效率可以提升到100%,EBC算法实现的总时间可缩短46%以上。该搬移机制已成功应用于JPEG2000星载遥感图像实时压缩系统中,系统的实时性达到了设计要求。 相似文献
905.
906.
907.
齿轮是机械传动中重要的传力构件,由于测试手段和试验方法的限制,齿轮设计的时候多采用静强度设计理论,无法准确地反应实际情况下的动态特性,导致齿轮的实际寿命和设计寿命有较大的差距。基于多体动力学仿真分析方法建立履带式车辆传动系统虚拟行驶试验平台,获得齿轮在不同工况下的动载荷谱。基于疲劳分析软件MSC.Fatigue获得了齿轮的疲劳寿命,进而改变齿轮的结构参数分析不同结构齿轮的疲劳寿命,得到其疲劳寿命随不同结构参数变化的规律,进行齿轮疲劳寿命的灵敏度分析,为齿轮的结构优化作了一定的探索。 相似文献
908.
909.
910.
针对海量电磁数据中雷达信号难以进行快速准确分选的问题,提出一种新的聚类分选方法,即改进k-means算法的Map Reduce并行化实现方法。通过引入初始聚类中心个数k1、最大聚类中心个数kmax和距离门限rt3个参数,克服了k-means算法需要事先确定k值和易受孤立点影响的局限;基于Hadoop平台实现了对改进k-means算法的Map Reduce并行化,克服了k-means算法串行实现时间复杂度高的局限。最后,实验表明改进k-means算法取得了更高的分选准确率,Map Reduce并行化后具有良好的加速比和扩展性,能够很好地对海量电磁数据中雷达信号进行高效分选。 相似文献