首页 | 本学科首页   官方微博 | 高级检索  
文章检索
  按 检索   检索词:      
出版年份:   被引次数:   他引次数: 提示:输入*表示无穷大
  收费全文   130192篇
  免费   2512篇
  国内免费   449篇
化学   58796篇
晶体学   1032篇
力学   8294篇
数学   39670篇
物理学   25361篇
  2023年   420篇
  2022年   396篇
  2021年   674篇
  2020年   809篇
  2019年   780篇
  2018年   11113篇
  2017年   10764篇
  2016年   7587篇
  2015年   2395篇
  2014年   1888篇
  2013年   3580篇
  2012年   7427篇
  2011年   13921篇
  2010年   7752篇
  2009年   7773篇
  2008年   9688篇
  2007年   11729篇
  2006年   3237篇
  2005年   4138篇
  2004年   3974篇
  2003年   3916篇
  2002年   2770篇
  2001年   1248篇
  2000年   1113篇
  1999年   758篇
  1998年   666篇
  1997年   590篇
  1996年   755篇
  1995年   469篇
  1994年   506篇
  1993年   503篇
  1992年   480篇
  1991年   431篇
  1990年   416篇
  1989年   413篇
  1988年   347篇
  1987年   346篇
  1986年   305篇
  1985年   471篇
  1984年   454篇
  1983年   311篇
  1982年   411篇
  1981年   380篇
  1980年   355篇
  1979年   343篇
  1978年   314篇
  1976年   293篇
  1975年   269篇
  1974年   271篇
  1973年   276篇
排序方式: 共有10000条查询结果,搜索用时 546 毫秒
141.
142.
143.
144.
145.
What strategy should a football (soccer, in American parlance) club adopt when deciding whether to sack its manager? This paper introduces a simple model assuming that a club's objective is to maximize the number of league points that it scores per season. The club's strategy consists of three choices: the length of the honeymoon period during which it will not consider sacking a new manager, the level of the performance trapdoor below which the manager get the sack, and the weight that it will give to more recent games compared to earlier ones. Some data from the last six seasons of the English Premiership are used to calibrate the model. At this early stage of the research, the best strategy appears to have only a short honeymoon period of eight games (much less than the actual shortest period of 12 games), to set the trapdoor at 0.74 points per game, and to put 47% of the weight on the last five games. A club adopting this strategy would obtain on average 56.8 points per season, compared to a Premiership average of 51.8 points.  相似文献   
146.
147.
148.
We consider estimation of loss for generalized Bayes or pseudo-Bayes estimators of a multivariate normal mean vector, θ. In 3 and higher dimensions, the MLEX is UMVUE and minimax but is inadmissible. It is dominated by the James-Stein estimator and by many others. Johnstone (1988, On inadmissibility of some unbiased estimates of loss,Statistical Decision Theory and Related Topics, IV (eds. S. S. Gupta and J. O. Berger), Vol. 1, 361–379, Springer, New York) considered the estimation of loss for the usual estimatorX and the James-Stein estimator. He found improvements over the Stein unbiased estimator of risk. In this paper, for a generalized Bayes point estimator of θ, we compare generalized Bayes estimators to unbiased estimators of loss. We find, somewhat surprisingly, that the unbiased estimator often dominates the corresponding generalized Bayes estimator of loss for priors which give minimax estimators in the original point estimation problem. In particular, we give a class of priors for which the generalized Bayes estimator of θ is admissible and minimax but for which the unbiased estimator of loss dominates the generalized Bayes estimator of loss. We also give a general inadmissibility result for a generalized Bayes estimator of loss. Research supported by NSF Grant DMS-97-04524.  相似文献   
149.
On effectiveness of wiretap programs in mapping social networks   总被引:1,自引:0,他引:1  
Snowball sampling methods are known to be a biased toward highly connected actors and consequently produce core-periphery networks when these may not necessarily be present. This leads to a biased perception of the underlying network which can have negative policy consequences, as in the identification of terrorist networks. When snowball sampling is used, the potential overload of the information collection system is a distinct problem due to the exponential growth of the number of suspects to be monitored. In this paper, we focus on evaluating the effectiveness of a wiretapping program in terms of its ability to map the rapidly evolving networks within a covert organization. By running a series of simulation-based experiments, we are able to evaluate a broad spectrum of information gathering regimes based on a consistent set of criteria. We conclude by proposing a set of information gathering programs that achieve higher effectiveness then snowball sampling, and at a lower cost. Maksim Tsvetovat is an Assistant Professor at the Center for Social Complexity and department of Public and International Affairs at George Mason University, Fairfax, VA. He received his Ph.D. from the Computation, Organizations and Society program in the School of Computer Science, Carnegie Mellon University. His dissertation was centered on use of artificial intelligence techniques such as planning and semantic reasoning as a means of studying behavior and evolution of complex social networks, such as these of terrorist organizations. He received a Master of Science degree from University of Minnesota with a specialization in Artificial Intelligence and design of Multi-Agent Systems, and has also extensively studied organization theory and social science research methods. His research is centered on building high-fidelity simulations of social and organizational systems using concepts from distributed artificial intelligence and multi-agent systems. Other projects focus on social network analysis for mapping of internal corporate networks or study of covert and terrorist orgnaizations. Maksim’s vita and publications can be found on Kathleen M. Carley is a professor in the School of Computer Science at Carnegie Mellon University and the director of the center for Compuational Analysis of Social and Organizational Systems (CASOS) which has over 25 members, both students and research staff. Her research combines cognitive science, social networks and computer science to address complex social and organizational problems. Her specific research areas are dynamic network analysis, computational social and organization theory, adaptation and evolution, text mining, and the impact of telecommunication technologies and policy on communication, information diffusion, disease contagion and response within and among groups particularly in disaster or crisis situations. She and her lab have developed infrastructure tools for analyzing large scale dynamic networks and various multi-agent simulation systems. The infrastructure tools include ORA, a statistical toolkit for analyzing and visualizing multi-dimensional networks. ORA results are organized into reports that meet various needs such as the management report, the mental model report, and the intelligence report. Another tool is AutoMap, a text-mining systems for extracting semantic networks from texts and then cross-classifying them using an organizational ontology into the underlying social, knowledge, resource and task networks. Her simulation models meld multi-agent technology with network dynamics and empirical data. Three of the large-scale multi-agent network models she and the CASOS group have developed in the counter-terrorism area are: BioWar a city-scale dynamic-network agent-based model for understanding the spread of disease and illness due to natural epidemics, chemical spills, and weaponized biological attacks; DyNet a model of the change in covert networks, naturally and in response to attacks, under varying levels of information uncertainty; and RTE a model for examining state failure and the escalation of conflict at the city, state, nation, and international as changes occur within and among red, blue, and green forces. She is the founding co-editor with Al. Wallace of the journal Computational Organization Theory and has co-edited several books and written over 100 articles in the computational organizations and dynamic network area. Her publications can be found at: http://www.casos.cs.cmu.edu/bios/carley/publications.php  相似文献   
150.
A tomographic reconstruction method based on Monte Carlo random searching guided by the information contained in the projections of radiographed objects is presented. In order to solve the optimization problem, a multiscale algorithm is proposed to reduce computation. The reconstruction is performed in a coarse-to-fine multigrid scale that initializes each resolution level with the reconstruction of the previous coarser level, which substantially improves the performance. The method was applied to a real case reconstructing the internal structure of a small metallic object with internal components, showing excellent results.  相似文献   
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号