全文获取类型
收费全文 | 189篇 |
免费 | 2篇 |
国内免费 | 3篇 |
专业分类
化学 | 3篇 |
力学 | 4篇 |
数学 | 177篇 |
物理学 | 10篇 |
出版年
2023年 | 4篇 |
2022年 | 3篇 |
2021年 | 6篇 |
2020年 | 4篇 |
2019年 | 4篇 |
2018年 | 8篇 |
2017年 | 1篇 |
2016年 | 6篇 |
2015年 | 1篇 |
2014年 | 7篇 |
2013年 | 6篇 |
2012年 | 6篇 |
2011年 | 10篇 |
2010年 | 13篇 |
2009年 | 21篇 |
2008年 | 14篇 |
2007年 | 8篇 |
2006年 | 10篇 |
2005年 | 10篇 |
2004年 | 8篇 |
2003年 | 8篇 |
2002年 | 5篇 |
2001年 | 6篇 |
2000年 | 4篇 |
1999年 | 3篇 |
1998年 | 2篇 |
1997年 | 4篇 |
1996年 | 2篇 |
1995年 | 1篇 |
1994年 | 4篇 |
1993年 | 1篇 |
1991年 | 2篇 |
1990年 | 1篇 |
1983年 | 1篇 |
排序方式: 共有194条查询结果,搜索用时 890 毫秒
21.
We present theory and algorithms for the equality constrained indefinite least squares problem, which requires minimization of an indefinite quadratic form subject to a linear equality constraint. A generalized hyperbolic QR factorization is introduced and used in the derivation of perturbation bounds and to construct a numerical method. An alternative method is obtained by employing a generalized QR factorization in combination with a Cholesky factorization. Rounding error analysis is given to show that both methods have satisfactory numerical stability properties and numerical experiments are given for illustration. This work builds on recent work on the unconstrained indefinite least squares problem by Chandrasekaran, Gu, and Sayed and by the present authors. 相似文献
22.
攀登伪蒙特卡罗积分法是由伪蒙特卡罗与蒙特卡罗方法混合而成的一种新方法,它体现了两者的优点.本文研究这种积分法在Sobolev空间和Korobov空间中的随机化误差.我们证明攀登(λ,t,m,s)-网积分法在这两个空间中的随机化误差的渐近阶为n^3/2[logn]^(s-1)/2。 相似文献
23.
Jens S. Frederiksen Kim S. Larsen John Noga Patchrawat Uthaisombut 《Journal of Algorithms in Cognition, Informatics and Logic》2003,48(2):407
When messages, which are to be sent point-to-point in a network, become available at irregular intervals, a decision must be made each time a new message becomes available as to whether it should be sent immediately or if it is better to wait for more messages and send them all together. Because of physical properties of the networks, a certain minimum amount of time must elapse in between the transmission of two packets. Thus, whereas waiting delays the transmission of the current data, sending immediately may delay the transmission of the next data to become available even more. We propose a new quality measure and derive optimal deterministic and randomized algorithms for this on-line problem. 相似文献
24.
Yannis?MarinakisEmail author Athanasios?Migdalas Panos?M.?Pardalos 《Computational Optimization and Applications》2005,32(3):231-257
In this paper, we present the application of a modified version of the well known Greedy Randomized Adaptive Search Procedure
(GRASP) to the TSP. The proposed GRASP algorithm has two phases: In the first phase the algorithm finds an initial solution
of the problem and in the second phase a local search procedure is utilized for the improvement of the initial solution. The
local search procedure employs two different local search strategies based on 2-opt and 3-opt methods. The algorithm was tested
on numerous benchmark problems from TSPLIB. The results were very satisfactory and for the majority of the instances the results
were equal to the best known solution. The algorithm is also compared to the algorithms presented and tested in the DIMACS Implementation Challenge that was organized by David Johnson. 相似文献
25.
Several improvements are made to an algorithm of Higham and Smith for computing the matrix cosine. The original algorithm
scales the matrix by a power of 2 to bring the ∞-norm to 1 or less, evaluates the [8/8] Padé approximant, then uses the double-angle
formula cos (2A)=2cos 2A−I to recover the cosine of the original matrix. The first improvement is to phrase truncation error bounds in terms of ‖A2‖1/2 instead of the (no smaller and potentially much larger quantity) ‖A‖. The second is to choose the degree of the Padé approximant to minimize the computational cost subject to achieving a desired
truncation error. A third improvement is to use an absolute, rather than relative, error criterion in the choice of Padé approximant;
this allows the use of higher degree approximants without worsening an a priori error bound. Our theory and experiments show
that each of these modifications brings a reduction in computational cost. Moreover, because the modifications tend to reduce
the number of double-angle steps they usually result in a more accurate computed cosine in floating point arithmetic. We also
derive an algorithm for computing both cos (A) and sin (A), by adapting the ideas developed for the cosine and intertwining the cosine and sine double angle recurrences.
AMS subject classification 65F30
Numerical Analysis Report 461, Manchester Centre for Computational Mathematics, February 2005.
Gareth I. Hargreaves: This work was supported by an Engineering and Physical Sciences Research Council Ph.D. Studentship.
Nicholas J. Higham: This work was supported by Engineering and Physical Sciences Research Council grant GR/T08739 and by a
Royal Society–Wolfson Research Merit Award. 相似文献
26.
Marta?EsoEmail author Soumyadip?Ghosh Jayant?Kalagnanam Laszlo?Ladanyi 《Journal of Heuristics》2005,11(2):147-173
Consider a marketplace operated by a buyer who wishes to procure large quantities of several heterogeneous products. Suppliers submit price curves for each of the commodities indicating the price charged as a function of the supplied quantity. The total amount paid to a supplier is the sum of the prices charged for the individual commodities. It is assumed that the submitted supply curves are piecewise linear as they often are in practice. The bid evaluation problem faced by the procurer is to determine how much of each commodity to buy from each of the suppliers so as to minimize the total purchase price. In addition to meeting the demand, the buyer may impose additional business requirements that restrict which contracts suppliers may be awarded. These requirements may result in interdependencies between the commodities which lead to suboptimal results if the commodities are traded in independent auctions rather than simultaneously. Even without the additional business constraints the bid evaluation problem is NP-hard. The main contribution of our study is a flexible column generation based heuristics that provides near-optimal solutions to the procurer’s bid evaluation problem. Our method scales very well due to the Branch-and-Price technology it is built on. We employ sophisticated rounding and local improvement heuristics to obtain quality solutions. We also developed a test data generator that produces realistic problems and allows control over the difficulty level of the problems using parameters. 相似文献
27.
《Random Structures and Algorithms》2018,53(2):289-307
A result of Spencer states that every collection of n sets over a universe of size n has a coloring of the ground set with of discrepancy . A geometric generalization of this result was given by Gluskin (see also Giannopoulos) who showed that every symmetric convex body with Gaussian measure at least , for a small , contains a point where a constant fraction of coordinates of y are in . This is often called a partial coloring result. While the proofs of both these results were inherently non‐algorithmic, recently Bansal (see also Lovett‐Meka) gave a polynomial time algorithm for Spencer's setting and Rothvoß gave a randomized polynomial time algorithm obtaining the same guarantee as the result of Gluskin and Giannopoulos. This paper contains several related results which combine techniques from convex geometry to analyze simple and efficient algorithms for discrepancy minimization. First, we prove another constructive version of the result of Gluskin and Giannopoulos, in which the coloring is attained via the optimization of a linear function. This implies a linear programming based algorithm for combinatorial discrepancy obtaining the same result as Spencer. Our second result suggests a new approach to obtain partial colorings, which is also valid for the non‐symmetric case. It shows that every (possibly non‐symmetric) convex body , with Gaussian measure at least , for a small , contains a point where a constant fraction of coordinates of y are in . Finally, we give a simple proof that shows that for any there exists a constant c > 0 such that given a body K with , a uniformly random x from is in cK with constant probability. This gives an algorithmic version of a special case of the result of Banaszczyk. 相似文献
28.
A conic integer program is an integer programming problem with conic constraints. Many problems in finance, engineering, statistical
learning, and probabilistic optimization are modeled using conic constraints. Here we study mixed-integer sets defined by
second-order conic constraints. We introduce general-purpose cuts for conic mixed-integer programming based on polyhedral
conic substructures of second-order conic sets. These cuts can be readily incorporated in branch-and-bound algorithms that
solve either second-order conic programming or linear programming relaxations of conic integer programs at the nodes of the
branch-and-bound tree. Central to our approach is a reformulation of the second-order conic constraints with polyhedral second-order
conic constraints in a higher dimensional space. In this representation the cuts we develop are linear, even though they are
nonlinear in the original space of variables. This feature leads to a computationally efficient implementation of nonlinear
cuts for conic mixed-integer programming. The reformulation also allows the use of polyhedral methods for conic integer programming.
We report computational results on solving unstructured second-order conic mixed-integer problems as well as mean–variance
capital budgeting problems and least-squares estimation problems with binary inputs. Our computational experiments show that
conic mixed-integer rounding cuts are very effective in reducing the integrality gap of continuous relaxations of conic mixed-integer
programs and, hence, improving their solvability.
This research has been supported, in part, by Grant # DMI0700203 from the National Science Foundation. 相似文献
29.
30.
Integrality gap of the hypergraphic relaxation of Steiner trees: A short proof of a 1.55 upper bound
Recently, Byrka, Grandoni, Rothvoß and Sanità gave a 1.39 approximation for the Steiner tree problem, using a hypergraph-based linear programming relaxation. They also upper-bounded its integrality gap by 1.55. We describe a shorter proof of the same integrality gap bound, by applying some of their techniques to a randomized loss-contracting algorithm. 相似文献