Results 1  10
of
2,084,031
Variance reduction in simulation of loss models
 Operations Research
, 1999
"... We propose a new estimator of steadystate blocking probabilities for simulations of stochastic loss models that can be much more efficient than the natural estimator (ratio of losses to arrivals). The proposed estimator is a convex combination of the natural estimator and an indirect estimator base ..."
Abstract

Cited by 10 (8 self)
 Add to MetaCart
of two component estimators are very different, the variance reduction from the optimal convex combination is about 1 − ρ 2, where ρ is the correlation between the component estimators. For loss models, the variances of the natural and indirect estimators are very different under both light and heavy
Discrete Choice Methods with Simulation
, 2002
"... This book describes the new generation of discrete choice methods, focusing on the many advances that are made possible by simulation. Researchers use these statistical methods to examine the choices that consumers, households, firms, and other agents make. Each of the major models is covered: logi ..."
Abstract

Cited by 1277 (20 self)
 Add to MetaCart
moments, and the method of simulated scores. Procedures for drawing from densities are described, including variance reduction techniques such as antithetics and Halton draws. Recent advances in Bayesian procedures are explored, including the use of the Metropolis– Hastings algorithm and its variant Gibbs
Simulating Physics with Computers
 SIAM Journal on Computing
, 1982
"... A digital computer is generally believed to be an efficient universal computing device; that is, it is believed able to simulate any physical computing device with an increase in computation time of at most a polynomial factor. This may not be true when quantum mechanics is taken into consideration. ..."
Abstract

Cited by 601 (1 self)
 Add to MetaCart
A digital computer is generally believed to be an efficient universal computing device; that is, it is believed able to simulate any physical computing device with an increase in computation time of at most a polynomial factor. This may not be true when quantum mechanics is taken into consideration
Modeling and simulation of genetic regulatory systems: A literature review
 JOURNAL OF COMPUTATIONAL BIOLOGY
, 2002
"... In order to understand the functioning of organisms on the molecular level, we need to know which genes are expressed, when and where in the organism, and to which extent. The regulation of gene expression is achieved through genetic regulatory systems structured by networks of interactions between ..."
Abstract

Cited by 729 (15 self)
 Add to MetaCart
for the modeling and simulation of genetic regulatory networks will be indispensable. This paper reviews formalisms that have been employed in mathematical biology and bioinformatics to describe genetic regulatory systems, in particular directed graphs, Bayesian networks, Boolean networks and their generalizations
A Fast Algorithm for Particle Simulations
, 1987
"... this paper to the case where the potential (or force) at a point is a sum of pairwise An algorithm is presented for the rapid evaluation of the potential and force fields in systems involving large numbers of particles interactions. More specifically, we consider potentials of whose interactions a ..."
Abstract

Cited by 1145 (19 self)
 Add to MetaCart
this paper to the case where the potential (or force) at a point is a sum of pairwise An algorithm is presented for the rapid evaluation of the potential and force fields in systems involving large numbers of particles interactions. More specifically, we consider potentials of whose interactions are Coulombic or gravitational in nature. For a the form system of N particles, an amount of work of the order O(N 2 ) has traditionally been required to evaluate all pairwise interactions, un F5F far 1 (F near 1F external ), less some approximation or truncation method is used. The algorithm of the present paper requires an amount of work proportional to N to evaluate all interactions to within roundoff error, making it where F near (when present) is a rapidly decaying potential con
The Macroscopic Behavior of the TCP Congestion Avoidance Algorithm
, 1997
"... In this paper, we analyze a performance model for the TCP Congestion Avoidance algorithm. The model predicts the bandwidth of a sustained TCP connection subjected to light to moderate packet losses, such as loss caused by network congestion. It assumes that TCP avoids retransmission timeouts and alw ..."
Abstract

Cited by 648 (18 self)
 Add to MetaCart
In this paper, we analyze a performance model for the TCP Congestion Avoidance algorithm. The model predicts the bandwidth of a sustained TCP connection subjected to light to moderate packet losses, such as loss caused by network congestion. It assumes that TCP avoids retransmission timeouts
Determining the Number of Factors in Approximate Factor Models
, 2000
"... In this paper we develop some statistical theory for factor models of large dimensions. The focus is the determination of the number of factors, which is an unresolved issue in the rapidly growing literature on multifactor models. We propose a panel Cp criterion and show that the number of factors c ..."
Abstract

Cited by 538 (29 self)
 Add to MetaCart
In this paper we develop some statistical theory for factor models of large dimensions. The focus is the determination of the number of factors, which is an unresolved issue in the rapidly growing literature on multifactor models. We propose a panel Cp criterion and show that the number of factors
Image denoising using a scale mixture of Gaussians in the wavelet domain
 IEEE TRANS IMAGE PROCESSING
, 2003
"... We describe a method for removing noise from digital images, based on a statistical model of the coefficients of an overcomplete multiscale oriented basis. Neighborhoods of coefficients at adjacent positions and scales are modeled as the product of two independent random variables: a Gaussian vecto ..."
Abstract

Cited by 514 (17 self)
 Add to MetaCart
vector and a hidden positive scalar multiplier. The latter modulates the local variance of the coefficients in the neighborhood, and is thus able to account for the empirically observed correlation between the coefficient amplitudes. Under this model, the Bayesian least squares estimate of each
Modeling Term Structures of Defaultable Bonds
, 1999
"... This article presents convenient reducedform models of the valuation of contingent claims subject to default risk, focusing on applications to the term structure of interest rates for corporate or sovereign bonds. Examples include the valuation of a creditspread option ..."
Abstract

Cited by 651 (34 self)
 Add to MetaCart
This article presents convenient reducedform models of the valuation of contingent claims subject to default risk, focusing on applications to the term structure of interest rates for corporate or sovereign bonds. Examples include the valuation of a creditspread option
A Compositional Approach to Performance Modelling
, 1996
"... Performance modelling is concerned with the capture and analysis of the dynamic behaviour of computer and communication systems. The size and complexity of many modern systems result in large, complex models. A compositional approach decomposes the system into subsystems that are smaller and more ea ..."
Abstract

Cited by 746 (102 self)
 Add to MetaCart
Performance modelling is concerned with the capture and analysis of the dynamic behaviour of computer and communication systems. The size and complexity of many modern systems result in large, complex models. A compositional approach decomposes the system into subsystems that are smaller and more
Results 1  10
of
2,084,031