Results 1  10
of
200,554
Spacetime codes for high data rate wireless communication: Performance criterion and code construction
 IEEE TRANS. INFORM. THEORY
, 1998
"... We consider the design of channel codes for improving the data rate and/or the reliability of communications over fading channels using multiple transmit antennas. Data is encoded by a channel code and the encoded data is split into n streams that are simultaneously transmitted using n transmit ant ..."
Abstract

Cited by 1782 (28 self)
 Add to MetaCart
We consider the design of channel codes for improving the data rate and/or the reliability of communications over fading channels using multiple transmit antennas. Data is encoded by a channel code and the encoded data is split into n streams that are simultaneously transmitted using n transmit
Marching cubes: A high resolution 3D surface construction algorithm
 COMPUTER GRAPHICS
, 1987
"... We present a new algorithm, called marching cubes, that creates triangle models of constant density surfaces from 3D medical data. Using a divideandconquer approach to generate interslice connectivity, we create a case table that defines triangle topology. The algorithm processes the 3D medical d ..."
Abstract

Cited by 2696 (4 self)
 Add to MetaCart
slice connectivity, surface data, and gradient information present in the original 3D data. Results from computed tomography (CT), magnetic resonance (MR), and singlephoton emission computed tomography (SPECT) illustrate the quality and functionality of marching cubes. We also discuss improvements that decrease
A LinearTime Heuristic for Improving Network Partitions
, 1982
"... An iterative mincut heuristic for partitioning networks is presented whose worst case computation time, per pass, grows linearly with the size of the network. In practice, only a very small number of passes are typically needed, leading to a fast approximation algorithm for mincut partitioning. To d ..."
Abstract

Cited by 524 (0 self)
 Add to MetaCart
An iterative mincut heuristic for partitioning networks is presented whose worst case computation time, per pass, grows linearly with the size of the network. In practice, only a very small number of passes are typically needed, leading to a fast approximation algorithm for mincut partitioning
Trade Liberalization, Exit, and Productivity Improvements: Evidence from Chilean Plants
 Review of Economic Studies
, 2002
"... This paper empirically investigates the effects of liberalized trade on plant productivity in the case of Chile. Chile presents an interesting setting to study this relationship since it underwent a massive trade liberalization that significantly exposed its plants to competition from abroad during ..."
Abstract

Cited by 555 (16 self)
 Add to MetaCart
in the estimates of the input coefficients required to construct a productivity measure. I explicitly incorporate plant exit in the estimation to correct for the selection problem induced by liquidated plants. These methodological aspects are important in obtaining a reliable plantlevel productivity measure based
A theory of timed automata
, 1999
"... Model checking is emerging as a practical tool for automated debugging of complex reactive systems such as embedded controllers and network protocols (see [23] for a survey). Traditional techniques for model checking do not admit an explicit modeling of time, and are thus, unsuitable for analysis of ..."
Abstract

Cited by 2651 (32 self)
 Add to MetaCart
using finitely many realvalued clock variables. Automated analysis of timed automata relies on the construction of a finite quotient of the infinite space of clock valuations. Over the years, the formalism has been extensively studied leading to many results establishing connections to circuits
An experimental comparison of three methods for constructing ensembles of decision trees
 Bagging, boosting, and randomization. Machine Learning
, 2000
"... Abstract. Bagging and boosting are methods that generate a diverse ensemble of classifiers by manipulating the training data given to a “base ” learning algorithm. Breiman has pointed out that they rely for their effectiveness on the instability of the base learning algorithm. An alternative approac ..."
Abstract

Cited by 610 (6 self)
 Add to MetaCart
approach to generating an ensemble is to randomize the internal decisions made by the base algorithm. This general approach has been studied previously by Ali and Pazzani and by Dietterich and Kong. This paper compares the effectiveness of randomization, bagging, and boosting for improving the performance
Lag length selection and the construction of unit root tests with good size and power
 Econometrica
, 2001
"... It is widely known that when there are errors with a movingaverage root close to −1, a high order augmented autoregression is necessary for unit root tests to have good size, but that information criteria such as the AIC and the BIC tend to select a truncation lag (k) that is very small. We conside ..."
Abstract

Cited by 558 (14 self)
 Add to MetaCart
framework in which the movingaverage root is local to −1 to document how the MIC performs better in selecting appropriate values of k. In montecarlo experiments, the MIC is found to yield huge size improvements to the DF GLS and the feasible point optimal PT test developed in Elliott, Rothenberg and Stock
Constructing Free Energy Approximations and Generalized Belief Propagation Algorithms
 IEEE Transactions on Information Theory
, 2005
"... Important inference problems in statistical physics, computer vision, errorcorrecting coding theory, and artificial intelligence can all be reformulated as the computation of marginal probabilities on factor graphs. The belief propagation (BP) algorithm is an efficient way to solve these problems t ..."
Abstract

Cited by 585 (13 self)
 Add to MetaCart
that is exact when the factor graph is a tree, but only approximate when the factor graph has cycles. We show that BP fixed points correspond to the stationary points of the Bethe approximation of the free energy for a factor graph. We explain how to obtain regionbased free energy approximations that improve
Mining Sequential Patterns: Generalizations and Performance Improvements
 RESEARCH REPORT RJ 9994, IBM ALMADEN RESEARCH
, 1995
"... The problem of mining sequential patterns was recently introduced in [3]. We are given a database of sequences, where each sequence is a list of transactions ordered by transactiontime, and each transaction is a set of items. The problem is to discover all sequential patterns with a userspecified ..."
Abstract

Cited by 759 (5 self)
 Add to MetaCart
The problem of mining sequential patterns was recently introduced in [3]. We are given a database of sequences, where each sequence is a list of transactions ordered by transactiontime, and each transaction is a set of items. The problem is to discover all sequential patterns with a user
Fibonacci Heaps and Their Uses in Improved Network optimization algorithms
, 1987
"... In this paper we develop a new data structure for implementing heaps (priority queues). Our structure, Fibonacci heaps (abbreviated Fheaps), extends the binomial queues proposed by Vuillemin and studied further by Brown. Fheaps support arbitrary deletion from an nitem heap in qlogn) amortized tim ..."
Abstract

Cited by 739 (18 self)
 Add to MetaCart
time and all other standard heap operations in o ( 1) amortized time. Using Fheaps we are able to obtain improved running times for several network optimization algorithms. In particular, we obtain the following worstcase bounds, where n is the number of vertices and m the number of edges
Results 1  10
of
200,554