Results 1  10
of
41,947
Actions as spacetime shapes
 IN ICCV
, 2005
"... Human action in video sequences can be seen as silhouettes of a moving torso and protruding limbs undergoing articulated motion. We regard human actions as threedimensional shapes induced by the silhouettes in the spacetime volume. We adopt a recent approach [14] for analyzing 2D shapes and genera ..."
Abstract

Cited by 651 (4 self)
 Add to MetaCart
Human action in video sequences can be seen as silhouettes of a moving torso and protruding limbs undergoing articulated motion. We regard human actions as threedimensional shapes induced by the silhouettes in the spacetime volume. We adopt a recent approach [14] for analyzing 2D shapes
The space complexity of approximating the frequency moments
 JOURNAL OF COMPUTER AND SYSTEM SCIENCES
, 1996
"... The frequency moments of a sequence containing mi elements of type i, for 1 ≤ i ≤ n, are the numbers Fk = �n i=1 mki. We consider the space complexity of randomized algorithms that approximate the numbers Fk, when the elements of the sequence are given one by one and cannot be stored. Surprisingly, ..."
Abstract

Cited by 845 (12 self)
 Add to MetaCart
, it turns out that the numbers F0, F1 and F2 can be approximated in logarithmic space, whereas the approximation of Fk for k ≥ 6 requires nΩ(1) space. Applications to data bases are mentioned as well.
Reducing the Space Requirement of Suffix Trees
 Software – Practice and Experience
, 1999
"... We show that suffix trees store various kinds of redundant information. We exploit these redundancies to obtain more space efficient representations. The most space efficient of our representations requires 20 bytes per input character in the worst case, and 10.1 bytes per input character on average ..."
Abstract

Cited by 145 (12 self)
 Add to MetaCart
We show that suffix trees store various kinds of redundant information. We exploit these redundancies to obtain more space efficient representations. The most space efficient of our representations requires 20 bytes per input character in the worst case, and 10.1 bytes per input character
Space/Time Tradeoffs in Hash Coding with Allowable Errors
 Communications of the ACM
, 1970
"... this paper tradeoffs among certain computational factors in hash coding are analyzed. The paradigm problem considered is that of testing a series of messages onebyone for membership in a given set of messages. Two new hash coding methods are examined and compared with a particular conventional h ..."
Abstract

Cited by 2097 (0 self)
 Add to MetaCart
hashcoding method. The computational factors considered are the size of the hash area (space), the time required to identify a message as a nonmember of the given set (reject time), and an allowable error frequency
Probabilistic Roadmaps for Path Planning in HighDimensional Configuration Spaces
 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION
, 1996
"... A new motion planning method for robots in static workspaces is presented. This method proceeds in two phases: a learning phase and a query phase. In the learning phase, a probabilistic roadmap is constructed and stored as a graph whose nodes correspond to collisionfree configurations and whose edg ..."
Abstract

Cited by 1277 (120 self)
 Add to MetaCart
nodes. The method is general and easy to implement. It can be applied to virtually any type of holonomic robot. It requires selecting certain parameters (e.g., the duration of the learning phase) whose values depend on the scene, that is the robot and its workspace. But these values turn out
A Faster Algorithm for Betweenness Centrality
 Journal of Mathematical Sociology
, 2001
"... The betweenness centrality index is essential in the analysis of social networks, but costly to compute. Currently, the fastest known algorithms require #(n ) time and #(n ) space, where n is the number of actors in the network. ..."
Abstract

Cited by 554 (5 self)
 Add to MetaCart
The betweenness centrality index is essential in the analysis of social networks, but costly to compute. Currently, the fastest known algorithms require #(n ) time and #(n ) space, where n is the number of actors in the network.
Maximum Likelihood Linear Transformations for HMMBased Speech Recognition
 COMPUTER SPEECH AND LANGUAGE
, 1998
"... This paper examines the application of linear transformations for speaker and environmental adaptation in an HMMbased speech recognition system. In particular, transformations that are trained in a maximum likelihood sense on adaptation data are investigated. Other than in the form of a simple bias ..."
Abstract

Cited by 570 (68 self)
 Add to MetaCart
) constrained, which requires the variance transform to have the same form as the mean transform (sometimes referred to as featurespace transforms). Reestimation formulae for all appropriate cases of transform are given. This includes a new and efficient "full" variance transform and the extension
Projection Pursuit Regression
 Journal of the American Statistical Association
, 1981
"... A new method for nonparametric multiple regression is presented. The procedure models the regression surface as a sum of general smooth functions of linear combinations of the predictor variables in an iterative manner. It is more general than standard stepwise and stagewise regression procedures, ..."
Abstract

Cited by 550 (6 self)
 Add to MetaCart
, does not require the definition of a metric in the predictor space, and lends itself to graphical interpretation.
Depth first search and linear graph algorithms
 SIAM JOURNAL ON COMPUTING
, 1972
"... The value of depthfirst search or "backtracking" as a technique for solving problems is illustrated by two examples. An improved version of an algorithm for finding the strongly connected components of a directed graph and ar algorithm for finding the biconnected components of an undirect ..."
Abstract

Cited by 1406 (19 self)
 Add to MetaCart
of an undirect graph are presented. The space and time requirements of both algorithms are bounded by k 1V + k2E d k for some constants kl, k2, and k a, where Vis the number of vertices and E is the number of edges of the graph being examined.
Computing semantic relatedness using Wikipediabased explicit semantic analysis
 In Proceedings of the 20th International Joint Conference on Artificial Intelligence
, 2007
"... Computing semantic relatedness of natural language texts requires access to vast amounts of commonsense and domainspecific world knowledge. We propose Explicit Semantic Analysis (ESA), a novel method that represents the meaning of texts in a highdimensional space of concepts derived from Wikipedi ..."
Abstract

Cited by 562 (9 self)
 Add to MetaCart
Computing semantic relatedness of natural language texts requires access to vast amounts of commonsense and domainspecific world knowledge. We propose Explicit Semantic Analysis (ESA), a novel method that represents the meaning of texts in a highdimensional space of concepts derived from
Results 1  10
of
41,947