Results 1  10
of
906,771
FeatureRich PartofSpeech Tagging with a Cyclic Dependency Network
 IN PROCEEDINGS OF HLTNAACL
, 2003
"... We present a new partofspeech tagger that demonstrates the following ideas: (i) explicit use of both preceding and following tag contexts via a dependency network representation, (ii) broad use of lexical features, including jointly conditioning on multiple consecutive words, (iii) effective ..."
Abstract

Cited by 693 (23 self)
 Add to MetaCart
We present a new partofspeech tagger that demonstrates the following ideas: (i) explicit use of both preceding and following tag contexts via a dependency network representation, (ii) broad use of lexical features, including jointly conditioning on multiple consecutive words, (iii
SupportVector Networks
 Machine Learning
, 1995
"... The supportvector network is a new learning machine for twogroup classification problems. The machine conceptually implements the following idea: input vectors are nonlinearly mapped to a very highdimension feature space. In this feature space a linear decision surface is constructed. Special pr ..."
Abstract

Cited by 3703 (35 self)
 Add to MetaCart
The supportvector network is a new learning machine for twogroup classification problems. The machine conceptually implements the following idea: input vectors are nonlinearly mapped to a very highdimension feature space. In this feature space a linear decision surface is constructed. Special
Imagenet classification with deep convolutional neural networks.
 In Advances in the Neural Information Processing System,
, 2012
"... Abstract We trained a large, deep convolutional neural network to classify the 1.2 million highresolution images in the ImageNet LSVRC2010 contest into the 1000 different classes. On the test data, we achieved top1 and top5 error rates of 37.5% and 17.0% which is considerably better than the pr ..."
Abstract

Cited by 1010 (11 self)
 Add to MetaCart
Abstract We trained a large, deep convolutional neural network to classify the 1.2 million highresolution images in the ImageNet LSVRC2010 contest into the 1000 different classes. On the test data, we achieved top1 and top5 error rates of 37.5% and 17.0% which is considerably better than
Orthogonal matching pursuit: Recursive function approximation with applications to wavelet decomposition
 in Conference Record of The TwentySeventh Asilomar Conference on Signals, Systems and Computers
, 1993
"... In this paper we describe a recursive algorithm to compute representations of functions with respect to nonorthogonal and possibly overcomplete dictionaries of elementary building blocks e.g. aiEne (wa.velet) frames. We propoeea modification to the Matching Pursuit algorithm of Mallat and Zhang (199 ..."
Abstract

Cited by 637 (1 self)
 Add to MetaCart
recursively. where fk is the current approximation, and Rkf the current residual (error). Using initial values ofR0f = 1, fo = 0, and k = 1, the MP algorithm is comprised of the following steps,.,.41) Compute the innerproducts {(Rkf,z)}. (H) Find flki such that (III) Set, I(R*f,1:n 1+,)l asupl
Bayesian Analysis of Stochastic Volatility Models
, 1994
"... this article is to develop new methods for inference and prediction in a simple class of stochastic volatility models in which logarithm of conditional volatility follows an autoregressive (AR) times series model. Unlike the autoregressive conditional heteroscedasticity (ARCH) and gener alized ARCH ..."
Abstract

Cited by 601 (26 self)
 Add to MetaCart
this article is to develop new methods for inference and prediction in a simple class of stochastic volatility models in which logarithm of conditional volatility follows an autoregressive (AR) times series model. Unlike the autoregressive conditional heteroscedasticity (ARCH) and gener alized
The neural basis of human error processing: Reinforcement learning, dopamine, and the errorrelated negativity
 PSYCHOLOGICAL REVIEW 109:679–709
, 2002
"... The authors present a unified account of 2 neural systems concerned with the development and expression of adaptive behaviors: a mesencephalic dopamine system for reinforcement learning and a “generic ” errorprocessing system associated with the anterior cingulate cortex. The existence of the error ..."
Abstract

Cited by 430 (20 self)
 Add to MetaCart
The authors present a unified account of 2 neural systems concerned with the development and expression of adaptive behaviors: a mesencephalic dopamine system for reinforcement learning and a “generic ” errorprocessing system associated with the anterior cingulate cortex. The existence
Features of similarity.
 Psychological Review
, 1977
"... Similarity plays a fundamental role in theories of knowledge and behavior. It serves as an organizing principle by which individuals classify objects, form concepts, and make generalizations. Indeed, the concept of similarity is ubiquitous in psychological theory. It underlies the accounts of stimu ..."
Abstract

Cited by 1455 (2 self)
 Add to MetaCart
of stimulus and response generalization in learning, it is employed to explain errors in memory and pattern recognition, and it is central to the analysis of connotative meaning. Similarity or dissimilarity data appear in di¤erent forms: ratings of pairs, sorting of objects, communality between associations
Modeling and Forecasting Realized Volatility
, 2002
"... this paper is built. First, although raw returns are clearly leptokurtic, returns standardized by realized volatilities are approximately Gaussian. Second, although the distributions of realized volatilities are clearly rightskewed, the distributions of the logarithms of realized volatilities are a ..."
Abstract

Cited by 549 (50 self)
 Add to MetaCart
: a fractionallyintegrated Gaussian vector autoregression (VAR) . Importantly, our approach explicitly permits measurement errors in the realized volatilities. Comparing the resulting volatility forecasts to those obtained from currently popular daily volatility models and more complicated high
The irreducibility of the space of curves of given genus
 Publ. Math. IHES
, 1969
"... Fix an algebraically closed field k. Let Mg be the moduli space of curves of genus g over k. The main result of this note is that Mg is irreducible for every k. Of course, whether or not M s is irreducible depends only on the characteristic of k. When the characteristic s o, we can assume that k ~ ..."
Abstract

Cited by 506 (2 self)
 Add to MetaCart
approach however is closest to Severi's incomplete proof ([Se], Anhang F; the error is on pp. 344345 and seems to be quite basic) and follows a suggestion of Grothendieck for using the result in char. o to deduce the result in char. p. The basis of both Severi's and Grothendieck's ideas
Loopy belief propagation for approximate inference: An empirical study. In:
 Proceedings of Uncertainty in AI,
, 1999
"... Abstract Recently, researchers have demonstrated that "loopy belief propagation" the use of Pearl's polytree algorithm in a Bayesian network with loops can perform well in the context of errorcorrecting codes. The most dramatic instance of this is the near Shannonlimit performanc ..."
Abstract

Cited by 676 (15 self)
 Add to MetaCart
Abstract Recently, researchers have demonstrated that "loopy belief propagation" the use of Pearl's polytree algorithm in a Bayesian network with loops can perform well in the context of errorcorrecting codes. The most dramatic instance of this is the near Shannon
Results 1  10
of
906,771