Results 11  20
of
713,503
Some Evidence on the Importance of Sticky Prices
 JOURNAL OF POLITICAL ECONOMY
, 2004
"... We examine the frequency of price changes for 350 categories of goods and services covering about 70 % of consumer spending, based on unpublished data from the BLS for 1995 to 1997. Compared with previous studies we find much more frequent price changes, with half of goods' prices lasting less ..."
Abstract

Cited by 734 (15 self)
 Add to MetaCart
We examine the frequency of price changes for 350 categories of goods and services covering about 70 % of consumer spending, based on unpublished data from the BLS for 1995 to 1997. Compared with previous studies we find much more frequent price changes, with half of goods' prices lasting less
Realtime human pose recognition in parts from single depth images
 In In CVPR, 2011. 3
"... We propose a new method to quickly and accurately predict 3D positions of body joints from a single depth image, using no temporal information. We take an object recognition approach, designing an intermediate body parts representation that maps the difficult pose estimation problem into a simpler p ..."
Abstract

Cited by 550 (19 self)
 Add to MetaCart
perpixel classification problem. Our large and highly varied training dataset allows the classifier to estimate body parts invariant to pose, body shape, clothing, etc. Finally we generate confidencescored 3D proposals of several body joints by reprojecting the classification result and finding
Network Coding for Large Scale Content Distribution
"... We propose a new scheme for content distribution of large files that is based on network coding. With network coding, each node of the distribution network is able to generate and transmit encoded blocks of information. The randomization introduced by the coding process eases the scheduling of bloc ..."
Abstract

Cited by 497 (6 self)
 Add to MetaCart
We propose a new scheme for content distribution of large files that is based on network coding. With network coding, each node of the distribution network is able to generate and transmit encoded blocks of information. The randomization introduced by the coding process eases the scheduling
Basic objects in natural categories
 COGNITIVE PSYCHOLOGY
, 1976
"... Categorizations which humans make of the concrete world are not arbitrary but highly determined. In taxonomies of concrete objects, there is one level of abstraction at which the most basic category cuts are made. Basic categories are those which carry the most information, possess the highest categ ..."
Abstract

Cited by 856 (1 self)
 Add to MetaCart
Categorizations which humans make of the concrete world are not arbitrary but highly determined. In taxonomies of concrete objects, there is one level of abstraction at which the most basic category cuts are made. Basic categories are those which carry the most information, possess the highest
Probabilistic Latent Semantic Indexing
, 1999
"... Probabilistic Latent Semantic Indexing is a novel approach to automated document indexing which is based on a statistical latent class model for factor analysis of count data. Fitted from a training corpus of text documents by a generalization of the Expectation Maximization algorithm, the utilized ..."
Abstract

Cited by 1207 (11 self)
 Add to MetaCart
Probabilistic Latent Semantic Indexing is a novel approach to automated document indexing which is based on a statistical latent class model for factor analysis of count data. Fitted from a training corpus of text documents by a generalization of the Expectation Maximization algorithm, the utilized
Boosting a Weak Learning Algorithm By Majority
, 1995
"... We present an algorithm for improving the accuracy of algorithms for learning binary concepts. The improvement is achieved by combining a large number of hypotheses, each of which is generated by training the given learning algorithm on a different set of examples. Our algorithm is based on ideas pr ..."
Abstract

Cited by 516 (15 self)
 Add to MetaCart
We present an algorithm for improving the accuracy of algorithms for learning binary concepts. The improvement is achieved by combining a large number of hypotheses, each of which is generated by training the given learning algorithm on a different set of examples. Our algorithm is based on ideas
Estimating the Support of a HighDimensional Distribution
, 1999
"... Suppose you are given some dataset drawn from an underlying probability distribution P and you want to estimate a "simple" subset S of input space such that the probability that a test point drawn from P lies outside of S is bounded by some a priori specified between 0 and 1. We propo ..."
Abstract

Cited by 766 (29 self)
 Add to MetaCart
propose a method to approach this problem by trying to estimate a function f which is positive on S and negative on the complement. The functional form of f is given by a kernel expansion in terms of a potentially small subset of the training data; it is regularized by controlling the length
Eliciting selfexplanations improves understanding
 Cognitive Science
, 1994
"... Learning involves the integration of new information into existing knowledge. Generoting explanations to oneself (selfexplaining) facilitates that integration process. Previously, selfexplanation has been shown to improve the acquisition of problemsolving skills when studying workedout examples. ..."
Abstract

Cited by 556 (22 self)
 Add to MetaCart
. This study extends that finding, showing that selfexplanation can also be facilitative when it is explicitly promoted, in the context of learning declarative knowledge from an expository text. Without any extensive training, 14 eighthgrade students were merely asked to selfexplain after reading each line
SelfSimilarity Through HighVariability: Statistical Analysis of Ethernet LAN Traffic at the Source Level
 IEEE/ACM TRANSACTIONS ON NETWORKING
, 1997
"... A number of recent empirical studies of traffic measurements from a variety of working packet networks have convincingly demonstrated that actual network traffic is selfsimilar or longrange dependent in nature (i.e., bursty over a wide range of time scales)  in sharp contrast to commonly made tr ..."
Abstract

Cited by 740 (24 self)
 Add to MetaCart
traffic modeling assumptions. In this paper, we provide a plausible physical explanation for the occurrence of selfsimilarity in LAN traffic. Our explanation is based on new convergence results for processes that exhibit high variability (i.e., infinite variance) and is supported by detailed statistical
An Efficient Boosting Algorithm for Combining Preferences
, 1999
"... The problem of combining preferences arises in several applications, such as combining the results of different search engines. This work describes an efficient algorithm for combining multiple preferences. We first give a formal framework for the problem. We then describe and analyze a new boosting ..."
Abstract

Cited by 707 (18 self)
 Add to MetaCart
The problem of combining preferences arises in several applications, such as combining the results of different search engines. This work describes an efficient algorithm for combining multiple preferences. We first give a formal framework for the problem. We then describe and analyze a new boosting algorithm for combining preferences called RankBoost. We also describe an efficient implementation of the algorithm for certain natural cases. We discuss two experiments we carried out to assess the performance of RankBoost. In the first experiment, we used the algorithm to combine different WWW search strategies, each of which is a query expansion for a given domain. For this task, we compare the performance of RankBoost to the individual search strategies. The second experiment is a collaborativefiltering task for making movie recommendations. Here, we present results comparing RankBoost to nearestneighbor and regression algorithms.
Results 11  20
of
713,503