Results 1 -
3 of
3
Learning where to sample in structured prediction
- In Artificial Intelligence and Statistics (AISTATS
, 2015
"... Abstract In structured prediction, most inference algorithms allocate a homogeneous amount of computation to all parts of the output, which can be wasteful when different parts vary widely in terms of difficulty. In this paper, we propose a heterogeneous approach that dynamically allocates computat ..."
Abstract
-
Cited by 1 (1 self)
- Add to MetaCart
Abstract In structured prediction, most inference algorithms allocate a homogeneous amount of computation to all parts of the output, which can be wasteful when different parts vary widely in terms of difficulty. In this paper, we propose a heterogeneous approach that dynamically allocates computation to the different parts. Given a pre-trained model, we tune its inference algorithm (a sampler) to increase test-time throughput. The inference algorithm is parametrized by a meta-model and trained via reinforcement learning, where actions correspond to sampling candidate parts of the output, and rewards are loglikelihood improvements. The meta-model is based on a set of domain-general metafeatures capturing the progress of the sampler. We test our approach on five datasets and show that it attains the same accuracy as Gibbs sampling but is 2 to 5 times faster.
Carnegie Mellon
"... Recently, several Web-scale knowledge harvesting systems have been built, each of which is compe-tent at extracting information from certain types of data (e.g., unstructured text, structured tables on the web, etc.). In order to determine the re-sponse to a new query posed to such systems (e.g., is ..."
Abstract
- Add to MetaCart
Recently, several Web-scale knowledge harvesting systems have been built, each of which is compe-tent at extracting information from certain types of data (e.g., unstructured text, structured tables on the web, etc.). In order to determine the re-sponse to a new query posed to such systems (e.g., is sugar a healthy food?), it is useful to integrate opinions from multiple systems. If a response is desired within a specific time budget (e.g., in less than 2 seconds), then maybe only a subset of these resources can be queried. In this paper, we ad-dress the problem of knowledge integration for on-demand time-budgeted query answering. We propose a new method, AskWorld, which learns a policy that chooses which queries to send to which resources, by accommodating varying bud-get constraints that are available only at query (test) time. Through extensive experiments on real world datasets, we demonstrate AskWorld’s capability in selecting most informative resources to query within test-time constraints, resulting in improved perfor-mance compared to competitive baselines. 1
Training for Fast Sequential Prediction Using Dynamic Feature Selection
"... We present paired learning and inference algorithms for significantly reducing computation and increasing speed of the vector dot products in the classifiers that are at the heart of many NLP components. This is accomplished by partition-ing the features into a sequence of templates which are ordere ..."
Abstract
- Add to MetaCart
(Show Context)
We present paired learning and inference algorithms for significantly reducing computation and increasing speed of the vector dot products in the classifiers that are at the heart of many NLP components. This is accomplished by partition-ing the features into a sequence of templates which are ordered such that high confidence can often be reached using only a small fraction of all features. Pa-rameter estimation is arranged to maximize accuracy and early confidence in this sequence. We present experiments in left-to-right part-of-speech tagging on WSJ, demonstrating that we can preserve accuracy above 97 % with over a five-fold re-duction in run-time. 1