• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Tools

Sorted by:
Try your query at:
Semantic Scholar Scholar Academic
Google Bing DBLP
Results 1 - 10 of 35,327
Next 10 →

Table 1: Compression of natural language alphabets

in Is Huffman Coding Dead?
by A. Bookstein, S.T. Klein 1993
"... In PAGE 3: ... Here, it is often desirable to extend the alphabet to incorporate n-grams that include the frequent characters; in the expanded alphabet, the highest probability will be reduced and Hu man coding could be e ective. That Hu man coding is e ective for databases of natural text is shown in the rst two columns of Table1 , which is discussed in detail below. There we nd the Hu man cost (de ned as... In PAGE 6: ...and 735 bigrams), the distribution has been computed using the database of The Re- sponsa Retrieval Project (RRP) [14] of about 40 million Hebrew and Aramaic words; the distribution for Italian, Portuguese and Spanish (26 letters each) can be found in Gaines [18], and for Russian (32 letters) in Herdan [23]. The results are summarized in Table1 , the two last lines corresponding to the bigrams. The rst two colums list the average codeword length of Hu man codes and arith- metic codes respectively, and the third column gives the increase of the former over the latter in percent.... In PAGE 10: ...rom 4.16 to 5.60! We thus see that arithmetic codes give worse compression in this case, even without considering the overhead caused by EOF. Table 3 summarizes an experiment in which we took the probability distributions of English, German, Finnish and French (as in Table1 ), adding to them the character distribution in Gadsby, and checked what happens if they are mutually interchanged. The rows correspond to the distributions which are used to generate the codewords (the assumed distribution), and the columns correspond to the distribution that actually occurs (the true distribution).... In PAGE 17: ... To store the alphabet with methods B, C and D one needs 26, 32 and 3 bytes respectively, to which 13 bytes have to be added for the lengths of the Hu man codewords, or 52 bytes for the probabilities for arithmetic codes. From Table1 we know that the average loss per character by using Hu man instead of arithmetic codes is 0.0251 bits, so the text has to be at least of length 12431 characters to justify the excess of these 39 bytes for methods B, C and D.... ..."
Cited by 13

Table 6: Retrieval results of n-gram stemmers using WEST natural language queries

in Corpus-Based Stemming using Co-occurrence of Word Variants
by Jinxi Xu, W. Bruce Croft 1998
Cited by 60

Table 1: Complementary Interface Technologies: Direct manipulation and natural language

in Enhancing Multimedia Interfaces With Intelligence
by Michael Wilson 1995
"... In PAGE 5: ... Why Multiple Modes ? The justification for using multiple output media to present information are given elsewhere in this book. The motivation for trying to use multiple input modes rather than relying on direct manipulation or command languages alone is that individual modes have different strengths and weaknesses as illustrated in Table1 (after Cohen, 1992). ------------------------ Table 1 about here ------------------------ The objective of using multiple input modes is to allow the user to utilise the strengths of each mode while providing mechanisms for overcoming the weaknesses of each.... In PAGE 5: ... The motivation for trying to use multiple input modes rather than relying on direct manipulation or command languages alone is that individual modes have different strengths and weaknesses as illustrated in Table 1 (after Cohen, 1992). ------------------------ Table1 about here ------------------------ The objective of using multiple input modes is to allow the user to utilise the strengths of each mode while providing mechanisms for overcoming the weaknesses of each. In conjunction with multiple output media the use of the corresponding input media will provide maximum engagement with the information.... ..."
Cited by 5

Table 1: Complementary Interface Technologies: Direct manipulation and natural language

in Earnshaw & Vince Academic Press Book Chapter- Jan 1994 ENHANCING MULTIMEDIA INTERFACES WITH INTELLIGENCE
by Michael Wilson
"... In PAGE 5: ...Why Multiple Modes ? The justification for using multiple output media to present information are given elsewhere in this book. The motivation for trying to use multiple input modes rather than relying on direct manipulation or command languages alone is that individual modes have different strengths and weaknesses as illustrated in Table1 (after Cohen, 1992). ------------------------ Table 1 about here ------------------------ The objective of using multiple input modes is to allow the user to utilise the strengths of each mode while providing mechanisms for overcoming the weaknesses of each.... In PAGE 5: ... The motivation for trying to use multiple input modes rather than relying on direct manipulation or command languages alone is that individual modes have different strengths and weaknesses as illustrated in Table 1 (after Cohen, 1992). ------------------------ Table1 about here ------------------------ The objective of using multiple input modes is to allow the user to utilise the strengths of each mode while providing mechanisms for overcoming the weaknesses of each. In conjunction with multiple output media the use of the corresponding input media will provide maximum engagement with the information.... ..."

Table 2: Levels of language models for document retrieval Object Base model Residual Full model

in General Terms
by Djoerd Hiemstra
"... In PAGE 4: ...4 Model Summary We have identified mixture models that can be used at three stages in the retrieval process to infer parsimonious language models. Table2 shows the models used in these three stages and the relation between them [24]. In each case in this table the base model is externally defined (in- dependent of the object under consideration).... In PAGE 4: ... In initial query formulation, we fit the RM(GM) to the only information that we have about the request, namely its original text. In [24], it was assumed that the search process would involve trying out the level 3 model on every document (shown as 3a in Table2 ), against a null hypothesis which would be the level 2 model derived at indexing time. This comparison would again have to appeal to parsimony.... ..."

Table 2: Levels of language models for document retrieval Object Base model Residual Full model

in Parsimonious Language Models for Information Retrieval
by Djoerd Hiemstra, Stephen Robertson, Hugo Zaragoza
"... In PAGE 4: ...4 Model Summary We have identified mixture models that can be used at three stages in the retrieval process to infer parsimonious language models. Table2 shows the models used in these three stages and the relation between them [?]. In each case in this table the base model is externally defined (in- dependent of the object under consideration).... In PAGE 4: ... In initial query formulation, we fit the RM(GM) to the only information that we have about the request, namely its original text. In [?], it was assumed that the search process would involve trying out the level 3 model on every document (shown as 3a in Table2 ), against a null hypothesis which would be the level 2 model derived at indexing time. This comparison would again have to appeal to parsimony.... ..."

Table 1. CAPTION

in A simple algorithm for complete motion planning . . .
by Gokul Varadhan, Shankar Krishnan, T. V. N. Sriram, Dinesh Manocha
"... In PAGE 10: ... We used the C++ programming language with the GNU g++ compiler under the Linux operating system. Table1 highlights the performance of our algorithm on these models. The execution times are on a 2 GHz Pentium IV PC with a GeForce 4 graphics card and 1 GB RAM.... In PAGE 10: ... The execution times are on a 2 GHz Pentium IV PC with a GeForce 4 graphics card and 1 GB RAM. Table1 provides a breakup of the total time in computing pairwiseconvexMinkowskisums,constructingaconnectivity roadmap and performing the graph search in the connectivity graph. It shows that most of the time is spent on roadmap con- struction and a very small fraction of the total time is spent on graph search and pairwise Minkowski sum computation.... ..."

Table 5: The retrieved objects

in Dempster-Shafer's Theory of Evidence applied to Structured Documents: Modelling Uncertainty
by Mounia Lalmas

Table 1 shows an illustrative list of text mining products and applications based on the text refining and knowledge distillation functions as well as the intermediate form adopted. The text mining products/applications can be roughly organized into two groups. One group focuses on document organization, visualization, and navigation. The other group focuses on text analysis functions, notably, information retrieval, information extraction, categorization, and summarization. While we see that most text mining systems provide natural language processing (NLP) functions, few, if any, have integrated data mining functions for knowledge distillation across concepts or objects.

in Text Mining: Promises And Challenges
by Ah-hwee Tan 1999
"... In PAGE 4: ... Table1 : A list of selected text mining products and applications based on the text refining and knowledge distillation functions as well as the intermediate form... ..."
Cited by 1

Table 3 Proposed controlled vocabulary for therapeutic music retrieval

in Facilitating Retrieval of Sound Recordings for Use By Professionals Treating Children
by Dena L. Belvin, Brian Sturm 2007
"... In PAGE 32: ... A controlled vocabulary and natural language are two tools that can be implemented for this search process. Examples of controlled vocabulary and accompanying natural language are in Table3 . The natural language area will need to be frequently updated based on input from the professionals using the system.... ..."
Next 10 →
Results 1 - 10 of 35,327
Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University