• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Tools

Sorted by:
Try your query at:
Semantic Scholar Scholar Academic
Google Bing DBLP
Results 1 - 10 of 1,489
Next 10 →

Table 1: Error functions for back propagation. This table, derived from Rumelhart et al. (1995), gives the appropriate error function (E), activation function (ai) for the output nodes, and two relevant derivatives from Equation 3 for three commonly assumed probability distributions. All summations range over all the output nodes.

in An Anytime Approach To Connectionist Theory Refinement: Refining The Topologies Of Knowledge-Based Neural Networks
by David William Opitz 1995
"... In PAGE 25: ... To maximize lnP (DjN), we need to know the distribution that the network apos;s outputs are to learn. Table1 shows three probability distributions that are commonly assumed in practice, and their appropriate error functions (Rumelhart et al., 1995).... In PAGE 88: ... Table1 0: Test-set error from a ten-fold cross validation. Table (a) shows the results from running three learners without the domain theory; Table (b) shows the results of running three learners with the domain theory.... In PAGE 88: ...5% 34.7% (b) Generating Non-KNN Ensembles Table1 0a presents the results from the case where the learners randomly create the topology of their networks (i.... In PAGE 88: ...opology of their networks (i.e., they do not use the domain theory). Table1 0a apos;s rst row, best-network, results from a single-layer neural network where, for each fold, I trained 20 networks containing between 0 and 100 (uniformly) hidden nodes and used a validation set to choose the best network. The next row, bagging, contains the results of applying the bagging algorithm to standard, single-hidden-layer networks, where the number of hidden nodes is randomly set between 0 and 100 for each network.... In PAGE 89: ...can produce large alterations in the predictions of the networks, thereby leading to an e ective ensemble. The bottom row of Table1 0a, Addemup, contains the results of a run of Addemup where its initial population (of size 20) is randomly generated using Regent apos;s method for creating networks when no domain theory is present (refer to Section 4.3.... In PAGE 89: ... Again, each ensemble con- tains 20 networks. The rst row of Table1 0b contains the generalization results of the Kbann algorithm, while the next row, Kbann-bagging, contains the results of the en- semble where each individual network in the ensemble is the Kbann network trained on a di erent partition of the training set. Even though each of these networks start with the same topology and \large quot; initial weight settings (i.... In PAGE 89: ...e., bagging in Table1 0a). The next two rows result from the Regent algorithm.... In PAGE 89: ... The rst row, Regent-best- network, contains the results from the single best network output by Regent, while the next row, Regent-combined, contains the results of simply combining, using Equa- tion 21, the networks in Regent apos;s nal population. Last chapter, I showed the e ective- ness of Regent-best-network, and comparing it with the results in Table1... In PAGE 90: ... Notice that simply combining the networks of Regent apos;s nal population (Regent-combined) decreases the test-set error over the single-best network picked by Regent. Addemup, the nal row of Table1 0b, mainly di ers from Regent-combined in two ways: (a) its tness function (i.e.... In PAGE 90: ...e., a repeat of Addemup from Table1 0b). The results show that, while reweighting the examples during training usually helps, Addemup gets most of its generalization power from its tness function.... In PAGE 91: ... Table1 1: Test-set error on the lesion studies of Addemup. Due to the inherent similarity of each algorithm and the lengthy run-times limiting the number of runs to a ten-fold cross-validation, the di erence between the lesions of Addemup is not signi cant at the 95% con dence level.... In PAGE 92: ...show that more diversity is needed when generating an e ective ensemble. There are two main reasons why I think the results of Addemup in Table1 0b are especially encouraging: (a) by comparing Addemup with Regent-combined, I explicitly test the quality of my tness function and demonstrate its e ectiveness, and (b) Addemup is able to e ectively utilize background knowledge to decrease the error of the individual networks in its ensemble, while still being able to create enough diversity among them so as to improve the overall quality of the ensemble. 5.... In PAGE 119: ... Table1 2: Domain theory for the arti cial chess problem. The OR in the following rules means that only one of the two antecedents needs to be satis ed in order for the rule to be satis ed.... In PAGE 124: ... Table1 3: Domain theory for nding promoters.... In PAGE 128: ... Table1 5: Domain theory for nding ribosome-binding sites.... In PAGE 130: ... Table1 6: Part 1 of the domain theory for nding transcription-termination sites; see Table 17 for the remainder of this theory.... In PAGE 131: ... Table1 7: Part 2 of the domain theory for nding transcription-termination sites; Table 16 contains the rst half of this theory. % A region upstream from the site should be rich with A apos;s and T apos;s.... In PAGE 131: ... loop-even :- @-2= quot;TTTT quot;. % As in Table1 6, this rule encodes the notion that a longer stem is more stable. Like % before, the consequent of this rule, stem-even, is true if the sum of the \weighted quot; % true antecedents is greater than 6.... ..."
Cited by 18

Table 1: Features of Traditional AI and Agent AI

in Active User Interfaces
by Scott M. Brown, Eugene Santos, Jr. 1990
"... In PAGE 9: ... [104] also does an excellent job of distinguishing \traditional AI quot; from the study of autonomous agents (what we apos;ll term agent AI). Table1 summarizes the main di erences traditional AI and agent AI. [104] discusses the basic problems AI researchers have with adaptive autonomous agents, which includes interface agents.... ..."
Cited by 12

Table 12. AI Matrix

in Discretization of Multidimensional Web Data for Informative Dense Regions Discovery
by Edmond H. Wu, Michael K. Ng, Andy M. Yip, Tony F. Chan 2004
"... In PAGE 10: ... The advantage of AI is that it considers the time factor. Table12 is the AI matrix, which is calculated by AD matrix (see Table 9) and page staying time (see Table 8). In this example, we use two-way discretization method, setting k = 2.... ..."
Cited by 1

Table 1. AIS Parameters

in An Artificial Immune System for Misbehavior Detection in Mobile Ad-Hoc Networks with Virtual Thymus, Clustering, Danger Signal and Memory Detectors
by Jean-yves Le Boudec 2004
"... In PAGE 11: ... Here we explain how the AIS works by putting a node equipped with the AIS in the network scenarios typical for invocation of certain blocks or concepts. For the details, please see the pseudo code of the six AIS building blocks (Appendix B) and the values of the used parameters ( Table1 in Section 5). Bootstrap.... In PAGE 18: ... Clustering is used in all the experiments, as we already have shown its advantage over simple matching in [2]. The values of the system parameters used in the simulation are given in Table1 . The same default values are used in all the experiments.... ..."
Cited by 5

Table Ai, then, is of the form

in Fixed Parameter Algorithms for Planar Dominating Set and Related Problems
by Jochen Alber, Hans L. Bodlaender, Henning Fernau, Rolf Niedermeier 2000
Cited by 39

Table AI Questionnaire

in unknown title
by unknown authors 1995
Cited by 1

Table AI. Model parameters.

in Architectural Specification for Massively Parallel Computers: An . . .
by Ron Brightwell, William Camp, Benjamin Cole, Erik DeBenedictis, Robert Leland, James Tomkins, Arthur B. MacCabe

TABLE AI RESPONDENT FONCTION

in Determinants of IT Adoption, Organizational Change and the Nature of the Firm - An Empirical Study of French Survey Data
by Rachel Bocquet, Olivier Brossard, An French, Survey Data

Table 1. Shared Properties of AIS Niches versus Typical AI Niches

in An Architecture for Adaptive Intelligent Systems
by Barbara Hayes-Roth
"... In PAGE 6: ...As illustrated by ICU monitoring and summarized in Table1 , AIS niches are considerably more demanding than the niches occupied by typical AI agents. First, AIS niches require performance of several diverse tasks, sometimes concurrently and often interacting.... ..."

Table 2. The AI2 algorithm.

in Skill acquisition via transfer learning and advice taking
by Lisa Torrey, Jude Shavlik, Trevor Walker, Richard Maclin 2006
"... In PAGE 5: ... It then applies both the transfer advice and the user advice to learning in the target task. Table2 summarizes the AI2 algorithm in high-level pseudocode. Figure 2 illustrates the transfer part of this algorithm with an example from RoboCup.... ..."
Cited by 6
Next 10 →
Results 1 - 10 of 1,489
Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University