• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Tools

Sorted by:
Try your query at:
Semantic Scholar Scholar Academic
Google Bing DBLP
Results 1 - 10 of 809,773
Next 10 →

Web Server Workload Characterization: The Search for Invariants (Extended Version)

by Martin F. Arlitt, Carey L. Williamson , 1996
"... The phenomenal growth in popularity of the World Wide Web (WWW, or the Web) has made WWW traffic the largest contributor to packet and byte traffic on the NSFNET backbone. This growth has triggered recent research aimed at reducing the volume of network traffic produced by Web clients and servers, b ..."
Abstract - Cited by 470 (6 self) - Add to MetaCart
, by using caching, and reducing the latency for WWW users, by using improved protocols for Web interaction. Fundamental to the goal of improving WWW performance is an understanding of WWW workloads. This paper presents a workload characterization study for Internet Web servers. Six different data sets

Learning the Kernel Matrix with Semi-Definite Programming

by Gert R. G. Lanckriet, Nello Cristianini, Laurent El Ghaoui, Peter Bartlett, Michael I. Jordan , 2002
"... Kernel-based learning algorithms work by embedding the data into a Euclidean space, and then searching for linear relations among the embedded data points. The embedding is performed implicitly, by specifying the inner products between each pair of points in the embedding space. This information ..."
Abstract - Cited by 780 (22 self) - Add to MetaCart
Kernel-based learning algorithms work by embedding the data into a Euclidean space, and then searching for linear relations among the embedded data points. The embedding is performed implicitly, by specifying the inner products between each pair of points in the embedding space. This information

Principled design of the modern web architecture

by Roy T. Fielding, Richard N. Taylor - ACM Trans. Internet Techn
"... The World Wide Web has succeeded in large part because its software architecture has been designed to meet the needs of an Internet-scale distributed hypermedia system. The modern Web architecture emphasizes scalability of component interactions, generality of interfaces, independent deployment of c ..."
Abstract - Cited by 507 (14 self) - Add to MetaCart
of components, and intermediary components to reduce interaction latency, enforce security, and encapsulate legacy systems. In this paper, we introduce the Representational State Transfer (REST) architectural style, developed as an abstract model of the Web architecture to guide our redesign and definition

Crowds: Anonymity for Web Transactions

by Michael K. Reiter, Aviel D. Rubin - ACM Transactions on Information and System Security , 1997
"... this paper we introduce a system called Crowds for protecting users' anonymity on the worldwide -web. Crowds, named for the notion of "blending into a crowd", operates by grouping users into a large and geographically diverse group (crowd) that collectively issues requests on behalf o ..."
Abstract - Cited by 831 (12 self) - Add to MetaCart
this paper we introduce a system called Crowds for protecting users' anonymity on the worldwide -web. Crowds, named for the notion of "blending into a crowd", operates by grouping users into a large and geographically diverse group (crowd) that collectively issues requests on behalf

DBpedia: A Nucleus for a Web of Open Data

by Christian Bizer, Georgi Kobilarov, Jens Lehmann, Zachary Ives - Proc. 6th Int’l Semantic Web Conf , 2007
"... Abstract DBpedia is a community effort to extract structured informa-tion from Wikipedia and to make this information available on the Web. DBpedia allows you to ask sophisticated queries against datasets derived from Wikipedia and to link other datasets on the Web to Wikipedia data. We describe the ..."
Abstract - Cited by 619 (36 self) - Add to MetaCart
Abstract DBpedia is a community effort to extract structured informa-tion from Wikipedia and to make this information available on the Web. DBpedia allows you to ask sophisticated queries against datasets derived from Wikipedia and to link other datasets on the Web to Wikipedia data. We describe

Focused crawling: a new approach to topic-specific Web resource discovery

by Soumen Chakrabarti, Martin van den Berg, Byron Dom , 1999
"... The rapid growth of the World-Wide Web poses unprecedented scaling challenges for general-purpose crawlers and search engines. In this paper we describe a new hypertext resource discovery system called a Focused Crawler. The goal of a focused crawler is to selectively seek out pages that are relevan ..."
Abstract - Cited by 628 (10 self) - Add to MetaCart
The rapid growth of the World-Wide Web poses unprecedented scaling challenges for general-purpose crawlers and search engines. In this paper we describe a new hypertext resource discovery system called a Focused Crawler. The goal of a focused crawler is to selectively seek out pages

Irrelevant Features and the Subset Selection Problem

by George H. John, Ron Kohavi, Karl Pfleger - MACHINE LEARNING: PROCEEDINGS OF THE ELEVENTH INTERNATIONAL , 1994
"... We address the problem of finding a subset of features that allows a supervised induction algorithm to induce small high-accuracy concepts. We examine notions of relevance and irrelevance, and show that the definitions used in the machine learning literature do not adequately partition the features ..."
Abstract - Cited by 741 (26 self) - Add to MetaCart
We address the problem of finding a subset of features that allows a supervised induction algorithm to induce small high-accuracy concepts. We examine notions of relevance and irrelevance, and show that the definitions used in the machine learning literature do not adequately partition the features

Gapped Blast and PsiBlast: a new generation of protein database search programs

by Stephen F. Altschul, Thomas L. Madden, Alejandro A. Schäffer, Jinghui Zhang, Zheng Zhang, Webb Miller, David J. Lipman - NUCLEIC ACIDS RESEARCH , 1997
"... The BLAST programs are widely used tools for searching protein and DNA databases for sequence similarities. For protein comparisons, a variety of definitional, algorithmic and statistical refinements described here permits the execution time of the BLAST programs to be decreased substantially while ..."
Abstract - Cited by 8393 (85 self) - Add to MetaCart
The BLAST programs are widely used tools for searching protein and DNA databases for sequence similarities. For protein comparisons, a variety of definitional, algorithmic and statistical refinements described here permits the execution time of the BLAST programs to be decreased substantially while

A saliency-based search mechanism for overt and covert shifts of visual attention

by Laurent Itti, Christof Koch , 2000
"... ..."
Abstract - Cited by 608 (32 self) - Add to MetaCart
Abstract not found

Social Network Sites: A Definition A History of Social Network Sites Previous Scholarship

by Nicole B. Ellison
"... ..."
Abstract - Cited by 915 (10 self) - Add to MetaCart
Abstract not found
Next 10 →
Results 1 - 10 of 809,773
Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University