Results 1  10
of
91
Predicting the Drape of Woven Cloth Using Interacting Particles
, 1994
"... We demonstrate a physicallybased technique for predicting the drape of a wide variety of woven fabrics. The approach exploits a theoretical model that explicitly represents the microstructure of woven cloth with interacting particles, rather than utilizing a continuum approximation. By testing a cl ..."
Abstract

Cited by 121 (5 self)
 Add to MetaCart
We demonstrate a physicallybased technique for predicting the drape of a wide variety of woven fabrics. The approach exploits a theoretical model that explicitly represents the microstructure of woven cloth with interacting particles, rather than utilizing a continuum approximation. By testing a cloth sample in a Kawabata fabric testing device, we obtain data that is used to tune the model's energy functions, so that it reproduces the draping behavior of the original material. Photographs, comparing the drape of actual cloth with visualizations of simulation results, show that we are able to reliably model the unique largescale draping characteristics of distinctly different fabric types. iii Figure 1.1: Draping cloth objects 1 Introduction The vast number of uses for cloth are mirrored in the extraordinary variety of types of woven fabrics. These range from the most exquisite fine silks, to the coarsest of burlaps, and are woven from such diverse fibers as natural wool and synth...
Offering a PrecisionPerformance Tradeoff for Aggregation Queries over Replicated Data
, 2000
"... Strict consistency of replicated data is infeasible or not required by many distributed applications, so current systems often permit stale replication,inwhich cached copies of data values are allowed to become out of date. Queries over cached data return an answer quickly, but the stale answer ..."
Abstract

Cited by 92 (8 self)
 Add to MetaCart
Strict consistency of replicated data is infeasible or not required by many distributed applications, so current systems often permit stale replication,inwhich cached copies of data values are allowed to become out of date. Queries over cached data return an answer quickly, but the stale answer may be unboundedly imprecise. Alternatively, queries over remote master data return a precise answer, but with potentially poor performance. To bridge the gap between these two extremes, we propose a new class of replication systems called TRAPP (Tradeoff in Replication Precision and Performance). TRAPP systems give each user finegrained control over the tradeoff between precision and performance: Caches store ranges that are guaranteed to bound the current data values, instead of storing stale exact values. Users supply a quantitative precision constraint along with each query. To answer a query, TRAPP systems automatically select a combination of locally cached bounds and exact master data stored remotely to deliver a bounded answer consisting of a range that is no wider than the specified precision constraint, that is guaranteed to contain the precise answer, and that is computed as quickly as possible. This paper defines the architecture of TRAPP replication systems and covers some mechanics of caching data ranges. It then focuses on queries with aggregation, presenting optimization algorithms for answering queries with precision constraints, and reporting on performance experiments that demonstrate the finegrained control of the precisionperformance tradeoff offered by TRAPP systems.
Automatic Structures
 IN PROC. 15TH IEEE SYMP. ON LOGIC IN COMPUTER SCIENCE
, 1999
"... We study definability and complexity issues for automatic and wautomatic structures. These are, in general, infinite structures but they can be finitely presented by a collection of automata. Moreover, they admit effective (in fact automatic) evaluation of all firstorder queries. Therefore, automa ..."
Abstract

Cited by 89 (7 self)
 Add to MetaCart
We study definability and complexity issues for automatic and wautomatic structures. These are, in general, infinite structures but they can be finitely presented by a collection of automata. Moreover, they admit effective (in fact automatic) evaluation of all firstorder queries. Therefore, automatic structures provide an interesting framework for extending many algorithmic and logical methods from finite structures to infinite ones. We explain the notion of (w)automatic structures, give examples, and discuss the relationship to automatic groups. We determine the complexity of model checking and query evaluation on automatic structures for fragments of firstorder logic. Further, we study closure properties and definability issues on automatic structures and present a technique for proving that a structure is not automatic. We give modeltheoretic characterisations for automatic structures via interpretations. Finally we discuss the composition theory of automatic structures and pro...
Racer: A Core Inference Engine for the Semantic Web
, 2003
"... In this paper we describe Racer, which can be considered as a core inference engine for the semantic web. The Racer inference server o#ers two APIs that are already used by at least three di#erent network clients, i.e., the ontology editor OilEd, the visualization tool RICE, and the ontology dev ..."
Abstract

Cited by 74 (0 self)
 Add to MetaCart
In this paper we describe Racer, which can be considered as a core inference engine for the semantic web. The Racer inference server o#ers two APIs that are already used by at least three di#erent network clients, i.e., the ontology editor OilEd, the visualization tool RICE, and the ontology development environment Protege 2. The Racer server supports the standard DIG protocol via HTTP and a TCP based protocol with extensive query facilities. Racer currently supports the web ontology languages DAML+OIL, RDF, and OWL.
A Query Language Based on the Ambient Logic
, 2001
"... The ambient logic is a modal logic proposed to describe the structural and computational properties of distributed and mobile computation. The structural part of the ambient logic is, essentially, a logic of labeled trees, hence it turns out to be a good foundation for query languages for semistruct ..."
Abstract

Cited by 63 (11 self)
 Add to MetaCart
The ambient logic is a modal logic proposed to describe the structural and computational properties of distributed and mobile computation. The structural part of the ambient logic is, essentially, a logic of labeled trees, hence it turns out to be a good foundation for query languages for semistructured data, much in the same way as first order logic is a fitting foundation for relational query languages. We define here a query language for semistructured data that is based on the ambient logic, and we outline an execution model for this language. The language turns out to be quite expressive. Its strong foundations and the equivalences that hold in the ambient logic are helpful in the definition of the language semantics and execution model.
Constraint Programming and Database Query Languages
 In Proc. 2nd Conference on Theoretical Aspects of Computer Software (TACS
, 1994
"... . The declarative programming paradigms used in constraint languages can lead to powerful extensions of Codd's relational data model. The development of constraint database query languages from logical database query languages has many similarities with the development of constraint logic programmin ..."
Abstract

Cited by 60 (3 self)
 Add to MetaCart
. The declarative programming paradigms used in constraint languages can lead to powerful extensions of Codd's relational data model. The development of constraint database query languages from logical database query languages has many similarities with the development of constraint logic programming from logic programming, but with the additional requirements of data efficient, setatatime, and bottomup evaluation. In this overview of constraint query languages (CQLs) we first present the framework of [41]. The principal idea is that: "the ktuple (or record) data type can be generalized by a conjunction of quantifierfree constraints over k variables". The generalization must preserve various language properties of the relational data model, e.g., the calculus/algebra equivalence, and have time complexity polynomial in the size of the data. We next present an algebra for dense order constraints that is simpler to evaluate than the calculus described in [41], and we sharpen some of...
Finitely Representable Databases
, 1995
"... : We study classes of infinite but finitely representable databases based on constraints, motivated by new database applications such as geographical databases. We formally define these notions and introduce the concept of query which generalizes queries over classical relational databases. We prove ..."
Abstract

Cited by 55 (8 self)
 Add to MetaCart
: We study classes of infinite but finitely representable databases based on constraints, motivated by new database applications such as geographical databases. We formally define these notions and introduce the concept of query which generalizes queries over classical relational databases. We prove that in this context the basic properties of queries (satisfiability, containment, equivalence, etc.) are nonrecursive. We investigate the theory of finitely representable models and prove that it differs strongly from both classical model theory and finite model theory. In particular, we show that most of the well known theorems of either one fail (compactness, completeness, locality, 0/1 laws, etc.). An immediate consequence is the lack of tools to consider the definability of queries in the relational calculus over finitely representable databases. We illustrate this very challenging problem through some classical examples. We then mainly concentrate on dense order databases, and exhibit...
An Experimental CLP Platform for Integrity Constraints and Abduction
 In Proceedings of FQAS2000, Flexible Query Answering Systems: Advances in Soft Computing series
, 2000
"... Integrity constraint and abduction are important in queryanswering systems for enhanced query processing and for expressing knowledge in databases. A straightforward characterization of the two is given in a subset of the language CHR _ , originally intended for writing constraint solvers to be a ..."
Abstract

Cited by 35 (13 self)
 Add to MetaCart
Integrity constraint and abduction are important in queryanswering systems for enhanced query processing and for expressing knowledge in databases. A straightforward characterization of the two is given in a subset of the language CHR _ , originally intended for writing constraint solvers to be applied for CLP languages. This subset has a strikingly simple computational model that can be executed using existing, Prologbased technology. Together with earlier results, this confirms CHR _ as a multiparadigm platform for experimenting with combinations of topdown and bottomup evaluation, disjunctive databases and, as shown here, integrity constraint and abduction 1 Introduction Constraint logic programming (CLP) [10] is established as an extension to logic programming that adds higher expressibility and in some cases more efficient query evaluation. CLP has also given rise to a field of constraint databases [14]. In the present paper, we suggest CLP techniques applied for defini...
Variable Independence and Aggregation Closure
 IN ACM SYMPOSIUM ON PRINCIPLES OF DATABASE SYSTEMS
, 1996
"... We discuss the issue of adding aggregation to constraint databases. Previous work has shown that, in general, adding aggregates to constraint databases results in languages that are not closed. We show that by imposing a natural restriction, called variable independence (which is a generalization of ..."
Abstract

Cited by 31 (10 self)
 Add to MetaCart
We discuss the issue of adding aggregation to constraint databases. Previous work has shown that, in general, adding aggregates to constraint databases results in languages that are not closed. We show that by imposing a natural restriction, called variable independence (which is a generalization of the assumptions underlying the classical relational model of data) on the schema, we can guarantee that a restricted version of the language with aggregation is closed. We illustrate our approach in the context of linear constraint databases.
Automatic construction of simple artifactbased workflows
 In: Proc. of the 12th Int. Conf. on Database Theory (ICDT 2009
, 2009
"... Almost all medium and largescale businesses rely on electronic workflow systems to manage their business processes. A key challenge is to enable the easy reuse and modification of these workflow schemas and their pieceparts, so that they can be adapted to new business situations. This paper desc ..."
Abstract

Cited by 26 (2 self)
 Add to MetaCart
Almost all medium and largescale businesses rely on electronic workflow systems to manage their business processes. A key challenge is to enable the easy reuse and modification of these workflow schemas and their pieceparts, so that they can be adapted to new business situations. This paper describes an approach for automatic construction (and thus, evolution) of a workflow schema that satisfies a specified condition (or “goal”), starting from a set of basic building block services (or “tasks”). We use a workflow model based on “business artifacts”, which represent key (real or conceptual) business entities, and include both the businessrelevant data about them and a specification of their lifecycle, that is, how they can evolve over time as they move through the workflow as the result of services being applied to them. This paper uses a declarative form of artifactcentric workflow. The