Results 1  10
of
45
The Lazy Lambda Calculus
 Research Topics in Functional Programming
, 1990
"... Introduction The commonly accepted basis for functional programming is the calculus; and it is folklore that the calculus is the prototypical functional language in puri ed form. But what is the calculus? The syntax is simple and classical; variables, abstraction and application in the pure cal ..."
Abstract

Cited by 258 (3 self)
 Add to MetaCart
(Show Context)
Introduction The commonly accepted basis for functional programming is the calculus; and it is folklore that the calculus is the prototypical functional language in puri ed form. But what is the calculus? The syntax is simple and classical; variables, abstraction and application in the pure calculus, with applied calculi obtained by adding constants. The further elaboration of the theory, covering conversion, reduction, theories and models, is laid out in Barendregt's already classical treatise [Bar84]. It is instructive to recall the following crux, which occurs rather early in that work (p. 39): Meaning of terms: rst attempt The meaning of a term is its normal form (if it exists). All terms without normal forms are identi ed. This proposal incorporates such a simple and natural interpretation of the calculus as
On reductionbased process semantics
 in Proceedings of FSTTCS ’93, LNCS 761
, 1995
"... Abstract. A formulation of semantic theories for processes which is based on reduction relation and equational reasoning is studied. The new construction can induce meaningful theories for processes, both in strong and weak settings. The resulting theories in many cases coincide with, and sometimes ..."
Abstract

Cited by 167 (26 self)
 Add to MetaCart
(Show Context)
Abstract. A formulation of semantic theories for processes which is based on reduction relation and equational reasoning is studied. The new construction can induce meaningful theories for processes, both in strong and weak settings. The resulting theories in many cases coincide with, and sometimes generalise, observationbased formulation of behavioural equivalence. The basic construction of reductionbased theories is studied, taking a simple name passing calculus called $\nu$calculus as an example. Results on other calculi are also briefly discussed. 1
The Lazy Lambda Calculus in a Concurrency Scenario (Extended Abstract)
 Information and Computation
, 1994
"... ) Davide Sangiorgi LFCS  Department of Computer Science Edinburgh University Edinburgh  EH9 3JZ  UK Abstract The use of lambda calculus in richer settings, possibly involving parallelism, is examined in terms of its effect on the equivalence between lambda terms. We concentrate here on Abra ..."
Abstract

Cited by 70 (9 self)
 Add to MetaCart
) Davide Sangiorgi LFCS  Department of Computer Science Edinburgh University Edinburgh  EH9 3JZ  UK Abstract The use of lambda calculus in richer settings, possibly involving parallelism, is examined in terms of its effect on the equivalence between lambda terms. We concentrate here on Abramsky's lazy lambda calculus and we follow two directions. First, the lambda calculus is studied within a process calculus by examining the equivalence $ induced by Milner's encoding into the calculus. We give exact operational and denotational characterizations for $. Secondly, we examine Abramsky's applicative bisimulation when the lambda calculus is augmented with (wellformed) operators, i.e. symbols equipped with reduction rules describing their behaviour. Then, maximal discrimination is obtained when all operators are considered; we show that this discrimination coincides with the one given by $ and that the adoption of certain nondeterministic operators is sufficient and necessary...
Strong Stability and the Incompleteness of Stable Models for λCalculus
 ANNALS OF PURE AND APPLIED LOGIC
, 1999
"... We prove that the class of stable models is incomplete with respect to pure λcalculus. More precisely, we show that no stable model has the same theory as the strongly stable version of Park's model. This incompleteness proof can be adapted to the continuous case, giving an incompleteness proo ..."
Abstract

Cited by 23 (0 self)
 Add to MetaCart
We prove that the class of stable models is incomplete with respect to pure λcalculus. More precisely, we show that no stable model has the same theory as the strongly stable version of Park's model. This incompleteness proof can be adapted to the continuous case, giving an incompleteness proof for this case which is much simpler than the original proof by Honsell an Ronchi della Rocca. Moreover, we isolate a very simple finite set, F , of equations and inequations, which has neither a stable nor a continuous model, and which is included in Th(P fs ) and in T
Eager normal form bisimulation
 In Proc. 20th Annual IEEE Symposium on Logic in Computer Science
, 2005
"... Abstract. Normal form bisimulation is a powerful theory of program equivalence, originally developed to characterize LévyLongo tree equivalence and Boehm tree equivalence. It has been adapted to a range of untyped, higherorder calculi, but types have presented a difficulty. In this paper, we prese ..."
Abstract

Cited by 20 (8 self)
 Add to MetaCart
Abstract. Normal form bisimulation is a powerful theory of program equivalence, originally developed to characterize LévyLongo tree equivalence and Boehm tree equivalence. It has been adapted to a range of untyped, higherorder calculi, but types have presented a difficulty. In this paper, we present an account of normal form bisimulation for types, including recursive types. We develop our theory for a continuationpassing style calculus, JumpWithArgument (JWA), where normal form bisimilarity takes a very simple form. We give a novel congruence proof, based on insights from game semantics. A notable feature is the seamless treatment of etaexpansion. We demonstrate the normal form bisimulation proof principle by using it to establish a syntactic minimal invariance result and the uniqueness of the fixed point operator at each type.
Discrete Normalization and Standardization in Deterministic Residual Structures
 In ALP '96 [ALP96
, 1996
"... . We prove a version of the Standardization Theorem and the Discrete Normalization (DN) Theorem in stable Deterministic Residual Structures, which are Abstract Reduction Systems with axiomatized residual relation, and model orthogonal rewrite systems. The latter theorem gives a strategy for construc ..."
Abstract

Cited by 10 (3 self)
 Add to MetaCart
. We prove a version of the Standardization Theorem and the Discrete Normalization (DN) Theorem in stable Deterministic Residual Structures, which are Abstract Reduction Systems with axiomatized residual relation, and model orthogonal rewrite systems. The latter theorem gives a strategy for construction of reductions L'evyequivalent (or permutationequivalent) to a given, finite or infinite, regular (or semilinear) reduction, based on the neededness concept of Huet and L'evy. This and other results of this paper add to the understanding of L'evyequivalence of reductions, and consequently, L'evy's reduction space. We demonstrate how elements of this space can be used to give denotational semantics to known functional languages in an abstract manner. 1 Introduction Long ago, Curry and Feys [CuFe58] proved that repeated contraction of leftmostoutermost redexes in any normalizable term eventually yields its normal form, even if the term possesses infinite reductions as well. The reaso...
A Filter Model for Concurrent λCalculus
 SIAM J. COMPUT
, 1998
"... Type free lazy λcalculus is enriched with angelic parallelism and demonic nondeterminism. Callbyname and callbyvalue abstractions are considered and the operational semantics is stated in terms of a must convergence predicate. We introduce a type assignment system with intersection and union ty ..."
Abstract

Cited by 9 (2 self)
 Add to MetaCart
Type free lazy λcalculus is enriched with angelic parallelism and demonic nondeterminism. Callbyname and callbyvalue abstractions are considered and the operational semantics is stated in terms of a must convergence predicate. We introduce a type assignment system with intersection and union types and we prove that the induced logical semantics is fully abstract.
A Categorytheoretic characterization of functional completeness
, 1990
"... . Functional languages are based on the notion of application: programs may be applied to data or programs. By application one may define algebraic functions; and a programming language is functionally complete when any algebraic function f(x 1 ,...,x n ) is representable (i.e. there is a constant a ..."
Abstract

Cited by 8 (1 self)
 Add to MetaCart
. Functional languages are based on the notion of application: programs may be applied to data or programs. By application one may define algebraic functions; and a programming language is functionally complete when any algebraic function f(x 1 ,...,x n ) is representable (i.e. there is a constant a such that f(x 1 ,...,x n ) = (a . x 1 . ... . x n ). Combinatory Logic is the simplest typefree language which is functionally complete. In a sound categorytheoretic framework the constant a above may be considered as an "abstract gödelnumber" for f, when gödelnumberings are generalized to "principal morphisms", in suitable categories. By this, models of Combinatory Logic are categorically characterized and their relation is given to lambdacalculus models within Cartesian Closed Categories. Finally, the partial recursive functionals in any finite higher type are shown to yield models of Combinatory Logic. ________________ (+) Theoretical Computer Science, 70 (2), 1990, pp.193211. A p...
Discrimination by Parallel Observers: the Algorithm
 LICS '97 , IEEE COMP. SOC
, 1998
"... The main result of the paper is a constructive proof of the following equivalence: two pure λterms are observationally equivalent in the lazy concurrent λcalculus iff they have the same LévyLongo trees. An algorithm which allows to build a context discriminating any two pure λterms with diffe ..."
Abstract

Cited by 8 (3 self)
 Add to MetaCart
The main result of the paper is a constructive proof of the following equivalence: two pure λterms are observationally equivalent in the lazy concurrent λcalculus iff they have the same LévyLongo trees. An algorithm which allows to build a context discriminating any two pure λterms with different LévyLongo trees is described. It follows that contextual equivalence coincides with behavioural equivalence (bisimulation) as considered by Sangiorgi. Another consequence is that the discriminating power of concurrent lambda contexts is the same as that of BoudolLaneve's contexts with multiplicities.