Re: Arrays and Assoc
> >But I think we can have the cake and eat it too, if we get rid of the > >restriction (which I never liked) that operators beginning with : must be a > >constructor: just define > >a := b = (a,b) > > Unfortunately that won't work if := had been used in patterns. I think > backward compatibility is an issue. The standard technique of supporting > Assoc but with compiler warnings will probably have to be used. Excuse my previous message. I misunderstood John's comment. Yes, := in patterns would be problematic. Something like import Prelude renaming ((,) to (:=)) could have helped, if it wasn't that is was forbidden in several ways. (First (,) is not allowed (I think it should!!!), second you'd want both (,) and (:=) visible.) -- Lennart
Re: Arrays and Assoc
> >But I think we can have the cake and eat it too, if we get rid of the > >restriction (which I never liked) that operators beginning with : must be a > >constructor: just define > >a := b = (a,b) > > Unfortunately that won't work if := had been used in patterns. Nonsense. Of course constructors can be arbitrary symbols (and identifiers), it just makes the compiler a little more complicated. -- Lennart
Recursive type synonyms
While we are proposing things, here's a further suggestion. The great thing about this suggestion is that it only *removes* a restriction, and makes the language definition simpler. It is fully backward compatible. The suggestion is: Remove the restriction that type synonym declarations must not be recursive. In other words, one could write things like type Stream a = (a, Stream a) which is equivalent to the type (a, (a, (a, ...))). The only reason we included the restriction at the time was (a) it makes unification easier to implement (b) it was more standard (c) there didn't seem any compelling reason *not* to include the restriction. Guy Steele has since pointed out several compelling examples where it would be *much* easier not to have the restriction, and I've encountered a few myself. Let's trash it! The obvious way to go is for someone to implement it first, to make sure it's not difficult. Mark Jones, have you tried this in Gofer yet? Cheers, -- P --- Professor Philip Wadler[EMAIL PROTECTED] Department of Computing Sciencetel: +44 41 330 4966 University of Glasgow fax: +44 41 330 4913 Glasgow G12 8QQ, SCOTLAND
Warren's proposed type signature syntax
Although Warren's suggestion is logical, it's not in any way standard. (Most of the places where we have two ways of doing something, they are both standard.) It might be a nice idea, but I'm not at all convinced it's nice enough to be worth putting into Haskell. -- P
Re: ADTs and strictness
Gerald Ostheimer notes that in Abramsky and Ong's lazy lambda calculus that (\x -> bottom) differs from bottom. That's correct. But just because they call it `lazy' doesn't mean that it really is the essence of laziness. I prefer to use the more neutral name `lifted lambda calculus' for their calculus. An example of a perfectly good lazy language in which neither products nor functions are lifted is Miranda (a trademark of Research Software Limited). Hope this clarifies things, -- P
Re: Lifted products
Oops! I should have underlined in my last message where I wrote `newtype' instead of `datatype'. As a result, Simon seems to have completely misunderstood my proposal. Sorry about that. Simon seems to think I am proposing that if one writes datatype T a_1 ... a_k = C t_1 ... t_n that one gets unlifted tuples. I am *not* proposing this. What I propose is that if one writes newtype T a_1 ... a_k = C t_1 ... t_n then one gets unlifted tuples. I'm not stuck on the keyword `newtype', anything other than `datatype' will do. Simon writes of my true proposal (which he mistakenly labels an alternative) `I like it not'. But doesn't say why. In particular, he seems not to have hoisted on board that my proposal is just a *generalisation* of his proposal to write newtype T a_1 ... a_k = C t. to declare a type isomorphic to an existing type. In particular, if one wants to create a type `New a' isomorphic to an existing type, Simon would write (by his latest proposal) datatype Data a => New a = MakeNew a whereas I would write newtype New a = MakeNew a So my alternative is simpler in some ways. Simon also notes that strictness declarations don't seem sensible for unlifted constructors. Indeed. Ban them. (Again, this is an argument against something I never proposed.) I think Simon's other points about ~ patterns are spurious. But I don't want to rebut them, because now that I've pointed out that he misunderstood my proposal, perhaps he no longer holds to them. Simon (or anyone else), if you have further arguments against what I *did* propose, please raise them again and I'll answer. All in the spirit of a quest for the perfect Haskell! -- P
Re: Recursive type synonyms
To illustrate a need for recursive type synonyms, Joe suggests the example: | nil f g = f | cons x xs f g = g x xs | | fold f z xs = xs z (\x xs -> f x (fold f z xs)) Indeed, this doesn't type check in Haskell, and recursive type synonyms would fix it. (So would general recursive types if we defined List a b as a synonym for mu l . b -> (a -> l -> b) -> b.) However, these definitions don't quite correspond to the standard encoding of data structures as functions. In polymorphic lambda calculus, we might define a type of lists: List a = Forall b . b -> (a -> b -> b) -> b nil = /\a. /\b. \n:b. \c:(a -> b -> b). n cons = /\a. \x:a. \xs:List a. /\b. \n:b. \c:(a -> b -> b). c x (xs b n c) Projecting back into a Hindley/Milner type system for Haskell gives: type List a b = b -> (a -> b -> b) -> b nil :: List a b nil f g= f cons :: a -> List a b -> List a b cons x xs f g = g x (xs f g) which is accepted by the type system (the price of the reduced polymorphism being the need to carry the return type `b' around as an extra part of the type of the lists). We can still define a fold function: fold :: b -> (a -> b -> b) -> List a b -> b fold z f xs = xs z f and then, if we define example = cons 1 (cons 2 (cons 3 (cons 4 nil))), we can find that fold 0 (+) example = 10 and fold 1 (*) example = 24. So, I can't do the example exactly as Joe gave it without using recursive types, but there is another way to do the same kind of things that doesn't need recursive types or, in particular, recursive type synonyms. Mark
Re: Recursive type synonyms
Phil suggests that Haskell 1.3 might: | Remove the restriction that type synonym | declarations must not be recursive. | | In other words, one could write things like | | type Stream a = (a, Stream a) | | which is equivalent to the type (a, (a, (a, ...))). I have some reservations about this. As things stand, a type synonym is just an abbreviation for another type. Synonyms are convenient because they allow us to write shorter, more informative types. But, if necessary, we can always eliminate them from a given type expression. The proposed change would give us types that (a) are not currently available to the Haskell 1.2 programmer, and (b) cannot be expressed without using a type synonym. Of course, one argument for dropping the restriction is that the new types introduced would actually be useful to the Haskell 1.3 programmer. However, I find (b) a little worrying. To get some idea for the problems that we're dealing with, consider the definition of a stream of zeros: zeros = (0, zeros) What type should we assign to zeros? A Haskell 1.2 system will reject this because it requires zeros :: t where t = (Int,t), which will not get past the occurs check in the unifier. I suppose that we could allow this if the user (i.e. programmer) gave an explicit type signature such as: zeros :: Stream Int Of course, we can't expect the definition to type check without an explicit type declaration unless a new type synonym is generated just for the purpose. Dropping the restriction on recursive type synonyms gives a poor man's version of recursive types. Maybe we could find a way to support full recursive types instead, using explicit mu's in type expressions? For example, the stream type might become: type Stream a = mu s . (a, s) This seems much more elegant and general. | Guy Steele has since pointed out several compelling examples | where it would be *much* easier not to have the restriction, | and I've encountered a few myself. I've played with some of Guy's examples and found that the recursive types can be avoided for some of the specific applications that he has in mind. I certainly would be interested to see more examples, and I'd also be surprised if they couldn't be dealt with using mu types. | The obvious way to go is for someone to implement it first, to | make sure it's not difficult. Mark Jones, have you tried this | in Gofer yet? In fact, the current version of Gofer actually requires the complete expansion of all type synonyms during static analysis, so recursive type synonyms would be rather difficult for me to implement. On the other hand, I suspect that it may be possible to get somewhere with proper recursive mu types. I've seen several papers about recursive types, but I'm not sure if I know of any work about type inference in the presence of recursive types; perhaps somebody can remind me? Informally, it's easy enough to see how mu types would be produced during unification. I suppose that there could also be an awkward interaction with overloading, but my intuition tells me that this is unlikely. It might also be worth mentioning that constructor classes actually give you a weak form of the mu operator. For example, we can define: data Mu f = In (f (Mu f)) However, with this treatment, you have to write an explicit `In' operation. For example, if I define: zeros = In (0, zeros) then the type inference system tells me that zeros :: Mu ((,) Int). What we'd really like is for the `coercion' from f (Mu f) to Mu f to be inferred automatically. (And, with recent discussions in mind, I'd also like the In constructor to be strict, or defined by Phil's newtype construct, so that it is actually an isomorphism ...) Looking forward to further comments, Mark
Re: Arrays and Assoc
>But I think we can have the cake and eat it too, if we get rid of the >restriction (which I never liked) that operators beginning with : must be a >constructor: just define >a := b = (a,b) Unfortunately that won't work if := had been used in patterns. I think backward compatibility is an issue. The standard technique of supporting Assoc but with compiler warnings will probably have to be used. --- >I'm not exactly sure what you mean here. It is allready possible to define >arrays by self-reference in Haskell. Haskell arrays are strict in the indices. That is, the whole of the defining list is consumed and the indices examined before the array becomes available. Thus, a recursive array definition in which the *index calculation* depends on the earlier part of the array gives bottom. The current definition allows for a recursive definition so long as it is only the values of the array elements which depend on the array. This is not always sufficient. --- >Let me just remind people what the LML arrays does: > >example: >lmlarray 1 3 f list = >array [ 1:= f [ x | (1,x) <- list], >2:= f [ x | (2,x) <- list], >3:= f [ x | (3,x) <- list] > ] >where array is like the ordinary Haskell array constructor function. > ... >It seems to me that it is a bit more general to apply f to the entire >list accumulated at each index, rather than as an operator for foldr. If you want the list you can supply (:) and []. If not, you supply the operations, and the intermediate list never gets built. John.
Re: Recursive type synonyms
Phil writes, | While we are proposing things, here's a further suggestion. | The great thing about this suggestion is that it only *removes* | a restriction, and makes the language definition simpler. | It is fully backward compatible. | | The suggestion is: | | Remove the restriction that type synonym | declarations must not be recursive. | | In other words, one could write things like | | type Stream a = (a, Stream a) | | which is equivalent to the type (a, (a, (a, ...))). Hear, hear! I've also run across a need for this: nil f g = f cons x xs f g = g x xs fold f z xs = xs z (\x xs -> f x (fold f z xs)) fold doesn't type, but this would do the trick: type List a b = b -> (a -> List a b -> b) -> b nil :: List a b cons :: a -> List a b -> List a b (If you like existential types, replace each "List a b" above by "List a".) As it is, the closest I can come is data List a b = List (b -> (a -> List a b -> b) -> b) nil = List const cons x xs = List (\f g -> g x xs) fold f z (List xs) = xs z (\x xs -> f x (fold f z xs)) This is particularly bothersome, given than List is lifted. ;-) --Joe
Re: Arrays and Assoc
John Launchbury says: > 1. We should get rid of Assoc. > > When explaining my programs to other people I find this is a point of > confusion. Imagine exaplaining array construction, "When I define an array, > the comprehension produces a list of index/value pairs, only they are not > written as pairs--these's this special type called Assoc. Oh, and don't be > confused by :=. That's not assignment. It is an infix pairing operator." > All of this is entirely unnecessary. Pairs have been used in maths for > decades to represent exactly this sort of thing. I simply do not believe > that [Assoc a b] provides me with any better information than [(a,b)]. > Worse, I often find myself having to redefine standard pair functions on > elements of Assoc. I agree. If I recall correctly, the := to be used in array comprehensions was a consession to the FORTRAN/Id/Sisal community, so that array comprehensions would look more like they were used to. But := is a bit unintuitive if you're thinking e.g. FORTRAN: a = array[1 := 2, 2 := 4] does *not* mean 1 is assigned to 2, etc! But I think we can have the cake and eat it too, if we get rid of the restriction (which I never liked) that operators beginning with : must be a constructor: just define a := b = (a,b) [ While I'm at it: we should also get rid of the lower/uppercase restrictions on constructor/nonconstructor names. ] > 2. Arrays should be lazier. > > I'm expecting Lennart to agree with me here as LML has the Right Thing. I > am convinced that there is no semantic problem with this, and I think that > even Simon isn't horrified at the implementation implications. The ability > to define arrays by self reference is just as important as it is for lists. I'm not exactly sure what you mean here. It is allready possible to define arrays by self-reference in Haskell. > I am assuming that the fact that lazy indexes provide a better match with > laziness elsewhere is clear, but I am willing to expand on this point if > someone wants. > > 3. AccumArray should mimic foldr, not foldl. > > This is tied up with the last point. The only advantage I can see with the > present scheme would be if the array element could be used as the > accumulator while the array was under construction. However, as arrays are > non-strict in their *elements* this seems to be of no benefit. It seems to > me highly sensible that the structure of the computation at each point > should reflect the structure of the input sequence (i.e. the elements are > in the same order). Furthermore, if a lazy operation is used (such as (:)) > then the result becomes available early (assuming point 2. above). > Again I wholeheartedly agree. Let me just remind people what the LML arrays does: example: lmlarray 1 3 f list = array [ 1:= f [ x | (1,x) <- list], 2:= f [ x | (2,x) <- list], 3:= f [ x | (3,x) <- list] ] where array is like the ordinary Haskell array constructor function. In the implementation, the filtering needs to be done only once and not n times, where n is the size of the array. [ If anyone wants to know how this is done, I could expand on this. ] It seems to me that it is a bit more general to apply f to the entire list accumulated at each index, rather than as an operator for foldr. -- Thomas
Re: ADTs and strictness
> I thought this inequality was one of the distinguishing characteristics of > lazy functional programming relative to the standard lambda-calculus. To > quote from Abramsky's contribution to "Research Topics in Functional > Programming", Addison-Wesley 1990: > >Let O == (\x.xx)(\x.xx) be the standard unsolvable term. Then > >\x.O = O > >in the standard theory, since \x.O is also unsolvable; but \x.O is in >weak head normal form and hence should be distinguished from O in our >"lazy" theory. Yes, internally \x.O != O, but since the only thing you can do with a function is to apply it these two are observationally equivalent. Adding seq (or strict constructors) would constitute another way of using function (checking for _|_) and would thus distinguish them. I think this is all right, but it makes eta conversion invalid. -- Lennart
Re: Arrays and Assoc
> 1. We should get rid of Assoc. I agree wholeheartedly! Do we have tp consider backwards compat? > 2. Arrays should be lazier. I agree again. But I think both kinds should be provided. > 3. AccumArray should mimic foldr, not foldl. Right! -- Lennart
ADTs and strictness
I have been following this discussion with interest and I'd like some clarification. Wadler writes: > But just because they call it `lazy' doesn't mean that it really is > the essence of laziness. What is really been called `lazy' and how is the `essence of laziness' defined? Also, forgive my ignorance, but what does it mean that 'products or functions are lifted'? Thanks, Sergio Antoy Dept. of Computer Science Portland State University P.O.Box 751 Portland, OR 97207 voice +1 (503) 725-3009 fax +1 (503) 725-3211 internet [EMAIL PROTECTED]
Re: ADTs and strictness
> So, as Lennart says, if we allow constructors to be strict in functions > then we have to change the semantics to distinguish _|_ from (\x -> _|_). > I, for one, am deeply reluctant to do so; I certainly have no good handle on > the consequences of doing so. Does anyone else? I thought this inequality was one of the distinguishing characteristics of lazy functional programming relative to the standard lambda-calculus. To quote from Abramsky's contribution to "Research Topics in Functional Programming", Addison-Wesley 1990: Let O == (\x.xx)(\x.xx) be the standard unsolvable term. Then \x.O = O in the standard theory, since \x.O is also unsolvable; but \x.O is in weak head normal form and hence should be distinguished from O in our "lazy" theory. Gerald
Arrays and Assoc
Here are three comments directed particularly at Haskell 1.3 people, but obviously open to general feedback. 1. We should get rid of Assoc. When explaining my programs to other people I find this is a point of confusion. Imagine exaplaining array construction, "When I define an array, the comprehension produces a list of index/value pairs, only they are not written as pairs--these's this special type called Assoc. Oh, and don't be confused by :=. That's not assignment. It is an infix pairing operator." All of this is entirely unnecessary. Pairs have been used in maths for decades to represent exactly this sort of thing. I simply do not believe that [Assoc a b] provides me with any better information than [(a,b)]. Worse, I often find myself having to redefine standard pair functions on elements of Assoc. 2. Arrays should be lazier. I'm expecting Lennart to agree with me here as LML has the Right Thing. I am convinced that there is no semantic problem with this, and I think that even Simon isn't horrified at the implementation implications. The ability to define arrays by self reference is just as important as it is for lists. I am assuming that the fact that lazy indexes provide a better match with laziness elsewhere is clear, but I am willing to expand on this point if someone wants. 3. AccumArray should mimic foldr, not foldl. This is tied up with the last point. The only advantage I can see with the present scheme would be if the array element could be used as the accumulator while the array was under construction. However, as arrays are non-strict in their *elements* this seems to be of no benefit. It seems to me highly sensible that the structure of the computation at each point should reflect the structure of the input sequence (i.e. the elements are in the same order). Furthermore, if a lazy operation is used (such as (:)) then the result becomes available early (assuming point 2. above). John.
Type signatures
Folks, Warren Burton makes what appears to me to be a Jolly Sensible suggestion about the syntax of type signatures. Haskell already has many dual ways of doing things (let/where, case/pattern-matching). Warren proposes an alternative syntax for type signatures. Simon --- Forwarded Message Date:Fri, 01 Oct 93 11:30:10 -0800 From:Warren Burton <[EMAIL PROTECTED]> To: [EMAIL PROTECTED] cc: [EMAIL PROTECTED] Subject: Re: ADTs in Haskell Simon, I agree with your comments about ADTs in Haskell. However, your comments brought to mind another question. Do you know why Haskell allows > f a b c = exp which almost means the same thing as > f = \a -> \b -> \c -> exp (ignoring the monomorphism restriction), but does not allow > f Int Char (Stk Thing) :: [Thing] for > f :: Int -> Char -> Stk Thing -> [Thing] When teaching functional programming I always find the > f :: Int -> Char -> Stk Thing -> [Thing] form confusing for students, particularly when the function is defined using the > f a c b = exp form. [..omitted...] --- End of Forwarded Message
Lifted products
I don't like Phil's suggestion to have non-lifted products: * It messes up the uniform semantics for algebraic data types (all lifted). For example a) You have to explain that f ~(z,a) = ... is the same as f (z,a) = ... but g ~(z:a) = ... is NOT the same as f (z:a) = ... b) You have to explain that if f (Foo y) = ... then f is strict if Foo is one of a multi-constructor data type, but non-strict otherwise (unless "..." is strict in y!) (Unless non-lifted products are a different construct, which complicates the language, and * An alternative is, I suppose, to have both standard, lifted algebraic data types, and a new form of data construction, namely non-lifted tuples. I like it not! * Lennart says that if the non-lifted products can also have strictness annotations then it requires parallel evaluation. I think it's rather amazing that one can implement non-lifted products without parallelism; doing so in the presence of strictness annoations makes my head hurt. I bet Lennart is right. Efficiency was not the only reason for having lifted tuples; semantic uniformity was a major one. Incidentally, a much less invasive way to achieve what Phil wants would be to say that there's a ~ stuck on every pattern from a single-constructor data type (or built-in tuple type?). Myself, I'd dislike this, esp if there was no way to "undo" it and recover strict matching, but it solves Phil's problem without adding new data types. Simon
re ADTs etc.
I think there is another problem with having strict constructors. It messes up parametricity even more than fix does. There are two reasons why this would be a shame: * Parametricity is cool. It lets you prove lots of interesting theorems for very little work. These theorems help with program transformation and the like. * Some compilers use parametricity. In particular, the justification for cheap deforestation method (foldr-build) comes from parametricity. If parametricity is weakened too much the transformation may become unsafe. One way to introduce strictness is to use overloading and have a class Strict with an operation strict : a -> a defined for each type in the class (not including functions unless their semantics changes, nor unlifted products if they get introduced). Then a strict constructor would have a class restriction and these would provide the standard mediation for parametricity. John.
ADTs and strictness
(This message assumes we head for the strictness-annotation-on-constructor-arg solution. I'll respond to Phil's comments in my next msg.) The problem with polymorphic strictness ~~~ John asks what the problem is with strict constructor args. As Lennart and Kevin say, the problem only really arises with function types; for example data Foo = MkFoo !(Int -> Int) Operationally the idea is that you evaluate the function before building the constructor. That places some new constraints on implementations, but I suspect it can always be done. More seriously, as Lennart says, Haskell says that _|_ = (\x -> _|_). Now, there is no way to find out whether the function given as an argument to MkFoo is a function which always returns bottom. Consider case MkFoo (\x -> a complicated calculation involving x, which always fails to terminate)of MkFoo f -> 0 If the implementation just "evaluates the function" and then wraps it in a MkFoo, then the result of this expression is just 0. But if _|_ = (\x -> _|_), and MkFoo really is strict, then the result should be _|_. So, as Lennart says, if we allow constructors to be strict in functions then we have to change the semantics to distinguis _|_ from (\x -> _|_). I, for one, am deeply reluctant to do so; I certainly have no good handle on the consequences of doing so. Does anyone else? The problem shows up if a constructor is strict in a polymoprhic position: data Baz a = MkBaz !a !a (consider Baz (Int -> Int)) All this applies equally to polymorphic seq too, of course. An alternative ~~ We already have a good mechanism for dealing with problems like this; it's called overloading. Suppose we had a class class Data a where seq :: a -> b -> b -- Other things too? There would be an instance for class Data on every algebraic data type, automatically derived. Then we could write data Data a => Baz a = MkBaz !a !a and everything is fine, because now Baz can only be applied to data types, not functions. And we get seq too. The annotation in the MkBaz can be explained by translation to seq. Implementations are free to implement seq with a single batch of polymorphic code if they want, of course. Ain't that easy? The only tiresome thing is having to write Data a => in places where you want a strictness annotation on a polymorphic constructor arg. But I don't mind that one bit. The only infelicity is that in the special case of single constructors with a single strict arg (ie the kind we need for ADTs) there is no need for the arg to be in class data: data Abstract a = MkAbstract !a is perfectly ok semantically and pragmatically. I suppose one could allow the (Data a =>) constraint to be omitted in this special case. Or give a different syntax for ADT decls, as I suggested before. Simon
Arrays and Assoc
John Launchbury makes the suggestion, inter alia, that Haskell 1.3 `should get rid of Assoc.' Reading some of the followup messages, I see that there is some division on this point. Those closer to the scientific applications community, such as Nikhil and Joe Fasel's acquaintances, seem to be warmed by the familiar sight of `:=', whereas the more pure-mathematically motivated commentators seem to find the (assuredly equivalent) pair constructor more congenial. There have also been some noises about compatibility, since adopting John's suggestion will definitely stop old code dead in its tracks (namely, in the type-checker). Clearly, what's needed to satisfy all parties and make Haskell 1.3 the rousing success that it deserves to be is to introduce a class `Associator' with methods `key', `image', `associate', `toPair', `toAssoc'. Then the array prelude functions could be redefined in terms of the class by (1) pattern-matching on `toAssoc assoc' instead of `assoc' for each variable assoc :: Assoc, and (2) replacing explicit applications of the constructor `:=' by `associate'. I don't think user code would have to change, but users might wonder about the new inferred type constraints on their array code. Of course, to recover efficiency, all Haskell implementors will have to treat the class `Associator' specially so that no dictionary usage is actually produced (as long as the users haven't perversely introduced their own instances, which suggests some wondrous new interpretations of the concept `array'). I intended this message to be humorous when I started, but I'm beginning to think this is a reasonable approach to such matters. So let's generalize with wild abandon: what would be the consequences of automatically deriving an class abstraction for _every_ Haskell data type? Even function types are eligible via the abstract operation `apply'. What new vistas now unfold? - Dan Rabin I must Create a System Department of Computer Scienceor be enslav'd by another Man's. P.O. Box 208285 I will not Reason & Compare: New Haven, CT 06520-8285 my business is to Create. [EMAIL PROTECTED] -- William Blake, `Jerusalem' -
Re: Arrays and Assoc
John Launchbury says, | Here are three comments directed particularly at Haskell 1.3 people, but | obviously open to general feedback. | | 1. We should get rid of Assoc. | | When explaining my programs to other people I find this is a point of | confusion. Imagine exaplaining array construction, "When I define an array, | the comprehension produces a list of index/value pairs, only they are not | written as pairs--these's this special type called Assoc. Oh, and don't be | confused by :=. That's not assignment. It is an infix pairing operator." | All of this is entirely unnecessary. Pairs have been used in maths for | decades to represent exactly this sort of thing. I simply do not believe | that [Assoc a b] provides me with any better information than [(a,b)]. | Worse, I often find myself having to redefine standard pair functions on | elements of Assoc. Mea maxima culpa. I must admit that the reason for introducing Assoc was syntactic. Making a semantic distinction between pairs and assocs for a syntactic purpose should have set off alarms; somehow, I managed to ignore them. At the time this decision was made, arrays and array syntax were something of a contentious issue. Even the use of infix ! for indexing was a source of anguish for potential users of arrays, and the fear was that pair syntax in "array comprehensions" would be unwieldy, particularly for multidimensional arrays. Consider a matrix of pairs (a typical construction in scientific mesh algorithms). Lennart asks whether we should be concerned about an upward compatibility problem. Thomas suggests that we could drop the syntactic restrictions on constructor and nonconstructor symbols and define (:=) as a pairing function. That almost does the job, but there are some programs that pattern-match Assocs. Also, I think there will be objection in some quarters to dropping the separation of name spaces. Here are two more possibilities: 2. Provide a way to declare synonyms for constructors, and use it to equate := with (,). 3. Don't provide such a general facility, but hack in := as a special case (rather like prefix minus). | 2. Arrays should be lazier. | | I'm expecting Lennart to agree with me here as LML has the Right Thing. I | am convinced that there is no semantic problem with this, and I think that | even Simon isn't horrified at the implementation implications. The ability | to define arrays by self reference is just as important as it is for lists. | I am assuming that the fact that lazy indexes provide a better match with | laziness elsewhere is clear, but I am willing to expand on this point if | someone wants. I agree, but I also agree with Lennart that both sorts of arrays are needed. The historical context again: Accumulators had been added to Id because too many scientific programs couldn't live without them (or else effects). Pragmatically, the accumulations in these programs were almost always sums. (histogramming, Monte Carlo tallying) People needed to be convinced that this could be done efficiently. | 3. AccumArray should mimic foldr, not foldl. | | This is tied up with the last point. The only advantage I can see with the | present scheme would be if the array element could be used as the | accumulator while the array was under construction. However, as arrays are | non-strict in their *elements* this seems to be of no benefit. It seems to | me highly sensible that the structure of the computation at each point | should reflect the structure of the input sequence (i.e. the elements are | in the same order). Furthermore, if a lazy operation is used (such as (:)) | then the result becomes available early (assuming point 2. above). | | John. | Agreed again. The historical reason for the choice of foldl should be evident from the remarks above. Since all of these decisions had to do with Id arrays, I'm pleased to hear from Nikhil that pH people are thinking along the same lines as John and Lennart. Consensus! --Joe
Re: re. Arrays and Assoc
Nikhil says, | Thomas Johnsson says: | | >If I recall correctly, the := to be used in array comprehensions was a | >consession to the FORTRAN/Id/Sisal community, so that array comprehensions | >would look more like they were used to. | | Both Arvind and I think this is notation is awful, and I don't recall | either of us ASKING for it, so this was probably someone else's idea | of a ``concession'' to the Id community! | | Nikhil All right! I'm sorry! ;-) As I recall, Nikhil is right that neither he nor Arvind asked for this. Some scientific programmers of my acquaintance did, though. Id uses = for this purpose, together with square brackets around the index. This, of course, was not possible for Haskell. The motivation was not so much a "concession" to the Id community, as a concern for the readability of [((i,j), (f i j, g i j)) | versus [(i,j) := (f i j, g i j) | or Id's {matrix (1,N),(1,N) | [i,j] = (f i j, g i j) || (if I have that somewhere close to right). The use of := for pairing (or if you like, binding, or single-assignment) rather that assignment did have a precedent in Val and Sisal. All this syntax may seem of little consequence now, but at the time, there was a genuine concern about the unpalatability of some choices of syntax to a large community of programmers. --Joe
re. Arrays and Assoc
Thomas Johnsson says: >If I recall correctly, the := to be used in array comprehensions was a >consession to the FORTRAN/Id/Sisal community, so that array comprehensions >would look more like they were used to. Both Arvind and I think this is notation is awful, and I don't recall either of us ASKING for it, so this was probably someone else's idea of a ``concession'' to the Id community! Nikhil
re. Arrays and Assoc
Two of John Launchbury's suggestions for Haskell 1.3 would mesh well with the pH (parallel Haskell) effort: >1. We should get rid of Assoc. > >When explaining my programs to other people I find this is a point of >confusion. Imagine exaplaining array construction, "When I define an array, >the comprehension produces a list of index/value pairs, only they are not >written as pairs--these's this special type called Assoc. Oh, and don't be >confused by :=. That's not assignment. It is an infix pairing operator." >All of this is entirely unnecessary. Pairs have been used in maths for >decades to represent exactly this sort of thing. I simply do not believe >that [Assoc a b] provides me with any better information than [(a,b)]. >Worse, I often find myself having to redefine standard pair functions on >elements of Assoc. In designing pH, we have been ``anguished'' by the fact that := had already been used for an unnecessary and unintuitive purpose. I agree that Assoc is just a point of confusion and one should use ordinary pairs instead. >2. Arrays should be lazier. > >I'm expecting Lennart to agree with me here as LML has the Right Thing. I >am convinced that there is no semantic problem with this, and I think that >even Simon isn't horrified at the implementation implications. The ability >to define arrays by self reference is just as important as it is for lists. >I am assuming that the fact that lazy indexes provide a better match with >laziness elsewhere is clear, but I am willing to expand on this point if >someone wants. In designing pH, we were going to adopt the lazier semantics and depart from Haskell semantics; this suggestion would bring them back together. Nikhil