On Wednesday, 16 October 2013 at 20:33:23 UTC, H. S. Teoh wrote:
On Wed, Oct 16, 2013 at 09:45:09PM +0200, Daniel Davidson wrote:
On Wednesday, 16 October 2013 at 19:12:48 UTC, Dicebot wrote:
[...]
>I think any usage of immutable with types/entities not >initially
>designed for immutability is an potential mistake and in that
>sense it is good that change has broken the user code. Same >goes >for operating on immutable entity in generic code as if it is >a
>value type without actually checking it via introspection.

I don't disagree. Now, what does it mean to "initially design for immutability" and what are the guidelines. I thought that was kind of what we were talking about here. How to effectively use const and
immutable.


Thanks - this is most helpful.

I'd say that user code should use const, not immutable (unless the library provides a way of constructing immutable instances of the type, of course), because immutable makes assumptions about implementation details, which should not be known unless you break encapsulation. As a user of a properly encapsulated type, I can't make any guarantees about its immutability; the best I can do is to promise I don't change it
myself -- i.e., const.


The distinction between user code and the rest (non-user code?) is not enough for me to get clear guidance. I am writing user code, the vast majority will be in libraries because I find that helpful. I would hate to have a separate set of rules for my code and the code I use. We should strive for guidelines that work in general - as in this is a good way to do D.

But leaving "user code" aside, I like the argument and the vote for const over immutable.


Immutable should be used in library code to provide strong guarantees to the user: since you're the one responsible for implementing the type, you're in the position to make guarantees about its uniqueness (and
hence, immutability).


I don't see the benefit of separating usage of types, usually via functions where mutability guarantees adorn the types, and implementation guarantees in a world with turtles all the way down. Maybe if those guarantees are totally hidden from user code via encapsulation then the impact of immutable and the difficulties of dealing with it are lessened. But then is the immutable guarantee for the user or just the developer of the encapsulated non-user code?

How about this... here is a plea for ideas on a good use of immutable for a type with potentially mutable aliasing. String I think is a great use of immutable but char will likely never have aliasing introduced to it. At this stage I want the information for educational purposes, because based on this thread and my experimentation - Ali's presentation guidelines aside - I am about to abandon immutable altogether. The fight is too hard and my skills too weakened.


[...]
One general idea that was presented by Ali is that an immutable
parameter means that not only are you guaranteeing that you will not change your data, but also that no one else, even in another thread will. That sounds appealing. After all, who doesn't want the data they are using in a function to not change from underneath them?

I'm actually wary of this view, to be honest. An immutable parameter means you expect your *caller* to provide you with a value that cannot be changed, not by you, nor by anybody else, ever. Sure, it's nice to have, but that imposes a rather high bar on your callers. They have to be responsible to guarantee that whatever they hand to you cannot be changed by anything or anyone else, at any time. If they can do this,
then great; if not, they won't be able to call your function.


I now agree with you on this, especially since it goes with my new guideline of don't use immutable.


Suppose you have highly structured, deeply nested reference data you read from a nosql DB. Surely there is opportunity and some benefit to use immutable? The result of the query is just a read only data
source. But then that data will be consumed - presumably by
functions with either immutable or const parameters. If all
signatures use const then when can you benefit from the fact that the data is really immutable (by the time it gets to a function with
const parm the fact that it really was/is immutable is lost.

I'm of the view that code should only require the minimum of assumptions it needs to actually work. If your code can work with mutable types, then let it take a mutable (unqualified) type. If your code works without modifying input data, then let it take const. Only if your code absolutely will not work correctly unless the data is guaranteed to
never change, ever, by anyone, should it take immutable.


I don't have the instincts yet to really buy this. Perhaps a specific example would be helpful.

I think most of the time `foo(ref const(T) t)` is written such that it is assumed t is never changed during the span of foo and the compiler helps. While it is possible to, in a single thread call out to code that also has a handle to shared state so t could be accidentally or purposely modified elsewhere, it is probably a rare design goal.

Take the query example. A big mongo db query result comes back and is deserialized into a large web of json like data - lists, dictionaries at many levels. Is that a good reason then for `foo(ref immutable(QueryResult) qr)`. Surely while you are working on query result you don't want it to change. I am ambivalent here because immutable sounds appealing. Any iteration over a hash in the data set must be immutable. I think this is the expectation most have with all const or immutable types taken in a function. Sticking with const just makes life easier.


I'm not sure what "benefits" you get from requiring immutable when the code doesn't really need to depend on immutability. It just makes the code harder to use (you have to make sure whatever you pass to it is immutable, and sometimes that's not easy to guarantee, and would require a lot of copying). Immutable only benefits you when the code *can't* work correctly unless the data is guaranteed never to change, ever. Hash table keys come to mind -- if you compute the hash value of the key and use that to determine which slot to put the data into, it would be very bad if somebody else mutated that key via a mutable reference after the fact -- now your AA is broken because the hash value no longer matches the key. By requiring an immutable key, you ensure that this never
happens.


I like the AA example - since it sounds like a good use for immutable.

Note that in D, everything is thread-local by default unless explicitly made shared, so using immutable to guarantee other threads won't mutate
the data isn't really necessary.


T

Reply via email to