Paul Cantrell
http://innig.net
Email: username "cantrell", domain name "pobox.com"
I recommend executing this file, then reading it alongside its output.
Alteratively, you can give yourself a sort of Ruby test by deleting all the comments, then trying to guess the output of the code!
A closure is a block of code which meets three criteria:
- It can be passed around as a value and
- executed on demand by anyone who has that value, at which time
- it can refer to variables from the context in which it was created (i.e. it is closed with respect to variable access, in the mathematical sense of the word "closed").
(The word "closure" actually has an imprecise meaning, and some people don't think that criterion 1 is part of the definition. I think it is.)
Closures are a mainstay of functional languages, but are present in many other languages as well (e.g. Java's anonymous inner classes). You can do cool stuff with them: they allow deferred execution, and some elegant tricks of style.
Ruby is based on the "principle of least surprise," but I had a really nasty surprise in my learning process. When I understood what methods like "each" were doing, I thought, "Aha! Ruby has closures!" But then I found out that a function can't accept multiple blocks -- violating the principle that closures can be passed around freely as values.
This document details what I learned in my quest to figure out what the deal is.
def example(num)
puts
puts "------ Example {num} ------"
end
Blocks are like closures, because they can refer to variables from their defining context:
def thrice
yield
yield
yield
end
x = 5
puts "value of x before: {x}"
thrice { x += 1 }
puts "value of x after: {x}"
A block refers to variables in the context it was defined, not the context in which it is called:
def thrice_with_local_x
x = 100
yield
yield
yield
puts "value of x at end of thrice_with_local_x: {x}"
end
x = 5
thrice_with_local_x { x += 1 }
puts "value of outer x after: {x}"
A block only refers to existing variables in the outer context; if they don't exist in the outer, a block won't create them there:
thrice do # note that {...} and do...end are completely equivalent
y = 10
puts "Is y defined inside the block where it is first set?"
puts "Yes." if defined? y
end
puts "Is y defined in the outer context after being set in the block?"
puts "No!" unless defined? y
OK, so blocks seem to be like closures: they are closed with respect to variables defined in the context where they were created, regardless of the context in which they're called.
But they're not quite closures as we've been using them, because we have no way to pass them around: "yield" can only refer to the block passed to the method it's in.
We can pass a block on down the chain, however, using &:
def six_times(&block)
thrice(&block)
thrice(&block)
end
x = 4
six_times { x += 10 }
puts "value of x after: {x}"
So do we have closures? Not quite! We can't hold on to a &block and call it later at an arbitrary time; it doesn't work. This, for example, will not compile:
def save_block_for_later(&block)
saved = █
end
But we can pass it around if we use drop the &, and use block.call(...) instead of yield:
def save_for_later(&b)
@saved = b # Note: no ampersand! This turns a block into a closure of sorts.
end
save_for_later { puts "Hello!" }
puts "Deferred execution of a block:"
@saved.call
@saved.call
But wait! We can't pass multiple blocks to a function! As it turns out, there can be only zero or one &block_params to a function, and the ¶m must be the last in the list.
None of these will compile:
def f(&block1, &block2) ...
def f(&block1, arg_after_block) ...
f { puts "block1" } { puts "block2" }
What the heck?
I claim this single-block limitation violates the "principle of least surprise." The reasons for it have to do with ease of C implementation, not semantics.
So: are we screwed for ever doing anything robust and interesting with closures?
Actually, no. When we pass a block ¶m, then refer to that param without the ampersand, that is secretly a synonym for Proc.new(¶m):
def save_for_later(&b)
@saved = Proc.new(&b) # same as: @saved = b
end
save_for_later { puts "Hello again!" }
puts "Deferred execution of a Proc works just the same with Proc.new:"
@saved.call
We can define a Proc on the spot, no need for the ¶m:
@saved_proc_new = Proc.new { puts "I'm declared on the spot with Proc.new." }
puts "Deferred execution of a Proc works just the same with ad-hoc Proc.new:"
@saved_proc_new.call
Behold! A true closure!
But wait, there's more.... Ruby has a whole bunch of things that seem to behave like closures, and can be called with .call:
@saved_proc_new = Proc.new { puts "I'm declared with Proc.new." }
@saved_proc = proc { puts "I'm declared with proc." }
@saved_lambda = lambda { puts "I'm declared with lambda." }
def some_method
puts "I'm declared as a method."
end
@method_as_closure = method(:some_method)
puts "Here are four superficially identical forms of deferred execution:"
@saved_proc_new.call
@saved_proc.call
@saved_lambda.call
@method_as_closure.call
So in fact, there are no less than seven -- count 'em, SEVEN -- different closure-like constructs in Ruby:
- block (implicitly passed, called with yield)
- block (&b => f(&b) => yield)
- block (&b => b.call)
- Proc.new
- proc
- lambda
- method
Though they all look different, some of these are secretly identical, as we'll see shortly.
We already know that (1) and (2) are not really closures -- and they are, in fact, exactly the same thing. Numbers 3-7 all seem to be identical. Are they just different syntaxes for identical semantics?
No, they aren't! One of the distinguishing features has to do with what "return" does.
Consider first this example of several different closure-like things without a return statement. They all behave identically:
def f(closure)
puts
puts "About to call closure"
result = closure.call
puts "Closure returned: {result}"
"Value from f"
end
puts "f returned: " + f(Proc.new { "Value from Proc.new" })
puts "f returned: " + f(proc { "Value from proc" })
puts "f returned: " + f(lambda { "Value from lambda" })
def another_method
"Value from method"
end
puts "f returned: " + f(method(:another_method))
But put in a "return," and all hell breaks loose!
begin f(Proc.new { return "Value from Proc.new" }) rescue Exception => e puts "Failed with {e.class}: {e}" end
The call fails because that "return" needs to be inside a function, and a Proc isn't really quite a full-fledged function:
example 11
def g result = f(Proc.new { return "Value from Proc.new" }) puts "f returned: " + result never executed "Value from g" never executed end
puts "g returned: {g}"
Note that the return inside the "Proc.new" didn't just return from the Proc -- it returned all the way out of g, bypassing not only the rest of g but the rest of f as well! It worked almost like an exception.
This means that it's not possible to call a Proc containing a "return" when the creating context no longer exists:
example 12
def make_proc_new begin Proc.new { return "Value from Proc.new" } this "return" will return from make_proc_new ensure puts "make_proc_new exited" end end
begin puts make_proc_new.call rescue Exception => e puts "Failed with {e.class}: {e}" end
(Note that this makes it unsafe to pass Procs across threads.)
A Proc.new, then, is not quite truly closed: it depends in the creating context still existing, because the "return" is tied to that context.
Not so for lambda:
example 13
def g result = f(lambda { return "Value from lambda" }) puts "f returned: " + result "Value from g" end
puts "g returned: {g}"
And yes, you can call a lambda even when the creating context is gone:
example 14
def make_lambda begin lambda { return "Value from lambda" } ensure puts "make_lambda exited" end end
puts make_lambda.call
Inside a lambda, a return statement only returns from the lambda, and flow continues normally. So a lambda is like a function unto itself, whereas a Proc remains dependent on the control flow of its caller.
A lambda, therefore, is Ruby's true closure.
As it turns out, "proc" is a synonym for either "Proc.new" or "lambda." Anybody want to guess which one? (Hint: "Proc" in lowercase is "proc.")
example 15
def g result = f(proc { return "Value from proc" }) puts "f returned: " + result "Value from g" end
puts "g returned: {g}"
Hah. Fooled you.
The answer: Ruby changed its mind. If you're using Ruby 1.8, it's a synonym for "lambda." That's surprising (and also ridiculous); somebody figured this out, so in 1.9, it's a synonym for Proc.new. Go figure.
I'll spare you the rest of the experiments, and give you the behavior of all 7 cases:
"return" returns from caller:
1. block (called with yield)
2. block (&b => f(&b) => yield)
3. block (&b => b.call)
4. Proc.new
5. proc in 1.9
"return" only returns from closure:
5. proc in 1.8
6. lambda
7. method
---------------------------- Section 4: Closures and Arity ----------------------------
The other major distinguishing of different kinds of Ruby closures is how they handle mismatched arity -- in other words, the wrong number of arguments.
In addition to "call," every closure has an "arity" method which returns the number of expected arguments:
example 16
puts "One-arg lambda:" puts (lambda {|x|}.arity) puts "Three-arg lambda:" puts (lambda {|x,y,z|}.arity)
...well, sort of:
puts "No-args lambda: " puts (lambda {}.arity) This behavior is also subject to change in 1.9. puts "Varargs lambda: " puts (lambda {|*args|}.arity)
Watch what happens when we call these with the wrong number of arguments:
example 17
def call_with_too_many_args(closure) begin puts "closure arity: {closure.arity}" closure.call(1,2,3,4,5,6) puts "Too many args worked" rescue Exception => e puts "Too many args threw exception {e.class}: {e}" end end
def two_arg_method(x,y) end
puts; puts "Proc.new:"; call_with_too_many_args(Proc.new {|x,y|}) puts; puts "proc:" ; call_with_too_many_args(proc {|x,y|}) puts; puts "lambda:" ; call_with_too_many_args(lambda {|x,y|}) puts; puts "Method:" ; call_with_too_many_args(method(:two_arg_method))
def call_with_too_few_args(closure) begin puts "closure arity: {closure.arity}" closure.call() puts "Too few args worked" rescue Exception => e puts "Too few args threw exception {e.class}: {e}" end end
puts; puts "Proc.new:"; call_with_too_few_args(Proc.new {|x,y|}) puts; puts "proc:" ; call_with_too_few_args(proc {|x,y|}) puts; puts "lambda:" ; call_with_too_few_args(lambda {|x,y|}) puts; puts "Method:" ; call_with_too_few_args(method(:two_arg_method))
Yet oddly, the behavior for one-argument closures is different....
example 18
def one_arg_method(x) end
puts; puts "Proc.new:"; call_with_too_many_args(Proc.new {|x|}) puts; puts "proc:" ; call_with_too_many_args(proc {|x|}) puts; puts "lambda:" ; call_with_too_many_args(lambda {|x|}) puts; puts "Method:" ; call_with_too_many_args(method(:one_arg_method)) puts; puts "Proc.new:"; call_with_too_few_args(Proc.new {|x|}) puts; puts "proc:" ; call_with_too_few_args(proc {|x|}) puts; puts "lambda:" ; call_with_too_few_args(lambda {|x|}) puts; puts "Method:" ; call_with_too_few_args(method(:one_arg_method))
Yet when there are no args...
example 19
def no_arg_method end
puts; puts "Proc.new:"; call_with_too_many_args(Proc.new {||}) puts; puts "proc:" ; call_with_too_many_args(proc {||}) puts; puts "lambda:" ; call_with_too_many_args(lambda {||}) puts; puts "Method:" ; call_with_too_many_args(method(:no_arg_method))
For no good reason that I can see, Proc.new, proc and lambda treat a single argument as a special case; only a method enforces arity in all cases. Principle of least surprise my ass.
---------------------------- Section 5: Rant ----------------------------
This is quite a dizzing array of syntactic options, with subtle semantics differences that are not at all obvious, and riddled with minor special cases. It's like a big bear trap from programmers who expect the language to just work.
Why are things this way? Because Ruby is:
(1) designed by implementation, and (2) defined by implementation.
The language grows because the Ruby team tacks on cool ideas, without maintaining a real spec apart from CRuby. A spec would make clear the logical structure of the language, and thus help highlight inconsistencies like the ones we've just seen. Instead, these inconsinstencies creep into the language, confuse the crap out of poor souls like me who are trying to learn it, and then get submitted as bug reports. Something as fundamental as the semantics of proc should not get so screwed up that they have to backtrack between releases, for heaven's sake! Yes, I know, language design is hard -- but something like this proc/lambda issue or the arity problem wasn't so hard to get right the first time. Yammer yammer.
---------------------------- Section 6: Summary ----------------------------
So, what's the final verdict on those 7 closure-like entities?
"return" returns from closure
True closure? or declaring context...? Arity check?
--------------- ----------------------------- -------------------
- block (called with yield) N declaring no
- block (&b => f(&b) => yield) N declaring no
- block (&b => b.call) Y except return declaring warn on too few
- Proc.new Y except return declaring warn on too few
- proc <<< alias for lambda in 1.8, Proc.new in 1.9 >>>
- lambda Y closure yes, except arity 1
- method Y closure yes
The things within each of these groups are all semantically identical -- that is, they're different syntaxes for the same thing:
1. block (called with yield)
2. block (&b => f(&b) => yield)
-------
3. block (&b => b.call)
4. Proc.new
5. proc in 1.9
-------
5. proc in 1.8
6. lambda
-------
7. method (may be identical to lambda with changes to arity checking in 1.9)
Or at least, this is how I think it is, based on experiment. There's no authoritative answer other than testing the CRuby implementation, because there's no real spec -- so there may be other differences I haven't discovered.
The final verdict: Ruby has four types of closures and near-closures, expressible in seven syntactic variants. Not pretty. But you sure sure do cool stuff with them! That's up next....
This concludes the "Ruby sucks" portion of our broadcast; from here on, it will be the "Ruby is awesome" portion.
---------------------------- Section 7: Doing Something Cool with Closures ----------------------------
Let's make a data structure containing all of the Fibonacci numbers. Yes, I said all of them. How is this possible? We'll use closures to do lazy evaluation, so that the computer only calculates as much of the list as we ask for.
To make this work, we're going to use Lisp-style lists: a list is a recursive data structure with two parts: "car," the next element of the list, and "cdr," the remainder of the list.
For example, the list of the first three positive integers is [1,[2,[3]]]. Why? Because:
[1,[2,[3]]] <--- car=1, cdr=[2,[3]] [2,[3]] <--- car=2, cdr=[3] [3] <--- car=3, cdr=nil
Here's a class for traversing such lists:
example 20
class LispyEnumerable include Enumerable
def initialize(tree)
@tree = tree
end
def each
while @tree
car,cdr = @tree
yield car
@tree = cdr
end
end
end
list = [1,[2,[3]]] LispyEnumerable.new(list).each do |x| puts x end
So how to make an infinite list? Instead of making each node in the list a fully built data structure, we'll make it a closure -- and then we won't call that closure until we actually need the value. This applies recursively: the top of the tree is a closure, and its cdr is a closure, and the cdr's cdr is a closure....
example 21
class LazyLispyEnumerable include Enumerable
def initialize(tree)
@tree = tree
end
def each
while @tree
car,cdr = @tree.call <--- @tree is a closure
yield car
@tree = cdr
end
end
end
list = lambda{[1, lambda {[2, lambda {[3]}]}]} same as above, except we wrap each level in a lambda LazyLispyEnumerable.new(list).each do |x| puts x end
example 22
Let's see when each of those blocks gets called: list = lambda do puts "first lambda called" [1, lambda do puts "second lambda called" [2, lambda do puts "third lambda called" [3] end] end] end
puts "List created; about to iterate:" LazyLispyEnumerable.new(list).each do |x| puts x end
Now, because the lambda defers evaluation, we can make an infinite list:
example 23
def fibo(a,b) lambda { [a, fibo(b,a+b)] } <---- this would go into infinite recursion if it weren't in a lambda end
LazyLispyEnumerable.new(fibo(1,1)).each do |x| puts x break if x > 100 we don't actually want to print all of the Fibonaccis! end
This kind of deferred execution is called "lazy evaluation" -- as opposed to the "eager evaluation" we're used to, where we evaluate an expression before passing its value on. (Most languages, including Ruby, use eager evaluation, but there are languages (like Haskell) which use lazy evaluation for everything, by default! Not always performant, but ever so very cool.)
This way of implementing lazy evaluation is terribly clunky! We had to write a separate LazyLispyEnumerable that knows we're passing it a special lazy data structure. How unsatisfying! Wouldn't it be nice of the lazy evaluation were invisible to callers of the lazy object?
As it turns out, we can do this. We'll define a class called "Lazy," which takes a block, turns it into a closure, and holds onto it without immediately calling it. The first time somebody calls a method, we evaluate the closure and then forward the method call on to the closure's result.
class Lazy def initialize(&generator) @generator = generator end
def method_missing(method, *args, &block) evaluate.send(method, *args, &block) end
def evaluate @value = @generator.call unless @value @value end end
def lazy(&b) Lazy.new &b end
This basically allows us to say:
lazy {value}
...and get an object that looks exactly like value -- except that value won't be created until the first method call that touches it. It creates a transparent lazy proxy object. Observe:
example 24
x = lazy do puts "<<< Evaluating lazy value >>>" "lazy value" end
puts "x has now been assigned" puts "About to call one of x's methods:" puts "x.size: {x.size}" <--- .size triggers lazy evaluation puts "x.swapcase: {x.swapcase}"
So now, if we define fibo using lazy instead of lambda, it should magically work with our original LispyEnumerable -- which has no idea it's dealing with a lazy value! Right?
example 25
def fibo(a,b) lazy { [a, fibo(b,a+b)] } end
LispyEnumerable.new(fibo(1,1)).each do |x| puts x end
Oops! That didn't work. What went wrong?
The failure started in this line of LispyEnumerable (though Ruby didn't report the error there):
car,cdr = @tree
Let's zoom in on that result, and see what happened:
example 26
car,cdr = fibo(1,1) puts "car={car} cdr={cdr}"
Here's the problem. When we do this:
x,y = z
...Ruby calls z.respond_to?(to_a) to see if z is an array. If it is, it will do the multiple assignment; if not, it will just assign x=z and set y=nil.
We want our Lazy to forward the respond_to? call to our fibo list. But it doesn't forward it, because we used the method_missing to do the proxying -- and every object implements respond_to? by default, so the method isn't missing! The respond_to? doesn't get forwarded; instead, out Lazy says "No, I don't respond to to_a; thanks for asking." The immediate solution is to forward respond_to? manually:
class Lazy def initialize(&generator) @generator = generator end
def method_missing(method, *args, &block) evaluate.send(method, *args, &block) end
def respond_to?(method) evaluate.respond_to?(method) end
def evaluate @value = @generator.call unless @value @value end end
And now our original Lispy enum can work:
example 27
LispyEnumerable.new(fibo(1,1)).each do |x| puts x break if x > 200 end
Of course, this only fixes the problem for respond_to?, and we have the same problem for every other method of Object. There is a more robust solution -- frightening, but it works -- which is to undefine all the methods of the Lazy when it's created, so that everything gets forwarded.
And guess what? There's already a slick little gem that will do it:
http://moonbase.rydia.net/software/lazy.rb/
Read the source. It's fascinating.
---------------------------- Section 8: Wrap-Up ----------------------------
So sure, this was all entertaining -- but is it good for anything?
Well, suppose you have an object which requires a network or database call to be created, or will use a lot of memory once it exists. And suppose that it may or may not be used, but you don't know at the time it's created whether it will be. Making it lazy will prevent it from consuming resources unless it needs to. Hibernate does this to prevent unnecessary DB queries, and it does it with more or less arbitrary Java objects (i.e. unlike ActiveRecord, it doesn't depend on a base class to do its lazy loading). Ruby can do the same thing, but with a lot less code!
That's just an example. Use your imagination.
If you're a functional langauge geek, and enjoyed seeing Ruby play with these ideas from Lisp and Haskell, you may enjoy this thread:
http://redhanded.hobix.com/inspect/curryingWithArity.html
OK, I'll stop making your brain hurt now. Hope this has been a bit enlightening! The experience of working it out certainly was for me.
Paul