Unix famously represents all content as byte sequences. This was a
great step forward, offering a way of representing arbitrary
information without forcing an interpretation on it.
However, it is not enough. Unix is an incomplete design. Supporting
only byte sequences, and nothing else, has caused wasted effort,
code duplication, and bugs.
Text is an obvious example of the problem
Consider just one data type: text. It has a zillion character sets and
encoding schemes. Each application must decide, on its own, which
encoding of which character set is being used for a given file.
When applications get this wrong, both obvious bugs like
Mojibake and subtler flaws
like the
IDN homograph attack
result.
Massive duplication of code and effort
Lack of system support for text yields massive code duplication.
Rather than having a system-wide, comprehensive model of text
representation, encoding, display, input, collation, and comparison,
each programming language and application must fend for itself.
Because it is difficult and time consuming to properly handle text,
developers tend to skimp on text support. Where a weakness is
identified, it must be repaired in each application individually
rather than at the system level. This is itself difficult and time
consuming.
Inconsistent treatment
Finally, dealing only with byte sequences precludes consistent user
interface design.
The user cannot simply work with a file containing text; they must
make a decision about which encoding to use. Woe betide them if they
choose incorrectly.
A consistent approach would separate the question of text encodings
entirely from application-specific UIs. System UI for transcoding
would exist in one place, common to all applications.
User frustration
A tiny fraction of the frustration this kind of thing causes is
recorded in Thunderbird’s
bug 117236.
Notice that it took fourteen years to be fixed.
Ubiquitous problem
This Thunderbird change is just one example. Each and every
application suffers the same problems, and must have its text support
repaired, upgraded, and enhanced independently.
It’s not only a Unix problem. Windows and OS X are just as bad. They,
too, offer no higher-level model than byte sequences to their
applications. Even Android is a missed opportunity.
Learn from the past
Systems like Smalltalk, for all their flaws, offer a
higher-level model to programmers and users alike. In many cases, the
user never need learn about text encoding variations.
Instead, the system can separate text from its encoding.
Where encoding is relevant to the user, there can be a single place
to work with it. Contrast this with the many places encoding leaks
into application UIs today, just one of which is shown in the
Thunderbird example above.
It’s not just text
Text is just one example. Pictures are another. You can probably think
of more.
Our operating systems do not support sharing of high-level
abstractions of data between documents or applications.
An operating system with a mechanism for doing so would take a great
burden off both programmers and users.
Let’s start thinking about what a better modern operating system
would look like.
Programming language designers often need to experiment with syntax
for their new language features. When it comes to Javascript, we rely
on language preprocessors, since altering a Javascript engine
directly is out of the question if we want our experiments to escape
the lab.
Ohm is “a library and domain-specific language for parsing and
pattern matching.” In this post, I’m going to use it as a Javascript
language preprocessor. I’ll build a simple compiler for ES5 extended
with a new kind of for loop, using Ohm and the ES5 grammar
included with it.
All the code in this post is available in a Github repo.
Our toy extension: “for five”
We will add a “for five” statement to ES5, which will let us write
programs like this:
forfiveasx{console.log("We have had",x,"iterations so far");}
The new construct simply runs its body five times in a row, binding a
loop variable in the body. Running the program above through our
compiler produces:
for(varx=0;x<5;x++){console.log("We have had",x,"iterations so far");}
Extending the ES5 grammar
We write our extension to the ES5 grammar in a new file for5.ohm as
follows:
For5 <: ES5 {
IterationStatement += for five as identifier Statement -- for5_named
five = "five" ~identifierPart
as = "as" ~identifierPart
keyword += five
| as
}
Let’s take this a piece at a time. First of all, the declaration For5
<: ES5 tells Ohm that the new grammar should be called For5, and
that it inherits from a grammar called ES5. Next,
IterationStatement += for five as identifier Statement -- for5_named
extends the existing ES5 grammar’s
IterationStatement
nonterminal with a new production that will be called
IterationStatement_for5_named.
Finally, we define two new nonterminals as convenient shorthands for
parsing the two new keywords, and augment the existing
keyword
definition:
five = "five" ~identifierPart
as = "as" ~identifierPart
keyword += five
| as
There are three interesting points to be made about keywords:
First of all, making something a keyword rules it out as an
identifier. In our extended language, writing var five = 5 is a
syntax error. Define new keywords with care!
We make sure to reject input tokens that have our new keywords as a
prefix by defining them as their literal text followed by
anything that cannot be parsed as a part of an identifier,
~identifierPart. That way, the compiler doesn’t get confused by,
say, fivetimes or five_more, which remain valid identifiers.
By making sure to extend keyword, tooling such as syntax
highlighters can automatically take advantage of our extension, if
they are given our extended grammar.
Translating source code using the new grammar
First, require the ohm-js NPM module and its included ES5 grammar:
Next, load our extended grammar from its definition in for5.ohm, and
compile it. When we compile the grammar, we pass in a namespace that
makes the ES5 grammar available under the name our grammar expects,
ES5:
Finally, we define the translation from our extended language to plain
ES5. To do this, we extend a semantic function, modifiedSource,
adding a method for each new production rule. Ohm automatically uses
defaults for rules not mentioned in our extension.
Each parameter to the IterationStatement_for5_named method is a
syntax tree node corresponding positionally to one of the tokens in
the definition of the parsing rule. Accessing the asES5 attribute of
a syntax tree node computes its translated source code. This is done
with recursive calls to the modifiedSource attribute where required.
Our compiler is, at this point, complete. To use it, we need code to
feed it input and print the results:
>compileExtendedSource("for five as x { console.log(x); }");'for (var x = 0; x < 5; x++) { console.log(x); }'
Discussion
This style of syntactic extension is quite coarse-grained: we must
translate whole compilation units at once, and must specify our
extensions separately from the code making use of them. There is no
way of adding a local syntax extension scoped precisely to a block
of code that needs it (known to Schemers as
let-syntax).
For Javascript, sweet.js offers a more Schemely
style of syntax extension than the one explored in this post.
Mention of sweet.js leads me to the thorny topic of
hygiene. Ohm is a
parsing toolkit. It lets you define new concrete syntax, but doesn’t
know anything about scope, or about how you intend to use identifiers.
After all, it can be used for languages that don’t necessarily even
have identifiers. So when we write extensions in the style I’ve
presented here, we must write our translations carefully to avoid
unwanted capture of identifiers. This is a tradeoff: the broad
generality of Ohm’s parsing in exchange for less automation in
identifier handling.
Ohm’s extensible grammars let us extend any part of the language, not
just statements or expressions. We can specify new comment syntax, new
string syntax, new formal argument list syntax, and so on. Because Ohm
is based on
parsing expression grammars,
it offers
scannerless parsing.
Altering or extending a language’s lexical syntax is just as easy as
altering its grammar.
Conclusion
We have defined an Ohm-based compiler for an extension to ES5
syntax, using only a few lines of code. Each new production rule
requires, roughly, one line of grammar definition, and a short method
defining its translation into simpler constructs.
You can try out this little compiler, and maybe experiment with your
own extensions, by cloning its Github repo.
I ran some very quick-and-dirty informal experiments on my Acer C720
laptop, which is running a 64-bit Debian system and has 2GB RAM and a
two-core Celeron 2955U at 1.40GHz.
I measured approximate alist and hash table performance for
fixnum keys, using eq? as the lookup predicate (so assq, hasheq) (fixnum program)
length-64 byte vector keys, using equal? as the lookup predicate (so assoc, hash) (byte-vector program)
Each chart below has four data series:
probe/alist, average time taken to search for a key that is present in an alist
probe/alist/missing, average time taken to search for a key that is not present in an alist
probe/hasheq or probe/hash, average time taken to search for a key that is present in a hash table
probe/hasheq/missing or probe/hasheq/missing, average time taken to search for a key that is not present in a hash table
Fixnum keys
Here are average timings for fixnum keys:
Things to note:
Alists are here always faster than hasheq tables for 7 keys or
fewer, whether the key is present or not.
When the key is present in the lookup table, alists are on
average faster up to around 14 keys or so.
Length-64 byte vector keys
Here are average timings for length-64 random byte vector keys:
Things to note:
Alists are here always faster than hasheq tables for 4 keys or
fewer, whether the key is present or not.
When the key is present in the lookup table, alists are on
average faster up to around 16 keys or so.
Conclusions
Alists will be faster when you have very few keys - for eq?, around
seven or fewer, or for equal?, perhaps only as many as four,
depending on the size of each key.
If you expect with high probability that a given key will be present
in the table, the picture changes slightly: then, alists may be faster
on average up to around perhaps fifteen keys. Specifics of the
insertion order of your keys will naturally be very important in this
case.
gRPC looks very interesting. From a quick
browse of the site, it looks like it differs from CORBA primarily in
that
It is first-order.
It eschews exceptions.
It supports streaming requests and/or responses.
(That’s setting aside differences between protobufs and GIOP.)
It’s the first point that I think is likely to be the big win. Much of
the complexity I saw with CORBA was to do with trying to pass object
(i.e. service endpoint) references back and forth in a transparent way.
Drop that misfeature, and everything from the IDL to the protocol to the
frameworks to the error handling to the implementations of services
themselves will be much simpler.
The way streaming is integrated is interesting too. There’s a clear
separation between (finite) data, including lists/arrays, in the
protobuf message-definition language, and (possibly non-finite) behavior
in the gRPC service-definition language. Streams, being coinductive, fit
naturally in the service-definition part.
From time to time, one stumbles across a website that has gone out of
its way to make it difficult to save images displayed on the page. A
common tactic is to disable right-click for the elements concerned.
From time to time, I make binary snapshot downloads of the RabbitMQ
plugins that I have developed. You will be able to find the most
recent downloads from this page. I sign the downloadable files with
my public key.
Google will split their portion of the global chat network off from
everyone else on the 16th of February.
This balkanising move means that you won’t be able to use a
Google-based chat account to contact people using non-Google chat
account addresses anymore.
If you’re still using a Google-based account for chat, I recommend
getting a chat account with a provider committed to open
interoperability.
You can use almost any chat client to connect; some web-based ones do
exist but it’s probably simpler and less bother to download a chat
client to run on your computer or cellphone directly.
It’s also easy to connect using your own DNS domain
For those of you who have DNS domains of your own, and would like a
yourname@yourdomain.com chat address, I heartily recommend
http://hosted.im/. All you do is add a couple of DNS records,
register at the site, and you have your own personal chat domain. You
can create accounts for yourself and your friends there, and have full
interoperability and federation with the global chat network.
Resist Balkanisation; support independent providers
The XMPP global chat network is the only open, federated chat
network still standing. As such, it deserves our support against the
hostile moves of large companies like Google, Apple, and Microsoft.
Supporting smaller chat account providers and/or running our own
servers is a great way to help keep the internet commons healthy.
In this post, I’ll build a monad
library for Racket, using its
generic interfaces
feature in place of Haskell’s type classes.
There have been lots of demonstrations of monads in dynamically-typed
languages before; notably,
Oleg’s article on monads in Scheme. However, I haven’t
yet (Update:see below) seen one that smoothly
allows for return-type polymorphism.
This post introduces (1) coercions between monad representations and
(2) “undecided” quasi-monadic placeholder values. Taken together,
these give a Haskell-like feel for monads in a dynamically-typed
language.
The techniques are illustrated using Racket, but aren’t specific to
Racket.
The Challenge
How can we write monadic programs like this, where monad types are
implicit, just as they are in comparable Haskell programs?
(run-io(do(mdisplay"Enter a number: ")n<-mreadall-n<-(return(for/list[(in)]i))evens<-(return(doi<-all-n#:guard(even?i)(returni)))(returnevens)))
The overall computation uses do-notation at IO type, while the
evens list uses do-notation at List type.
The problem
Simple monomorphic monads are easy. Here’s the List monad in Racket:
This is a fine definition, as far as it goes, but we’ve just run into
the first problem:
Monads behave differently at different types
Simply defining fail, return and bind as functions won’t work.
We need some kind of dynamic method dispatch. Haskell does this by
dispatching using
dictionary passing.
But there’s a deeper problem, too:
Monads don’t learn their types until it’s too late
How should we choose which dictionary to use? It’s easy with bind,
because the specific monad being bound is always given as bind’s
first argument. The bind implementation can be a method on the
monad’s class.
This doesn’t work for return or fail. When they’re called, they
don’t know what kind of monad they should produce. That only becomes
clear later, when their results flow into a bind.
Haskell uses the type system. We have to do it at runtime, on a
case-by-case basis.
A solution
To address these problems, we explicitly represent the dictionary we
need:
(structmonad-class(failer;; -> (M a)returner;; a -> (M a)binder;; (M a) (a -> (M b)) -> (M b)coercer));; N (M a) -> (N a)
The failer, returner and binder methods correspond to the monad
operations we met above, but coercer is new. It lets us dynamically
reinterpret a monad at a different monadic type. We’ll see its use
below.
The not-coercable function is for use when it’s not possible to
reinterpret the type of a monad, which is the overwhelmingly common
case. It checks to make sure we’re already at the expected type, and
raises an exception if we’re not.
This declares a new generic interface, gen:monad, with a single
method, monad->monad-class. It supplies implementations of
gen:monad for the builtin types null? and pair?, in each case
returning the List monad-class when asked.
Now we can implement bind and coerce, not just for a single monad, but for all
monads:
They use monad->monad-class to find the relevant dictionary, extract
the right method from the dictionary, and delegate to it.
Note that bind dispatches on its first argument, while coerce
dispatches on its second argument: monad instances get to decide how
they are bound and how (or if!) they are to be interpreted at other
monad types.
Next, we introduce neutral quasi-monad instances for when return
and fail don’t know which monad type to use. Because we might bind
such a quasi-monad before it collapses into a particular type, we need
to define a quasi-monad to represent binds in superposition, too.
The Return method dictionary looks very similar to the
implementation of the
identity monad,
with the addition of a return-struct box around the carried value:
The bind implementation is a direct statement of one of the monad
laws,
return x >>= f ≡ f x
The coerce implementation rebuilds the
monad using the return implementation from the target monad.
The Return coerce implementation gives us the equivalent of
return-type polymorphism, letting us avoid sprinkling monad type
annotations throughout our code.
Likewise, coerce for a suspended-bindre-binds its parameters, after first coercing
them into the now-known monad.
You might have noticed a call to coerce in the bind implementation
for List above:
(λ(xsf)(append-map(λ(x)(coerceList(fx)))xs))
This is necessary because f might simply (return 1), which is one
of these neutral quasi-monad instances that needs to be told what kind
of thing it should be before it can be used. On the other hand, f
might have returned (list 1), in which case List’s coercer method
will notice that no conversion needs to be done.
The IO Monad
Racket is an eager, imperative language, and so doesn’t need an IO
monad the way Haskell does. But we can define one anyway, to see what
it would be like.
It’s also useful as a template for implementing unusual control
structures for monadic embedded domain-specific languages.
Values in the IO type denote sequences of delayed actions. We’ll
represent these using two structure types, one for return and one
for bind:
We now have all the pieces we need to run the challenge program given
above.
The challenge comes from the end of Oleg’s article on
monadic programming in Scheme. His example uses many different types
of monad together, using the monomorphic technique he developed:
(IO::run(IO::>>(put-line"Enter a number: ")(IO::>>=(read-int)(λ(n)(IO::>>=(IO::return(for/list[(in)]i))(λ(all-n)(IO::>>=(IO::return(List:>>=all-n(λ(i)(if(even?i)(List::returni)(List::fail"odd")))))(λ(evens)(IO::returnevens)))))))))
With our polymorphic monads, we are able to use the generic
do-notation macro defined above, and write instead
(run-io(do(mdisplay"Enter a number: ")n<-mreadall-n<-(return(for/list[(in)]i))evens<-(return(doi<-all-n#:guard(even?i)(returni)))(returnevens)))
which compares favourably to the Haskell
main=doputStr"Enter a number: "n<-getIntallN<-return[0..n-1]evens<-return$doi<-allNguard$i`mod`2==0returnireturnevens
Implementation
A somewhat fleshed-out implementation of these ideas is available
here:
It uses
prop:procedure,
making monad-classes directly callable and allowing (coerce List
foo) to be written as (List foo).
It uses the
double-dispatch pattern
to let monad instances flexibly adapt themselves to each other in
coercions. For example, lists, streams and sets can all be
considered interconvertible.
It uses Racket’s built-in exception type to represent pending
failures instead of the simple structure sketched above. This way,
failures may carry stack traces and an informative error message.
It includes a simple State monad, with sget and sput
operations.
Related Work
I based this work on some
experiments I ran back in 2006
in writing a monadic library for Scheme. Racket offers much better
support for object-oriented/generic programming (among many other
things!) than vanilla Scheme does, which makes for smoother
integration between my monad library and the rest of the language.
Ben Wolfson pointed me to his
monad library for clojure in a
reddit thread
about this post. It seems to be using a very similar approach, with
“neutral” carriers for monadic values that stay neutral until a later
point in the program. The main difference seems to be that with my
approach, no “run” step is needed for some monads, since monadic
values carry their bind implementation with them. Ben’s library
supplies the bind implementation at “run” time instead.
Update, 20150127: Paul Khuong pointed me to
his Scheme monad implementation
based around
delimited continuations
in
this twitter conversation.
It doesn’t have dynamic-dispatch for automatically determining the
appropriate bind implementation, but I suspect it could be added;
perhaps in his join operator, which makes an interesting contrast to
the technique explored in this post. The really nice thing about the
delimited continuation approach to monads is that the return is
implicit. You never mention return explicitly in monadic code.
That makes a big difference, because return-type polymorphism is only
needed to handle explicit returns! By placing the return right
next to the appropriate run operation, the delimited continuation
technique neatly sidesteps the problem.
Conclusions
Haskell’s return is generic: it injects a value into any monad at
all. Which specific monad is used depends on the context. This is
impossible to directly implement in general without static analysis.
Lacking such static analyses in dynamically-typed languages, achieving
the equivalent of Haskell’s return is challenging, even though
implementing particular monomorphic variations is straightforward.
The trick is to make return produce a kind of superposition which
doesn’t collapse to a specific monad type until it touches a monad
that already knows what type it is.
In this post, I’ve shown how this kind of monad “coercion” works well
to give the kind of return-type polymorphism we need to make monads
feel in Racket similar to the way they do in Haskell.
No implementations are given for their failer and
returner methods, since they are only quasi-monad classes, and
it’s not possible for user code to ever be in the Fail or Bind
“monads”. ↩
“Giving a professional illustrator a goal for a poster usually
results in what was desired. If one tries this with an artist, one
will get what the artist needed to create that day. Sometimes we
make, to have, sometimes to know and express.” — Alan Kay
I want to set up Postfix SMTP authentication so I have two different
passwords to authenticate with: one main one that gives me access to
both SMTP and IMAP, for my own use with trusted clients, and a
separate one that gives me the ability to send messages via SMTP
only, and that will not work for IMAP.
The reason is that gmail will not let me send mail as an auxiliary
mail identity through its own SMTP servers anymore, and requires a
username and password to authenticate itself to my own SMTP servers.
Since I’m unhappy giving gmail my main email password, I decided a
second gmail-specific password would be a good idea.
This turns out to be difficult.
Postfix will authenticate either against Dovecot, or using Cyrus-SASL.
Cyrus-SASL can talk to a number of things, and one of them is a SQL
database, but it won’t let you use crypt for the passwords stored
in the database. That’s a showstopper there. The other alternative is
to back Cyrus-SASL with PAM, but that involves figuring out PAM.
Painful, and another link in the (already long) chain:
Postfix → Cyrus-SASL → PAM → Database.
I’ve asked on the Freenode #cyrus channel about the possibility of
getting a patch for crypted SQL-database passwords accepted, but no
replies yet.
There are a couple of patches out there that get close to what I want;
the best (but still not quite right) is this
one. It needs work to support other crypt
schemes and to extract salt properly.
Perhaps I will give in and learn how to configure PAM…
Update: At the suggestion of the kind people on #cyrus, I’ve
decided instead to set up a separate account on my email server for
gmail to log into, and simply forward any mail that’s delivered there
(by accident, presumably) to my main account.