The tokenizer didn't handle $$, but relied on token pasting of two $ tokens. This broke after the improvements in 9bb46df4.
9bb46df4