Edd Mann Developer

Rewriting the santa-lang interpreter in Rust, Part 2 - Runtimes

With the core language library now written I could go about building the desired (delivery) runtimes. In this article within the series I will delve into my experience integrating the core language library with these chosen runtimes, documenting some of the challenges that I faced along the way.

The TypeScript implementation of the santa-lang interpreter included a CLI, Web and Lambda runtimes. I decided to tackle these in this order of importance, as I required at least the CLI to begin testing the interpreter on real solutions.

CLI

I intentionally decided to write the CLI with minimal package dependencies. There are many useful crates available to help assist in CLI development, but I did not feel like I needed the additional features they provided at this time. I did however use a crate to provide the ability to start up an interactive shell via readline support. The runtime itself included I/O external functions puts and read which had feature parity with the TypeScript implementation. This included POSIX support for reading files from disk, fetching contents from HTTP endpoints, and the AoC schema I had devised when first designing santa-lang. I was able to provide test coverage for this delivery using the assert_cmd crate. This provided me with confidence that I had correctly integrated the library with the delivery.

CLI Runner
CLI REPL

To help aid in performance benchmarking I added the ability to compile the runtime with pprof code profiling (via a profile feature flag). The chosen library also allowed me to produce flame graphs, which were used extensively throughout development to help pinpoint areas where performance could be improved. Further discussion on performance will be documented in the next article within the series.

Web (WASM)

The TypeScript interpreter being transpiled into JavaScript made it already well suited for Web platform usage, with very little additional effort on my part. To provide use of this interpreter within the Web-based code editor and Browsers/Node in-general I was required to leverage WebAssembly (WASM). Having had experience doing this before I was again surprised at how easy it was to achieve. Not only was I able to trivially expose the language as a WASM library, but additionally provide the ability to include user-land JavaScript functions as external functions (similar to the TypeScript implementation), as shown below:

import { evaluate } from '@eddmann/santa-lang-wasm';

evaluate('[1, 2, 3] |> map(_ + 1) |> sum');

evaluate('puts("Hello, world")', { puts: console.log.bind(console) });

The runtime required me to enable the serde feature support which I had included within the core language library, to make it possible to represent compatible data-structures between JavaScript and santa-lang. Being enabled on a per-runtime basis ensured that any additional overheads would not impact other runtimes. The Runner instance used the Time trait with a specific WASM implementation to determine the execution time of solutions. Similar to the CLI, I was able to garner confidence of the desired behaviour by using a testing crate which allowed me to verify correctness of the compiled library within both Node and Chrome-browser environments.

With very little changes I was then able to update the Web-based code editor to use this new interpreter. On top of this, I had begun documenting the language specification and features it had to offer using MkDocs. I decided to additionally add executable examples using this library, which exercised use of the runtime in a different context.

Web Editor

Lambda

I had initially developed the Lambda runtime in TypeScript for use within last year’s Secret Santa Allocation project. I wanted to provide feature parity and implement such a runtime in Rust. To achieve this I borrowed ideas from the Node version, conforming to the Lambda runtime API and expected request/response lifecycle. I used the lightweight ureq crate to handle HTTP transport between the compiled bootstrap and Lambda runtime environment. Due to Lambda’s heavy reliance on JSON for transport, I again enabled the serde feature support (similar to the WASM runtime) that had been added to the core language library to handle converting santa-lang data-structures to and from their JSON equivalent. I was pleasantly surprised how performant this runtime was in practise, with very little overheads and a small binary size.

Lambda Runtime

PHP extension

One evening I spotted an interesting crate which provided an abstraction around creating PHP extensions within Rust. As all I had been thinking about over the past several months was santa-lang, the first thing that came into my mind was could I make a runtime for it?! So I set about creating a simple extension that provided user-land PHP with the ability to evaluate santa-lang code, along with AoC solutions.

I had limited experience in writing PHP extensions in the past and really enjoyed how this crate abstracted away any need of a C-macros and zend_ magic. It was interesting constructing PHP associative arrays (Zend Hash Tables) for each of the exposed PHP functions return values. This certainly isn’t the most useful runtime, but it does document how easy it is to expose Rust-based code in a higher-level language, without the need to resort to C.

<?php

$solution = file_get_contents(__DIR__ . '/solution.santa');

santa_aoc_run($solution, cwd: __DIR__);

santa_aoc_test($solution);

santa_evaluate('1.. |> filter(_ % 2) |> take(3);');

Along with providing this PHP extension support, there is future scope to expose the interpreter within Python. This crate looks to expose a similar level of abstraction as the PHP one.

Jupyter Kernel

Over the past several months I have been introduced to the power of Jupyter Notebook and the fast REPL it provides. This is in most part due to every AI/LLM YouTube video/or article I see using it (due to the authors coming from Data Science backgrounds). Similar to the PHP extension, my mind wandered into the possibility of creating a custom Kernel to expose the santa-lang language as an additional runtime.

Thanks to yet another well-thought-out crate, this task proved to be rather seamless. There was one caveat however, in that the crate used asynchronous execution to handle language server requests. This caused issues with the Rc<RefCell<Environment>> type that is required to model script state during evaluation. As santa-lang’s core language is designed to be single-threaded I was required to add unsafe Sendable and Sync trait implementations to a wrapper SharedEnvironment type to appease the compiler (a big no-no!).

This is certainly not production ready, but as a proof-of-concept it was fun to build. It also make me interested to see how I could possibly use a notebook to help develop Advent of Code solutions in the future.

Jupyter Kernel

What’s next?

Now that all the desired runtimes had been implemented it was finally time to investigate performance considerations! In the next post within the series I will document how I went about benchmarking the two implementations, greatly improving performance and highlighting interesting findings along the way.