Babashka and fzf are two tools that are consistently part of my workflow, regardless of what I'm doing. I've briefly wrote about them in my previous post, PhuzQL: A GraphQL Fuzzy Finder, where I combined them together with Pathom to create a simple GraphQL explorer. In short, Babashka lets you write Clojure scripts with a fast startup time and fzf is an interactive picker with fuzzy text matching to save you some keystrokes.
Babashka includes a tasks feature that is somewhat like a Makefile replacement, allowing you to define named tasks in a bb.edn file and then execute them with bb <task name> [optional params]. bb-fzf just lets you pick from your list of tasks interactively - think of it as an ergonomic autocomplete.
Check out the README in the bb-fzf repo for more details on installation and usage. There's always room for improvement, but this is sufficient for my needs for now. In the future I might add a preview window and more robust argument handling with the help of babashka.cli.
We’ll dive into pattern matching using the core.match library in Clojure
This powerful tool allows you to handle complex data structures in a declarative and concise way, making your code more readable and maintainable. Let’s explore what pattern matching is, how to use core.match, and why it’s a valuable addition to your Clojure toolkit, complete with practical examples.
What Is Pattern Matching?
Pattern matching is a technique where you match data against predefined patterns, extracting components or triggering actions based on the match. It’s a staple in functional programming languages like Haskell and Scala, and in Clojure, the core.match library brings this capability to the table. Unlike basic conditionals like if or cond, pattern matching lets you express "what" you want to happen rather than "how" to check conditions step-by-step. It’s especially useful for:
Parsing structured data (e.g., JSON or EDN).
Handling data variants or tagged unions.
Simplifying nested conditional logic.
Setting Up core.match
To get started, you’ll need to add core.match to your project. The latest version available is 1.1.0 (as per the GitHub repository). Here’s how to include it:
For Leiningen (in project.clj):
:dependencies [[org.clojure/core.match "1.1.0"]]
For deps.edn: {:deps {org.clojure/core.match {:mvn/version "1.1.0"}}}
This is destructuring made more powerful and flexible.
Practical Example: Parsing Commands
Let’s apply pattern matching to a real-world scenario—parsing commands. Suppose we have commands like [:add 5] or [:mul 2 3]:
(defnprocess-command[cmd](matchcmd[:addx](str"Adding "x)[:subx](str"Subtracting "x)[:mulxy](str"Multiplying "x" by "y):else"Unknown command"))(println(process-command[:add5]))(println(process-command[:mul23]))
This is much cleaner than a chain of cond statements, and the intent is immediately clear.
Advanced Features: Guards and Nested Patterns
core.match goes beyond basics with features like guards ":guard fn" and nested pattern matching. Here’s an example with a guard:
This declarative approach makes the state transitions crystal clear.
What do you think? Have you tried core.match in your projects? Let me know in the comments—I’d love to hear your experiences! Stay tuned for the next part of our series, where we’ll uncover more Clojure awesomeness.
Hey everyone! Welcome back to Clojure Is Awesome! After a bit of a break, I’m thrilled to be back with Part 18, diving into one of Clojure’s superpowers: interoperability with Java. If you’ve ever worked with Java or wondered how Clojure can leverage the vast Java ecosystem, this post is for you. We’ll explore how Clojure seamlessly uses Java libraries, frameworks, and even builds graphical interfaces—all while keeping things functional, clean, and practical.
What is Interoperability and Why Does It Matter?
Interoperability is the ability of a language or system to interact smoothly with another. In Clojure, this means it can:
Import and instantiate Java classes.
Call methods on Java objects.
Integrate with JVM libraries and frameworks.
Why is this important? Clojure runs on the JVM (Java Virtual Machine), giving it access to a mature ecosystem filled with battle-tested tools. This enables:
Reusing existing Java code.
Leveraging robust libraries for tasks like data manipulation, logging, or processing.
Integrating Clojure into legacy Java projects or using it alongside modern systems.
In short, interoperability makes Clojure a versatile choice for everything from simple scripts to complex enterprise applications.
Importing and Using Java Classes in Clojure
Let’s start with the basics: importing Java classes and creating instances in Clojure.
Example 1: Working with java.util.Date
Here, we import the Date class and create a function to get the current date.
(nsclojure-is-awesome.interop(:import[java.utilDate]))(defnget-current-date"Returns the current date using java.util.Date."[](Date.))(println"Current date:"(get-current-date))
(:import [java.util Date]): Imports the Date class from the java.util package.
(Date.): Creates a new instance by calling the default constructor.
Example 2: Manipulating an ArrayList
Now, let’s create an ArrayList, add elements, and interact with its methods.
(nsclojure-is-awesome.interop(:import[java.utilArrayList]))(defncreate-and-populate-list"Creates an ArrayList, adds elements, and returns the list."[](let[lista(ArrayList.)](.addlista"Clojure")(.addlista"Java")lista))(println"List:"(create-and-populate-list))
(ArrayList.): Instantiates a new ArrayList.
(.add lista "Clojure"): Calls the add method on the lista instance.
Using Popular Java Libraries
Clojure can integrate with widely used Java libraries. Let’s use Apache Commons Lang for string manipulation.
Example 3: Reversing Strings with Apache Commons
First, add the dependency to your project.clj (if using Leiningen):
Now, create a function that reverses a string using StringUtils.
(nsclojure-is-awesome.interop(:import[org.apache.commons.lang3StringUtils]))(defnreverse-string"Reverses a string using Apache Commons' StringUtils."[s](StringUtils/reverses))(println"Reversed string:"(reverse-string"Clojure Is Awesome!!!"))
StringUtils/reverse: Calls the static method reverse from StringUtils.
Integrating with Java Frameworks: Spring Boot
Clojure also integrates with frameworks like Spring Boot, enabling robust web applications. Let’s create a simple REST endpoint.
Example 4: REST Endpoint with Spring Boot and Clojure
Create a Spring Boot project with a Java class that delegates to a Clojure function.
Define the Clojure function to be called.
Clojure function (src/clojure/myapp/core.clj):
(nsmyapp.core)(defngreet"Returns a greeting message."[name](str"Hello, "name"!"))
Spring Boot Java class (src/java/com/example/demo/DemoApplication.java):
Spring Boot calls the greet function from the myapp.core namespace using the clojure.java.api.
The /greet?name=John endpoint returns "Hello, John!".
This basic example shows how Clojure can be used within a Spring Boot application.
Building Graphical Interfaces with Swing
Clojure can also create graphical interfaces using Java libraries like Swing. Let’s make a simple window with a button.
Example 5: Window with Button in Swing
(nsclojure-is-awesome.interop(:import[javax.swingJFrameJButton][java.awt.eventActionListener]))(defncreate-gui"Creates a window with a button that prints a message when clicked."[](let[frame(JFrame."Clojure Example")button(JButton."Click Here")](.addActionListenerbutton(reifyActionListener(actionPerformed[__](println"Button clicked!"))))(.addframebutton)(.setSizeframe300200)(.setVisibleframetrue)))(create-gui)
reify: Implements the ActionListener **interface to react to button clicks.
**Result: A window appears with a button that prints a message to the console when clicked.
Practical Tips for Interoperability
To wrap up, here are some useful tips when working with Clojure and Java:
Dependency Management:
Add Java libraries in project.clj or deps.edn. Example: :dependencies [[org.apache.commons/commons-lang3 "3.12.0"]]
Exception Handling:
Use try/catch to capture Java exceptions:
Type Hints:
Improve performance by avoiding reflection:
(defnmy-function[^Strings](.toUpperCases))
Static Methods:
Call them directly with Class/method, like StringUtils/reverse.
Documentation:
Refer to Java documentation to understand the classes and methods used.
Conclusion
In Part 18, we’ve explored the interoperability between Clojure and Java, showcasing how Clojure leverages the Java ecosystem elegantly. We’ve covered everything from using simple classes like Date and ArrayList to integrating with libraries like Apache Commons, frameworks like Spring Boot, and building graphical interfaces with Swing. Interoperability is one of Clojure’s key strengths, bringing the best of both worlds together.
What about you—have you tried combining Clojure and Java in your projects? Share your experiences in the comments! Next time, we’ll dive into another exciting corner of Clojure’s universe. Until then, keep coding with style!
This is part of a series of posts exploring programming with Rama, ranging from interactive consumer apps, high-scale analytics, background processing, recommendation engines, and much more. This tutorial is self-contained, but for broader information about Rama and how it reduces the cost of building backends so much (up to 100x for large-scale backends), see our website.
Like all Rama applications, the example in this post requires very little code. It’s easily scalable to millions of reads/writes per second, ACID compliant, high performance, and fault-tolerant from how Rama incrementally replicates all state. Deploying, updating, and scaling this application are all one-line CLI commands. No other infrastructure besides Rama is needed. Comprehensive monitoring on all aspects of runtime operation is built-in.
In this post, I’ll explore building the backend for a real-time collaborative text editor like Google Docs or Etherpad. The technical challenge of building an application like this is conflict resolution. When multiple people edit the same text at the same time, what should be the result? If a user makes a lot of changes offline, when they come online how should their changes be merged in to a document whose state may have diverged significantly?
There are many approaches for solving this problem. I’ll show an implementation of “operational transformations” similar to the one Google Docs uses as described here. Only incremental changes are sent back and forth between server and client, such as “Add text ‘hello’ to offset 128” or “Remove 14 characters starting from offset 201”. When clients send a change to the server, they also say what version of the document the change was applied to. When the server receives a change from an earlier document version, it applies the “operational transformation” algorithm to modify the addition or removal to fit with the latest document version.
Code will be shown in both Clojure and Java, with the total code being about 120 lines for the Clojure implementation and 160 lines for the Java implementation. Most of the code is implementing the “operational transformation” algorithm, which is just plain Clojure or Java functions, and the Rama code handling storage/queries is just 40 lines. You can download and play with the Clojure implementation in this repository or the Java implementation in this repository.
Operational transformations
The idea behind “operational transformations” is simple and can be understood through a few examples. Suppose Alice and Bob are currently editing a document, and Alice adds "to the " at offset 6 when the document is at version 3, like so:
However, when the change gets to the server, the document is actually at version 4 since Bob added "big " to offset 0:
Applying Alice’s change without modification would produce the string “big heto the llo world”, which is completely wrong. Instead, the server can transform Alice’s change based on the single add that happened between versions 3 and 4 by pushing Alice’s change to the right by the length of "big ". So Alice’s change of “Add ‘to the ’ at offset 6” becomes “Add ‘to the ’ at offset 10”, and the document becomes:
Now suppose instead the change Bob made between versions 3 and 4 was adding " again" to the end:
In this case, Alice’s change should not be modified since Bob’s change was to the right of her addition, and applying Alice’s change will produce “hello to the world again”.
Transforming an addition against a missed remove works the same way: if Bob had removed text to the left of Alice’s addition, her addition would be shifted left by the amount of text removed. If Bob removed text to the right, Alice’s addition is unchanged.
Now let’s take a look at what happens if Alice had removed text at an earlier version, which is slightly trickier to handle. Suppose Alice removes 7 characters starting from offset 2 when the document was “hello world” at version 3. Suppose Bob had added "big " to offset 0 between versions 3 and 4:
In this case, Alice’s removal should be shifted right by that amount to remove 7 characters starting from offset 6. Now suppose Bob had instead added "to the " to offset 6:
In this case, text had been added in the middle of where Alice was deleting text. It’s wrong and confusing for Alice’s removal to delete text that wasn’t in her version of the document. The correct way to handle this is split her removal into two changes: remove 3 characters from offset 13 and then remove 4 characters from offset 2. This produces the following document:
The resulting document isn’t legible, but it’s consistent with the conflicting changes that Alice and Bob were making at the same time without losing any information improperly.
Let’s take a look at one more case. Suppose again that Alice removes 7 characters from offset 2 when the document was “hello world” at version 3. This time, Bob had already removed one character starting from offset 3:
In this case, one of the characters Alice had removed was already removed, so Alice’s removal should be reduced by one to remove 6 characters from offset 2 instead of 7, producing:
There’s a few more intricacies to how operational transformations work, but this gets the gist of the idea across. Since this post is really about using Rama to handle the storage and computation for a real-time collaborative editor backend, I’ll keep it simple and only handle adds and removes. The code can easily be extended to handle other kinds of document edits such as formatting changes.
The Rama code will make use of two functions that encapsulate the “operational transformation” logic, one to apply a series of edits to a document and the other to transform an edit against a particular version against all the edits that happened until the latest version. These functions are as follows:
The “apply edits” function updates a document with a series of edits, and the “transform edit” function implements the described “operational transformation” algorithm. The “transform edit” function returns a list since an operational transformation on a remove edit can split that edit into multiple remove edits, as described in one of the cases above.
Indexed datastores in Rama, called PStates (“partitioned state”), are much more powerful and flexible than databases. Whereas databases have fixed data models, PStates can represent infinite data models due to being based on the composition of the simpler primitive of data structures. PStates are distributed, durable, high-performance, and incrementally replicated. Each PState is fine-tuned to what the application needs, and an application makes as many PStates as needed. For this application, we’ll make two PStates: one to track the latest contents of a document, and one to track the sequence of every change made to a document in its history.
Here’s the PState definition for the latest contents of a document:
This PState is a map from a 64-bit document ID to the string contents of the document. The name of a PState always begins with
$$
, and this is equivalent to a key/value database.
Here’s the PState tracking the history of all edits to a document:
This declares the PState as a map of lists, with the key being the 64-bit document ID and the inner lists containing the edit data. The inner list is declared as “subindexed”, which instructs Rama to store the elements individually on disk rather than the whole list read and written as one value. Subindexing enables nested data structures to have billions of elements and still be read and written to extremely quickly. This PState can support many queries in less than one millisecond: get the number of edits for a document, get a single edit at a particular index, or get all edits between two indices.
Let’s now review the broader concepts of Rama in order to understand how these PStates will be materialized.
Rama concepts
A Rama application is called a “module”. In a module you define all the storage and implement all the logic needed for your backend. All Rama modules are event sourced, so all data enters through a distributed log in the module called a “depot”. Most of the work in implementing a module is coding “ETL topologies” which consume data from one or more depots to materialize any number of PStates. Modules look like this at a conceptual level:
Modules can have any number of depots, topologies, and PStates, and clients interact with a module by appending new data to a depot or querying PStates. Although event sourcing traditionally means that processing is completely asynchronous to the client doing the append, with Rama that’s optional. By being an integrated system Rama clients can specify that their appends should only return after all downstream processing and PState updates have completed.
A module deployed to a Rama cluster runs across any number of worker processes across any number of nodes, and a module is scaled by adding more workers. A module is broken up into “tasks” like so:
A “task” is a partition of a module. The number of tasks for a module is specified on deploy. A task contains one partition of every depot and PState for the module as well as a thread and event queue for running events on that task. A running event has access to all depot and PState partitions on that task. Each worker process has a subset of all the tasks for the module.
Coding a topology involves reading and writing to PStates, running business logic, and switching between tasks as necessary.
Implementing the backend
Let’s start implementing the module for the collaborative editor backend. The first step to coding the module is defining the depot:
This declares a Rama module called “CollaborativeDocumentEditorModule” with a depot called
*edit-depot
which will receive all new edit information. Objects appended to a depot can be any type. The second argument of declaring the depot is called the “depot partitioner” – more on that later.
To keep the example simple, the data appended to the depot will be
defrecord
objects for the Clojure version and
HashMap
objects for the Java version. To have a tighter schema on depot records you could instead use Thrift, Protocol Buffers, or a language-native tool for defining the types. Here are the functions that will be used to create depot data:
(defn mk-add-edit [id version offset content] (->Edit id version offset (->AddText content)))
(defn mk-remove-edit [id version offset amount] (->Edit id version offset (->RemoveText amount)))
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19
publicstaticMap makeAddEdit(long id, int version, int offset, String content){ Map ret =newHashMap();
ret.put("type", "add");
ret.put("id", id);
ret.put("version", version);
ret.put("offset", offset);
ret.put("content", content); return ret; }
publicstaticMap makeRemoveEdit(long id, int version, int offset, int amount){ Map ret =newHashMap();
ret.put("type", "remove");
ret.put("id", id);
ret.put("version", version);
ret.put("offset", offset);
ret.put("amount", amount); return ret; }
Each edit contains a 64-bit document ID that identifies which document the edit should apply.
Next, let’s begin defining the topology to consume data from the depot and materialize the PStates. Here’s the declaration of the topology with the PStates:
This defines a stream topology called “core”. Rama has two kinds of topologies, stream and microbatch, which have different properties. In short, streaming is best for interactive applications that need single-digit millisecond update latency, while microbatching has update latency of a few hundred milliseconds and is best for everything else. Streaming is used here because a collaborative editor needs quick feedback from the server as it sends changes back and forth.
Notice that the PStates are defined as part of the topology. Unlike databases, PStates are not global mutable state. A PState is owned by a topology, and only the owning topology can write to it. Writing state in global variables is a horrible thing to do, and databases are just global variables by a different name.
Since a PState can only be written to by its owning topology, they’re much easier to reason about. Everything about them can be understood by just looking at the topology implementation, all of which exists in the same program and is deployed together. Additionally, the extra step of appending to a depot before processing the record to materialize the PState does not lower performance, as we’ve shown in benchmarks. Rama being an integrated system strips away much of the overhead which traditionally exists.
Let’s now add the code to materialize the PStates:
The code to implement the topology is less than 20 lines, but there’s a lot to unpack here. The business logic is implemented with dataflow. Rama’s dataflow API is exceptionally expressive, able to intermix arbitrary business logic with loops, conditionals, and moving computation between tasks. This post is not going to explore all the details of dataflow as there’s simply too much to cover. Full tutorials for Rama dataflow can be found on our website for the Java API and for the Clojure API.
Let’s go over each line of this topology implementation. The first step is subscribing to the depot:
This subscribes the topology to the depot
*edit-depot
and starts a reactive computation on it. Operations in dataflow do not return values. Instead, they emit values that are bound to new variables. In the Clojure API, the input and outputs to an operation are separated by the
:>
keyword. In the Java API, output variables are bound with the
.out
method.
Whenever data is appended to that depot, the data is emitted into the topology. The Java version binds the emit into the variable
*edit
and then gets the fields “id” and “version” from the map into the variables
*id
and
*version
, while the Clojure version captures the emit as the variable
*edit
and also destructures its fields into the variables
*id
and
*version
. All variables in Rama code begin with a
*
. The subsequent code runs for every single emit.
Remember that last argument to the depot declaration called the “depot partitioner”? That’s relevant here. Here’s that image of the physical layout of a module again:
The depot partitioner determines on which task the append happens and thereby on which task computation begins for subscribed topologies. In this case, the depot partitioner says to hash by the “id” field of the appended data. The target task is computed by taking the hash and modding it by the total number of tasks. This means data with the same ID always go to the same task, while different IDs are evenly spread across all tasks.
Rama gives a ton of control over how computation and storage are partitioned, and in this case we’re partitioning by the hash of the document ID since that’s how we want the PStates to be partitioned. This allows us to easily locate the task storing data for any particular document.
The next line fetches the current version of the document:
The
$$edits
PState contains every edit applied to the document, and the latest version of a document is simply the size of that list. The PState is queried with the “local select” operation with a “path” specifying what to fetch. When a PState is referenced in dataflow code, it always references the partition of the PState that’s located on the task on which the event is currently running.
Paths are a deep topic, and the full documentation for them can be found here. A path is a sequence of “navigators” that specify how to hop through a data structure to target values of interest. A path can target any number of values, and they’re used for both transforms and queries. In this case, the path navigates by the key
*id
to the list of edits for the document. The next navigator
view
runs a function on that list to get its size. The Clojure version uses Clojure’s
count
function, and the Java version uses the
Ops.SIZE
function. The output of the query is bound to the variable
*latest-version
. This is a fast sub-millisecond query no matter how large the list of edits.
The next few lines run the “operational transformation” algorithm if necessary to produce the edits to be applied to the current version of the document:
First, an “if” is run to check if the version of the edit is the same as the latest version on the backend. If so, the list of edits to be applied to the document is just the edit unchanged. As mentioned before, the operational transformation algorithm can result in multiple edits being produced from a single edit. The Clojure version produces the single-element list by calling
vector
, and the Java version does so with the
Ops.TUPLE
function. The list of edits is bound to the variable
*final-edits
.
The “else” branch of the “if” handles the case where the edit must be transformed against all edits up to the latest version. The “local select” on
$$edits
fetches all edits from the input edit’s version up to the latest version. The navigator to select the sublist,
srange
in Clojure and
sublist
in Java, takes in as arguments a start offset (inclusive) and end offset (exclusive) and navigates to the sublist of all elements between those offsets. This sublist is bound to the variable
*missed-edits
.
The function shown before implementing operational transformations,
transform-edit
in Clojure and
CollaborativeDocumentEditorModule::transformEdit
in Java, is then run on the edit and all the missed edits to produce the new list of edits and bind them to the variable
*final-edits
.
Any variables bound in both the “then” and “else” branches of an “if” conditional in Rama will be in scope after the conditional. In this case,
*final-edits
is available after the conditional.
*missed-edits
is not available since it is not bound in the “then” branch. This behavior comes from Rama implicitly “unifying” the “then” and “else” branches.
The next bit of code gets the latest document and applies the transformed edits to it:
The “local select” fetches the latest version of the document from the
$$docs
PState. The second navigator in the path,
nil->val
in Clojure and
nullToVal
in Java, handles the case where this is the first ever edit on the document. In that case the document ID does not exist in this PState. In that case the key navigation by
*id
would navigate to
null
, so the next navigator instead navigates to the empty string in that case.
The next line runs the previously defined “apply edits” function to apply the transformed edits to produce the new version of the document into the variable
*new-doc
.
The two PStates are updated with the “local transform” operation. Like “local select”, a “local transform” takes in as input a PState and a “path”. Paths for “local transform” navigate to the values to change and then use special “term” navigators to update them.
The first “local transform” navigates into
$$docs
by the document ID and uses the “term val” navigator to set the value there to
*new-doc
. This is exactly the same as doing a “put” into a hash map.
The second “local transform” appends the transformed edits to the end of the list of edits for this document. It navigates to the list by the key
*id
and then navigates to the “end” of the list. More specifically, the “end” navigator navigates to the empty list right after the overall list. Setting that empty list to a new value appends those elements to the overall list, which is what the final “term val” navigator does.
This entire topology definition executes atomically – all the PState queries, operational transformational logic, and PState writes all happen together and nothing else can run on the task in between. This is a result of Rama colocating computation and storage, which will be explored more in the next section.
The power of colocation
Let’s take a look at the physical layout of a module again:
Every task has a partition of each depot and PState as well as an executor thread for running events on that task. Critically, only one event can run at a time on a task. That means each event has atomic access to all depots and PState partitions on that task. Additionally, those depot and PState partition are local to the JVM process running that event so interactions with them are fully in-process (as opposed to the inter-process communication used with databases).
A traditional database handles many read and write requests concurrently, using complex locking strategies and explicit transactions to achieve atomicity. Rama’s approach is different: parallelism is achieved by having many tasks in a module, and atomicity comes from colocation. Rama doesn’t have explicit transactions because transactional behavior is automatic when computation is colocated with storage.
When writing a topology in a module, you have full control over what constitutes a single event. Code runs synchronously on a task unless they explicitly go asynchronous, like with partitioners or yields.
This implementation for a collaborative editor backend is a great example of the power of colocation. The topology consists of completely arbitrary code running fine-grained logic for the “operational transformational” logic and manipulating multiple PStates, and nothing special needed to be done to get the necessary transactional behavior.
When you use a microbatch topology to implement an ETL, you get even stronger transactional behavior. All microbatch topologies are cross-partition transactions in every case, no matter how complex the logic.
You can read more about Rama’s strong ACID semantics on this page.
Query topology to fetch document and version
The module needs one more small thing to complete the functionality necessary for a real-time collaborative editor backend. When the frontend is loaded, it needs to load the latest document contents along with its version. The contents are stored in the
$$docs
PStates, and the version is the size of the list of edits in the
$$edits
PState. So we need to read from both those PStates atomically in one event.
If you were to try to do this with direct PState clients, you would be issuing one query to the client for the
$$edits
PState and one query to the client for the
$$docs
PState. Those queries would run as separate events, and the PStates could be updated in between the queries. This would result in the frontend having incorrect state.
Rama provides a feature called “query topologies” to handle this case. Query topologies are exceptionally powerful, able to implement high-performance, real-time, distributed queries across any or all of the PStates of a module and any or all of the partitions of those PStates. They’re programmed with the exact same dataflow API as used to program ETLs.
For this use case, we only need to query two PStates atomically. So this is a simple use case for query topologies. The full implementation is:
This declares a query topology named “doc+version” that takes in one argument
*id
as input. It declares the return variable
*ret
, which will be bound by the end of the topology execution.
The next line gets the query to the task of the module containing the data for that ID:
ClojureJava
1
(|hash *id)
1
.hashPartition("*id")
The line does a “hash partition” by the value of
*id
. Partitioners relocate subsequent code to potentially a new task, and a hash partitioner works exactly like the aforementioned depot partitioner. The details of relocating computation, like serializing and deserializing any variables referenced after the partitioner, are handled automatically. The code is linear without any callback functions even though partitioners could be jumping around to different tasks on different nodes.
When the first operation of a query topology is a partitioner, query topology clients are optimized to go directly to that task. You’ll see an example of invoking a query topology in the next section.
The next two lines atomically fetch the document contents and version:
These two queries are atomic because of colocation, just as explained above. Fetching the latest contents of a document is simply a lookup by key, and fetching the latest version is simply the size of the list of edits.
The next line packages these two values into a single object:
Finally, here’s the last line of the query topology:
ClojureJava
1
(|origin)
1
.originPartition();
The “origin partitioner” relocates computation to the task where the query began execution. All query topologies must invoke the origin partitioner, and it must be the last partitioner invoked.
Let’s now take a look at how a client would interact with this module.
Interacting with the module
Here’s an example of how you would get clients to
*edit-depot
,
$$edits
, and the
doc+version
query topology, such as in your web server:
A “cluster manager” connects to a Rama cluster by specifying the location of its “Conductor” node. The “Conductor” node is the central node of a Rama cluster, which you can read more about here. From the cluster manager, you can retrieve clients to any depots, PStates, or query topologies for any module. The objects are identified by the module name and their name within the module.
Here’s an example of appending an edit to the depot:
ClojureJava
1
(foreign-append! edit-depot (mk-add-edit 12335"to the "))
1
editDepot.append(makeAddEdit(123, 3, 5, "to the "));
This uses the previously defined helper functions to create an add edit for document ID 123 at version 3, adding “to the ” to offset 5.
Here’s an example of querying the
$$edits
PState for a range of edits:
This looks no different than invoking any other function, but it’s actually executing remotely on a cluster. This returns a hash map containing the doc and version just as defined in the query topology.
Workflow with frontend
Let’s take a look at how a frontend implementation can interact with our completed backend implementation to be fully reactive. The approach is the same as detailed in this post. Architecturally, besides Rama you would have a web server interfacing between Rama and web browsers.
The application in the browser need to know as soon as there’s an update to the document contents on the backend. This is easy to do with Rama with a reactive query like the following:
“Proxy” is similar to the “select” calls already shown. It takes in a path specifying a value to navigate to. Unlike “select”, “proxy”:
Returns a ProxyState which is continuously and incrementally updated as the value in that PState changes. This has a method “get” on it that can be called at any time from any thread to get its current value.
Can register a callback function that’s invoked as soon as the value changes in the module.
The time to invoke the callback function after being changed on the PState is less than a millisecond, and it’s given three arguments: the new value (what would be selected if the path was run from scratch), a “diff” object, and the previous value the last time the callback function was run. The diff isn’t needed in this case, but it contains fine-grained information about how the value changed. For example, if the path was navigating to a set object, the diff would contain the specific info of what elements were added and/or removed from the set. Reactive queries in Rama are very potent, and you can read more about them here.
For this use case, the browser just need to know when the version changes so it can take the following actions:
Fetch all edits that it missed
Run operational transformational algorithm against any pending edits it has buffered but hasn’t sent to the server yet
The browser contains four pieces of information:
The document contents
The latest version its contents are based on
Edits that have been sent to the server but haven’t been acknowledged yet
Edits that are pending
The basic workflow of the frontend is to:
Buffer changes locally
Send one change at a time to the server. When it’s applied, run operational transformation against all pending edits, update the version, and then send the next pending change to the server if there is one.
The only time the full document contents are ever transmitted between server and browser are on initial load when the query topology is invoked to atomically retrieve the document contents and version. Otherwise, only incremental edit objects are sent back and forth.
Summary
There’s a lot to learn with Rama, but you can see from this example application how much you can accomplish with very little code. For an experienced Rama programmer, a project like this takes only a few hours to fully develop, test, and have ready for deployment. The Rama portion of this is trivial, with most of the work being implementing the operational transformation algorithm.
As mentioned earlier, there’s a Github project for the Clojure version and for the Java version containing all the code in this post. Those projects also have tests showing how to unit test modules in Rama’s “in-process cluster” environment.
You can get in touch with us at consult@redplanetlabs.com to schedule a free consultation to talk about your application and/or pair program on it. Rama is free for production clusters for up to two nodes and can be downloaded at this page.
Back in 2019, I shared the first community survey results for
CIDER, the beloved Clojure interactive
development environment for Emacs. Five years later, we’re back to take the
pulse of the ecosystem and see how things have evolved.
In this post, we’ll explore the results of the 2024 State of CIDER Survey,
compare them to the 2019 edition, and reflect on the progress and the road
ahead.
Who Are CIDER Users Today?
Experience With CIDER
In 2019, most users had between 1-5 years of experience with CIDER. Fast forward
to 2024, and we now see a significant portion with 5+ years under their
belts — a great sign of CIDER’s stability and staying power. Newcomers are still
joining, but the user base has clearly matured.
I guess also has to do with the fact that Clojure is not growing as much as before
and few new users are joining the Clojure community. For the record:
550 took part in CIDER’s survey in 2019
330 took part in 2024
I’ve observed drops in the rate of participation for State of Clojure surveys as well.
Prior Emacs Usage
The majority of respondents were Emacs users before trying CIDER in both surveys. A fun new entry in 2024: “It’s complicated” — we see you, Vim-converts and editor-hoppers!
Tools, Setups, and Installs
Emacs & CIDER Versions
Users have generally moved in sync with CIDER and Emacs development. CIDER 1.16 is now the dominant version, and Emacs 29/30 are common. This reflects good community alignment with upstream tooling.
Probably by the time I wrote this article most people are using CIDER 1.17 (and 1.18 is right around the corner). Those results embolden us to be slightly more aggressive with adopting newer Emacs features.
Installation Methods
Package.el remains the most popular method of installation, although straight.el has carved out a notable niche among the more config-savvy crowd.
Nothing really shocking here.
Usage Patterns
Professional Use
Just like in 2019, around half of the respondents use CIDER professionally. The remaining half are hobbyists or open-source tinkerers — which is awesome.
Upgrade Habits
There’s a visible shift from “install once and forget” toward upgrading with major releases or as part of regular package updates. CIDER’s release cadence seems to be encouraging healthier upgrade practices.
Used Features
This was the biggest addition to the survey in 2024 and the most interesting to me. It confirmed my long-held suspicions that most people use only the most basic CIDER functionality.
I can’t be sure why is that, but I have a couple of theories:
Information discovery issues
Some features are somewhat exotic and few people would benefit from them
I have plans to address both points with better docs, video tutorials and gradual removal of some features that add more complexity than value. CIDER 1.17 and 1.18 both
make steps in this direction.
Community & Documentation
Documentation Satisfaction
Documentation continues to score well. Most users rate it 4 or 5 stars, though there’s always room for growth, especially in areas like onboarding and advanced features.
From my perspective the documentation can be improved a lot (e.g. it’s not very consistent, the structure is also suboptimal here and there), but that’s probably not a big
issue for most people.
Learning Curve
The majority of users rate CIDER’s learning curve as moderate (3-4 out of 5), consistent with the complexity of Emacs itself. Emacs veterans may find it smoother, but newcomers still face a bit of a climb.
I keep advising people not to try to learn Emacs and Clojure at the same time!
Supporting CIDER
While more users and companies are aware of ways to support CIDER (like
OpenCollective or GitHub Sponsors), actual support remains low. As always, a
small donation or contribution can go a long way to sustaining projects like
this. As a matter of fact the donations for CIDER and friends have dropped a lot
of since 2022, which is quite disappointing given all the efforts me and other
contributors have put into the project.
Conclusion
CIDER in 2024 is a mature, well-loved tool with a seasoned user base. Most users
are professionals or long-time hobbyists, and satisfaction remains high. If
you’re reading this and you’re new to CIDER — welcome! And if you’re a
long-timer, thank you for helping build something great.
Thanks to everyone who participated in the 2024 survey. As always, feedback and
contributions are welcome — and here’s to many more years of productive, joyful
hacking with Emacs and CIDER.
I spent one week with Claude Code, vibe coding two apps in Clojure. Hours and $134.79 in. Let me tell you what I got out of that.
Claude Code is a new tool from Anthropic. You cd to the folder, run claude and a CLI app opens.
You tell it what to do and it starts working. Every time it runs a command, it lets you decide whether to do the task, or do something else. You can tell it to always be allowed to do certain tasks which is useful for things like moving in the directory, changing files, running tests.
Task 1: Rewrite app from Python to Clojure
We have a small standalone app here implemented in Python. It’s an app that serves as a hub for many ETL other jobs. Every job is run, the result is collected, packed as JSON and sent into the API for further processing.
So I copied:
the app that should be converted from Python to Clj (basically job runner, data collector)
source code of the API that receives data
source code of workers that process received data
I explained exactly what I want to achieve and started iterating.
Also, I created a CLAUDE.md file which explained the purpose of the project, how to use Leiningen (yes, I still use Leiningen for new projects; I used Prismatic Schema in this project too). It also contained many additional instructions like adding schema checking, asserts, tests.
An excerpt from CLAUDE.md file.
## Generating data
Double check correctness of files you generate. I found you inserting 0 instead of } in JSON and “jobs:” string to beginning of lines in yml files, making them invalid.
The app is processing many Git repositories and most of the time is spent waiting for jobs to finish.
One innovation Claude Code did itself was creating 4 working copies for each repo and running 8 repositories in parallel. This means, the app was processing all work 32* times faster. I had to double check if it is really running my jobs because of how fast the result app was.
There were minor problems, like generating invalid JSON files for unit tests, or having problems processing outputs from commands and wrapping them in JSON.
I was about 6 hours in and 90% finished.
There were still some corner cases, where jobs got stuck. Or when jobs took much longer than they should. Or when some commits in repositories weren’t processed.
I instructed Claude Code to write more tests, testing scripts and iterate to resolve this. I wanted to make sure this will be a real drop in replacement for existing Python app that will take the same input config, will call API compatible, and will be just 30* faster.
This is where I ran into a problem. Claude Code worked for hours creating and improving scripts, adding more tests. For an app that’s maybe 1000 lines of code, I ended up with 2000 lines of tests and 2000 lines of various testing scripts.
At that time, I also ran into problems with bash. Claude Code generated code that required associative arrays.
I started digging, why does this app even need associative arrays in bash. To my big disappointment, I found that over 2 days, Claude Code copied more and more logic from Clojure to bash.
It got to the point where all logic was transported from Clojure to bash.
My app wasn’t running Clojure code anymore! Everything was done in bash. That’s why it all ended up with 2000 lines of shell code.
That was the first time in my life, when I told LLM I was disappointed by it. It quickly apologized and started moving back to Clojure.
Unfortunately, after a few more hours, I wasn’t able to really get all the tests passing and my test tasks running at the same time. I abandoned my effort for a while.
Task 2: CLI-based DAW in Clojure
When I was much younger, I used to make and DJ music. It all started with trackers in DOS.
Most of the world was using FastTracker II, but my favourite one was Impulse Tracker.
Nowadays, I don’t have time to do as much music as in the past. But I got an idea. How hard might it be to build my own simple music making software.
Criteria I gave to Claude Code:
Build it in Clojure
Do not rewrite it in another language (guess how I got to this criteria?)
Use Overtone as an underlying library
Use CLI interface
Do not use just 80×25 or any small subset. Use the whole allocated screen space.
The control should be based on Spacemacs. So for example loading a sample would be [SPC]-[S]-[L].
Spacemacs [SPC]- based approach is awesome. Our HR software Frankie uses a Spacemacs-like approach too (adding new candidates with [SPC]-[C]-[A]). So it seems to me, it’s only natural to extend this approach to other areas.
Imagine, a music making software that is inspired by Spacemacs!
I called it spaceplay and let Claude Code cook.
After a while, I had a simple interface, session mode, samples, preferences, etc.
So now, I got to the point where I wanted to add tests and get the sound working.
One option was to hook directly on the interface. Just like every [SPC] command sequence dispatches an event, I could just simulate those presses and test, if the interface reacts.
The issue is, when you have a DAW, you have many things happening at the same time. I didn’t want to test only things that are related to the user doing things.
And at the same time, I wanted my DAW to be a live set environment, where the producer can make music. It is an environment with many moving parts. So it made sense to me to make a testing API. I exposed the internal state via API and let tests manipulate it.
This was a bit of the problem for Claude.
However, Claude Code is extremely focused on text input and output. There’s no easy way to explain how to attach to sound output and test it. Even overtone test folder doesn’t spend a lot of time doing this overtone/test/overtone at master · overtone/overtone.
I wanted to test Claude Code in vibe coding mode. I didn’t want to do a lot of code review, or fixing code for it, like I do with Cursor.
So we ended up in a cycle where I wanted it to get sound working & test it and it was failing at this task.
Conclusion
After 6000–7000 lines of code (most of it Clojure, some of it bash) generated, $134.79 spent, and 12 hours invested, I came to a conclusion.
Cursor is a much better workflow for me. AI without overview is still not there to deliver maintainable apps. All of those people who are vibecoding are going to throw those apps away later, or are going to regret it.
AI is absolutely perfect in laying out in the first 90%. Unfortunately, the second 90% it takes to finish the thing.
Apps, I have successfully finished with LLMs, were code reviewed by humans constantly, they were architected by human (me), they were deployed to production early and big parts were implemented, or refactored by hand.
I will be happy to try Claude Code again in 6 months from now. Until then, I got back to Cursor and Calva.
Clojure macros have two modes: avoid them at all costs/do very basic stuff, or go absolutely crazy.
Here’s the problem: I’m working on Humble UI’s component library, and I wanted to document it. While at it, I figured it could serve as an integration test as well—since I showcase every possible option, why not test it at the same time?
This is what I came up with: I write component code, and in the application, I show a table with the running code on the left and the source on the right:
It was important that code that I show is exactly the same code that I run (otherwise it wouldn’t be a very good test). Like a quine: hey program! Show us your source code!
This macro accepts code AST and emits a pair of AST (basically a no-op) back and a string that we serialize that AST to.
This is what I consider to be a “normal” macro usage. Nothing fancy, just another day at the office.
Unfortunately, this approach reformats code: while in the macro, all we have is an already parsed AST (data structures only, no whitespaces) and we have to pretty-print it from scratch, adding indents and newlines.
I tried a couple of existing formatters (clojure.pprint, zprint, cljfmt) but wasn’t happy with any of them. The problem is tricky—sometimes a vector is just a vector, but sometimes it’s a UI component and shows the structure of the UI.
And then I realized that I was thinking inside the box all the time. We already have the perfect formatting—it’s in the source file!
So what if... No, no, it’s too brittle. We shouldn’t even think about it... But what if...
What if our macro read the source file?
Like, actually went to the file system, opened a file, and read its content? We already have the file name conveniently stored in *file*, and luckily Clojure keeps sources around.
In any other language, this would’ve been a project. You’d need a parser, a build step... Here—just ten lines of code, on vanilla language, no tooling or setup required.
Sometimes, a crazy thing is exactly what you need.
Following the maturing of the Noj toolkit for Clojure data science, we are planning a workshop for people who are curious to learn the Clojure language for data analysis. Please share this page broadly with your friends and groups who may be curious to learn Clojure at this occasion.
The SciNoj Light conference schedule is emerging these days, with a fantastic set of talks. We want a broader audience to feel comfortable joining, and thus we wish to run a prep workshop one week earlier.
Following the maturing of the Noj toolkit for Clojure data science, we are planning a free online workshop for people who are curious to learn the Clojure language for data analysis. Please share this page broadly with your friends and groups who may be curious to learn Clojure at this occasion.
The SciNoj Light conference schedule is emerging these days, with a fantastic set of talks. We want a broader audience to feel comfortable joining, and thus we wish to run a prep workshop one week earlier.
A deep overview of Clojure Sublimed, Socket REPL, Sublime Executor, custom color scheme, clj-reload and Clojure+. We discuss many usability choices, implementation details, and broader observations and insights regarding Clojure editors and tooling in general.
Hey folks! I&aposve spent the past quarter working on jank&aposs error messages. I&aposve focused on reaching parity with Clojure&aposs error reporting and improving upon it where possible. This has been my first quarter spent working on jank full-time and I&aposve been so excited to sit at my desk every morning and get hacking. Thank you to all of my sponsors and supporters! You help make this work possible.
The pull toward JavaScript has never been stronger. While ClojureScript remains an extremely expressive language, the JavaScript ecosystem continues to explode with tools like v0, Subframe & Paper generating entire UI trees and even full websites.
I found the feedback loop of these tools extremely quick and often use v0 to prototype specific components or interactions.
To benefit from these new tools and development experiences in an existing ClojureScript codebase you have two options:
Rewrite all the code to CLJS
Somehow use it as JS
In reality what I do is a bit of both. I mostly translate components to UIx but sometimes will use JavaScript utility files as is. This post is about that second part.
(I’ll probably write about the first part soon as well!)
The shadow-cljs JS import toolchain
shadow-cljs, the de facto frontend for the ClojureScript compiler, has built-in support for importing JavaScript .js files directly into your ClojureScript codebase.
Recently this was helpful when I wanted to add a custom d3-shape implementation to a codebase. I experimented in v0 until I had the desired result, leaving me with rounded_step.js:
// rounded_step.js
function RoundedStep(context, t, radius) {
this._context = context;
this._t = t; // transition point (0 to 1)
this._radius = radius; // corner radius
}
RoundedStep.prototype = {
// ... implementation details full of mutable state
};
const roundedStep = function (context) {
return new RoundedStep(context, 0.5, 5);
}
export { roundedStep };
Now this code would be kind of annoying (and not very valuable) to rewrite to ClojureScript. I tried briefly but eventually settled on just requiring the JS file directly:
Note the path /app/atoms/charts/rounded_step - shadow-cljs understands this refers to a JavaScript file in your source tree and will look for it in on the classpath.
Assuming you have :paths “src” then the file would be at src/app/atoms/charts/rounded_step.js.
When to use JavaScript directly
While I generally will still translate components to UIx (using these instructions) using plain JS can be nice in a few cases:
Code relying on mutability - some library APIs may expect it and it’s usually a bit annoying and perhaps even error prone to articulate in CLJS
Hard to translate syntax constructs - spreading operators, async/await, etc.
Performance - If you want to drop down a level to squeeze out higher performance
“Leaf nodes” only, meaning you can’t require CLJS from JS and things like that. (Fine for my use cases.)
Making it work
Generally when dealing with JS libraries, the following has been helpful for my workflow:
Use js-interop libraries - applied-science/js-interop and cljs-bean make working with JavaScript objects more ergonomic
Use literal objects - The j/lit macro makes passing complex configuration objects cleaner
The payoff
The real benefit? You get to use the best of both worlds:
ClojureScript's expressive syntax, immutable data structures and functional approach where and when you want it
Plug in JavaScript snippets when it makes sense
Less friction when adopting new JavaScript tools
Some folks will be arguing for pure ClojureScript solutions to everything. But in today's landscape, embracing JavaScript interop is the pragmatic choice.
After all, sometimes the best code is the code you don't have to write.