This document is a prologue to the Gall guide. If you've worked though Hoon School (or have otherwise learned the basics of Hoon), you'll likely be familiar with generators, but not with all the other parts of the Arvo operating system or the way it fits together. We'll go over the basic details here so you're better oriented to learn Gall agent development. We'll not go into the internal workings of the kernel much, but just what is necessary to understand it from the perspective of userspace.
Arvo and its Vanes
Arvo is the Urbit OS and kernel which is written in Hoon, compiled to Nock, and executed by the runtime environment and virtual machine Vere. Arvo has eight kernel modules called vanes: Ames, Behn, Clay, Dill, Eyre, Gall, Iris, and Jael.
Arvo itself has its own small codebase in /sys/arvo.hoon
which primarily
implements the transition function (State, Event) -> (State, Effects)
for
events injected by the runtime. It also handles inter-vane messaging, the
scry system, and a couple of other things. Most of
the heavy lifting is done by the vanes themselves - Arvo itself typically just
routes events to the relevant vanes.
Each vane has its own state. Gall's state contains the agents it's managing, Clay's state contains all the desks and their files, Jael's state contains all its PKI data, etc. All the vanes and their states live in Arvo's state, so Arvo's state ultimately contains the entire OS and its data.
Here's a brief summary of each of the vanes:
-
Ames: This is both the name of Urbit's networking protocol, as well as the vane that handles communications over it. All inter-ship communications are done with Ames, but you'd not typically deal with it directly in a Gall agent because Gall itself handles it for you.
-
Behn: A simple timer vane. Behn lets your Gall agent set timers which go off at the time specified and notify your agent.
-
Clay: Filesystem vane. Clay is a revision-controlled, typed filesystem with a built-in build system. Your agent's source code lives in Clay. Your agent's source code and relevant files are automatically built and loaded upon installation, so your Gall agent itself would not need to interact with Clay unless you specifically wanted to read and write files.
-
Dill: Terminal driver vane. You would not typically interact with Dill directly; printing debug messages to the terminal is usually done with hinting runes and functions rather than tasks to Dill, and CLI apps are mediated by a sub-module of the
%hood
system agent called%drum
. CLI apps will not be touched on in this guide, but there's a separate CLI Apps guide which covers them if you're interested. -
Eyre: Webserver vane. App web front-ends are served via Eyre. It's possible to handle HTTP requests directly in a Gall agent (see the Eyre Guide for details), but usually you'd just serve a front-end glob via the
%docket
agent, so you'd not typically have your agent deal with Eyre directly. -
Gall: App management vane; this is where your agent lives.
-
Iris: Web client vane. If you want your agent to query external web APIs and the like, it's done via Iris. Oftentimes web API interactions are spun out into threads to avoid complicating the Gall agent itself, so a Gall agent would not necessary deal with Iris directly, even if it made use of external APIs.
-
Jael: Key infrastructure vane. Jael keeps track of PKI data for your ship and other ships on the network. Jael's data is most heavily used by Ames, and since Gall handles Ames communications for you, you'd not typically deal with Jael directly unless your were specifically writing something that made use of its data.
Userspace
Gall agents live in "userspace" as opposed to "kernelspace". Kernelspace is Arvo and its eight vanes. Userspace is primarily Gall agents, generators, threads, front-ends, and all of their related files in Clay. The distinction looks something like this:
By and large, Gall is the userspace vane - the majority of userspace is either Gall agents, or things used by Gall agents. Apart from the agents themselves, there's also:
-
Generators: These are basically scripts. You'll likely already be familiar with these from Hoon School. Aside from learning exercises, their main use is to make interacting with Gall agents easier from the dojo. Rather than having to manually craft
%poke
s to agents, generators can take a simpler input, reformat it into what the agent actually expects, and poke the agent with it. When you do something like:dojo|wipe
in the dojo, you're actually running the/gen/dojo/wipe.hoon
generator and poking the%dojo
agent with its output. -
Threads: While generators are for strictly synchronous operations, threads make it easy to implement sequences of asynchronous operations. Threads are managed by the
%spider
agent. They can be used as mere scripts like generators, but their main purpose is for performing complex IO. For example, suppose you need to query some external web API, then with the data in its response you make another API call, and then another, before finally having the data you need. If one of the API calls fails, your Gall agent is potentially left in a strange intermediary state. Instead, you can put all the IO logic in a separate thread which is completely atomic. That way the Gall agent only has to deal with the two conditions of success or failure. Writing threads is covered in a separate guide, which you might like to work through after completing the Gall Guide. -
Front-end: Web UIs. It's possible for Gall agents to handle HTTP requests directly and dynamically produce responses, but it's also possible to have a static glob of HTML, CSS, Javascript, images, etc, which are served to the client like an ordinary web app. Such front-end files are typically managed by the
%docket
agent which serves them via Eyre. The software distribution guide covers this in detail, and you might like to work through it after completing the Gall Guide.
The Filesystem
On an ordinary OS, you have persistent disk storage and volatile memory. An application is launched by reading an executable file on disk, loading it into memory and running it. The application will maybe read some more files from disk, deserialize them into data structures in memory, perform some computations and manipulate the data, then serialize the new data and write it back to disk. This process is necessary because persistent storage is too slow to operate on directly and the fast memory is wiped when it loses power. The result is that all non-emphemeral data is ultimately stored as files in the filesystem on disk. Arvo on the other hand is completely different.
Arvo has no concept of volatile memory - its whole state is assumed to be persistent. This means it's unnecessary for a Gall agent to write its data to the filesystem or read it in from the filesystem - an agent can just modify its state in situ and leave it there. The urbit runtime writes events to disk and backs up Arvo's state on the host OS to ensure data integrity but Arvo itself isn't concerned with such details.
The result of this total persistence is that the filesystem—Clay—does not have
the same fundamental role as on an ordinary OS. In Arvo, very little of its data
is actually stored in Clay. The vast majority is just in the state of Gall
agents and vanes. For example, none of the chat messages, notebooks, etc, in the
Groups app exist in Clay - they're all in the state of the %graph-store
agent.
For the most part, Clay just stores source code.
Clay has a few unique features - it's a typed filesystem, with all file types
defined in mark
files. It's revision controlled, in a similar way to git. It
also has a built-in build system (formerly a separate vane called Ford, but was
merged with Clay in 2020 to make atomicity of upgrades easier). We'll look at
some of these features in more detail later in the guide.
Desk Anatomy
The fundamental unit in Clay is a desk. Desks are kind of like git repositories.
By default, new urbits come with the following desks included: %base
, %garden
, %landscape
, %webterm
and %bitcoin
.
%base
- This desk contains the kernel as well as some core agents and utilities.%garden
- This desk contains agents and utilities for managing apps, and the home screen that displays other app tiles.%landscape
- This desk contains everything for the Groups app.%webterm
- This desk is for the web dojo app.%bitcoin
- This desk is for the bitcoin wallet app and bitcoin provider.
You'll typically also have a %kids
desk, which is just a copy of %base
from
upstream that sponsored ships (moons in the case of a planet, planets in the
case of a star) sync their %base
desk from. Any third-party apps you've
installed will also have their own desks.
Desks are typically assumed to store their files according to the following directory structure:
desk
├── app
├── gen
├── lib
├── mar
├── sur
├── sys
├── ted
└── tests
app
: Gall agents.gen
: Generators.lib
: Libraries - these are imported with the/+
Ford rune.mar
:mark
files, which are filetype definitions.sur
: Structures - these typically contain type definitions and structures, and would be imported with the/-
Ford rune.sys
: Kernel files and standard library. Only the%base
desk has this directory, it's omitted entirely in all other desks.ted
: Threads.tests
: Unit tests, to be run by the%test
thread. This is often omitted in distributed desks.
This directory hierarchy is not strictly enforced, but most tools expect things to be in their right place. Any of these folders can be omitted if they'd otherwise be empty.
As mentioned, the %base
desk alone includes a /sys
directory containing the
kernel and standard libraries. It looks like this:
sys
├── arvo.hoon
├── hoon.hoon
├── lull.hoon
├── vane
│ ├── ames.hoon
│ ├── behn.hoon
│ ├── clay.hoon
│ ├── dill.hoon
│ ├── eyre.hoon
│ ├── gall.hoon
│ ├── iris.hoon
│ └── jael.hoon
└── zuse.hoon
arvo.hoon
: Source code for Arvo itself.hoon.hoon
: Hoon standard library and compiler.lull.hoon
: Mostly structures and type definitions for interacting with vanes.vane
: This directory contains the source code for each of the vanes.zuse.hoon
: This is an extra utility library. It mostly contains cryptographic functions and functions for dealing with web data like JSON.
The chain of dependency for the core kernel files is hoon.hoon
-> arvo.hoon
-> lull.hoon
-> zuse.hoon
. For more information, see the Filesystem
Hierarchy documentation.
In addition to the directories discussed, there's a handful of special files a
desk might contain. All of them live in the root of the desk, and all are
optional in the general case, except for sys.kelvin
, which is mandatory.
sys.kelvin
: Specifies the kernel version with which the desk is compatible.desk.bill
: Specifies Gall agents to be auto-started upon desk installation.desk.ship
: If the desk is being republished, the original publisher can be specified here.desk.docket-0
: Configures the front-end, tile, and other metadata for desks which include a home screen app.
Each desk must be self-contained; it must include all the mark
s, libraries,
threads, etc, that it needs. The one exception is the kernel and standard
libraries from the %base
desk. Agents, threads and generators in other desks
all have these libraries available to them in their subject.
APIs
You should now have a general idea of the different parts of Arvo, but how does a Gall agent interact with these things?
There are two basic ways of interacting with other parts of the system: by scrying into them, and by passing them messages and receiving messages in response. There are also two basic things to interact with: vanes, and other agents.
- Scries: The scry system allows you to access the state of other agents and
vanes in a read-only fashion. Scries can be performed from any context with
the dotket (
.^
) rune. Each vane has "scry endpoints" which define what you can read, and these are comprehensively documented in the Scry Reference of each vane's section of the Arvo documentation. Agents define scry endpoints in the+on-peek
arm of their agent core. Scries can only be done on the local ship; it is not yet possible to perform scries over the network (but this functionality is planned for the future). There is a separate guide to scries which you might like to read through for more details. - Messages:
- Vanes: Each vane has a number of
task
s it can be passed andgift
s it can respond with in its respective section oflull.hoon
. These might do all manner of things, depending on the vane. For example, Iris might fetch an external HTTP resource for you, Clay might read or build a specified file, etc. Thetask
s andgift
s of each vane are comprehensively documented in the API Reference of each vane's section of the Arvo documentation. - Agents: These can be
%poke
d with some data, which is a request to perform a single action. They can also be%watch
ed, which means to subscribe for updates. We'll discuss these in detail later in the guide.
- Vanes: Each vane has a number of
Here's a simplified diagram of the ways an agent can interact with other parts of the system:
Things like on-poke
are arms of the agent core. Don't worry about their
meaning for now, we'll discuss them in detail later in the guide.
Inter-agent messaging can occur over the network, so you can interact with
agents on other ships as well as local ones. You can only talk to local vanes,
but some vanes like Clay are able to make requests to other ships on your
behalf. Note this summary is simplified - vanes don't just talk in task
s and
gift
s in all cases. For example, requests from HTTP clients through Eyre (the
webserver vane) behave more like those from agents than vanes, and a couple of
other vanes also have some different behaviours. Agent interactions are also a
little more complicated, and we'll discuss that later, but the basic patterns
described here cover the majority of cases.
Environment Setup
Before proceeding with the Gall Guide, you'll need to have an appropriate text
editor installed and configured, and know how to work with a fake ship for
development. Best practices are described in the environment setup
guide. Example agents and other code throughout
this guide will just be committed to the %base
desk of a fake ship, but it's a
good idea to have a read through that guide for when you begin work on your own
apps.