Notes
by Matt Stubbs
How to start Phoenix with a node name
26 March 2025
It can be handy to start Phoenix with a node name and cookie. This is often used in distributed systems in production, but I find it handy to use in development with Livebook.
Rather than starting Phoenix with the usual mix phx.server
I use:
elixir --name my_app@127.0.0.1 --cookie dev -S mix phx.server
replacing my_app
with something more appropriate. I leave cookie as dev
because I'm not concerned about security when running locally.
Once the server is running, I connect to it from Livebook:

- Click Runtime Settings, then Configure.
- Enter the
[email protected]
and the cookie into the panel.
Once connected I can interact with the process from Livebook.
Porting a vim theme to Zed with o3-mini
21 February 2025
I came across an editor theme called Everforest. It's available in dark and light variants, and I wanted to try it out.
It's available for vim, but only the dark mode has been ported to Zed.
How hard would it be to get an LLM to create the light theme for me?
To do a quick port I used gitingest to create a big chunk of text with everything from the original vim theme respository.
I also grabbed a copy of the dark theme JSON to give the model an example of the output I wanted.
Then I wrote a quick prompt:
Convert the vim theme Everforest light to work in Zed. I already have the
Everforest dark converted. Create light.json with the same structure and
keys as the dark theme, but use the colour values from the light theme. Here
is the dark theme:
```json
<copy of dark.json>
```
and here is the original theme:
<git respositroy contents>
I saved the cotents to a text file and passed it to o3-mini
using llm
:
cat instructions.txt | llm -m o3-mini
This thought for a while, then produced a JSON file which I saved into my local Zed themes path at ~/.config/zed/themes
.
And it worked! It had the correct structure and pretty good colour values. I needed to tweak a few values to get it right. The status bar was too dark, and the predictive text wasn't quite right.
Considering I've never created a Zed theme before this was a quick way to get a working version up and running, and even if I’d createit by hand I would expect to spend some time tweaking values.
How much did it cost? I checked the usage report in my OpenAI account to estimate how much this cost. About 45k tokens: 33,100 input tokens and 12,778 output tokens. At $1.10/million input tokens and $4.40/million output tokens this cost about $0.08. Pretty good! I could probably get the input token count down a bit by excluding some extraneous files from the GitHub repo. It’s marginal though, and for a one-off task saving 1 cent isn’t worth it.
The original Everforest theme repo includes examples screenshots of the theme in vim. I wonder if including a screenshot would help?
I updated the prompt to include:
I've included a screenshot of the light theme being used in the vim editor.
and ran the prompt again:
cat instructions.txt | llm -m o3-mini -a \
https://user-images.githubusercontent.com/37491630/206063932-6bd60bef-8d2a-40d7-86b9-b284f2fea7b0.png
but o3-mini
doesn’t support image attachments! Currently four OpenAI models have vision capability: o1
, gpt-4o
, gpt-4o-mini
, and gpt-4-turbo
.
How well would this work with gpt-4o
?
I tried it:
cat instructions.txt | llm -m 4o -a \
https://user-images.githubusercontent.com/37491630/206063932-6bd60bef-8d2a-40d7-86b9-b284f2fea7b0.png
This time I got a response back almost immediately. I’m guessing this because o3-mini
is a reasoning model and gpt-4o
is not. o3-mini
spends some time reasoning and planning before returning a response. gpt-4o
just starts completions right away.
The result was better than the first one from o3-mini
. This time the status bar was the correct colour. Only a few small changes but definitley better.
The cost for this one was 32,189 input tokens and 1,952 output tokens. I was surprised to see a lower number of input tokens when the prompt was slighly longer and I’d included an image this time. I assume this is down to a difference in the models and the way they each tokenise content.
No reasoning means less output tokens. gpt-4o
is a more than twice the cost of o3-mini
. Even with the lower number of output tokens the cost was slighly higher at $0.10.
This was a fun experiment. Not only was it quick to get something working but the price/quality ratio was excellent.
Indexing and querying with embeddings and Livebook
20 February 2025
In his video Vector Search Demystified: Embedding and Reranking Zack Siri uses Elixir and Livebook to demonstate how to index and query 100,000 embeddings.
It's a good introduction to preparing data with Explorer and indexing embeddings with Livebook. Zack uses hnswlib which provides an in memory vector store. This is fast and ideal for testing and iterating.
The reranking stage was particularly interesting. Zak uses a reranking model (BAAI/bge-reranker-v2-m3) to improve the order of results from the vector search. I hadn't come across this variety of models before and I'm keen to learn more.
I asked Perplexity to create an introduction to reranking models and it described them like this:
Reranker models, also termed cross-encoders, are specialized neural networks designed to evaluate the relevance of documents to a query through direct pairwise comparisons. Unlike first-stage retrieval systems that rely on vector similarity, rerankers process the raw text of both the query and candidate documents, generating a relevance score that reflects their semantic alignment. This capability allows rerankers to address limitations in initial retrievals, such as mismatches caused by polysemy or syntactic variations.
The Livebook used in the video is available on Github.