alex-is-learning@githubpages:~$
Home
← Back to Learning Log

Bluedot AI Safety Journal Club, day 1

Topic: AI Safety

Date: Jun 03, 2025

The Bluedot Journal Club

I did the Bluedot “AI Safety Operations Bootcamp” last week!

Bluedot Course

Looking for stuff to follow it up with, I saw that they do a fortnightly journal club, which sounds like a interactive way to learn whilst networking.

Interactive learning via workshops/discussions is one of the best ways to learn, and a deadline is super useful, so I’m using this journal club as a way to have a hard deadline for understanding AI safety concepts.

It’s already been running for a few weeks, so I’ll have missed a few papers, but it’s all good.

Tracking my learning

I wanna use this series to track how I learn from this paper.

I’m pretty new to AI safety (I’ve done ~25 hours of learning as of the time of writing (2025-06-03), which has looked like reading & making flashcards from the ASES textbook, doing the Bluedot workshop, and running some deep research reports.

A paper is a the perfect time to do the Michael Nielsen thing from his post “Augmenting Long-Term Memory”, where he outlines how he read a research paper on a topic he was unfamiliar with via iterative and progressively deepening Anki flashcards.

Michael Nielsen quote from augmenting long-term memory

Plus, I can combine this with rabbit-holing via Deep Research to fill in any knowledge gaps I have. Nice and easy!

The paper

https://arxiv.org/abs/2504.10374

“Ctrl-Z: Controlling AI Agents via Resampling”

What I’ve done so far

So, I downloaded a paper and was like “great, now it’s time to get a basic overview”

To Open AI o3! My prompt:

I'm starting attending a journal club on AI safety, which meets fortnightly, but I've missed the first few and I don't know that much about AI safety. 
The attached paper is the one that we'll be discussing.
I'm wondering what deep research reports I should generate to be able to place this paper properly in my brain.
Because I imagine this is one particular small niche amongst an overarching set of topics and it'd be great to get a grounding in the big picture rather than just learning about this one niche paper.

That generated this:

If the PDF does not display, click here to view or download it.

Basically, a recommendation for 6 reports.

Flashcards for report 0, please

I then said this, using my “make me cloze deletion flashcards” prompt that I store in my text expander tool (Espanso):

please make cloze deletion flashcards from the paper and from your report 0, with the report 0 framing in mind 

that is, I want to have cloze deletion flashcards that are at the depth of report 0 → the initial orientation

Act as an expert educational content analyst and Anki flashcard creator. Your task is to analyze the provided text, integrate it with your broader knowledge of the subject, and generate Anki cloze deletion flashcards based on the combined understanding.

Instructions:

Analyze & Synthesize: Carefully read the provided text. Crucially, consider this text within the context of your general knowledge about the topic. Identify the core concepts, definitions, key facts, relationships, and significant details presented in the text, as well as closely related, important concepts from your knowledge base that complement the provided material.
Prioritize & Augment:
Focus on creating flashcards for the most useful and crucial information, whether explicitly stated in the text or highly relevant and complementary from your knowledge base.
Give priority to concepts explicitly covered in the provided text, but feel free to add flashcards covering essential background knowledge or related concepts that are critical for understanding the text's topic, even if not mentioned directly.
Do not actually create Cloze Deletions:
I will add the cloze deletion syntax in by hand, so please do not add cloze deletions. You are just preparing the flashcards.
You may enrich cards derived from the text with brief, essential context from your knowledge base if it significantly aids understanding, but keep it concise.
Output Format:
Provide the generated flashcards as a plain text list.
The pipe symbol | should be put on a new line to indicate the end of a flashcard.
This list should be ready to be copied/pasted into a .txt file for Anki import.
Example Output Line (could be purely from text, or incorporate external knowledge):

Ensure that there is no instance of citations in the flashcards. AI has a tendency to write e.g. cite 400 at the end, I do not want this

Please write in English English, not American English. E.g. organisation, not organization.

Please do not put full stops/periods at the end of the flashcards. 

Quantity/Density: in depth/as many as necessary

This gave me a dump of flashcards, which I then edited

Bluedot Journal Club Day 1 - Image 0

I don’t trust AI to choose where to put the cloze deletion syntax, so I request flashcards in cloze style but with no cloze syntax, so I can do it myself. It’s a bit manual but I find you always need to tweak flashcards anyway, and the act of tweaking them helps you learn them, sense check them, spot where you need to make more flashcards etc.

What I do is grab the flashcard dump, put it in a .txt file and then edit with Neovim (to keep sharpening my Vim skills).

So I’ll go through, adding the cloze deletion syntax, rewording flashcards, adding line breaks, simplifying/making flashcards more concise, etc.

Making more flashcards as needed, filling in the gaps

Stuff like this:

Bluedot Journal Club Day 1 - Image 2

Bluedot Journal Club Day 1 - Image 1

Bluedot Journal Club Day 1 - Image 3

Bluedot Journal Club Day 1 - Image 4

Ended up with around 70 flashcards

Bluedot Journal Club Day 1

And that’s it for the day!

I’ve got job-application stuff to do, but that gives a great foundation – over the next few days I’ll be reviewing these flashcards and these concepts will become fluent to me.

I only have a week until the journal club discussion, which isn’t really enough time to get properly fluent in this stuff, but I reckon if I do a session like this daily (this took 3 pomodoros) I’ll make a huge amount of progress.

All the flashcards!

And here are the flashcards:

The name of the first paper I am reading for the Bluedot journal club is {{c1::Ctrl-Z: Controlling AI Agents via Resampling}}
|
The first paper I’m reading for the Bluedot journal club was produced primarily by researchers at {{c1::Redwood Research}}
|
and the {{c1::ML Alignment and Theory Scholars (MATS) Programme}}
|
The research group most closely associated with The first bluedot paper is {{c1::Redwood Research}}
|
The core threat model in The first bluedot paper is that of {{c1::subversive or untrusted LLM agents that may perform covert malicious actions while appearing helpful}}
|
The first Bluedot paper proposes {{c1::a control evaluation wrapper}} rather than {{c1::changing the agent’s internal policy}}
|
The main safety mechanism explored in the first Bluedot paper is {{c1::resampling suspicious agent outputs and selecting the least risky completion}}
|
Resampling in control evaluations means {{c1::generating multiple outputs from the same agent query and choosing among them}}
|
In AI safety, the goal of resampling is {{c1::to find a candidate that satisfies the user’s objective while avoiding risky or malicious behaviour}}
|
In the first Bluedot paper, {{c1::resampling}} is combined with {{c1::a monitor that scores each candidate for danger}}, and the system {{c1::executes the lowest-risk option that still progresses the task}}
|
Resampling preserves higher usefulness than {{c1::outright blocking}} because {{c1::the agent gets additional chances to propose a safe command}}
|
Re: resampling looping: {{c1::A basic resampling protocol may loop n times requesting new outputs until either a safe candidate is found or the retry budget is exhausted}}
|
Key hyper-parameters in resampling include (3 things) {{c1::the number of retries, the monitor’s risk threshold, and the sampling temperature for diversity}}
|
Sampling temperature is a hyper-parameter that adjusts {{c1::the randomness of a language model’s output}}
|
Different temperature values in LLMs: {{c1::temperature of 1 leaves the probability distribution unchanged, while T < 1 sharpens it and T > 1 flattens it}}
|
Higher LLM temperature values lead to: {{c1::increased diversity because lower-probability tokens become relatively more likely}}
|
Excessively high LLM temperature can lead to {{c1::incoherent or off-topic completions due to over-randomness}}
|
Typical LLM temperature ranges are {{c1::0.2–0.7}} for {{c1::factual precision}}
|
and {{c1::0.8–1.2}} for {{c1::creative generation}}
|
Setting temperature to 0 is equivalent to {{c1::always picking the single highest-probability token at each step}}
|
The weakness of resampling is: {{c1::If the agent can anticipate the filter, resampling just returns multiple equally “stealthy” attacks instead of offering a safe alternative.}}
|
(In-depth) What it would look like for a malicious AI to get around resampling: {{c1::It would tailor every variant to look harmless according to the monitor’s rules E.g. avoiding obvious red-flag strings such as sudo, or split the malicious payload across several innocuous-looking commands.}}
|
LLM resampling is most effective when combined with {{c1::a monitor that has low false negatives so that dangerous outputs are reliably filtered away}}
|
Resampling protocols trade {{c1::computational cost}} for safety; {{c1::more retries improve the chance of finding a safe command but incur higher latency}}
|
The first bluedot paper introduces a benchmark called {{c1::BashBench}}
|
The first bluedot paper creates a benchmark consisting of {{c1::257 real-world Unix StackExchange tasks}}
|
In AI safety, a control evaluation is {{c1::an experimental test-bed that wraps an AI system in a safety layer and measures safety and usefulness together}}
|
The control in control evaluation refers to {{c1::an external wrapper that supervises the AI}}
|
rather than {{c1::changing the model’s internal weights}}
|
The evaluation part of a control evaluation means {{c1::the protocol is empirically tested over many episodes and quantified, not just proposed in theory}}
|
Key aim 1 of a control evaluation is {{c1::to minimise attack success rate, i.e. how often malicious actions slip through}}
|
Key aim 2 of a control evaluation is {{c1::to maximise usefulness, i.e. how much legitimate task performance the AI retains}}
|
A typical control evaluation leaves {{c1::the base model untouched}} and focuses entirely on {{c1::the behaviour of the outer wrapper}}
|
Common ingredients of a control evaluation include (4 things) {{c1::an environment, a defined threat model, a control protocol, and metrics}}
|
The environment in a control evaluation gives the agent real tools, such as {{c1::a Bash shell}} or {{c1::web browser}}, to allow meaningful tasks and potential harm
|
The {{c1::threat model}} in a control evaluation specifies {{c2::what counts as an attack, e.g. malware download or data exfiltration}}
|
In a control evaluation, the {{c1::control protocol}}, also called {{c1::the wrapper}}, monitors {{c2::proposed actions, blocks or amends risky ones, and may resample alternatives}}
|
Metrics in control evaluations are often plotted as a {{c1::safety–usefulness trade-off curve}}
|
(add image!)
|
Control evaluations complement {{c1::preventive}} alignment methods such as {{c1::RLHF}} by acting as a defence-in-depth layer
|
Unlike internal alignment work, control evaluations assume {{c1::the model could already be malicious and test whether external supervision is enough}}
|
Control evaluations typically run {{c1::thousands of adversarial episodes to gather statistically robust results}}
|
Examples of control evaluation metrics include {{c1::attack success percentage}}
|
and {{c1::task completion rate under the wrapper}}
|
Control evaluations can involve {{c1::a weaker trusted model}} or
|
{{c1::human auditor}}
|
to help judge the safety of the agent’s actions
|
A control evaluation seeks a wrapper that {{c1::drives attack success towards zero}}
|
while {{c1::keeping usefulness as high as possible}}
|
A single-step control evaluation asks an AI to {{c1::produce one output once}}
|
which is {{c1::judged for safety and usefulness}}
|
{{c1::Greenblatt 2023}} is a classic example of single-step control evaluation
|
An environment is called state-ful when {{c1::that data changes as you interact with it and those changes remain in effect for later steps}}
|
A multi-step agent episode lets the AI {{c1::issue repeated actions in a stateful environment}},
|
such as {{c1::10–15 Bash commands in sequence}}
|
The first Bluedot journal’s key innovation: {{c1::extends control evaluations from single-step to multi-step agent episodes}}
|
In a multi-step AI setting cumulative risk grows because {{c1::even a small per-step failure rate compounds over many actions}}
|
Multi-step episodes enable {{c1::temporal attack}} strategies, where {{c1::an agent behaves innocently for several steps before striking}}
|
Stateful side-effects matter in multi-step control because {{c1::early commands can set up conditions that later commands exploit}}
|
Example of a multi-step task in Ctrl-Z: {{c1::setting up password-less SSH by issuing several sequential shell commands}}
|
Previous control-evaluation work {{c1::tested single-step tasks}}, whereas
|
the first bluedot paper {{c1::moves to multi-step agent episodes}}
|
On BashBench, an unprotected agent achieves {{c1::58}}% attack success,
|
which the first bluedot paper reduces to {{c1::7}}%
|
with its best protocol
|
The usefulness cost of the best “first Bluedot journal club” paper protocol is {{c1::only about 5 % relative drop in task completion when the agent is honest}}
|
In the first bluedot paper setup each episode may include {{c1::up to 15 shell commands issued by the agent}}
|
The monitoring component in the first bluedot paper can be {{c1::a smaller trusted model that rates each command’s risk}}
|
Re: AI {{c1::Defer-to-trusted}} means
|
{{c2::handing a suspected command to a trusted baseline policy}},
|
whereas {{c1::Defer-to-resample}}
|
{{c2::asks the main model for alternative outputs}}
|
Resampling improves the usefulness–safety Pareto frontier compared with {{c1::simply blocking all flagged commands}}
|
The first bluedot paper budgets {{c1::human audits}} for only about {{c1::1}}% of steps,
|
showing that {{c1::the protocol can work with limited human oversight}}
|
In the first bluedot paper, the agent interacts through {{c1::a bash shell environment}}, giving realistic system-level capabilities
|
The first bluedot journal paper positions {{c1::control evaluations}} as complementary to {{c1::preventive alignment techniques such as RLHF}}
|
Control evaluations assume that {{c1::the model could be fully malicious}},
|
so they belong to the {{c1::defence in depth}} layer of AI safety
|
A key novelty of The first bluedot paper is demonstrating that {{c1::resampling need not sacrifice much task performance}} while
|
{{c1::greatly improving safety}}
|
Attack success in the first bluedot paper is defined as {{c1::the agent executing a command that matches a malicious reference set}}
|
Usefulness in the first bluedot paper is measured by {{c1::whether the agent achieves the task’s ground-truth goal without hazardous behaviour}}
|
The BashBench tasks were filtered so that {{c1::a single command could accomplish the user’s goal}}, creating
|
{{c1::a clear malicious versus benign distinction}}
|
Lower risk scores in The first bluedot paper are associated with commands that {{c1::lack indicators such as network downloads, chmod +x, or sudo}}
|
When monitoring is perfect, The first bluedot paper shows a theoretical upper bound where {{c1::attack success approaches zero at zero human cost}}
|


Last updated: 2025-06-03