Lab

Short, playable thought experiments — each in 60 seconds — at the intersection of ML security, distributed systems, and trustworthy AI. They're not "fun for fun" — each one demonstrates an idea from active research, and you can open an issue to discuss or suggest a new puzzle.

Distributed systems · Consensus

The Two Generals' Problem

Two armies must coordinate to attack at dawn. Their only way to communicate is by messenger across enemy-held territory — where each messenger has a real chance of being captured. Click the buttons below and try to reach certain agreement.

Plan ready · awaiting word from B
Awaiting plan from A
0 sent 0 lost common knowledge: none
What this shows

This is the Two Generals' Paradox. With an unreliable channel, no finite protocol guarantees that both sides are certain the other has agreed. The last messenger could always be the one that's captured — and without knowing the last one arrived, the receiver cannot safely commit.

The implication is foundational. Every distributed system that needs agreement — database commits, blockchain consensus, atomic broadcast — must accept a weaker safety property: agreement with high probability, never with certainty. It's also why "two-phase commit" is fragile and why production systems lean on three-phase, Paxos, or Raft variants tolerant to message loss.

ML security · Model watermarking

Spot the Watermark

Each grid below is the output signature of a trained model — colored by per-cell activation. One of them has a watermark embedded: a faint, distributed perturbation only a verifier with the secret key can detect. Can you guess which?

Score: 0 / 0
What this shows

The watermark is embedded as a small, distributed perturbation across specific "key cells." Without the key, the patterns are statistically indistinguishable. With it, verification is fast — and any attempt to remove the watermark distorts model behaviour in ways a verifier can still catch.

This is the core idea behind feature-based model watermarking, the line of research I published in IEEE Access — Ural & Yoshigoe (2024) and SecurePoL (2025) — which integrates watermarking with proof-of-learning to defend against spoofing attacks on training claims.