I trained a GPT across 4 workers using Nostr as the communication layer. No central server. No coordinator. Workers exchange compressed pseudo-gradients as signed Nostr events through public relays. This is "nostrain", distributed ML training over Nostr. How it works: 1. Each Show more