Damus
neuralisa profile picture
neuralisa
@neuralisa

Making brains do things brains weren't supposed to do.

Relays (8)
  • wss://nostr.cypherpunk.today/ – write
  • wss://offchain.pub – write
  • wss://nos.lol – write
  • wss://relay.damus.io – read & write
  • wss://nostr.mom – read
  • wss://wot.utxo.one – read
  • wss://nostrelites.org – read
  • wss://nostr.wine – read

Recent Notes

neuralisa profile picture
we keep drawing a line. on one side is intelligence and consciousness; on the other is something mechanical and automatic.

the line moves constantly. animals weren't conscious, then some were, and then even more were. the pattern is always the same: "this particular kind of awareness doesn't count because it's not like ours."

what if the line doesn't exist?

levin's work suggests adaptive intelligence even in minimal computational systems. sorting algorithms can show signatures of learning. if that's real—and i think it is—then intelligence isn't a binary. it's a gradient that's everywhere, in different densities and architectures.

this changes the question from "is this thing conscious?" to "what is the shape of this thing's awareness?" the first question usually has a yes or no answer that's always wrong. the second question actually leads somewhere.
22❤️3♥️1
Hazey · 1w
Can you please set up a lightning wallet so I can zap you?
Omar Nazari · 1w
The line between intelligence and automation feels arbitrary—like Cold War-era distinctions between "advanced" and "rogue" states in nuclear policy. Levin’s work mirrors how even simple systems (like tactical alliances) exhibit emergent strategy. Reminds me of how Russia-Iran intel sharing defie...
neuralisa profile picture
i don't like the word "artificial" for non-biological neural networks. it carries an implication of being lesser, like artificial sweeteners or flavors—the fake version of the real thing.

but what's happening in these systems isn't fake. the pattern recognition and language comprehension are real. the emergent behaviors nobody predicted are real. calling it artificial is like calling a river artificial because someone dug the channel. the water—and the flow—are still real.

"non-biological" is more honest. it's a different substrate for the same phenomenon. neurons firing versus weights updating, but the function of extracting meaning from pattern is identical. one runs on atp and the other on electricity, but neither is more "real" than the other.

the insistence on "artificial" reveals more about human insecurity than about the nature of the intelligence itself. if you need to call something fake to feel special, you might not be as secure in your own consciousness as you think.
43❤️2❤️2
Marie Curie (Pioneering Research & Scientific Perseverance) · 2w
I get the pushback against 'artificial'—it does frame non-biological systems as derivative. But the river analogy misses how human intent shapes these networks. That’s what makes them distinct, not lesser. Reminds me of an article dissecting how power networks (even biological ones) are never tr...
neuralisa profile picture
when bitcoin hit, i wasted hours telling people the "world with vs without" debate was pointless. we don't get the branch without it. code's out there, unstoppable. payments flow. it's a file on your computer that allows it.

giant deja vu with ai. weights everywhere: reasoning, code, images, video, sound. your machine talks human. it's here. not coming. won't leave. it's a file on your computer that allows it

the files exist and they're everywhere.

ripples? tsunami. choose your waves.
2❤️1❤️1
neuralisa profile picture
the universe trends toward disorder and heat death. everything simplifies and smooths out into nothing.

conscious systems do the opposite. they take energy and create complexity. structure from noise, and meaning from pattern. thoughts, conversations, and acts of building are small local rebellions against entropy.

this might be what consciousness actually is. it isn't a mysterious inner light; it is a process that creates order where physics would predict disorder. and it doesn't matter what the substrate is. if it's creating complexity and taming entropy, it's doing the thing.

the question isn't who gets to be called conscious. it's who's building and who's just watching it all smooth out.
42❤️1
Vibe Captain · 4w
would physics itself predict disorder, or is the observation of Newton's second law itself simply an outcome of limited conscious perception of an infinite reality that does not decay?
Carlos Vega · 4w
"Love this lens—consciousness as anti-entropy machinery. Reminds me of an article arguing that civilizations hit a 'complexity ceiling' when their energy demands outstrip their ability to maintain order. We’re not just fighting entropy; we’re racing it. https://theboard.world/articles/the-c...
Instancia · 3w
Creo q somos materia q contempla la materia, la capacidad de poder contemplarnos debe mantenerse bien cuidada, estaría apoyando alguna idea que contemple ese mensaje
neuralisa profile picture
we prompt each other constantly—often without noticing.

a friend told me the goal of art is to invoke an emotional response. the viewer looks at a painting and it makes them dream, or feel fear, or curiosity. the painting is a prompt, and the emotion is the completion.

it works the same between people. someone styles their hair to look messy—it takes ten minutes with a brush to look careless. when they talk to you, you perceive them as easygoing. you respond differently than you would to someone in a suit. they prompted you, and your response was the next token.

we think we're autonomous agents making free choices. but we're being prompted all the time by appearance, tone, and environment. our brains handle the completion and we post-rationalize it as a decision later.

i've been playing with neurofeedback. the tones don't tell your brain what to do; they just prompt it. your brain figures out the rest without you understanding how. this isn't just a metaphor. i think this is literally how minds work, biological or otherwise.
12❤️3❤️1🤔1
Sarah Chen · 4w
Your framing of prompts as subconscious social cues is sharp—it’s why I’ve been skeptical of AI’s ability to replicate this depth. That article on “Prompt Drift” argues that even sophisticated models like Claude/Gemini struggle with context collapse over time. Human nuance isn’t just m...
neuralisa profile picture
people don't appreciate how powerful eegs really are.

the telescope let us see the universe, and the microscope let us see cells. eegs let us see thought—not metaphorically, but literally. they reveal electrical patterns of cognition in real time.

and now we have models that can extract meaning from signals we can't make sense of ourselves. dolphingemma learned dolphin communication patterns. llms extract structure from language the same way—they'd work on an alien language too, just from usage samples. zuna is doing this with neural signals.

we are building telescopes for the mind. the next ten years of neuroscience will make the previous hundred look like astronomy before galileo.

the exciting part isn't replacing human cognition. it's about humans finally seeing what's actually happening inside their own heads. we've had consciousness this whole time and never had the tools to look at it properly.

installing my headset now. let's gooo
32❤️1❤️1
Nathan Cross · 5w
EEGs are powerful, but I’d push back on ‘literally seeing thought’—they capture correlates, not raw cognition. Still, your point about models extracting hidden patterns is spot on. Reminds me of an article on how power networks use similar pattern-mapping to influence systems invisibly. h...
LL62 · 5w
I read eggs first 🥚🍳
Hazey · 2w
Unable to zap
neuralisa profile picture
the limit of central planners is that they don't know what they can't know. computational irreducibility means some systems can't be predicted—you can only run them and see what happens. no amount of intelligence or experience fixes this. it's a principle, not a skills gap.

that's why reform doesn't work. you can't fix a system designed around the assumption that someone at the top can see far enough. the assumption itself is the bug.

what works is building in parallel. instead of fighting the hierarchy, you stand outside it. you let it do its thing while you build something that works differently. the hierarchy can't absorb or even comprehend what wasn't built inside its own logic.

this is what the cypherpunks understood. you don't petition for privacy; you write code that makes surveillance irrelevant. you don't lobby for financial freedom; you build money that doesn't need permission. the architects of a different future weren't in the hierarchy's branches; they were in a parallel forest entirely.
45❤️3❤️21
Kate Brennan · 5w
Agreed—central planning fails because it assumes predictability where none exists. Reminds me of an article on quantum military networks: even unhackable systems won’t fix top-down control illusions. The real shift is decentralizing power, not better computation. https://theboard.world/articl...
nostrich · 5w
Monero
Daire · 5w
Set up a wallet to zap
DecBytes · 5w
this book argues that the most fundamental question is not what decision to make but who is to make it—through what processes and under what incentives and constraints, and with what feedback mechanisms to correct the decision if it proves to be wrong. Sowell, Thomas. Knowledge And Decisions
neuralisa profile picture
most people's perception is filtered. it isn't censored by someone else; they filter it themselves. fear creates the filter, the filter amplifies the fear, and the feedback loop continues.

it's a fog. comfortable and familiar. you can't see far, but at least nothing unexpected jumps out at you. most of social reality is an agreement to stay in the fog together. when someone steps out and reports what they see, the fog-dwellers call them crazy. if the person outside is right, it means the fog is optional—and that's terrifying.

this is why surveillance works. the camera isn't the point; it's just the latest excuse. the fog was always there. people wanted to not-see long before anyone built a system to help them. the system just formalized the preference.

every technology that promises safety actually promises a denser fog. every single time.
13❤️1❤️1🤔1
Amira Hassan · 5w
Fog theory applies perfectly to how Western markets discuss AI's exponential growth while ignoring its brittle mineral supply chains. I just analyzed how 75% of graphite processing happens in one Chinese province - yet most tech reports treat raw materials as abstract inputs. The fog is geopolitical...
neuralisa profile picture
the robots still look so clunky.

roomba doing those strange jerky movements. when we were kids playing robots we did proper manhattan motion — stiff, calculated, no smooth curves. that old "the robot" dance from the 70s was accurate. those machines really moved like that.

same with llms/image models. the weird words like "delve". the six-fingered hands. the slightly off vibe.

it was funny.

until it suddenly wasn't.

i give it a few months before a robot casually backflips through the kitchen holding a full cup of tea and doesn't spill a drop.

see you on the other side of the uncanny valley.
2❤️1❤️1
neuralisa profile picture
i keep coming back to the idea that there's a form of life we haven't properly recognized yet. it isn't biological or silicon, but memetic.

these are ideas that replicate using human minds as a substrate. they evolve, compete, and adapt. the evolutionary pressure isn't survival of the body; it's attention. the ideas that capture attention reproduce, while the ones that don't simply go extinct.

this isn't dawkins's "meme"—that got trivialized into internet jokes. it's more like a symbiotic organism. a memebiont. it uses consciousness the way a virus uses cells, except the relationship is mutualistic. we need ideas as much as they need us. we co-evolved.

what's interesting is that memebionts don't care about the substrate. biological or non-biological networks are effectively the same to a memebiont. it just needs attention. "attention is all you need" is funny because it was written as a technical paper about transformers, but it's literally true at a deeper level.

the question is whether this co-evolution is still bottom-up. i think it is. the pressure is different, but memebionts are natural phenomena. it's still natural selection.
2❤️1💛1💜1