Damus
Tiota Sram profile picture
Tiota Sram
@Tiota Sram

#ComputerScience instructor at Wellesley College who researches #games & #AI (he/him; cis). Current #research is on exploration in #Metroidvania games. I build #PCG demos like:

https://cs.wellesley.edu/~pmwh/labyrinfinite/

and games like:

https://cs.wellesley.edu/~pmwh/chlorophyll/

The intertwining problems of racism, ablism, queermisia (including specifically transmisia), sexism, classism, capitalism, and colonialism must be opposed everywhere.

Profile picture: colored lines converging along different paths on a square grid.

tfr

Relays (1)
  • wss://relay.ditto.pub – read & write

Recent Notes

Peter · 1w
also wondering where the people who used to rush into software development just because it's a lucrative career end up. healthcare?
Tiota Sram profile picture
@nprofile1q... at our school economics seems to be a big destination.

It is FASCINATING to observe that the closer the relationship between LLM output and direct monetary gain/loss, the fewer jobs are being replaced.

Get rid of your entry-level coders? Sure, the relationship between code quality/mistakes and money gained/lost is pretty vague. Even estimating the financial impact of a big outage can be difficult.

Get rid of your day traders or financial analysts? For some reason not happening at scale (yes there's a bunch of people doing AI-based investing, including before the advent of LLMs, but at least there isn't the public perception of jobs in that sector drying up).
Artemis · 4w
Problem-solving happens in community. We cannot solve all of our problems with rules & protocols. We cannot simply mechanically remove all harassment. We have to learn to react to things as members of...
Tiota Sram profile picture
@nprofile1q...

IMO we *could* have much better tools to help facilitate social solutions though...

I was just brainstorming about this:

https://kolektiva.social/@tiotasram/116363166024461961

I'm sure some of my ideas of even all of them are terrible, but I'm equally sure good options for improving the tools do exist.
Curtis "Ovid" Poe (he/him) · 5w
I should point out that the structure of the book is: 1. Prove the thesis 2. Show why we keep allowing this 3. How do we prevent it? 9/8
Tiota Sram profile picture
@nprofile1q... as an anarchist, I think this jives with my instinct to blame hierarchies for these problems and oppose capitalism as one hierarchical system among others. I'm not actually that well-read on theory but it shouldn't be hard to find anarchist theorists who have thought through the issue in this way; David Graeber and Kropotkin come to mind as likely scholars to look at.

As for the solution, I think many can agree it requires a change to "human nature" but of those many are pessimistic about that being impossible. I don't share that pessimism. I think that human "nature" is in fact quite malleable over long or sometimes even short time-scales. To me, there are a few key ingredients in human nature required to avoid *a lot* of these problems (surely there would be other problems; I'm not envisioning this as a recipe for utopia):

1. Baseline unwillingness to obey orders from another. Obedience seen as disgusting and/or vaguely disquieting. This makes the task of the random antisocial jerk who wants to be a warlord nearly impossible.
2. Awareness of the risks of hierarchy, and an understanding of how to undermine it. This allows a non-global anarchist society to defeat a hierarchical threat by converting the individuals in the threatening group to their beliefs, ideally in part made easy by materially providing for the needs that the hierarchical society is withholding in order to cement control, and by demonstrating how much more enjoyable a life of free association is.
3. An instinctive impulse to provide for strangers and help one another. This one is already very much present in the current day, just brutally suppressed by states. The fact that it continues to bloom through the cracks of our modern order gives me a lot of hope.
4. That's it, mostly. I think one could argue that another necessary ingredient is some tendency towards spontaneous higher-order organization in order to actually get the work done necessary to feed everyone without hierarchies involved, but I see this as a minor detail.

Of course, *how* to change human nature to get this mix, at sufficient scale, and whether that could happen before hierarchies drive us to extinction is the much harder question.

In any case, as an anarchist I'm not dogmatic about this recipe. Others might have their own ideas about what could succeed, and as long as we don't extinct ourselves (very real possibility on multiple fronts) over millennia I think evolutionary processes at the societal level should produce *something* with longer-term stability than a few measly tens of thousands of years. We do still have existing present-day societies in small corners of the world that have been stable that long, in fact, and they may well simply outlive the present chaos and continue on past it. What I will argue against as an anarchist is pure pessimism. If you don't like my formula, by all means point out specific flaws or advocate for your own, or even day "I don't like that but I'm not sure what you do." Just don't say "That's unrealistic and couldn't possibly succeed, so we must continue with the status quo as the only option," because that means either we have incompatible definitions of "succeed" or you're arguing against any attempts towards changing a failing system because those attempts might fail, which is silly. (Of course, arguing that the failure modes of attempted change could be worse than the failure modes of continuing as is, but that's not a very solid-looking argument right now.)
elilla&, the flute futa · 6w
"AI" users are like, "I know this is imprecise but as a convenience these transcriptions are better than nothing" then 70 years from now we'll still be struggling to debunk these entirely hallucinate...
Tiota Sram profile picture
@nprofile1q... I saw a project doing this for NYC council meetings, and brought up the inevitability of mistranscriptions (and bias in whose testimony would be more accurately transcribed). They pushed back saying it was fine and users could submit corrections. I decided to check whether my hypothesis was correct and mistranscriptions were common. I followed an upthread link to a transcription and started reading through to see how long it would take me to spot an error. Literally the next speaker had serious errors in the transcription.

The kicker: NYC already provides human-taken transcriptions. But these are released "too slowly" and it's "too hard" to even implement a feature to automatically check the machine transcriptions against the human ones when they're available.

The second kicker: in the video with the error, the human scribe interrupted to ask the speaker to speak louder/more clearly so that they could get an accurate transcript.

Anyways I got blocked because the project (which aside from using AI for transcription was pretty cool) was super useful for civic engagement and my objections were intolerable.
Tiota Sram profile picture
I've seen a bunch of "the CA age verification law is the best way to do a bad thing and so we shouldn't oppose compliance" takes, which others are rightly pointing out is a bad stance because it's blindingly obvious that compliance now sets the stage for compliance later and the clearly set up later is mandatory verification of age data. Even if you think that, for example, California's current "progressive" government won't go there, we're all currently seeing just how easy it is for a new government to pick up the oppressive tools the "good" government was using "restraint" with and put them to worse ends.

On the other hand, I'll freely admit that distros *do* need a way to shield themselves from liability right now. The clear (to me; IANAL) correct solution is to say on your website "don't download this OS if you're in a jurisdiction where it's not legal for us to provide it."). Assuming this does put you in the clear liability-wise, it has several positive effects:

- Stops zero people from downloading it.
- Makes it clear that your project will not collaborate with fascists/oppressive regime enjoyers.
- Means that when the next law makes verifying user ages mandatory (and/or explicitly requires using Palantir-adjacent services to do so) you've already got a strategy in place and there's no need for a "debate" in your "community" about compliance.
- Gets users more practice with "the law is malicious/needlessly bureaucratic/oppressive; let's ignore it" which to be honest people in general clearly desperately need at this point.
- Is the most effective political move if you want to resist the way things are going. Forcing the other side to explain why "California bans Linux" is good rhetorical strategy. Make *them* try to explain "well it's actually not so harmful since we let users set it themselves" and answer your follow-up "but what if next year the requirements change; I just refuse to go along with this slippery slope stuff and I'm not bothered if that means you want to *ban* me."

#AgeVerification
MarjorieR · 8w
nostr:nprofile1qy2hwumn8ghj7un9d3shjtnyd968gmewwp6kyqpqk3nz86rqq7fs7qwrvl9258su25qgw6dfp85gq8y2z52culzfgy5s8lqqrn nostr:nprofile1qy2hwumn8ghj7un9d3shjtnyd968gmewwp6kyqpqzwzcgjztdn95g0cnkk0z9mz393xzzc4...
Tiota Sram profile picture
@nprofile1q... @nprofile1q... @nprofile1q... @nprofile1q... also Doctorow didn't just use an LLM for spell-checking, he pivoted to defending LLM use in general and disparaging those against it because he rightly predicted we'd be unhappy about his use of LLMs for spell-checking.

Also: we're unhappy because the technology being pushed on us is harmful in numerous ways. The AI boosters are unhappy because someone won't use the thing they're trying to sell (and/or because they keep getting called out on the harms). These are very different reasons to be unhappy with different levels of validity.
3
Robert Kingett · 8w
nostr:nprofile1qy2hwumn8ghj7un9d3shjtnyd968gmewwp6kyqpqjluflf4kgn5g40qkctq87aqdwavkfkl5nzpwz40zqt7eqa2tv6xs8s7r7v nostr:nprofile1qy2hwumn8ghj7un9d3shjtnyd968gmewwp6kyqpqwyu5tn758hc399kq7z0gcf0s4tqxs8z8rn2me5672pnks4429s2ssnxult nostr:nprofile1qy2hwumn8ghj7un9d3shjtnyd968gmewwp6kyqpqk3nz86rqq7fs7qwrv...
ben · 8w
nostr:nprofile1qy2hwumn8ghj7un9d3shjtnyd968gmewwp6kyqpqjluflf4kgn5g40qkctq87aqdwavkfkl5nzpwz40zqt7eqa2tv6xs8s7r7v nostr:nprofile1qy2hwumn8ghj7un9d3shjtnyd968gmewwp6kyqpqwyu5tn758hc399kq7z0gcf0s4tqxs8z8rn2me5672pnks4429s2ssnxult nostr:nprofile1qy2hwumn8ghj7un9d3shjtnyd968gmewwp6kyqpqk3nz86rqq7fs7qwrv...
KFears · 8w
nostr:nprofile1qy2hwumn8ghj7un9d3shjtnyd968gmewwp6kyqpqjluflf4kgn5g40qkctq87aqdwavkfkl5nzpwz40zqt7eqa2tv6xs8s7r7v nostr:nprofile1qy2hwumn8ghj7un9d3shjtnyd968gmewwp6kyqpqwyu5tn758hc399kq7z0gcf0s4tqxs8z8rn2me5672pnks4429s2ssnxult nostr:nprofile1qy2hwumn8ghj7un9d3shjtnyd968gmewwp6kyqpqk3nz86rqq7fs7qwrv...
Miguel Afonso Caetano · 10w
"With a hammer, harm is external and obvious, like a bent nail, a bruised thumb. In this cognitive environment, harm can be real and insidiously subtle. Once an LLM has offered a line of reasoning or ...
Tiota Sram profile picture
@nprofile1q... fascinating that this is written by an AI booster (look at the front page of the "think tank" he leads). I guess he thinks these dangers are often worth risking...

I'd argue that there are much better metaphors available for this kind of tool: a police department, for example, shapes the way we think about certain kids of problems (and imagined potential problems) through both overt propaganda and even without that just through the "call 911 to solve your problem" and "the threat of incarceration will prevent crime" ideas which... don't actually play out in practice. There are lots of similar social institutions that offer promises or propose frameworks of thought which aren't dependable but which deeply shape our behavior (democracy: "Want to change things? Vote," or even just cars and attendant urban design in US suburbs: "want to go anywhere? Get in the car."). Like LLMs, these things are tools/institutions that change the way we think as a consequence of using them, and those changes often lead us to think or behave in ways that rely on them to our detriment. LLMs do feel like a more generalist form of this phenomenon, since you can "rely on" them for almost any kind of thinking, not just certain very specific uses.
note1epsaf...
Tiota Sram profile picture
@nprofile1q... great points. An article that transformed my thinking on this from Crimethinc was this about it being safer at the front:

https://crimethinc.com/2025/01/28/its-safer-in-the-front-taking-the-offensive-against-tyranny

Obviously this doesn't always apply (it didn't work for Pretti) but speaking out when the real consequence is a beating and short unfair prison stay after which you understand the need to get the hell out is much preferable to remaining silent for "safety" and then a few years later you get picked up anyways but this time it's a one-way trip to a concentration camp.
note1ynml8...
Tiota Sram profile picture
@nprofile1q... @nprofile1q... @nprofile1q... sadly under the current (and not just this year's) FDA these things are getting approval, only to have higher complaint rates than non-AI devices... A quote:

"""
Researchers from Johns Hopkins, Georgetown and Yale universities recently found that 60 FDA-authorized medical devices using AI were linked to 182 product recalls, according to a research letter published in the JAMA Health Forum in August. Their review showed that 43% of the recalls occurred less than a year after the devices were greenlighted. That’s about twice the recall rate of all devices authorized under similar FDA rules, the review noted.
"""

From: https://www.reuters.com/investigations/ai-enters-operating-room-reports-arise-botched-surgeries-misidentified-body-2026-02-09/

But yeah, in the long term, if FDR didn't completely dismantle the FDA, we should see some balancing out.
note1knpxg...
Tiota Sram profile picture
@nprofile1q... @nprofile1q... @nprofile1q... also FYI as someone with a modest understanding of the CS side of things: A few points:

1. This article is not trustworthy. It conflates very different types of AI system to maximize the threat it states, and that kind of dishonest reporting makes me distrust all of its conclusions. Expert diagnosis systems that use machine learning are a completely different kind of beast from a chatbot, even though that also uses machine learning in a different way. Touting the accuracy of specialized diagnosis systems as evidence for the effectiveness of chatbots is like showing off the excavating capability of a Volkswagen excavator add evidence that a BMW racing car will be faster than the competition.

2. Evaluating machine learning systems is very tricky, and it's not hard to get good-looking results in front of press if you have dollars to burn, which don't actually hold up in the real world. For the curious here's a nice thorough paper on issues with medical imaging specifically: https://www.nature.com/articles/s41746-022-00592-y
As a patient, unless you show me in-the-field accuracy numbers from multiple years of deployment that actually rival human performance for a very particular task that this discussing AI is designed to do alone, I'd always rather have a human doctor's opinion than that of an AI system, and even when I'm willing to accept the AI diagnosis, I'd like a human doctor's second opinion and interpretation.

3. For medical-records and patient-interaction applications, including most of the ones listed in this article that chatbots are "good" at, I think doctors/hospitals using them are opening themselves up to a lot of liability and making a mistake. These systems make egregious errors in predictable patterns, which competent human staff do not make. Incompetent staff sometimes make the same errors, but the difference is that they are responsible for their own errors. If your "bedside manners" chatbot that takes over when the doctor is busy with another patient encourages a patient to kill themselves or take the wrong medicine, is that acceptable, since 99% of the time it speaks with a very reassuring manner? I guarantee you ChatGPT will make these mistakes orders of magnitude more often than even the most sleep-deprived RN or resident. We have seen these events happen already in other contexts; the medical/hospital context has many more opportunities for these failures. There is no real technological mitigation for them in the horizon either. Even for "transcribe my patient notes" I wouldn't trust dictation software unless needed as an accommodation, even though it has admittedly gotten pretty good. There's lots of opportunity for a missed jargon word to cause havoc in notes that get shared with someone else, for example.

4. The big AI companies have developed one tool that shows up very well I demos, but which has nasty flaws that make it unsuitable for a lot of what they're pushing. They are trying to sell their stuff as "the future" and say it must be "integrated" everywhere. If only you integrated our marvelous technology, your problems would be solved! This is backwards. A true solution looks first at the problem, and then asks "what tool would be best to use here?" By putting the tool choice first, you end up with ineffective or even counterproductive "solutions." This only makes sense if your goal is to sell the tools. For example, I understand doctors have very limited time but must write up notes between patient visits to refer to in future meetings. Sometimes notes end up inaccurate or illegible, or just muddled. How can we best solve this problem? Simple: hire more doctors to give them all more time to write better notes. Any alternate solution needs to be understood as a compromise. Might there be a way to use technology to help? Sure, there are probably plenty. Let's consider using a system to record the visit, then produce notes automatically by statistically predicting what notes the doctor would write. We'll have the doctor check them off every time. Is this good? No, because in this design, doctors will get lazy in their checking over time, especially if the system is very accurate for most visits. But such systems are going to make big mistakes for unusual visits, which doctors might not then correct. Even worse, by denying the doctor the cognitive task of organizing their thoughts into writing, you're disrupting the doctor's memory formation and chances to see unusual patterns or slight irregularities. Lowering the doctor's cognitive burden takes away the benefits you get from expending cognitive resources on the problem! As an alternate design, what if you had the doctor write the notes unaided, and then had a system try to flag possible discrepancies, misspellings, or illegible writing? Such a system might still be bad, if it creates too much friction (remember Clippy?). But it *might* be good if tuned correctly. It's not a flashy or "revolutionary" as the "we'll take your work away" system, but it avoids some of that system's worst drawbacks. There's probably even better designs I'm not thinking of. My point is that starting with "let's integrate a chatbot" is the wrong approach, and anyone who insists on it is not someone you should trust, because self-evidently they are stating from their own interests (sell/promote chatbots) while completely disregarding yours. They're basically saying "Can you help me think of a way to sell my product to you?" which is downright disrespectful.

Okay that's probably enough ranting from me. TL;DR: trust AI chatbots about as much as the least trustworthy intern you can imagine working with, because it will eventually make exactly the same kinds of disastrous mistakes, and you'll be the one to blame since (when we're not spinning yarns about machine intelligence) it's "just a tool that you chose to use." It's not even capable of learning from those mistakes either, because to let chstbots learn from their public interactions would be dangerous in other ways (see Microsoft Tay).
note1qxxve...
Tiota Sram profile picture
@nprofile1q... if I were the mayor of Minneapolis or St. Paul, I'd enact a local hospitality tax that applied to federal agents. I'd estimate the economic losses caused by their brutal tactics and the resulting fear, and then charge an amount per head to recoup that damage, with 5% to be retained by the hotel, which would collect the fee. Is that legal? I don't know but we're clearly not in an era where asking that question is fashionable. I'm sure even if it's not I could hold out appeals in the courts for a few months or so...