I write about technology at theluddite.org

  • 3 Posts
  • 120 Comments
Joined 1 year ago
cake
Cake day: June 7th, 2023

help-circle

  • Totally agreed. I didn’t mean to say that it’s a failure if it doesn’t properly encapsulate all complexity, but that the inability to do so has implications for design. In this specific case (as in many cases), the error they’re making is that they don’t realize the root of the problem that they’re trying to solve lies in that tension.

    The platform and environment are something you can shape even without an established or physical community.

    Again, couldn’t agree more! The platform is actually extremely powerful and can easily change behavior in undesirable ways for users, which is actually the core thesis of that longer write up that I linked. That’s a big part of where ghosting comes from in the first place. My concern is that thinking you can just bolt a new thing onto the existing model is to repeat the original error.


  • This app fundamentally misunderstands the problem. Your friend sets you up on a date. Are you going to treat that person horribly. Of course not. Why? First and foremost, because you’re not a dick. Your date is a human being who, like you, is worthy and deserving of basic respect and decency. Second, because your mutual friendship holds you accountable. Relationships in communities have an overlapping structure that mutually impact each other. Accountability is an emergent property of that structure, not something that can be implemented by an app. When you meet people via an app, you strip both the humanity and the community, and with it goes the individual and community accountability.

    I’ve written about this tension before: As we use computers more and more to mediate human relationships, we’ll increasingly find that being human and doing human things is actually too complicated to be legible to computers, which need everything spelled out in mathematically precise detail. Human relationships, like dating, are particularly complicated, so to make them legible to computers, you necessarily lose some of the humanity.

    Companies that try to whack-a-mole patch the problems with that will find that their patches are going to suffer from the same problem: Their accountability structure is a flat shallow version of genuine human accountability, and will itself result in pathological behavior. The problem is recursive.


  • Journalists actually have very weird and, I would argue, self-serving standards about linking. Let me copy paste from an email that I got from a journalist when I emailed them about relying on my work but not actually citing it:

    I didn’t link directly to your article because I wasn’t able to back up some of the claims made independently, which is pretty standard journalistic practice

    In my opinion, this is a clever way to legitimize passing off research as your own, which is definitely what they did, up to and including repeating some very minor errors that I made.

    I feel similarly about journalistic ethics for not paying sources. That’s a great way to make sure that all your sources are think tank funded people who are paid to have opinions that align with their funding, which is exactly what happens. I understand that paying people would introduce challenges, but that’s a normal challenge that the rest of us have to deal with every fucking time we hire someone. Journalists love to act like people coming forth claiming that they can do X or tell them about Y is some unique problem that they face, when in reality it’s just what every single hiring process exists to sort out.



  • Investment giant Goldman Sachs published a research paper

    Goldman Sachs researchers also say that

    It’s not a research paper; it’s a report. They’re not researchers; they’re analysts at a bank. This may seem like a nit-pick, but journalists need to (re-)learn to carefully distinguish between the thing that scientists do and corporate R&D, even though we sometimes use the word “research” for both. The AI hype in particular has been absolutely terrible for this. Companies have learned that putting out AI “research” that’s just them poking at their own product but dressed up in a science-lookin’ paper leads to an avalanche of free press from lazy credulous morons gorging themselves on the hype. I’ve written about this problem a lot. For example, in this post, which is about how Google wrote a so-called paper about how their LLM does compared to doctors, only for the press to uncritically repeat (and embellish on) the results all over the internet. Had anyone in the press actually fucking bothered to read the paper critically, they would’ve noticed that it’s actually junk science.


  • Sounds very doable! My friend has an old claw foot tub that he lights a fire under. If you want something a little less country, you can buy on demand electric or propane water heaters and hook your hose up, though I’d expect the electric one wouldn’t be able to keep up at 120v. Hardest part of this project is probably moving the tub. I say go for it!






  • I know that this kind of actually critical perspective isn’t point of this article, but software always reflects the ideology of the power structure in which it was built. I actually covered something very similar in my most recent post, where I applied Philip Agre’s analysis of the so-called Internet Revolution to the AI hype, but you can find many similar analyses all over STS literature, or throughout just Agre’s work, which really ought to be required reading for anyone in software.

    edit to add some recommendations: If you think of yourself as a tech person, and don’t necessarily get or enjoy the humanities (for lack of a better word), I recommend starting here, where Agre discusses his own “critical awakening.”

    As an AI practitioner already well immersed in the literature, I had incorporated the field’s taste for technical formalization so thoroughly into my own cognitive style that I literally could not read the literatures of nontechnical fields at anything beyond a popular level. The problem was not exactly that I could not understand the vocabulary, but that I insisted on trying to read everything as a narration of the workings of a mechanism. By that time much philosophy and psychology had adopted intellectual styles similar to that of AI, and so it was possible to read much that was congenial – except that it reproduced the same technical schemata as the AI literature. I believe that this problem was not simply my own – that it is characteristic of AI in general (and, no doubt, other technical fields as well). T


  • I’ve now read several of these from wheresyoured.at, and I find them to be well-researched, well-written, very dramatic (if a little ranty), but ultimately stopping short of any structural or theoretical insight. It’s right and good to document the shady people inside these shady companies ruining things, but they are symptoms. They are people exploiting structural problems, not the root cause of our problems. The site’s perspective feels like that of someone who had a good career in tech that started before, say, 2014, and is angry at the people who are taking it too far, killing the party for everyone. I’m not saying that there’s anything inherently wrong with that perspective, but it’s certainly a very specific one, and one that I don’t particularly care for.

    Even “the rot economy,” which seems to be their big theoretical underpinning, has this problem. It puts at its center the agency of bad actors in venture capital becoming overly-obsessed with growth. I agree with the discussion about the fallout from that, but it’s just lacking in a theory beyond “there are some shitty people being shitty.”






  • Just because a postcapitalist world should have a battery for every house does not make batteries in and of themselves solarpunk. The story surrounding the battery, in this case, the branding, is actually precisely what matters, because solarpunk is explicitly about speculative futures. It’s a genre of science fiction that creates an optimistic and green aesthetic to aid in imagining a postcapitalist world. Posting a link to a currently existing consumer grade technology with consumerist branding is, by definition, not solarpunk.

    “A good science fiction story should be able to predict not the automobile but the traffic jam.” You’re posting the automobile. Science fiction is about the social context of the technology as much if not more than about the technology itself.

    Again, I’m not saying that personal batteries are bad, or have no part in a postcapitalist future.