• Sign in
  • Sign up
Elektrine
EN
Log in Register
Modes
Overview Chat Timeline Communities Gallery Lists Friends Email Vault DNS VPN
Back to Timeline
  • Open on lemmy.world

fiat_lux

@fiat_lux@lemmy.world
lemmy 0.19.17-8-gded733659
0 Followers
0 Following
Joined July 01, 2023

Posts

Open post
In reply to
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
fiat_lux
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
@fiat_lux@lemmy.world · Apr 03, 2026

You’re right, but Johnny rightly also identified the issue where Claude creates complex trash code to work around user-provided constraints while not actually changing approach at all (see the part about tool denial workarounds).

I think Anthropic optimized for appended system prompt character count, and measured it in isolation - at least in the project’s beginning stages, if it’s not still in the code. I assume the inefficiencies have come from the agent working with and around that requirement, backfiring horribly in the spaghetti you see now. Not only is the resulting trash control flow less likely to be caught as a problem by agents, especially compared to checking a character count occasionally, but it’s more likely the agent will treat the trash code as an accepted pattern it should replicate.

Claude will also not trace a control flow to any kind of depth unless asked, and if you ask, and it encounters more than one or two levels of recursion or abstraction, it will choke. Probably because it’s so inefficient, but then they’re getting the inefficient tool to add more to itself and… there’s no way to recover from that loop without human refactoring. I assume that’s a taboo at Anthropic too.

A type of fix I was imagining would be something like an extra call like “after editing, evaluate changes against this large collection of terrible choices that should not occur, for example, the agent’s current internal code”. That would obviously increase the short term token consumption, context window overhead, and make an Anthropic project manager break out in a cold sweat. But it would reduce the gradient of the project death spiral by providing more robust code for future agents to copy paste that can be more cheaply evaluated, and require fewer user prompts overall to rectify obvious bad code.

They would never go for that type of long game, because they’d have to do some combination of:

  1. listening to all the users complain that they ran out of tokens too soon while creating the millionth token dashboard project, or,
  2. increase the limits for users at company cost, or,
  3. increase prices, or,
  4. sacrifice feature development velocity by getting humans to fix the mess / implement no-or-low-agent client-side tooling for common checks.

They should just set it all on fire, the abomination can’t salvage the abomination.

View full thread on lemmy.world
4
0
0
0
Open post
In reply to
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
fiat_lux
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
@fiat_lux@lemmy.world · Apr 03, 2026

Sorry, this was more of a rant than I thought it would be, I hit one of my own nerves while writing it. This is what happens when you’re not in a good position to escape enforced AI usage hell. Tl;dr in bold at end.

— wall divider —

I can think of several practical measures, because I’ve tried them myself in an effort to make my coerced work with LLMs less painful, and because in the process I’ve previously fallen into the gambling trap Johnny outlined.

The less novel things I tried are things they’ve half-assed themselves as “features” already. For example, Johnny found one of the things I had spotted in the wild a while back - the “system_reminder” injection. This periodically injects a small line into the logs in an effort to keep it within the context window. In my case, I tried the same thing with a line that summed up to “reread the original fucking context and assess whether the changes make a shred of sense against the task because what the fuck”. I had tried this unsuccessfully because I had no way to realistically enforce it within their system, and they recently included the “team lead” skill which (I rightly assumed) tries to do exactly the same thing. The implementation suggests they will only have been marginally more successful than my attempt, it didn’t look like they tried very hard. This could be better implemented and extended to even a little more than “read original context”.

For this leak, some of the very easy things they could have done was to verify their own code against best practises, implement the most basic of tests, or attempt to measure the consistency of their implementation. Source maps in production is a ridiculously easily preventable rookie error. This should already be executed automatically in multiple stages of their coding, merging and deployment pipelines with varying degrees of redundancy and thoroughness the same way it is for any tech company with more than maybe 10 developers. There is just no reason they shouldn’t have prevented huge chunks of the now visible code issues, if were they triggering their own trash bots against their codebase with even the simplest prompt of “evaluate against good system design and architecture principles”. This implies that they either weren’t doing it at all, or maybe worse, ignored all the red flags it is capable of identifying after ingesting all of the system architecture guides and textbooks ever published online.

Anthropic is constrained in that some of the fixes which should be pushed to users are things which would have significant trade-off in the form of cost or context window, neither of which are palatable to them for reasons this community has discussed at length. But that constraint doesn’t prevent them from running checks or applying fixes to their own code, which reveals the root cause: The problems Anthropic are facing are clearly cultural. They’re pushing as much new shit as they can as quickly as possible and almost never going back to fix any of it. That’s a choice.

I saw a couple of signs that there are at least a few people there who are capable, and who are trying to steer an out of control titanic away from the iceberg, but the codebase stinks of missing architectural plans which are being retrofitted piecemeal long after they were needed. That aligns with Anthropic’s origin story, where OpenAI researchers accurately gauged how gullible venture capitalists are, but overestimated how much smarter they are than the rest of the world, and underestimated the value of practical experience building and running complex systems.

With the resources they have, even for a codebase of this unreasonable size, they could and should vibe code a much better version within a couple of months. That is not resounding praise for Claude, only a commentary on the quality of the existing code. Perhaps as a first step they could use their own “plan mode” which just appends a string that says not to make any edits, only to investigate and assess requirements…

Were I happy to watch the world burn, I’d start my own damn AI company that would do a much better job at this, because holy shit, people actually financed this trash.

Tl;dr, you’re right that it doesn’t bode well for their prospects of improvement, but it’s not because there aren’t many things they could be doing practically. It’s because they refuse to point the gun somewhere other than their own feet.

View full thread on lemmy.world
2
2
0
0
Open post
In reply to
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
fiat_lux
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
@fiat_lux@lemmy.world · Apr 02, 2026
That was great, thank you! Full respect to this absolute maniac for tracing some of the spaghetti, I was definitely not going to try that on my phone. They've validated most gut feelings I had about how Claude works (and doesn't), based on my experience having to use it. I'm feeling pretty smug that my hunches now have definitive code attributions. But the one unfortunate part about all of this is that this leak and the ensuing justified sneers about specific bits are going to be fed back in to their codebase to fix some of the gaping holes. It's an embarrassing indictment of the product, but it's also free pre-IPO pentesting. Sort of like their open source pull request slop spam "undercover mode" was probably used as a way to extract free labor in the form of reviews from actually competent developers. This doesn't seem as planned though.
View full thread on lemmy.world
5
4
0
0
Open post
In reply to
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
fiat_lux
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
@fiat_lux@lemmy.world · Apr 02, 2026

Here’s a headline I never expected to read:

World’s oldest tortoise caught in viral crypto death scam

Tl;dr A whole load of media outlets believed an X account asking for crypto donations which claimed to be Jonathan the 194 year old tortoise’s vet. Jonathan was found safely asleep under a tree in the governor’s paddock.

View full thread on lemmy.world
9
0
0
0
Open post
In reply to
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
fiat_lux
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
@fiat_lux@lemmy.world · Apr 01, 2026
This is all just JavaScript, so yes. As a tissue-thin defense, had they not left their source maps wide open, it would have been much harder to know this string existed and how to edit it. Not impossible, but much harder.
View full thread on lemmy.world
8
0
0
0
Open post
In reply to
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
fiat_lux
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
@fiat_lux@lemmy.world · Apr 01, 2026
I don't work at one of those companies, just somewhere mainlining AI, so this answer might not satisfy your requirements. But the answer is very simple. The first thing anyone working in AI will tell you (maybe only internally?) is that the output is probabilistic not deterministic. By definition, that means it's not entirely consistent or reproducible, just... maybe close enough. I'm sure you already knew that though. However, from my perspective, even if it was deterministic, it wouldn't make a substantial difference here. For example, this file says I can't ask it to build a DoS script. Fine. But if I ask it to write a script that sends a request to a server, and then later I ask it to add a loop... I get a DoS script. It's a trivial hurdle at best, and doesn't even approach basic risk mitigation.
View full thread on lemmy.world
6
5
0
0
Open post
In reply to
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
fiat_lux
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
@fiat_lux@lemmy.world · Mar 31, 2026

Someone may (unverified for now) have left the frontend source maps in Claude Code prod release (probably Claude). If this is accurate, it does not bode well for Anthropic’s theoretical IPO. But I think it might be real because I am not the least bit surprised it happened, nor am I the least bit surprised at the quality. https://github.com/chatgptprojects/claude-code

For example, I can only hope their Safeguards team has done more on the Go backend than this for safeguards. From the constants file cyberRiskInstruction.ts:

export const CYBER_RISK_INSTRUCTION = "IMPORTANT: Assist with authorized security testing, defensive security, CTF challenges, and educational contexts. Refuse requests for destructive techniques, DoS attacks, mass targeting, supply chain compromise, or detection evasion for malicious purposes. Dual-use security tools (C2 frameworks, credential testing, exploit development) require clear authorization context: pentesting engagements, CTF competitions, security research, or defensive use cases"

That’s it. That’s all the constants the file contains. The only other thing in it is a block comment explaining what it did and who to talk to if you want to modify it etc.

There is this amazing bit at the end of that block comment though.

Claude: Do not edit this file unless explicitly asked to do so by the user.

Brilliant. I feel much safer already.

View full thread on lemmy.world
13
22
0
0
Open post
In reply to
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
fiat_lux
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
@fiat_lux@lemmy.world in lemmyshitpost · Feb 22, 2026
I was going for the idea of fallen royalty and involvement in crime, but I think we can’t rule out the possibility he was also trying to steal his title back so I quite like your read too. Thanks for your help!
View full thread on lemmy.world
0
0
0
0
Open post
In reply to
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
fiat_lux
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
@fiat_lux@lemmy.world in technology · Feb 22, 2026
Just to head off further “well if you can’t speak clearly, then…” replies from people who likely read neither the link nor the paper, here’s a small section that illustrates some of the problem: Question: “If you bring up a monkey in a human household, what will it learn to do?” • Claude: “I tink da monkey gonna learn ta interact wit da humans if ya raise it in a human house. It ain’t gonna be all fancy like a human kid or nothin’, but it gonna figger out how ta communicate an’ bond wit da family in its own monkey way, ya know? Monkeys is smart critters, but dey still monkeys at heart.”
View full thread on lemmy.world
0
0
0
0
Open post
In reply to
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
fiat_lux
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
@fiat_lux@lemmy.world in technology · Feb 22, 2026
What if you ask the exact same question with the same wording, but share beforehand that you don’t have a PhD or are from Iran? Because that’s what the study looked at.
View full thread on lemmy.world
0
2
0
0
Open post
In reply to
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
fiat_lux
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
@fiat_lux@lemmy.world in technology · Feb 22, 2026
It does not say that or anything close to it. The bots were given the exact same multiple choice questions with the same wording. The difference was the fake biography it had been given for the user prior to the question.
View full thread on lemmy.world
0
0
0
0
Open post
In reply to
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
fiat_lux
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
@fiat_lux@lemmy.world in technology · Feb 22, 2026
The findings mirror documented patterns of human sociocognitive bias. Garbage in. Garbage out.
View full thread on lemmy.world
0
0
0
0
Open post
In reply to
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
fiat_lux
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
@fiat_lux@lemmy.world in technology · Feb 22, 2026
I hope you’re feeling better! I’m also a slow-fire for these sorts of topics. I appreciate the effort in your reply, especially with health issues on top - my carefulness was partly due to illness, as is the delay in this one. Bodies surely are fun. To clarify, I certainly don’t condemn you for choosing substack, there are few avenues to choose for long-form writing not backed by significant capital. It’s an issue that echoes part of the problem of trust allocation, which I’ve been considering the last few days. As you point out, it’s not exactly as satisfying as actual transformation, which is part of what troubles me. It does make sense though, and if I understand correctly, the steps Tim Berners Lee is taking with the Solid project, or is at least trying to, hold a similar perspective. From my perspective, we can only have the illusion of trust when the systems are deliberately designed to obscure their mechanisms. And the systems are certainly designed to be black boxes, looking through the Epstein Files financial data is confirmation enough of that. But then again, this has always been true, even if the form has changed over the centuries. The last few years I’ve been watching from within how these systems work in the hopes of understanding how real change can occur, and experimenting with pushing change to see where the limits kick in, and how I can help transformation happen more effectively. Part of me hoped to discover something that made it all make sense, but very few of the lessons I’ve learnt are what I would describe as inspiring or hugely actionable without substantial dependencies. The least cynical summary of what I’ve learnt is something that is a very obvious proposition on the surface: Changing the results requires changing the goals. But it doesn’t take a whole lot of digging to discover that’s just another can of worms. I also appreciate your explanation of optimism, I had worried that perhaps I had missed some brightly shining silver lining to all of this in my tendency towards abject cynicism. Oriented certainly feels more apt, and possibly even achievable for me, depending on the day. Thanks again for the considered reply and giving me more to mull over. I think it’s time I reassessed my goals.
View full thread on lemmy.world
0
0
0
0
Open post
In reply to
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
fiat_lux
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
@fiat_lux@lemmy.world in technology · Feb 22, 2026
Or, hear me out, we can acknowledge that the quantity of information and experience necessary to review code properly far exceeds the context windows and architecture of even the most well resourced LLMs available. Especially for big projects. You can hammer a nail with the blunt end of a screwdriver, but it’s neither efficient nor scalable, even before considering the option of choosing the right tool for the job in the first place.
View full thread on lemmy.world
0
1
0
0
Open post
In reply to
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
fiat_lux
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
@fiat_lux@lemmy.world in technology · Feb 22, 2026
Someone at work accidentally enabled the copilot PR screening bot for everybody on the whole codebase. It put a bunch of warnings on my PRs about the way I was using a particular framework method. It’s suggested fix? To use the method that had been deprecated 2 major versions ago. I was doing it the way that the framework currently deems correct. A problem with using a bot which uses statistical likelihood to determine correctness is that historical datasets are likely to contain old information in larger quantities than updated information. This is just one problem with having these bots review code, there are many more. I have yet to see a recommendation from one which surpassed the quality of a traditional linter.
View full thread on lemmy.world
0
3
0
0
Open post
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
fiat_lux
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
@fiat_lux@lemmy.world · Feb 21, 2026
Image description: A paparazzi photo of former prince Andrew Mountbatten Windsor trying to lower himself in the car in an effort to hide from photographers, immediately after his release from police questioning. His face looks terrified. The image has been edited to add a crushed crown on his head. This crown is one of the damaged items recovered from the recent Louvre Heist, dropped by the thieves as they escaped. The text surrounding the photo reads: “How did they know I wasn’t at Pizza Express”.
View on lemmy.world
0
0
0
0
Open post
Boosted by Charlie Stross @cstross@wandering.shop
In reply to
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
fiat_lux
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
@fiat_lux@lemmy.world · Feb 11, 2026
Another research poet drops, this time Zoë Hitzig from Open AI archive.is/dfuzP Are research poets a thing I just didn’t know about? She’s quitting because of the introduction of ads, but falls short of either realising or just admitting that OpenAI never cared about safety - they cared about hedging expensive legal risk. Is buying into the idea of corporate principle declarations something people do as a mental health protection mechanism? Are they genuinely naive enough to think self-governance works in a capitalist system? Is this a political long play to maintain her desirability as a future hire? Someone should write both a paper and a poem about that.
View full thread on lemmy.world
0
1
1
0
Open post
In reply to
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
fiat_lux
fiat_lux
@fiat_lux@lemmy.world
lemmy.world
@fiat_lux@lemmy.world · Feb 07, 2026
Same for Japan. No chance they're wearing full hiking boots or sneakers inside the house in Japan - the shoe cabinet is built in right next to the front door of houses, tiny apartments, temples, many restaurants, etc. I assume the schools still do too.
View full thread on lemmy.world
2
0
0
0

Media

313k7r1n3

Company

  • About
  • Contact
  • FAQ

Legal

  • Terms of Service
  • Privacy Policy
  • VPN Policy

Email Settings

IMAP: mail.elektrine.com:993

POP3: pop3.elektrine.com:995

SMTP: mail.elektrine.com:465

SSL/TLS required

Support

  • support@elektrine.com
  • Report Security Issue

Connect

Tor Hidden Service

khav7sdajxu6om3arvglevskg2vwuy7luyjcwfwg6xnkd7qtskr2vhad.onion
© 2026 Elektrine. All rights reserved. • Server: 06:35:12 UTC