Egregoros

Signal feed

Timeline

Post

Remote status

Context

2

LLM generated comments:

# Create connection to i3
i3 = i3ipc.Connection()

...makes me think of this ancient blogpost from the 2000s about not writing useless comments that's probably disappeared from the Internet by now:

/*****************************
*     Adds 1 to x                         *
******************************/
x = x + 1

Replies

13
@djsumdog @Zergling_man humans make up function names that don't exist too, if you don't let them run the code, don't let them look anything up and ask them to code a few thousands LOC on paper. you can solve most of these issues by giving the LLM ground truth that it can check its work with (i.e., a compiler it can run, a test suite it can run, for network software: a few server implementations it can test against).
@lain @djsumdog @Zergling_man the best fix for hallucinations is better documentation I found web3 libraries like ethers.js hallucinate all the time because they change everything constantly and don’t document it well. Before ai I couldn’t get it right myself. After ai I solved a couple problems I set aside for literally two years bc I couldn’t read their dog shit JavaScript source and figure it out

They're not hallucinations ... the random code generator has no intent. It can't "lie" or "hallucinate." It's randomly guessing the wrong next token.

I hate this human personification bullshit to the weighted random code machine. If you treat them as "not all that great," you might get something useful from them. If you're in the "I haven't written code in 3 months and pass all my code reviews" camp, your co-workers are incompetent and your job will be replaced in 5 years by someone who has to clean up everything.

I usually refuse most agents when they ask to run commands. they don't make i easy to just give the LLM feedback if you run the command (or something close) yourself. The few times I've allowed an agent to run build/test commands on a difficult problem, it usually goes into a big loop that just eats through tokens (even with Claude Opus vi Copilot .. because it's free paid for by my job)

I get Opus 4.5 and GPT-5.2 from work. GPT52 is cheaper (Opus ix 3x usage). I had something recently that ate up a few days worth of tokens in a loop before I stopped it ... but that might have been Claude Sonnet now that I think about it :bunthink: