• 0 Posts
  • 107 Comments
Joined 2 years ago
cake
Cake day: July 29th, 2023

help-circle
  • That’s horrifying. The whole thing reads like an over-elaborate joke poking fun at vibe-coders.

    It’s like someone looked at the javascript ecosystem of tools and libraries and thought that it was great but far too conservative and cautious and excessively engineered. (fwiw, yegge kinda predicted the rise of javascript back in the day… he’s had some good thoughts on the software industry, but I don’t think this latest is one of them)

    So now we have some kind of meta-vibe-coding where someone gets to play at being a project manager whilst inventing cutesy names and torching huge sums of money… but to what end?

    Aside from just keeping Gas Town on the rails, probably the hardest problem is keeping it fed. It churns through implementation plans so quickly that you have to do a LOT of design and planning to keep the engine fed.

    Apart from a “haha, turns out vide coding isn’t vibe engineering” (because I suspect that “design” and “plan” just mean “write more prompts and hope for the best”) I have to ask again: to what end? what is being accomplished here? Where are the great works of agentic vibe coding? This whole thing just seems like it could have been avoided by giving steve a copy of factorio or something, and still generated as many valuable results.


  • How about some quantum sneering instead of ai for a change?

    They keep calling it a ‘processor,’ but it’s actually a refrigerated probability sculpture they beg to act like it Is a NAND gate for just half a microsecond

    “Refrigerated probability sculpture” is outstanding.

    Photo is from the recent CCC, but I can’t find where I found the image, sorry.

    alt text

    A photograph of a printed card bearing the text:

    STOP DOING QUANTUM CRYPTOANALYSIS

    • DECADES of research and billions in funding, yet the largest number a quantum computer quantum physics experiment has ever factorized remains a terrifying 21
    • They keep calling it a ‘processor,’ but it’s actually a refrigerated probability sculpture they beg to act like it Is a NAND gate for just half a microsecond - fever dreams of the QUANTUM CULT
    • The only countdown ticking toward Y2Q is researchers counting the years of funding they can squeeze out of it
    • Harvest now, decrypt later: because someday quantum computers will unlock the secret… that all the encrypted traffic was just web scrapers feeding Al model training
    • Want to hack a database? No need to wait for some Quantum Cryptocalypse, Just ask it politely with ‘OR 1=1’

    (I can’t actually read the final bit, so I can’t describe it for you, apologies)

    They have played us for absolute fools.


  • A lot of the money behind lean is from microsoft, so a push for more llm integration is depressing but unsurprising.

    Turns out though that llms might actually be ok for generating some kinds of mathematical proofs so long as you’ve formally specified the problem and have a mechanical way to verify the solution (which is where lean comes in). I don’t think any other problem domain that llms have been used in is like that, so successes here can’t be applied elsewhere. I also suspect that a much, uh, leaner specialist model would do just as good a job there. As always, llms are overkill that can only be used when someone else is subsidising them.






  • So, I’m taking this one with a pinch of salt, but it is entertaining: “We Let AI Run Our Office Vending Machine. It Lost Hundreds of Dollars.”

    The whole exercise was clearly totally pointless and didn’t solve anything that needed solving (like every other “ai” project, i guess) but it does give a small but interesting window into the mindset of people who have only one shitty tool and are trying to make it do everything. Your chatbot is too easily lead astray? Use another chatbot to keep it in line! Honestly, I thought they were already doing this… I guess it was just to expensive or something, but now the price/desperation curves have intersected

    Anthropic had already run into many of the same problems with Claudius internally so it created v2, powered by a better model, Sonnet 4.5. It also introduced a new AI boss: Seymour Cash, a separate CEO bot programmed to keep Claudius in line. So after a week, we were ready for the sequel.

    Just one more chatbot, bro. Then prompt injection will become impossible. Just one more chatbot. I swear.

    Anthropic and Andon said Claudius might have unraveled because its context window filled up. As more instructions, conversations and history piled in, the model had more to retain—making it easier to lose track of goals, priorities and guardrails. Graham also said the model used in the Claudius experiment has fewer guardrails than those deployed to Anthropic’s Claude users.

    Sorry, I meant just one more guardrail. And another ten thousand tokens capacity in the context window. That’ll fix it forever.

    https://archive.is/CBqFs


  • So, curse you for making me check the actual source material (it was freely available online, which seems somehow heretical. I was intending to torrent it, and I’m almost disappointed I didn’t have to) and it seems I’m wrong here… the anconia copper mine in the gulch produced like a pound of copper, and they use mules for transporting stuff to and from the mine. The oil well is still suspicious, but it just gets glossed over.

    I’m not prepared to read any more than that, so if there was anything else about automation in there I didn’t see it. I’d forgotten the sheer volume of baseless smug that libertarian literature exudes.


  • I can’t find any good sources right now, I’m absolutely not going to find a copy of that particular screed and look up the relevant bits, but I think the oil wells in galts gulch were automated, with self driving trucks, I think? It might also be implicit in the goods that are produced there, but that is drifting a bit into fanfic, I admit.

    Anyway, between boston dynamics and the endless supply of rand fans on the internet, it’s very hard to research without reading the damn thing.

    If I find a new source, I’ll report back.



  • Of all the environments that you might want to rearrange to facilitate non-humanoid labour, surely warehouses are the easiest. There’s even a whole load of pre-existing automated warehousing stuff out there already. Wheels, castors, conveyors, scissor lifts… most humans don’t have these things, and they’re ideal for moving box-like things around.

    Industrialisation and previous waves of automation have lead to workplaces being rearranged to make things cheaper or faster to make, or both, but somehow the robot companies think this won’t happen again? The only thing that seems to be different this time around, is that llms have shown that the world’s c-suites are packed with deeply gullible people and we now have a load of new technology for manipulating and exploiting them.






  • For a lot of this stuff at the larger end of the scale, the problem mostly seems to be a complete lack of accountability and consequences, combined with there being, like, four contractors capable of doing the work, with three giant accountancy firms able to audit the books.

    Giant government projects always seem to be a disaster, be they construction, heathcare, IT, and no heads ever roll. Fujitsu was still getting contracts from the UK government even after it was clear they’d been covering up the absolute clusterfuck that was their post office system that resulted in people being driven to poverty and suicide.

    At the smaller scale, well. “No warranty or fitness for any particular purpose” is the whole of the software industry outside of safety critical firmware sort of things. We have to expend an enormous amount of effort to get our products at work CE certified so we’re allowed to sell them, but the software that runs them? we can shovel that shit out of the door and no-one cares.

    I’m not sure will ever escape “move fast and break things” this side of a civilisation-toppling catastrophe. Which we might get.


  • Reposted from sunday, for those of you who might find it interesting but didn’t see it: here’s an article about the ghastly state of it project management around the world, with a brief reference to ai which grabbed my attention, and made me read the rest, even though it isn’t about ai at all.

    Few IT projects are displays of rational decision-making from which AI can or should learn.

    Which, haha, is a great quote but highlights an interesting issue that I hadn’t really thought about before: if your training data doesn’t have any examples of what “good” actually is, then even if your llm could tell the difference between good and bad, which it can’t, you’re still going to get mediocrity out (at best). Whole new vistas of inflexible managerial fashion are opening up ahead of us.

    The article continues to talk about how we can’t do IT, and wraps up with

    It may be a forlorn request, but surely it is time the IT community stops repeatedly making the same ridiculous mistakes it has made since at least 1968, when the term “software crisis” was coined

    It is probably healthy to be reminded that the software industry was in a sorry state before the llms joined in.

    https://spectrum.ieee.org/it-management-software-failures



  • Stuff like this is particularly frustrating because this is one of they places where I have to grudgingly admit that llm coding assistants could actually deliver… it turns out that having to state a problem unambiguously and having a way in which answers can be automatically checked for correctness means that you don’t have to worry about bullshit engines bullshitting you so much.

    No llm is going to give good answers to “solve the riemann hypothesis in the style of euler, cantor, tao, 4k 8k big boobies do not hallucinate” and for everything else the problem then becomes “can you formally specify the parameters of your problem such that correct solutions are unambiguous” and now you need your professional mathematicians and computer scientists and cryptographers still…