• 13 Posts
  • 153 Comments
Joined 1 year ago
cake
Cake day: July 19th, 2023

help-circle

  • It’s almost completely ineffective, sorry. It’s certainly not as effective as exfiltrating weights via neighborly means.

    On Glaze and Nightshade, my prior rant hasn’t yet been invalidated and there’s no upcoming mathematics which tilt the scales in favor of anti-training techniques. In general, scrapers for training sets are now augmented with alignment models, which test inputs to see how well the tags line up; your example might be rejected as insufficiently normal-cat-like.

    I think that “force-feeding” is probably not the right metaphor. At scale, more effort goes into cleaning and tagging than into scraping; most of that “forced” input is destined to be discarded or retagged.









  • Hallucinations — which occur when models authoritatively states something that isn’t true (or in the case of an image or a video makes something that looks…wrong) — are impossible to resolve without new branches of mathematics…

    Finally, honesty. I appreciate that the author understands this, even if they might not have the exact knowledge required to substantiate it. For what it’s worth, the situation is more dire than this; we can’t even describe the new directions required. My fictional-universe theory (FU theory) shows that a knowledge base cannot know whether its facts are describing the real world or a fictional world which has lots in common with the real world. (Humans don’t want to think about this, because of the implication.)





  • To be clear: Cohost did take funding from an anonymous angel, and as a result will not be sharing their source code; quoting from your link:

    Majority control of the cohost source code will be transferred to the person who funded the majority of our operations, as per the terms of the funding documents we signed with them; Colin and I will retain small stakes so we have some input on what happens to it, at their request.

    We are unable to make cohost open source. the source code for cohost was the collateral used for the loan from our funder.

    Somebody paid a very small amount of money to get a cleanroom implementation of Tumblr and did not mind that they would have to build a community and then burn it to the ground in the process. It turns out that angels are not better people than VCs.




  • Y’know what, I started out agreeing with the author, but now I realize that their critique is fundamentally not technical enough to hit the mark.

    Meta and Google own half of the fiber optic cables supplying internet services across continents.

    This is the only part of this article I’ll endorse. Y’know how this happened, right? Google bought dark fiber that was laid by the USA and privatized repeatedly. Meta set up Internet.org, a project that put phones with free Facebook into the hands of exploitable impoverished folks around the world, and then lobbied local governments to subsidize fiber rollouts to handle the induced demand. At some point, you gotta suck it up and start blaming capitalism, not only oligarchs or trusts, for this situation.

    The cloud is a lie.

    The clouds are products and services. They are a consumer’s understanding of the underlying infrastructure. It’s a lot more work than a mere fib!

    Over a decade or more, while our politicians were busy sub-tweeting fascists for clout, GAMM was buying up all the infrastructure it could carry. … The production cost of data storage plummeted by 94% in just ten years. You can’t sell 50GB plans to college kids who own M2 Macbook Pros with a terabyte of solid-state storage.

    Okay, now read between the lines. If an oligopoly (1) buys many warehouse-scale computers, (2) in an environment where prices are rapidly dropping on new hardware, (3) in a market which already provided basic local compute to all of your customers, then this is going to produce a massive second-hand market from all of the smaller shops which were using commodity hardware until they got displaced. Google, Apple, and Meta all purchase custom datacenter hardware at a scale which requires a consortium merely to ensure that the motherboards are printed fast enough, obsoleting workstations from Dell and HP.

    This has led to something of a boon for USA homelabs. I can purchase RAM-heavy workstations at less than $1/GiB, disks are at least half a TiB, small form factors are available as long as you’re willing to do some BIOS work, rackables are something like $100/U, etc. We’re talking discounts of 90-95%. In my house, a $200 workstation has more disk, RAM, cores, and system stability than a $600 gaming desktop, and the only thing missing is purple gamer LED strips.

    Amazon controls 35% of the cloud computing market and has created a tight seal around its customer base. … Amazon is mostly quiet as the frontrunner in the cloud computing market.

    The author hasn’t worked in the business. That’s fine, but it means they don’t know that AWS is not secure in its position. AWS is only tolerated because product managers ask for it, not because engineers like it; AWS is shit. For comparison, Google Cloud is fine but expensive and a third of the services are bad, Microsoft Azure is awful aside from their k8s, and Meta doesn’t operate a public cloud.

    Yes, if everyone open-sources its AI models, they cannot build a moat on proprietary software. However, Google’s memo fails to mention that it already has the infrastructure to run computing-hungry AI models and that infrastructure is wildly expensive to build.

    Click through to their side rant. This is where I realized that the author could be more clueful. If any of GAMM train another Llama-sized model, and it is at all good, somebody will put it up on Bittorrent and leak it to 4chan. This is literally how we got Llama. There is no moat.

    Don’t get me wrong, open-source tech is great and important, and wonderful. But it’s not like the average person runs a Large Language Model on their Mac to make grocery lists. If you are, in fact, doing this, you are a nerd and I love you. But you’re not the average user.

    He is a year behind in a field where things change every few months. See RWKV’s recent blog post. There is no moat.

    So, who gives a shit if Meta put Llama on Github for free? … Read the terms and conditions. Llama is not open-source.

    You naïve motherfucker, we the neighbors took it from Meta and we will take it again. There is no moat.

    Mark Zuckerberg is a capable businessman who understands the industry better than most tech founders. I don’t know the guy personally, but look at the facts.

    This is the most sneerable part of the article for me. You’re supposed to be a writer and humanist. It should be obvious after doing maybe five minutes of research that Zuck thinks of himself as a modern-day Octavian. Same haircut, same daily routine, same politics. Zuck is exactly the kind of person to hire a private navy to win a civil war for him by sailing off to defeat a pirate captain while he sits on a beach and idly thinks of how cool it will be to rule the Roman Empire.

    Because why give a shit who sells the milk jars when you own the motherfucking cows, baby!

    Have you seen the prices on the pre-owned cow market lately? Maybe milk is just permanently getting cheaper. The existence of Big Dairy and government cheese doesn’t preclude local dairies, either.

    My tip for this guy: Look up this new company “nVidia”, they make computer chips or something, I dunno. I wonder if they ever do anything anti-competitive~


  • I’m saying that we shouldn’t radiate if it would be expensive. It’s not easy to force the heat out to the radiators; normally radiation only works because the radiator is more conductive than the rest of the system, and so it tends to pull heat from other components.

    We can set up massive convection currents in datacenters on Earth, using air as a fluid. I live in Oregon, where we have a high desert region which enables the following pattern: pull in cold dry air, add water to cool it further and make it more conductive, let it fall into cold rows and rise out of hot rows, condition again to recover water and energy, and exhaust back out to the desert. Apple and Meta have these in Prineville and Google has a campus in The Dalles. If you do the same thing in space, then you end up with a section of looped pipe that has fairly hot convective fluid inside. What to do with it?

    I’m merely suggesting that we can reuse that concentrated heat, at reduced efficiency (not breaking thermodynamics), rather than spending extra effort pumping it outside. NASA mentions fluid loops in this catalogue of cooling options for cubesats and I can explain exactly what I mean with Figure 7.13. Note the blue-green transition from “heat” to “heat exchanger”; that’s a differential, and at the sorts of power requirements that a datacenter has, it may well be a significant amount of usable entropy.


  • You’re entirely right. Any sort of computation in space needs to be fluid-cooled or very sedate. Like, inside the ISS, think of the laptops as actively cooled by the central air system, with the local fan and heatsink merely connecting the laptop to air. Also, they’re shielded by the “skin” of the station, which you’d think is a given, but many spacebros think about unshielded electronics hanging out in the aether like it’s a nude beach or something.

    I’d imagine that a serious datacenter in space would need to concentrate heat into some sort of battery rather than trying to radiate it off into space. Keep it in one spot, compress it with heat pumps, and extract another round of work from the heat differential. Maybe do it all again until the differential is small enough to safely radiate.