• 2 Posts
  • 35 Comments
Joined 1 year ago
cake
Cake day: August 29th, 2023

help-circle





  • Some nitpicks. some of which are serious are some of which are sneers…

    consternating about the policy implications of Sam Altman’s speculative fan fiction

    Hey, the fanfiction is actually Eliezer’s (who in turn copied it from older scifi), Sam Altman just popularized it as a way of milking the doom for hype!

    So, for starters, in order to fit something as powerful as ChatGPT onto ordinary hardware you could buy in a store, you would need to see at least three more orders of magnitude in the density of RAM chips—​leaving completely aside for now the necessary vector compute.

    Well actually, you can get something close to as powerful on a personal computer… because the massive size of ChatGPT and the like don’t actually improve their performance that much (the most useful thing I think is the longer context window?).

    I actually liked one of the lawfare AI articles recently (even though it did lean into a light fantasy scenario)… https://www.lawfaremedia.org/article/tort-law-should-be-the-centerpiece-of-ai-governance . Their main idea is that corporations should be liable for near-misses. Like if it can be shown that the corporation nearly caused a much bigger disaster, they get fined in accordance with the bigger disaster. Of course, US courts routinely fail to properly penalize (either in terms of incentives of in terms of compensation) corporations for harms they actually cause, so this seems like a distant fantasy to me.

    AI has no initiative. It doesn’t want anything

    That’s next on the roadmap though, right? AI agents?

    Well… if the way corporations have tried to use ChatGPT has taught me anything, its that they’ll misapply AI in any and every way that looks like it might save or make a buck. So they’ll slap an API to a AI it into a script to turn it into an “agent” despite that being entirely outside the use case of spewing words. It won’t actually be agentic, but I bet it could cause a disaster all the same!



  • First of all. You could make facts a token value in an LLM if you had some pre-calculated truth value for your data set.

    An extra bit of labeling on your training data set really doesn’t help you that much. LLMs already make up plausible looking citations and website links (and other data types) that are actually complete garbage even though their training data has valid citations and website links (and other data types). Labeling things as “fact” and forcing the LLM to output stuff with that “fact” label will get you output that looks (in terms of statistical structure) like valid labeled “facts” but have absolutely no guarantee of being true.




  • I chose to have children, be a father and a husband, live an honest industrious life as an example to my offspring, and attempt to preserve my way of life through them.

    Wow, just a few words off the 14 words.

    I find it kind of irritating how someone that doesn’t familiarize themselves with white supremacists rhetoric and methods might manage to view that phrase innocuously. But it really isn’t that hard to see through the bullshit once you’ve familiarized themselves with the most basic dog whistles and slogans.


  • Wow… I took a look at that link before reading the comments/explanations here, and I was briefly confused why they were hating on him so much, before I realized he isn’t radical right wing enough for them.

    Eh, you’re a gay furry ex-Mormon (which is like a triple strike against you in my book) but I still like you well enough.

    It is almost sad seeing TWG trying to appeal to these people that fundamentally hate him… except he could just admit themotte is a cesspit and abandon it. But that would involve admitting that sneerclub (and David Gerard specifically) was right about the sort of people that lurked around SCC and later concentrated within themotte, so I think he’s going to keep making himself suffer.

    TW knows about the propaganda war, but has very different objectives to you. Much harder to balance ones too: he needs enough Progress for surrogate gaybies, but not too much that white gay guys can’t get the good lawyer jobs.

    Wow, I feel really gross agreeing with a motte poster, but they’ve called out TWG pretty effectively. TWG at least knows he needs things progressive enough he doesn’t end up against the wall for being gay, ex-Mormon and furry (as he describes himself), yet he wants to flirt with the alt-right!

    and in case I was in danger of forgetting what the motte really is…

    Yes, we’ve all thrown our hat in the ring in different ways. I chose to have children, be a father and a husband, live an honest industrious life as an example to my offspring, and attempt to preserve my way of life through them.

    sure buddy, you just need to “secure the future for your people and your children”… Yeah I know the rest of the words that go in that slogan.



  • I am probably giving most of them too much credit, but I think some of them took the Bitter Lesson and learned the wrong things from it. LLMs performed better than originally expected just off context, and (apparently) scaled better with bigger model and more training than expected, so now they think they just need to crank up the size and tweak things slightly (i.e. “prompt engineering” and RLHF) and don’t appreciate the limits built into the entire approach.

    The annoying thing about another winter is that it would probably result in funding being cut for other research. And laymen don’t appreciate all the academic funding that goes into research for decades before an approach becomes interesting and viable enough to scale up and commercialize (and then overhyped and oversold before some more modest practical usages become common, and relabeled as something other than AI).

    Edit: or more cynically, the leaders and hype-men know that algorithmic advances aren’t an automatic dump money in, get out disruptive product process, so they don’t bother putting as much monetary investment or hype into algorithmic advances. Like compare the attention paid towards Yann LeCunn talking about algorithmic developments vs. Sam Altman promising grad student level LLMs (as measured by a spurious benchmark) in two years.



  • Careful, if you present the problem and solution that way, AI tech bros will try pasting a LLM and a Computer Algebra System (which already exist) together, invent a fancy buzzword for it, act like they invented something fundamentally new, and then devise some benchmarks that break typical LLMs but their Frankenstein kludge can ace, and then sell the hype (actual consumer applications are luckily not required in this cycle but they might try some anyway).

    I think there is some promise to the idea of an architecture similar to a LLM with components able to handle math like a CAS. It won’t fix a lot of LLM issues but maybe some fundamental issues (like ability to count or ability to hold an internal state) will improve. And (as opposed to an actually innovative architecture) simply pasting LLM output into CAS input and then the CAS output back into LLM input (which, let’s be honest, is the first thing tech bros will try as it doesn’t require much basic research improvement), will not help that much and will likely generate an entirely new breed of hilarious errors and bullshit (I like the term bullshit instead of hallucination, it captures the connotation errors are of a kind with the normal output).


  • Well, if they were really “generalizing” just from training on crap tons of written text, they could implicitly develop a model of letters in each token based on examples of spelling and word plays and turning words into acronyms and acrostic poetry on the internet. The AI hype men would like you to think they are generalizing just off the size of their datasets and length of training and size of the models. But they aren’t really “generalizing” that much (and even examples of them apparently doing any generalizing are kind of arguable) so they can’t work around this weakness.

    The counting failure in general is even clearer and lacks the excuse of unfavorable tokenization. The AI hype would have you believe just an incremental improvement in multi-modality or scaffolding will overcome this, but I think they need to make more fundamental improvements to the entire architecture they are using.