• 3 Posts
  • 149 Comments
Joined 2 years ago
cake
Cake day: August 4th, 2023

help-circle







  • What would the “bot that finds bots larping as people” do exactly? Ban them? Block or mute them? File reports? DM an admin about them?

    If it’s just for pointing out suspected LLM-generated material, I think humans would be better at that than bots would be, and could block, mute, or file reports as necessary.

    Also, are you saying you intend to make a bot that posts LLM-generated drivel or a bot that detects LLM-generated drivel?








  • I’m definitely excited for this technology to start getting into slicers. In the meantime, I might have occasion to want so much strength in a part that I’d go to the trouble of using a script.

    I currently use Cura, but I’m disgusted with Cura and looking to switch to PrusaSlicer. Cura’s a great slicer, but a terrible program. I use Raspberry Pis as desktop systems frequently. Cura used to work on ARM, but doesn’t any more. I’m also switching my main x86_64 box to Gentoo. It seems like they’ve added just tons of ridiculous libraries as dependencies to Cura that make it so hard to build Cura, the Gentoo devs have given up trying. Cura also doesn’t play nice with Wayland. And it will only run on an old version of Python, which makes getting it to run on a modern system challenging. In short, the slicing isn’t the problem. It’s getting it installed and running on your system of choice.

    So, given that I’m probably switching to PrusaSlicer soon anyway, I’ll be in just the right place to start using scripts for PrusaSlicer/Orca/etc like this one. Hopefully this feature makes it into PrusaSlicer upstream soon as well.

    (I do say I’m probably switching to PrusaSlicer. I don’t really have a good grasp on what features I’ve depended on in Cura are absent in Prusaslicer. Like, does it have tree supports? Support blocking? Top surface ironing? Not that all of those things are deal breakers, but some might end up being a big deal to me. And if so, I might have to go to the trouble of wrangling building Cura or holding off on switching to Gentoo or running Cura in Docker or something. We’ll see.)

    Final thought:

    "Fuck Patents. All my homies hate patents."




  • TootSweet@lemmy.worldtoOpen Source@lemmy.mlIf we had libre AI
    link
    fedilink
    English
    arrow-up
    13
    ·
    edit-2
    3 months ago

    The GPL family of licenses was designed to cover code specifically. AI engines are code and are covered in most jurisdictions by copyright. (Disclaimer: I know a lot less about international intellectual property law than about U.S. intellectual property law. But I’m pretty confident what I’ll say here is at least true of the U.S…) But you don’t really have a functional generative AI system without weights. And it’s not clear that weights are covered by any particular branch of intellectual property in any particular jurisdiction. (And if they are, it’s not clear that the legal entity who trained the engine owns those rights on those weights rather than the rights holders who hold rights to the materials being used as training data.) It’s the weights that would make for any biases or purposefully nefarious output. Nothing that isn’t covered by intellectually property can meaningfully be said to be “licensed”, really. Under the AGPLv3 or any other license. To speak of something not covered by any intellectual (or non-intellectual, I suppose) property as “licensed” is just kindof nonsensical.

    Like, since Einstein’s General Relativity isn’t covered by any intellectual property, it’s not possible for General Relativity to be “licensed”. Similarly, unless some law is passed making LLM weights covered by, say, copyright law, one can’t speak of those weights being “licensed”.

    By the way, there are several high-profile cases of companies like Meta releasing LLMs that you can run locally and calling them “Open Source” when there’s nothing “Open Source” about them. As in, they don’t distribute the source code of LLaMa at all. That’s exactly the opposite of “Open Source” and the weights aren’t code and can’t really be said to be “Open Source”. More info here.

    Now, all that said, I don’t think there’s actually any inherent benefit to LLMs, AGPLv3 or otherwise, so I don’t have any interest even in AGPLv3 engines. But I’m all for more software being licensed AGPLv3. I just don’t think AGPLv3 is a concept that applies to any portion of LLMs aside from the engine.


  • AIs (well, LLMs, at least) aren’t coded, though. The engine is coded, but then they just throw training data at it until it starts parrotting the training data.

    Humans can create scripts around the LLMs. Scripts that filter certain stuff out of the training data (though that can involve some pretty tricky natural language processing and can never really account for everything) or scripts that watch responses for certain keywords or whatever and either preempt the response from getting to the user or try to get the LLM to generate a different, more acceptable answer.

    I think for poisoning to work well, we’d have to be creative, keep shifting our tactics, and otherwise do things in ways that can sneak past the LLMs’ babysitters. It would be a bit of an arms race, but I don’t think it’s as doomed from the start as you seem to think it is.


  • When I bought my current car, I read the privacy policy and it says that they’ll record anything in the cabin of the car they damned well please and upload it to the mothership(/car manufacturer/Subaru).

    For a while, I adopted the practice of repeating disparaging things about Subaru while I drove. I’ve kindof gotten away from the practice lately. What I really ought to do is find and unplug the OnStar MOBO to kill its internet connection. I’ll do that one of these days.

    As for what you’re talking about, I don’t think LLMs (typically?) learn by your interaction with them, right? Like, they take a lot of data, churn it through the algorithm, and produce a set of weights that are then used with the ending to produce hallucinations. And it’s very possible (probable, actually) that for the next generation of the LLM, they’ll use the prompts you used in the previous generation as more training data. So, yeah, what you’re getting at would work, but I don’t think it would work until the release of the next major version of the LLM.

    I dunno. I could be wrong about some of my assumptions in that last paragraph, though. Definitely open to correction.


  • TootSweet@lemmy.worldtolinuxmemes@lemmy.worldOrwelluan
    link
    fedilink
    English
    arrow-up
    0
    ·
    3 months ago

    Yeah, I’m planning to switch from Arch to Gentoo. Systemd isn’t the only reason, but it’s a big one.

    (Yes, I know about Artix, but it’s… kindof a Frankenstein’s monster, still mostly depending on the Arch repos and still with certain relics of Systemd. Or at least it was when I last tried it.)