Of course, I’m not in favor of this “AI slop” that we’re having in this century (although I admit that it has some good legitimate uses but greed always speaks louder) but I wonder if it will suffer some kind of piracy, if it is already suffering or people simple are not interested in “pirated AI”

  • Kaboom@reddthat.com
    link
    fedilink
    English
    arrow-up
    0
    ·
    10 days ago

    You can just run Automatic1111 locally if you want to generate images. I don’t know what the text equivalent is though, but I’m sure there’s one out there.

    There’s no real need for pirate ai when better free alternatives exist.

  • daniskarma@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    0
    ·
    edit-2
    9 days ago

    Not pirated. But my country, Spain, released an open AI model completely for free. Everything is open. The training data the models and everything. It’s supposedly ethically trained with open data(I have not personally dig in the training data but it’s there published).

    It’s focused on spanish and regional languages of spain. But I think it can also do things in English.

    Not piracy per se, as it’s completely legal. But there’s something you don’t depend on any bussiness to run.

  • Aceticon@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    0
    ·
    edit-2
    10 days ago

    I’m pretty sure those things are trained on content which was obtained without paying royalties to the creators, hence by definition pirated content - so that would count as “piracy around them”.

    On the opposite side, as far as I know the things created with Generative AI so far can’t be copyrighted, hence by definition can’t be pirated as they’ve always belonged to the Public Domain.

    As for the engines themselves, there are good fully open source options out there which can be locally installed (if you have enough memory in your graphics card) and there seem to be thriving communities around it (at least it looks like it from what bit I dipped into that stuff so far). I’m not sure if it’s at all possible to pirate the closed source engines since I expect those things are designed to be deployed to very specific server farm architectures.

    • OminousOrange@lemmy.ca
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      9 days ago

      There are quite a few options for running your own LLM. Ollama makes it fairly easy to run (with a big selection of models - there’s also Hugging Face with even more models to suit various use cases) and OpenWebUI makes it easy to operate.

      Some self-hosting experience doesn’t hurt, but it’s pretty straightforward to configure if you follow along with Networkchuck in this video.

    • mindbleach@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      0
      ·
      9 days ago

      Training is transformative use. Sluicing data through a pile of linear algebra, to mechanically distill the essence of words like “fantasy,” is not what copyright protects against.

      • Lime Buzz (fae/she)@beehaw.org
        link
        fedilink
        English
        arrow-up
        0
        ·
        10 days ago

        Nope, there’s a difference there in that they aren’t taking something from ordinary people who need the money in order to survive. Actors, producers, directors etc have already been paid and besides, hollywood etc aren’t exactly using that money to give back to society in any meaningful way most of the time.

        • notfromhere@lemmy.ml
          link
          fedilink
          English
          arrow-up
          0
          ·
          10 days ago

          They did not take money from anyone. Are ‘t we on the priacy community? What is with the double standards? It’s theft if it’s against the Little Guy™ but it’s civil copyright violation if it’s against the Corpos?

    • catloaf@lemm.ee
      link
      fedilink
      English
      arrow-up
      0
      ·
      10 days ago

      How is what they’re doing different from, say, an IPTV provider?

  • PerogiBoi@lemmy.ca
    link
    fedilink
    English
    arrow-up
    0
    ·
    10 days ago

    There already is. You can download copies of AI that are similar or better than ChatGPT from hugging face. I run different models locally to create my own useless AI slop without paying for anything.

      • PerogiBoi@lemmy.ca
        link
        fedilink
        English
        arrow-up
        0
        ·
        10 days ago

        No because that is just an API that can run LLMs locally. GPT4All is an all in one solution that can run the .gguf file. Same with kobold ai.

    • Domi@lemmy.secnd.me
      link
      fedilink
      English
      arrow-up
      0
      ·
      9 days ago

      Which model would you say is better than GPT-4? All I tried are cool but are not quite on GPT-4 level.

      • antipiratgruppen@lemmy.dbzer0.com
        link
        fedilink
        English
        arrow-up
        0
        ·
        7 days ago

        The very newly released Deepseek R1 “reasoning model” from China beats OpenAI’s o1 model on multiple areas, it seems – and you can even see all the steps of the pre-answering “thinking” that’s hidden from the user in o1. It’s a huge model, but it (and the paper about it) will probably positively impact future “open source” models in general, now the “thinking” cat’s outta the bag. Though, it can’t think about Tiananmen Square or Taiwan’s autonomy – but many derivative models will probably be modified to effectively remove such Chinese censorship.

        • Domi@lemmy.secnd.me
          link
          fedilink
          English
          arrow-up
          0
          ·
          edit-2
          6 days ago

          Just gave Deepseek R1 (32b) a try. Except for the censorship probably the closest to GPT-4 so far. The chain-of-thought output is pretty interesting, sometimes even more useful than the actual response.

  • JohnBrownsBussy2 [she/her, they/them]@hexbear.net
    link
    fedilink
    English
    arrow-up
    0
    ·
    10 days ago

    Weight leaks for semi-open models have been fairly common in the past. Meta’s LLaMa1.0 model was originally closed source, but the weights were leaked and spread pretty rapidly (effectively laundered through finetunes and merges), leading to Meta embracing quasi-open source post-hoc. Similarly, most of the anime-style Stable Diffusion 1.5 models were based on NovelAI’s custom finetune, and the weights were similarly laundered and became ubiquitous.

    Those incidents were both in 2023. Aside from some of the biggest players (OpenAI, Google, Anthropic, and I guess Apple kinda), open weight releases (usually not open source) have been become the norm (even for frontier models like DeepSeek-V3, Qwen 2.5 and Llama 3.1), so piracy in that case is moot (although it’s easy to assume that use non-compliant with licenses is also ubiquitous). Leakage of currently closed frontier models would be interesting from an academic and journalistic perspective, for being able to dig into the architecture and assess things like safety and regurgitation outside of the online service shell, but those frontier models would require so much compute that they’d be unusable by individual actors.