• NewNewAugustEast@lemmy.zip
    link
    fedilink
    English
    arrow-up
    23
    ·
    edit-2
    4 hours ago

    Copilot? You mean the AI with terms of service that are in bold and explicit: “for entertainment purposes only”?

    Which is why its in the title and not the article? EntertainBait?

  • CanIFishHere@lemmy.ca
    link
    fedilink
    English
    arrow-up
    32
    arrow-down
    4
    ·
    5 hours ago

    AI is here, another tool to use…the correct way. Very reasonable approach from Torvalds.

    • Newsteinleo@infosec.pub
      link
      fedilink
      English
      arrow-up
      6
      ·
      4 hours ago

      I don’t have a problem with LLMs as much as the way people use them. My boss has offloaded all of his thinking to LLMs to the point he can’t fix a sentence in a slide deck without using an LLM.

      It’s the people that try to use LLMs for things outside their domain of expertise that really cause the problems.

      • NotMyOldRedditName@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        edit-2
        24 minutes ago

        It’s the people that try to use LLMs for things outside their domain of expertise that really cause the problems.

        That seems to general. Im a mobile developer and sometimes I need a simple script outside my knowledge area. I needed to scrape a website recently, not for anything serious, but to save me time. Claude wrote it and it works. Its probably trash code, but it works and it helped. But you wouldn’t want me using Claude to do important work outside my specific area of focus either or im sure Id cause problems.

      • InternetCitizen2@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        ·
        58 minutes ago

        This is a big point. People need to understand that the LLMs are more like a fancy graphing calculator; they are very good and handle multiple things, but its on you to understand why the calculation is meaningful. At a certain point no one wants to see your long division or factorial. We want the results and for students and professionals to focus on the concept.

      • CanIFishHere@lemmy.ca
        link
        fedilink
        English
        arrow-up
        1
        ·
        2 hours ago

        Very frustrating for sure. Like any tool, it’s up to humans to know when the tool is useful.

        • filcuk@lemmy.zip
          link
          fedilink
          English
          arrow-up
          1
          ·
          25 minutes ago

          Partly a marketing issue.
          Companies keep advertising their new AI’s as destroyers of worlds, and something that’s too dangerous to even release.
          As with anything else, the average user will not have but the most surface level understanding of the tool

    • null@lemmy.zip
      link
      fedilink
      English
      arrow-up
      4
      ·
      4 hours ago

      Clickbait got me. No mention of “Yes copilot” which I assumed was a joke anyway.

  • gandalf_der_12te@discuss.tchncs.de
    link
    fedilink
    English
    arrow-up
    23
    arrow-down
    2
    ·
    edit-2
    7 hours ago

    I agree. If AI becomes outlawed, it will simply be used without other people knowing about it.

    This approach, at least, means that people will label AI-generated code as such.

    • emmy67@lemmy.world
      link
      fedilink
      English
      arrow-up
      15
      ·
      6 hours ago

      Maybe. There’s still strong disapproval around it. I can imagine many will still hide it.

  • null@lemmy.org
    link
    fedilink
    English
    arrow-up
    32
    arrow-down
    2
    ·
    9 hours ago

    Ah, the solution that recognizes there’s no way to eliminate AI from the supply chain after it’s already been introduced.

    • sunbeam60@feddit.uk
      link
      fedilink
      English
      arrow-up
      7
      arrow-down
      9
      ·
      edit-2
      8 hours ago

      You make it sound as if there was another choice if just people had better principles. Prey tell us, what would you have done, now. Not in the past, now.

      • null@lemmy.org
        link
        fedilink
        English
        arrow-up
        11
        ·
        6 hours ago

        That wasn’t my intent. This is me saying, “of course that’s what they’re going to do because there’s nothing else they can do.”

      • Feyd@programming.dev
        link
        fedilink
        English
        arrow-up
        8
        ·
        6 hours ago

        You’re agreeing with the comment you replied to. Why the fuck are you trying to be so smug???

  • Blue_Morpho@lemmy.world
    link
    fedilink
    English
    arrow-up
    178
    arrow-down
    1
    ·
    13 hours ago

    The title of the article is extraordinary wrong that makes it click bait.

    There is no “yes to copilot”

    It is only a formalization of what Linux said before: All AI is fine but a human is ultimately responsible.

    " AI agents cannot use the legally binding “Signed-off-by” tag, requiring instead a new “Assisted-by” tag for transparency"

    The only mention of copilot was this:

    “developers using Copilot or ChatGPT can’t genuinely guarantee the provenance of what they are submitting”

    This remains a problem that the new guidelines don’t resolve. Because even using AI as a tool and having a human review it still means the code the LLM output could have come from non GPL sources.

    • TheOctonaut@piefed.zip
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      2
      ·
      2 hours ago

      the LLM output could have come from non-GPL sources

      Fundamentally not how LLMs work, it’s not a database of code snippets.

    • marlowe221@lemmy.world
      link
      fedilink
      English
      arrow-up
      51
      ·
      edit-2
      12 hours ago

      Yeah, that’s also my question. Partially because I am a former-lawyer-turned-software-developer… but, yeah. How are the kernel maintainers supposed to evaluate whether a particular PR contains non-GPL code?

      Granted, this was potentially an issue before LLMs too, but nowhere near the scale it will be now.

      (In the interests of full disclosure, my legal career had nothing to do with IP law or software licensing - I did public interest law).

      • wonderingwanderer@sopuli.xyz
        link
        fedilink
        English
        arrow-up
        6
        ·
        6 hours ago

        If it’s flagged as “assisted by <LLM>” then it’s easy to identify where that code came from. If a commercial LLM is trained on proprietary code, that’s on the AI company, not on the developer who used the LLM to write code. Unless they can somehow prove that the developer had access to said proprietary code and was able to personally exploit it.

        If AI companies are claiming “fair use,” and it holds up in court, then there’s no way in hell open-source developers should be held accountable when closed-source snippets magically appear in AI-assisted code.

        Granted, I am not a lawyer, and this is not legal advice. I think it’s better to avoid using AI-written code in general. At most use it to generate boilerplate, and maybe add a layer to security audits (not as a replacement for what’s already being done).

        But if an LLM regurgitates closed-source code from its training data, I just can’t see any way how that would be the developer’s fault…

      • Alex@lemmy.ml
        link
        fedilink
        English
        arrow-up
        21
        ·
        12 hours ago

        They don’t, just like they don’t with human submitted stuff. The point of the Signed-off-by is the author attests they have the rights to submit the code.

    • anarchiddy@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      9
      arrow-down
      1
      ·
      12 hours ago

      Yup.

      I would also just point out that this doesnt change the legal exposure to the Linux kernel to infringing submissions from before the advent of LLMs.

  • catlover@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    41
    ·
    11 hours ago

    I’d still be highly sceptical about pull requests with code created by llms. Personally what I noticed is that the author of such pr doesn’t even read the code, and i have to go through all the slop

    • kcuf@lemmy.world
      link
      fedilink
      English
      arrow-up
      14
      ·
      10 hours ago

      Ya I’m finding myself being the bad code generator at work as I’m scattered across so many things at the moment due to attrition and AI can do a lot of the boilerplate work, but it’s such a time and energy sink to fully review what it generates and I’ve found basic things I missed that others catch and shows the sloppiness. I usually take pride in my code, but I have no attachment to what’s generated and that’s exposing issues with trying to scale out using this

      • Repple (she/her)@lemmy.world
        link
        fedilink
        English
        arrow-up
        12
        ·
        edit-2
        9 hours ago

        Same. There’s reduction in workforce, pressure to move faster, and no good way to do that without sloppiness. I have never been this down on the industry before; it was never great, but now it’s terrible.

        • Danitos@reddthat.com
          link
          fedilink
          English
          arrow-up
          5
          ·
          8 hours ago

          Some thought I had the other day :LLM is supposed to make us more productive, say by 20%. Have you won a 20% pay rise since you adopted it? I haven’t

      • Feyd@programming.dev
        link
        fedilink
        English
        arrow-up
        2
        arrow-down
        5
        ·
        6 hours ago

        Just fucking stop using it? Wtf? Tell you boss to pound sand! They’re going to blame you when it goes south anyway so you might as well stay honest.

    • jj4211@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      1
      ·
      8 hours ago

      I suspect the answer will be that such large requested as you frequently see with LLM codegen will just be rejected.

      Already I see changes broken up and suggested bit by bit, so I presume the same best practice applies.

    • terabyterex@lemmy.world
      link
      fedilink
      English
      arrow-up
      4
      arrow-down
      9
      ·
      edit-2
      9 hours ago

      Did we all forget about stackoverflow?

      Peopleblindly copy/pasted from there all the time.

      • Railcar8095@lemmy.world
        link
        fedilink
        English
        arrow-up
        6
        ·
        9 hours ago

        Couple of years back I got a PR at work that used a block of code that read a CSV, used some stream method to covert it to binary to then feed it to pandas to make a dataframe. I don’t remember the exact steps it did, but was just crazy when pd.read_csv existed.

        On a hunch I pasted the code in google and found an exact match on overflow for a very weird use case on very early pandas.

        I’m lucky and if people send obvious shit at work I can just cc their manager, but I fell for the volunteers at large FOSS projects, or even paid employees.

  • theherk@lemmy.world
    link
    fedilink
    English
    arrow-up
    124
    arrow-down
    4
    ·
    13 hours ago

    Seems like a reasonable approach. Make people be accountable for the code they submit, no matter the tools used.

    • ell1e@leminal.space
      link
      fedilink
      English
      arrow-up
      23
      arrow-down
      1
      ·
      12 hours ago

      If the accountability cannot be practically fulfilled, the reasonable policy becomes a ban.

      What good is it to say “oh yeah you can submit LLM code, if you agree to be sued for it later instead of us”? I’m not a lawyer and this isn’t legal advice, but sometimes I feel like that’s what the Linux Foundation policy says.

      • ViatorOmnium@piefed.social
        link
        fedilink
        English
        arrow-up
        39
        arrow-down
        1
        ·
        12 hours ago

        But this was already the case. When someone submitted code to Linux they always had to assume responsibility for the legality of the submitted code, that’s one of the points of mandatory Signed-off-by.

        • badgermurphy@lemmy.world
          link
          fedilink
          English
          arrow-up
          4
          arrow-down
          15
          ·
          11 hours ago

          But now, even the person submitting the license-breaching content may be unaware that they are doing that, so the problem is surely worse now that contributors can easily unwittingly be on the wrong side of the law.

          • Traister101@lemmy.today
            link
            fedilink
            English
            arrow-up
            35
            arrow-down
            1
            ·
            11 hours ago

            That’s their problem. If they are using an LLM and cannot verify the output they shouldn’t be using an LLM

            • hperrin@lemmy.ca
              link
              fedilink
              English
              arrow-up
              1
              ·
              2 hours ago

              Nobody can verify that the output of an LLM isn’t from its training data except those with access to its training data.

            • jj4211@lemmy.world
              link
              fedilink
              English
              arrow-up
              3
              ·
              8 hours ago

              Problem is that broadly most GenAI users don’t take that risk seriously. So far no one can point to a court case where a rights holder successfully sued someone over LLM infringement.

              The biggest chance is getty and their case, with very blatantly obvious infringement. They lost in the UK, so that’s not a good sign.

            • badgermurphy@lemmy.world
              link
              fedilink
              English
              arrow-up
              3
              arrow-down
              12
              ·
              11 hours ago

              It is their problem until the second they submit it, then it is the project’s problem. You can lay the blame for the bad actions wherever you want, but the reality is that the work of verifying the legality and validity of these submissions if being abdicated, crippling projects under increased workloads going through ever more submissions that amount to junk.

              What is the solution for that? The fact that is the fault of the lazy submitter doesn’t clean up the mess they left.

              • Traister101@lemmy.today
                link
                fedilink
                English
                arrow-up
                10
                arrow-down
                1
                ·
                10 hours ago

                Frankly I expect the kernel dudes to be pretty good about this, their style guides alone are quite strick and any funny business in a PR that isn’t marked correctly is I think likely a ban from making PRs at all. How it worked beforehand, as already stated by others is the author says “I promise this follows the rules” and that’s basically the end of it. Giving an official avenue for generated code is a great way to reduce the negatives of it that’ll happen anyway. We know this from decades of real life experience trying to ban things like alcohol or drugs, time after time providing a legal avenue with some rules makes things safer. Why wouldn’t we see a similar effect here?

                • badgermurphy@lemmy.world
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  ·
                  3 hours ago

                  I do think that some projects will fare better than others, particularly ones like you mentioned, where the team is robust and capable of handling the filtering of increased submissions from these new sources.

                  I believe we are going to end up having to see some new mechanism for project submissions to deal with the growing imbalance between submission volume and work hours available for review, as became necessary when viruses, malware, and spam first came into being. It has quickly become incredibly easy for anyone to make a PR, but not at all easier to review them, so something is going to have to give in the FOSS world.

    • hperrin@lemmy.ca
      link
      fedilink
      English
      arrow-up
      8
      arrow-down
      4
      ·
      11 hours ago

      No, it’s not a reasonable approach. Make people be the authors of the code they submit is reasonable, because then it can be released under the GPL. AI generated code is public domain.

      • theherk@lemmy.world
        link
        fedilink
        English
        arrow-up
        8
        arrow-down
        5
        ·
        11 hours ago

        I suppose there should be no code generators, assemblers, compilers, linkers, or lsp’s then either? Just etching 1’s and 0’s?

        • hperrin@lemmy.ca
          link
          fedilink
          English
          arrow-up
          1
          ·
          5 hours ago

          The copyright office has made it explicitly clear that those tools do not interfere with the traditional elements of authorship, and that the use of LLMs does. So, if you don’t want to take my word for it, take the US Copyright Office’s word for it.

          • theherk@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            ·
            edit-2
            2 hours ago

            As the agency overseeing the copyright registration system, the Office has extensive experience in evaluating works submitted for registration that contain human authorship combined with uncopyrightable material, including material generated by or with the assistance of technology. It begins by asking “whether the ‘work’ is basically one of human authorship, with the computer [or other device] merely being an assisting instrument, or whether the traditional elements of authorship in the work (literary, artistic, or musical expression or elements of selection, arrangement, etc.) were actually conceived and executed not by man but by a machine.” In the case of works containing AI-generated material, the Office will consider whether the AI contributions are the result of “mechanical reproduction” or instead of an author’s “own original mental conception, to which [the author] gave visible form.” The answer will depend on the circumstances, particularly how the AI tool operates and how it was used to create the final work. This is necessarily a case-by-case inquiry. If a work’s traditional elements of authorship were produced by a machine, the work lacks human authorship and the Office will not register it For example, when an AI technology receives solely a prompt from a human and produces complex written, visual, or musical works in response, the “traditional elements of authorship” are determined and executed by the technology—not the human user. Based on the Office’s understanding of the generative AI technologies currently available, users do not exercise ultimate creative control over how such systems interpret prompts and generate material. Instead, these prompts function more like instructions to a commissioned artist—they identify what the prompter wishes to have depicted, but the machine determines how those instructions are implemented in its output. For example, if a user instructs a text-generating technology to “write a poem about copyright law in the style of William Shakespeare,” she can expect the system to generate text that is recognizable as a poem, mentions copyright, and resembles Shakespeare’s style. But the technology will decide the rhyming pattern, the words in each line, and the structure of the text. When an AI technology determines the expressive elements of its output, the generated material is not the product of human authorship. As a result, that material is not protected by copyright and must be disclaimed in a registration application.

            In other cases, however, a work containing AI-generated material will also contain sufficient human authorship to support a copyright claim. For example, a human may select or arrange AI-generated material in a sufficiently creative way that “the resulting work as a whole constitutes an original work of authorship.” Or an artist may modify material originally generated by AI technology to such a degree that the modifications meet the standard for copyright protection. In these cases, copyright will only protect the human-authored aspects of the work, which are “independent of ” and do “not affect” the copyright status of the AI-generated material itself.

            This policy does not mean that technological tools cannot be part of the creative process. Authors have long used such tools to create their works or to recast, transform, or adapt their expressive authorship. For example, a visual artist who uses Adobe Photoshop to edit an image remains the author of the modified image, and a musical artist may use effects such as guitar pedals when creating a sound recording. In each case, what matters is the extent to which the human had creative control over the work’s expression and “actually formed” the traditional elements of authorship.

            https://www.copyright.gov/ai/ai_policy_guidance.pdf

            What this makes clear is that it certainly isn’t black or white as you say. Nevertheless, automation converting an input to an output, simply cannot be the only mechanism used in determining authorship.

            And that wouldn’t change my statement anyway, but rather supports it. The person submitting a patch must be accountable for its contents.

            An outright ban would need to carefully define how an input gets converted to an output, and that may not be so clear. To be effectively clear, one would have to potentially end the use of many tools that have been used for many years in the kernel, including snippet generation, spelling and grammar correction, IDE autocompleting. So such a reductive view simply will not suffice.


            Additionally, copywritability and licenseability are wholly different questions. And it does not violate GPL to include public domain content, since the license applies to the aggregate work.

            • hperrin@lemmy.ca
              link
              fedilink
              English
              arrow-up
              1
              ·
              edit-2
              2 hours ago

              If a work’s traditional elements of authorship were produced by a machine, the work lacks human authorship and the Office will not register it For example, when an AI technology receives solely a prompt from a human and produces complex written, visual, or musical works in response, the “traditional elements of authorship” are determined and executed by the technology—not the human user. Based on the Office’s understanding of the generative AI technologies currently available, users do not exercise ultimate creative control over how such systems interpret prompts and generate material. Instead, these prompts function more like instructions to a commissioned artist—they identify what the prompter wishes to have depicted, but the machine determines how those instructions are implemented in its output. For example, if a user instructs a text-generating technology to “write a poem about copyright law in the style of William Shakespeare,” she can expect the system to generate text that is recognizable as a poem, mentions copyright, and resembles Shakespeare’s style. But the technology will decide the rhyming pattern, the words in each line, and the structure of the text. When an AI technology determines the expressive elements of its output, the generated material is not the product of human authorship. As a result, that material is not protected by copyright and must be disclaimed in a registration application.

              That seems very clear to me. Generative AI output is not human authored, and therefore not copyrighted.

              The policy I use also makes very clear the definition of AI generated material:

              https://sciactive.com/human-contribution-policy/#Definitions

              I’m not exactly sure how you can possibly think there is an equivalence between a tool like a spelling and grammar checker and a generative AI, but there’s a reason the copyright office will register works that have been authored using spelling and grammar checkers, but not works that have been authored using LLMs.

              • theherk@lemmy.world
                link
                fedilink
                English
                arrow-up
                1
                ·
                2 hours ago

                Just read the next two paragraphs. Don’t just stop because you got to something that you like. The equivalence I draw is clear. You don’t like it, and that’s okay. But one would have to clarify exactly what the ban entails, and that wouldn’t be as clear as you might think. LLM’s only, transformers specifically, what about graph generation, other ML models? Is it just ML? If so, is that because a matrix lattice was used to get from input to output? Could other deterministic math functions trigger the same ban? What is a spell checker used RNG to select best replacement from a list of correct options? What if a compiler introduces an assembled output with an optimization not of the authors writing?

                Do you see why they say “The answer will depend on the circumstances, particularly how the AI tool operates and how it was used to create the final work. This is necessarily a case-by-case inquiry”?

                And that still affects copywriteability, not license compliance.

                • hperrin@lemmy.ca
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  ·
                  edit-2
                  1 hour ago

                  Do you want to explain to me what, in those two paragraphs, means that the use of spell checkers and LLMs is equivalent with regard to copyrightability? It seems like those paragraphs make it clear that the use of spell checkers is not the same as LLMs.

                  The policy I use bans “generative AI model” output. Generative AI is a pretty well defined term:

                  https://en.wikipedia.org/wiki/Generative_AI

                  https://www.merriam-webster.com/dictionary/generative AI

                  If you have trouble determining whether something is a generative AI model, you can usually just look up how it is described in the promotional materials or on Wikipedia.

                  Type: Large language model, Generative pre-trained transformer

                  - https://en.wikipedia.org/wiki/Claude_(language_model)

                  I never said it violates GPL to include public domain code. I’m not sure where you got that from. What I said is that public domain code can’t really be released under the GPL. You can try, but it’s not enforceable. As in, you can release it under that license, but I can still do whatever I want with it, license be damned, because it’s public domain.

                  I did that with this vibe coded project:

                  https://github.com/hperrin/gnata

                  I just took it and rereleased it as pubic domain, because that’s what it is anyway.

      • ziproot@lemmy.ml
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        2
        ·
        9 hours ago

        Isn’t that the rule? The author has to be a human?

        The new guidelines mandate that AI agents cannot use the legally binding “Signed-off-by” tag, requiring instead a new “Assisted-by” tag for transparency. Ultimately, the policy legally anchors every single line of AI-generated code and any resulting bugs or security flaws firmly onto the shoulders of the human submitting it.

  • 404found@lemmy.zip
    link
    fedilink
    English
    arrow-up
    11
    arrow-down
    2
    ·
    10 hours ago

    I don’t understand the full picture here, but the person who is submitting AI slop will be held accountable. Never a company.

    So if a company is pushing staff to us AI to complete projects faster and their code ends up being AI slop when submitted, only the person working for the company will be held responsible.

    I’m not sure what the repercussions are here but hopefully it’s not a large fine. Those fines could add up quick if the person is submitting code all the time and doesn’t know they are messing up.

    • Wispy2891@lemmy.world
      link
      fedilink
      English
      arrow-up
      25
      ·
      10 hours ago

      Which fines, this is just an internal rule in an organization.

      At most can be rightfully banned from contributing

      It someone is contributing with code that doesn’t really understand, then shouldn’t contribute

      • 404found@lemmy.zip
        link
        fedilink
        English
        arrow-up
        4
        ·
        9 hours ago

        Ah okay got it now. Thanks. I didn’t understand it all the way. My comment is irrelevant

    • truthfultemporarily@feddit.org
      link
      fedilink
      English
      arrow-up
      20
      arrow-down
      9
      ·
      13 hours ago

      Where does slop start? If you use auto complete and it is just adding a semicolon or some braces, is it slop? Is producing character by character what you would have wrote yourself slop?

      How about using it for debugging?

      • hperrin@lemmy.ca
        link
        fedilink
        English
        arrow-up
        8
        ·
        11 hours ago

        You don’t need AI to autocomplete code. We’ve had autocomplete for over 30 years.

      • ell1e@leminal.space
        link
        fedilink
        English
        arrow-up
        12
        arrow-down
        3
        ·
        edit-2
        12 hours ago

        If you would have written it yourself the same way, why not write it yourself? (And there was autocomplete before the age of LLMs, anyway.)

        The big problems start with situations where it doesn’t match what you would have written, but rather what somebody else has written, character by character.

      • BoxOfFeet@lemmy.world
        link
        fedilink
        English
        arrow-up
        8
        arrow-down
        4
        ·
        13 hours ago

        To me, it starts at anything beyond correcting spelling for individual words or adding punctuation. I don’t even want it suggesting quick reply phrases.

        Is producing character by character what you would have wrote yourself slop?

        Yes.

      • badgermurphy@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        11 hours ago

        There’s the rub. When establishing laws and guidelines, every term must be explicitly defined. Lack of specificity in these definitions is where bad-faith actors hide their misdeeds by technically obeying the letter of the law due to its vagueness, while flagrantly violating its spirit.

        Its why today, in the USA, corporations are legally people when its convenient, and not when its not, and the expenditure of money is governments protected “free speech”.

      • FauxLiving@lemmy.world
        link
        fedilink
        English
        arrow-up
        5
        arrow-down
        14
        ·
        edit-2
        12 hours ago

        There is a certain brand of user (who may or may not be a human) who draws the venn of ‘AI slop’ and ‘AI output’ as a circle.

        They’ve taken the extremist position that AI should be uninvented and any use of AI is the worst thing that could possibly happen to any project and they’ll have an entire grab bag of misinformation-based memes to shotgun at you. Engaging with these people is about as productive as trying to convince a vaccine denier that vaccines don’t cause autism.

        I’m not saying that the user you replies to believes this, but the comment they wrote is indistinguishable from the comments of such a user.

        e: I’d also like to point out that these users are very much attracted to low-effort activism. This is why you see comments like mind being heavily downvoted but not many actual replies. They want to try to influence the discussion but don’t have the capability or motivation to step into the ring, so to speak, and defend their opinions.

        • ell1e@leminal.space
          link
          fedilink
          English
          arrow-up
          10
          arrow-down
          1
          ·
          edit-2
          13 hours ago

          It’s less extremist if you look at how easily these LLMs will just plagiarize 1:1, apparently:

          https://github.com/mastodon/mastodon/issues/38072#issuecomment-4105681567

          Some see “AI slop” as “identified by the immediate problems of it that I can identify right away”.

          Many others see “AI slop” as bringing many more problems beyond the immediate ones. Then seeing LLM output as anything but slop becomes difficult.

          • FauxLiving@lemmy.world
            link
            fedilink
            English
            arrow-up
            5
            arrow-down
            4
            ·
            12 hours ago

            It’s extremist to take the fact that you CAN get plagiaristic output and to conclude that all other output is somehow tainted.

            You personally CAN quote copyrighted music and screenplays. If you’re an artist then you also CAN produce copyright violating works. None of these facts taint any of the other things that you produce that are not copyright or plagiarized.

            In this situation, and in the current legal environment, the responsibility to not produce illegal and unlicensed code is on the human. The fact that the tool that they use has the capability to break the law does not mean that everything generated by it is tainted.

            Photoshop can be used to plagiarize and violate copyright too. It would be just as absurd to declare all images created with Photoshop are somehow suspect or unusable because of the capability of the tool to violate copyright laws.

            The fact that AI can, when specifically prompted, produce memorized segments of the training data has essentially no legal weight in any of the cases where it has been argued. It is a fact that is of interest to scientists who study how AI represent knowledge internally and not any kind of foundation for a legal argument against the use of AI.

            • badgermurphy@lemmy.world
              link
              fedilink
              English
              arrow-up
              2
              arrow-down
              1
              ·
              11 hours ago

              Sure, but if they can be demonstrated to ever plagiarize without attribution, and the default user behavior is to pencil-whip the output, which it is, then it becomes statistically certain that users are unwittingly plagiarizing other works.

              Its like using a tool that usually bakes cookies, but every once in a great while, it knocks over the building its in. It almost never does that, though.

              • FauxLiving@lemmy.world
                link
                fedilink
                English
                arrow-up
                2
                arrow-down
                1
                ·
                10 hours ago

                Plagiarism and copyright violation are two different things, one is ethical and the other is legal.

                Copyright has a body of case law which helps determine when a work significantly infringes on the copyrighted work of another. Plagiarism has no body of law at all, it is an ethical construct and not a legal one.

                You can plagiarize something that has no copyright protection and you can infringe on copyright protection without plagiarizing. They’re not interchangeable concepts.

                In your example, some institutions would not allow such a device to operate on their property but it would not be illegal to operate and the liability would be on the person and not on the oven.

                To further strain the metaphor, Linus is saying that you can use (possibly) exploding ovens, because he isn’t taking a moral stance on the topic, but you are responsible for the damages if they cause any because the legal systems require that this be the case.

        • hperrin@lemmy.ca
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          2
          ·
          edit-2
          2 hours ago

          According to the US Copyright Office, AI generated material cannot be copyrighted (unless of course it’s plagiarized copyrighted code). That’s reason enough to leave it out of the kernel. If the kernel’s license becomes unenforceable because of public domain code, the kernel is tainted.

          Edit: I don’t know why people are downvoting this. It’s literally just the truth: https://sciactive.com/human-contribution-policy/#More-Information

          • FauxLiving@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            arrow-down
            1
            ·
            10 hours ago

            Copyright and License terms are two different categories of law. Copyright is an idea created and enforced by the laws of the country which has jurisdiction. Licenses are a contract between two parties and is covered by contract law.

            A thing can be unable to be protected by copyright and also protected by the terms of the license that it is provided under. If a project contains copyrighted code that does not mean that you cannot be held to the terms of the license. Your use of licensed works is granted under the agreement that you follow the terms of the license. You cannot be held liable for copyright violations for using the code, but using the code in a manner that is not allowed by the license makes you liable for violation of the contract that is the license agreement.

            • hperrin@lemmy.ca
              link
              fedilink
              English
              arrow-up
              1
              ·
              5 hours ago

              I think you’re misunderstanding what I’m saying. Any portions of the kernel that are public domain can be used by anyone for any purpose without following the terms of the GPL. AI generated code is public domain. To make sure all parts of the kernel are protected by the GPL, public domain code should not be accepted unless absolutely necessary.

    • femtek@lemmy.blahaj.zone
      link
      fedilink
      English
      arrow-up
      10
      arrow-down
      2
      ·
      13 hours ago

      I mean I don’t use copilot but a self hosted Claude at work for debugging and creating templates. I still run thru and test it. I’m only doing crossplane, kyverno, kubernetes infra things though and I started without it so I have an understanding. Now running their someone’s crossplane composition written in go and I asked them about this error and he just said get the AI to fix it was worrying since his last day is next week.

    • chilicheeselies@lemmy.world
      link
      fedilink
      English
      arrow-up
      4
      arrow-down
      3
      ·
      11 hours ago

      Its only slop if you accept slop. What i mean is that it cna and does generate perfectly fine code. It also generates code that is ok, but needs a human touch. It also generates verbose garbage.

      Its only slop if you approve the slop. Its perfectly fine to let it generate the boilerplate of what you want, and tweak it. If its prompted well enough, you get less slop.

      Ultimately I am with Linus on this one. The genie is out of the bottle. Use it responsibly.

  • menas@lemmy.wtf
    link
    fedilink
    English
    arrow-up
    4
    ·
    9 hours ago

    Ecological, social, economic issues and the answer is on the legal site. FOSS as usual I guess

  • hperrin@lemmy.ca
    link
    fedilink
    English
    arrow-up
    13
    arrow-down
    4
    ·
    11 hours ago

    This is a bad move. The GPL license cannot be enforced on AI generated code.

    • terabyterex@lemmy.world
      link
      fedilink
      English
      arrow-up
      4
      arrow-down
      3
      ·
      10 hours ago

      Thats not true. The new article being shoved down lemmy’s throat is not correct. They site court cases and come to bad conclusions

        • terabyterex@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          1
          ·
          4 hours ago

          The copyright office never saif gpl could not be enforced. Thats s conclusion made. Hell even in what you linked the requirement for this is that ai had to be a “substantial” part. The linux teamsaid they would take submissions that were assisted but not all out generated. But to argue a point, lets pretend that an entire pull request was ai generated. That is only a small part of thr linux kernel since the kernel is what is licensed. A sma amount of uncopywrited code cant invalidate the whole project, which the license is on.

          But regardless, the copyright office never said anything about enforcement of gpl. T very clear said code with no meaningful human involvement, which isnt the case here. So nothing establishes what you said true. Its all leaping to comclusions that cant be leaped to.

          • hperrin@lemmy.ca
            link
            fedilink
            English
            arrow-up
            2
            ·
            3 hours ago

            The copyright office said material generated by AI is not copyrighted, even if that material is subsequently revised by the AI through additional prompts. That includes code. The GPL can only be used on copyrighted code. It is a copyleft license because it uses copyright law as a mechanism to enforce its terms. If you believe you can enforce a license on public domain material, that’s simply a gross misunderstanding of copyright law.

            Yes, it will hopefully be a very small part of the kernel, but what happens thirty years from now if the kernel is all AI generated code? It may be a slippery slope, but it’s a valid slippery slope. The more the kernel is AI generated, the less of it the license can cover.

      • hperrin@lemmy.ca
        link
        fedilink
        English
        arrow-up
        2
        ·
        5 hours ago

        Sure, you can license it whatever you want, but I can too, because it’s public domain. And neither of us can enforce those license terms on the other, because again, it’s public domain.

        • FauxLiving@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          2
          ·
          7 hours ago

          The status of generated code is ‘uncopyrightable’, which can be licensed.

          Copyright law determines the copyright status and contract law enforces the terms of contracts. They are two separate issues.

          If someone licenses you to use their AI generated code and you violate the license agreement, it doesn’t matter that they don’t have a claim under copyright law. They have a claim under contract law due to you violating the terms of the license (which is a contract).

            • hperrin@lemmy.ca
              link
              fedilink
              English
              arrow-up
              1
              arrow-down
              1
              ·
              edit-2
              5 hours ago

              Distributing under the GPL is a software license agreement which is absolutely a contract:

              A software license agreement is a legal contract that grants you permission to use software without transferring ownership. The software creator retains intellectual property rights while giving you specific usage rights under defined terms and conditions.

              - https://ironcladapp.com/journal/contracts/software-license-agreement

            • FauxLiving@lemmy.world
              link
              fedilink
              English
              arrow-up
              1
              arrow-down
              1
              ·
              7 hours ago

              That is the FSF’s position, but the case law has examples of cases where it was allowed to be treated by a contract.

              SFC v. Vizio, the Software Freedom Conservancy sued Vizio as a third-party beneficiary of the GPL as a contract, and the court allowed the case to proceed on that theory.

              • eleijeep@piefed.social
                link
                fedilink
                English
                arrow-up
                1
                ·
                7 hours ago

                Because in that case the copyright holder is the arbitrator of the terms under which their copyrighted material can be used and reproduced. If they did not own the copyright then any “license” would not be worth the paper it was written on and no judge would allow it to be treated as an implicit contract.