AI-Generated Code is Causing Outages and Security Issues in Businesses - eviltoast
  • reka@lemmy.world
    link
    fedilink
    English
    arrow-up
    73
    arrow-down
    7
    ·
    2 months ago

    As stated in the article, this has less to do with using AI, more to do with sloppy code reviews and code quality enforcement. Bad code from AI is just the latest version of mindlessly pasting from Stack Overflow.

    I encourage jrs to use tools such as Phind for solving problems but I also expect them to understand what they’re submitting and be ready to defend it no differently to any other PR. If they’re submitting code they don’t understand that’s incredibly unprofessional and I would come down very hard on them. They don’t do this though because we don’t hire dickheads.

    • Wrench@lemmy.world
      link
      fedilink
      English
      arrow-up
      21
      ·
      edit-2
      2 months ago

      Shift-left eliminated the QA role.

      Now we have AI generated shit code, with devs that don’t understand the low level details of both the language, and the specifics of the generated code.

      So we basically have content entry (ai inputs) and extremely shitty QA bundled into the “developer” role.

      As a 20 year veteran of the industry, people keep asking me if I think AI will make developers obsolete. I keep telling them “maybe some day, but today’s LLMs are not it. The AI bubble is going to burst, and a few legit use cases will make it through”

    • technocrit@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      15
      arrow-down
      1
      ·
      edit-2
      2 months ago

      Bad code from AI is just the latest version of mindlessly pasting from Stack Overflow.

      Humans literally can not scan all of SO to make a huge copypasta.

      It takes much more time, effort, and thought to find various solutions on SO and patch them together into something that works well.

    • MonkderVierte@lemmy.ml
      link
      fedilink
      English
      arrow-up
      15
      arrow-down
      2
      ·
      edit-2
      2 months ago

      Yeah but… i asked chatgpt once how to style something in asciidoctors style.yml. It proposed me html syntax (some inline stuff can be done with html tags in asciidoctor, if output is html). After the usual apology, it suggested some wrong yaml. Third try, because formatting was wrong, it mixed them both.

      I mean, sure, some niche usecase in a somewhat obscure (lots of moving parts) lightweight markup. But still, this was a lesson.

    • BlameTheAntifa@lemmy.world
      link
      fedilink
      English
      arrow-up
      9
      ·
      2 months ago

      this has less to do with using AI, more to do with sloppy code reviews and code quality enforcement.

      They are the same picture.

    • forrcaho@lemmy.world
      link
      fedilink
      English
      arrow-up
      4
      ·
      2 months ago

      We used to have these shit developers and I accepted a lot of bad code back then – if it actually worked – because otherwise “code review” is full-on training, which is an entire other job from the one I was hired to do.

      The client ditched that contracting firm, and the devs I work with now are worth putting in time on code review with – but damn, we got hella shit code in our codebase to deal with now. Some of it got tossed, some of it … we live with.