Signs of undeclared ChatGPT use in papers mounting - eviltoast
    • FaceDeer@kbin.social
      link
      fedilink
      arrow-up
      2
      ·
      1 year ago

      So you’re telling me that the code it generated for me wasn’t good and useful, and that when I told it to correct errors it actually did introduce new errors and restore old ones, contrary to what I just said? Guess all that stuff I got done using its help didn’t actually get done after all and I’m descending ever deeper into a world of delusion, thinking my projects are finished and working when in fact they aren’t.

      Obviously if you’re trying to get it to use APIs from after 2021 that’s not going to work. It also won’t bake you a cake if you ask it to. Use tools for the tasks they’re good for, don’t use them for things you know they can’t do.

      • boredazfcuk@mastodon.social
        link
        fedilink
        arrow-up
        0
        ·
        1 year ago

        @FaceDeer @floofloof @henfredemars @PoisonedPrisonPanda @sab @Vilian the 500 char limit made me pick the first failure that spring to mind. Maybe you forget, but AI wasn’t trained on “good” data. It was trained “all” data and large amounts of that is plain wrong. People with problems pasting blocks of code and responses correcting a single line. ChatGPT isn’t smart enough to merge those into a single block of working code.

        • FaceDeer@kbin.social
          link
          fedilink
          arrow-up
          1
          ·
          1 year ago

          You have to tell ChatGPT that you want good code, then.

          I’m actually serious. If you just ask for something generic, it’ll assume you want something generic. If you ask it for something that’s “high efficiency, well commented and maintainable” then it’s going to know you wanted that and give you something more along those lines. Just like if you asked it for something “that looks crappy and sloppy, like an amateur wrote it.”

          Very often when people complain about ChatGPT’s “style” or say they can immediately spot something that “sounds like” ChatGPT it’s because they’re not giving it good directions. It can’t read your mind. Yet.

          • boredazfcuk@mastodon.social
            link
            fedilink
            arrow-up
            0
            ·
            1 year ago

            @FaceDeer @floofloof @henfredemars @PoisonedPrisonPanda @sab @Vilian nope. I ask for highly precise stuff. When I say, “I’m no coder,” it’s coz I use interpreters and don’t compile “real” code, plus it only accounts for 10% of my day job. ChatGPT maybe useful for Hello World, Towers of Babel or other stuff it scraped from udemy, but when you ask it to assist in automating complex production systems, it really falls down.

            • FaceDeer@kbin.social
              link
              fedilink
              arrow-up
              1
              ·
              1 year ago

              Maybe ChatGPT just hates you personally, then.

              You’re saying “it can’t work for anyone because it doesn’t work for me!” And I’m saying “well, it worked for me, so maybe you’re using it wrong.”

              You can’t insist it’s not working for me because it did. I’m not disputing that it didn’t work for you, all I can suggest is reading up a bit on prompt engineering to see if you can find out what you’re doing differently.