Google apologizes for ‘missing the mark’ after Gemini generated racially diverse Nazis - eviltoast

Google apologizes for ‘missing the mark’ after Gemini generated racially diverse Nazis::Google says it’s aware of historically inaccurate results for its Gemini AI image generator, following criticism that it depicted historically white groups as people of color.

  • crimsonpoodle@pawb.social
    link
    fedilink
    English
    arrow-up
    3
    ·
    9 months ago

    Just run ollama locally and download uncensored versions— runs on my m1 MacBook no problem and is at the very least comparable to chatgpt3. Unsure for images though, but there should be some open source options. Data is king here, so the more you use a platform the better its AI gets (generally) so don’t give the corporations the business.

    • FinishingDutch@lemmy.world
      link
      fedilink
      English
      arrow-up
      4
      ·
      9 months ago

      I’ve never even heard of that, so I’m definitely going to check that out :D I’d much prefer running my own stuff rather than sending my prompts to god knows where. Big tech already knows way yoo much about us anyway.

      • intensely_human@lemm.ee
        link
        fedilink
        English
        arrow-up
        2
        ·
        9 months ago

        I love teaching GPT-4. I’ve given permission for them to use my conversations with it as part of future training data, so I’m confident what I teach it will be taken up.

    • intensely_human@lemm.ee
      link
      fedilink
      English
      arrow-up
      2
      ·
      9 months ago

      How powerful is ollama compared to say GPT-4?

      I’ve heard GPT-4 uses an enormous amount of energy to answer each prompt. Are the models runnable on personal equipment once they’re trained?

      I’d love to have an uncensored AI

      • crimsonpoodle@pawb.social
        link
        fedilink
        English
        arrow-up
        1
        ·
        9 months ago

        Llama2 is pretty good but there are a ton of different models which have different pros and cons, you can see some of them here: https://ollama.com/library . However I would say that as a whole models are generally slightly less polished compared to chat gpt.

        To put it another way: when things are good they’re just as good, but when things are bad the AI will start going off the rails, for instance holding both sides on the conversation, refusing to answer, just saying goodbye, etc. More “wild westy” but you can also save the chats and go back to them so there are ways to mitigate, and things are only getting better.