I am extremely curious what the general take around here is on the Singulairty - eviltoast

First, let me say that what broke me from the herd at lesswrong was specifically the calls for AI pauses. That somehow ‘rationalists’ are so certain advanced AI will kill everyone in the future (pDoom = 100%!) that they need to commit any violent act needed to stop AI from being developed.

The flaw here is that there’s 8 billion people alive right now, and we don’t actually know what the future is. There are ways better AI could help the people living now, possibly saving their lives, and essentially eliezer yudkowsky is saying “fuck em”. This could only be worth it if you actually somehow knew trillions of people were going to exist, had a low future discount rate, and so on. This seems deeply flawed, and seems to be one of the points here.

But I do think advanced AI is possible. And while it may not be a mainstream take yet, it seems like the problems current AI can’t solve, like robotics, continuous learning, module reuse - the things needed to reach a general level of capabilities and for AI to do many but not all human jobs - are near future. I can link deepmind papers with all of these, published in 2022 or 2023.

And if AI can be general and control robots, and since making robots is a task human technicians and other workers can do, this does mean a form of Singularity is possible. Maybe not the breathless utopia by Ray Kurzweil but a fuckton of robots.

So I was wondering what the people here generally think. There are “boomer” forums I know of where they also generally deny AI is possible anytime soon, claim GPT-n is a stochastic parrot, and make fun of tech bros as being hypesters who collect 300k to edit javascript and drive Teslas*.

I also have noticed that the whole rationalist schtick of “what is your probability” seems like asking for “joint probabilities”, aka smoke a joint and give a probability.

Here’s my questions:

  1. Before 2030, do you consider it more likely than not that current AI techniques will scale to human level in at least 25% of the domains that humans can do, to average human level.

  2. Do you consider it likely, before 2040, those domains will include robotics

  3. If AI systems can control robotics, do you believe a form of Singularity will happen. This means hard exponential growth of the number of robots, scaling past all industry on earth today by at least 1 order of magnitude, and off planet mining soon to follow. It does not necessarily mean anything else.

  4. Do you think that mass transition where most human jobs we have now will become replaced by AI systems before 2040 will happen

  5. Is AI system design an issue. I hate to say “alignment”, because I think that’s hopeless wankery by non software engineers, but given these will be robotic controlling advanced decision-making systems, will it require lots of methodical engineering by skilled engineers, with serious negative consequences when the work is sloppy?

*“epistemic status”: I uh do work for a tech company, my job title is machine learning engineer, my girlfriend is much younger than me and sometimes fucks other dudes, and we have 2 Teslas…

  • BrickedKeyboard@awful.systemsOP
    link
    fedilink
    English
    arrow-up
    1
    arrow-down
    2
    ·
    1 year ago

    Just I think to summarize your beliefs: rationalists are wrong about a lot of things and assholes. And also the singularity (which predates yuds existence) is not in fact possible by the mechanism I outlined.

    I think this is a big crux here. It’s one thing if its a cult around a false belief. It’s kind of a problem to sneer at a cult if the core S of it happens to be a true law of nature.

    Or an analogy. I think gpt-4 is like the data from the Chicago pile. That data was enough to convince the domain experts then a nuke was going to work to the point they didn’t test Fat Man, you believe not. Clearly machine generality is possible, clearly it can solve every problem you named including, with the help of humans, ordering every part off digikey and loading the pick and place and inspecting the boards and building the wire harnesses and so on.

    • froztbyte@awful.systems
      link
      fedilink
      English
      arrow-up
      12
      ·
      1 year ago

      Just I think to summarize your beliefs

      don’t be puttin’ words in my mouth yo

      rationalists

      this is a big set of very many people and lots of details

      are wrong about a lot of things

      many of them about many things, yes

      and assholes

      some, provably

      And also the singularity (which predates yuds existence) is not in fact possible by the mechanism I outlined

      whether it’s the wet dream of kurzweil or yud or whoever else, doesn’t matter? but as to the details… you’re engaging with this like the rats do (yes, told you, you only half escaped). you “set the example”, and then “test the details”

      just … don’t?

      the siren song of this is “okay what if I change the details of the experiment slightly?”

      we’ve had the trolley problem for ages, doesn’t mean it’s just “solved”. you won’t manage to “solve” whether the singularity can happen or not here, for the same reason

      • BrickedKeyboard@awful.systemsOP
        link
        fedilink
        English
        arrow-up
        1
        arrow-down
        2
        ·
        edit-2
        1 year ago

        I wanted to know what you know and I don’t. If rationalists are all scammers and not genuinely trying to be, per the name ‘lesswrong’ in their view of reality, what’s your model of reality. What do you know? So far unfortunately I haven’t seen anything. Sneer club’s “reality model” seems to be “whatever the mainstream average person knows + 1 physicist”, and it exists to make fun of the mistakes of rationalists and I assume ignores any successes if there are any.

        Which is fine, I guess? Mainstream knowledge is probably usually correct. It’s just that I already know it, there’s nothing to be learned here.

            • self@awful.systemsM
              link
              fedilink
              English
              arrow-up
              5
              ·
              1 year ago

              it is interesting how after their temporary ban, their focus shifted to “all you can see is the mainstream” meshed with “the governments and stock markets of the world agree with me, like, in secret”, as if these weren’t conflicting ideas. this is tragically reminiscent of the thought processes of several conspiracy theorists I’ve known

              • froztbyte@awful.systems
                link
                fedilink
                English
                arrow-up
                4
                ·
                1 year ago

                and the sudden “I may be too autistic” drops were super wtf too

                massive overall tone change without any revision of their position (even the barest acknowledgement of reflection was a “nuh-uh, still think I’m right” post)

          • BrickedKeyboard@awful.systemsOP
            link
            fedilink
            English
            arrow-up
            1
            arrow-down
            3
            ·
            1 year ago

            which is fine. the bigger topic is, could you leave a religion if the priest’s powers were real*, even if the organization itself was questionable?

            *real as in generally held to be real by all the major institutions in the world you are in. Most world governments and stock market investors are investing in AI, they believe they will get an ROI somehow.

        • froztbyte@awful.systems
          link
          fedilink
          English
          arrow-up
          4
          ·
          edit-2
          1 year ago

          Okay I realize this esteemed po(a)ster already got shown the door but I happened to read this reply days ago and it kept bugging me

          On the face of it, “What you know and I don’t” was a weird phrase to pick, in context of their later claims of “I work in the industry [and thus I totes know things]”

          The thing that really bugs me about it is that it’s as though this person is (apparently? willingly? by choice? something?) incapable of (a certain class of[0]) subjective (value) judgements. And even as I type that out I realize that’s almost certainly a hallmark of these types (only uncertain because I’ve never had to spend thought on that previously).

          [0] - they’ll go with “cold hard fact” but only their[1] cold hard fact

          [1] - this I also want to rantpost on, but probably needs some prep to give it a full NSFW-worthy posting

        • earthquake@lemm.ee
          link
          fedilink
          English
          arrow-up
          4
          ·
          1 year ago

          Mainstream knowledge is probably usually correct. It’s just that I already know it

          Still reflecting on this incredible claim.

    • gerikson@awful.systems
      link
      fedilink
      English
      arrow-up
      8
      ·
      1 year ago

      Or an analogy. I think gpt-4 is like the data from the Chicago pile. That data was enough to convince the domain experts then a nuke was going to work to the point they didn’t test Fat Man, you believe not.

      Are you mixing up Fat Man and Little Boy? Because Fat Man was an implosion-type bomb, just like the Trinity device. Little Boy was a gun-type. From vague memories of Rhode’s book, they wanted implosion types to maximize Pu weight to kiloton ratio, but it was much less straightforward than a gun-type bomb.

    • BernieDoesIt@kbin.social
      link
      fedilink
      arrow-up
      4
      ·
      1 year ago

      I think gpt-4 is like the data from the Chicago pile. That data was enough to convince the domain experts then a nuke was going to work to the point they didn’t test Fat Man, you believe not.

      Whoa whoa whoa there! I’m the contrarian that thinks that gpt is clearly more that just plagiarizing things, but it’s still just a step above Mad Libs in terms of intelligence. It’s not clear that you could get it to be smarter than a goldfish, let alone a human being. It’s just really good at stringing words together in a way that sounds good.