The problem with AI alignment is that humans aren't aligned - eviltoast

I’m sure there are some AI peeps here. Neural networks scale with size because the number of combinations of parameter values that work for a given task scales exponentially (or, even better, factorially if that’s a word???) with the network size. How can such a network be properly aligned when even humans, the most advanced natural neural nets, are not aligned? What can we realistically hope for?

Here’s what I mean by alignment:

  • Ability to specify a loss function that humanity wants
  • Some strict or statistical guarantees on the deviation from that loss function as well as potentially unaccounted side effects
  • JunctionSystem@lemmy.world
    link
    fedilink
    arrow-up
    11
    ·
    1 year ago

    C: AGI is possible. If it weren’t, we wouldn’t exist. The laws of physics permit the creation of conscious agents, therefore it is possible for one to be deliberately engineered.

    • Rhaedas@kbin.social
      link
      fedilink
      arrow-up
      3
      ·
      1 year ago

      Good counterpoint. I was more inferring not possible for us to figure out rather than impossible. We may simply be running out of time more than anything. Maybe that’s why the top people are so eager to push into unknowns (aside from the profitability of course). They see we have a small window of high tech opportunity and being too cautious and slow we’ll miss it. Naturally a big assumption is that AGI will be aligned and able to help us fix things, rather than the often portrayed versions that decide we are the problem, or that the problems are too insurmountable and the AI turns itself off.