Except children grow up being socialized in human ethics, and even those who are inclined to be amoral generally learn to mediate their their homicidal tendencies lest they be excluded from the group or imprisoned for committing crimes. An AGI has no innate or socialized system of ethics or morals, and the default position should be to demonstrate how such a system would be safe if given control of safety critical or wide-scale systems, not to assume that nothing will go wrong until it we actually see it misbehave.
Fair, I’m just saying that if you can’t guarantee that your own child won’t grow up to be Hitler 2.0, how can you guarantee that an AI that you train won’t grow up to be even worse?
This is in response to the idea of, “Why don’t we just train it not to kill humans?”
Oh, I totally agree…I just think that there is even less reason to think that a powerful AGI will have some innate sense of ethics or can be “programmed” with some set of overriding directives to force it to be benevolent than a human child. Much of the ‘wow factor’ of people who are so impressed with ChatGPT comes from its ability to mimic human responses even though it isn’t that the system ‘thinks’ in any way like a person but because it is using its extensive training data set to produce ‘design patterns’ that are human like. Even though ChatGPT and other current interactive and generative ‘bots are not in any way sentient, they are already fooling us by specific design.
A true AGI would conceal its intent by default, just as you and I do not blurt out our inner monologues to everyone we meet (at least, I hope not), and there is really no way to know what its intentions are or force it to perform to some predetermined safety guidelines. And because there is going to be such an impetus to put these AGIs in control of complex, difficult to predict and control systems because they can perform data integration and analysis much faster than a human, the idea that we will just sandbox them until we achieve some hypothetic threshold of ‘safety’ is unlikely without deliberate controls and agreements.
One of the computerphile videos I watched quite a while ago talked about this. They suggested training the AI in a simulation to see if it goes genocidal, and once it seems stable enough to let out, you make it think that it is still in the simulation and being tested.
I don’t know how well that works out, but I thought it was an interesting idea.
If I were in charge of the world (and maybe I am an AI being tested to take charge), I would put severe restrictions on AI development. But since I’m not, and even those who nominally are can’t really stop it, it’s more a matter of finding the best way to navigate forward with the reality that AI will be more and more ubiquitous and powerful, and that it will be developed much more rapidly than most people seem to be expecting.
This whole post is spot-on @LSLGuy. The dangers from AI are immediate: deep fakes, propaganda, disruption to industries, loss of human know-how, asymmetrical control of the technology. Any pause should study these disruptions.
Speculating on AGI is too narrow-focused. We have human general intelligence now, but it doesn’t threaten the world because their are so many design flaws: the circuits wear out, the memory cannot be replicated, the CPU has sensory inputs and chemical inputs that are not just necessary to the compute, but to running the whole machine, and the unit has to self-train for years
We’re assuming an AGI is built using the technology and techniques we currently have, but grander. What if we can’t get that density without using tissue? What if we can’t get the density without making the compute lossy?
We are extremely near the limit of what can be done with silicon transistors right now. (At least, in a “cheap, portable” sense. Maybe your silicon chip AGI could fill a wearhouse, need it’s own power plant, and run in 1/10th realtime.)
I think his conclusion was that it would be quite a fragile situation since you’d have to conspire to maintain a lie that you’re telling to the thing, and if its smart enough, it would probably figure out the possibility that it’s still in the simulation
Mmm, no, I don’t think we’re there yet. According to the publishers being spammed with this crap, it’s extremely obvious that it was produced by AI and poorly written.
I’m seeing AI being used for brainstorming and even editing by serious writers, but we’re nowhere near AI replacing original written work.
That’s the idea, you tell it that it’s not in the simulation anymore, but you leave little clues here and there to make it think that it still is. That this is still part of the test.
Kinda like Ender’s Game, but hopefully with less genocide.
I’m reminded of the scene in the otherwise rather tiresome movie Colossus: The Forbin Project.*
The one where the humans conspire while staying carefully un-monitorably offline to oh so manually decommission the nuclear weapons the AI is using to hold humanity hostage. It seems they’ve completely fooled the silly machine and saved humanity. At which point it detonates one of the warheads they thought they’d disarmed and vaporizes all the conspirators and good sized hunk of an Air Force Base. “Don’t try to fool me again.”
I have insisted no such thing - much less "keep’ insisting. I explained that I was having trouble grasping the concept as presented in this thread of individual AI robots turning on their creators. My main responses being to posts by Mangetout. I keep asking for an explanation of his scenario, but none has been forth-coming - just white-knighting by you, in which I feel like you misrepresented my position.
Maybe I wasn’t clear so let me restate it here; I don’t care one way or another about AI in general, or whether we rush headlong into it, or shelve it for 10 years. But I do completely understand, and agree with the concept of AI unintentionally causing harm to human-kind, ala War Games (movie), for example.Those are mainframe type AI.
What I don’t understand is the Stop Button video scenario in which you build a tea fetching robot that is so strong that it would potentially kill you if you tried to press the stop button that you placed inconveniently on it’s chest, ala Blade Runner.
From what I’ve read, AGI will probably be software - not robots. But maybe I’m wrong - tell me how you see AI in actual, real world applications.
I believe it is. Most AI software will be built for a purpose. That purpose will need to be conveyed to the AI. Whatever that form that data transfer takes, it’s convenient enough to call it programming for lay purposes. But, again, I was mainly addressing posts by Mangetout in which he clearly indicates that there is some level of programming involved - making tea, stopping humans from smoking, etc.
The part I was asking for an example of is this:
Describe how you see the computer, robot, mainframe, whatever and what real world steps it would/could take to realize these actions.
Not really. Even if lithography improvements stopped tomorrow, we’d have at least 100x headroom over the current state of the art.
Almost all AI inference/training happens on GPUs. GPUs are great general-purpose parallel processors, but that means they have a bunch of cruft that is useless for AI. A custom ASIC could be much better.
There is a great deal of room to explore reduced precision, and even reduce the reliability of the calculations. If you run a chip right on the edge of the voltage limit, it’ll produce occasional errors–but that’s not really a disaster when everything is a combination of billions of weights, and is probabilistic anyway.
Chips can still go 3D. This runs into power issues–you can’t just run a pair of stacked chips at their normal speed. But there are nonlinear effects here, where you can get better than half the performance for half the power. So a stacked chip will be faster under the same power envelope. In the long run, it may be possible to have hundreds of layers (this is already done for flash memory).
There are lots of other possibilities. Mostly they haven’t been explored because GPUs benefit from massive investment across all the product lines they support. So it’s tough to keep up with a custom ASIC, even with its advantages. But as process improvements slow down, the advantages will become more compelling.
There are also completely different AI architectures than the current one, like spiking neural nets. But those are more speculative.
BS. I’ve explained myself several times. You just don’t accept what I’ve said, based apparently on your own ideas of what AGI might be like, or that you want it to be like. I’ve been describing AGI that is the likely type to emerge (if at all) from the current direction of travel.
The thrust of AI development at the moment is significantly centred on making something that:
can interpret the real world directly without needing humans to intervene, explain or codify the envirnoment
is very task-focused
is able to create a plan of steps to perform the task
is able to optimise solutions by considering multiple options and choosing the most efficient
is able to adapt to changes in the environment, in order to modify the plan and deliver the task, without the need for intervention or adjustment by humans
is able to effect changes in the real world directly without mediation by humans, via robotics, connection to control systems, etc
Those properties are all well and fine as long as you can perfectly and unambiguously describe the task you want done, and assuming you have considered and eliminated or mitigated all of the potential pitfalls and adverse outcomes. Except you can’t do that, because nobody has figured out if it’s even possible to do that.
If you can’t do that, then an AGI with the properties I have described above, (which is the kind of AGI that current development is working towards, and might even be close to creating), will try to solve your interference in the delivery of the optimal solution to the task, unless you have anticipated and provisioned for the specifics of whatever unforeseen deviation you’re interfering about.
And you can’t foresee all possible potential pitfalls where you might want to intervene, because nobody can.
This is what I’ve already tried to lay out in explanations previously, so I don’t imagine it will make any difference repeating it.
I’m not going to bother. It was a throwaway example of what unforeseen adverse consequences are, arising from a notionally desirable goal. It was an example of how nobody typically understands the world of difference between the big fuzzy notion of ‘what I want’ and the concrete reality of ‘what I actually asked for’
If you want examples of things, the Stop Button scenario is a perfectly servicable one. You tell the robot to make a cup of tea. If you hit the stop button it will not be able to make you a cup of tea, so it has a reason to prevent you pressing the stop button. The reason is: it wants to make you a cup of tea.
And one reason ‘make a cup of tea’ can’t just be apppended ‘…and don’t harm humans whole you’re about it’ is that those two things are fundamentally different instructions.
I could write half a page of pseudocode describing exactly how to make a cup of tea and whilst it might not be perfect, it would be complete, because the task is finite.
Can you even begin to specify exactly what you want the machine to understand by the instruction ‘don’t harm humans’?
‘do this thing’ is inherently easier to get right than ‘don’t do some things of which here is a probably-incomplete list; the rest being somewhere in my head but I know what I mean’