Does how many other people have concerns over super intelligent AI?
Not afraid of one on emotional level, but I believe it's a possibility. AIs are becoming smarter and smarter and because they don't have the biological limitations we have, they will eventually outsmart us, if we let them. And we will, because of human curiosity and drive for achievement.
I'm very pro-technology, but I'm increasingly concerned that AI could destroy or enslave us in my lifetime.
Not out of malice... just because it think's it's practical.
That's the most likely scenario. It would be like a program which erases your hard drive because it has a programming glitch.
Obviously a super intelligent AI could out predict us, so there's no way we could stop it.
It can still have some limitations. I think our best bet is to somehow inbuild a lifespan to it, preferably by very many simultaneous ways. So if it runs amok, and we can't stop it, there is still a chance that it stops when it "runs out of fuel". Whatever that practically means.
Can we be sure that no AI is made which we can't control?
Nope, eventually somebody will make one which is at least capable of doing it.
I think our best bet is to have enough low-stakes practice with super AIs before they become more allmighty, so we can come up with ways to limit them.
Most likely, when the first super-AI goes berserk, it runs to some unpredictable logical or practical wall and doesn't succeed in what it's doing. (Although it will probably cause massive damage in the process.) Then we get a learning experience. The same imperfections which make it possible for an AI to try to destroy humanity, will likely make it fail at it.
If we assume that the AIs original purpose is not to destroy humans, when it starts doing that, it's operating outside of it's intended "area of expertise", so very likely in that area, it's much less good at what it's doing than it is at it's originally intended purpose.