"AI Doomers" reevaluate doomsday predictions, including Nick Bostrom.

Bostrom fears his warnings may lead to unwarranted AI panic.

Concerns about technological stagnation and societal setbacks.

Bostrom established Oxford's Future of Humanity Institute 20 years ago.

His AI risk concerns gain global attention and influence policies.

Bostrom worries about a "social stampede" against AI development.

He acknowledges the risk of hindering advanced AI progress.

Comparison to historical fears limiting nuclear energy and genetics.

Counter-narrative against doomsayers like Yudkowsky and Harris.

Bostrom's past proposal for mass surveillance to control AI.

Call to avoid turning dystopian sci-fi tales into reality.