Advisory Board

Michael Anissimov

Chapter six of The Singularity Is Near : When Humans Transcend Biology by Ray Kurzweil began with the following quote from Michael Anissimov

One of the biggest flaws in the common conception of the future is that the future is something that happens to us, not something we create.

He has also said

I cannot emphasize this enough. If an existential disaster occurs, not only will the possibilities of extreme life extension, sophisticated nanotechnology, intelligence enhancement, and space expansion never bear fruit, but everyone will be dead, never to come back. This would be awful. Because we have so much to lose, existential risk is worth worrying about even if our estimated probability of occurrence is extremely low.
It is not the funding of life extension research projects that immortalists should be focusing on. It should be projects that decrease the risk of existential risk. By default, once the probability of existential risk is minimized, life extension technologies will be developed and applied. There are powerful economic and social imperatives in that direction, but few towards risk management. Existential risk creates a ‘loafer problem’ — we always expect someone else to do it. I assert that this is a dangerous strategy and should be discarded in favor of making prevention of such risks a central focus.

Michael Anissimov writes and speaks on futurist issues, especially the relationships between accelerating change, nanotechnology, existential risk, transhumanism, and the Singularity. His popular blog Accelerating Future discusses these issues regularly.
Michael was a founding director of the nonprofit Immortality Institute, the first organization focused on the abolition of nonconsensual death. He is a member of the World Transhumanist Association, an associate of the Institute for Accelerating Change, and a member of the Center for Responsible Nanotechnology’s Global Task Force. Michael was Cofounder and Director of the Singularity Summit and was Media Director of the Machine Intelligence Research Institute.
His central fields of study outside of futurism are normative rationality, models for judgement under uncertainty, and heuristics and biases research. His Concise Introduction to Heuristics and Biases is the second Google result for the term.
He has authored More Dangers From Molecular Nanotechnology, What is Transhumanism?, 10 Simple Ways You Can Help the Technological Singularity, Deconstructing Asimov’s Laws, and Fountains of Youth: Hacking the Maximum Lifespan.
As a science and technology writer, Michael contributes to the Q&A website WiseGeek. He has also authored dozens of papers on transhumanism, futurism, and the Singularity. Over three hundred pages of his writing is online, much of which can be found at his personal website. Michael has been presenting on futurist issues since 2001, and has given talks to audiences at technology and philosophy conferences in San Francisco, Las Vegas, Los Angeles, and at Yale University.
A leading voice on the technological Singularity, Michael was quoted multiple times in Ray Kurzweil’s 2005 book The Singularity is Near: When Humans Transcend Biology. He lives in San Francisco, CA.
Read the transcript of his Future Blogger interview. Listen to Michael at the Singularity University, on FastForward Radio, and on The Future And You.