Jeffrey HerrlichThe article Existential Risk and Fermi’s Paradox said
The odds may be so heavily stacked against us that the probability of success is only 0.0000000000001% for any given civilization (or worse). That doesn’t mean that we can’t possibly be that one civilization. And it doesn’t mean we shouldn’t try. What if the goal is possible (albeit very remotely possible) but all civilizations decide to give up prematurely. That would ultimately make this an entirely pointless Universe.
I’m starting to believe more and more, that a very large “fraction” of the paradox, is that an evolved intelligence like us is simply extremely rare in this Universe.
Jeffrey Herrlich is the author of this article and a Singularitarian
who has been aware of the
Singularity hypothesis and its implications since 2005
The Singularity is Near by
He has a deep
desire to witness progress toward the mitigation of existential
In particular, he believes that the vigorous pursuit of “Friendly
as advocated by SIAI, is the most promising pathway toward a
future for all sentient beings originating from Earth.
Jeff is a Sustaining Donor of SIAI and is currently a university student pursuing a bachelor’s degree in Computer Science.
He says “If we are not careful with the design of the first Strong AI, one possibility is that the AI will pursue a course which we humans would consider both of no value to humanity and of no value to the AI. A classic illustration is a Strong AI, without independent motivations, that ceaselessly pursues a trivial goal such as tiling the solar system with optimal paper clips, and destroying humanity in the process. It is for this reason among others that I believe that the design of a ‘Friendly AI’ is a win : win situation both for humanity and for the Strong AI.”