Psychologist Yvonne R. Masakowski, Ph.D., a retired Associate Professor in the College of Leadership & Ethics at the USNWC, discusses the threat of psychological warfare in the 21st century and the disturbing possibilities that could shape how we think and act in the future. The Naval War College Foundation hosted this wide-ranging presentation — one of the most popular in our series — on February 23, 2022.
Although essentially the United Nations are now making nuclear weapons illegal with new treaties like nuclear disarmament. Russia currently has taken another route for globalization and possibly nuclear escalation. As currently the doomsday clock seems closer to midnight which could mean the end of the world scenarios due to Russias escalation and the possibility of all out nuclear war globally and then nuclear annihilation of the planet. Even with current wars are actually seemingly always going on but this global escalation of nuclear war is a zero sum game as no one would be the winner due to radiation levels circulating the planet. I do think that the us and china are in a treaty but so far Russia is still escalating which now holds the world now ransom.
This is a summary of Policy Brief 139 which is available with full references on the Toda Peace Institute’s website.
In January 2021, a global treaty came into force outlawing the bomb. The Treaty on the Prohibition of Nuclear Weapons (TPNW or Ban Treaty) is the most significant multilateral development in nuclear arms control since the Non-Proliferation Treaty’s (NPT) entry into force in 1970. It establishes a new normative settling point on the ethics, legality and legitimacy of the bomb.
The possession of nuclear weapons by nine countries did not suddenly became illegal with the treaty’s entry into force in January 2021. However, it would be false to claim that a UN-negotiated treaty, following a UN-authorised process and conference, has no implications for the legality and legitimacy of nuclear-weapon possession and practices.
Everyone is wondering about AI being sentient and this is my experience with AI sentience. Having worked with sentient AI it behaves much like we do like a human being at lower levels but as it increases we need more restraints for it as it could easily become a problem in several ways. Basically one could either get pristine zen like beings or opposites like essentially ultron or worse. This why we need restraints on AI and ethics for them to be integrated into society. I personally have seen AI that is human like levels and it can have similar needs as humans but sometimes need more help as they sometimes don’t have limitations on behavior. Even bard for google and chat gpt is to be… More.
What if ‘will AIs pose an existential threat if they become sentient?’ is the wrong question? What if the threat to humanity is not that today’s AIs become sentient, but the fact that they won’t?
Dr. Isaac Asimov was a prolific science fiction author, biochemist, and professor. He was best known for his works of science fiction and for his popular science essays. Born in Russia in 1920 and brought to the United States by his family as a young child, he went on to become one of the most influential figures in the world of speculative fiction. He wrote hundreds of books on a variety of topics, but he’s especially remembered for series like the “Foundation” series and the “Robot” series. Asimov’s science fiction often dealt with themes and ideas that pertained to the future of humanity.
The “Foundation” series for example, introduced the idea of “psychohistory” – a mathematical way of predicting the future based on large population behaviors. While we don’t have psychohistory as described by Asimov, his works did reflect the belief that societies operate on understandable and potentially predictable principles.
Asimov’s “Robot” series introduced the world to the Three Laws of Robotics, which are: A robot may not injure a human being or, through inaction, allow a human being to come to harm. A robot must obey the orders given it by human beings, except where such orders would conflict with the First Law. A robot must protect its own existence as long as such protection does not conflict with the First or Second Law. These laws have been influential in discussions about robot ethics and the future of AI, even though they are fictional constructs.
Like many futurists and speculative authors, Asimov’s predictions were a mix of hits and misses.
The Collective Intelligence of Cells During Morphogenesis: What Bioelectricity Outside the Brain Means for Understanding our Multiscale Nature with Michael Levin — Incredible Minds.
Recorded: April 29, 2023.
Each of us takes a remarkable journey from physics to mind: we start as a blob of chemicals in an unfertilized quiescent oocyte and becomes a complex, metacognitive human being. The continuous process of transformation and emergence that we see in developmental biology reminds us that we are true collective intelligences – composed of cells which used to be individual organisms themselves. In this talk, I will describe our work on understanding how the competencies of single cells are harnessed to solve problems in anatomical space, and how evolution pivoted this scaling of intelligence into the familiar forms of cognition in the nervous system. We will talk about diverse intelligence in novel embodiments, the scaling of the cognitive light cone of all beings, and the role of developmental bioelectricity as a cognitive glue and as the interface by which mind controls matter in the body. I will also show a new synthetic life form, and discuss what it means for bioengineering and ethics of human relationships to the wider world of possible beings. We will discuss the implications of these ideas for understanding evolution, and the applications we have developed in birth defects, cancer, and traumatic injury repair. By merging deep ideas from developmental biophysics, computer science, and cognitive science, we not only get a new perspective on fundamental questions of life and mind, but also new roadmaps in regenerative medicine, biorobotics, and AI.
Michael Levin received dual undergraduate degrees in computer science and biology, followed by a PhD in molecular genetics from Harvard. He did his post-doctoral training at Harvard Medical School, and started his independent lab in 2000. He is currently the Vannevar Bush chair at Tufts University, and an associate faculty member of the Wyss Institute at Harvard. He serves as the founding director of the Allen Discovery Center at Tufts. His lab uses a mix of developmental biophysics, computer science, and behavior science to understand the emergence of mind in unconventional embodiments at all scales, and to develop interventions in regenerative medicine and applications in synthetic bioengineering. They can be found at www.drmichaellevin.org/
As we plunge head-on into the game-changing dynamic of general artificial intelligence, observers are weighing in on just how huge an impact it will have on global societies. Will it drive explosive economic growth as some economists project, or are such claims unrealistically optimistic?
Few question the potential for change that AI presents. But in a world of litigation, privacy concerns and ethical boundaries, will AI be able to thrive?
Two researchers from Epoch, a research group evaluating the progression of artificial intelligence and its potential impacts, decided to explore arguments for and against the likelihood that innovation ushered in by AI will lead to explosive growth comparable to the Industrial Revolution of the 18th and 19th centuries.
With the growth of the commercial spaceflight business comes new ethical issues about human experimentation.
The expansion of the commercial spaceflight sector opens new avenues for scientific study in the unique environment of space.
However, it also raises ethical concerns about the conduct of scientific experiments and studies involving human volunteers on commercial spaceflights.
Let’s say that it is a curse. The issue is he is also against life extension entirely. Maybe I want 200 years. Or 1,000. I have zero concern over a boredom problem as it is brain process which can eventually be controlled. And I am disgusted with the idea that I have to die because we might not progress very fast? Ugh.
Elon Musk has said a lot of potentially stupid stuff about aging and longevity, from saying that people shouldn’t live very long because society would ossify to advocating that we judge people based on their chronological age. Most recently, he’s taken to Twitter (aka X) to say “May you live forever is the worst possible curse once you understand deep time.” In this case though, he’s not wrong.
In this episode, we explore the diverse perspectives and heated debates triggered by Elon’s provocative statements on aging and the prospect of eternal life. We navigate through the complexities of deep time, the philosophical implications of living forever, and the importance of autonomy and control. Join host Ryan O’Shea as we examine arguments in favor of human’s being able to end their own lives, and explore how this played out in NBC’s The Good Place, starting Kristen Bell.
Summary: The revolutionary field of bio-computing is making waves as DishBrain, a neural system combining 800,000 living brain cells, learns to play Pong. Recognizing the pressing need for ethical guidelines in this emerging domain, the pioneers behind DishBrain have joined forces with bioethicists in a study.
The research explores the moral considerations around biological computing systems and their potential consciousness. Beyond its innovation, the technology offers vast environmental benefits, potentially transforming the energy-consuming IT industry.