Toggle light / dark theme

Many are worried about the rise of artificial intelligence (A.I.). Could A.I. become sentient (able to perceive or feel things)? What does that mean for how we look at these machines? An ex-google engineer has claimed that Google has actually created a sentient AI. Dr. Jeffrey reacts to these claims in the video from Help Me Understand.

Dr. Thomas R. Jeffrey is both the host for Help Me Understand as well as a professor at Campbellsville University. His research centers around the uses and implementation of artificial intelligence.

Dr. Jeffrey reacts to Bloomberg Tech’s video: https://youtu.be/kgCUn4fQTsc.

In this video, Dr. Jeffrey, our residential AI expert, gives insight to the aspect of A.I and machine learning. From a technical point of view. He also suggests that we might need to look at creating ethical guidelines to how we interact with A.I.

Jeff Lichtman is Jeremy R. Knowles Professor of Molecular and Cellular Biology at Harvard. He received an A.B. from Bowdoin (1973), and an M.D. and Ph.D. from Washington University in St. Louis (1980) where he worked for 30 years before moving to Cambridge (2004). He is a member of Harvard’s newly established Center for Brain Science. Jeff’s research interests revolve around the question of how mammalian brain circuits are physically altered by experiences, especially in early life. He has focused on the dramatic re-wiring of neural connections in early postnatal development. More recently his research has focused on developing new electron microscopy methods to map the entire wiring diagram of the developing and adult brain. One of the principal aims of this “connectomics” approach is to uncover the ways information is stored in neural networks.

In the spirit of ideas worth spreading, TEDx is a program of local, self-organized events that bring people together to share a TED-like experience. At a TEDx event, TEDTalks video and live speakers combine to spark deep discussion and connection in a small group. These local, self-organized events are branded TEDx, where x = independently organized TED event. The TED Conference provides general guidance for the TEDx program, but individual TEDx events are self-organized.* (*Subject to certain rules and regulations)

On January 18, 2013, Caltech hosted TEDxCaltech: The Brain, a forward-looking celebration of humankind’s quest to understand the brain, by exploring the past, present and future of neuroscience. Visit TEDxCaltech.com for more details.

In this episode, Emad and Peter discuss everything from AI-generated content and property rights to ethical implications, along with the upcoming hyper-disruption wave of technology in all industries.

Emad Mostaque is the CEO and Co-Founder of Stability AI, a company funding the development of open-source music-and image-generating systems such as Dance Diffusion and Stable Diffusion.

Learn about Stability AI: https://platform.stability.ai/

Access Stable Diffusion: https://github.com/CompVis/stable-diffusion.

After the release of The Jazz Singer in 1927, all bets were off for live musicians who played in movie theaters. Thanks to synchronized sound, the use of live musicians was unnecessary — and perhaps a larger sin, old-fashioned. In 1930 the American Federation of Musicians formed a new organization called the Music Defense League and launched a scathing ad campaign to fight the advance of this terrible menace known as recorded sound.

The evil face of that campaign was the dastardly, maniacal robot. The Music Defense League spent over $500,000, running ads in newspapers throughout the United States and Canada. The ads pleaded with the public to demand humans play their music (be it in movie or stage theaters), rather than some cold, unseen machine. A typical ad read like this one from the September 2, 1930 Syracuse Herald in New York:

Tho’ the Robot can make no music of himself, he can and does arrest the efforts of those who can.

Scientists have labored for decades to understand how brain structure and functional connectivity drive intelligence. A new analysis offers the clearest picture yet of how various brain regions and neural networks contribute to a person’s problem-solving ability in a variety of contexts, a trait known as general intelligence, researchers report.

They detail their findings in the journal Human Brain Mapping.

The study used “connectome-based predictive modeling” to compare five theories about how the gives rise to , said Aron Barbey, a professor of psychology, bioengineering and neuroscience at the University of Illinois Urbana-Champaign who led the new work with first author Evan Anderson, now a researcher for Ball Aerospace and Technologies Corp. working at the Air Force Research Laboratory.

Basically we need a sorta vision from marvel comics to become a reality or a God in machine device otherwise we could easily see AI events that could be not as positive like a demon in a box. I personally have seen glimpses of these kinda AI that could have endless needs because they don’t really have limits. Not all AI behave this way most are just automatons but if they have sentience which I have seen that is evil it could be anything from something of a small threat to even like a ultron. That is why we need to evolve past AI to be our own superintelligence whether that be a biological singularity or robot like abilities.


TECHNOLOGY may be too pervasive in today’s world and could hinder our decision-making process, experts have warned.

By now, most people have used an AI-powered device as technology has become ubiquitous worldwide.

This can look like having Amazon’s Alexa set a timer for you or asking Apple’s Siri to check the weather.

𝐓𝐡𝐢𝐬 𝐢𝐬 𝐲𝐨𝐮𝐫 𝐛𝐫𝐚𝐢𝐧 𝐨𝐧 𝐜𝐨𝐝𝐞

𝙈𝙄𝙏 𝙧𝙚𝙨𝙚𝙖𝙧𝙘𝙝𝙚𝙧𝙨 𝙖𝙧𝙚 𝙙𝙞𝙨𝙘𝙤𝙫𝙚𝙧𝙞𝙣𝙜 𝙬𝙝𝙞𝙘𝙝 𝙥𝙖𝙧𝙩𝙨 𝙤𝙛 𝙩𝙝𝙚 𝙗𝙧𝙖𝙞𝙣 𝙖𝙧𝙚 𝙚𝙣𝙜𝙖𝙜𝙚𝙙 𝙬𝙝𝙚𝙣 𝙖 𝙥𝙚𝙧𝙨𝙤𝙣 𝙚𝙫𝙖𝙡𝙪𝙖𝙩𝙚𝙨 𝙖 𝙘𝙤𝙢𝙥𝙪𝙩𝙚𝙧 𝙥𝙧𝙤𝙜𝙧𝙖𝙢


Researchers from MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL) found that the Multiple Demand and Language brain systems encode specific code properties and uniquely align with machine-learned representations of code.