Toggle light / dark theme

Artificial Intelligence & Robotics Tech News For October 2022

Deep Learning AI Specialization: https://imp.i384100.net/GET-STARTED
AI News Timestamps:
0:00 New AI Robot Dog Beats Human Soccer Skills.
2:34 Breakthrough Humanoid Robotics & AI Tech.
5:21 Google AI Makes HD Video From Text.
8:41 New OpenAI DALL-E Robotics.
11:31 Elon Musk Reveals Tesla Optimus AI Robot.
16:49 Machine Learning Driven Exoskeleton.
19:33 Google AI Makes Video Game Objects From Text.
22:12 Breakthrough Tesla AI Supercomputer.
25:32 Underwater Drone Humanoid Robot.
29:19 Breakthrough Google AI Edits Images With Text.
31:43 New Deep Learning Tech With Light waves.
34:50 Nvidia General Robot Manipulation AI
36:31 Quantum Computer Breakthrough.
38:00 In-Vitro Neural Network Plays Video Games.
39:56 Google DeepMind AI Discovers New Matrices Algorithms.
45:07 New Meta Text To Video AI
48:00 Bionic Tech Feels In Virtual Reality.
53:06 Quantum Physics AI
56:40 Soft Robotics Gripper Learns.
58:13 New Google NLP Powered Robotics.
59:48 Ionic Chips For AI Neural Networks.
1:02:43 Machine Learning Interprets Brain Waves & Reads Mind.

What Meta’s Galactica missteps mean for GPT-4

Check out the on-demand sessions from the Low-Code/No-Code Summit to learn how to successfully innovate and achieve efficiency by upskilling and scaling citizen developers. Watch now.

Like Rodin’s The Thinker, there was plenty of thinking and pondering about the large language model (LLM) landscape last week. There were Meta’s missteps over its Galactica LLM public demo and Stanford CRFM’s debut of its HELM benchmark, which followed weeks of tantalizing rumors about the possible release of OpenAI’s GPT-4 sometime over the next few months.

The online chatter ramped up last Tuesday. That’s when Meta AI and Papers With Code announced a new open-source LLM called Galactica, that it described in a paper published on Arxiv as “a large language model for science” meant to help scientists with “information overload.”

A simpler path to better computer vision

Before a machine-learning model can complete a task, such as identifying cancer in medical images, the model must be trained. Training image classification models typically involves showing the model millions of example images gathered into a massive dataset.

However, using real image data can raise practical and : The images could run afoul of copyright laws, violate people’s privacy, or be biased against a certain racial or ethnic group. To avoid these pitfalls, researchers can use image generation programs to create for model training. But these techniques are limited because expert knowledge is often needed to hand-design an image generation program that can create effective training data.

Researchers from MIT, the MIT-IBM Watson AI Lab, and elsewhere took a different approach. Instead of designing customized image generation programs for a particular training task, they gathered a dataset of 21,000 publicly available programs from the internet. Then they used this large collection of basic image generation programs to train a computer vision model.

Is GPT-4 going to be Artificial General Intelligence? + GPT-4 Release Date

Use my link to get 50% off of your Skillshare subscription: https://skl.sh/AINews50 — Invest in yourself this holiday season taking advantage of Skillshare’s best deal of the year! Valid for a limited time only.
GPT-4 is about to be released to the public and is supposedly close to what some may consider Artificial General Intelligence. In this video, I’ll talk all about the possible applications and abilities which GPT4 will have.

TIMESTAMPS:
00:00 Passing the Turing Test.
02:41 Why such a big Secret?
03:58 What is GPT-4 going to be?
06:32 Last Words.

#ai #gpt4 #openai

Dr Renée Deehan — VP, Science & AI, InsideTracker — Evidence-Based And Actionable Wellness Solutions

Evidence-Based And Actionable Health, Wellness And Longevity Solutions — Dr. Renee DeHaan, Ph.D. — VP, Science & AI, InsideTracker


Dr. Renée Deehan, Ph.D. is the VP of Science & Artificial Intelligence at InsideTracker (https://www.insidetracker.com/), and leads a science team that builds and mines the world’s largest data set of blood, DNA, fitness tracking and phenotypic data from healthy people, creating evidence-based solutions that are simple, clear, and actionable.

Dr. Deehan has spent her career working in the precision medicine and personalized nutrition domains, previously serving as the VP of Computational Biology & Translational Informatics at QuartzBio and as the VP of Biology and Bioinformatics at PatientsLikeMe, the world’s largest integrated community, health management, and real-world data platform.

At PatientsLikeMe, Dr. Deehan was responsible for data and knowledge engineering, AI/machine-learning, and translational biology functions that drove infrastructure and consumer & business product development. She was also the Principal Investigator for the DigitalMe Ignite program, which collected longitudinal blood and patient-generated health data from over 5,000 at-home site visits from over 2,000 participants and was able to generate over 2 Million data points from the DigitalMe program.

Dr. Deehan also designed and cross-functionally implemented the first generation of an “advanced research platform”, capable of integrating survey and omics data for biomarker analysis, including ensemble-based machine-learning pipelines. Additionally, they developed an outsourced pipeline to support their wet-lab omics needs (DNA/RNAseq, proteomics, immune sequencing/antibody repertoire analysis, metabolomics, methylomics).

This AI Uses a Scan of Your Retina to Predict Your Risk of Heart Disease

They then used QUARTZ to analyze retinal images from 7,411 more people, these aged 48 to 92, and combined this data with information about their health history (such as smoking, statin use, and previous heart attacks) to predict their risk of heart disease. Participants’ health was tracked for seven to nine years, and their outcomes were compared to Framingham risk score (FRS) predictions.

A common tool for estimating heart disease risk, the FRS looks at age, gender, total cholesterol, high density lipoprotein cholesterol, smoking habits, and systolic blood pressure to estimate the probability someone will develop heart disease within a given span of time, usually 10 to 30 years.

The QUARTZ team compared their data to 10-year FRS predictions and said the algorithm’s accuracy was on par with that of the conventional tool.

Tesla reportedly places massive order of next-gen self-driving chips with TSMC

Those who follow Elon closely will notice that he is always trying to save a buck. For example, the AI chips used in current Teslas were made by Samsung instead of TSMC to save money. (TSMC makes the best chips but they are also the most expensive.) Another example of Elon saving money would be the mass layoffs at Twitter.

Well, it looks like Elon is really trying to get Full Self-Driving working and has decided that the next generation of FSD chips will be made by TSMC. He placed an order that is so big that Tesla will be the 7th largest customer of TSMC next year. He is going for 4/5 nm chips compared to the 14 nm chips he is using today.


Tesla has reportedly placed a massive order of chips for its next-gen Full Self-Driving (FSD) computer with Taiwan’s TSMC. The order is so large that it might make Tesla one of TSMC’s biggest customers.

Back in 2016, Tesla started building a team of chip architects led by legendary chip designer Jim Keller to develop its own silicon.

The goal was to design a super powerful and efficient chip to achieve self-driving in consumer vehicles without additional hardware like in custom-built autonomous vehicles operated by Waymo and Cruise.

Scientists Created an Artificial Neuron That Actually Retains Electronic Memories

The human brain is incredible.

Despite consuming the equivalent of just two bananas per day, this doesn’t stop it from executing unconscionably complex tasks with impressive efficiency. But a team of researchers has designed a way to build a prototype of an artificial neuron made of unbelievably thin graphene slits housing a single layer of water molecules, according to a new study published in the journal Science.