Toggle light / dark theme

Scientists cover robotic finger with living skin made from human cells

We’ve been hearing a lot about synthetic skins designed for robotic hands, which would give the devices more human-like qualities. Well, scientists in Japan have gone a step further, by covering a robotic finger in a self-healing skin made from live human cells.

Led by Prof. Shoji Takeuchi, a team at the University of Tokyo started by building an articulated motor-driven robotic finger, capable of bending and straightening like its human counterpart. That finger was then submerged in a cylinder filled with a solution made up of collagen and human dermal fibroblast cells – these are the main components of our skin’s connective tissues.

Due to its natural properties, that solution shrank and conformed to the contours of the finger, forming a seamless hydrogel coating. Next, the scientists added a layer of human epidermal keratinocyte cells, which constitute 90 percent of our epidermis (the outermost layer of skin). These formed a moisture-retaining/water-resistant barrier on top of the gel, and gave the finger a more natural texture.

LG AI Research’s First AI Artist, ‘Tilda’, Creates A New Sustainable Clothing Collection Made By Combining Digital Waste With Secondhand Denim and Materials

Numerous activities, including construction and demolition, mining and industrial activities, cooking and gardening, and others, generate a substantial amount of garbage. The amount of waste generated is directly proportional to consumption and production patterns.

In most cases, waste formation is the result of inefficient material utilization. Trends in the number, composition and impacts of these materials provide insight into the nation’s efficiency in using (and reusing) materials and resources. It also provides a better understanding of the effects of waste on human health and the environment.

According to surveys, 92 million tonnes of cloth are dumped as garbage each year worldwide. Estimates predict that this figure will likely exceed 130 million tonnes by 2030. When 200 tonnes of water used to make a single tonne of fabric is considered, it becomes clear that the end-to-end processes of the garment industry are severe threats to environmental initiatives.

Andrea De Souza — Eli Lilly — Leveraging Big Data & Artificial Intelligence For Unmet Medical Needs

Leveraging big data & artificial intelligence to solve unmet medical needs — andrea de souza — eli lilly & co.


Andrea De Souza, is Associate Vice President, Research Data Sciences and Engineering, at Eli Lilly & Company (https://www.lilly.com/) where over the past three years her work has focused around empowering the Lilly Research Laboratories (LRL) organization with greater computational, analytics-intense experimentation to raise the innovation of their scientists.

A former neuroscience researcher, Andrea’s portfolio career has included leadership assignments at the intersection of science, technology and business development. She has built and led informatics and scientific teams across the entire pharmaceutical value chain.

Most recently, Andrea focused on building the Pharma Artificial Intelligence market at NVIDIA. Through this experience she has traveled the world advising bio-pharmaceutical clients, academics, research institutes, and startups in the potential of machine learning and artificial intelligence across every discipline of the industry.

Prior to her role at NVIDIA, Andrea held leadership positions at the Broad Institute of Harvard and MIT, Amgen, and Roche.

Training a robot to recognize and pour water

Jeffrey DeanUnless you’re actively scrubbing the co2, that’s what happens when you recirculate air.

James FalkA carbonator?

Michael Taylor shared a link.


A horse, a zebra and artificial intelligence helped a team of Carnegie Mellon University researchers teach a robot to recognize water and pour it into a glass.

Water presents a tricky challenge for robots because it is clear. Robots have learned how to pour before, but previous techniques like heating the water and using a thermal camera or placing the glass in front of a checkerboard background don’t transition well to everyday life. An easier solution could enable servers to refill water glasses, robot pharmacists to measure and mix medicines, or robot gardeners to .

Wearable, waterproof sensors combine high sensitivity and location options

Wearable sensors—an important tool for health monitoring and for training artificial intelligence—can be waterproof or can measure more than one stimuli, but combining these factors while maintaining a high level of precision in the measurements is difficult. Researchers co-led by Huanyu “Larry” Cheng, assistant professor of engineering science and mechanics at Penn State, have created sensors that are waterproof, an important trait for exercise monitoring and for withstanding perspiration and all weather conditions; can measure temperature and motion on both small and large scales; and can be attached to distal arteries such as those located beneath the eyebrow or in a toe.

The results are available now online in the Chemical Engineering Journal ahead of publication in the journal’s September print edition.

“There are three aspects of this that are novel in combination: the underwater application, the ability to detect ultra-small vibrations and subtle motions and temperature changes, and the multiple options for sensor location, such as the eyebrow or toe,” Cheng said.

Biomimetic elastomeric robot skin has tactile sensing abilities

A team of researchers at Korea Advanced Institute of Science and Technology, working with one colleague from MIT and another from the University of Stuttgart, has developed a biomimetic elastomeric robot skin that has tactile sensing abilities. Their work has been published in the journal Science Robotics.

Roboticists continue to work on improving robot abilities and to make them more human-like. In this new effort, the researchers gave a the ability to detect such sensations as a pat, tickling, wind, or something stroking its surface. They accomplished this by partially imitating .

The new robot is multi-layered, like human skin, to allow for different functions. The top layer is made of a rubber-like polymer resembling human skin. Beneath that, the researchers added a hydrogel to imitate the human epidermis. They chose a hydrogel because it not only deforms when pressed, but jiggles when bumped. By embedding sensors to detect these reactions, the skin is able to sense things like a finger press by monitoring the pressure of the hydrogel and the direction of its movement. If something taps against it, the system senses and measures in the hydrogel to gauge what the tap felt like.

Chinese hacking group Aoqin Dragon quietly spied orgs for a decade

Žilvinas DeveikaIt’s much sooner than that. My prediction (that is almost 10 years old now) of an “early” appearance of a strong AGI is 2029. I am completely sure that it will either emerge or will already be there in 2030s.

24 Replies.

Marc O MonfilsAnd what strategies do we have in place to guarantee humanity’s continued relevance in the era of super intelligent machines?

Empathy? Never saved any tribe in the past…

Human services and interaction? For what, to keep our irrelevance engaged?… See more.

5 Replies.

View 9 more comments.

/* */