Toggle light / dark theme

The latest “machine scientist” algorithms can take in data on dark matter, dividing cells, turbulence, and other situations too complicated for humans to understand and provide an equation capturing the essence of what’s going on.


Despite rediscovering Kepler’s third law and other textbook classics, BACON remained something of a curiosity in an era of limited computing power. Researchers still had to analyze most data sets by hand, or eventually with Excel-like software that found the best fit for a simple data set when given a specific class of equation. The notion that an algorithm could find the correct model for describing any data set lay dormant until 2009, when Lipson and Michael Schmidt, roboticists then at Cornell University, developed an algorithm called Eureqa.

Their main goal had been to build a machine that could boil down expansive data sets with column after column of variables to an equation involving the few variables that actually matter. “The equation might end up having four variables, but you don’t know in advance which ones,” Lipson said. “You throw at it everything and the kitchen sink. Maybe the weather is important. Maybe the number of dentists per square mile is important.”

One persistent hurdle to wrangling numerous variables has been finding an efficient way to guess new equations over and over. Researchers say you also need the flexibility to try out (and recover from) potential dead ends. When the algorithm can jump from a line to a parabola, or add a sinusoidal ripple, its ability to hit as many data points as possible might get worse before it gets better. To overcome this and other challenges, in 1992 the computer scientist John Koza proposed “genetic algorithms,” which introduce random “mutations” into equations and test the mutant equations against the data. Over many trials, initially useless features either evolve potent functionality or wither away.

Machines are learning things fast and replacing humans at a faster rate than ever before. Fresh development in this direction is a robot that can taste food. And not only it can taste the food, it can do so while making the dish it is preparing! This further leads to the robot having the ability to recognise taste of the food in various stages of chewing when a human eats the food.

The robot chef was made by Mark Oleynik, a Russian mathematician and computer scientist. Researchers at the Cambride University trained the robot to ‘taste’ the food as it cooks it.

The robot had already been trained to cook egg omelets. The researchers at Cambridge University added a sensor to the robot which can recognise different levels of saltiness.

Jack in the Box has become the latest American food chain to experiment with automation, as it seeks to handle staffing challenges and improve the efficiency of its service.

Jack in the Box is one of the largest quick service restaurant chains in America, with more than 2,200 branches. With continued staffing challenges impacting its operating hours and costs, Jack in the Box saw a need to revamp its technology and establish new systems – particularly in the back-of-house – that improve restaurant-level economics and alleviate the pain points of working in a high-volume commercial kitchen.

Reimagining A Healthier Future for All — Dr. Pat Verduin PhD, Chief Technology Officer, Colgate, discussing the microbiome, skin and oral care, and healthy aging from a CPG perspective.


Dr. Patricia Verduin, PhD, (https://www.colgatepalmolive.com/en-us/snippet/2021/circle-c…ia-verduin) is Chief Technology Officer for the Colgate-Palmolive Company where she provides leadership for product innovation, clinical science and long-term research and development across their Global Technology Centers’ Research & Development pipeline.

Dr. Verduin joined Colgate Palmolive in 2007 as Vice President, Global R&D. Previously she served as Vice President, Scientific Affairs, for the Grocery Manufacturers Association, and from 2000 to 2006, she held the position of Vice President, Research & Development, at ConAgra Foods.

From search engines to voice assistants, computers are getting better at understanding what we mean. That’s thanks to language-processing programs that make sense of a staggering number of words, without ever being told explicitly what those words mean. Such programs infer meaning instead through statistics—and a new study reveals that this computational approach can assign many kinds of information to a single word, just like the human brain.

The study, published April 14 in the journal Nature Human Behavior, was co-led by Gabriel Grand, a graduate student in and computer science who is affiliated with MIT’s Computer Science and Artificial Intelligence Laboratory, and Idan Blank Ph.D. ‘16, an assistant professor at the University of California at Los Angeles. The work was supervised by McGovern Institute for Brain Research investigator Ev Fedorenko, a cognitive neuroscientist who studies how the uses and understands language, and Francisco Pereira at the National Institute of Mental Health. Fedorenko says the rich knowledge her team was able to find within computational language models demonstrates just how much can be learned about the world through language alone.

The research team began its analysis of statistics-based language processing models in 2015, when the approach was new. Such models derive meaning by analyzing how often pairs of co-occur in texts and using those relationships to assess the similarities of words’ meanings. For example, such a program might conclude that “bread” and “apple” are more similar to one another than they are to “notebook,” because “bread” and “apple” are often found in proximity to words like “eat” or “snack,” whereas “notebook” is not.

A growing group of startups and established logistics firms have created a multi-billion-dollar industry applying artificial intelligence and other cutting-edge… See more.


LONDON, May 3 (Reuters) — Over the last two years a series of unexpected events has scrambled global supply chains. Coronavirus, war in Ukraine, Brexit and a container ship wedged in the Suez Canal have combined to delay deliveries of everything from bicycles to pet food.

In response, a growing group of startups and established logistics firms has created a multi-billion dollar industry applying the latest technology to help businesses minimize the disruption.

Interos Inc, Fero Labs, KlearNow Corp and others are using artificial intelligence and other cutting-edge tools so manufacturers and their customers can react more swiftly to supplier snarl-ups, monitor raw material availability and get through the bureaucratic thicket of cross-border trade.

Join us on Patreon!
https://www.patreon.com/MichaelLustgartenPhD

Papers referenced in the video:
Life-Span Extension in Mice by Preweaning Food Restriction and by Methionine Restriction in Middle Age.
https://pubmed.ncbi.nlm.nih.gov/19414512/

Low methionine ingestion by rats extends life span.
https://pubmed.ncbi.nlm.nih.gov/8429371/

Fasting glucose level and all-cause or cause-specific mortality in Korean adults: a nationwide cohort study.