Toggle light / dark theme

Researchers, from biochemists to material scientists, have long relied on the rich variety of organic molecules to solve pressing challenges. Some molecules may be useful in treating diseases, others for lighting our digital displays, still others for pigments, paints, and plastics. The unique properties of each molecule are determined by its structure—that is, by the connectivity of its constituent atoms. Once a promising structure is identified, there remains the difficult task of making the targeted molecule through a sequence of chemical reactions. But which ones?

Organic chemists generally work backwards from the target molecule to the starting materials using a process called retrosynthetic analysis. During this process, the chemist faces a series of complex and inter-related decisions. For instance, of the tens of thousands of different chemical reactions, which one should you choose to create the target molecule? Once that decision is made, you may find yourself with multiple reactant molecules needed for the reaction. If these molecules are not available to purchase, then how do you select the appropriate reactions to produce them? Intelligently choosing what to do at each step of this process is critical in navigating the huge number of possible paths.

Researchers at Columbia Engineering have developed a based on reinforcement learning that trains a to correctly select the “best” reaction at each step of the retrosynthetic process. This form of AI provides a framework for researchers to design chemical syntheses that optimize user specified objectives such synthesis cost, safety, and sustainability. The new approach, published May 31 by ACS Central Science, is more successful (by ~60%) than existing strategies for solving this challenging search problem.

Read more

How might future changes in the structure of business and the nature of work impact the environment?

While governments around the world are wrestling with the potential for massive on-rushing technological disruption of work and the jobs market, few are extending the telescope to explore what the knock-on impacts might be for the planet. Here we explore some dimensions of the issue.

Although replacing humans with robots has a dystopian flavor, what, if any positives are there from successive waves of artificial intelligence (AI) and other exponentially developing technologies displacing jobs ranging from banker to construction worker? Clearly, the number of people working and the implications for commuting, conduct of their role and their resulting income-related domestic lifestyle all have a direct bearing on their consumption of resources and emissions footprint. However, while everyone wants to know the impact of smart automation, the reality is that we are all clueless as to the outcome over the next twenty years, as this fourth industrial revolution has only just started.

There is a dramatic variation in views on the extent to

Luminous Computing, a one-year-old startup, is aiming to build a photonics chip that will handle workloads needed for AI at the speed of light. It’s a moonshot and yet, the young company already has a number of high-profile investors willing to bet on the prospect.

The company has raised $9 million in a seed round led by Bill Gates, NEO’s Ali Partovi and Luke Nosek and Steve Oskoui of Gigafund.

The round also attracted other new investors, including Travis Kalanick’s fund 10100, BoxGroup, Uber CEO Dara Khosrowshahi, and Emil Michael as well as pre-seed investors Class 5 Global, Joshua Browder, Ozmen Ventures, Schox Investments and Third Kind Venture Capital.

Read more

Machines are mastering vision and language, but one sense they’re lagging behind on is touch. Now researchers have created a sensor-laden glove for just $10 and recorded the most comprehensive tactile dataset to date, which can be used to train machine learning algorithms to feel the world around them.

Dexterity would be an incredibly useful skill for robots to master, opening up new applications everywhere from hospitals to our homes. And they’ve been coming along in leaps and strides in their ability to manipulate objects, OpenAI’s cube juggling robotic hand being a particularly impressive example.

So far, though, they’ve had one hand tied behind their backs. Most approaches have relied on using either visual data or demonstrations to show machines how they should grasp objects. But if you look at how humans learn to manipulate objects, you realize that’s just one part of the puzzle.

Read more