Toggle light / dark theme

Would you want to know if you’re at risk of Alzheimer’s disease, for example?


The integration of sequencing into health care doesn’t fit very well in the model of how medicine is practiced today, but is well aligned with the future vision of health care that so many of us have — a vision that focuses upon prediction and prevention.

We imagine that personal genome sequencing could play a central role in bringing about a more personalized and participatory form of medicine — including a health care system where patients have more knowledge of their own risks and diagnoses and are empowered to act upon that information.

With that in mind, more of us are asking this question: Rather than focusing only on people with a suspected or diagnosed genetic disease, why not also use genome sequencing to help seemingly healthy people screen for all sorts of conditions, even diseases for which they have no known family history?

China has said that some vaccines for the novel coronavirus could be in clinical use next month as the number of global coronavirus cases soared past 100,000.

The country’s scientists are striving to develop immunisation products with five technologies simultaneously, according to officials.

‘We estimate that in April — in line with country’s relevant law and regulations – there is hope that some of the vaccines can enter the stage of clinical or emergency use,’ said Zheng Zhongwei, director of the Technical Development and Research Centre of China’s National Health Commission.

My editorial from today’s (3/18/19) Financial Times:

Far sooner than most people realise, the genetics revolution will transform the world within and around us. Although we think about genetic technologies primarily in the context of healthcare, these tools are set to change the way we make babies, the nature of the babies we make and, ultimately, our evolutionary trajectory as a species — and we are not remotely ready for what’s coming. Yet we must be, to optimise the benefits and minimise the potential harms of genetic technologies.

Scientists are now able to manipulate biology to a previously unimaginable degree. In the past year, we’ve seen two female mice having their own babies, dramatic increases in the precision of gene-editing tools, and the birth in China of the first gene-edited humans. As this science advances exponentially, however, the regulations guiding how it should best be used are struggling to keep up. If the applications race forward without appropriate guard rails, the danger increases that more scientists like He Jiankui, the Chinese biophysicist who genetically altered two girls, will put people’s health at risk. But if the regulations are hastily written before the issues are clear, are too strong or are not flexible enough, many people who would otherwise have benefited from applied genetic technologies will be condemned to unnecessary suffering or even death.

In a study of epilepsy patients, researchers at the National Institutes of Health monitored the electrical activity of thousands of individual brain cells, called neurons, as patients took memory tests. They found that the firing patterns of the cells that occurred when patients learned a word pair were replayed fractions of a second before they successfully remembered the pair. The study was part of an NIH Clinical Center trial for patients with drug-resistant epilepsy whose seizures cannot be controlled with drugs.

“Memory plays a crucial role in our lives. Just as are recorded as grooves on a record, it appears that our brains store memories in that can be replayed over and over again,” said Kareem Zaghloul, M.D., Ph.D., a neurosurgeon-researcher at the NIH’s National Institute of Neurological Disorders and Stroke (NINDS) and senior author of the study published in Science.

Dr. Zaghloul’s team has been recording electrical currents of drug-resistant epilepsy patients temporarily living with surgically implanted electrodes designed to monitor in the hopes of identifying the source of a patient’s seizures. This period also provides an opportunity to study neural activity during memory. In this study, his team examined the activity used to store memories of our past experiences, which scientists call episodic memories.

Powerful antibiotics and widespread sanitation practices have expanded lifespans across the industrialized world. But they have also come at a cost. Our microbiomes, or the trillions of microbes collectively working in our bodies to help regulate our immune system and food digestion, have lost much of its health-promoting bacteria because of our modern lifestyles and sanitation practices.

Scientists across the world are now looking to the planet’s few remaining pre-industrialized societies to see what industrialized guts have lost–and in doing so, could fundamentally change the way scientists think about germs. Thomas Morton heads to the Central African Republic to see this emerging field of microbiome science.

Check out VICE News for more: http://vicenews.com
Follow VICE News here:
Facebook: https://www.facebook.com/vicenews
Twitter: https://twitter.com/vicenews
Tumblr: http://vicenews.tumblr.com/
Instagram: http://instagram.com/vicenews
More videos from the VICE network: https://www.fb.com/vicevideo
#VICEonHBO

For the first time, doctors have attempted to cure blindness by gene-hacking a patient with CRISPR technology.

A team from Oregon Health & Science Institute injected three droplets of fluid that delivered the CRISPR DNA fragments directly into a patient’s eyeball, The Associated Press reports, in hopes that it will reverse a rare genetic condition called Leber congenital amaurosis, which causes blindness early in childhood.

“We literally have the potential to take people who are essentially blind and make them see,” Charles Albright, chief scientific officer of Editas Medicine, told the AP.

World health officials said Tuesday the mortality rate for COVID-19 is 3.4% globally, higher than previous estimates of about 2%.

“Globally, about 3.4% of reported COVID-19 cases have died,” WHO Director-General Tedros Adhanom Ghebreyesus said during a press briefing at the agency’s headquarters in Geneva. In comparison, seasonal flu generally kills far fewer than 1% of those infected, he said.

For the first time, doctors have attempted to cure blindness by gene-hacking a patient with CRISPR technology.

A team from Oregon Health & Science Institute injected three droplets of fluid that delivered the CRISPR DNA fragments directly into a patient’s eyeball, The Associated Press reports, in hopes that it will reverse a rare genetic condition called Leber congenital amaurosis, which causes blindness early in childhood.

“We literally have the potential to take people who are essentially blind and make them see,” Charles Albright, chief scientific officer of Editas Medicine, told the AP. Editas is one of the biotech companies that actually developed the treatment. “We think it could open up a whole new set of medicines to go in and change your DNA.”

Another important question is the extent to which continued increases in computational capacity are economically viable. The Stanford Index reports a 300,000-fold increase in capacity since 2012. But in the same month that the Report was issued, Jerome Pesenti, Facebook’s AI head, warned that “The rate of progress is not sustainable…If you look at top experiments, each year the cost is going up 10-fold. Right now, an experiment might be in seven figures but it’s not going to go to nine or 10 figures, it’s not possible, nobody can afford that.”

AI has feasted on low-hanging fruit, like search engines and board games. Now comes the hard part — distinguishing causal relationships from coincidences, making high-level decisions in the face of unfamiliar ambiguity, and matching the wisdom and commonsense that humans acquire by living in the real world. These are the capabilities that are needed in complex applications such as driverless vehicles, health care, accounting, law, and engineering.

Despite the hype, AI has had very little measurable effect on the economy. Yes, people spend a lot of time on social media and playing ultra-realistic video games. But does that boost or diminish productivity? Technology in general and AI in particular are supposed to be creating a new New Economy, where algorithms and robots do all our work for us, increasing productivity by unheard-of amounts. The reality has been the opposite. For decades, U.S. productivity grew by about 3% a year. Then, after 1970, it slowed to 1.5% a year, then 1%, now about 0.5%. Perhaps we are spending too much time on our smartphones.