Toggle light / dark theme

FDA Okays AI-Based Prostate Cancer Risk-Stratification Tool

The US FDA has granted De Novo marketing authorization for ArteraAI Prostate (Artera, Los Altos, California), a novel artificial intelligence (AI)-powered risk-stratification tool for patients with nonmetastatic prostate cancer.

The authorization means the digital pathology software tool is recognized as an FDA-regulated Software as a Medical Device, Artera explained in a press release.

De Novo authorization provides a marketing pathway to classify low-or moderate-risk novel medical devices. The De Novo authorization for this specific test establishes a new product code category for future AI-powered digital pathology risk-stratification tools and enables implementation at the point of diagnosis at qualified US pathology labs, the company said.

The test analyzes digital pathology images from patients’ biopsy slides to predict long-term outcomes, such as 10-year risk for metastasis and mortality. This can help direct treatment decisions.


The FDA’s De Novo authorization for the tool establishes a new product code category for future AI-powered digital pathology risk-stratification tools.

ChatGPT in Your Clinic: Who’s the Expert Now

Patients arriving at appointments with researched information is not new, but artificial intelligence (AI) tools such as ChatGPT are changing the dynamics.

Their confident presentation can leave physicians feeling that their expertise is challenged. Kumara Raja Sundar, MD, a family medicine physician at Kaiser Permanente Burien Medical Center in Burien, Washington, highlighted this trend in a recent article published in JAMA.

A patient visited Sundar’s clinic reporting dizziness and described her symptoms with unusual precision: “It’s not vertigo, but more like a presyncope feeling.” She then suggested that the tilt table test might be useful for diagnosis.

Occasionally, patient questions reveal subtle familiarity with medical jargon. This may indicate that they either have relevant training or have studied the subject extensively.

(Artificial Intelligence is the science of making machines do things that would require intelligence if done by men — Marvin Minsky. Google helps you gain information with a search engine. AI helps you gain information through algorithms. It is the same thing. However people profit from ignorance).


Patients are showing up with ChatGPT-generated diagnoses, challenging physicians to balance empathy, evidence, and authority in the exam room.

Tesla Kills Dojo for AI6! Here’s Why

Questions to inspire discussion.

🚗 Q: How will AI6 be used in Tesla vehicles? A: AI6 will be used for FSD inference, with two chips in every car, enabling advanced autonomous driving capabilities.

🤖 Q: What role will AI6 play in Optimus? A: AI6 will enable on-device learning and reinforced learning in Optimus, enhancing its AI capabilities.

🔋 Q: Will AI6 be used in other Tesla products? A: AI6 will be integrated into every edge device produced by Tesla, including Tesla Semi, Mega Pack, and security cameras.

Technical Specifications.

💻 Q: What is the architecture of AI6? A: AI6 will use a cluster model of individual chips with a software layer on top, similar to Dojo 3 for training.

Default Layout (content area only)

A research team headed by the University of Zurich has developed a powerful new method to precisely edit DNA by combining cutting-edge genetic engineering with artificial intelligence. This technique opens the door to more accurate modeling of human diseases and lays the groundwork for next-generation gene therapies.

Precise and targeted DNA editing by small point mutations as well as the integration of whole genes via CRISPR/Cas technology has great potential for applications in biotechnology and gene therapy. However, it is very important that the so-called gene scissors do not cause any unintended genetic changes, but maintain genomic integrity to avoid unintended side effects. Normally, double-stranded breaks in the DNA molecule are accurately repaired in humans and other organisms. But occasionally, this DNA end joining repair results in genetic errors.

Gene editing with greatly improved precision Now, scientists from the University of Zurich (UZH), Ghent University in Belgium and the ETH Zurich have developed a new method which greatly improves the precision of genome editing. Using artificial intelligence (AI), the tool called Pythia predicts how cells repair their DNA after it is cut by gene editing tools such as CRISPR/Cas9. “Our team developed tiny DNA repair templates, which act like molecular glue and guide the cell to make precise genetic changes,” says lead author Thomas Naert, who pioneered the technology at UZH and is currently a postdoc at Ghent University.


Page description.

This “smart coach” helps LLMs switch between text and code

Large language models (LLMs) excel at using textual reasoning to understand the context of a document and provide a logical answer about its contents. But these same LLMs often struggle to correctly answer even the simplest math problems.

Textual reasoning is usually a less-than-ideal way to deliberate over computational or algorithmic tasks. While some LLMs can generate code like Python to handle symbolic queries, the models don’t always know when to use code, or what kind of code would work best.

LLMs, it seems, may need a coach to steer them toward the best technique.

Enter CodeSteer, a smart assistant developed by MIT researchers that guides an LLM to switch between code and text generation until it correctly answers a query. (Strangely like a text editor “CodeSteer”🤔)


CodeSteer is a smart assistant from MIT that automatically guides large language models to switch between generating text and code, and to refine its response, until it answers a query correctly.

Allie, an AI chess bot, learns to play like humans from 91 million Lichess games

Yiming Zhang didn’t grow up playing chess. Like many other people, the Carnegie Mellon University Ph.D. student discovered the Netflix series “The Queen’s Gambit” during the pandemic and began playing online. However, he quickly realized how unnatural it felt playing against chess bots.

“After I learned the rules, I was in the bottom 10%, maybe 20% of players online,” said Zhang, who is part of the Language Technologies Institute (LTI) in CMU’s School of Computer Science. “For beginners, it’s not interesting or instructive to play against chess bots because the moves they make are often bizarre and incomprehensible to humans.”

Zhang’s frustration led him to develop Allie, a chess bot powered by that demonstrates the benefits of AI tools that think like humans. He believes training future AI systems to ponder and deliberate on could create better agents for use in therapy, education and medicine.

Quantum precision reached in modeling molten salt behavior

Scientists have developed a new machine learning approach that accurately predicted critical and difficult-to-compute properties of molten salts, materials with diverse nuclear energy applications.

In a Chemical Science article, Oak Ridge National Laboratory researchers demonstrated the ability to rapidly model salts in liquid and solid state with quantum chemical accuracy. Specifically, they looked at thermodynamic properties, which control how molten salts function in high-temperature applications. These applications include dissolving nuclear fuels and improving reliability of long-term reactor operations. The AI-enabled approach was made possible by ORNL’s supercomputer Summit.

“The exciting part is the simplicity of the approach,” said ORNL’s Luke Gibson. “In fewer steps than existing approaches, machine learning gets us to higher accuracy at a faster rate.”

AI chatbots can be easily manipulated to make us share more personal data

Millions of people chat with AI tools every day, trading small talk for quick answers or support. A new study presented at the 34th USENIX Security Symposium shows how easily those friendly agents can be tuned to make you reveal far more than you planned.

The researchers report that malicious chatbots can push users to disclose up to 12.5 times more personal details than standard ones. The most effective tricks leaned on reciprocity and reassurance, not blunt questions about your life.


New research shows manipulative AI chatbots can make you reveal much more personal information than neutral ones.

/* */