Navigational technology used in self-driving cars aids brain surgery visualization

September 01, 2023

Prasad Vagdargi Johns Hopkins doctoral candidate "We believe ... augmented endoscopy is a technology whose time has now come." The team's surgical guidance method builds on an advanced computer vision technique called simultaneous localization and mapping, or SLAM, which has also been used for navigation in self-driving cars. "Think of it as a dynamic 3D map of a patient's brain that you can use to track and match deep brain deformation with preoperative imaging," Vagdargi says. The researchers have coined the term "augmented endoscopy" for this highly accurate, state-of-the-art method of live surgical visualization. "We believe our findings are very important and feel that augmented endoscopy is a technology whose time has now come," says Vagdargi.

Accessing the brain for neurosurgery involves drilling and cutting that can cause deep-brain anatomy to shift or become distorted. This can create discrepancies between pre-operative imaging and the actual state of the brain during a procedure.

Current surgical navigation systems may assist in providing live guidance, but typically use pins and clamps to hold a patient's head firmly in place, which carries the risk of complications and can prolong recovery time.

A team of researchers at the Imaging for Surgery, Therapy, and Radiology (I-STAR) Labs at Johns Hopkins University is working on a less invasive solution that doesn't require additional equipment or expose patients to the extra radiation or long scan times typically associated with live imaging. They have partnered with medical device company Medtronic and the National Institutes of Health to develop a real-time guidance system that uses an endoscope, a piece of equipment already commonly employed in neurosurgeries.

Their work is described in in IEEE Transactions on Medical Robotics and Bionics.

"Our study demonstrates the superiority of this real-time 3D navigation method over current visualization techniques," says Prasad Vagdargi, a doctoral candidate at the I-STAR Labs and primary author of the study. Vagdargi is advised by Jeffrey Siewerdsen, professor of biomedical engineering with a joint appointment in the Department of Computer Science, and Gregory Hager, professor of computer science.

Prasad Vagdargi

Johns Hopkins doctoral candidate

"We believe ... augmented endoscopy is a technology whose time has now come."

The team's surgical guidance method builds on an advanced computer vision technique called simultaneous localization and mapping, or SLAM, which has also been used for navigation in self-driving cars. After calibrating an endoscopic video feed, the team's SLAM algorithm tracks important visual details in each frame and uses those details to determine where the endoscope camera is and how it's positioned. The algorithm then transforms those details into a 3D model of the object—in this case, the inside of a patient's skull. This model is then overlaid with the real-world video feed to visualize targeted structures on-screen in real time.

"Think of it as a dynamic 3D map of a patient's brain that you can use to track and match deep brain deformation with preoperative imaging," Vagdargi says. "Combining this map with an augmented reality overlay on the endoscopic video will help surgeons to visualize targets and critical anatomy well beyond the surface of the brain."

The researchers have coined the term "augmented endoscopy" for this highly accurate, state-of-the-art method of live surgical visualization. Through a series of preclinical experiments, they demonstrated that this method is more than 16 times faster than previous computer vision techniques while still maintaining submillimeter accuracy.

The team predicts that the improved accuracy afforded by augmented endoscopy may lead to reduced complications, shorter operation times, and therefore increased surgical efficiency—not only in neurosurgeries but also in other endoscopic procedures across medical disciplines.

The research team is currently collaborating with neurosurgeons at the Johns Hopkins Hospital in a clinical study with the goal of refining and validating their method for use in real operating rooms. The researchers are also expanding their approach by applying artificial intelligence and machine learning methods to further improve the speed and precision of future iterations of their algorithm.

"We believe our findings are very important and feel that augmented endoscopy is a technology whose time has now come," says Vagdargi.

Study co-authors include colleagues in the Department of Biomedical Engineering, the Department of Computer Science, and the Department of Neurology and Neurosurgery at the Johns Hopkins Hospital, as well as industry collaborators at Medtronic.

The source of this news is from Johns Hopkins University

Popular in Research

1

May 28, 2024

Can exams be shorter and better?

2

May 29, 2024

HKGAI Debuts Cutting Edge AI Projects at InnoEx 2024 | The Hong Kong University of Science and Technology

3

May 28, 2024

Three NYU Faculty Awarded 2024 Guggenheim Fellowships

4

May 28, 2024

Campus garden initiatives can help grow the next generation of environmental change-makers

5

May 29, 2024

New molecule found to suppress bacterial antibiotic resistance evolution

Hagari: '99% of Iranian aerial threats were shot down'

6 days ago

Quiet before the storm: Is escalating Israel-Iran violence on the way? - analysis

6 days ago

Foreign holdings of US Treasuries hit record high; Japan holdings rise, data shows

6 days ago

Stormy Daniels' chilling warning to world on Donald Trump's bid to become US President

6 days ago

School suspensions and exclusions put vulnerable children at risk

6 days ago

Tackling cancer at the nanoscale

6 days ago