Response differences in brain circuits when broadcasting (top row) or receiving signals (bottom row). Left: healthy controls; Middle: unresponsive wakefulness; Right: minimally conscious
Response differences in brain circuits when broadcasting (top row) or receiving signals (bottom row). Left: healthy controls; Middle: unresponsive wakefulness; Right: minimally conscious

Spanish researchers use brain modeling to identify necessary circuits of consciousness

Currently, patients with DoC are classified into several categories (coma, unresponsive wakefulness syndrome, and minimally conscious state) which describe their overall consciousness and awareness. “The diagnosis is mainly response-based: the doctor sits down with the patient and assesses their response to stimuli” explains Jitka Annen from the University of Liège. “However, this may not correspond to their underlying brain activity - patients with high activity may still be unable to react. It’s a heterogeneous group. We wanted to go a step beyond assessment and classification based on neuroimaging and instead look at the flow of information in their brains, to find common patterns associated with consciousness.”

The researchers focused on two DoC groups - unresponsive wakefulness syndrome (previously known as the vegetative state) and the minimally conscious state. After collecting fMRI data from each patient during resting state (i.e., patients were awake but no particular task was provided), they looked at the spontaneous and model-based perturbation of brain activity captured by the blood flow, such as signals and peaks. “Based on spontaneous peaks of activity, we evaluated the personalized connectivity of each patient’s brain, which can tell us how likely a signal is to travel from one point to the other,” says Gorka Zamora-López from Pompeu Fabra. “After we constructed a patient-specific computational model of their propagation patterns, we then trigger a signal in the model and see how the brain reacts. In particular, we look for which areas are more likely to respond to a signal; which areas are more likely to propagate it. Basically, we look at whether an area acts as an influencer or influenced.”

A marked distinction arises between the unresponsive wakefulness syndrome group and the minimally conscious state group, with the former not displaying activity in identifiable circuits. “The key difference is that in patients with unresponsive wakefulness syndrome, no region of the brain seems embedded in a functional network, they all display equally low activation. Meanwhile, distinct regions and circuits pop out in the brain models of people in the minimally conscious state: the thalamic-frontotemporal region when broadcasting signals, and the posterior cortical region when receiving them,” adds Rajanikant Panda of the University of Liège. 

These findings shed new light on disorders of consciousness and could lead to a more defined understanding of the mechanisms based on brain activity rather than behavioral responses. “I believe these results can potentially guide the clinicians to better understand what is going wrong in the information exchange and thus look for ways to reactivate those circuits” concludes Zamora-López.

BIDMC researchers test AI-powered chatbot's medical diagnostic ability

Generative artificial intelligence could serve as a promising tool to assist medical professionals

In a recent experiment published in JAMA, physician-researchers at Beth Israel Deaconess Medical Center (BIDMC) tested one well-known publicly available chatbot’s ability to make accurate diagnoses in challenging medical cases. The team found that the generative AI, Chat-GPT 4, selected the correct diagnosis as its top diagnosis nearly 40 percent of the time and provided the correct diagnosis in its list of potential diagnoses in two-thirds of challenging cases.

Generative AI refers to a type of artificial intelligence that uses patterns and information it has been trained on to create new content, rather than simply processing and analyzing existing data. Some of the most well-known examples of generative AI are so-called chatbots, which use a branch of artificial intelligence called natural language processing (NLP) that allows computers to understand, interpret and generate human-like language. Generative AI chatbots are powerful tools poised to revolutionize creative industries, education, customer service, and more. However, little is known about their potential performance in the clinical setting, such as complex diagnostic reasoning.

“Recent advances in artificial intelligence have led to generative AI models that are capable of detailed text-based responses that score highly in standardized medical examinations,” said Adam Rodman, MD, MPH, co-director of the Innovations in Media and Education Delivery (iMED) Initiative at BIDMC and an instructor in medicine at Harvard Medical School. “We wanted to know if such a generative model could ‘think’ like a doctor, so we asked one to solve standardized complex diagnostic cases used for educational purposes. It did really, really well.”

To assess the chatbot’s diagnostic skills, Rodman and colleagues used clinicopathological case conferences (CPCs), a series of complex and challenging patient cases including relevant clinical and laboratory data, imaging studies, and histopathological findings published in the New England Journal of Medicine for educational purposes.

Evaluating 70 CPC cases, the artificial intelligence exactly matched the final CPC diagnosis in 27 (39 percent) of cases. In 64 percent of the cases, the final CPC diagnosis was included in the AI’s differential – a list of possible conditions that could account for a patient’s symptoms, medical history, clinical findings, and laboratory or imaging results.

“While Chatbots cannot replace the expertise and knowledge of a trained medical professional, generative AI is a promising potential adjunct to human cognition in diagnosis,” said first author Zahir Kanjee, MD, MPH, a hospitalist at BIDMC and assistant professor of medicine at Harvard Medical School. “It has the potential to help physicians make sense of complex medical data and broaden or refine our diagnostic thinking. We need more research on the optimal uses, benefits, and limits of this technology, and a lot of privacy issues need sorting out, but these are exciting findings for the future of diagnosis and patient care.”

“Our study adds to a growing body of literature demonstrating the promising capabilities of AI technology,” said co-author Byron Crowe, MD, an internal medicine physician at BIDMC and an instructor in medicine at Harvard Medical School. “Further investigation will help us better understand how these new AI models might transform health care delivery.”

This work did not receive separate funding or sponsorship. Kanjee reports royalties for books edited and membership of a paid advisory board for medical education products not related to artificial intelligence from Wolters Kluwer, as well as honoraria for CME delivered from Oakstone Publishing. Crowe reports employment by Solera Health outside the submitted work. Rodman reports no conflicts of interest.  

Assistant Professor of Biomolecular Engineering Ali Shariati. (photo by Carolyn Lagattuta)
Assistant Professor of Biomolecular Engineering Ali Shariati. (photo by Carolyn Lagattuta)

UCSC prof Shariati develops deep-learning software that detects, tracks individual cells with high performance

Cell growth and division are two of the most fundamental and essential features of life, and closely monitoring cell changes over time can give scientists key insights into the dynamics of these biological processes. Time-lapse microscopy allows scientists to detect and track cells but produces huge amounts of data that are nearly impossible to sort through manually.

Now, however, powerful data processing capabilities of modern deep learning models offer techniques to sort through so much imaging data. Assistant Professor of Biomolecular Engineering Ali Shariati and doctoral student Abolfazl Zarageri together with several student researchers in the Shariati lab have developed and released a new deep learning model called “DeepSea,” one of the only tools with the ability to segment cells, track them and detect their division to follow lineages of cells. DeepSea, which is detailed in a new paper in Cell Reports Methods, is one of the highest-accuracy tools of its kind. 

DeepSea’s model training dataset, user-friendly software, and open-source code are available for use on the DeepSea website, and Shariati and his team of researchers have already used it to make discoveries about stem cell growth and division. 

“The model is more efficient, has fewer parameters, and both segmentation and tracking are integrated into a user-friendly software,” Shariati said. “The software allows you to train the model for any cell type of interest, paving the way for future discoveries.”

Time-lapse microscopy, which captures a series of images from a microscope over time, allows researchers to monitor single cells throughout an experiment to track phenomena such as differentiation — when stem cells become a specific type of cell — or change in shape and size over time. This can allow scientists to make new biological discoveries by measuring the dynamics of cell biological phenomena at the single-cell level.

Once the scientists have gathered images, they need to carry out two main tasks: segmentation, or identifying the borders of individual cells from each other and the background; and tracking, or following a cell from one frame to the next. From that point, the researchers can further investigate characteristics such as size, shape, texture, how they move and change their shape, and more. 

Manually sorting through microscopy images is tedious, time-consuming, and ultimately a task better suited for supercomputing, which is where DeepSea comes in. This efficient deep-learning model can perform segmentation in less than a second, and track cells with 98 percent accuracy. 

Enabling the software to detect cell division was a particularly unique and challenging aspect of this project, as there are few if any other situations in which artificial intelligence and computer vision must track one object transforming into two. 

“This is a very unusual problem for object tracking,” Shariati said. “If you want to track a car or something, the car will be moving around and you can use machine learning and computer vision to follow them as they move. But for cells, all of a sudden one object becomes two, and that's a fundamentally new problem that we needed to solve, and we were able to do so.” 

DeepSea is a generalizable model, meaning it can be used to track a variety of cell types. It uses a modified version of a popular model, 2D-UNET, with significantly fewer parameters to achieve both fast speeds and high accuracy. 

“We compared our model with some of the best cell segmentation models, and ours is now showing the best results in terms of precision, and speed, especially for these cell types,” said Zarageri, an electrical and computer engineering Ph.D. student in Shariati’s lab who led the creation of the software.

The researchers trained DeepSea using a dataset of images of cells manually segmented from their backgrounds, a time-intensive process as the images are often low-contrast and the cell bodies hard to make out. To aid in this process, the team developed another software tool to help crop, label, and edit the microscopy images of cells, which is also available at DeepSeas.org.

The training dataset included images of lung, muscle, and stem cells, meaning DeepSea achieves high precision across different cell types. More cell types can be added to future versions of the model. 

The researchers used DeepSea to study the size regulation of embryonic stem cells, which are the foundation of multicellular life and can differentiate into every other cell type. They came away with the discovery that embryonic stem cells, which are known to divide unusually fast, regulate their size so that smaller cells spend a longer time growing before producing the next generation of cells. 

“We found that if an embryonic stem cell is born small, they kind of know that they are small, so they spend more time growing before they go on and divide again,” Shariati said. “We do not know why and how exactly this happens, but at least that phenomenon is there.” 

In the future, the researchers plan to apply their existing software to gather data to study spatial relationships between cells, and how the cellular features are organized in 3-D patterns to form structures. 

The researchers also aim to resolve bottlenecks they have noticed in using their deep learning models, such as the lack of labeled images of cells that are used to train the models. They plan to use a class of machine learning frameworks called Generative Adversarial Networks (GANs) to create new synthetic data and images of cells that are already annotated to cut down on the time it takes to create labels. The researchers would then have large libraries of datasets of any cell type of interest with minimal human involvement.