Skip to content

Commit 026939b

Browse files
authored
Update 2025-6-20-ai-review.md
1 parent f25d784 commit 026939b

File tree

1 file changed

+26
-40
lines changed

1 file changed

+26
-40
lines changed

blog/2025-6-20-ai-review.md

Lines changed: 26 additions & 40 deletions
Original file line numberDiff line numberDiff line change
@@ -8,54 +8,53 @@ tags: [neurosynth]
88

99
We’re excited to share the release of a significant new innovation in Neurosynth Compose: **AI-Assisted Curation**. This new feature aims to simplify and accelerate custom neuroimaging meta-analyses, making it easier than ever to curate a set of studies for inclusion into a quantitative meta-analysis.
1010

11-
As Neurosynth Compose users well know, even with the help of a streamlined user interface, manually reviewing studies for inclusion into a meta-analysis is very time-consuming. In fact, a single systematic review can take hundreds of hours, which severely limits the kinds of research questions we can explore. For this very reason, the original Neurosynth platform, which used text-mining to group neuroimaging studies by keywords or topics, was a big step forward. However, the low hanging fruit from these methods have largely been picked. Additionally, crucial details about how a study was done or who participated (like sample size, age, patient group, and experimental specifics) are hard to extract automatically using simple methods such as text frequency or heuristics because they're often described inconsistently or not mentioned frequently in the text.
11+
Traditionally, reviewing studies for inclusion in a meta-analysis takes hundreds of hours, even with a user-friendly interface. The original Neurosynth platform helped by using text mining to group studies by keywords, but extracting crucial details—like participant demographics or task design—remained a challenge because such information is inconsistently reported.
1212

13-
Here, we aimed to leverage innovations in Zero-Shot Learning using Large Language Models (LLMs) and pair this with our platform for custom meta-analysis (Neurosynth Compose) to **make precise, automated meta-analysis of neuroimaging literature a reality.**
13+
With Large Language Models (LLMs) and zero-shot learning, we can now automatically extract structured study information directly from full-text articles. This makes **precise, automated meta-analysis of neuroimaging literature a reality.**
1414

15-
## **Large Language Models (LLMs) for Neuroscientific Information Extraction**
15+
## **LLMs for Neuroscientific Information Extraction**
1616

17-
At the heart of this effort is the recent ability for LLMs to understand language with little specialized training. Historically, developing AI models for scientific information extraction was difficult due to the large number of annotated examples required for training. For low-annotation fields like neuroimaging, that largely meant that state-of-the-art biomedical NLP models were out of reach.
17+
Historically, developing AI models for information extraction required a large number of annotated examples required for training, making it impossible to apply in low-data fields like neuroimaging. However, modern LLMs can interpret scientific text without extensive domain-specific training. It is now possible to automatically extract information from articles, even in areas where there are very few existing human-labeled examples. This approach, called "**zero-shot learning**", means modern LLMs can extract information without task-specific training.
1818

19-
However, recent advancements in **LLM transfer learning** have made it possible to automatically extract information from articles, even in areas where there are very few existing human-labeled examples. Newer LLMs that are trained on vast amounts of general text can be prompted to learn new information with no training data. This approach, called "**zero-shot learning**", means SOTA LLMs can extract information even if they haven't seen that exact type of task before.
20-
21-
Here, we use these models to **extract specific details directly from the full text of over 30,000 neuroimaging studies indexed in the NeuroStore database**. By carefully guiding the AI to focus on verifiable facts within the paper, we can reduce the chance of hallucinations, and allow us to verify how accurately key details are extracted. Using this information, we can build a large, structured collection of neuroscientific facts—including participant demographics, study designs, task information, and more—which is then seamlessly presented to you during the curation stage.
19+
By prompting LLMs carefully, we can extract verifiable details—like sample size, diagnosis, age range, tasks, and modalities—from more than 30,000 studies in the NeuroStore database. These details are displayed during curation, making it easier to compare studies and decide which to include.
2220

2321
![Figure 1](img/ai_extraction_workflow.png)
2422
Figure 1\. High-level overview of Zero Shot Information Extraction using LLM prompting
2523

26-
## **A Revamped Curation Experience**
24+
## **A New Curation Experience**
25+
26+
This information is presented in two key places within the revamped curation user interface: a **concise table view**, allowing for easy comparison across papers, and more **detailed individual study pages** where you can delve deeper into the extracted specifics for a single paper.
2727

28-
This information is presented in two key places within the revamped curation user interface: a **concise table view**, allowing for easy comparison across papers, and more **detailed individual study pages** where you can delve deeper into the extracted specifics for a single paper. This dramatically facilitates systematic literature review and helps you efficiently screen studies for eligibility into your meta-analysis research question, all within a fully web-based, PRISMA-compliant workflow (Note that for the PRISMA-workflow AI features are available in the Screening and Eligibility phases).
28+
This supports a faster, PRISMA-compliant screening and eligibility process, all within a web-based interface.
2929

3030
![Figure 2](img/ai_table_view.png)
3131
Figure 2\. Table view showing AI-extracted information (Task Name, Group names, Diagnosis), across three studies
3232

33-
By clicking on a row in the table view, you can see study-level meta-data and extracted features in more detail:
33+
By clicking on a row in the table view, you can see detailed study-level extracted information:
3434

3535
![Figure 3](img/ai_study_view.png)
3636
Figure 3\. Detailed study-evel AI-extracted information, showing Participant Demographics.
3737

38-
## **Iterative Approach to Validation and Development**
39-
40-
Our approach to information extraction is specifically focused on study details relevant for neuroimaging meta-analysis. We have developed specific extraction schemas that capture the nuanced details crucial for meta-analysis in this field. For each set of guidelines, a sample of studies is manually reviewed and tagged, and the automated extractions are checked for accuracy against these manual tags, both by numbers and by human review. This thorough process makes sure that when new extraction features are introduced to the platform, a reasonable level of accuracy can be established. In contrast with domain-general automated literature review platform and deep review platforms (e.g Elict, Perplexity, Google Notebook LM), the specific extraction schemas have been validated and aligned with expert-guided knowledge representations.
38+
## **Validated, Iterative Development**
39+
Unlike general-purpose review tools (e.g., Elicit, Perplexity, Google Notebook LM), our approach is grounded in *domain expert neuroimaging guidelines*. First, we developed expert-driven extraction schemas for neuroimaging. Next, each schema is validated by comparing AI outputs with manual annotations (treated as the gold standard), ensuring accuracy before release.
4140

4241
![Figure 4](img/iterative_workflow.png)
4342

4443
Figure 4\. Iterative annotation and prompt development workflow.
4544

46-
**This effort is an ongoing process that is open to community feedback.** The goal is to continuously refine and check our extraction guidelines for neuroimaging-specific study information that help researchers find and screen studies for inclusion into meta-analysis. This extracted data can then be used with the rest of our existing comprehensive meta-analysis ecosystem (i.e. Neurosynth Compose & NiMARE), to perform detailed meta-analyses.
45+
**This ongoing effort is open to community feedback.** The goal is to continuously refine and check our extraction guidelines for neuroimaging-specific study information that help researchers find and screen studies for inclusion into meta-analysis.
4746

48-
All of the studies annotated for Neurosynth Compose are sourced from PubMed Central using *pubget* and annotated using *labelbuddy*— a set of tools our group recently introduced by our group for literature mining (Dockes et al., 2024). All of the annotations we have generated are openly accessible under the [labelbuddy annotations GitHub repository](https://github.com/litmining/labelbuddy-annotations/)
47+
Annotated studies are sourced from PubMed Central using *pubget* and annotated using *labelbuddy*— a set of tools recently introduced by our group for literature mining (Dockes et al., 2024).
4948

50-
The extraction pipelines that are validated and iteratively developed using these annotations (and put into production on Neurosynth Compose), are also [openly available](https://github.com/neurostuff/neurostore-text-extraction/tree/main/ns\_extract/pipelines).
49+
Annotations are openly accessible in the [labelbuddy annotations GitHub repository](https://github.com/litmining/labelbuddy-annotations/), and [extraction pipelines](https://github.com/neurostuff/neurostore-text-extraction/tree/main/ns\_extract/pipelines) are also made available for transparency.
5150

5251
## **Initial Extraction Schemas**
5352

54-
At launch, we have extracted two schemas across the articles indexed by NeuroStore: *participant demographics* and *experimental details.* To begin, these schemas were extracted from the full text of articles using GPT-4— a model whos performance has already been established in previous internal validations.
53+
At launch, we support two schemas: *participant demographics* and *experimental details.*, extracted from the full text of articles using GPT-4—
5554

56-
### **Participant Demographics**
55+
### 1. **Participant Demographics**
5756

58-
Participant demographics were extracted for each experimental group in the study. LLM models were instructed to focus on values that were *explicitly* mentioned in the text
57+
Extracts demopgrahic details separately for each participant group:
5958

6059
| Field | Description |
6160
| ----- | ----- |
@@ -71,24 +70,18 @@ Participant demographics were extracted for each experimental group in the study
7170
| `age_maximum` | Highest age reported or upper bound of range. |
7271
| `age_median` | Median age, only if explicitly provided. |
7372

74-
**Preliminary Validation.**
75-
- We annotated over 220 articles for participant demographics.
76-
- We observed a high level of accuracy for most fields, notably for participant *count* (\<0.15 Mean Percentage Error).
77-
- In our annotated sample, we identified 100 individual participant groups with *diagnosis* labels (e.g. “schizophrenia”). Using BERTScore to quantitatively compare the extracted and annotated diagnoses, the best performing models achieved >0.8 F1-score, indicating moderate to high accuracy. (higher scores are better).
78-
- Qualitative analysis confirmed that LLMs are increasingly adept at capturing specific diagnostic information (e.g., "Autism Spectrum Disorder", "phobic prone", "eating disorders prone") and associating it correctly with relevant demographic data, even if the specific form differed from the manual annotation.
79-
80-
### **Experimental Details**
73+
Validation on 220 articles showed high accuracy, especially for participant counts (<15% error) and diagnosis classification (>0.8 F1-score using BERTScore, on subset of 100 studies with annotated *diagnosis*). Qualitative analysis confirmed that LLMs are increasingly adept at capturing specific diagnostic information (e.g., "Autism Spectrum Disorder", "phobic prone", "eating disorders prone") and associating them correctly with relevant demographic data, even if the specific form differed from the manual annotation.
8174

82-
The goal of this schema was to extract key details of the overall study, and the individual fMRI Tasks that were used.
75+
### 2. **Experimental Details**
8376

84-
For the overall study, the following was extracted:
77+
Captures study design and task information:
8578

8679
| Field | Description |
8780
| ----- | ----- |
8881
| `Modality` | Imaging modalities used (e.g., "fMRI-BOLD", "MEG", "PET"). |
8982
| `StudyObjective` | Brief summary of the study’s main research question or goal. |
9083

91-
For each fMRI task presented within the study, the following was extracted:
84+
For each fMRI task presented within the study, the following was also extracted:
9285

9386
| Field | Description |
9487
| ----- | ----- |
@@ -104,19 +97,12 @@ For each fMRI task presented within the study, the following was extracted:
10497
| `TaskDesign` | Task design type(s): Blocked, EventRelated, Mixed, or Other. |
10598
| `TaskDuration` | Total task duration (e.g., "10 minutes" or "600 seconds"). |
10699

107-
**Preliminary Validation**
108-
109-
We annotated 104 papers to validate study/task information, with the majority of these papers sourced from the NeuroVault collection.
110-
111-
* **Modality & RestingState:** Modality and Resting State fields demonstrated very high accuracy. For instance, with 94% accuracy for these fields using GPT 4\.
112-
* **TaskName and TaskDescription Accuracy:** TaskName is accurate for studies with a clearly defined task name (64/104 of studies), with a score of 0.9 (1-Levenshtein distance). For studies without a clearly defined task name, qualitative review of examples showed that the models often provided a coherent and plausible description of the task based on the provided context, even if it wasn't a direct match to a predefined label.
100+
Validation on 104 studies found high accuracy for modality and resting-state fields (94%) and strong performance for task information (1-Levenshtein distance of 0.9), particularly when task names were clearly reported in the original sudies (64% of studies). For studies without a clearly defined task name, qualitative review indicated GPT often provided a coherent and plausible description of the task based on the provided context.
113101

114102
This preliminary validation is just a first step. Stay tuned for a more comprehensive evaluation of AI-extracted neuroimaging features\!
115103

116-
# **Try AI-Assisted Curation now\!**
117-
118-
These new features are available for you to try out now on [compose.neurosynth.org](https://compose.neurosynth.org).
104+
# **Get Started**
119105

120-
Remember, this is an ongoing, iterative effort, and we have many more features planned on the horizon to increase the level of accuracy and transparency of these AI-extracted features. Feel free to suggest features that would be useful for us to extract.
106+
You can try AI-Assisted Curation now at [compose.neurosynth.org](https://compose.neurosynth.org).
121107

122-
We're always looking for ways to improve Neurosynth Compose, and your feedback is incredibly valuable\! If you have thoughts, questions, or suggestions about AI-Assisted Curation or any other feature, please don't hesitate to reach out. You can also engage with us and the broader community on [NeuroStars](https://neurostars.org/tag/neurosynth-compose), our discussion forum.
108+
This is an ongoing project: we’ll keep expanding schemas and refining accuracy. We welcome your feedback and ideas—join the conversation on [NeuroStars](https://neurostars.org/tag/neurosynth-compose), our discussion forum.

0 commit comments

Comments
 (0)