We design literature reviews across
the range of evidence needs
We are trusted leaders through publishing guidance on how to develop and communicate SLRs and advising on the optimal approach based on the evidence needed, including state-of-the-art AI assistance
Scoping
Understanding the evidence need and standards
Working within your constraints
Designing
Search strategy and methods
JCA- or PRISMA-compliant as needed
Reporting structure
Delivering
Quality-checked results
Reports that meet your needs
Collaborating
Working closely together
Managing time and budget




Data Extraction
We compared how Elicit AI performed against our
human experts when they extracted cost, humanistic
and study design data
With expert prompting, we saw strong performance
(78.0–95.0%) from the AI but a need for robust quality assurance protocols

and abstract screening workflow for supporting TLRs
with high screening burdens
Sensitivity and specificity were improved with systematic prompt refinement, reducing the number of missed relevant articles while enhancing overall accuracy,
but human oversight remained essential

We used AI to extract simple data at title/abstract screening and then ranked the studies
This allows us to adjust sensitivity and rapidly prioritize the most relevant studies, resulting in reports that include the bigger, better studies to rapidly give answers to questions and inform next steps

the best prompts for title/abstract screening
For each, we calculated how accurately the prompts
identified or excluded studies and found big differences between strategies
We are using these results to choose the best approach
based on the needs of each review





Data Extraction
We compared how Elicit AI performed against our human experts when they extracted cost, humanistic and study design data
With expert prompting, we saw strong performance (78.0–95.0%) from the AI but a need for robust quality assurance protocols

Sensitivity and specificity were improved with systematic prompt refinement, reducing the number of missed relevant articles while enhancing overall accuracy, but human oversight remained essential

We used AI to extract simple data at title/abstract screening and then ranked the studies
This allows us to adjust sensitivity and rapidly prioritize the most relevant studies, resulting in reports that include the bigger, better studies to rapidly give answers to questions and inform next steps

For each, we calculated how accurately the prompts identified or excluded studies and found big differences between strategies
We are using these results to choose the best approach based on the needs of each review

Connect with our experts
Get in touch to learn how we combine AI with scientific rigour for impactful results
Gemma Carter
Communications Director
gemma.carter@pharmagenesis.com
Polly Field
Communications Director
polly.field@pharmagenesis.com
Christian Eichinger
Scientific Director
christian@pharmagenesis.com
General enquiries
Want to find out more about AI-powered systematic literature review? Have questions that you’d like to ask? Or simply want a demo? Fill in your details below and we’ll be in touch to help you out.