Bio-SIEVE: Exploring Instruction Tuning Large Language Models for Systematic Review Automation
Medical systematic reviews can be very costly and resource intensive.
Tags:Medical and Health Paper and LLMsPICOPricing Type
- Pricing Type: Free
- Price Range Start($):
GitHub Link
The GitHub link is https://github.com/ambroser53/bio-sieve
Introduce
The project “Bio-SIEVE” explores the use of Large Language Models (LLMs) for automating literature screening in medical systematic reviews. The study focuses on training LLMs to perform abstract screening based on specific selection criteria. The best model developed, named Bio-SIEVE, outperforms both ChatGPT and traditional methods, displaying better generalization across medical domains. The study also investigates multi-task training but finds that single-task Bio-SIEVE performs better. The models, code, and dataset information are made available for reproducibility. The project’s models, training process, and evaluation on various datasets are detailed, highlighting its potential for streamlining biomedical systematic reviews.
Medical systematic reviews can be very costly and resource intensive.
Content
The adapter weights for the 4 best models trained as part of this project can be found and used from HuggingFace: Instruct Cochrane consists of 5 main splits as detailed in the table below: The dataset can be constructed from separate lists of DOIs, as described in data/README.md. Models are all trained using a modified version of the QLoRA training script (Dettmers et. al.). An example training script is found below with the necessary parameters to recreate out model from the dataset. Models are evaluated on four datasets: Test, Subsets, Safety-First and Irrelevancy. Test evaluates the performance on the raw cochrane reviews. Subsets allow for comparison with logistic regression baselines as it allows for k-fold cross validation while training per review, simluating the existing active learning methods in literature. Safety-First better approximates the include/exclude process on just abstracts and titles. The test set is the final decision based on full-text screening, hence it is not always possible to derive their decision from the abstract and title alone. Irrelevancy is based on the subsets, wherein abstracts from completely different reviews are tested to evaluate whether the model can exclude samples far from the decision boundary. Details on using the evaluations scripts can be found in evaluation/README.md.

Related

GlassOuse's assistive devices aim to enhance accessibility and usability for individuals with disabilities, allowing them to engage with technology in a more inclusive and independent manner. These devices can be particularly beneficial for people with limited mobility who seek alternative methods of interaction.







