--- title: "Using OpenAI's GPT API models for Title and Abstract Screening in Systematic Reviews" author: "Mikkel H. Vembye" date: "`r format(Sys.time(), '%Y.%m.%d')`" bibliography: AIscreenR.bib link-citations: true csl: apa.csl format: html: code-fold: false code-tools: true code-summary: "Show the code" toc: true vignette: | %\VignetteIndexEntry{Using GPT API Models For Screening} %\VignetteEngine{quarto::html} %\VignetteEncoding{UTF-8} execute: echo: fenced warning: false message: false knitr: opts_chunk: fig.pos: "H" fig.retina: 2 cache: FALSE R.options: knitr.graphics.auto_pdf: true width: 100 knitr.kable.NA: "-" dplyr.summarise.inform: FALSE scipen: 10 pillar.sigfig: 4 editor_options: chunk_output_type: console ---
Important note
*Always remember that title and abstract screeening with GPT API models can be case sensitive. Therefore, see [Vembye, Christensen, Mølgaard, & Schytt](https://psycnet.apa.org/record/2026-37236-001) [-@vembye_generative_2025] for an overview of how and when GPT API models can be used for title and abstract (TAB) screening (find the article record of Vembye et al. 2025 [here](https://psycnet.apa.org/record/2026-37236-001)). Our most recent results suggest that the `gpt-4o-mini` is an effective model for screening titles and abstracts with performances in many cases on par with the `gpt-4` and `gpt-5` models. This is a very cheap model (200 times cheaper than `gpt-4`). Therefore, to reduce costs, we recommendation always testing the performance of `gpt-4o-mini` before considering other models. Prompting `gpt-5` models can be different from `gpt-4` models. Therefore, see the [comparistion of model](https://mikkelvembye.github.io/AIscreenR/articles/comparing-reasoning-models.html) article on the AIscreenR webpage for guidance.*