Search
On-Site Program Calendar
Browse By Day
Browse By Time
Browse By Person
Browse By Room
Browse By Unit
Browse By Session Type
Search Tips
Change Preferences / Time Zone
Sign In
Bluesky
Threads
X (Twitter)
YouTube
The rapid advancement in artificial intelligence (AI) presents both challenges and opportunities for traditional academic research paradigms. This study investigates the potential of large language models (LLMs) in automating the literature screening task for meta-analyses. Using a database of 2,198 articles, we compared the performance of two popular LLMs, GPT-4.0 and DeepSeek-V3 with manual verification in automated screening. Results showed DeepSeek-V3 achieves 88.2% accuracy in abstract screening (vs. GPT-4.0’s 63.3%) with faster processing speed. The accuracy was improved in both models when screening full-text (30 articles), reaching 93.9% (DeepSeek-V3) and 92.0% (GPT-4.0). These findings underscore AI’s transformative role in evidence synthesis, enabling researchers to focus on higher-order intellectual tasks.