Search
On-Site Program Calendar
Browse By Day
Browse By Time
Browse By Person
Browse By Room
Browse By Unit
Browse By Session Type
Search Tips
Change Preferences / Time Zone
Sign In
Bluesky
Threads
X (Twitter)
YouTube
Self-report instruments are vital in psychological research. Yet, developing such instruments is labor-intensive, requiring initial item writing, and expert review. Large Language Models (LLMs) can streamline this process through the preliminary review of items. This study examines the ability of OpenAI, Google, and Anthropic models to evaluate 28 self-report loneliness items. Results show ChatGPT-4 outperformed all other models in identifying bias and complexity, but all models struggled with idiomatic and double-barreled language and items measuring a closely related construct. While additional prompt development should deliver more accurate results, these LLMs should be seen as a complement to expert item review. Future research should focus on concise, clear prompts exploring a variety of constructs to expedite the development of new instruments.