Ask a large language model (LLM) such as ChatGPT to summarize what people are saying about a topic, and although the model might summarize the facts efficiently, it might give a false impression of how people feel about the topic. LLMs play an increasingly large role in research, but rather than being a transparent window into the world, they can present and summarize content with a different tone and emphasis than the original data, potentially skewing research results.