Event JSON
{
"id": "b02030a1118b4918d3457319ba8eea4be72d680db5f11af3e605dae1afa1e961",
"pubkey": "22f0d290be692a4cc66d000f97257b257dcd81a64c76b633f5a1b3ab51738a61",
"created_at": 1718065281,
"kind": 1,
"tags": [
[
"t",
"voting"
],
[
"t",
"llm"
],
[
"t",
"ai"
],
[
"t",
"google"
],
[
"t",
"openai"
],
[
"proxy",
"https://flipboard.social/@TechDesk/112595126316518895",
"web"
],
[
"t",
"elections"
],
[
"proxy",
"https://flipboard.social/users/TechDesk/statuses/112595126316518895",
"activitypub"
],
[
"L",
"pink.momostr"
],
[
"l",
"pink.momostr.activitypub:https://flipboard.social/users/TechDesk/statuses/112595126316518895",
"pink.momostr"
]
],
"content": "A study has found that large language models, including Google’s Gemini and OpenAI’s ChatGPT, gave incorrect information 27% of the time when asked about voting and the 2024 U.S. election, and some models individually faired much worse.\n\nWhile some answers to researchers' questions came with warnings for users to check the information with their local election authorities, many did not, creating concerns that voters could arm themselves with incorrect information about candidates and voting law. NBC News has more.\n\nhttps://flip.it/L6SIjG\n\n#LLM #Voting #Elections #AI #Google #OpenAI",
"sig": "91edcebd8893648d85c320416c4bf81b9594e59b7d8f78349fac55a3b6fee1a0f442151f3e6534814e615aaefa8f6981f20fd52a4f32c10b63b5dde4142edf2d"
}