{
"id":"6289000af6778feb835eea6bd26636db78d4562128028bf7e460695698f9b30b",
"pubkey":"311ec1842831549f5f4ba7d82a5d9b21eb5f7969dc0267c04bd2aedf728affb7",
"created_at":1728721478,
"kind":1,
"tags": [
[
"p",
"36d32bc91218c7575ce505a9fc21196602edb0b5b955c51752e232941f34671b"
],
[
"e",
"a388e5c6bdcb6fefe6fee86f195b5fa63060545c030d89a76968f89d49ce4efd",
"",
"root",
"36d32bc91218c7575ce505a9fc21196602edb0b5b955c51752e232941f34671b"
],
[
"proxy",
"https://mastodon.social/@stevefenton/113293490845841991",
"web"
],
[
"proxy",
"https://mastodon.social/users/stevefenton/statuses/113293490845841991",
"activitypub"
],
[
"L",
"pink.momostr"
],
[
"l",
"pink.momostr.activitypub:https://mastodon.social/users/stevefenton/statuses/113293490845841991",
"pink.momostr"
],
[
"-"
]
],
"content":"I once worked on an OCR project, and the execs never grasped the fundamental problem of OCR (which IMO applies to LLMs):\n\nIf something is wrong 20% of the time\nAnd you don't know which 20%\nYou have to check 100% of it's work",
"sig":"f383318d6f10e925a8425f6b722b20beec5692a19a6a2760f4e92f2565bf7fd6fc29ef1ae2fe8ae2f2f8c27a694d183d9d63eeb14d7bebcde816a1e69c68eb5b"
}