Event JSON
{
"id": "556433ff24bac45e3f98935d3885a98bc4e3583c808f79714edc8feb1e7bd6ec",
"pubkey": "129f83898c7008d335771fe681ecf979e7767ad958c552ff85de962ba2f775be",
"created_at": 1733561489,
"kind": 1,
"tags": [
[
"t",
"meta"
],
[
"t",
"llama3"
],
[
"t",
"ai"
],
[
"t",
"opensource"
],
[
"t",
"llm"
],
[
"imeta",
"url https://social-cdn.vivaldi.net/system/media_attachments/files/113/610/685/779/170/910/original/3888e57185bfdfd2.jpeg",
"m image/jpeg",
"dim 1024x768",
"blurhash UC84utD*I9t6~qIVRPod%KM|o#f8s*RjxuWB"
],
[
"proxy",
"https://social.vivaldi.net/users/michabbb/statuses/113610685797215791",
"activitypub"
]
],
"content": "#Meta's #Llama3 3.3 70B: Advanced #AI Language Model Redefines Performance Standards š\n\nš¹ #Meta's latest #opensource #LLM features 70B parameters and 128K token context window, matching 405B model performance while requiring fewer resources\n\nhttps://social-cdn.vivaldi.net/system/media_attachments/files/113/610/685/779/170/910/original/3888e57185bfdfd2.jpeg",
"sig": "5fc54163a7394d3b5acd3215c938173cacc88af45bbc0fa9f9ae4d01cb609a3c2324cc4bfee2b9ad78f0ff943b23407e4ad4c81e782cee81bca1b5f521e6ac68"
}