They got the whole Twitter database. It’s kinda the same with Gemini. But somehow Meta isn’t catching up, maybe their llama 4 architecture isn’t that stable to train.
Llama 3.3 was good, tho. For the multimodal, llama 4 also use llama3.2 approach where the image and text is made into single model instead using CLIP or siglip.
They got the whole Twitter database. It’s kinda the same with Gemini. But somehow Meta isn’t catching up, maybe their llama 4 architecture isn’t that stable to train.
Or maybe Facebook data is even worse than Twitter?
Llama 3.3 was good, tho. For the multimodal, llama 4 also use llama3.2 approach where the image and text is made into single model instead using CLIP or siglip.