Li Junnan
@LiJunnan0409
422 Following • 2.1K Followers
Multimodal Chief Scientist @rhymes_ai_ • Aria • Ex Senior Research Manager @salesforce • BLIP series
RT @DongxuLi_: Great thread to learn more about Aria pretraining pipelines and data recipes at @rhymes_ai_
2 months ago
🧵 (5/6) Unlike typical post-training of language models, our post-training involves a large amount of 20B tokens due to the diversity of multimodal data (e.g., documents, charts, videos). Some domains remain uncovered, where we resort to Aria's generalization capabilities.
230 views • 0 likes • 2 months ago
xAlerts
Public Lists
Articles
Legal