TE
TechEcho
Home
24h Top
Newest
Best
Ask
Show
Jobs
English
GitHub
Twitter
Home
Mosaic trained a 1B parameter model on 440 GPUs for 200B tokens
2 points
by
ovaistariq
about 2 years ago
no comments
no comments