TECH NEWS
Monday 1 June 2020
OpenAI researchers debut GPT-3 language model trained with 175B parameters, far more than GPT-2's biggest version with 1.5B parameters (Khari Johnson/VentureBeat)
via Techmeme https://ift.tt/2zUSDGb
No comments:
Post a Comment
Newer Post
Older Post
Home
Subscribe to:
Post Comments (Atom)
No comments:
Post a Comment