Skip main navigation (Press Enter).
Log in
Toggle navigation
Log in
Community
Topic Groups
Champions
Directory
Program overview
Rising Champions
IBM Champions group
User Groups
Directory
Benefits
Events
Dev Days
Conference
Community events
User Groups events
All TechXchange events
Participate
TechXchange Group
Welcome Corner
Blogging
Member directory
Community leaders
Resources
IBM TechXchange
Community
Conference
Events
IBM Developer
IBM Training
IBM TechXchange
Community
Conference
Events
IBM Developer
IBM Training
Global AI and Data Science
×
Global AI & Data Science
Train, tune and distribute models with generative AI and machine learning capabilities
Group Home
Threads
4K
Blogs
907
Events
0
Library
370
Members
28.3K
View Only
Share
Share on LinkedIn
Share on X
Share on Facebook
Back to discussions
Expand all
|
Collapse all
sort by most recent
sort by thread
RoBERTa: a Robustly Optimized BERT Pretraining Approach
William Roberts
Wed September 04, 2019 12:55 AM
How do you optimize language model pre-training when training tends to be computationally expensive and ...
William Roberts
Mon September 23, 2019 06:31 PM
RoBERTa is surely going to drop out of SOTA, soon! ------------------------------ William Roberts ------------------------------ ...
1.
RoBERTa: a Robustly Optimized BERT Pretraining Approach
Like
William Roberts
Posted Wed September 04, 2019 12:55 AM
Edited by System Admin Fri January 20, 2023 04:09 PM
Reply
Options Dropdown
How do you optimize language model pre-training when training tends to be computationally expensive and executed on differing datasets? Maybe
RoBERTa
has the answers. Facebook's pre-training recipe appears to have greatly improved on BERT's bench-marking performance. What do you think is in store for RoBERTa?
Image from:
https://github.com/facebookresearch/LASER/blob/master/tasks/WikiMatrix/WikiMatrix-sizes.pdf
------------------------------
William Roberts
------------------------------
#GlobalAIandDataScience
#GlobalDataScience
2.
RE: RoBERTa: a Robustly Optimized BERT Pretraining Approach
Like
William Roberts
Posted Mon September 23, 2019 06:31 PM
Reply
Options Dropdown
RoBERTa is surely going to drop out of SOTA, soon!
------------------------------
William Roberts
------------------------------
Original Message
Original Message:
Sent: Wed September 04, 2019 12:54 AM
From: William Roberts
Subject: RoBERTa: a Robustly Optimized BERT Pretraining Approach
How do you optimize language model pre-training when training tends to be computationally expensive and executed on differing datasets? Maybe
RoBERTa
has the answers. Facebook's pre-training recipe appears to have greatly improved on BERT's bench-marking performance. What do you think is in store for RoBERTa?
Image from:
https://github.com/facebookresearch/LASER/blob/master/tasks/WikiMatrix/WikiMatrix-sizes.pdf
------------------------------
William Roberts
------------------------------
#GlobalAIandDataScience
#GlobalDataScience
×
New Best Answer
This thread already has a best answer. Would you like to mark this message as the new best answer?
Related Content
Data Science Community News | Volume 1, Issue 2
Christina Howell
Added Mon September 09, 2019
Blog Entry
Understanding Data in Transformers AI: Inputs, Outputs, and Training
Youssef Sbai Idrissi
Added Fri July 21, 2023
Blog Entry
Data Science Community News | Volume 1, Issue 1
Christina Howell
Added Mon August 05, 2019
Blog Entry
Data Science Community News | March
Christina Howell
Added Thu March 12, 2020
Blog Entry
Data Science Training in Mumbai
Richa Malhotra
Added Fri October 01, 2021
Discussion Thread
2
Powered by Higher Logic