Audience

AI developers

About ALBERT

ALBERT is a self-supervised Transformer model that was pretrained on a large corpus of English data. This means it does not require manual labelling, and instead uses an automated process to generate inputs and labels from raw texts. It is trained with two distinct objectives in mind. The first is Masked Language Modeling (MLM), which randomly masks 15% of words in the input sentence and requires the model to predict them. This technique differs from RNNs and autoregressive models like GPT as it allows the model to learn bidirectional sentence representations. The second objective is Sentence Ordering Prediction (SOP), which entails predicting the ordering of two consecutive segments of text during pretraining.

Integrations

Ratings/Reviews

Overall 0.0 / 5
ease 0.0 / 5
features 0.0 / 5
design 0.0 / 5
support 0.0 / 5

This software hasn't been reviewed yet. Be the first to provide a review:

Review this Software

Company Information

Google
Founded: 1998
United States
github.com/google-research/albert

Videos and Screen Captures

ALBERT Screenshot 1
Other Useful Business Software
Outgrown Windows Task Scheduler? Icon
Outgrown Windows Task Scheduler?

Free diagnostic identifies where your workflow is breaking down—with instant analysis of your scheduling environment.

Windows Task Scheduler wasn't built for complex, cross-platform automation. Get a free diagnostic that shows exactly where things are failing and provides remediation recommendations. Interactive HTML report delivered in minutes.
Download Free Tool

Product Details

Platforms Supported
Cloud
Training
Documentation

ALBERT Frequently Asked Questions

Q: What kinds of users and organization types does ALBERT work with?
Q: What languages does ALBERT support in their product?
Q: What type of training does ALBERT provide?

ALBERT Product Features