This article has multiple issues. Please help improve it or discuss these issues on the talk page. (Learn how and when to remove these messages)
|
The XLNet was an autoregressive Transformer designed as an improvement over BERT, with 340M parameters and trained on 33 billion words. It was released on 19 June, 2019, under the Apache 2.0 license.[1] It achieved state-of-the-art results on a variety of natural language processing tasks, including language modeling, question answering, and natural language inference.
Original author(s) | Google AI |
---|---|
Initial release | 19 June 2019 |
Repository | https://github.com/zihangdai/xlnet/ |
Type | |
License | Apache-2.0 |
Architecture
editThe main idea of XLNet is to model language autoregressively like the GPT models, but allow for all possible permutations of a sentence.[2] Concretely, consider the following sentence:
My dog is cute.
In standard autoregressive language modeling, the model would be tasked with predicting the probability of each word, conditioned on the previous words as its context:
We factorize the joint probability of a sequence of words using the chain rule:
For example, the sentence "My dog is cute" is factorized as:
Schematically, we can write it as
However, for XLNet, the model is required to predict the words in a randomly generated order. Suppose we have sampled a randomly generated order 3241, then schematically, the model is required to perform the following prediction task:
By considering all permutations, XLNet is able to capture longer-range dependencies and better model the bidirectional context of words.
Two-Stream Self-Attention
editTo implement permutation language modeling, XLNet uses a two-stream self-attention mechanism. The two streams are:
- Content stream: This stream encodes the content of each word, as in standard causally masked self-attention.
- Query stream: This stream encodes the content of each word in the context of what has came before. In more detail, it is a masked cross-attention mechanism, where the queries are from the query stream, and the key-value pairs are from the content stream.
The content stream uses the causal mask permuted by a random permutation matrix to .
The query stream uses the cross-attention mask , where the diagonal is subtracted away specifically to avoid the model "cheating" by looking at the content stream for what the current masked token is.
Like the causal masking for GPT models, this two-stream masked architecture allows the model to train on all tokens in one forward pass.
Training
editTwo models were released:[1][2]
- XLNet-Large, cased: 110M parameters, 24-layer, 1024-hidden, 16-heads
- XLNet-Base, cased: 340M parameters, 12-layer, 768-hidden, 12-heads.
It was trained on a dataset that amounted to 32.89 billion tokens after tokenization with SentencePiece. The dataset was composed of BooksCorpus, and English Wikipedia, Giga5, ClueWeb 2012-B, and Common Crawl.
It was trained on 512 TPU v3 chips, for 5.5 days. At the end of training, it still under-fitted the data, meaning it could have achieved lower loss with more training. It took 0.5 million steps with an Adam optimizer, linear learning rate decay, and a batch size of 8192.[3]
See also
editReference
edit- ^ a b "xlnet". GitHub. Retrieved 2 January 2024.
- ^ a b "Pretrained models — transformers 2.0.0 documentation". huggingface.co. Retrieved 2024-08-05.
- ^ Yang, Zhilin; Dai, Zihang; Yang, Yiming; Carbonell, Jaime; Salakhutdinov, Ruslan; Le, Quoc V. (2 January 2020). "XLNet: Generalized Autoregressive Pretraining for Language Understanding". arXiv:1906.08237 [cs.CL].