CANINE: Pre-training an Efficient Tokenization-Free Encoder for Language Representation. Clark, J. H., Garrette, D., Turc, I., & Wieting, J. arXiv:2103.06874 [cs], March, 2021. arXiv: 2103.06874Paper abstract bibtex Pipelined NLP systems have largely been superseded by end-to-end neural modeling, yet nearly all commonly-used models still require an explicit tokenization step. While recent tokenization approaches based on data-derived subword lexicons are less brittle than manually engineered tokenizers, these techniques are not equally suited to all languages, and the use of any fixed vocabulary may limit a model's ability to adapt. In this paper, we present CANINE, a neural encoder that operates directly on character sequences, without explicit tokenization or vocabulary, and a pre-training strategy that operates either directly on characters or optionally uses subwords as a soft inductive bias. To use its finer-grained input effectively and efficiently, CANINE combines downsampling, which reduces the input sequence length, with a deep transformer stack, which encodes context. CANINE outperforms a comparable mBERT model by 2.8 F1 on TyDi QA, a challenging multilingual benchmark, despite having 28% fewer model parameters.
@article{clark_canine_2021,
title = {{CANINE}: {Pre}-training an {Efficient} {Tokenization}-{Free} {Encoder} for {Language} {Representation}},
shorttitle = {{CANINE}},
url = {http://arxiv.org/abs/2103.06874},
abstract = {Pipelined NLP systems have largely been superseded by end-to-end neural modeling, yet nearly all commonly-used models still require an explicit tokenization step. While recent tokenization approaches based on data-derived subword lexicons are less brittle than manually engineered tokenizers, these techniques are not equally suited to all languages, and the use of any fixed vocabulary may limit a model's ability to adapt. In this paper, we present CANINE, a neural encoder that operates directly on character sequences, without explicit tokenization or vocabulary, and a pre-training strategy that operates either directly on characters or optionally uses subwords as a soft inductive bias. To use its finer-grained input effectively and efficiently, CANINE combines downsampling, which reduces the input sequence length, with a deep transformer stack, which encodes context. CANINE outperforms a comparable mBERT model by 2.8 F1 on TyDi QA, a challenging multilingual benchmark, despite having 28\% fewer model parameters.},
urldate = {2022-02-15},
journal = {arXiv:2103.06874 [cs]},
author = {Clark, Jonathan H. and Garrette, Dan and Turc, Iulia and Wieting, John},
month = mar,
year = {2021},
note = {arXiv: 2103.06874},
keywords = {Computer Science - Computation and Language, Computer Science - Machine Learning},
}
Downloads: 0
{"_id":"iTDX9tPqu6WYqycPx","bibbaseid":"clark-garrette-turc-wieting-caninepretraininganefficienttokenizationfreeencoderforlanguagerepresentation-2021","author_short":["Clark, J. H.","Garrette, D.","Turc, I.","Wieting, J."],"bibdata":{"bibtype":"article","type":"article","title":"CANINE: Pre-training an Efficient Tokenization-Free Encoder for Language Representation","shorttitle":"CANINE","url":"http://arxiv.org/abs/2103.06874","abstract":"Pipelined NLP systems have largely been superseded by end-to-end neural modeling, yet nearly all commonly-used models still require an explicit tokenization step. While recent tokenization approaches based on data-derived subword lexicons are less brittle than manually engineered tokenizers, these techniques are not equally suited to all languages, and the use of any fixed vocabulary may limit a model's ability to adapt. In this paper, we present CANINE, a neural encoder that operates directly on character sequences, without explicit tokenization or vocabulary, and a pre-training strategy that operates either directly on characters or optionally uses subwords as a soft inductive bias. To use its finer-grained input effectively and efficiently, CANINE combines downsampling, which reduces the input sequence length, with a deep transformer stack, which encodes context. CANINE outperforms a comparable mBERT model by 2.8 F1 on TyDi QA, a challenging multilingual benchmark, despite having 28% fewer model parameters.","urldate":"2022-02-15","journal":"arXiv:2103.06874 [cs]","author":[{"propositions":[],"lastnames":["Clark"],"firstnames":["Jonathan","H."],"suffixes":[]},{"propositions":[],"lastnames":["Garrette"],"firstnames":["Dan"],"suffixes":[]},{"propositions":[],"lastnames":["Turc"],"firstnames":["Iulia"],"suffixes":[]},{"propositions":[],"lastnames":["Wieting"],"firstnames":["John"],"suffixes":[]}],"month":"March","year":"2021","note":"arXiv: 2103.06874","keywords":"Computer Science - Computation and Language, Computer Science - Machine Learning","bibtex":"@article{clark_canine_2021,\n\ttitle = {{CANINE}: {Pre}-training an {Efficient} {Tokenization}-{Free} {Encoder} for {Language} {Representation}},\n\tshorttitle = {{CANINE}},\n\turl = {http://arxiv.org/abs/2103.06874},\n\tabstract = {Pipelined NLP systems have largely been superseded by end-to-end neural modeling, yet nearly all commonly-used models still require an explicit tokenization step. While recent tokenization approaches based on data-derived subword lexicons are less brittle than manually engineered tokenizers, these techniques are not equally suited to all languages, and the use of any fixed vocabulary may limit a model's ability to adapt. In this paper, we present CANINE, a neural encoder that operates directly on character sequences, without explicit tokenization or vocabulary, and a pre-training strategy that operates either directly on characters or optionally uses subwords as a soft inductive bias. To use its finer-grained input effectively and efficiently, CANINE combines downsampling, which reduces the input sequence length, with a deep transformer stack, which encodes context. CANINE outperforms a comparable mBERT model by 2.8 F1 on TyDi QA, a challenging multilingual benchmark, despite having 28\\% fewer model parameters.},\n\turldate = {2022-02-15},\n\tjournal = {arXiv:2103.06874 [cs]},\n\tauthor = {Clark, Jonathan H. and Garrette, Dan and Turc, Iulia and Wieting, John},\n\tmonth = mar,\n\tyear = {2021},\n\tnote = {arXiv: 2103.06874},\n\tkeywords = {Computer Science - Computation and Language, Computer Science - Machine Learning},\n}\n\n","author_short":["Clark, J. H.","Garrette, D.","Turc, I.","Wieting, J."],"key":"clark_canine_2021","id":"clark_canine_2021","bibbaseid":"clark-garrette-turc-wieting-caninepretraininganefficienttokenizationfreeencoderforlanguagerepresentation-2021","role":"author","urls":{"Paper":"http://arxiv.org/abs/2103.06874"},"keyword":["Computer Science - Computation and Language","Computer Science - Machine Learning"],"metadata":{"authorlinks":{}},"html":""},"bibtype":"article","biburl":"https://bibbase.org/zotero/mxmplx","dataSources":["aXmRAq63YsH7a3ufx"],"keywords":["computer science - computation and language","computer science - machine learning"],"search_terms":["canine","pre","training","efficient","tokenization","free","encoder","language","representation","clark","garrette","turc","wieting"],"title":"CANINE: Pre-training an Efficient Tokenization-Free Encoder for Language Representation","year":2021}