• precompute.net
  • 4archives
  • Literature Notes
  • Readings
precompute.net  Blog   About 
[Home]
[Tree]
[Tags]
[Tasks]
[Search]
About Me
4archives
0 Lifestyle Creep
How Much Lifestyle Creep Can You Afford
AWS Certified Machine Learning Specialty [44]
BASB [16]
BigScience [5]
Blog Pages [6]
Hypothesis [14]
Just Keep Buying
Literature Notes
Courses [4]
Readings
AN IMAGE IS WORTH 16X16 WORDS: TRANSFORMERS FOR IMAGE RECOGNITION AT SCALE
Article Lessions from my PhD
Redefining SOTA
Article Reflections on my (Machine Learning) PhD Journey
Scikit-learn Pitfalls
Article Top 10 Open Source MLOps Tools
Digital Zettelkasten
Book How to Take Smart Notes
How to Write a Paper with ZettleKasten
Zettlekasten Principles
DIT: SELF-SUPERVISED PRE-TRAINING FOR DOCUMENT IMAGE TRANSFORMER
Gu et al_2021_UniDoc
Learning with Signatures
Paper A Practical Survey on Faster and Lighter Transformers
Paper CANINE Pre-training an Efficient Tokenization-Free Encoder for Language Representation
Paper Charformer Fast Character Transformers via Gradient-based Subword Tokenization
data2vec: A General Framework for Self-supervised Learning in Speech, Vision and Language
Paper DeBERTa Decoding-enhanced BERT with disentangled attention
Paper Fastformer Additive Attention Can Be All You Need
Gradients without Backpropagation
Hidden Techinical Debt in Machine Learning Systems
Paper Language Models are Few-Shot Learners
LayoutReader Pre-training of Text and Layout for Reading Order Detection
Paper Muppet Massive Multi-task Representations with Pre-Finetuning
Paper Robust Open-Vocabulary Translation from Visual Text Representations
Paper Shortformer Better Language Modeling using Shorter Inputs
Paper Supervised Contrastive Learning for Pre-trained Language Model Fine-tuning
Paper Switch Transformers
Switch Transformers: Scaling to Trillion Parameter Models with Simple and Efficient Sparsity
UNIPELT: A Unified Framework for Parameter-Efficient Language Model Tuning
Post Embedding Layer might not be Necessary for Your Next NLP Project
Log in Python
Deduplication in Modern Large-scale LM Datasets
Research Two-sample Hypothesis Testing
Spelling Correction with Denoising Transformer
UNICORN on RAINBOW: A Universal Commonsense Reasoning Model on a New Multitask Benchmark
UniDoc Unified Pretraining Framework for Document Understanding
Unified Pretraining Framework for Document Understanding
Video Transformer in Transformer
Machine Learning Concepts [2]
Permanent Notes [24]
Privacy AI [2]
System Design [4]
aaa
blog [1]
coffee [9]
Testfile

Paper Charformer Fast Character Transformers via Gradient-based Subword Tokenization

Links to this page
  • 202110091321 Open Vocabulary

    Paper CANINE Pre-training an Efficient Tokenization-Free Encoder for Language Representation Paper Charformer Fast Character Transformers via Gradient-based Subword Tokenization