How to fine-tune a NLLB-200 model for translating a new languageNLLB is a translation model that supports 200 languages. I teach it one more language, Tyvan, and explain the code behind this update.Oct 17, 2023A response icon9Oct 17, 2023A response icon9
Published inTDS ArchiveCompressing unsupervised fastText modelsA python package to reduce word embeddings models by 300 times, with almost the same performance on downstream NLP tasks.Dec 14, 2021A response icon1Dec 14, 2021A response icon1
Published inTDS ArchiveHow to adapt a multilingual T5 model for a single languageLoad embeddings only for the tokens from your language to reduce model sizeMay 4, 2021A response icon7May 4, 2021A response icon7
Do you have to try to love math?If you don’t like and don’t understand math, does it mean you are stupid? Do you need to love math to achieve at…Feb 13, 2018Feb 13, 2018
Published inThe StartupA machine learning model to understand fancy abbreviations, trained on TolkienRecently I bumped into a question on Stackoverflow, how to recover phrases from abbreviations, e.g. turn “wtrbtl” into “water bottle”, and…Jan 14, 2018A response icon5Jan 14, 2018A response icon5