Login | Deutsch
  • TU
  • |
  • ULB
  • |
  • TUdata
    • English
    • Deutsch (Deutschland)
  •   Documentation
  •   FAQs
  •   Browse
    • All of TUdatalib
    • Communities
    • Titles
    • Authors
    • Type
    • By Submit Date
    • Subjects
    • DFG subject classifications
    • This Context
    • Titles
    • Authors
    • Type
    • By Submit Date
    • Subjects
    • DFG subject classifications
View Item 
  •   TUdatalib Home
  • Technische Universität Darmstadt
  • Forschungskooperationen
  • AIPHES Heidelberg
  • View Item
  •   TUdatalib Home
  • Technische Universität Darmstadt
  • Forschungskooperationen
  • AIPHES Heidelberg
  • View Item
JavaScript is disabled for your browser. Some features of this site may not work without it.

Browse

All of TUdatalibCommunitiesTitlesAuthorsTypeBy Submit DateSubjectsDFG subject classificationsThis CollectionTitlesAuthorsTypeBy Submit DateSubjectsDFG subject classifications

My Account

LoginRegister

Statistics

View Usage Statistics

BPEmb: Pre-trained Subword Embeddings in 275 Languages (LREC 2018)

Thumbnail
Author
Heinzerling, Benjamin
Metadata
Show full item record
Export
Text
BibTex
DataCite XML
JSON

Description

BPEmb is a collection of pre-trained subword unit embeddings in 275 languages, based on Byte-Pair Encoding (BPE). In an evaluation using fine-grained entity typing as testbed, BPEmb performs competitively, and for some languages better than alternative subword approaches, while requiring vastly fewer resources and no tokenization.

Subject

Computer and Information Science;subword embeddings;byte-pair encoding;multilingual

URI

https://doi.org/10.11588/data/V9CXPR

Collections

  • AIPHES Heidelberg [5]
Legal note | About TUdatalib | Contact | Privacy Policy | Agreement | Website analytics | OAI-PMH
 

 

Legal note | About TUdatalib | Contact | Privacy Policy | Agreement | Website analytics | OAI-PMH