Skip to main content

Blog

Learn About Our Meetup

5000+ Members

MEETUPS

LEARN, CONNECT, SHARE

Join our meetup, learn, connect, share, and get to know your Toronto AI community. 

JOB POSTINGS

INDEED POSTINGS

Browse through the latest deep learning, ai, machine learning postings from Indeed for the GTA.

CONTACT

CONNECT WITH US

Are you looking to sponsor space, be a speaker, or volunteer, feel free to give us a shout.

[DISCUSSION] Bert Token Embeddings

From Paper is easy to understand that BERT input is composed by Token Embeddings, Positional Encode, Sentence Encode. The last two are well-defined in BERT paper and in “Attention is all you need”. But Token embeddings is not clear how are build. Reading on Internet I found different opinions. For sure tokenization is performed using WordPiece Tokens and it’s easy understand how it splits words. But once you have the token id how BERT converts it in a Embedding?

submitted by /u/lor_v
[link] [comments]