Positional encoding: Key idea to parallelize compute in deep learning models

Anudha Mittal
Sep 30, 2024

For notes on tokenization, see https://medium.com/@anudhamittal/tokenization-in-nlp-9b36b5cc2590

STEPS for a Deep Learning Model for Text Data

1/ Tokenize text

choice of tokenization schemes: BytePair , Word Piece, Sentence Piece

2/ Embed text (Query, Key, Value)

choice of embedding model: many

3/ Attach a positional encoding to each token

keeps track of position and permits parallel training of tokens

Sign up to discover human stories that deepen your understanding of the world.

Free

Distraction-free reading. No ads.

Organize your knowledge with lists and highlights.

Tell your story. Find your audience.

Membership

Read member-only stories

Support writers you read most

Earn money for your writing

Listen to audio narrations

Read offline with the Medium app

No responses yet

Write a response