Add One Smoothing

preview_player
Показать описание
What do we do with words that are in our vocabulary (they are not unknown words)
but appear in a test set in an unseen context (for example they appear after a word
they never appeared after in training)? To keep a language model from assigning
zero probability to these unseen events, we’ll have to shave off a bit of probability
mass from some more frequent events and give it to the events we’ve never seen.
This modification is called smoothing or discounting. In this section and the following ones we’ll introduce a variety of ways to do smoothing: Laplace (add-one)
smoothing, add-k smoothing, stupid backoff, and Kneser-Ney smoothing.
Рекомендации по теме
Комментарии
Автор

I which when we calculate P( <s> | I ) then in this case what is the value of <s>

infopediaCreateor
Автор

9:01 - Sir isme V (Unique Vocabulary) kaise calculate karte hai?

thepriestofvaranasi