DLFVC - 10 - Input Data Normalization / Data Preprocessing

preview_player
Показать описание
Deep Learning, Normalization, Data Preprocessing
Рекомендации по теме
Комментарии
Автор

Great video and very helpful. That was the missing link in data preprocessing for me.
One follow up question about multi dimensional tensors. For example, I have a timeseries of sensor readings that I want to feed into a Transformer Encoder model. I want to treat all sensor readings per timestep as a feature vector. The dataset has then the shape (#samples, sequence_length, features)
Since the sensor readings are in different magnitudes, I absolutely need to normalize it. Should I normalize it individually over one sample (meaning over one sequence) to keep the temporal relationship, or should I normalize it over the whole dataset, like in your 2d example?
Second question: In some NLP tasks the word embeddings are also normalized but with L2 normalization such that the embedding vector is scaled to unit length (over last axis). You also presented a few more normalizations like min-max, max-abs and robust. In which scenario is it best to use what normalization? Can one combine multiple normalizations, like minmax and mean/var? Which normalization technique would you recommend, given the description of my dataset above?
Sorry for the long question, I couldn't find good information about normalization. Greetings.

flake
welcome to shbcf.ru