10 PySpark Product Based Interview Questions

preview_player
Показать описание
10 PySpark Product-Based Interview Questions

Question 1 -00:00:00
Question 2 -00:04:24
Question 3 -00:07:36
Question 4 -00:13:35
Question 5 -00:21:29
Question 6 -00:24:28
Question 7 -00:26:36
Question 8 -00:29:12
Question 9 -00:33:13
Question 10 -00:37:01

17 Hours Big Data Engineering Full Course

Hash Tags
#bigdata #dataengineering #pyspark
Рекомендации по теме
Комментарии
Автор

Thanks for the video, Please continue this pyspark interview videos.
Thanks Again.

abhishekn
Автор

Really a nice explanation with a clear shot... Thanks a lot... Please keep more videos on pyspark

Sandeep-blji
Автор

Only one video on the Pyspark Playlist ...
Pls post more!!

Vanmathi-eo
Автор

i feel we can solve the 7th question using window function row_number() as well

chandrarahul
Автор

Doubt in 1st question:
The delimiters in the data are ", " "\t" "|"
Then why did you use ", |\t|\|"
Please explain.

amitrou
Автор

Bro, thanks for your effort for sharing interview based real time questions and answers. can you please share realtime streaming (kafka and pyspark) based interview based questions and answers??.

Ameem-rwir
Автор

Instead of left anti we can use except

boreddykesavareddy
Автор

in interview if we solve the problems in SQL using sparksql will it be okay?

nlgqcct
Автор

bro, thanks for your inputs. below data is in a file. can you please help me how to handle this?. I got bit trouble on your one line string data with what i have in multiple rows with multiple delimiter.

Expected output
empid, fname, lname, sal, deptid
1, mohan, kumar, 5000, 100
2, karan, varadan, 3489, 101
3, kavitha, gandan, 6000, 102

hyderali-wlyi
Автор

bro, thanks for your inputs. below data is a file.can you please help me how to handle this in pyspark?

Expected output
empid, fname, lname, sal, deptid
1, mohan, kumar, 5000, 100
2, karan, varadan, 3489, 101
3, kavitha, gandan, 6000, 102

hyderali-wlyi