Apache Spark Interview Questions and Answer | Spark Coding Assessment | Using PySpark

preview_player
Показать описание
#apachespark #sparkinterview #bigdata
Apache Spark Interview Questions and Answer | Spark Coding Assessment | Using PySpark

Coding Assessment given to candidate in Technical Interview by famous Service based company in India. Try to Solve on your own.
================================

Steps to install Spark in Windows 10:

===============================

Blog on Dataset and Notebook for you to Try:

================================

Blog link to learn more on Spark:

Linkedin profile:

FB page:
Рекомендации по теме
Комментарии
Автор

Hi why are you using rdd2=rdd1.flatMap(lambda x:x.split(', ') i dont see any comma in the input file

nitika
Автор

Hi bro. Your explanation was beautiful . I need a video on how to deploy spark job? Please make a video on this.

praneethbhat
Автор

You can directly read this file as csv in pipe delimited. Why to use rdd and make it complex.

amiyaghosh
Автор

I have a doubt, how lambda works here does it pick whole df or it takes the value row by row

mohitkumar-coxt
Автор

Why do u use rdd which is mostly not used in industry and not optimised ?

dhivakarsathya
Автор

Hi bro. I tired this code using case class schema defining. But getting error

pavithrasrividhya