4 Recently asked Pyspark Coding Questions | Apache Spark Interview

preview_player
ะŸะพะบะฐะทะฐั‚ัŒ ะพะฟะธัะฐะฝะธะต

I have trained over 20,000+ professionals in the field of Data Engineering in the last 5 years.

๐–๐š๐ง๐ญ ๐ญ๐จ ๐Œ๐š๐ฌ๐ญ๐ž๐ซ ๐’๐๐‹? ๐‹๐ž๐š๐ซ๐ง ๐’๐๐‹ ๐ญ๐ก๐ž ๐ซ๐ข๐ ๐ก๐ญ ๐ฐ๐š๐ฒ ๐ญ๐ก๐ซ๐จ๐ฎ๐ ๐ก ๐ญ๐ก๐ž ๐ฆ๐จ๐ฌ๐ญ ๐ฌ๐จ๐ฎ๐ ๐ก๐ญ ๐š๐Ÿ๐ญ๐ž๐ซ ๐œ๐จ๐ฎ๐ซ๐ฌ๐ž - ๐’๐๐‹ ๐‚๐ก๐š๐ฆ๐ฉ๐ข๐จ๐ง๐ฌ ๐๐ซ๐จ๐ ๐ซ๐š๐ฆ!

"๐€ 8 ๐ฐ๐ž๐ž๐ค ๐๐ซ๐จ๐ ๐ซ๐š๐ฆ ๐๐ž๐ฌ๐ข๐ ๐ง๐ž๐ ๐ญ๐จ ๐ก๐ž๐ฅ๐ฉ ๐ฒ๐จ๐ฎ ๐œ๐ซ๐š๐œ๐ค ๐ญ๐ก๐ž ๐ข๐ง๐ญ๐ž๐ซ๐ฏ๐ข๐ž๐ฐ๐ฌ ๐จ๐Ÿ ๐ญ๐จ๐ฉ ๐ฉ๐ซ๐จ๐๐ฎ๐œ๐ญ ๐›๐š๐ฌ๐ž๐ ๐œ๐จ๐ฆ๐ฉ๐š๐ง๐ข๐ž๐ฌ ๐›๐ฒ ๐๐ž๐ฏ๐ž๐ฅ๐จ๐ฉ๐ข๐ง๐  ๐š ๐ญ๐ก๐จ๐ฎ๐ ๐ก๐ญ ๐ฉ๐ซ๐จ๐œ๐ž๐ฌ๐ฌ ๐š๐ง๐ ๐š๐ง ๐š๐ฉ๐ฉ๐ซ๐จ๐š๐œ๐ก ๐ญ๐จ ๐ฌ๐จ๐ฅ๐ฏ๐ž ๐š๐ง ๐ฎ๐ง๐ฌ๐ž๐ž๐ง ๐๐ซ๐จ๐›๐ฅ๐ž๐ฆ."

๐‡๐ž๐ซ๐ž ๐ข๐ฌ ๐ก๐จ๐ฐ ๐ฒ๐จ๐ฎ ๐œ๐š๐ง ๐ซ๐ž๐ ๐ข๐ฌ๐ญ๐ž๐ซ ๐Ÿ๐จ๐ซ ๐ญ๐ก๐ž ๐๐ซ๐จ๐ ๐ซ๐š๐ฆ -

In this session I have talked about 4 interview questions which were recently asked in pyspark coding interview.

I am sure this session is going to help all the big data enthusiasts.

#bigdata #dataengineering #pyspark
ะ ะตะบะพะผะตะฝะดะฐั†ะธะธ ะฟะพ ั‚ะตะผะต
ะšะพะผะผะตะฝั‚ะฐั€ะธะธ
ะะฒั‚ะพั€

Need more Pyspark Interview Solutions like this ๐Ÿ˜Š

souradeep.official
ะะฒั‚ะพั€

One of the great explanation so far on youtube. I wish i could afford your course :(

adityatomar
ะะฒั‚ะพั€

It will be great if you put questions in comment . Others can try without looking at solution first

venugopal-ncnz
ะะฒั‚ะพั€

00:03 Recently asked Pyspark Coding Questions
02:37 Writing and executing Pyspark pseudo code
05:21 Creating a Spark dataframe from input and performing group by aggregation
08:04 Using aggregation functions and collect list in Pyspark.
11:15 Spark SQL solution for creating DataFrame and running queries.
14:18 Understanding the data frame reader API for reading JSON and the usage of explode function
17:11 Creating a Spark dataframe and performing operations on it.
19:44 Converting string to date and performing group by in Pyspark DataFrame
22:32 Finding the average stock value using PySpark
25:38 Practice more on data frames for interviews
28:15 Practice more to gain confidence in writing correct syntax for Pyspark coding

singhjirajeev
ะะฒั‚ะพั€

One of the best interview series Thank you sumit sir .

sopankardile
ะะฒั‚ะพั€

Thank you Sir greatly explained, would be good if you can post data/schemas also in the decription box for us to query and do hands on. Thanks.! :)

shashankgupta
ะะฒั‚ะพั€

Very useful informative video which gives more confidence to the bigdata aspirants. Thanks Sumit.

veerugandhad
ะะฒั‚ะพั€

Sir...Share need more .. please continue this playlist

gudiatoka
ะะฒั‚ะพั€

We can apply distinct() too I guess for avoiding duplicate values in df.

rohit-llrj
ะะฒั‚ะพั€

What about remaining 10 questions on pyspark you told we are covering it in next video but still you not uploaded on YouTube and when you will upload it on YouTube we are waiting for remaining 10 questions on pyspark
Thank you โค

electricalsir
ะะฒั‚ะพั€

Hi Sumit,
Could you please create Video explaining pipelines on AWS Databricks End-End along with Orchestration of those.

prasoonvijay
ะะฒั‚ะพั€

You are doing a great job posting theseโค

praptijoshi
ะะฒั‚ะพั€

Nice explanation sir, kindly post scenario based questions

anjibabumakkena
ะะฒั‚ะพั€

in question number 2 = do we not need to remove duplicate as last can you please clear me on it ?

dlckymr
ะะฒั‚ะพั€

Hi Sir, can we not write in Spark sql in interview? As there is no difference in performance.

mdasif
ะะฒั‚ะพั€

Hello sir, how can I run pyspark code online, are you also using any online utilty to run pyspark code as shown in this video, could you please share the source, it would be very helpful.

jlcbcsj