Spark Scenario Based Question | Alternative to df.count() | Use Case For Accumulators | learntospark

preview_player
Показать описание
In this video, we will discuss about the scenario based question on Spark. We will see an alternative method to find the count of dataframe in Spark application. We will learn about the use case for accumulator using this Demo.

DataSet:

Blog link to learn more on Spark:

Linkedin profile:

FB page:
Рекомендации по теме
Комментарии
Автор

Can you please make these kind of videos more frequently..

sudheersatya
Автор

please provide databricks interview Q & A

smdimran
Автор

You are taking all the data to the driver, tat will create memory issues right. What about listeners to get the metrics?

radhakrishnanselvaraj
Автор

Bro can you share in spark scala and groupBy need to give all columns to find duplicates, when I try to use header inside groupBy not working, so I have give all column header separately within double quotes. Any solution?

maheshk
Автор

Can you make video on the spark partition how spark decide no of partitions while reading the data and while doing shuffle operation and how to decide the no the partitions and how to change the partitions.

ravikirantuduru
Автор

**IMP**
Hi.. are you giving any training to people who need real-time knowledge and exp? if so please let me know how to get in touch with you?

jagadeeswarap