Spark Scenario Based Question | Handle Nested JSON in Spark | Using Spark with Scala | LearntoSpark

preview_player
Показать описание
In this video, We will learn how to handle nested JSON file using Spark with Scala. This will be useful for your Spark interview preparation.

Blog link to learn more on Spark:

Blog to handle nested Json file using Spark

Linkedin profile:

FB page:
Рекомендации по теме
Комментарии
Автор

Great Shahul, I was searching for this scenario from last month. But you saved my time.

sivasankargoturu
Автор

Thank you Azarudeen. It was nice tutorial. Looking for more concepts on spark scala.

gowthambabu
Автор

Thanks a lot for making video in Scala 🙏

srinivasasameer
Автор

What if I want to read json data coming from Kafka topics...using Microsoft spark..json is complex and getting it through stream..Thanks for such wonderful knowledge sharing

syedawasama
Автор

Bro awesome work....keep adding more video's

DilliPH
Автор

Very nicely explained in this video as well as in your blog.
Just one suggestion ...could you please put your input file in blog or shared path where viewers can use it for practice purpose ?

sagarrode
Автор

What if JSON is just one column of data, rest of the columns are strings. For example a csv file with many columns and one of the column is nested json that can have malformed json. How to validate json column and read only valid json?

pshar
Автор

thanks for the session, can you please explain who to pull the specific columns, i mean df1.select("name", "year").show(), instead of df1.select("name", "education_flat.*").show().

RaviKumar-oyjq
Автор

I need to convert mock CSV datato JSON format . Please help me this

ManishKumar-sreu
Автор

How to handle nested json and flatten json using pyspark can you please explain

sravankumar
Автор

Hi this is really good tutorial....can you please let me know how to do reverse like I have CSV file now I want to create nested json from it...

keyurthakor
Автор

I was trying to read a proper json file. I got _corrupt_record ...
How to handle such cases

Nehaz_palette
Автор

How to calculate number of partitions required for a 10 GB of data, and for repartitioning and coalesce please help??

MrManish
Автор

Commands are same to some extent in both Pyspark and Scala.
Will learning both at a same time be beneficial or lead to confusion..?

Do comment as I am newbie to both Pyspark and Scala!

nehachopade