Delta Lake in Spark | Schema Evaluation using Delta | Session - 1 | LearntoSpark

preview_player
Показать описание
In this video, we will learn how schema evaluation problem is handled in Delta lake from Spark 3.0. We will have a Demo on the existing problem and Delta lake solution using PySpark.

Blog link to learn more on Spark:

Linkedin profile:

FB page:
Рекомендации по теме
Комментарии
Автор

Thanks for providing videos with examples. Really appreciate for your time. FYI, We can use merge schema with tradition approach as well, it’s available as part of 2.4. Please refer Apache spark site.

srinuch
Автор

If we dont have spark 3.0 then what might be the option apart from delta

vishalbhardwaj
Автор

Hi, Thanks for the video. How this issue resolved in traditional approach?

naveenkumarmaddala
Автор

Hi,
I am getting an error while tried to write with "delta" and in community edition, I am not getting Spark 3.0 rather I am getting spark 3.1.1

tanmoychowdhury
Автор

Really informative video bro!

BtW, are you tamil?

Shiva-kztn
Автор

Thanks for video can you create template for pyspark best practice to submit spark job on cluster

ravikirantuduru
Автор

Can you make video building pipeline of etl Kafka spark nifi etc and Thanks it was good.i find tutorials are useful and how to handle data skewness

ravikirantuduru
Автор

How to calculate number of partitions required for a 10 GB of data, and for repartitioning and coalesce please help??

MrManish
Автор

For those of you, who is trying to run from their local machine and getting error can't find class delta, create sparkSession using the below code and it will work.

import pyspark
spark = \
.config("spark.jars.packages", \
.config("spark.sql.extensions", \
.config("spark.sql.catalog.spark_catalog", \
.getOrCreate()

AtifImamAatuif