25. groupBy() in PySpark | Azure Databricks #spark #pyspark #azuredatabricks #azuresynapse #azure

preview_player
Показать описание
In this video, I discussed about groupBy() in Pyspark which helps to perform grouping of rows in dataframe.

Link for PySpark Playlist:

Link for PySpark Real Time Scenarios Playlist:

Link for Azure Synapse Analytics Playlist:

Link to Azure Synapse Real Time scenarios Playlist:

Link for Azure Data bricks Play list:

Link for Azure Functions Play list:

Link for Azure Basics Play list:

Link for Azure Data factory Play list:

Link for Azure Data Factory Real time Scenarios

Link for Azure Logic Apps playlist

#PySpark #Spark #databricks #azuresynapse #synapse #notebook #azuredatabricks #PySparkcode #dataframe #WafaStudies #maheer #azure
Рекомендации по теме
Комментарии
Автор

Thank you sir this use full for me today i'm struggle this scenario based interview question

sreenivasreddydadam
Автор

Hi sir, I have one doubt
I am loading tables from csv file to Sql serverless pool with the help of Azure Synapse analytics pipelines and notebook.
There, I am facing one issue:
My soure data I.e Csv File has more than 8000 to 50000 values in each rows in one of my column. I can see the complete data in Pyspark notebook.
But, I am unable to see my tables in ssms which is an serverless pool. Because the max char is 8000 by default in Serverless pool.
Suggest me an idea how to get my tables, which has more than 8000 chars in Serverlesspool.

Thanks.

ashokkumarm
Автор

How to use group by without using agg function

gauravguleria
Автор

Thank you bhaii...waiting for next video aggregate function

manu
Автор

For reference Can you please help us with the PDF of Pyspark tutorial

interviewpreparation
Автор

Subscribed!!! Wonderful Pyspark series. Please continue until advanced topics are covered.
Can you give me a real time use case where we prefer using groupByKey() over reduceByKey() in spark?

SatishKumar-fryv