Azure Databricks- Importing CSV file into DataBricks File System with PySpark Code

preview_player
Показать описание
Databricks File System (DBFS) is a distributed file system mounted into an Azure Databricks workspace and available on Azure Databricks clusters.

DBFS is an abstraction on top of scalable object storage

The default storage location in DBFS is known as the DBFS root.

/FileStore: Imported data files, generated plots, and uploaded libraries

/databricks-datasets: Sample public datasets.

/databricks-results: Files generated by downloading the full results of a query.

video contains example video with importing CSV file to DBFS and perform some transformation in it
Рекомендации по теме
join shbcf.ru