Sunday, May 15, 2022

Azure Databricks- Importing CSV file into DataBricks File System with PySpark Code

 

Azure DataBricks File System

Databricks File System (DBFS) is a distributed file system mounted into an Azure Databricks workspace and available on Azure Databricks clusters. 

DBFS is an abstraction on top of scalable object storage


The default storage location in DBFS is known as the DBFS root.

  /FileStore: Imported data files, generated plots, and uploaded libraries

  /databricks-datasets: Sample public datasets.

  /databricks-results: Files generated by downloading the full results of a query.

Below video contains example video with importing CSV file to DBFS and perform some transformation in it





0 comments:

Post a Comment

Global Certifications: