Import CSV File

%%pyspark ## Magic Command, set language
import pyspark.sql.functions as f

#Absolute OneLake Path
pathContainer = 'abfss://a3663432-1fc7-4d96-afa5-01491fe98a3c@onelake.dfs.fabric.microsoft.com/03894def-6a71-405a-ac13-4a744161b0e7'
pathFile =  pathContainer + '/Files/Orders.csv'

#Relative path
pathFile = 'Files/Orders.csv'

#Whole folder
pathFile = 'Files/*.csv'

## Read CSV File
df = spark.read.load(pathFile, format='csv', header=True,sep=";", inferSchema=True)

df = df.withColumn("OrderDate", f.to_date(df['OrderDate'], "MM/dd/yyyy"))
df = df.withColumn("Price", f.regexp_replace(df["Price"], ",", "").cast("double"))

df = df.withColumn("LineTotal", f.round( df["Price"] * df["Quantity"] ,2))


## SELECT top 5
display(df.schema)
display(df.limit(5))

Import Parquet File

#Relative path folder
pathFile ='Files/sales-parquet/*/*/*/*/*.parquet'
df = spark.read.load(pathFile, format='parquet')

#Recursive
pathFile = 'Files/sales-parquet/'
df = spark.read.option("recursiveFileLookup","true").load(pathFile, format='parquet')


## SELECT top 5
df.printSchema()
display(df.limit(5))

Leave a Reply

Your email address will not be published. Required fields are marked *

Post comment