How to Read & Write Avro Files into a PySpark DataFrame | Simple Guide

Apache Spark is a powerful tool for big data processing, and PySpark is the Python API for Spark. One of the widely used formats for data storage and exchange in big data applications is Avro. Avro is a row-oriented binary serialization format that provides rich data structures and a compact, fast, binary data format. In this guide, we’ll take an in-depth look at how to read & write Avro files into a DataFrame using PySpark.

Introduction to Avro

Avro, a project under Apache Software Foundation, offers an efficient and compact way to encode data. It supports defining schemas through JSON, and it is a primary format used in Apache Hadoop. Avro files are often preferred because of their robust support for schema evolution and ease of use. Avro also supports various data types and compression codecs, making it efficient for data storage and transmission.

Why Use Avro with Spark?

Avro’s compact format and schema evolution capabilities align perfectly with the distributed computing power of Apache Spark. Some benefits include:

  • Fast Data Processing: Avro files can be read and written efficiently using Spark.
  • Schema Evolution: Avro supports schema changes over time without breaking compatibility.
  • Binary Encoding: Avro is a binary format, making it very compact.
  • Interoperability: Avro can be used with different programming languages.

Installing Necessary Libraries

Before diving into the code, ensure you have all the necessary libraries. PySpark does not come with built-in support for Avro files. You need to add the Avro package to your Spark installation:


pip install pyspark

Additionally, you will need to include the Avro package when initializing your Spark session. If your Spark is running locally:


from pyspark.sql import SparkSession

spark = SparkSession.builder \
    .appName("AvroExample") \
    .config("spark.jars.packages", "org.apache.spark:spark-avro_2.12:3.2.1") \
    .getOrCreate()

Replace “3.2.1” with your appropriate Spark version if necessary.

Loading Avro Files into DataFrame

Basic Example of Reading Avro File

Let’s start with a simple example of reading an Avro file into a DataFrame. Assume we have an Avro file named `data.avro` stored in the local filesystem:


from pyspark.sql import SparkSession

# Initialize Spark Session
spark = SparkSession.builder \
    .appName("AvroExample") \
    .config("spark.jars.packages", "org.apache.spark:spark-avro_2.12:3.2.1") \
    .getOrCreate()

# Read Avro file
df = spark.read.format("avro").load("data.avro")

# Show DataFrame
df.show()

Output of the DataFrame will look like this:


+---+-----+-------+
| id| name| salary|
+---+-----+-------+
|  1| John| 50000 |
|  2| Jane| 60000 |
|  3| Mike| 55000 |
+---+-----+-------+

Specifying Schema

Reading Avro files without defining a schema can be convenient but may not always be efficient. For more control over the data types and structure, you can specify a schema using PySpark’s `StructType`:


from pyspark.sql.types import StructType, StructField, IntegerType, StringType

# Define schema
schema = StructType([
    StructField("id", IntegerType(), True),
    StructField("name", StringType(), True),
    StructField("salary", IntegerType(), True)
])

# Read Avro file with schema
df = spark.read.format("avro").schema(schema).load("data.avro")

# Show DataFrame
df.show()

This method ensures that the data is read with explicit data types, reducing the possibility of schema inference errors.

Reading Avro from S3

In many big data applications, data is stored in cloud storage, such as Amazon S3. To read Avro files from S3, you need to provide the S3 path and configure the AWS credentials:


import os

# Set AWS credentials
os.environ['AWS_ACCESS_KEY_ID'] = 'YOUR_ACCESS_KEY'
os.environ['AWS_SECRET_ACCESS_KEY'] = 'YOUR_SECRET_KEY'

# Read Avro file from S3
df = spark.read.format("avro").load("s3a://your-bucket/data.avro")

# Show DataFrame
df.show()

Handling Compression

Avro files support various compression codecs like Snappy, Deflate, and Bzip2. By default, Spark handles compression transparently, but you can specify the codec if needed:


# Read Avro file with specified compression codec
df = spark.read.format("avro").option("compression", "snappy").load("data_compressed.avro")

# Show DataFrame
df.show()

Writing DataFrame to Avro

In addition to reading, you can also write DataFrames to Avro format. This can be useful for data storage or transfer:


# Sample DataFrame
data = [(1, "John", 50000), (2, "Jane", 60000), (3, "Mike", 55000)]
columns = ["id", "name", "salary"]
df = spark.createDataFrame(data, columns)

# Write DataFrame to Avro
df.write.format("avro").save("output_data.avro")

You can specify various options such as compression while writing:


# Write DataFrame to Avro with compression
df.write.format("avro").option("compression", "snappy").save("output_data_compressed.avro")

Conclusion

Reading and writing Avro files into DataFrames in PySpark is a seamless process once you configure the necessary libraries and Spark session settings. Avro’s compact format pairs well with Spark’s computational capabilities, providing a robust solution for big data processing. By following this guide, you can efficiently read and write Avro files, handle schema specifications, and deal with compressed data.

With these tools, you’re well-equipped to leverage Avro and Spark to handle large-scale data processing tasks efficiently.

About Editorial Team

Our Editorial Team is made up of tech enthusiasts deeply skilled in Apache Spark, PySpark, and Machine Learning, alongside proficiency in Pandas, R, Hive, PostgreSQL, Snowflake, and Databricks. They're not just experts; they're passionate educators, dedicated to demystifying complex data concepts through engaging and easy-to-understand tutorials.

Leave a Comment

Your email address will not be published. Required fields are marked *

Scroll to Top