Skip to content

Exploration of two important strategies to make our data analysis faster and independent of the dataset size.

Notifications You must be signed in to change notification settings

teroxrr/Optimizing-Dataframes-and-Processing-in-Chunks

Repository files navigation

Optimizing Dataframes and Processing in Chunks

In this project we will explore two very important strategies for handling large datasets and dealing with limited hardware resources:

  • Optimizing Memory Footprint
  • Processing the Data in Chunks

As we will see, the combination of these two powerful approaches will significantly improve our access to the data, independently of its size.

Head over to Optimizing Dataframes and Processing in Chunks.ipynb to start reading.

Requirements

For this project we implemented the following tools:

  • Python
  • Pandas

About

Exploration of two important strategies to make our data analysis faster and independent of the dataset size.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published