Featured Post

8 Ways to Optimize AWS Glue Jobs in a Nutshell

Image
  Improving the performance of AWS Glue jobs involves several strategies that target different aspects of the ETL (Extract, Transform, Load) process. Here are some key practices. 1. Optimize Job Scripts Partitioning : Ensure your data is properly partitioned. Partitioning divides your data into manageable chunks, allowing parallel processing and reducing the amount of data scanned. Filtering : Apply pushdown predicates to filter data early in the ETL process, reducing the amount of data processed downstream. Compression : Use compressed file formats (e.g., Parquet, ORC) for your data sources and sinks. These formats not only reduce storage costs but also improve I/O performance. Optimize Transformations : Minimize the number of transformations and actions in your script. Combine transformations where possible and use DataFrame APIs which are optimized for performance. 2. Use Appropriate Data Formats Parquet and ORC : These columnar formats are efficient for storage and querying, signif

A Beginner's Guide to Pandas Project for Immediate Practice

Pandas is a powerful data manipulation and analysis library in Python that provides a wide range of functions and tools to work with structured data. Whether you are a data scientist, analyst, or just a curious learner, Pandas can help you efficiently handle and analyze data. 


Simple project for practice


In this blog post, we will walk through a step-by-step guide on how to start a Pandas project from scratch. By following these steps, you will be able to import data, explore and manipulate it, perform calculations and transformations, and save the results for further analysis. So let's dive into the world of Pandas and get started with your own project!


Simple Pandas project

Import the necessary libraries:


import pandas as pd

import numpy as np


Read data from a file into a Pandas DataFrame:


df = pd.read_csv('/path/to/file.csv')

Explore and manipulate the data:


View the first few rows of the DataFrame:


print(df.head())


Access specific columns or rows in the DataFrame:


print(df['column_name'])

print(df.iloc[row_index])


Iterate through the DataFrame rows:


for index, row in df.iterrows():

    print(index, row)


Sort the DataFrame by one or more columns:


df_sorted = df.sort_values(['column1', 'column2'], ascending=[True, False])


Perform calculations and transformations on the data:


df['new_column'] = df['column1'] + df['column2']


Save the manipulated data to a new file:

df.to_csv('/path/to/new_file.csv', index=False)

Remember to adjust the file paths and column names based on your project requirements. These steps provide a basic starting point for a Pandas project and can be expanded upon depending on the specific task or analysis you're working on.


Data sources for CSV files

Comments

Popular posts from this blog

How to Fix datetime Import Error in Python Quickly

How to Check Kafka Available Brokers

SQL Query: 3 Methods for Calculating Cumulative SUM