Now you can Instantly Chat with Atif!
About Me
- python
- pyspark
- airflow
- Linux/shell scripting
- AWS (S3, lambda, batch, ec2)
- HADOOP
- HIVE
- Data Engineering
- Data Orchestration
Skills
Positions
Portfolio Projects
Description
Technology:
Python, Pandas, Redshift, SPARK/pyspark, Meltano, api request, aws batch, aws lambda,
aws s3, aws ec2. airflow, snowflake
Responsibilities:
● Single-handedly responsible for extracting and loading data from different Platforms/ RDS like Mysql ,
postgres, api’s like Freshworks, surveysparrow, appsflyer , zendesk, clevertap etc
● Creating/Orchestrating data pipelines.
● Designing and implementing data ingestion pipelines involving different file formats like CSV json, parquet,
compressed files etc.
● Performing UT cases for developed codes ,logics and pipelines.
Description
Responsibilities:
● Created Data Quality Check pipeline using dagster.
● Worked on pipeline for ingesting different file formats like parquet , delimited , fixed width file to redshift.
● Brought down runtime of parquet file ingestion module by 10 times..
● Worked on a pipeline for loading data from Staging to Integration layer.
● Took active part in brainstorming sessions.