About Me
Having 4+ years of professional IT experience which includes experience in Big data, ETL and Mainframe related technologies.
Experience with software development process models like Agile and Waterfall methodologies.
Knowle...
Knowledge on Hadoop tools/frameworks, Sqoop, Spark core/PySpark, Hive.
Experience in insurance & health care domains (Have worked with State Farm and Anthem Inc.).
Used UNIX shell scripting to run the ETL flow to load LUW DB2 tables.
Good experience in creating real time data solutions using Apache Spark Core, Spark SQL.
Experience in using Avro, Parquet, ORC and JSON file formats.
Achieved Best Employee Award for my excellence in Development of Automation tool.
Shown Flexibility, quick learner, proactive attitude and amazing Team Player.
Show MoreSkills
Software Engineering
Web Development
Data & Analytics
Programming Language
Database
Development Tools
Operating System
Others
Positions
Portfolio Projects
Company
Customer Service Expectations
Description
Responsibilities: Worked closely with business customers for Requirement gatherings Developing Sqoop jobs with incremental load from heterogeneous RDBMS (IBM DB2) using native dB connectors Designed Hive repository with external tables, internal tables, partitions, ACID property and UDF for incremental data load of parsed data for analytical & operational dashboards Experienced in developing Hive Queries on different data formats like Text file, CSV file, Log files and leveraging time based partitioning yields improvement in performance using HiveQL Created Hive external tables for the data in HDFS and moved data from archive layer to business layer with hive transformations. Developed application using PySpark to determine the CSE Agents Worked on Revision management system to install the changes in production
Show More Show LessTools
GitCompany
Data migration for pceuidat & PCAS from FDWD2.0 to Hadoop
Description
Responsibilities: Analysis, Requirement gathering, Planning & designing Preparing the unload scripts for the existing 36 Atomic tables. Preparing scp scripts for the required 36 unloaded files to Hadoop Creating Clojure application to clean and extract data from HDFS
Show More Show LessTools
Git