Now you can Instantly Chat with Ruban!
About Me
Having 4+ years of professional IT experience which includes experience in Big data, ETL and Mainframe related technologies.
Experience with software development process models like Agile and Waterfall methodologies.
Knowle...
Knowledge on Hadoop tools/frameworks, Sqoop, Spark core/PySpark, Hive.
Experience in insurance & health care domains (Have worked with State Farm and Anthem Inc.).
Used UNIX shell scripting to run the ETL flow to load LUW DB2 tables.
Good experience in creating real time data solutions using Apache Spark Core, Spark SQL.
Experience in using Avro, Parquet, ORC and JSON file formats.
Achieved Best Employee Award for my excellence in Development of Automation tool.
Shown Flexibility, quick learner, proactive attitude and amazing Team Player.
Show MoreSkills
Portfolio Projects
Description
Responsibilities: Worked closely with business customers for Requirement gatherings Developing Sqoop jobs with incremental load from heterogeneous RDBMS (IBM DB2) using native dB connectors Designed Hive repository with external tables, internal tables, partitions, ACID property and UDF for incremental data load of parsed data for analytical & operational dashboards Experienced in developing Hive Queries on different data formats like Text file, CSV file, Log files and leveraging time based partitioning yields improvement in performance using HiveQL Created Hive external tables for the data in HDFS and moved data from archive layer to business layer with hive transformations. Developed application using PySpark to determine the CSE Agents Worked on Revision management system to install the changes in production
Show More Show LessDescription
Responsibilities: Analysis, Requirement gathering, Planning & designing Preparing the unload scripts for the existing 36 Atomic tables. Preparing scp scripts for the required 36 unloaded files to Hadoop Creating Clojure application to clean and extract data from HDFS
Show More Show Less