Ruban B.

Ruban B.

Software Developer

Manavalakurichi , India

Experience: 4 Years

Ruban

Manavalakurichi , India

Software Developer

19200 USD / Year

  • Notice Period: Days

4 Years

Now you can Instantly Chat with Ruban!

About Me

Having 4+ years of professional IT experience which includes experience in Big data, ETL and Mainframe related technologies.

 Experience with software development process models like Agile and Waterfall methodologies.

 Knowle...

 Knowledge on Hadoop tools/frameworks, Sqoop, Spark core/PySpark, Hive.

 Experience in insurance & health care domains (Have worked with State Farm and Anthem Inc.).

 Used UNIX shell scripting to run the ETL flow to load LUW DB2 tables.

 Good experience in creating real time data solutions using Apache Spark Core, Spark SQL.

 Experience in using Avro, Parquet, ORC and JSON file formats.

 Achieved Best Employee Award for my excellence in Development of Automation tool.

 Shown Flexibility, quick learner, proactive attitude and amazing Team Player.

Show More

Portfolio Projects

Description

Responsibilities:  Worked closely with business customers for Requirement gatherings  Developing Sqoop jobs with incremental load from heterogeneous RDBMS (IBM DB2) using native dB connectors  Designed Hive repository with external tables, internal tables, partitions, ACID property and UDF for incremental data load of parsed data for analytical & operational dashboards  Experienced in developing Hive Queries on different data formats like Text file, CSV file, Log files and leveraging time based partitioning yields improvement in performance using HiveQL  Created Hive external tables for the data in HDFS and moved data from archive layer to business layer with hive transformations.  Developed application using PySpark to determine the CSE Agents  Worked on Revision management system to install the changes in production

Show More Show Less

Description

Responsibilities:  Analysis, Requirement gathering, Planning & designing  Preparing the unload scripts for the existing 36 Atomic tables.  Preparing scp scripts for the required 36 unloaded files to Hadoop  Creating Clojure application to clean and extract data from HDFS

Show More Show Less