Now you can Instantly Chat with PRAFULLA KUMAR!
About Me
8+ Years(Total) of extensive IT experience in all phases of Software Development Life Cycle (SDLC) with skills in data analysis, design, development, testing and deployment of software systems for client/server multi-user business applications. 4+ Ye...testing and deployment of software systems for client/server multi-user business applications. 4+ Years of strong experience, working on Apache Hadoop ecosystem components like MapReduce, HDFS, HBase, Hive, Sqoop, Pig, Oozie, Zookeeper, Flume, Spark, Python with CDH4&5 distributions and EC2 cloud computing with AWS. Working closely with the stakeholders & solution architect, Ensuring architecture meets the business requirements, Building highly scalable, robust & fault-tolerant systems. Key participant in all phases of software development life cycle with Analysis, Design, Development, Integration, Implementation, Debugging, and Testing of Software Applications in client server environment Strong in Developing MapReduce applications, Configuring the Development Environment, Tuning Jobs and Creating MapReduce Workflows. Experience in performing data enrichment, cleansing, analytics, aggregations using Hive and Pig. Knowledge in Cloudera CDH4 and CDH5 distributions and Hortonworks (HDP Proficient in big data ingestion and streaming tools like Flume, Sqoop, Kafka, Storm and Kinesis Experience with different data formats like Json, Avro, parquet, RC and ORC and compressions like snappy & bzip. Experienced in analyzing data using HQL, PigLatin and extending HIVE and PIG core functionality by using custom UDFs. Good Knowledge/Understanding of NoSQL data bases and hands on work experience in writing applications on NoSQL databases like Hbase and MongoDB. Good knowledge on various scripting languages like Linux/Unixshell scripting and Python. Good knowledge of Dataware housing concepts and ETL processes. Importing / exporting data from RDBMS to HDFS for batch data process using SQOOP Configured Zookeeper to coordinate the servers in clusters to maintain the data consistency. Used Oozie and Control - M workflow engine for managing and scheduling Hadoop Jobs. Diverse experience in working with variety of Database like Oracle, MySql, Salesforce and Netezza. AWS provides a secure global infrastructure, plus a range of features that use to secure the data in the cloud Hands on experience on AWS cloud services (VPC, EC2, S3, RDS, Redshift, Data Pipeline, EMR, DynamoDB, WorkSpaces, Lambda, Kinesis, SNS, SQS) Good experience of AWS Elastic Block Storage (EBS), different volume types and use of various types of EBS volumes based on requirement. Ability to spin up different AWS Instances including EC2-Classing and EC2-VPC using cloud formation template Cognitive about designing, deploying and operating highly available, scalable and fault tolerant systems using Amazon Web Services (AWS). With the help of IAM created roles, users and groups and attached policies to provide minimum access to the resources, created topics in SNS to send notifications to subscribers as per the requirement. Implemented Amazon RDS multi-AZ for automatic failover and high availability at the database tier, created CloudFront distributions to serve content from edge locations to users so as to minimize the load on the frontend servers. Experienced in Performance Tuning and Query Optimization in AWS Redshift. Implemented POC to migrate map reduce programs into Spark transformations using Spark and Scala. Good knowledge in understanding Core Java and J2EE technologies such as Hibernate, JDBC, EJB, Servlets, JSP, JavaScript, Struts and spring. Experienced in using IDEs and Tools like Eclipse, Jenkins, Maven and IntelliJ. Familiar with data architecture including data ingestion pipeline design, Hadoop information architecture, data modeling and data mining, machine learning and advanced data processing. Experience optimizing ETL workflows. Experience on handling cluster when it is in Safe mode. Good knowledge of High-Availability, Fault Tolerance, Scalability, Database Concepts, System and Software Architecture, Security and IT Infrastructure. All the projects which I have worked for are Open Source Projects and has been tracked using JIRA. Load and transform large sets of structured, semi-structured and unstructured data using Hadoop ecosystems Good knowledge of single node and multi-node cluster setup Strong experience with good knowledge in SQL (Including Triggers, Stored Procedures). Experience in writing small to complex queries. Lead onshore & offshore service delivery functions to ensure end-to-end ownership of incidents and service requests. Getting in touch with the Junior developers and keeping them updated with the present cutting Edge technologies like Hadoop, Spark, SparkSQL. Highly motivated, subject oriented, has ability to work independently and as a part of the team with Excellent Technical, Analytical and Communication skills. Strong team player, ability to work independently and in a team as well, ability to adapt to a rapidly changing environment, commitment towards learning, Possess excellent communication, project management, documentation, interpersonal skills.
Show MorePortfolio Projects
Description
The crux of the Reconciliation Application is to perform matching between various feeds and to ensure that the data is consistent without any discrepancies in retail banking. The goal of this application is to describe a mechanism which allows providing an end to end solution on reconciliations to financial institutions, and to handle end to end recon process, starting from file retrieval, data transformation, and reconciliation to exception enrichment and resolution.
Show More Show LessDescription
In my team, we were responsible for front-end UI and Stored procedure enhancement certain code changes was happening in a periodical basis as per AHA (DRG Codes) so we need to change all the necessary code change for the incoming and Inter patients so that it could be effect to their bill to respect the AHA and Insurance process flow
Show More Show LessDescription
Synechron and Me:
- Worked for Innovation project (Fin-Labs)
- In Synechron Fin-lab I involved from beginning to deployment (on AWS) for all accelerators
- We are working for 15 + financial Accelerators for our external client like MSCI, JPMC HSBC more
- We involved Artificial Intelligence for all of our accelerators
- We used NLP/NLG Jflex, Hadoop, Spark, R for Data Extraction and process/Compute
Description
As per American Health Association (AHA) there is fraud of 5 million dollars annually in health insurance in US. This project is to identify possible fraudulent claims out of the total claims processed daily. We receive insurance claim data (OLTP) in 11 different files from auto adjudication system in X12 format with fixed layout format. We load the data into HDFS and have written multiple map reduce jobs to convert the X12 format into CSV format and load the data into Hive after creating tables. Hive join queries are used to fetch information from multiple tables. Query output is populated into temporary tables to perform more complex joins.
Show More Show LessVerifications
-
Profile Verified
-
Phone Verified
Preferred Language
-
English - Fluent
-
Hindi - Fluent
Available Timezones
BROWSE SIMILAR DEVELOPER
-
KULJIT S
Corporate Mgmt. – PMG, HR, legal, MIS/ERP
-
Daniel M
Mainframe Developer/Senior QA Analyst
-
Steven T
Have coded almost everything from firmware through apps, dev to valid to customer suppport
-
Mikhail B
Seniors Micro-Services consultant
-
James B
Managing Director, Sr PM, Consultant
-
James M
• Infrastructure Management/Project Mangement
-
RICHARD V
Electronics Circuit Design Engineer
-
Fred L
Executive Consultant
-
Jeffrey L
Cisco Network and Voice Engineer with heavy Linux / C / PHP / Python programming background
-
MICHAEL H
SENIOR MANUFACTURING TEST ENGINEER