Dileep K.

Dileep K.

Big Data Engineer

New Delhi , India

Experience: 11 Years

Dileep

New Delhi , India

Big Data Engineer

38400 USD / Year

  • Immediate: Available

11 Years

Now you can Instantly Chat with Dileep!

About Me

·        Worked primarily as a Data Engineer, responsible for requirements analysis, design, coding and coordinate with team members and deliver project.

·        As a Big Data engineer, worked with business and engineering...

·        As a Big Data engineer, worked with business and engineering teams to define information needs and develop solutions that supports desired business and technical                 capabilities / requirements.

·        Extensively worked on creating logical and physical data models using SAP Power Designer.

·        Experienced in working with large data sets, and distributed computing. Extensively used Apache spark (Scala) and HIVE.

·        Experienced in extracting data from multiple structured and semi structured feeds (XML) by building and maintaining scalable ETL pipelines on distributed software                     systems.

·        Participated in developing applications using Informatica Power Center, SQL, Python, UNIX Shell scripting.

·        Responsible for creating Data Control Tool frameworks to ingest data into data lakes.

·        Our team is responsible for maintaining Consumer lending Data Mart.

·        Participated in onboarding new data sources both into Data Mart and Data Lakes.

·        Participated in 2 POC’s in data science related work. Identifying fraudulent transactions using SMOTE technique and Identifying loan defaulters.

Show More

Portfolio Projects

Description

Wells Fargo & Company is an American international banking and financial services holding company headquartered in San Francisco, California. It is the world's largest bank by market capitalization and the third largest bank in the U.S. by assets. This project Escheatment is the transfer of properties to the state if an account is deemed dormant for a specific amount of time (i.e. dormancy period). The rules regarding the identification of a dormant account and dormancy period vary by state and can change at any time. The regulation is to declare assets as dormant based on return mail and/or account contact.

· Worked primarily as a Data Architect, responsible for requirements analysis, design, coordinate with team members and deliver projects.

· I work as a Lead data engineer in the project and conduct meetings with all stake holders to gather requirements and create BRD (Business Requirements Document).

· Daily responsibilities include preparing Business Requirements Documents, Functional Specifications Document and pass on them to the developers for development.

· Extensively worked on Scala, Informatica Power Center ETL tool, Unix Shell scripting.

· Conduct design and code reviews of the team.

· Review QA and UAT tests to ensure all requirements are tested.

· Involved in all phases of the application and have complete knowledge of project.

· Leading a team of 10 members, working closely with US counterparts on daily deliverables and tracking the progress of the project.

Show More Show Less

Description

SMS Retirement project is to retire the existing functionality of SMS application. Secondary Marketing System is an application that pools and delivers the loans to the trade desk. Using this information investors buy these pooled loans from Wells Fargo. Entities such as Fannie Mae and Freddie Mac also purchase mortgages on a secondary market. This is part of data ingestion to data lakes. We worked in Consumer APP to load the data into data lakes.

· Worked primarily as a Data Architect, responsible for requirements analysis, design, coding and coordinate with team members and deliver project.

· As a Big Data engineer, worked with business and engineering teams to define information needs and develop solutions that supports desired business and technical capabilities / requirements.

· Extensively worked on creating logical and physical data models using SAP Power Designer.

· Experienced in working with large data sets, and distributed computing. Extensively used Apache spark (Scala) and HIVE.

· Experienced in extracting data from multiple structured and semi structured feeds (XML) by building and maintaining scalable ETL pipelines on distributed software systems.

· Participated in developing applications using Informatica Power Center, SQL, Python, UNIX Shell scripting.

· Responsible for creating Data Control Tool frameworks to ingest data into data lakes.

· Our team is responsible for maintaining Consumer lending Data Mart.

· Participated in onboarding new data sources both into Data Mart and Data Lakes.

· Participated in 2 POC’s in data science related work. Identifying fraudulent transactions using SMOTE technique and Identifying loan defaulters.

· Responsible for setting up L1/L2/L3 teams for Data Lake application.

· Conducted design and code reviews with the team.

· Responsible for creating Test plans and review test reports.

· Review UAT tests to ensure all requirements are tested.

· Involved in all phases of the application and have complete knowledge of project.

· Leading a team of 10 members, working closely with US counterparts on daily deliverables and tracking the progress of the project.

· Review the design documents, ETL code and QA test reports.

Show More Show Less

Description

I worked as a Tech Lead in Cognizant Technology Solutions for Kaiser which is a health care client. Kaiser Permanenteis anintegrated managed health careconsortium, based inOakland, California,United States. Kaiser offers Health Plans and has Kaiser Foundation Hospitals in 8 states of USA. This project is to load claims from 5 different regions into Claims Data warehouse. The Claims Data warehouse (CDW) program provides central, consistent industry standard view of finalized data at an atomic level for all the regions to perform reporting, analytical and operational functions. Source system is flat files and the target system is Oracle Database.

· Worked as ETL Lead in delivering code for National Tapestry module.

· Involved from design phase of the application and have complete knowledge of end to end architecture of the project.

· Leading a team of 6 members, worked closely with onsite on daily deliverables and tracking the progress of the project.

· Extensively worked on Informatica Power Center ETL tool, SQL, PL/SQL, Unix Shell scripting.

· Reviewing the ETL code and test reports. Taking leadership of defect analysis and helping the team members on day to day activities.

· Created UNIX wrapper scripts using UNIX shell scripting and automated Informatica workflows through these UNIX wrapper scripts.

Show More Show Less

Description

I worked as an ETL Developer at Infosys Ltd for Fidelity Investments client. GPS analyses the customer portfolio and helps them in taking informed decisions on investments, initially the customer eligibility for the Portfolio services are screened. On successful screening of the customer profiles, the portfolio is analyzed on the financial instrument held by the customer based on which several recommendations are made to the customer through email.

· Extensively worked on Data Extraction, Transformation, and Loading with RDBMS and Flat files.

· Good real time experience in tuning mappings, identifying and resolving performance bottlenecks in various levels like sources, targets, mappings, and sessions.

· Worked as an ETL developer in the project. Skills like Informatica, ETL, SQL, PL/SQL, Python scripting, UNIX shell scripting are involved in this project.

· Created unit test cases to validate the data and technical process involved in the entire data flow process.

· Involved in activities like requirements gathering, design, development, implementation of ETL processes using Informatica Power Center.

Show More Show Less