Now you can Instantly Chat with Sunilkumar!
Having 9.5 years of experience in Information Technology with Technical and Functional experience in maintaining data warehouses/marts. Proficiency in Azure Data Factory and Tools, Azure Cloud components, Data warehousing, ETL process, OLAP system...
· Extensively used ELT methodology for Data Extraction, Loading and Transformation process in a corporate-wide Solution using Azure Data Factory,Data Lake, Databricks Delta Lake,Azure Sqldb,USQL ,Databricks,Pyspark and Powershell, AzureData Storage Explore components.
· Azure Storage component, Data Lake Store(Gen1,Gen2), Access policy, Container/Blob Access & Security Audit for Different operation.
.ETL Tool Informatica ,Unix ,Teradata MPP System
· AAD group Creation ,RBAC role on access provisioning and Security
Data & Analytics
Health care project
Health care project
This project delivered the initial set of Client OnPrem and Sales Force data feeds into the Azure Data Lake.
Working with fellow Microsoft partners, with the Business Intelligence/Analytics team, responsible for BI development and the surrounding infrastructure for the UK, USA, Brazil ,APAC ,Australia and South Africa.
Involved in end to end implementation like requirement gathering ,Functional discussion, Data Ingestion ,Data set & frequency setup documentation ,building ETL /ELT Logic for business users and the creation of front end reporting with mobile compatible dashboards using Power BI.
· Actively involved in developed pipelines using Azure Data Factory to ingest data from One Prem(File system),Sales Force Data, amongst others into the ADLS and Different Storage Account.
· These pipelines were delta enabled and had various activities (copy and Transformation) as the data moved from source to target. Worked with data frames to ingest, transform, analyze and help Front end team to visualize data.
· Staging – ELT stage layer where the data is loaded directly from Source files. In this step data are loaded from source files.
· Hierarchy File Build on Top of Dim Data– Pre-defined Files build on staging layer with user-friendly column names.
· In Pre-Core or Collated Layer ,The business logic and Transformation using Databricks,Pyspark and some case USQL Jobs & PowerShell script.
· PowerShell script creation for Storage Access policy check and Access provisioning, ARM template deployment.
· Code repository for the different pipelines, Scripts being maintained using VSTS.
· Involved in Performance optimization in USQL job using Index and Partitioning, applying required filter and conditional Logic.
· Used Assembly in USQL for Diagnostic Log analysis. Pipeline being developed to Track ADLS user access, Storage account user metrics and Power bi report being generated as per user requirement.Show More Show Less
This project was started with the existing OnPrem Sqlserver,File based data feeds into the Azure Data Lake first using in ADF v1.
Post ADF V2 GA , the existing jobs with the enhancement or change requests as required by business move to V2 version with Data Lake Store Gen2 & Databricks From Gen1 phase wise make sure client production system not impacted.
Client has it’s exist mixed datasets of relational or File based like Asset,Course offered ,Topic,Usages,Skillset details which all for learning prospect in existing on prem system which need to be moved to Cloud with required transformation.Show More Show Less