Snehal B.

Snehal B.

Big Data Developer with expertise in Spark/Scala

Mumbai , India

Experience: 5 Years

Snehal

Mumbai , India

Big Data Developer with expertise in Spark/Scala

27401.1 USD / Year

  • Notice Period: 60 Days

5 Years

Now you can Instantly Chat with Snehal!

About Me

Data engineer having 5+ years of experience in Big Data skills majorly on Spark with Scala. I had worked on variety of log analytics tools like ELK stack, Flume, Sqoop, Hive. Also have expereince working on cloud with AWS Lambda and EC2.

...

Show More

Portfolio Projects

Log analytics and near real time visualization and alerting using Hadoop & ELK stack

Company

Log analytics and near real time visualization and alerting using Hadoop & ELK stack

Role

Backend Developer

Description

 Contributed in enhancing the business processes within TCS organization by providing Log Analytics

using Hadoop and other big data tools. This saved lot of efforts and gained huge profits for organization.

  • Configured multiple log collectors like Flume, Logstash, Filebeat, Nxlog to collect logs from various sources.
  • Developed Hive scripts to process the data stored in HDFS. Also adopted optimization techniques to lessen the data processing time and achieve results faster.
  • For providing real time visualization of data used Elasticsearch NOSQL database and Kibana as visualization layer on top of it.
  • To notify user about 3 consecutive failure login attempts on a server, configured ElastAlert tool on top of Elasticsearch and send alert to user on occurrence of that event.

Show More Show Less

Predictive Modelling of Server Utilization using R

Company

Predictive Modelling of Server Utilization using R

Role

Backend Developer

Description

Deployed R for providing prediction values of several server parameters like CPU,Mem, Disk,etc.  This was useful for server administrators to know utilization of their servers in upcoming month.

  • Retrieved server utilization data from MySQL by integrating R with MySQL database.
  • After some fine tuning data in R, fed it to the ARIMA model in R to obtain predictions for same.

Show More Show Less

Skills

R Language MySQL

IS Data Warehouse

Company

IS Data Warehouse

Role

Backend Developer

Description

   Centralization of employee specific data in order to get all relevant information for a employee at single place. Performed data ingestion in hadoop from multiple data sources. Facilitated insightful daily analysis by comparing multiple datasets for various use cases like asset tracking, associates allocation,etc.

  • Developed Sqoop script to import data from various databases like MySQL, Postgres, Oracle, MS-SQL,etc in HDFS. Performed merging of daily incremental data with existing data using sqoop merge.
  • Created Hive queries to fine tune imported data and join it with other datasets and executed it using Spark-SQL.
  •  Provided design recommendations and thought leadership to other stakeholders that improved review processes and resolved technical problems.
  • Shared responsibility for administration of Hadoop, Hive, Spark.

Show More Show Less

Tools

sqoop putty Hive

Data Analytics

Company

Data Analytics

Role

Backend Developer

Description

This project was in BFSI domain. Client was Canadian multinational financial services organization. This project was in migration phase from Oracle based architecture to Big Data Platform.

  • Developed Scala code for processing data stored in hive tables using Spark. Processed data was used for report generation in Tableau for client.
  • For processing Capital Market data used Spark-SQL functions in scala code.

Show More Show Less

Skills

Apache Spark

Data Lake

Company

Data Lake

Role

Backend Developer

Description

This project was part of Customer Engagement Platform(CEP) initiative for healthcare client (American multinational biopharmaceutical company). They were creating data lake for centralizing all product, customer, professional and sales data.

  • For processing data stored in Hive, developed code using Snaplogic tool.
  • Implemented complex logic using Snaplogic’s limited functionality tools.

Show More Show Less

Tools

Snaplogic

Enterprise Analytics

Company

Enterprise Analytics

Role

Backend Developer

Description

This Project was for technology client.

  • Developed Scala code to process data stored in Hive.
  • Reduced BI layer downtime by performing hive optimization.
  • Created Talend jobs to launch Spark code in cluster for data processing.

 

Show More Show Less

Workflow Automation

Company

Workflow Automation

Role

Backend Developer

Description

This Project was for American architecture software client.

  • Tested AWS Sagemaker data science code in AWS EC2 instance.
  • Reduced code runtime using EC2.
  • Automated start/stop EC2 instance through Oozie-Jenkins workflow.

Show More Show Less

Global Piracy Analytics

Company

Global Piracy Analytics

Role

Backend Developer

Description

This Project is for American architecture software client to detect piracy of their product and convert them into legitimate users.

  • Developing scala code to detect pirate users and run the code on Spark cluster..
  • Automate the workflow using Oozie-Jenkins.

Show More Show Less

Tools

Qubole Jenkins