Now you can Instantly Chat with Sathya!
About Me
Certified Horton works Hadoop Developer.
Good knowledge on Pyspark, AWS S3 and AWS Athena and EMR Services.
Good Knowledge on UNIX, Shell Scripting, Python and SQL.
Good Ex...
Good Knowledge on UNIX, Shell Scripting, Python and SQL.
Good Experience in Writing Hive, Impala Queries and assisting in optimizing the jobs.
Experience in Development, Maintenance & Support projects
Big Data Skills : HDFS, MapReduce, Pig, Hive, Impala, Sqoop, Apache Kudu & Spark
Programming Languages: C, C++, SQL, Shell Scripting, Python
Databases: Oracle, Ingress, MYSQL, PostgressSQL
Cloud Technologies : AWS S3, AWS Athena, EMR, RedShift
Show MoreSkills
Portfolio Projects
Description
Involved in Hadoop Framework design until Implementation.
Understanding the business functionality & Analysis of business requirements.
Conducts root cause analysis on systems and database issues throughout the life cycle of the project.
Implemented Adhoc Patient Process in Pyspark using RDD’s and Data Frames.
Written HQL’s from existing PL/SQL procedures for batch processing.
Migrated Patient Matching Batch process from PL/SQL to Hadoop Using Impala and apache Kudu.
Written python programs to read files from HDFS and perform some aggregations and move back to
other HDFS path.
Developed Hive and Impala UDF’s which are required for the process.
Worked on Migrating existing Oracle Historical data to Hadoop using Sqoop and Informatica Work flows.
Worked on Performance tuning of the Hadoop process on different tools such as Hive, Impala and for
storing Archival data.
Worked on different compression formats and decided best approach to store Archival and as well as
batch processing Data.
Migrated Oracle Data to AWS S3 using Spark.
Performing some analytical Queries from AWS Athena from CLI which is integrated in Shell Scripting.
Worked on POC’s to check the connectivity from Hadoop to ETL tools (Pentaho and Informatica BDE).
Provided best practices to improve the Cluster Performance.
Show More Show LessDescription
- Involved in Hadoop Framework design until Implementation.
- Understanding the business functionality & Analysis of business requirements.
- Design and migrate the Mainframe module to Hadoop using Impala, Sqoop and Shell script.
- Conducts root cause analysis on systems and database issues throughout the life cycle of the project.
- Written HQL’s from existing PL/SQL procedures for batch processing.
- Developed Hive and Impala UDF’s which are required for the process.
- Worked on Performance tuning of the Hadoop process on different tools such as Hive, Impala and for storing Archival data.
- Provided best practices to improve the Cluster Performance.
Description
- Involved in Hadoop Framework design
- Understanding the business functionality & Analysis of business requirements.
- Extensively used pig script, complex type, grouping, flattening etc.
- Loading data to Hive tables and writing queries to process.
- Administrating the 23 nodes Hadoop cluster.
- Loading datasets to HDFS and writing Map Reduce jobs to mine the data.
- Loading of data to HBase using pig and Hive.
- Responsible for loading, extracting and validation of client data.
- Installed/Configured/Maintained cloudera Hadoop clusters for application development and Hadoop tools like Hive, Pig, HBase, Zookeeper and Flume.
- Wrote the shell scripts to monitor the health check of Hadoop daemon services and respond accordingly to any warning or failure conditions.
- Implemented NameNode backup using NFS. This was done for High availability.
- Use of sqoop to import and export data from HDFS to RDBMS and vice-versa.
- Supported in setting up QA environment and updating configurations for implementing scripts with Pig and Sqoop.
Description
- Involvement in development and implementation of Configuration Management standards, procedures and guidelines
- Worked on nearly 6 Projects end to end in Oracle BRM including Upgrade and Agile Projects.
- Technical, Operational and Functional application maintenance which involves writing PL/SQL scripts to resolve issues and to extract data.
- Providing Estimates to the Projects from Environment support and maintaining all the environments up-to-date and assigning them to the respective project as per requirement.
- Automating the process using Shell Script and perl.
- Creating the Clear case view, Branches and merging the code.
- Creating new Oracle BRM environments on UNIX servers.
- Solving incidents and fulfilling the service requests and honoring all SLAs.
- Deploying the code changes to the UNIX environment from clear case and also applying Database changes.
- Monitoring, Status reporting of errors and taking proactive steps to avoid errors.
- Participating in and reacting to, quarterly Project Health Reviews and Submitting Project Metrics.