loader image

Solutions Architect - Data Engineering

As a Solution Architect on our Data Engineering Team, your responsibilities will include:


  • Design, develop, and innovative Hadoop solutions; partner with our internal Infrastructure Architects and Data Engineers to build creative solutions to tough big data problems.  

     

  • Determine the project road map, select the best tools, assign tasks and priorities, and assume general project management oversight for performance, data integration, ecosystem integration, and security of big data solutions

     

  • Work across a broad range of technologies – from infrastructure to applications – to ensure the ideal Hadoop solution is implemented and optimized

     

  • Integrate data from a variety of data sources (data warehouse, data marts) utilizing on-prem or cloud-based data structures (AWS); determine new and existing data sources





  • Design and implement streaming, data lake, and analytics big data solutions


  • Create and direct testing strategies including unit, integration, and full end-to-end tests of data pipelines


  • Select the right storage solution for a project - comparing Kudu, HBase, HDFS, and relational databases based on their strengths


  • Utilize ETL processes to build data repositories; integrate data into Hadoop data lake using Sqoop (batch ingest), Kafka (streaming), Spark, Hive or Impala (transformation)


  • Partner with our Managed Services team to design and install on prem or cloud based infrastructure including networking, virtual machines, containers, and software


  • Determine and select best tools to ensure optimized data performance; perform Data Analysis utilizing Spark, Hive, and Impala


  • Mentor and coach Developers and Data Engineers. Provide guidance with project creation, application structure, automation, code style, testing, and code reviews



Qualifications



  • 5+ years previous experience as a Software Engineer, Data Engineer or Data Analytics


  • Expertise in core Hadoop technologies including HDFS, Hive and YARN.  


  • Deep experience in one or more ecosystem products/languages such as HBase, Spark, Impala, Solr, Kudu, etc


  • Expert programming experience in Java, Scala, or other statically typed programming language


  • Ability to learn new technologies in a quickly changing field


  • Strong working knowledge of SQL and the ability to write, debug, and optimize distributed SQL queries


  • Excellent communication skills including proven experience working with key stakeholders and customers


  • Ability to translate “big picture” business requirements and use cases into a Hadoop solution, including ingestion of many data sources, ETL processing, data access and consumption, as well as custom analytics


  • Experience scoping activities on large scale, complex technology infrastructure projects


  • Customer relationship management including project escalations, and participating in executive steering meetings



Position

Software Architect


Must have Skills

  • SQL

    Beginner

  • Scala

    Beginner

  • Java (All Versions)

    Beginner

  • Apache Spark

    Beginner

  • HBase

    Beginner

  • Hadoop

    Beginner

Client Payroll

Up to 450 K/Year USD (Annual salary)

Fully Remote

Cancel
Cancel

Active

Skip

Solutions Architect - Data Engineering

As a Solution Architect on our Data Engineering Team, your responsibilities will include:


  • Design, develop, and innovative Hadoop solutions; partner with our internal Infrastructure Architects and Data Engineers to build creative solutions to tough big data problems.  

     

  • Determine the project road map, select the best tools, assign tasks and priorities, and assume general project management oversight for performance, data integration, ecosystem integration, and security of big data solutions

     

  • Work across a broad range of technologies – from infrastructure to applications – to ensure the ideal Hadoop solution is implemented and optimized

     

  • Integrate data from a variety of data sources (data warehouse, data marts) utilizing on-prem or cloud-based data structures (AWS); determine new and existing data sources





  • Design and implement streaming, data lake, and analytics big data solutions


  • Create and direct testing strategies including unit, integration, and full end-to-end tests of data pipelines


  • Select the right storage solution for a project - comparing Kudu, HBase, HDFS, and relational databases based on their strengths


  • Utilize ETL processes to build data repositories; integrate data into Hadoop data lake using Sqoop (batch ingest), Kafka (streaming), Spark, Hive or Impala (transformation)


  • Partner with our Managed Services team to design and install on prem or cloud based infrastructure including networking, virtual machines, containers, and software


  • Determine and select best tools to ensure optimized data performance; perform Data Analysis utilizing Spark, Hive, and Impala


  • Mentor and coach Developers and Data Engineers. Provide guidance with project creation, application structure, automation, code style, testing, and code reviews



Qualifications



  • 5+ years previous experience as a Software Engineer, Data Engineer or Data Analytics


  • Expertise in core Hadoop technologies including HDFS, Hive and YARN.  


  • Deep experience in one or more ecosystem products/languages such as HBase, Spark, Impala, Solr, Kudu, etc


  • Expert programming experience in Java, Scala, or other statically typed programming language


  • Ability to learn new technologies in a quickly changing field


  • Strong working knowledge of SQL and the ability to write, debug, and optimize distributed SQL queries


  • Excellent communication skills including proven experience working with key stakeholders and customers


  • Ability to translate “big picture” business requirements and use cases into a Hadoop solution, including ingestion of many data sources, ETL processing, data access and consumption, as well as custom analytics


  • Experience scoping activities on large scale, complex technology infrastructure projects


  • Customer relationship management including project escalations, and participating in executive steering meetings



Job Type

Client Payroll


Positions

Software Architect


Must have Skills

  • SQL

    Beginner

  • Scala

    Beginner

  • Java (All Versions)

    Beginner

  • Apache Spark

    Beginner

  • HBase

    Beginner

  • Hadoop

    Beginner

Up to 450 K/Year USD (Annual salary)

Longterm (Duration)

Fully Remote

Skip

Brandon S

| United States