Senior Data Engineer | India Remote | Online Jobs | Optimhire

Senior Data Engineer | India Remote

INTRODUCTIONDashboards are where data goes to die. If you’ve ever built a dashboard, you’ve visited its graveyard. Today’s businesses are generating data faster than ever before, and while leaders want to leverage this data effectively, it's clear that automation is needed to fully understand and utilize such large amounts of data. Unfortunately, the existing solution – more dashboards – isn’t going to cut it. Dashboards are checked occasionally, tell a stale static story, and come with no guarantees.This is a great opportunity to build from the ground up. We are looking for founding team members to build a generation defining product in this huge market. We believe diversity and perspective drive innovation and we are building a culture where you drive meaningful action and difference is valued. As a member of the team, you'll have the opportunity to build this revolutionary platform and culture with us.
THE ROLEJoin our early-stage startup, as our Senior Data Engineer to drive the customer experience of a new category-creating and ground-breaking product in the Data Analytics space. The ideal candidate is a passionate, yet pragmatic thinker and maker that can think broadly across platforms and domains, but then dive deep into the details. You should be passionate about working strategically with executives and other senior leaders to create and deliver big, visionary and customer-focused experiencesAs a member of the v1.0 launch team, you will shape this entire product from scratch backed by a strong product aesthetic. This is as “early ideation and creation phase” of a startup as it gets – as autonomous as it gets – as FUN as it gets – if you thrive on the energy of creating something out of nothing.RESPONSIBILITIES

  • Coding, scripting, templating with Python/Scala and SQL
  • Building production solutions with services on public cloud and open source
  • You will build and own large additions to our data engineering framework, charged with finding ways to create and improve scalable and reliable tables and central data pipelines
  • You will build and own scalable, efficient, and well-tested data engineering solutions using Spark, Druid, and a mature collection of open source technologies.
  • You will be involved in the day-to-day operations of the team, including maintaining and improving our scalable production pipelines, current tools & scripts and supporting full-stack data scientists
  • You will have autonomy to help shape the future of data engineering by bringing your ideas on improving and automating what we do and how we do it
  • In-depth knowledge and experience in creating a proof-of-concept with the help of necessary scripting languages and other tools for testing the concepts in practice
  • Close collaboration with internal stakeholders world-class data scientists, analysts and engineers on how to solve data and business problems in a scalable way
  • Working on overlap between AI, Software and Data Engineering.
  • A passionate technologist with a brilliant growth hacker mindset
  • Experience in building out scalable data engineering capabilities and creating data pipelines for Internet companies’ data scale (millions of users/billions of events daily).
  • Good knowledge of agile development processes as well as experience in the close integration of development and operation in the cloud (DevOps, CI/CD, CT)
  • Experience with modern tools for collaboration in an agile environment (e.g., GitHub, JIRA, Confluence)
  • Work in global teams, across time zones and be ready to support customer deployments of software you create.
  • A code craftsman who likes to follow the scout rule “Always leave the campground cleaner than you found it” while writing good quality code, with best practices for design documentation.
  • Proficient in cleaning and wrangling the data while and enjoy working on distributed data processing tools (e.g., Spark, Flink, Dask)
  • Not just be familiar with AWS and potentially also GCP and Azure. understands the importance of cloud and know the way around it and want to learn that.
  • Experience with workflow management tools (e.g., Airflow, Metaflow, Prefect).
  • Excited to make customers successful and feel pride of ownership in the solutions enabled by your work
  • Experience in working autonomously and taking ownership of projects.
  • Ability to think globally, devising and building solutions to meet many needs rather than completing individual projects or task
Nice to Have
  • Experience with stream processing technologies (e.g., Kinesis, Kafka).
  • Familiar with profiling code and performance optimizations.
  • Open-source contributions and/or personal software projects.
  • Self-motivated and comfortable dealing with the ambiguity of V1 and incubation work.
  • Using a startup mentality to discover and utilize technologies to rapidly bring the project to life.
Tech StackPythonSQL and NoSQL databases (Redshift, DRUID, MySQL, Postgres, MongoDB etc.)Airflow, ZookeeperAWS (most important), GCP, AzureDocker, KubernetesMicroservices for APIsCICD, test driven development
PREFERRED QUALIFICATIONSAid in attracting and growing an emerging and talented engineering team

Job Type



Data Scientists

Must have Skills

  • Python - 5 Years
  • Scala
  • SQL - 5 Years
  • AWS - 3 Years
  • Docker - 3 Years
  • Microservices
  • CI/CD
  • Zookeeper
  • Data Engineer - 5 Years


english - Fluent

2 - 37 K/Year USD (Annual salary)

Longterm (Duration)

Fully Remote


Avantika Jadhav