Craft Recruiting
Craft Recruiting
...

Lead Data Platform Engineer

Job Title

Ingeniero de datos

Sector

Sector informático

Lugar

San Francisco, U.S.

Salario

Remuneración bruta 25000.00 - 35000.00 € Anual

Job Type

Indefinido

Horario de trabajo

Tiempo completo

Job Description

We are looking for our first fully-dedicated Data Platform Engineer to collaborate with our growing team of Data Scientists and Analysts. As the lead data platform engineer you will be responsible for architecting our next gen data processing applications and reporting systems.

Being the first dedicated data platform engineer, you will have the opportunity to build wholly new systems and services. We have some of the foundational elements in place for a modern data stack (Redshift, dbt, Segment, Kubernetes, Docker, Airbyte, Mode) and an active data science practice. We are looking for someone who can determine the next phase of the roadmap for the data platform and to lead us in building it -- for example, when do we want a different orchestration tool? When should we look to be less dependent on our current (3rd party) libraries for client app event tracking? We are looking to invest in ensuring that the core infrastructure on which all of our research depends is scalable, tested, robust, and able to evolve to meet our constantly changing data science needs.

We are looking for a lead engineer who enjoys being product-minded in the sense that they own a product from beginning to end by designing, constructing, integrating, testing, documenting, and supporting their creations.

Location: Cambridge, MA San Francisco, CA - NYC

Compensation: Includes competitive salary, company stock options and health benefits.

REMOTE UNTIL JANUARY 2022 then partial remote/office after.

As a lead data platform engineer you will be responsible for:

  • Designing, building, and supporting our next gen data processing applications and reporting systems through some combination of Python, Ruby, SQL, R, Go, and determining what tools we should build and what tools we should buy
  • Following software engineering development practices for building scalable and highly secure applications / services
  • Crafting optimal data processing architecture and systems for new data and ETL pipelines and driving the recommendation for improvements and modifications to existing data and ETL pipelines. (While you will certainly contribute to production ETL workflows, we expect you to spend more time building the tools and systems to enable data scientists, analysts, and other engineers to build the majority of the workflows.)
  • Collaborating with infrastructure teams to improve data processing CI / CD practices
  • Evangelizing high quality data engineering practices towards building data infrastructure, pipelines at scale and fostering the next-gen state of art data warehouse
  • Analyzing extremely large data sets (tens of millions to billions of records) to identify, evaluate and prioritize new opportunities to grow and optimize the business through analytics and data science


A strong candidate should be:

  • Highly proficient in SQL
  • Familiar with Python, Ruby, and/or Golang, and should have deep proficiency with at least one of those languages.
  • Experienced working with data pipelines in a cloud-native environment (bonus points for AWS experience)
  • Able to write, test, ship, and maintain clean production code within a collaborative and version-controlled (git) codebase.


Company

In case you offer a job
Crea una oferta de trabajo Create a company page