Data Engineer

Location: Postcode 1018 in Amsterdam Centrum

*** Mention DataYoshi when applying ***

DataChef is looking for an experienced and pragmatic Data Engineer (Scala/Python)

Jouw rol

We are looking for an experienced, pragmatic data engineer to join our team and help designing and building a green field data mesh platform on AWS.

The mission is building a scalable data foundation for one of the largest distribution companies in world; we are looking for A-players who are ready for the challenge and fun of creating such a unique foundation and stream processing backend using AWS services (Lambda, Glue, Lake formation,...), Spark (Scala/Python) and Kafka.

This is a consultant position so we are looking for an all-round candidate and not only an AWeSome engineer.

Jouw profiel

  • Have a real click with Our Core Values. If you’re nodding emphatically while reading it, you’ll probably fit right in, in which case, we can’t wait to hear from you. If your inner voice says bla bla bla while reading it, we want to save your and our time by not proceeding with interviews.
  • Are good consultants: have excellent communication skills to simplify and present concepts to other people. Show them how the future might be and help them participate in creating it. They don’t assume others know, so make everything (meetings, decisions, thoughts, code, etc) explicit & traceable.
  • Are doers, not talkers: we are a small team and our individual performance directly impacts the team outcome; so you need not only be taking initiatives but actually finishing what you’ve started. We are looking for a level 4+ problem solver.
  • Can demonstrate solid technical skills in Scala and/or Python (deep understanding of language internals, profiling, testing methods). Prior experience with ZIO is a big plus (DataChef is a contributor to the project ??)
  • Have 2+ years of hands-on experience with AWS services like Lambda functions, EMR, ElasticSearch, Lake formation and Glue.
  • Have a solid understanding (and preferably experience) of building pub-sub and asynchronous systems using Apache Kafka or any other messaging API like SQS, Kinesis, Celery, RabbitMQ, ActiveMQ.
  • Design & code defensively for the harsh real world and not for the happy path “Hello World” scenarios. They know missing, late and low-quality raw data is a fact and pipelines failures, replay/re-process are the norm, not a drama.
  • Can ingest new data sources (via REST APIs, file sharing etc) and deal with the ever-changing schemas.
  • Can analyze algorithms complexity and know data structures beyond “List and Stack” and pros/cons of using each for a problem.
  • Have been using Linux/macOS, Docker and git in collaborative workflows for 3+ years.
  • Are fast movers: Our culture is “Go, go and go faster”. Of course, you will break things by running fast which is understood and even appreciated. Just focus on learning fast and changing fast. And yes, we believe in agility and a distilled interpretation of agile manifesto

Nice to have

  • Have 5+ years of experience not only with developing greenfield projects from scratch but in an operational and live environment with strict high availability requirements.
  • Make quality a requirements issue: it is not enough to deliver something that works sometimes/maybe, we are building a mission-critical data platform. We love people who care about their craft and are proud about the quality of their code. Prior experience with the Great Expectations (DataChef is a contributor to the project ??) and/or Deequ is a big plus.
  • Write clean code that’s testable, maintainable, solves the right problem and does it well.
  • Know how to instrument their code for just-enough logging, better monitoring and easier debugging when it goes to production and operational environment.
  • Believe in DRY! They “Don’t Repeat Themselves”. Have allergies for any kind of waste, manual repetitive not-automated task; well… after manually doing it a few times!
  • Understand CAP theorem and also know how to design a resilient and partition tolerant service and its associated costs and tradeoffs.
  • Know about the latest developments in the Big Data community, combined with the ability to decide which of these are most relevant to our business and translate them into opportunities.
  • Contributed to open source: Make us happy with those green dots on Github!

Wat we bieden

  • Annual gross salary of €43,500 to €65,500, depending on your seniority and role in the kitchen!
  • 5% of DataChef's annual EBITDA will be distributed equally between outperforming chefs who have been with the company for 6+ months in the last year
  • If you have been with DataChef for less than 6 months you will recieve 50% of what other chefs get
  • Social security
  • Holiday allowance (25 days per year)
  • Reimbursement of commute costs
  • Company credit card (Bunq). You can spend up to €200/year for your trainings and software you didn't find in our shared resources
  • Paid and extended maternity/paternity leave (up to 4 weeks)
  • MacBook pro
  • Access to exclusive video and workshop trainings on AWS every month (on top of a, LinuxAcademy and other subscriptions that we have)
  • MasterClass membership
  • Holiday gift: at the end of the year, DataChef gives everyone and their family a surprise thank-you gift
  • Charitable Gifts Match: When you donate money to a qualified charity, DataChef will make a matching gift to that charity on your behalv (up to €500/year)
  • 100% of Hardware & Software: Whatever hardware, software, or services you need to do your job are always 100% on us. No red tape, no questions asked.

Over ons

  • DataChef: We're small, profitable, self-funded and growing company based in Amsterdam. If you believe that data can (and must) change the quality of life in companies (and humans who run those companies) then you will find yourself in your tribe here at DataChef.
  • We are a consultancy/agency, focused on develop and delivering top quality projects on Big Data and machine learning and on AWS platform.
  • Behind the scenes, we are working on a SaaS product and aim at becoming a 100% product company somewhere in 2023.
  • 100% Open and Transparent company: Our role models are not giant corporates but relatively small yet happily successful companies like Basecamp, Buffer and ahrefs. We even share our salaries and benefits publicly - in case you're curious:
    • Salaries and Benefits (in Europe ????)
  • We thrive on technical excellence by hiring only the best and seeing ourselves at the beginning of the same success path as Databricks and Elasticsearch but only 10 years younger!
  • DataChef is an Equal Opportunity Employer – Minority / Female / Disability / Veteran / Gender Identity / Sexual Orientation.

*** Mention DataYoshi when applying ***

Offers you may like...

  • Grubhub Holdings, Inc.

    Senior Data Engineer
    New York, NY 10018
  • Splunk

    Data Engineer, DevOps - Remote OK
    Sun Prairie, WI 53590
  • Cognizant Technology Solutions

    Senior Data Engineer (Remote) - Advanced Technolog...
    Chicago, IL 60601
  • Lessonly

    Senior Data Engineer
    Indianapolis, IN 46202
  • Numerator

    Sr. Data Engineer
    Chicago, IL 60606