Data Engineer

Company:
Location: New York, NY

*** Mention DataYoshi when applying ***

Data Engineer - Hearst Newspapers

With more than 3,000 employees across the nation, Hearst Newspapers publishes 24 dailies and 52 weeklies, including the Houston Chronicle, San Francisco Chronicle, San Antonio Express-News and Albany Times Union. It also operates digital marketing services and directories businesses under the LocalEdge brand.

We are looking for driven individuals to join our team of passionate data engineers and data scientists in creating Hearst Newspaper’s next generation of data products and capabilities. Candidates should possess a deep knowledge of modern data processing technology stacks and database systems including SQL/NoSQL, alternatively a strong background in software engineering is also preferred. We are a small and efficient team building scalable data pipelines on cloud platform for data analysis and reporting process and providing a solution with machine learning and data mining techniques.

Responsibilities:

Build reusable products and systems that can effectively ingest data to leverage across technology, business intelligence, and advertising operations teams.

Support new feature developments between data sources for our own ingestion, reporting, and analysis platforms.

Collaborate with data analysts and project managers to identify and use the data that is most relevant to the problem.

Develop, deploy, and maintain the data products and systems.

Work closely with data scientists to transform data and integrate algorithms and models into the production system.

Qualifications:

Bachelor’s degree in Computer Science or equivalent practical experience. Master’s degree a plus.

3+ years of solid experience in building ETL/ELT pipelines. Familiarity with Airflow is a plus

Fluent in Python and SQL. Excellent working knowledge with relational databases.

Knowledge with object-oriented or functional programming skills.

Experience with bash shell scripts, UNIX/Linux commands.

Experience with cloud platform offerings. Google GCP preferred.

Ability to quickly learn modern data technologies.

Experience with traditional agile methodology.

Bonuses:

Natural Language Processing (NLP) and/or Machine Learning(ML).

Knowledge of Big Data technologies such as Spark, Kafka, and Apache Solr

Experience with BI tools like Looker and PowerBI.

Experience with CI/CD pipelines in Data Oriented systems

Docker / Container / VM experience

Terraform / Ansible / other code-as-infrastructure



Degree Level :

*** Mention DataYoshi when applying ***

Offers you may like...

  • Glassdoor

    Data Engineer (Remote)
    Washington, DC
  • Homethrive

    Data Engineer
    Remote
  • YouGov

    Data Engineer
    New York, NY
  • drc

    Senior Full Stack Developer / Data Engineer - REMO...
    Maple Grove, MN
  • Relatus

    Senior Data Engineer
    Remote