Data Engineer - Hearst Newspapers
With more than 3,000 employees across the nation, Hearst Newspapers publishes 24 dailies and 52 weeklies, including the Houston Chronicle, San Francisco Chronicle, San Antonio Express-News and Albany Times Union. It also operates digital marketing services and directories businesses under the LocalEdge brand.
We are looking for driven individuals to join our team of passionate data engineers and data scientists in creating Hearst Newspaper’s next generation of data products and capabilities. Candidates should possess a deep knowledge of modern data processing technology stacks and database systems including SQL/NoSQL, alternatively a strong background in software engineering is also preferred. We are a small and efficient team building scalable data pipelines on cloud platform for data analysis and reporting process and providing a solution with machine learning and data mining techniques.
Responsibilities:
Build reusable products and systems that can effectively ingest data to leverage across technology, business intelligence, and advertising operations teams.
Support new feature developments between data sources for our own ingestion, reporting, and analysis platforms.
Collaborate with data analysts and project managers to identify and use the data that is most relevant to the problem.
Develop, deploy, and maintain the data products and systems.
Work closely with data scientists to transform data and integrate algorithms and models into the production system.
Qualifications:
Bachelor’s degree in Computer Science or equivalent practical experience. Master’s degree a plus.
3+ years of solid experience in building ETL/ELT pipelines. Familiarity with Airflow is a plus
Fluent in Python and SQL. Excellent working knowledge with relational databases.
Knowledge with object-oriented or functional programming skills.
Experience with bash shell scripts, UNIX/Linux commands.
Experience with cloud platform offerings. Google GCP preferred.
Ability to quickly learn modern data technologies.
Experience with traditional agile methodology.
Bonuses:
Natural Language Processing (NLP) and/or Machine Learning(ML).
Knowledge of Big Data technologies such as Spark, Kafka, and Apache Solr
Experience with BI tools like Looker and PowerBI.
Experience with CI/CD pipelines in Data Oriented systems
Docker / Container / VM experience
Terraform / Ansible / other code-as-infrastructure