This is a Hybrid Role - Office location is Denver, CO or Raleigh/Durham, NC
Do you want to be a part of big data in education? Pearson – the world’s leading learning company – has an exciting opportunity for a mid-level Data Engineer on the forefront of big data collection and curation.
Imagine all the interactions that a learner, educator or administrator has every day. Multiply that by the millions of learners, educators, and administrators around the world. All those interactions and events generate data. Now you can begin to understand the scale of data at Pearson.
You can have an impact on learners around the world by contributing to the curated data eco-system that is the foundation for a wide range of vital platform and business processes here at Pearson.
The Digital Analytics Cloud (DAC) Data team is the heart of Pearson’s data environment. As a member of this team, you’ll be responsible for ingesting, transforming, and curating data in our data lake. We use multiple technologies including AWS Snowflake and Google BigQuery. Our data lake has over 300 terabytes of data generated by application platforms, user interactions, and system monitors. Our consumers include applications, data science, machine learning, monitors, dashboards and reports.
Responsibilities:
- Design, code, unit test, and deploy data processes for ingestion, transformation, or curation of big data while keeping data security and privacy in mind.
- Actively participate in requirements and design meetings.
- Enjoy a healthy technical debate but know how to collaboratively solve problems that move everyone forward.
- Communicate well in-person, over email, and via instant messaging both within the team and with other development teams.
- Work efficiently and collaboratively in an Agile (Scrum or Kanban) team environment.
- Directly and meaningfully impact the experience of students, institutions, professors and administrators.
- Be able to organize your workload based on team priorities.
- Be willing to learn new skills whether they are directly tied to software development, software testing, server maintenance, or release and delivery.
Qualifications
- The ideal candidate will be detail-oriented, self-directed, self-motivated, with a strong capacity for working successfully and flexibly with members across the organization.
- 2+ years of data engineering or ELT experience.
- Experience and practice with one or more of the following technologies:
- Snowflake
- AWS Redshift, DMS or, Kinesis
- Google BigQuery
- Databricks
- Commercial or open source ETL / ETL tools
- Experience and practice with SQL.
- Experience or interest working in cloud-based environments preferred.
- Experience or interest in Python, Java or Scala.
- Experience with automated deployments a plus.
- Solid foundation in computer science, with experience in OO design and development and multiple data structures.
- Experience with data modeling a plus.
- Easy to work with, stays confident and optimistic in face of challenges.
- Strong communication skills and the ability to communicate technical concepts to non-technical people.
Job: TECHNOLOGY
Organization: Corporate Strategy & Technology
Schedule: FULL\_TIME
Req ID: 4610
\#LI-REMOTE