The City of Detroit is seeking a Data Engineer interested in public service and data for social good. This role will support data transformation, integration, data cataloging, documentation, and publishing data to our open data portal, including innovative data work being done via the City’s Digital Inclusion effort. This position is part of a small and collaborative team that will be working closely to develop open data portal functionality improvements, public engagement strategies, training resources, and utilization of the portal for applied statistics
As a Data Engineer, you will contribute to the team's ETL/ELT processes. You will manage Open Data related data pipelines, written in Python and executed using a tool such as Airflow or Prefect, which extract data from departmental/vendor source systems into a small data warehouse. These pipelines further transform the data into models which are usable by end users or user-facing systems.
This position is funded through appropriations from the American Rescue Plan Act and is part of the larger Digital Inclusion effort. The Office of Digital Inclusion sees the operation of a robust open data portal as one of the ways we move Digital Inclusion towards Digital Empowerment. Ultimately, the City not only wants to get residents access to the internet, but they also want to find ways to engage and empower them with information once they are online. Open Data is an important tool for building trust between the City and its residents, but in order to be inclusive, the data needs to be presented in a way that is approachable for a variety of audiences and levels of technical expertise.
Examples of Duties
What you'll do:
- Develop data pipelines to extract, transform and load data from a variety of systems.
- Develop scripts or database models to clean and integrate data
- Design data schemas and database architecture
- Meet with team members and contractors to ensure alignment and buy in as well as technical feasibility.
- Identify and implement ways to test data to ensure quality and reliability
- Develop new data sources with team members or departmental clients
- Help the team develop trainings, lead focus groups, and identify and execute analytics and research projects that will benefit public users
- Bachelors in computer science, data science, information management, or database administration.
- 4 years of experience in systems analysis, computer science engineering, programming, information security management, data analysis, information science, or geographical information systems analysis is preferred but not required
Minimum tech skills:
- You should be skilled in both Python and SQL (ideally PostgreSQL)
- Experience with a pipeline/job scheduling/orchestration framework such as Airflow or Prefect
- Experience integrating large disparate data sources in a variety of formats.
- Experience working with REST APIs (typically as a source of data)
Would be great if:
- Experience with dbt (data build tool) to handle data modelling, transformation, and documentation
- Experience working with geospatial data, particularly in PostGIS or Esri/ArcGIS Online
- Experience with cloud data warehouses (Snowflake, Redshift, BigQuery)
- Experience working with Oracle, SQL Server
- Experience with Smartsheet and Box
- Experience working in Agile teams
- Working knowledge of web based applications and user experience best practices.
Appointment term is based on availability of grant funds