Arobas Personnel is looking for a Data Engineer for a contract at one of its clients located in Toronto.
*Remote : Hybrid-Mode – 3 days/week in office*
What does your day to day look like?
Data ingestion pipeline using Spark, Hive, Map Reduce, Impala. And development of workflows in GCP data platforms starting in Q2.
Develop and operate our data pipeline & infrastructure
Responsible for the building, deployment, and maintenance of mission critical analytics solutions that process data quickly at big data scales
Works directly with business analysts and data scientists to understand, develop data-driven dashboards and solutions to support their use cases
Owns one or more key components of the infrastructure. Works to identify gaps and improving the platform’s quality, robustness, maintainability, and speed
Contributes design, code, configurations, and documentation for components that manage data ingestion, real time streaming, batch processing, data extraction, transformation, and loading across multiple game franchises.
What we’re looking for?
Experience utilizing relational concepts, RDBMS systems, and data design techniques (SQL server, Teradata, Oracle)
Strong background on the Hadoop ecosystem and its tools such as Spark, Oozie, Pig, Hive, Map Reduce, Impala
Experience in developing data pipelines and workflows in GCP data platforms, services and applications (Pub/Sub, Storage, Dataflow, Dataproc BQ, Airflow, Composer).
Hands-on skills with CI/CD and build automation tooling (Chef, Terraform, Ansible, GitOps, etc.)