Responsibilities:
- Design, build, launch and maintain efficient and reliable large-scale batch and real-time data pipelines with data processing frameworks
- Integrate and collate data silos in a manner which is both scalable and compliant
- Responsible for developing backend APIs & working on databases to support the applications
- Drive enterprise data foundation requirements of Data Warehousing, Data Lake
- Acquire, store, govern and process large datasets of structured/unstructured data
- Monitor and optimise infrastructure and pipeline
- Ensure Data Quality through continuous improvement and monitoring
Requirements:
- Professional Certification in IT, Information Management and/or Computer Science
- Experience with big data processing tools such as Talend, Informatica or Pentaho.
- Experience with batch and streaming jobs to ingest and process data
- Experience with Cloud environments such as AWS, GCP, Azure
- Experience with other NoSQL databases such as GreenPlum
- Knowledgeable about system design, data structure and algorithms
- Ability to work well in a team environment and develop rapport with business function managers
- Ability to manage multiple priorities and timelines
- Possess initiative, highly motivated and able to innovate and work independently if needed
All applicants will be notified on whether they are shortlisted or not within 4 weeks of the closing date of this job posting.
Successful candidates will be given a 2-year contract.