Hadoop Developer/Data Engineer (44322)

Company:
Location: Paraná, Entre Ríos

*** Mention DataYoshi when applying ***

Envie o currículo apenas no site petra-group.com.br/vagas
Develop & Support data pipelines/ETL feeds/applications such as reading data from external sources, merge data, transform data, perform data enrichment and load into target data destinations.

Primary Job Functions:

  • Technical support and development expertise for data pipeline such as Hadoop/Spark workflows & Kafka Spark Streaming real time data pipelines.

  • Develop or enhance Hadoop workflows including data ingestion, extraction and processing of structured and unstructured data.

  • Applying analytical experience with database to write complex queries, query optimization, debugging, Troubleshooting data issue on Hbase and Hive.

  • Support ongoing activities ingestion of additional data or new data sources, processing and monitoring incremental flows, managing of Solr collections, reprocessing data.

  • Datalake knowledge with experience of migrating use cases on-Prem HDP clusters to CDP Platform.(Knowledge on Cloudera distribution migration)

  • Optimizing existing Hadoop workflow and working with Stakeholders for data gap related issue.

  • Engage in proof of concepts, technical demos and interaction with customers and other technical teams

  • Work to utilize common CI/CD best practices and Working with Continuous Integration Tools for the big data use cases.


Requisitos

Proficiency

Primary skills:
  • Experience with Spark components Like Spark (Data frame/Dataset) and Streaming.
  • Experience with programming language like Scala and/or Python.
  • Experience with Apache Solr Search engine.
  • Experience with NO-SQL database like Hbase and Apache Phoenix.
  • Experience with Hive Query Language and optimization.
  • Sound Knowledge in performance tuning and optimization(Spark, Hbase & Hive).
  • Sound Working knowledge in Kafka and Spark Streaming.
  • Experience with Job scheduling & monitoring with Oozie.
  • Good Experience in SQL based language.
  • Experience in UNIX or UNIX-like systems.
  • Ability to interface with customers, technology partners, testing, architecture and analysis groups.

Secondary Skills:
  • Good to have NIFI knowledge.
  • Familiarity with Azure DevOps and GIT tools.
  • Experience with SCRUM or other similar Agile framework
  • Strong verbal and written communication skills.
  • Ability to work with data scientists and develop seamless workflows.
  • Self-starter and individual contributor.
  • Strong problem Solving skills.
  • Able to create Technical Documentation for the use cases.

*** Mention DataYoshi when applying ***

Offers you may like...

  • Techwave Consulting Inc.

    Hadoop Data Engineer
    San Diego, CA 92101
  • PBT Group

    Hadoop Data Engineer
    Johannesburg, Gauteng
  • PBT Group

    Hadoop Data Engineer
    Sandton, Gauteng
  • Cognizant Technology Solutions

    Hadoop Data Analyst
    Hillsboro, OR