Data Engineering
Skill Set
- Strong knowledge in DataBricks
- Strong in PySpark and Python
- Strong in SQL/MySQL, Relational data modelling
- Strong Notebooks environment experience (Jupyter, DataBricks)
- Fluent in complex, distributed and massively parallel cloud systems (AWS, GCP, AZURE)
- Experience with workflow orchestration tools (Airflow, Jenkins)
- Basics of AWS/Azure technologies
- Proficient in using Git & Tera Forms
- Experience collecting requirements from Partners and choosing the right technologies to meet end to end data flow requirements that are required.
- Strong analytical and problem solving skills.
- Strong written and verbal communication skills
- Description
Responsible for all stages of design and development for complex products and platforms, including solution design, analysis, coding, testing, and integration. Exercises independent judgment within generally defined policies and practices to identify and select a solution. Ability to handle most unique situations. Designs and establishes secure and performant data architectures, enhancements, updates, and programming changes for portions and subsystems of data product pipelines, repositories, or models for structured/unstructured data.
- Responsibilities
- High level of initiative, with an ability to plan and manage tasks, ability to work collaboratively with a group of peers, both within and outside one’s own group.
- Designs, analyses, programs, debugs, troubleshoots, and modifies software applications for enhancements and new products
- Education and Experience
- Bachelor’s or master’s degree in electrical and Electronics / Electronics and Communication / Electronics & Instrumentation/Computer Science/Information Science