3-6 years of Big Data experience with good experience in building data processing applications using Hadoop, Spark and NoSQL DB and Hadoop streaming.
Expertise in one or more programming languages like Java, Scala or Python and in unix scripting.
Expertise in using query languages such as SQL, Hive, Sqoop and SparkSQL.
Expertise in storage and process optimization techniques in Hadoop and Spark. Experience in using tools like Jenkins for CI, Git for version Control and Exposure to Google Cloud (GCP) data components such as Cloud Data Flow, Cloud Data Proc, BigQuery and BigTable is preferred
Strong problem-solving, communication and articulation skill
Good knowledge of Python.
Strong in SQL (SQL queries, T-SQL, PL/SQL) and hands on coding.
Experience in a variety of relevant technologies including Streaming, NoSQL databases, Hive, Presto, Docker, Micro services architecture.
Familiarity with AWS Data and Analytics technologies such as Glue, Athena, Redshift, Spectrum, Data Pipeline.