Experience in design, development, and maintenance and support of Big Data Analytics using Hadoop Ecosystem components like HDFS, Hive, Pig, HBase, Zookeeper, Map Reduce, and oozie.
Strong working experience with ingestion, storage, querying, processing and analysis of big data.
Extensive Experience on working with Hadoop Architecture and the components of Hadoop - Map Reduce, HDFS, Job Tracker, Task Tracker, Name Node and Data Node.
Having exposure to faster processing of large datasets using Apache Spark and distributed messaging and multi-processing of Kafka.
Hands-on experience in writing and tweaking Map Reduce programs according to the requirement.
Good experience with Productionalizing Apache Hadoop, Apache Spark and Apache Kafka applications (such as monitoring, debugging and performance tuning).
Experience in writing Pig scripts, hive scripts, pig UDF, hive UDF for ETL process workflows.
Familiar with importing and exporting data using Sqoop into HDFS and have experience in fine tuning the sqoop actions for terabytes of data.
Experience with SQL, PL/SQL and database concepts and NoSQL databases.