Job Description :
- Must have sound knowledge of Hadoop architecture and Apache Spark framework
- Experience in Data ingestion services, Sqoop, Flume, Kafka, NiFi
- Must be familiar with Hortonworks, Cloudera and AWSAzureGCPDatabricks
- Familiarity with one or more SQL-on-Hadoop technology (Hive, Pig, Impala, Spark SQL, Presto).
- Strong hands-on experience in integrating multiple databases like Oracle, SQL Server, PostgreSQL, Teradata, SQL server etc.
- High-end Troubleshooting and Communication skills.
- Understanding of Apache SparkHadoop and the Data Analytics ecosystem.
- Experience with one or more relevant tools (Sqoop, Flume, Kinesis, Kafka, Oozie, Hue, Zookeeper, Ranger, Elasticsearch, Avro, Parquet).
- Writing high-performance, reliable and maintainable code. - Should have 1-2 yrs. of experience working on AWSAzureGCPDatabricks
- Should have strong knowledge of either of the programming languages (JavaPythonScala)
- At least one full project lifecycle with design and development experience in SCRUM or AGILE methodology
- Good experience in the development of cloud-native applications.
- Good understanding of the No-SQL database and Timeseries database.
- Should have very good communication skills and should interact with client stakeholders on daily basis & provides daily status updates to reporting leads.
- Experience on Jira tool
- At least one full project lifecycle with design and development experience in SCRUM or AGILE methodology
- Virtualization techniques - Dockerization and Kubernetes(Experience with docker environments) desired
- JavaJ2EE knowledge desired