Sparksoft

Hadoop Developer

  • Job Type: Full Time
  • Industry Type: IT Sector
  • Industry Location: Columbia
  • Experience: NA
  • No. of Positions: 1
  • Primary Skills: Hadoop Developer AWS Databricks Snowflake HDFS
  • Secondary Skills: Hive Spark Big Data
  • Job Location: Columbia, Maryland
  • Posted Date: Posted today
Job Description

Job Title: Hadoop Developer

Location: Columbia-MD

Duration : Long term with Sparksoft

 

Job Summary:

This position will be supporting and working on one of Sparksoft’s technical projects. The ideal candidate will have experience in developing data ingestion and transformation ETL processes for analytical data loads, from a technical perspective.


Responsibilities:

• Selecting and integrating any Big Data tools and frameworks required to provide requested capabilities.

• Transition of legacy ETLs with Java and Hive queries to Spark ETLs.

• Design, develop, test and release ETL solutions including data quality validations and metrics that follow data governance and standardization best practices.

• Design, develop, test and release ETL mappings, mapplets, workflows using Streamsets, Java MapReduce, Spark and SQL.

• Performance tuning of end-to-end ETL integration processes.

• Monitoring performance and advising any necessary infrastructure changes

• Analyze and recommend optimal approach for obtaining data from diverse source systems.

• Work closely with the data architects, who maintain the data models, including data dictionaries/metadata registry.

• Interface with business stakeholders to understand requirements and offer solutions.

Required Skills:

• Proficient understanding of distributed computing principles and hands on experience in Big Data Analytics and development

• Good knowledge of Hadoop and Spark ecosystems including HDFS, Hive, Spark, Yarn, MapReduce and Sqoop

• Experience in designing and developing applications in Spark using Scala that work with different file formats like Text, Sequence, Xml, parquet and Avro

• Experience of using build tools Ant, SBT Maven

• Strong SQL coding; understanding of SQL and No SQL statement optimization/tuning.

• Ability to lead designing and implementation of ETL data pipelines.

• Experience developing data quality checks and reporting to verify ETL rules and identify data anomalies.

• AWS development using big data technologies.

• Techniques for testing ETL data pipelines either manual or using tools.

• AWS cloud certified, CMS experience, Databricks and Snowflake experience a plus.


Education/Experience Level:

• Bachelor’s Degree with 5 years’ experience or 10+ years of experience in the software development field.

• 5+ years of Bigdata ETL development experience.

• 4+ years of AWS big data experience.

• 3+ years of experience developing data validation checks and quality reporting.

• 4+ years of experience tuning Spark/Java coding, SQL and No SQL.

Sparksoft is a certified Capability Maturity Model Integration (CMMI) SVC and DEV Level 3, ISO 9001:2015, ISO 27001:2013, HUBZone, 8(a), Small Disadvantaged Business (SDB), Women-Owned Small Business (WOSB), and Small, Women-owned, Minority-owned (SWaM), and MBE/DBE/SBE consulting firm. With our focused mission “to ignite innovation, inspire transformation, and implement digital solutions for a healthier nation”, we specialize in 6 specific digital health services: Test Automation, Cloud Services, DevOps Delivery, Cyber Security, Data Science, and Human-Centered Design. Since 2004, our exceptionally skilled people, proven leadership, and optimized processes all work together relentlessly to continuously push for more efficient solutions.

Sparksoft is an Affirmative Action/Equal Opportunity Employer and does not discriminate against any applicant for employment or employee because of race, color, religion, sex, sexual orientation, gender identity, national origin, age, disability, protected veteran status, or any other characteristic prohibited under Federal, State, or local laws.

Relevant Job Openings
Tableau developer
Azure Data Architect with Talend
Node JS Developer
Solution Architect
Java Developer
Big Data Developers with Scala