Share this Job

Go Back

Hadoop Developer/ Architect (Strong in Java and Spark)

Boston
2018-10-10 16:12:07
Job Type: Contract

Responsibilities

•Work closely with business and IT teams to clarify and refine functional data requirement specifications.

•Translate detailed business requirements into optimal database and ETL solutions.

•Define and deliver consistent ETL standards, methodologies, guidelines & techniques.

•Diagnose complex problems, including performance issues, and driving to resolution.

•Collaborate with Data Analysts to develop coordinated plans to meet business requirements.

•Assist data analysts and end users across all functional areas in identifying long-term, strategic data needs for the enterprise, defining how the data is related to the business processes and developing a conceptual and functional model of the data

•Develop a strong understanding of health care data.

•Develop the production process to automate the Hadoop jobs and create documentation to transition the process to the Production team.

•Provide technical support to User Acceptance and Integration testing environments by researching suspected anomalies and periodically refreshing the environments.

•Performance tuning and optimization of processes in Hadoop environment.

 

Required Technical Skills

•3+ years of experience with big data technologies like Hadoop (Hortonworks, Cloudera)   Hive, Spark SQL, Kafka, Flume, HBase, Spark, Sqoop, and Impala.

•1 year of proven experience working with, processing and managing large data sets (multi TB scale). 

•2+ year experience in coding shell scripting in Python/Java and Pig. 

•1+ year experience with NoSQL databases, Spark SQL.

•Advanced ANSI SQL query writing and data modeling is must.

•3+ years of professional software development experience using Java/Scala with exposure to Junit and Mockito.

•Building APIs and Services against Hadoop based platform

•Experience building reusable framework using Java Framework and UDFs.

•Experience working in CICD environment using DevOps tools like GitHub, Jenkins, Jacoco, Maven etc.

•Experience in Data Lake or Data Warehousing ETL development is plus.

•Excellent oral and written communication.

•Ability to handle multiple priorities and meet deadlines.

•Experience with Cloudera distributions is plus.

•Healthcare domain knowledge is a plus

•Excellent verbal and written communication

 

Education

•    Bachelor’s degree or higher in Computer Science or related field.



Key Skills:
-