Data Engineer

Role: Data Engineer

Location: Emeryville, CA

Type: Contract

 

Job Description:

  

·       Build and optimize performance of Scala-Spark batch jobs using AWS EMR

·       Build and optimize performance of Kafka, NIFI and other components of
real-time pipelines

·       Participate and contribute to the design, architecture and
development of high quality data-lake data models

·       Build data pipeline orchestration

·       Ensuring code quality and conformance of code to applicable rules,
norms and relevant best

·       practices

·       Coaching, mentoring and developing new hires and less experienced
developers to work on Big Data projects

 

Skills

·       Strong understanding of distributed systems and distributed
computation.

·       Strong working knowledge in Scala and Spark, Kafka

·       Good knowledge in AWS services like S3, EMR, Glue, SageMaker, Lambda,
ECS, DMW and Athena.

·       Demonstrated working knowledge in Spark, Kafka, NiFi, 

·       Demonstrated working knowledge of data modeling

·       Experience in designing and running Unit and Integration testing

·       Hands-on exposure to data store technologies like MongoDB, DynamoDB,
Postgres and Redshift.

·       Knowledge of TDD & BDD methodologies, tools and practices

 

Reference : Data Engineer jobs
Source: http://jobrealtime.com/jobs/technology/data-engineer_i9847

Leave a comment