Friday, 16 February 2018

Title: DEVELOPER Location: SanJose CA, Duration:  6 Months Key skills required for the job are: Apache Spark-L3 (Mandatory) Scala programming-L3 Hadoop-L3 Minimum experience: 3 - 5 years Bachelors or Master s Degree in Computer Science  1 -2 years Apache Spark experience.  2 years of experience designing and implementing data ingestion and transformation for big data platforms. (Spark, Sqoop, Kafka, etc.)  Proven track record...
Title: DEVELOPER
Location: SanJose CA,

Duration:  6 Months

Key skills required for the job are:

Apache Spark-L3 (Mandatory)
Scala programming-L3
Hadoop-L3

Minimum experience: 3 - 5 years

Bachelors or Master s Degree in Computer Science 
1 -2 years Apache Spark experience. 
2 years of experience designing and implementing data ingestion and transformation for big data platforms. (Spark, Sqoop, Kafka, etc.) 
Proven track record designing highly parallelized data ingestion and transformation jobs in Spark including Spark Streaming. 
Production experience working with Apache Spark clusters. 
Knowledge of Java, Scala and how they are used in big data projects. 
Ability to create & implement data engineering best practices for the full software development life cycle, including coding standards, code reviews, source control management, documentation, build  processes, automated testing, and operations. 
Demonstrated ability to achieve stretch goals in a very innovative and fast paced environment. 
Demonstrated ability to learn new technologies quickly and independently. 
Excellent verbal and written communication skills, especially in technical communications. 
Strong inter-personal skills and a desire to work collaboratively. 
Desire and ability to mentor other developers.
Sound knowledge in Hadoop technology, Hive, HBASE, Spark using Scala.
 Experience with ETL tools, Spark, Hadoop, Hive, HBASE on Hadoop distribution.
 Develop streaming/real-time based complex event processing on Hadoop framework. Interface with different databases (SQL, NoSQL). Manage data quality, by reviewing data for errors or mistakes from data input, data transfer, or storage limitations.
 Perform data management to ensure data definitions, lineage and source are suitable for analysis, undertanding of dimensional modeling, Data Warehousing fundamentals.
 Proficient in Java programming, in addition to development IDE tool - Eclipse.
 Sound knowledge of Unix Scripting and Tidal Enterprise Scheduler.
 Good to have knowledge of Cisco Process and methodologies.

Related Posts:

0 comments:

Post a Comment

Blog Archive

Contributors

GemSoft Tech Solutions. Powered by Blogger.

Recent Posts