Dieses Jobangebot ist archiviert und steht nicht mehr zur Verfügung.
Vakante Jobangebote finden Sie unter Projekte.
Vakante Jobangebote finden Sie unter Projekte.
Lead Data Engineer - Hadoop/Kafka/Spark
Eingestellt von Orcan Intelligence
Gesuchte Skills: Engineer, Python, Client, Linux
Projektbeschreibung
Lead Data Engineer - Hadoop/Kafka/Spark
For our prestigious client, on a major Data Transformation program are currently looking for a a Lead Data Engineer with excellent skills on Hadoop/Kafka/Spark.,
TO QUALIFY YOU MUST HAVE THE FOLLOWING SKILL SET:
- Minimum 4 years relevant devops and data wrangling experience in a (big) data environment
- Experience in big data Hadoop ecosystem: (with some of the component of the Hadoop ecosystem)
- Storage: HDFS, MongoDB, PostgreSQL, T HBase, Cassandra
- Tools: Kafka, Mesos, Docker, Spark, Hive, YARN,
- Programming knowledge in Scala, Python is a plus
- Excellent knowledge of Linux environment (system admin knowledge needs to be high)
- Knowledge of continuous development/integration pipelines including rules to test/validate code (git, Jenkins, test framework)
TASKS & RESPONSIBILITIES:
- Data pipelines starting from RDBMS with event capturing, transfer into KAFKA broker, consuming the events from the cluster with Spark, Spark streaming, generating metadata tables on Hive metastore, and generating data marts that will be exposed on Solr,HBase, Impala.
- CDC and Stream processing inside the Hadoop Stack
English speaking, no other language required.
If you have the required skill sand interested to apply, please send your CV now for immediate consideration
For our prestigious client, on a major Data Transformation program are currently looking for a a Lead Data Engineer with excellent skills on Hadoop/Kafka/Spark.,
TO QUALIFY YOU MUST HAVE THE FOLLOWING SKILL SET:
- Minimum 4 years relevant devops and data wrangling experience in a (big) data environment
- Experience in big data Hadoop ecosystem: (with some of the component of the Hadoop ecosystem)
- Storage: HDFS, MongoDB, PostgreSQL, T HBase, Cassandra
- Tools: Kafka, Mesos, Docker, Spark, Hive, YARN,
- Programming knowledge in Scala, Python is a plus
- Excellent knowledge of Linux environment (system admin knowledge needs to be high)
- Knowledge of continuous development/integration pipelines including rules to test/validate code (git, Jenkins, test framework)
TASKS & RESPONSIBILITIES:
- Data pipelines starting from RDBMS with event capturing, transfer into KAFKA broker, consuming the events from the cluster with Spark, Spark streaming, generating metadata tables on Hive metastore, and generating data marts that will be exposed on Solr,HBase, Impala.
- CDC and Stream processing inside the Hadoop Stack
English speaking, no other language required.
If you have the required skill sand interested to apply, please send your CV now for immediate consideration
Projektdetails
Geforderte Qualifikationen
-
Kategorie:
IT Entwicklung, Ingenieurwesen/Technik