Основные требования
· Bachelor's degree with at least 2 years of experience building and operating big data processing pipelines and/or query solutions.
· Proficient understanding of distributed computing principles.
· An ability to functionally explain the utility and application of particular big data processing and query capabilities.
· Demonstrated ability to manage a Hadoop cluster, and associated services.
· Proficiency with Hadoop, MapReduce, and HDFS.
· Experience with big data ETL frameworks and techniques such as Flume, Oozie, or Sqoop.
· Experience with Big data interactive query technologies such as Spark, Impala, Hive, or Presto.
· Experience with the Elastic ecosystem - elasticsearch, kibana, marvel, watcher, and shield.
· Programming experience, ideally in Python or Java.
· Excellent written and oral communication skills.
Дополнительные (желательные) требования
· Experience tuning Elasticsearch for ingest and query performance.
· Experience securing Elasticsearch and other big data technologies for the intelligence community.
· Experience with messaging/queuing technologies like RabbitMQ or Kafka.
· Experience with version control platforms (git, svn, etc.)
· Prior experience in support of law enforcement or intelligence related big data solution development.
Знание иностранного языка и степень владения им
Знания английского – свободное владение (устный, письменный)
Должностные обязанности
· Use creativity and knowledge of the possible to build the data processing pipeline and data stores to support applications delivering actionable intelligence insights from massive data sets.
· Recommend big data technologies and associated infrastructure applicable to specific problem sets, and demonstrate their efficacy.
· Implement ETL pipelines to automate the ingestion, transformation, and augmentation of both structured and unstructured data sources, and provide instrumentation for pipeline operations.
· Implement, configure, and operate big data technologies, and tune for performance at scale.
· Participate as an engineer on an agile development team, writing, testing, and deploying code to implement changes to existing data pipelines or construct new ones.
Условия:
Работа в Москве (удаленная на проектах в США)
Участие в очень крупных проектах
Контакты: VZhurin@bellintegrator.ru
· Bachelor's degree with at least 2 years of experience building and operating big data processing pipelines and/or query solutions.
· Proficient understanding of distributed computing principles.
· An ability to functionally explain the utility and application of particular big data processing and query capabilities.
· Demonstrated ability to manage a Hadoop cluster, and associated services.
· Proficiency with Hadoop, MapReduce, and HDFS.
· Experience with big data ETL frameworks and techniques such as Flume, Oozie, or Sqoop.
· Experience with Big data interactive query technologies such as Spark, Impala, Hive, or Presto.
· Experience with the Elastic ecosystem - elasticsearch, kibana, marvel, watcher, and shield.
· Programming experience, ideally in Python or Java.
· Excellent written and oral communication skills.
Дополнительные (желательные) требования
· Experience tuning Elasticsearch for ingest and query performance.
· Experience securing Elasticsearch and other big data technologies for the intelligence community.
· Experience with messaging/queuing technologies like RabbitMQ or Kafka.
· Experience with version control platforms (git, svn, etc.)
· Prior experience in support of law enforcement or intelligence related big data solution development.
Знание иностранного языка и степень владения им
Знания английского – свободное владение (устный, письменный)
Должностные обязанности
· Use creativity and knowledge of the possible to build the data processing pipeline and data stores to support applications delivering actionable intelligence insights from massive data sets.
· Recommend big data technologies and associated infrastructure applicable to specific problem sets, and demonstrate their efficacy.
· Implement ETL pipelines to automate the ingestion, transformation, and augmentation of both structured and unstructured data sources, and provide instrumentation for pipeline operations.
· Implement, configure, and operate big data technologies, and tune for performance at scale.
· Participate as an engineer on an agile development team, writing, testing, and deploying code to implement changes to existing data pipelines or construct new ones.
Условия:
Работа в Москве (удаленная на проектах в США)
Участие в очень крупных проектах
Контакты: VZhurin@bellintegrator.ru