Advanced Analytical Advisor

  • 37441
  • Non-Life - Data Science
  • |
  • Quebec, Canada
  • |
  • May 27, 2020
  • |
  • French
  • Analyze and interpret the business requirements in terms of advanced analytical capacity development and implement adequate solutions to meet them.
  • Develop pipelines to extract, load and convert data gathered from a wide variety of sources for data lake systems.
  • Develop state-of-the-art tools to prepare, model, profile and clean the data that scientists can use to create predictive models and machine-learning algorithms.
  • Contribute to operationalizing and maintaining existing data pipelines.
  • Work with data lake architects on the ecosystem’s road map, automation and orchestration.
  • Design the documentation required to support solutions.
  • Recommend solutions to business partners by using data storage techniques and modern Big Data platforms.
  • Interact with data experts and data scientists in the implementation of complex data pipelines in a Big Data ecosystem.
  • Serve as an ambassador to promote and simplify data lake platform capacity and services.
  • Support the data architecture experts in the development of data pipeline architecture patterns.
  • Ensure advanced analytical solutions are secure.
  • Bachelor’s degree in computer science or in a related field.
  • At least six years of relevant experience, including two years working on Big Data solutions.
  • Recognized experience in developing Big Data solutions.
  • Experiences with Spark Streaming, Kafka Streaming and Spark HBase.
  • Experience with Linux environments.
  • Experience and familiarity with Agile methodologies and tools.
  • Proficiency in developing data pipelines using tools and frameworks such as Apache NiFi, Spark, Kafka, Hive and HBase.
  • Proficiency in development and query languages, and in SQL, Java, Scala and Python data processing libraries.
  • Solid practical knowledge of how to use CI/CD.
  • Knowledge of Hortonworks and/or Cloudera platforms and business solutions.
  • General knowledge of data formats such as Avro, Parquet and ORC.
  • Knowledge of Cloud Azure and/or GCP platforms.
  • Knowledge of NoSQL solutions.
  • Familiarity with the development of microservices.