Data engineering is your passion, and you love to make sure that data can be turned into valuable assets? Are you keen on creating data products that enable teams and organizations to increase their productivity? Do you have solid experience in designing and maintaining data architectures? You have an engineering mindset and love to analyze complex datasets? Do you love to work in an agile environment and deliver in short iterations? If this sounds interesting to you, then it is YOU we need in our team!
Tasks
- Design, develop, optimize, and maintain data architecture;
- Design and maintain ingestion of multiple data sources;
- Analyze, manipulate, transform, and process large and complex datasets;
- Enable training and running machine learning models;
- Building real-time data pipelines;
- Help our customers to become cloud-native and data-driven companies;
- Support your team with active knowledge transfer;
- Be part of and influence the introduction of new tools, methodologies, and techniques;
- Work in an agile environment and cross cross-functional team.
Requirements
At heart you are a passionate team player, who respects the opinions of your colleagues, as:
- You have at minimum 3 years of experience as a Data Engineer;
- You have at minimum 2 years of experience in either Python or Scala and SQL;
- You have a bachelor in Computer Science, Data Science, Data Engineering, or a relevant subject such as mathematics or physics;
- You have experience in semantic modelling of complex data landscapes and are familiar with concepts of Data Lake, Data Warehouse, Data Vault, and Data Mart;
- You have a deep understanding of various data stores, both structured and unstructured, and their capabilities (i.e. distributed filesystems, SQL, and noSQL data stores).
- You know exactly how to structure data pipelines for reliability, scalability, and optimal performance;
- You are comfortable working with analytics processing engines (i.e. Spark, Flink);
- You have worked with many different storage formats and know when to use which (i.e. JSON, Parquet, ORC);
- You speak fluent English;
- You know how to be the best team player;
- You have an eye for details and an ace in documenting your work;
- You base your decisions on metrics;
- You are very structured, and you set the benchmark for quality;
- You are open to new technologies.
Maybe you already have a bit of experience in the following:
- You have worked with one or more cloud technologies: Azure Synapse, Snowflake, AWS Athena, Google BigQuery;
- You have experience with building real-time data pipelines, by using some of the tools like: Azure Stream Analytics, Amazon Kinesis, Google Cloud Dataflow, Kafka, RabbitMQ.