Internship in Data Engineering | Adaltas

Job Description:

Data is a valuable business asset. Some call it the new oil. A data engineer collects, transforms, and refines raw data into information that can be used by business analysts and data scientists.

As part of your internship, you will be trained in various aspects of data engineering. You will build a real-time, end-to-end data flow ingestion pipeline that combines metric collections, data cleansing and aggregation, storage across multiple data warehouses, (near) real-time analytics through key dashboard impact metrics, and usage. machine learning models applied to weak signal prediction and detection.

You will participate in application architecture and pipeline implementation to go into production. You will join an agile team led by a Big Data expert.

In addition, at the end of the internship, you will receive a certificate from a Cloud provider and a Databricks certification.

Presentation of the company

Adaltas specializes in data processing and storage. We work on-premise and in the cloud to run Big Data platforms and strengthen our clients’ teams in architecture, operations, data engineering, data science and DevOps. A partner with Cloudera and Databricks, we are also open source contributors. We invite you to browse our website and our many technical publications to learn more about Adaltas.


  • Collection of system and application metrics
  • Distributed data warehouse provisioning with OLAP-type columnar storage
  • Cleaning, enrichment, aggregation of data flows
  • Real-Time Analysis in SQL
  • Creating dashboards
  • Generating machine learning models with the MLOps cycle
  • Deployment in Azure cloud infrastructure and on-premises

Expected qualifications

  • Engineering school, graduate practice
  • Analytical and structural
  • Autonomous and curious
  • You are an open-minded person who likes to share, communicate and learn from others
  • Good knowledge of Python, Spark and Linux systems

You will be responsible for designing the technical architecture. We are looking for someone who has or will develop skills in the following tools and solutions:

All additional experiences are valuable.

Additional information

  • Location: Boulogne-Billancourt, France
  • Languages: French or English
  • Start: February 2022
  • Duration: 6 months
  • Telecommuting: 2 days per week available

Available equipment

A laptop with the following specifications:

  • 32 GB of RAM
  • 1TB SSD
  • 8c/16t processor

A cluster that is composed of:

  • 3x 28c/56t Intel Xeon Scalable Gold 6132
  • 3x 192TB RAM DDR4 ECC 2666MHz
  • 3x 14 SSD 480GB SATA Intel S4500 6Gbps

Kubernetes cluster and Hadoop cluster.


  • Salary 1200 €/month
  • Restaurant tickets
  • Transport ticket
  • Participation in one international conference

Past conferences we’ve attended include KubeCon, hosted by the CNCF Foundation, the Open Source Summit from the Linux Foundation, and Fosdem.

For any further information requests and to submit your application, please contact David Worms:

Source link