Detalhes da oferta
Descrição:
We are looking for a Data Engineer to build and maintain scalable platforms for data storage, processing, and orchestration. The ideal candidate implements performant pipelines and integration
solutions while remaining technology-agnostic, aligning technical details with the 'big picture' to drive business success and ensuring high-quality data flow for all stakeholders.
O que terás de fazer
- Design, build, and maintain scalable data platforms;
- Collect, process, and analyze large and complex data sets from various sources;
- Develop and implement data processing workflows using data processing framework technologies such as Spark, and Apache Beam;
- Collaborate with cross-functional teams to ensure data accuracy and integrity;
- Ensure data security and privacy through proper implementation of access controls and data encryption;
- Extraction of data from various sources, including databases, file systems, and APIs;
- Monitor system performance and optimize for high availability and scalability.
O que precisas de garantir
- Experience with cloud platforms and services for data engineering (GCP);
- Proficiency in programming languages like Python, Java, or Scala;
- Use of Big Data Tools as Spark, Flink, Kafka, Elastic Search, Hadoop, Hive, Sqoop, Flume, Impala, Kafka Streams and Connect, Druid, etc.;
- Knowledge of data modeling and database design principles;
- Familiarity with data integration and ETL tools (e.g., Apache Kafka, Talend);
- Understanding of distributed systems and data processing architectures;
- Strong SQL skills and experience with relational and NoSQL databases;
- Familiarity with cloud platforms and services for data engineering (e.g., AWS S3,Azure Data Factory);
- Experience with version control tools such as Git.
O que te proporcionamos
- Direct employment contract with the client
- 100% remote work model
Responsabilidades
O que terás de fazer
- Design, build, and maintain scalable data platforms;
- Collect, process, and analyze large and complex data sets from various sources;
- Develop and implement data processing workflows using data processing framework technologies such as Spark, and Apache Beam;
- Collaborate with cross-functional teams to ensure data accuracy and integrity;
- Ensure data security and privacy through proper implementation of access controls and data encryption;
- Extraction of data from various sources, including databases, file systems, and APIs;
- Monitor system performance and optimize for high availability and scalability.
Requisitos
O que precisas de garantir
- Experience with cloud platforms and services for data engineering (GCP);
- Proficiency in programming languages like Python, Java, or Scala;
- Use of Big Data Tools as Spark, Flink, Kafka, Elastic Search, Hadoop, Hive, Sqoop, Flume, Impala, Kafka Streams and Connect, Druid, etc.;
- Knowledge of data modeling and database design principles;
- Familiarity with data integration and ETL tools (e.g., Apache Kafka, Talend);
- Understanding of distributed systems and data processing architectures;
- Strong SQL skills and experience with relational and NoSQL databases;
- Familiarity with cloud platforms and services for data engineering (e.g., AWS S3,Azure Data Factory);
- Experience with version control tools such as Git.
Condições
O que te proporcionamos
- Direct employment contract with the client
- 100% remote work model
| Distrito | Lisboa |
|---|---|
| Local | Portugal |
| Sector de actividade | TI, Telecomunicações e Comunicação |
| Tipo de contrato | Full-time |
| Tipo de oferta | Emprego |
| Contactos |
Clan
Av. D. João II, n.º 36, 2º Piso Norte, 201/202 1998-017 Lisboa |