Data Engineering Specialist (ETL, SQL, Python) - Onsite Geneva
Randstad (Switzerland) Ltd.
Date de publication :
16 avril 2025Taux d'activité :
100%- Lieu de travail :Geneva
- Salaire estimation de jobs.ch:Se connecter pour voir l’estimation de jobs.ch
Résumé de l'emploi
Randstad Digital Switzerland recherche un spécialiste en ingénierie des données. Rejoignez une équipe dynamique pour un projet international.
Tâches
- Collaborer avec l'équipe de sécurité de l'information pour garantir la qualité des données.
- Identifier des outils de visualisation de données et créer des présentations détaillées.
- Développer des produits de données, y compris des pipelines et des transformations.
Compétences
- Diplôme universitaire en informatique, ingénierie ou domaine connexe.
- Excellente maîtrise de SQL, Python et des outils de développement.
- Bonnes compétences en communication et en travail d'équipe.
Est-ce utile ?
job details
Randstad Digital Switzerland is seeking a highly skilled and motivated Data Engineering Specialist to join our team and work on an exciting project with a prestigious international organization. This role involves enhancing and developing security metrics within a large-scale IP Analytics Platform, utilizing cutting-edge cloud technologies and agile methodologies.
Responsibilities:
-
Collaborate with the Information Security team to define and implement data quality assurance processes.
-
Identify suitable data visualization tools and create detailed visualization and presentation layers (Tableau or Amazon QuickSight).
-
Identify and document data structures for additional data sources and ensure ongoing data drift monitoring.
-
Assist Information Security experts in defining and extracting relevant features for data analysis.
-
Collaborate with business areas to improve the performance of central collation of big datasets.
-
Develop data products, including pipelines, jobs, and transformations.
-
Create use case diagrams, data, and process flow diagrams, and user guides.
-
Design and implement Security Metrics presentation/visualization dashboards with drill-down capabilities.
-
Ensure solution evolution in response to changing data sources and emerging technologies.
Requirements:
-
A first-level university degree in information security, computer science, engineering, mathematics, business, or a related discipline.
-
Expert knowledge and experience in developing and implementing ETL jobs for data warehouses using SQL and Python.
-
Strong knowledge of software development languages and tools, including SQL, Python, and Spark.
-
Good knowledge and experience with AWS or Azure Big Data tools (e.g., Glue, Athena, Redshift, Kinesis, Data bricks, Azure analytics, Data explorer) and cloud-based storage and functions (S3, Blob, Lambda, Azure Functions).
-
Expert knowledge and experience with data engineering tools and methodologies (Data warehouse, data lake, star schema).
-
Good knowledge and experience with AWS Cloud Formation or Terraform.
-
Knowledge of CI/CD concepts, particularly AWS CodePipeline, CodeBuild, CodeDeploy.
-
Knowledge of provisioning data APIs and information security concepts and terminology.
-
Excellent written and verbal communication skills, with the ability to articulate complex technical ideas to non-technical stakeholders.
-
Confident communicator and team player with strong organizational and interpersonal skills.
-
Personal drive, ownership, and accountability to meet deadlines and achieve results.
-
Proficient user of Git, with familiarity with Jira and Bitbucket.
Desirable:
-
Additional certifications such as Certified Amazon Web Services (AWS) Solutions Architect and AWS Certified Data Analytics Specialty.
-
Experience in the implementation of or demonstrable familiarity with the Gartner ODM framework.
-
Experience of working with Tableau and Amazon QuickSight would be an advantage.
Ideally, you will be working from our client’s offices in Geneva, although a hybrid working arrangement (nearshore or Switzerland) is possible.