+

Вход

Въведи своя e-mail и парола за вход, ако вече имаш създаден профил в DEV.BG/Jobs

Забравена парола?
+

Създай своя профил в DEV.BG/Jobs

За да потвърдите, че не сте робот, моля отговорете на въпроса, като попълните празното поле:

101+53 =
+

Забравена парола

Въведи своя e-mail и ще ти изпратим твоята парола

Една от всички 167 обяви за Data Science в София

Виж всички
Adastra
GCP Data Engineer
ApplyКандидатствай

Обявата е публикувана в следните категории

+
  • Anywhere
  • Съобщи проблем Megaphone icon

Съобщи за проблем с обявата

×

    Какво не е наред с обявата?*
    Моля опиши ни, къде е проблемът:
    За да потвърдите, че не сте робот, моля отговорете на въпроса, като попълните празното поле:
    Tech Stack / Изисквания

    We are seeking a GCP Data Engineer responsible for designing and implementing end-to-end data products within the Google Cloud Platform (GCP) environment. The successful candidate will oversee ingestion, transformation, loading to final structures, and the exposition of the data according to agreed requirements. This role entails the delivery of infrastructure solutions for assigned GCP data applications throughout the complete use case lifecycle, from identifying and documenting big data use case requirements to implementing and rolling out solutions in production.

    Responsibilities:

    • Cooperate with data analysts and business stakeholders to ensure requirements are well understood.
    • Collaborate with data architects and other team members to design the solution for the data products in GCP.
    • Automate and orchestrate data pipelines in GCP.
    • Design and develop data ingestion and processing/transformation frameworks leveraging GCP core and related technologies.
    • Participate in architectural design and solution implementation of large-scale operational and analytic use cases.
    • Design, develop, and integrate ETL/ELT data pipelines.
    • Address non-functional requirements such as performance, security, scalability, continuous integration, migration, and compatibility.
    • Document the implemented solution according to agreed standards.
    • Mentor other team members on implementation aspects in GCP.
    • Ensure that governance and security requirements are met during the implementation.
    • Implement generic features of the data platform and policies to be applied according to requirements.
    • Implement interfaces of data products for data consumers.
    • Apply CI/CD practices during the implementation.
    • Assume ownership from design through to the performance of the feature in production.
    • Ensure fully automated testing by designing and writing automated unit, integration, and acceptance tests.

    Candidate Requirements:

    • Qalification in Computer Science or equivalent technical experience.
    • Good analytical and design thinking.
    • Strong practical experience with data infrastructure & processing & storage components in GCP (GCS, BigQuery, Cloud Functions, Composer, Cloud Run, Dataflow, Dataproc).
    • Google Cloud Professional Data Engineer certification is advantageous.
    • Strong SQL, Python, and Scala skills, with experience in Java or R being a plus.
    • Experience in API development, API product expertise, API design patterns, and API Security.
    • Knowledge of big data open source technologies such as Hadoop, Workflow Managers such as NiFi, Kafka, Druid, Hive, Storm, Ignite, Kudu.
    • Understanding of different file formats (e.g., AVRO, ORC, Parquet, etc.) and data sources moving data into and out of HDFS.
    • Familiarity with issue tracking (Jira), source code management (Bitbucket), Continuous Integration tools (Jenkins), Linux/Mac OS administration, build tools, package management and testing framework.
    • CI/CD experience is a plus.
    • Knowledge of Terraform for Infrastructure as Code (IaC) is a plus.
    • Strong Unix and scripting skills.
    • Excellent communication skills.

    Technologies you will use:

    • Python
    • Spark, Scala
    • Google Cloud Products
    • Dataproc
    • Dataflow
    • Databricks
    • Pub/Sub, Kafka
    • BigQuery
    • Hadoop, Hive, Hbase
    • Cloud Composer
    • DevOps /Terraform