Job Description:
- Design, develop, and maintain data pipelines, ETL processes, and data warehouses.
- Ensure data quality, integrity, and reliability throughout the data pipeline.
- Work with GCP services such as Google Cloud Storage, BigQuery, Dataflow, and Pub/Sub to architect and implement scalable and efficient data processing solutions.
- Monitor and troubleshoot data pipeline issues.
- Implement security and access controls to protect sensitive data and ensure compliance with data privacy regulations.
- Document data processes, data flows, and architectural decisions.
Requirements:
- At least 1 year of experience in designing and building data pipelines and ETL processes.
- Bachelor’s degree in Engineering or a related field; a Master’s degree is a plus.
- Strong proficiency in Google Cloud Platform (GCP) services such as BigQuery, Dataflow, Pub/Sub, Composer/Airflow, and Cloud Storage.
- Solid programming skills in languages such as Python, Java, or Scala.
- Experience with data modeling, data warehousing concepts, and relational databases.
- Proficiency in SQL and database design principles.
- Familiarity with data integration and ETL tools.
- Familiarity with cloud platforms (e.g., AWS, Azure, Google Cloud).
- Strong understanding of data governance, data security, and data quality best practices.