- Bachelor degree from computer science or related fields, or equivalent software engineering experience.
- Proficiency in Pyhton Programming Language
- Experience in dataset processing and feature engineering using tools such as numpy, pandas, scikit-learn
- Understanding of SQL & non SQL
- Experience with data engineering tools such as spark, Hadoop, talend, nosal, pentaho, informatica, airflow, presto, hive, kafka, IBM datastage or Microsoft SSIS
- Proficiency in source control i.e Git
- Understanding and Experience of cloud native technologies as AWS, GCP or Azure is even better
- Understand ETL Concept, automation scheduling & data pipelines internal
- Have a good understanding of data modelling, data lake, data warehouse & big data stacks
- Have a good understanding on programming concept, especially python & SQL
Job Descriptions:
Design OLTP Operational menggunakan big data tools (ex: kafka, sparks, etc), design proses integrasi data lake menggunakan (ex: pentaho, SSIS, etc) dan mendesign data streaming data warehouse menggunakan (ex: kafka, spark) dan reporting data.