Job Description:
- Being able to understand and develop ANSI SQL queries;
- Being aware of handling other semi-structured data formats – JSON, XMLs, HTMLs etc;
- Being comfortable in querying tools for distributed systems such as Hive, Impala and other SQL clients (RDBMS, Python, Scala etc.);
- Being comfortable in handling transforming data to suite business needs;
- Being able to handle tools for data transformation and/or ETL through tool-based and code-based development – ex. Informatica or Talend for tool based / Spark or MapReduce framework in Python or Scala/Java for code-based development;
- Having a background on data warehousing and reporting technologies
Requirements
Qualifications:
- 5+ years of experience in data engineering or a similar role.
- Bachelor's degree in computer science, Information Technology, or a related field.
- Proficiency in SQL and experience with relational databases.
- Strong programming skills in Python.
- Experience with ETL tools and data pipeline frameworks (e.g., Apache Airflow, Talend, Informatica).
- Knowledge of data warehousing solutions like Snowflake or Redshift.
- Good communication and stakeholder management skills
- Root cause analysis
- Business requirements gathering
- Capable of leading cross functional teams in order to deliver projects and initiatives