Job Description
Job Responsibilities:
- Understand the business requirements, processes, and system functionality to develop data design and flow.
- Utilize best practices for standard naming conventions, coding practices, architecture, and ETL guidelines to ensure consistency in data models and department outputs.
- Develop and maintain the approved “Enterprise Business Terminology,” “Definition Document,” “Business Glossary,” and integration with metadata.
- Implement processes and logic to extract, transform, and load data across one or more data stores from various sources.
- Optimize data integration platform performance under increasing data sources and volumes.
- Include source validation, target validation, quality checks, technical metadata, and logging for each ETL Process.
- Review, support, control daily extraction window, maintain process flows, schedule jobs, and identify dependencies.
- Automate ETL process to minimize manual intervention and ensure data availability and quality.
- Maintain, configure, and monitor Data Engineering and ETL tools.
Job Qualifications:
- Experience with Big Data platforms and EDW.
- Hands-on experience with various input files (CSV, JSON, etc.).
- Strong knowledge of emerging technologies and tools.
- Strong problem-solving skills.
- Advanced knowledge in SQL, Database Concepts, and Design.
- Advanced knowledge in Data Warehouse Concepts and Design, Dimensional Modeling.
- Advanced knowledge in Programming Languages.
- Advanced experience in ETL Tools (Microsoft SSIS and Informatica).
- Advanced knowledge in Microsoft Excel.
- Strong technical communication and writing skills.
- Advanced understanding of the Banking business.
- 2+ years of experience in data engineering.
- At least two years of ETL design, development, and performance tuning on RDBMS like SQL Server.
- Bachelor's Degree in Computer Science, Computer Engineering, or IS/MIS.
Good To Have:
- Testing and debugging experience.
- Experience on Big Data platforms using open-source tools like Sqoop, Spark, Hive SQL, and Pig.
- Exposure to cloud solutions like AWS or Azure.
- Experience implementing solutions from requirements gathering to functional production deployment.
- Interface with key customers from all functional areas.
- Advanced knowledge of OLAP and analytical reporting principles.
- Advanced knowledge of Software Development.
- Advanced knowledge of Linux/Unix Shell Scripting.
- Presentation and Communication skills.
Tech Stack:
- Databases: Postgres, MySQL
- DWH: AWS Redshift, BigQuery, Azure Synapse, Snowflake
- ETL: Airflow, PySpark, AWS Glue, Airbyte, Apache NiFi, Mage
- Data Modeling: DBT, Databricks
- Reporting tool: Metabase, PowerBI, Tableau, Looker Studio
- Cloud: AWS, GCP, Azure
- Tools: Docker, Kubernetes