Job Summary:
The Data Engineer will play a pivotal role in the development, maintenance, and optimization of our data infrastructure, focusing on data integration, ETL/ELT processes, and management of Data Warehouses, and Lakehouse.
The successful candidate will be responsible for building and maintaining data pipelines using Databricks, Synapse, and Fabric environments, using tools like DBT and Databricks pipelines. Key responsibilities include ensuring system readiness in terms of security, performance, and health, completing data loads, and implementing data modeling to support various business domains. This hands-on role demands strong technical expertise alongside excellent communication and collaboration skills.
Major Responsibilities:
- Develop Python, SQL, and PySpark-based applications and data flows within Databricks.
- Build and maintain data pipelines using DBT and Databricks, ensuring efficient and scalable data processes.
- Design and implement real-time and batch data processing systems to support analytics, reporting, and business needs.
- Monitor and analyze data pipelines for performance, reliability, cost, and efficiency.
- Proactively address any issues or bottlenecks to ensure smooth operations.
- Discover opportunities for process improvements, including redesigning pipelines and optimizing data delivery mechanisms.
- Manage and maintain Azure cloud services and monitor alerts to ensure system availability and performance.
Minimum Requirements:
- Demonstrated ability writing SQL scripts (required).
- Some experience with Python (strong plus).
- Exposure to Databricks is a significant advantage.
- Experience working in a cloud environment (strong plus).
- Experience with DBT Core or DBT Cloud is a major plus.
- Ability to quickly learn and absorb existing and new data structures.
- Excellent interpersonal and communication skills (both written and verbal).
- Ability to work independently and collaboratively within a team.
Preferred Additional Requirements
- Experience with cloud platforms (Azure preferred).
- Databricks Data Engineer Certification or similar.
- Software development
Key Working Relationships:
- Data Team
- Enterprise systems owners and technical and analytics teams under their leadership