Jobs Career Advice Signup
X

Send this job to a friend

X

Did you notice an error or suspect this job is scam? Tell us.

  • Posted: Oct 6, 2023
    Deadline: Not specified
    • @gmail.com
    • @yahoo.com
    • @outlook.com
  • Never pay for any CBT, test or assessment as part of any recruitment process. When in doubt, contact us

    ENGIE is a leading global energy company that builds its businesses around a model based on responsible growth to take on energy transition challenges. We provide individuals, cities and businesses innovative solutions based on our expertise in 4 key sectors: independent power production, natural gas, renewable energy and energy efficiency services to a low...
    Read more about this company

     

    Data Engineer

    Purpose/Mission 

    • This position will be part of the Global Data team that is based across Germany, Uganda, Kenya, and Nigeria. 
    • You will report to the Head of Data, and work closely with data scientists, devops team, and software engineers.
    • This is an incredible opportunity for a talented individual to join a high-performing team that is passionate about pioneering expanded financial services to off-grid customers at the base of the pyramid.
    • Key responsibilities will include building, maintaining, and ensuring scalability of data pipelines between MySQL databases which service our in-house applications, IoT data delivered from devices, PBX, our in-house ticketing system, and our data lake used for analytics across the company. 
    • You would also be responsible for building and optimizing pipelines to deliver data in real time to our field team mobile application to allow data-informed decisions to be made in the field, as well as working with members of the data team to ensure high code quality and database design. 
    • Your work will make a meaningful impact by enabling EEA to continuously innovate on how we support our customers in their solar kit experience and repayment journey.

    Responsibilities

    • Work with data and software teams to design, build, and support critical data pipelines using airflow for modeling, ETL, and analytics.
    • Optimize data storage between Redshift, s3, and other storage solutions to support data analytics, modeling, archiving and data integrity.
    • Develop logging and visualization KPI Dashboards with Grafana or another tool to score the efficiency of business processes.
    • Containerize models with Docker and Kubernetes to serve real-time financial information to field teams.
    • Work with software engineers and devops team to optimize performance of in-house applications which communicate data through APIs and other means.
    • Maintain and Develop Tools for unit testing, streamlined ETL, and other processes for the Data Team.
    • Mentor data scientists and analysts on best coding practices through code reviews and discussions.

    Knowledge and skills 

    Experience:

    • 3+ years of industry experience
    • Experience building infrastructure to support streaming or offline data
    • Extensive programming experience in Python/Scala/Java
    • Experience with SQL in addition to one or more of Spark/Hadoop/Hive/HDFS
    • Experience with implementing unit and integration testing
    • Ability to gather requirements and communicate with stakeholders across data, software, and platform teams
    • Ability to develop a strategic vision for data pipelining and infrastructure
    • AWS Certification is a plus
    • Strong communication across data, devops, and software team members
    • Sense of adventure and willingness to dive in, think big, and execute with a team

    Qualifications:

    • Degree in computer science or related field

    Language(s): 

    • English
    • French, Portuguese, or German is a plus

    Technology:

    • Linux-based systems
    • Knowledge of Amazon Web Services (AWS) and its services, such as, but not limited to, Cloudwatch, RDS, Redshift, Lambda, EMR, S3, SQS, EC2
    • Python, jupyter notebooks
    • Airflow (or other workflow management systems, such as Luigi)
    • Docker, Kubernetes or other container like tools
    • Streaming tools such as Kafka, Kinesis
    • Knowledge of Hetzner is a plus

    Method of Application

    Interested and qualified? Go to Engie Africa on jobs.engie.com to apply

    Build your CV for free. Download in different templates.

  • Send your application

    View All Vacancies at Engie Africa Back To Home

Subscribe to Job Alert

 

Join our happy subscribers

 
 
Send your application through

GmailGmail YahoomailYahoomail