Search by job, company or skills

qnbeyond plus

Data Engineer

new job description bg glownew job description bg glownew job description bg svg
  • Posted 9 hours ago
  • Be among the first 10 applicants
Early Applicant

Job Description

About the Role

We are looking for a motivated Data Engineer to join our Data Management and AI team and contribute to building scalable data platforms that power analytics and AI initiatives across our digital banking ecosystem.

In this role, you will design and develop data pipelines, streaming infrastructure, and data warehouse solutions that enable reliable and high-performance data processing.

This is an excellent opportunity for a junior data engineer looking to grow within a fast-paced fintech environment while working on modern data technologies.

Team Overview

The Data Management and AI team works across the entire technology stack within QNBeyond Plus.

Our team collaborates with engineering, analytics, and product teams to manage, process, and analyze data that enables innovative digital banking experiences for our customers.

Key Responsibilities

Data Pipeline Development

  • Design and build scalable data pipelines for batch and real-time processing.
  • Implement ETL/ELT workflows integrating data from APIs, databases, and event streams.

Streaming & Event Processing

  • Develop and maintain streaming data pipelines using tools such as Kafka or similar platforms.
  • Ensure low-latency, reliable, and fault-tolerant data delivery.

Data Transformation

  • Develop data enrichment and transformation workflows to convert raw data into business-ready datasets.
  • Implement SQL transformations across layered architectures (raw → bronze → silver → gold).

Data Warehouse Development

  • Design and maintain data warehouse models and schemas.
  • Optimize query performance and storage design for analytical workloads.

Data Quality & Reliability

  • Implement data validation and quality checks.
  • Monitor data pipeline performance and ensure reliability.

Collaboration

  • Work closely with engineering, analytics, and product teams to deliver production-ready data solutions.

What We Are Looking For

You must be:

  • Self-motivated, proactive, and results-driven
  • A team player who enjoys collaboration and learning
  • A problem solver who approaches challenges analytically
  • Customer-oriented, committed to delivering high-quality data solutions
  • Detail-oriented, especially regarding data quality and compliance
  • Adaptable and eager to grow in a fast-paced fintech environment

Required Qualifications

Education

  • Bachelor's degree in Computer Science, Computer Engineering, or related field

Technical Knowledge

  • Understanding of computer systems and cloud environments
  • Understanding of network protocols and system connectivity

Systems Skills

  • Experience using Linux command line tools
  • Familiarity with SSH, SFTP, Bash, IP, ports, and networking concepts

Data Engineering Skills

  • Experience developing and optimizing ETL/ELT pipelines
  • Experience with tools such as Informatica PowerCenter, Kafka, NiFi, or similar

Data & Analytics

  • Strong SQL skills
  • Understanding of data warehousing concepts and schema design
  • Experience with data modeling and data lifecycle management

Programming

  • Programming experience using Python
  • Scripting experience using Bash or similar tools

Experience

  • 2+ years of experience working with ETL or data engineering workflows
  • Experience with Informatica PowerCenter and MDM
  • Experience developing and scaling data models within modern data warehouse architectures
  • Experience implementing data enrichment and transformation logic
  • Experience optimizing queries and storage design for analytics workloads
  • At least 1 year of experience in fintech, regulated environments, or PCI/DSS compliance is preferred
  • Familiarity with performance monitoring tools

More Info

Job Type:
Industry:
Function:
Employment Type:

About Company

Job ID: 145571815

Similar Jobs