ThetaRay Logo

ThetaRay

Senior Data Engineer

Posted 22 Days Ago
Be an Early Applicant
In-Office or Remote
Hiring Remotely in Madrid, Comunidad de Madrid
Senior level
In-Office or Remote
Hiring Remotely in Madrid, Comunidad de Madrid
Senior level
As a Senior Data Engineer, you will design, implement, and optimize data pipelines, support data scientists, and build data tools for analytics.
The summary above was generated by AI

About ThetaRay:

ThetaRay is a trailblazer in AI-powered Anti-Money Laundering (AML) solutions, offering cutting-edge technology to fintechs, banks, and regulatory bodies worldwide. Our mission is to enhance trust in financial transactions, ensuring compliant and innovative business growth. 

Our technology empowers customers to expand into new markets and introduce groundbreaking products.

Thetaray is a culture-driven company. Our values are at the heart of our success. By joining us, you'll have the opportunity to embody these values and inspire others through your actions.

Why Join ThetaRay?

At ThetaRay, you'll be part of a dynamic global team committed to redefining the financial services sector through technological innovation. You will contribute to creating safer financial environments and have the opportunity to work with some of the brightest minds in AI, ML, and financial technology. We offer a collaborative, inclusive, and forward-thinking work environment where your ideas and contributions are valued and encouraged.

Join us in our mission to revolutionize the financial world, making it safer and more trustworthy for millions worldwide. Explore exciting career opportunities at ThetaRay – where innovation meets purpose.

We are looking for a Senior Data Engineer to join our growing team of data experts. As a Senior Data Engineer, you will be responsible for designing, implementing, and optimizing data pipeline flows within the ThetaRay system. You will support our data scientists with the implementation of the relevant data flows based on the data scientist’s features design and construct complex rules to detect money laundering activity.

The ideal candidate has experience in building data pipelines and data transformations and enjoys optimizing data flows and building them from the ground up. They must be self-directed and comfortable supporting multiple production implementations for various use cases.


Responsibilities

  • Implement and maintain data pipeline flows in production within the ThetaRay system
  • Design and implement data flows for ThetaRay products
  • Building a Machine Learning big data pipeline
  • Create data tools for analytics and data scientist team members that assist them in building and optimizing our product into an innovative industry leader
  • Work with product, R&D, data, and analytics experts to strive for greater functionality in our systems
  • Travel to customer locations both domestically and abroad
  • Build and manage technical relationships with customers and partners
  • 6+ years of experience as a data engineer - must
  • Hands-on experience working with Apache Spark, with Pyspark or Scala - must
  • Hands-on experience with SQL - must
  • Hands-on experience with version-control tools such as GIT - must
  • Experience working with and optimizing big data pipelines and architectures
  • Hands-on experience with data transformation, validations, cleansing, and ML feature engineering - must
  • BSc degree or higher in Computer Science, Statistics, Informatics, Information Systems, Engineering, or another quantitative field - must
  • Strong analytic skills related to working with structured and semi-structured datasets - must
  • Business-oriented and able to work with external customers and cross-functional teams - must
  • Fluent in English - must

Nice to have

  • Experience with Linux
  • Experience in building Machine Learning lifecycle
  • Experience with Jupyter
  • Experience with workflow automation platforms such as Airflow. N8N
  • Experience with Microservices architecture components, including Docker and Kubernetes
  • Experience in GitHub Copilot, Codex, etc
  • Hands-on experience with Apache Hadoop Ecosystem including Hive, Impala, Hue, HDFS, Sqoop etc.
  • Hands-on experience with Agentic AI

Top Skills

Spark
Git
Pyspark
Scala
SQL

Similar Jobs

4 Days Ago
Remote
6 Locations
Senior level
Senior level
Artificial Intelligence • Information Technology • Consulting
The Senior Data Engineer will build scalable APIs, design and maintain ETL/ELT pipelines, and collaborate with ML Engineers to create robust data solutions.
Top Skills: AirflowApache DruidAWSAws CloudformationDagsterDbtKafkaPythonSparkSQLTerraform
11 Days Ago
Remote
28 Locations
Senior level
Senior level
Information Technology
Lead the design and implementation of strategic data architecture, develop and maintain data pipelines, ensure data governance and quality, and mentor team members.
Top Skills: Apache FlinkSparkAWSBigQueryDbtGCPPythonRedshiftScalaSnowflakeSQL
16 Days Ago
In-Office or Remote
6 Locations
Senior level
Senior level
Artificial Intelligence • Information Technology • Consulting
Join Provectus as a Senior Data Engineer, working on ML infrastructure, data engineering projects, and collaborating with a multidisciplinary team to resolve technical challenges.
Top Skills: AirflowApache DruidAWSAws CloudformationDagsterDatabricksDbtFastapiFlaskKafkaLookerPower BIPythonQuicksightSnowflakeSparkSQLTableauTerraform

What you need to know about the Manchester Tech Scene

Home to a £5 billion digital ecosystem, including MediaCity, which consists of major players like the BBC, ITV and Ericsson, Manchester is one of the U.K.'s top digital tech hubs, at the forefront of advancements in film, television and emerging sectors like as e-sports, while also fostering a community of professionals dedicated to pushing creative and technological boundaries.

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account