Posted on 
Jun 19, 2024

Principal Data Engineer - Data Platform & Ingress

Mid-Senior ICs
Engineering, Data Science + Analytics
Movable Ink
Movable Ink
Movable Ink
Series D
251-1000
Marketing & Marketing Tech

Movable Ink is a software company that provides marketers with technology and expert services to create unique customer experiences.

Job Description

As a Principal Data Engineer you will help drive the direction of our Data Warehouse.  You will be hands-on with teammates across all departments enabling them to access the data and empowering teams to make data-driven decisions about the direction of our business.  You will play a pivotal role in owning the infrastructure and code for our data pipelines within our Data Platform. You will be responsible for designing, implementing, and optimizing ingestion pipelines, enabling seamless extraction and loading of data from various sources such as customer warehouses, business systems, and events messaging buses. Movable Ink collects campaign data resulting from Billions of requests served each day.  Come and help us manage and make sense of the massive amount of data we’re ingesting! 

Responsibilities:

---------------------

  • Partner with internal operations teams to identify, collect, and integrate data from various business systems, ensuring comprehensive and accurate data capture
  • Design, implement, and maintain robust data pipelines that feed data into our Data Platform, ensuring high performance, scalability, and reliability
  • Ensure data pipelines adhere to best practices and are optimized for performance and scalability
  • Conduct thorough testing of data pipelines to validate data accuracy and integrity
  • Monitor data pipelines, troubleshoot any issues that arise, and make improvements to these issues where applicable
  • Establish and track SLAs for data processing and delivery, ensuring timely and reliable access to data for all users
  • Become a mentor for less experienced team members, and establish patterns and practices that can be followed to increase quality, accuracy, and efficiency of solutions produced by the team
  • Work with other teams in order to ensure access to data corresponds with company policies, and ensure data access, processing, and storage is in compliance with regulatory (e.g. GDPR, CCPA, etc.) requirements

Qualifications:

-------------------

  • 12+ years of professional experience in data engineering, software engineering, database administration, business intelligence, or related field with 8+ years of that experience as a Data Engineer with a focus on cloud-based Data Warehouse platforms (Redshift, Snowflake, Firebolt, BigQuery).  We currently use Redshift.
  • Elite-level understanding on how to work with and optimize multi-petabyte, mission-critical databases, ensuring high availability, performance, and reliability informed by a strong understanding of database internals
  • Elite-level proficiency with Python and SQL languages, and significant experience building robust data pipelines with these languages
  • Elite-level proficiency in using, deploying and managing at least one data pipeline orchestration tool/framework such as Apache Airflow, Prefect, etc.  We currently use Apache Airflow.
  • Significant experience in building solutions that comply with regulatory requirements such as GDPR and CCPA
  • Significant experience in designing and implementing solutions that can support both batch and real-time data consumption models
  • Significant experience in building solutions that implement data security best practices within an AWS environment
  • Significant experience in providing technical leadership, setting best practices, and successfully driving the adoption of new technologies and methodologies within a fast-moving organization
  • Significant data modeling experience spanning more than one data modeling paradigm (e.g. Data Vault, Kimball/Ross, Inmon)
  • Experience working in an Agile/Scrum environment, has experience working with technical managers and product owners/manager to break down high-level requirements into actionable cards
  • Experience working with streaming platforms such as Apache Kafka and Apache Pulsar
  • Excellent communication skills for effective collaboration with cross-functional teams
Movable Ink scales content personalization for marketers through data-activated content generation and AI decisioning. The world’s most innovative brands rely on Movable Ink to maximize revenue, simplify workflow and boost marketing agility. Headquartered in New York City with close to 600 employees, Movable Ink serves its global client base with operations throughout North America, Central America, Europe, Australia, and Japan.

As a Principal Data Engineer you will help drive the direction of our Data Warehouse.  You will be hands-on with teammates across all departments enabling them to access the data and empowering teams to make data-driven decisions about the direction of our business.  You will play a pivotal role in owning the infrastructure and code for our data pipelines within our Data Platform. You will be responsible for designing, implementing, and optimizing ingestion pipelines, enabling seamless extraction and loading of data from various sources such as customer warehouses, business systems, and events messaging buses. Movable Ink collects campaign data resulting from Billions of requests served each day.  Come and help us manage and make sense of the massive amount of data we’re ingesting! 

Responsibilities:

  • Partner with internal operations teams to identify, collect, and integrate data from various business systems, ensuring comprehensive and accurate data capture
  • Design, implement, and maintain robust data pipelines that feed data into our Data Platform, ensuring high performance, scalability, and reliability
  • Ensure data pipelines adhere to best practices and are optimized for performance and scalability
  • Conduct thorough testing of data pipelines to validate data accuracy and integrity
  • Monitor data pipelines, troubleshoot any issues that arise, and make improvements to these issues where applicable
  • Establish and track SLAs for data processing and delivery, ensuring timely and reliable access to data for all users
  • Become a mentor for less experienced team members, and establish patterns and practices that can be followed to increase quality, accuracy, and efficiency of solutions produced by the team
  • Work with other teams in order to ensure access to data corresponds with company policies, and ensure data access, processing, and storage is in compliance with regulatory (e.g. GDPR, CCPA, etc.) requirements

Qualifications:

  • 12+ years of professional experience in data engineering, software engineering, database administration, business intelligence, or related field with 8+ years of that experience as a Data Engineer with a focus on cloud-based Data Warehouse platforms (Redshift, Snowflake, Firebolt, BigQuery).  We currently use Redshift.
  • Elite-level understanding on how to work with and optimize multi-petabyte, mission-critical databases, ensuring high availability, performance, and reliability informed by a strong understanding of database internals
  • Elite-level proficiency with Python and SQL languages, and significant experience building robust data pipelines with these languages
  • Elite-level proficiency in using, deploying and managing at least one data pipeline orchestration tool/framework such as Apache Airflow, Prefect, etc.  We currently use Apache Airflow.
  • Significant experience in building solutions that comply with regulatory requirements such as GDPR and CCPA
  • Significant experience in designing and implementing solutions that can support both batch and real-time data consumption models
  • Significant experience in building solutions that implement data security best practices within an AWS environment
  • Significant experience in providing technical leadership, setting best practices, and successfully driving the adoption of new technologies and methodologies within a fast-moving organization
  • Significant data modeling experience spanning more than one data modeling paradigm (e.g. Data Vault, Kimball/Ross, Inmon)
  • Experience working in an Agile/Scrum environment, has experience working with technical managers and product owners/manager to break down high-level requirements into actionable cards
  • Experience working with streaming platforms such as Apache Kafka and Apache Pulsar
  • Excellent communication skills for effective collaboration with cross-functional teams
Receive Tech Ladies'
newest jobs in your inbox,
every week.

Join Tech Ladies for full-access to the job board, member-only events, and more!

If you're already a member, we haven't forgotten you. We promise. It's a new system. If you fill out the form once, it'll remember you going forward. Apologies for the inconvenience.

No items found.
Angular JS
Angular JS
AWS
AWS
CSS
CSS
AWS Redshift
AWS Redshift
Django
Django
Ember.js
Ember.js
Docker
Docker
Golang
Golang
GraphQL
GraphQL
HTML
HTML
Hadoop
Hadoop
Google Cloud Platform (GCP)
Google Cloud Platform (GCP)
JavaScript
JavaScript
TensorFlow
TensorFlow
Scikit
Scikit
React
React
React Native
React Native
Engineering
Engineering
Data Science + Analytics
Data Science + Analytics
Remote
Remote