Movable Ink scales content personalization for marketers through data-activated content generation and AI decisioning. The world’s most innovative brands rely on Movable Ink to maximize revenue, simplify workflow and boost marketing agility. Headquartered in New York City with close to 600 employees, Movable Ink serves its global client base with operations throughout North America, Central America, Europe, Australia, and Japan.
As a Principal Data Engineer you will help drive the direction of our Data Warehouse. You will be hands-on with teammates across all departments enabling them to access the data and empowering teams to make data-driven decisions about the direction of our business. You will play a pivotal role in owning the infrastructure and code for our data pipelines within our Data Platform. You will be responsible for designing, implementing, and optimizing ingestion pipelines, enabling seamless extraction and loading of data from various sources such as customer warehouses, business systems, and events messaging buses.
Movable Ink collects campaign data resulting from Billions of requests served each day. Come and help us manage and make sense of the massive amount of data we’re ingesting!
Responsibilities:
- Partner with internal operations teams to identify, collect, and integrate data from various business systems, ensuring comprehensive and accurate data capture
- Design, implement, and maintain robust data pipelines that feed data into our Data Platform, ensuring high performance, scalability, and reliability
- Ensure data pipelines adhere to best practices and are optimized for performance and scalability
- Conduct thorough testing of data pipelines to validate data accuracy and integrity
- Monitor data pipelines, troubleshoot any issues that arise, and make improvements to these issues where applicable
- Establish and track SLAs for data processing and delivery, ensuring timely and reliable access to data for all users
- Become a mentor for less experienced team members, and establish patterns and practices that can be followed to increase quality, accuracy, and efficiency of solutions produced by the team
- Work with other teams in order to ensure access to data corresponds with company policies, and ensure data access, processing, and storage is in compliance with regulatory (e.g. GDPR, CCPA, etc.) requirements
Qualifications:
- 12+ years of professional experience in data engineering, software engineering, database administration, business intelligence, or related field with 8+ years of that experience as a Data Engineer with a focus on cloud-based Data Warehouse platforms (Redshift, Snowflake, Firebolt, BigQuery). We currently use Redshift.
- Elite-level understanding on how to work with and optimize multi-petabyte, mission-critical databases, ensuring high availability, performance, and reliability informed by a strong understanding of database internals
- Elite-level proficiency with Python and SQL languages, and significant experience building robust data pipelines with these languages
- Elite-level proficiency in using, deploying and managing at least one data pipeline orchestration tool/framework such as Apache Airflow, Prefect, etc. We currently use Apache Airflow.
- Significant experience in building solutions that comply with regulatory requirements such as GDPR and CCPA
- Significant experience in designing and implementing solutions that can support both batch and real-time data consumption models
- Significant experience in building solutions that implement data security best practices within an AWS environment
- Significant experience in providing technical leadership, setting best practices, and successfully driving the adoption of new technologies and methodologies within a fast-moving organization
- Significant data modeling experience spanning more than one data modeling paradigm (e.g. Data Vault, Kimball/Ross, Inmon)
- Experience working in an Agile/Scrum environment, has experience working with technical managers and product owners/manager to break down high-level requirements into actionable cards
- Experience working with streaming platforms such as Apache Kafka and Apache Pulsar
- Excellent communication skills for effective collaboration with cross-functional teams
The base pay range for this position is $230,000 - $250,000 USD/ year. The base pay offered may vary depending on job-related knowledge, skills, and experience. Stock options and other incentive pay may be provided as part of the compensation package, in addition to a full range of medical, financial, and/or other benefits, depending on the position ultimately offered.Studies have shown that women, communities of color, and historically underrepresented people are less likely to apply to jobs unless they meet every single qualification.
We are committed to building a diverse and inclusive culture where all Inkers can thrive. If you’re excited about the role but don’t meet all of the abovementioned qualifications, we encourage you to apply. Our differences bring a breadth of knowledge and perspectives that makes us collectively stronger.We welcome and employ people regardless of race, color, gender identity or expression, religion, genetic information, parental or pregnancy status, national origin, sexual orientation, age, citizenship, marital status, ethnicity, family or marital status, physical and mental ability, political affiliation, disability, Veteran status, or other protected characteristics.
We are proud to be an equal opportunity employer.