SR Data Engineer (MySQL Aurora to NoSQL MongoDB) - Canada Citizens Only

<p><strong>SR. Data Engineer (Monolith-to-Microservices)<em> </em></strong><u>SQL</u>: MySQL/Aurora/Redshift --2-- <u>NoSQL</u>: MongoDB/DocDB </p><p><br></p><p><em>Fulltime position with Vacation & Holidays & Company Benefits Program – Remote (Not Quebec) – Must be </em><strong><em>Canadian Citizen </em></strong></p><p><br></p><p><em>The company is striving to be the #1 Software for Fitness & Wellness Businesses. They offer a cloud-based business management software that is used by leading fitness gyms, yoga studios, and wellness centers. The platform is trusted by more than </em><strong><em>5,000</em></strong><em> businesses and has more than 15 million users. From a bootstrap start-up to one of North America’s </em><strong><em>fastest-growing</em></strong><em> </em><strong><em>SaaS</em></strong><em> companies, the company helps entrepreneurs grow, manage, and streamline their businesses to drive more revenue with a reliable, intuitive software. The company is experiencing dramatic growth and is building a suite of four</em><strong><em> Ai modules</em></strong><em> that will streamline client communications, bookings & payments, growth marketing, and business owner coaching. The company is headquartered in </em><strong><em>Toronto, Ontario, Canada </em></strong></p><p><br></p><p><strong><em>Monolith-to-Microservices Transition:</em></strong><em> Databases 1) </em><strong><em><u>SQL</u></em></strong><em>: </em><strong><em>MySQL, Aurora, Redshift</em></strong><em> -and- 2) </em><strong><em><u>NoSQL</u></em></strong><em>: </em><strong><em>MongoDB, DocumentDB</em></strong></p><p><br></p><p><strong><em><u>Tech Stack</u> - </em></strong><em><u>LLMs</u>: OpenAI, Gemini • <u>Voice</u>: Twilio, LiveKit, STT/TTS • <u>Backend</u>: Python, Node/TypeScript, REST/GraphQL • <u>Data</u>: Aurora Serverless (MySQL), Redshift, S3/Glue • <u>Infrast</u>: Docker, Kubernetes, GitHub Actions • <u>Mess’g</u>: Twilio, SendGrid, FCM, SES • <u>Integrations</u>: Nuvei/Paragon payments, Salesforce/HubSpot/Zapier. </em></p><p><br></p><p>This seasoned <strong>Senior Data Engineer</strong> will help lead the modernization of our data infrastructure as we transition from a tightly coupled monolithic system to a scalable, microservices-based architecture. This role is central to <strong>decoupling legacy database</strong> structures, enabling domain-driven service ownership, and powering real-time analytics, operational intelligence, and AI initiatives across our platform. Position will work closely with solution architects and domain owners to design resilient pipelines and data models that reflect business context and support scalable, secure, and auditable data access for internal and external consumers. </p><p><br></p><p><strong><u>Key Responsibilities</u></strong></p><ul><li><strong>Monolith-to-Microservices Data Transition:</strong> Lead the decomposition of monolithic database structures into <strong>domain-aligned schemas</strong> that enable service independence and ownership.</li><li><strong>Pipeline Development & Migration:</strong> Build and optimize ETL/ELT workflows using <strong>Python, PySpark/Spark, AWS Glue, </strong>and <strong>dbt</strong>, including schema/data mapping and transformation from on-prem and cloud legacy systems into data lake and warehouse environments.</li><li><strong>Domain Data Modeling:</strong> Define logical and physical <strong>domain-driven data models</strong> (star/snowflake schemas, data marts) to serve cross-functional needs, BI, operations, streaming, and ML.</li><li><strong>Legacy Systems Integration:</strong> Design strategies for <strong>extracting, validating, and restructuring</strong> data from legacy systems with embedded logic and incomplete normalization.</li><li><strong>Database Management:</strong> Administer, optimize, and scale<strong> SQL </strong>(<strong>MySQL, Aurora, Redshift</strong>) and <strong>NoSQL (MongoDB)</strong> platforms to meet high-availability and low-latency needs.</li><li><strong>Cloud & Serverless ETL:</strong> Leverage <strong>AWS Glue Catalog, Crawlers, Lambda</strong>, and <strong>S3 </strong>to manage and orchestrate modern, cost-efficient data pipelines.</li><li><strong>Data Governance & Compliance:</strong> Enforce best practices around cataloging, lineage, retention, access control, and security, ensuring compliance with <strong>GDPR, CCPA, PIPEDA</strong>, and internal standards.</li><li><strong>Monitoring & Optimization:</strong> Implement observability (CloudWatch, logs, metrics) and performance tuning across Spark, Glue, and Redshift workloads.</li><li><strong>Stakeholder Collaboration:</strong> Work with architects, analysts, product managers, and data scientists to define, validate, and prioritize requirements.</li><li><strong>Documentation & Mentorship:</strong> Maintain technical documentation (data dictionaries, migration guides, schema specs) and mentor junior engineers in engineering standards. </li></ul><p><br></p><p><strong><u>REQUIRED Qualifications </u></strong></p><ul><li>5+ years in data engineering with a proven record in <strong>modernizing legacy data systems</strong> and <strong>driving</strong> <strong>large-scale migration</strong> initiatives.</li><li><u>Cloud ETL Expertise:</u> Proficient in <strong>AWS Glue, Apache Spark/PySpark</strong>, and modular transformation frameworks like <strong>dbt</strong>.</li><li><u>Data Modeling:</u> Strong grasp of <strong>domain-driven design</strong>, bounded contexts, and BI-friendly modeling approaches (<strong>star/snowflake/data vault</strong>).</li><li><u>Data Migration:</u> Experience with <strong>full lifecycle migrations</strong> including schema/data mapping, reconciliation, and exception handling.</li><li><u>Databases:</u> 1) <strong><u>SQL</u></strong>: <strong>MySQL, Aurora, Redshift</strong> -and- 2) <strong><u>NoSQL</u></strong>: <strong>MongoDB, DocumentDB</strong></li><li><u>Programming:</u> Strong <strong>Python</strong> skills for data wrangling, pipeline automation, and <strong>API </strong>interactions.</li><li><u>Data Architecture:</u> Hands-on with <strong>data lakes</strong>, warehousing strategies, and hybrid cloud data ecosystems.</li><li><u>Compliance & Security:</u> Track record implementing governance, data cataloging, encryption, retention, lineage, and <strong>RBAC</strong>.</li><li><u>DevOps Practices:</u> <strong>Git, CI/CD</strong> pipelines, <strong>Docker</strong>, and test automation for data pipelines.</li><li><strong><u>Must be Canadian Citizen</u></strong>. REMOTE - But we can NOT hire anyone living in Quebec. Ideal candidate lives near Toronto on EDT. </li></ul><p><br></p><p><strong><u>PREFERRED Qualifications </u></strong></p><ul><li>Degreed, and experience with <u>streaming data platforms</u> like <strong>Kafka, Kinesis,</strong> or <strong>CDC tools</strong> such as <strong>Debezium</strong>.</li><li>Familiarity with <u>orchestration platforms</u> like <strong>Airflow </strong>or <strong>Prefect</strong></li><li>Background in analytics, <strong>data modeling for AI/ML pipelines</strong>, or ML-ready data preparation</li><li>Understanding of <u>cloud-native data services</u> (<strong>AWS Glue, Redshift, Snowflake, BigQuery</strong>, etc.)</li><li>Strong written and verbal communication skills. Self-starter with ability to navigate ambiguity and legacy system complexity.</li><li>Exposure to generative <strong>AI, LLM fine-tuning</strong>, or feature store design is a plus.</li></ul>

Back to blog