About RealPlay We’re an established and growing online gaming company operating in the sweepstakes casino space. At RealPlay, we believe in delivering fun, fair and engaging experiences to our players—powered by data-driven insights and a culture of continuous improvement. About the Role We’re building out a dedicated Data & Analytics team—and we’re looking for a Data Engineer to own the design, implementation and operation of our data platform. You will work hands-on to build scalable, reliable data pipelines and models using BigQuery and dbt, laying the groundwork for advanced analytics, reporting, and ML workloads as we scale. If you love turning raw event streams into clean, well-documented data assets and collaborating with analysts, scientists and product teams, this is your chance to make a big impact. Requirements * 3+ years of professional experience building data pipelines and warehouses. * Strong proficiency in SQL, especially in a cloud data warehouse context (BigQuery preferred). * Hands-on experience with dbt (data modeling, testing, documentation). * Solid programming skills in Python (for data integration, scripting, custom transformations). * Familiarity with workflow orchestration tools (e.g., Apache Airflow, Prefect, or similar). * Experience with version control (Git), and setting up CI/CD for data projects. * Deep understanding of data warehousing concepts (star/snowflake schemas, slowly changing dimensions, incremental loads). * Excellent communication skills and a collaborative mindset.
Responsibilities * Pipeline Development: Design, build and maintain end-to-end ETL/ELT pipelines in BigQuery and orchestrate them via dbt (and, where needed, Airflow or other schedulers). * Вata Modeling: Author modular, testable dbt models that follow best practices (e.g. star schemas, incremental builds), ensuring data is performant and maintainable. * Performance & Cost Optimization: Monitor and optimize query performance, table partitioning/clustering, and storage costs within BigQuery. * Data Quality & Observability: Implement testing, monitoring and alerting frameworks (dbt tests, dataflow metrics, BigQuery INFORMATION_SCHEMA checks) to ensure data accuracy and reliability. * Collaboration: Partner closely with Data Analysts, Data Scientists and Product Managers to understand reporting requirements, translate them into technical solutions, and deliver timely data assets. * Documentation & Governance: Maintain clear documentation of data sources, pipeline logic, schema definitions and data lineage. Help establish standards for naming, versioning and access control. * Automation & CI/CD: Build and maintain CI/CD pipelines for dbt projects and supporting Python/SQL code (Git integration, automated testing, deployment to production) together with the DevOps team. * Troubleshooting & Support: Quickly investigate and resolve data pipeline failures, schema changes upstream, and ad-hoc data requests from stakeholders.
Would be a plus * Experience on Google Cloud Platform beyond BigQuery (Pub/Sub, Dataflow, Cloud Storage). * Knowledge of containerization and infrastructure as code (Docker, Terraform, Kubernetes). * Exposure to real-time or streaming data architectures. * Familiarity with orchestration in a microservices environment. * Background in gaming, sweepstakes, or digital marketing analytics.