Role Type: Contract position. Engagement options available on W2 or C2C basis.
Location: Primarily remote. Occasional onsite presence may be required based on project needs. Preference will be given to candidates within driving distance of Woonsocket, RI.
Job Title: BigQuery Developer / GCP Data Engineer
Job Summary:
We are looking for a motivated and detail-oriented BigQuery Developer with hands-on experience in Google Cloud Platform to support and enhance our enterprise data warehouse and analytics solutions. The ideal candidate will have strong SQL , BigQuery and Python development experience, along with working knowledge of Dataflow, Cloud Composer, GitHub, and CI/CD practices. This role requires strong analytical skills, problem-solving ability, and effective communication to work with cross-functional teams.
Key Responsibilities
· BigQuery Development (Primary Focus)
· Develop, maintain, and optimize BigQuery datasets, tables, views, procedures and queries.
· Write efficient and scalable SQL for reporting and analytics.
· Implement partitioning and clustering to improve query performance.
· Support data warehouse design and data modeling activities.
· Monitor query performance and optimize cost usage.
· Troubleshoot and resolve data-related issues in BigQuery.
· Support data validation and quality checks.
Data Pipeline Development
· Develop and maintain batch pipelines using Python and Google Cloud Dataflow (Apache Beam).
· Load, transform, and integrate data from various sources into BigQuery.
· Work on ETL/ELT processes and ensure reliable data processing.
· Assist in debugging and performance tuning of pipelines.
Workflow Orchestration
· Develop and maintain workflows using Cloud Composer (Apache Airflow).
· Integrate workflows with Tidal Job Scheduler for enterprise scheduling.
· Monitor production jobs and support issue resolution.
Version Control & CI/CD
· Use GitHub for source control and collaboration.
· Contribute to CI/CD pipelines using GitHub Actions.
· Follow best practices for code versioning and peer reviews.
Collaboration & Communication
· Work closely with data analysts, business users, and technical teams.
· Translate business requirements into efficient BigQuery solutions.
· Document data flows, technical designs, and operational processes.
· Provide production support as needed.
· 3–5 years of experience in data engineering or data development.
· Strong hands-on experience with BigQuery.
· Strong SQL skills (joins, aggregations, window functions, performance tuning).
· Experience with Google Cloud Platform (GCP).
· Experience building batch pipelines using Python and Dataflow.
· Experience with Cloud Composer (Airflow).
· Working knowledge of GitHub and GitHub Actions.
· Experience with enterprise job schedulers such as Tidal.
· Understanding of data warehousing concepts.
· Strong analytical and problem-solving skills.
· Good verbal and written communication skills.