← All Jobs
Posted Apr 15, 2026

BigQuery Developer / GCP Data Engineer

Apply Now
Description Role Type: Contract position. Engagement options available on W2 or C2C basis. Location: Primarily remote. Occasional onsite presence may be required based on project needs. Preference will be given to candidates within driving distance of Woonsocket, RI. Job Title: BigQuery Developer / GCP Data Engineer Job Summary: We are looking for a motivated and detail-oriented BigQuery Developer with hands-on experience in Google Cloud Platform to support and enhance our enterprise data warehouse and analytics solutions. The ideal candidate will have strong SQL , BigQuery and Python development experience, along with working knowledge of Dataflow, Cloud Composer, GitHub, and CI/CD practices. This role requires strong analytical skills, problem-solving ability, and effective communication to work with cross-functional teams. Key Responsibilities · BigQuery Development (Primary Focus) · Develop, maintain, and optimize BigQuery datasets, tables, views, procedures and queries. · Write efficient and scalable SQL for reporting and analytics. · Implement partitioning and clustering to improve query performance. · Support data warehouse design and data modeling activities. · Monitor query performance and optimize cost usage. · Troubleshoot and resolve data-related issues in BigQuery. · Support data validation and quality checks. Data Pipeline Development · Develop and maintain batch pipelines using Python and Google Cloud Dataflow (Apache Beam). · Load, transform, and integrate data from various sources into BigQuery. · Work on ETL/ELT processes and ensure reliable data processing. · Assist in debugging and performance tuning of pipelines. Workflow Orchestration · Develop and maintain workflows using Cloud Composer (Apache Airflow). · Integrate workflows with Tidal Job Scheduler for enterprise scheduling. · Monitor production jobs and support issue resolution. Version Control & CI/CD · Use GitHub for source control and collaboration. · Contribute to CI/CD pipelines using GitHub Actions. · Follow best practices for code versioning and peer reviews. Collaboration & Communication · Work closely with data analysts, business users, and technical teams. · Translate business requirements into efficient BigQuery solutions. · Document data flows, technical designs, and operational processes. · Provide production support as needed. Requirements · 3–5 years of experience in data engineering or data development. · Strong hands-on experience with BigQuery. · Strong SQL skills (joins, aggregations, window functions, performance tuning). · Experience with Google Cloud Platform (GCP). · Experience building batch pipelines using Python and Dataflow. · Experience with Cloud Composer (Airflow). · Working knowledge of GitHub and GitHub Actions. · Experience with enterprise job schedulers such as Tidal. · Understanding of data warehousing concepts. · Strong analytical and problem-solving skills. · Good verbal and written communication skills.
Interested in this role?Apply on iHire