
GCP Big Data +Java+ Apache Beam (RARR Job 5497)
Job Skills
Job Description
We are seeking a skilled GCP Developer with expertise in Apache Beam to design, build, and optimize scalable data pipelines and cloud-based solutions. The ideal candidate will have hands-on experience with Google Cloud Platform services and a strong background in big data processing, distributed systems, and data engineering practices.
Key Responsibilities:
-
Design, develop, and optimize data pipelines using Apache Beam and related frameworks.
-
Work extensively with Google Cloud services (BigQuery, Dataflow, Pub/Sub, Cloud Storage, Composer, etc.).
-
Implement data integration, transformation, and processing workflows for high-volume datasets.
-
Ensure performance, reliability, and scalability of data pipelines.
-
Collaborate with cross-functional teams including architects, data engineers, and business stakeholders.
-
Troubleshoot and resolve issues in large-scale distributed systems.
-
Follow best practices for CI/CD, DevOps, and cloud-native deployments.
Required Skills:
-
3–8 years of professional experience in data engineering / cloud development.
-
Strong expertise in Apache Beam (batch and streaming pipelines).
-
Hands-on experience with GCP services (Dataflow, BigQuery, Pub/Sub, GCS).
-
Proficiency in Python / Java / Scala for pipeline development.
-
Experience with SQL, ETL design, and data modeling.
-
Knowledge of CI/CD pipelines and version control (Git, Jenkins, Cloud Build, etc.).
-
Solid understanding of cloud security, monitoring, and logging practices.
Preferred Qualifications:
-
Experience in real-time data streaming and event-driven architectures.
-
Familiarity with Terraform / Infrastructure as Code for cloud deployments.
-
Exposure to Apache Kafka, Spark, or Airflow.
-
GCP Certification (Data Engineer, Professional Cloud Developer, or equivalent).