Job Overview
As a GCP Data Engineer, you’ll help design and develop robust, scalable data solutions that power real-time analytics and business insights. You’ll work with cutting-edge open source tools and Google Cloud technologies to transform data into business value.
Key Responsibilities
- Build and maintain data-ingestion frameworks and real-time processing solutions
- Design data transformation pipelines using Kafka, Apache Spark (SQL, Scala, Java), Python, Hadoop, Hive, Presto, Druid, Airflow
- Optimize and support BigQuery for data warehousing, ensuring high performance and efficiency
- Develop data models and metrics using LookML
- Configure and manage DataProc clusters for running Hadoop/Spark jobs on GCP
- Collaborate with teams to operationalize data analytics tools for enterprise use
- Implement and validate data governance policies, including data lineage, quality checks, and classification
- Provide system support as part of a team rotation
- Follow SDLC processes to design, deploy, and deliver reliable solutions
Skills Required
Strong hands-on experience with:
- Big Data technologies: Kafka, Apache Spark, Hadoop ecosystem
- Cloud tools: BigQuery, DataProc on GCP
- Data modeling: LookML
- Scripting & development: Python, Java, Scala
- Workflow tools: Airflow
Understanding of:
- BigQuery architecture and performance optimization
- Data governance and data quality best practices
- SDLC processes for delivering production-ready solutions
Qualifications
Bachelor’s degree in Computer Science, Information Systems, or related field
Experience working on projects involving software development life cycles (SDLC)
Why You’ll Love Working With Us
- Be part of a collaborative team that truly believes in its mission
- Work with cutting-edge tools and tackle meaningful technical challenges
- Enjoy a growth-driven, people-first culture that celebrates innovation and initiative
- Thrive in an environment where your work drives real-world impact and success