UST is looking for a talented GCP Data Engineer with 5 to 10 years of experience to join our team and play a crucial role in designing and implementing efficient data solutions on the Google Cloud Platform (GCP). The ideal candidate should possess strong data engineering skills, expertise in GCP services, and proficiency in data processing technologies, particularly PySpark.
Responsibilities:
Data Pipeline Development:
§ Design, implement, and optimize end-to-end data pipelines on GCP, focusing on scalability and performance.
§ Develop and maintain ETL workflows for seamless data processing.
GCP Cloud Expertise:
§ Utilize GCP services such as BigQuery, Cloud Storage, and Dataflow for effective data engineering.
§ Implement and manage data storage solutions on GCP.
Data Transformation with PySpark:
§ Leverage PySpark for advanced data transformations, ensuring high-quality and well-structured output.
§ Implement data cleansing, enrichment, and validation processes using PySpark.
Requirements: