Data Engineer-Data Platforms-AWS
IBM
**Introduction**
In this role, you'll work in one of our IBM Consulting Client Innovation Centers (Delivery Centers), where we deliver deep technical and industry expertise to a wide range of public and private sector clients around the world. Our delivery centers offer our clients locally based skills and technical expertise to drive innovation and adoption of new technology. A career in IBM Consulting is rooted by long-term relationships and close collaboration with clients across the globe. You'll work with visionaries across multiple industries to improve the hybrid cloud and AI journey for the most innovative and valuable companies in the world. Your ability to accelerate impact and make meaningful change for your clients is enabled by our strategic partner ecosystem and our robust technology platforms across the IBM portfolio; including Software and Red Hat. Curiosity and a constant quest for knowledge serve as the foundation to success in IBM Consulting. In your role, you'll be encouraged to challenge the norm, investigate ideas outside of your role, and come up with creative solutions resulting in ground breaking impact for a wide network of clients. Our culture of evolution and empathy centers on long-term career growth and development opportunities in an environment that embraces your unique skills and experience
**Your role and responsibilities**
* Design, construct, install, test, and maintain highly scalable data management systems using big data technologies such as Apache Spark (with focus on Spark SQL) and Hive. Manage and optimize our data warehousing solutions, with a strong emphasis on SQL performance tuning. Implement ETL/ELT processes using tools like Talend or custom scripts, ensuring efficient data flow and transformation across our systems.
* Utilize AWS services including S3, EC2, and EMR to build and manage scalable, secure, and reliable cloud-based solutions.
* Develop and deploy scripts in Linux environments, demonstrating proficiency in shell scripting. Utilize scheduling tools such as Airflow or Control-M to automate data processes and workflows.
* Implement and maintain metadata-driven frameworks, promoting reusability, efficiency, and data governance. Collaborate closely with DevOps teams utilizing SDLC tools such as Bamboo, JIRA, Bitbucket, and Confluence to ensure seamless integration of data systems into the software development lifecycle.
* Communicate effectively with both technical and non-technical stakeholders, for handover, incident management reporting, etc
**Required technical and professional expertise**
* Demonstrated expertise in Big Data Technologies, specifically Apache Spark (focus on Spark SQL) and Hive.
* Extensive experience with AWS services, including S3, EC2, and EMR.
* Strong expertise in Data Warehousing and SQL, with experience in performance optimization
* Experience with ETL/ELT implementation (such as Talend)
* Proficiency in Linux, with a strong background in shell scripting
**Preferred technical and professional experience**
* Familiarity with scheduling tools like Airflow or Control-M.
* Experience with metadata-driven frameworks.
* Knowledge of DevOps tools such as Bamboo, JIRA, Bitbucket, and Confluence. Excellent communication skills and a willing attitude towards learning
Confirm your E-mail: Send Email
All Jobs from IBM