Cloud Data Engineer bei Permian Supply Chain CC
Permian Supply Chain CC · Houston, Vereinigte Staaten Von Amerika · Onsite
- Professional
- Optionales Büro in Houston
3+ years of expertise in building scalable distributed data processing solutions with Azure Data Lake Storage and Azure Synapse Data Warehouse, understanding of the functionalities - Identity Management, Security, Data Governance, DevOps, and Operations on Azure Platform. Expert in managing data models, Pyspark and Python code, data pipelines, data load failure issues, and optimization of existing ML spark jobs for better performance.
Responsibilities
Skills and Abilities:
Ability to:
- Build and manage Data Pipelines
- Build and manage data models
- MS Fabric/Azure Synapse/Data Factory
- Azure Data Lake Storage /ADLS Gen2
- Oracle PL/Sql, TSQL (Queries, Functions, Stored Procedures)
- Python and Pyspark
Knowledge:
- Python, Pyspark
- Azure DevOps
Minimum Education and Certification:
- Bachelor's degree
Minimum Experience:
- 3+ years of IT and Data Warehouse experience
Leadership:
- Customer-focused professional who is motivated to drive the creation of great data engineering platforms.
- Contribute to the continued evolution of Corporate Analytics Platform
- Responsible for managing a growing cloud-based data ecosystem.
Problem Solving and Innovation:
- Ability to define and create complex data models to build insights, predictions, and innovations from data.
- Expected to identify opportunities and to provide ideas, methods and innovations to enhance teamwork, efficiency and quality.
This job description is a summary of essential job functions. It is not intended as an employment contract, nor is it intended to describe all duties someone in this position may perform. All employees are expected to perform tasks, as assigned by supervisor, regardless of job title or routine job duties.
Qualifications
3+ years of expertise in building scalable distributed data processing solutions with Azure Data Lake Storage and Azure Synapse Data Warehouse, understanding of the functionalities - Identity Management, Security, Data Governance, DevOps, and Operations on Azure Platform.
Expert in managing data models, Pyspark and Python code, data pipelines, data load failure issues, and optimization of existing ML spark jobs for better performance.