- Professional
- Office in Tamil Nadu
- Design, develop, and implement scalable data pipelines for batch and real-time processing.
- Collaborate with data engineering and analytics teams to understand data requirements and deliver effective solutions.
- Optimize ETL/ELT workflows using Python and PySpark to handle large-scale datasets.
- Write advanced SQL queries to process, transform, and analyze data efficiently.
- Handle JSON data for integration, serialization, and deserialization in distributed systems.
- Utilize Azure cloud services to deploy, manage, and maintain data solutions.
- Troubleshoot and resolve issues in data pipelines and workflows.
- Proven experience in designing and implementing data integration processes.
- Advanced proficiency in Python, SQL, and JSON.
- Hands-on experience with PySpark technologies and distributed computing frameworks.
- Practical expertise in Azure cloud services (e.g., Azure Data Lake, Azure Synapse etc.).
- Strong problem-solving and analytical skills.
- Ability to work independently and collaboratively in a fast-paced environment.
- Familiarity with additional cloud platforms (GCP or AWS) is a plus.
- Experience with CI/CD tools and data versioning.
- Knowledge of data modeling and big data technologies