Platzhalter Bild

Hybrid IN_Senior Associate_Hadoop Developer_Advisory Corporate_Advisory_Bangalore at PwC

PwC ·  Bengaluru Millenia, India · Hybrid

Apply Now Logo

Exponent – Ace your tech interviews with mock sessions and expert coaches from top companies.

Sponsored by Exponent

Line of Service

Advisory

Industry/Sector

Not Applicable

Specialism

SAP

Management Level

Senior Associate

Job Description & Summary

A career within Data and Analytics services will provide you with the opportunity to help organisations uncover enterprise insights and drive business results using smarter data analytics. We focus on a collection of organisational technology capabilities, including business intelligence, data management, and data assurance that help our clients drive innovation, growth, and change within their organisations in order to keep up with the changing nature of customers and technology. We make impactful decisions by mixing mind and machine to leverage data, understand and navigate risk, and help our clients gain a competitive edge.

*Why PWC
At PwC, you will be part of a vibrant community of solvers that leads with trust and creates distinctive outcomes for our clients and communities. This purpose-led and values-driven work, powered by technology in an environment that drives innovation, will enable you to make a tangible impact in the real world. We reward your contributions, support your wellbeing, and offer inclusive benefits, flexibility programmes and mentorship that will help you thrive in work and life. Together, we grow, learn, care, collaborate, and create a future of infinite experiences for each other. Learn more about us.
At PwC, we believe in providing equal employment opportunities, without any discrimination on the grounds of gender, ethnic background, age, disability, marital status, sexual orientation, pregnancy, gender identity or expression, religion or other beliefs, perceived differences and status protected by law. We strive to create an environment where each one of our people can bring their true selves and contribute to their personal growth and the firm’s growth. To enable this, we have zero tolerance for any discrimination and harassment based on the above considerations. 

Responsibilities:

  • 6-7+ years of Experience in Hadoop or any Cloud Bigdata components (specific to the Data Engineering role), Hadoop hands on (airflow, oozie, hive, hdfs, sqoop, pig, map, reduce )
  • 4+ years of Exp in Spark (Spark Batch, Streaming, Mlib  etc). Candidates should possess proficiency in utilizing the Apache Spark framework.
  • 6-7+ years of experience in Python programming language.
  • 4+ years of experience with pyspark data transformation (json,csv,rdbms,stream) pipeline design , development and deployment with kubernate/onprem platform (not cloud based).
  • 2+ years of experience in designing and implementing data workflows with Apache Airflow.
  • Kafka or equivalent Cloud Bigdata components (specific to the Data Engineering role)
  • Exposure in Oracle, MySql, SQL Server, DB2, Teradata, SPARK SQL, POSTGRES SQL, Spark SQL
  • Unix/Shell Scripting experience.
  • Cloud technologies GCP preferable.

Additional Requirements:

  • Exposure to Large enterprise data
  • Experience in application support and maintenance of spark applications
  • Experience in optimize and tune the performance to handle large and medium scale data volume with spark.
  • Experience in performance tuning techniques for large-scale data processing.
  • Experience working with Continuous Integration/Continuous Deployment tools
  • Experience working on project(s) involving the implementation of solutions applying development life cycles (SDLC)
  • Adherence to clean coding principles: Candidates should be capable of producing code that is devoid of bugs and can be easily understood and replicated by other developers.
  • Strong teamwork abilities: developers typically collaborate closely with data scientists and other backend developers. Therefore, candidates should exhibit excellent communication and collaboration skills.

Mandatory skill sets:

Hadoop, Hive, Pyspark, Python

Preferred skill sets:

Hadoop, Hive, Pyspark, Python

Years of experience required:

5 - 8

Education qualification:

B.Tech / M.Tech / MBA / MCA

Education (if blank, degree and/or field of study not specified)

Degrees/Field of Study required: Bachelor of Engineering, Master of Business Administration, Master of Engineering

Degrees/Field of Study preferred:

Certifications (if blank, certifications not specified)

Required Skills

Hadoop Framework

Optional Skills

Accepting Feedback, Accepting Feedback, Active Listening, Agile Scalability, Amazon Web Services (AWS), Analytical Thinking, Apache Hadoop, Azure Data Factory, Communication, Creativity, Data Anonymization, Database Administration, Database Management System (DBMS), Database Optimization, Database Security Best Practices, Data Engineering, Data Engineering Platforms, Data Infrastructure, Data Integration, Data Lake, Data Modeling, Data Pipeline, Data Quality, Data Transformation, Data Validation {+ 18 more}

Desired Languages (If blank, desired languages not specified)

Travel Requirements

Available for Work Visa Sponsorship?

Government Clearance Required?

Job Posting End Date

Apply Now

Other Jobs