Apply on
At PwC, we specialize in providing cutting-edge data solutions that drive business success. We partner with clients across diverse industries to understand their data needs, design robust data architectures, and implement scalable data solutions. Our team is dedicated to solving complex data challenges and delivering strategic solutions that align with our clients' goals.
As a Data Engineer, you will be responsible for designing, developing, and managing data architecture and pipelines. You will collaborate with solution architect, devops engineer, data scientist, and other stakeholders to create reliable and efficient data systems using cutting-edge technologies such as GenAI, Machine Learning, etc.
In this role, you will also get the opportunity to enhance your expertise across a broad spectrum of data-related disciplines such as data governance, data strategies, data science, and project management.
Key Responsibilities
- Design, implement, and maintain scalable data pipelines to import, process, and integrate data from various sources.
- Build and optimize data warehousing solutions to support business intelligence and analytics needs.
- Develop and manage ETL processes to ensure timely and accurate data transformation and loading.
- Monitor and enhance data quality, implementing data validation and cleansing processes to ensure accuracy and consistency.
- Partner with solution architect, business analyst, devops engineer, data scientist, and other stakeholders to understand data requirements and provide technical solutions.
- Tune and optimize data processing and query performance to handle large datasets efficiently.
- Create and maintain comprehensive documentation for data pipelines, processes, and systems.
- Stay current with industry trends and emerging technologies to continuously improve data engineering practices.
Experience And Qualifications
- Bachelor’s Degree in Computer Science, Information Technology, Engineering or a similar field.
- Minimum 2-3 years of experience as a Data Engineer or in a related role, with a proven track record in data pipeline development and data management.
- Proficient in cloud platforms (e.g. AWS Glue, AWS Lambda, AWS Redshift, Azure Data Factory, Databricks, Synapse Analytic and BI tools like SSIS, SSAS).
- Strong programming skills in Python, Java, or Scala. Experience with data processing frameworks and libraries.
- Familiarity with ETL tools, data warehousing solutions, and data visualization tools is advantageous.
- Knowledge of containerization and orchestration tools (e.g., Docker, Kubernetes).
- Excellent analytical and troubleshooting skills with the ability to resolve complex data-related issues.
- Strong verbal and written communication skills. Ability to work effectively with cross-functional teams.
- Understanding of data governance, data privacy, and security best practices.