
Overview
Specific Duties and Responsibilities:
Key Responsibilities:
● Design, develop, and implement data pipelines and ETL processes using Microsoft Fabric. ● Work with SQL and PySpark to extract, transform, and load data from various sources.
● Develop and maintain data models and schemas to support business analytics and reporting.
● Collaborate with data analysts, data scientists, and other stakeholders to understand data requirements and deliver solutions.
● Ensure data quality, integrity, and security throughout the data lifecycle.
● Optimize data processing workflows for performance and scalability.
● Stay updated with the latest advancements in data engineering technologies and best practices. Key Skills:
● Microsoft Fabric: Enthusiasm for working within the newly launched Microsoft Fabric environment for data engineering tasks.
● SQL: Strong knowledge of SQL for querying databases, data manipulation, and optimizing queries.
● Python (PySpark): Proficiency in Python, with experience in using PySpark for distributed data processing.
● Data Warehousing: Understanding of data warehouse concepts, including data modeling, ETL processes, and data architecture.
● Problem-Solving: Excellent analytical and problem-solving skills to address data engineering challenges.
● Communication: Strong verbal and written communication skills, with the ability to collaborate effectively with stakeholders.
Qualifications:
● Bachelor's degree in Computer Science, Engineering, Data Science, or a related field.
● 2-3 years of experience as a Data Engineer or in a similar role.
● Experience with cloud platforms and services is a plus.
● Familiarity with additional data engineering tools and programming languages is a bonus.
Job Type: Full-time
Pay: ₹1,000,000.00 - ₹2,000,000.00 per year
Benefits:
- Health insurance
- Provident Fund
Schedule:
- Day shift
Supplemental Pay:
- Performance bonus
- Yearly bonus
Work Location: In person