
Overview
Location: Hyderabad / Remote (with quarterly visits)
Notice Period: 30 Days
Experience: 5+ Years
Responsibilities:
Design and develop business-critical backend systems using high-quality data pipelines.
Assemble large, complex datasets that meet functional and non-functional business requirements.
Build infrastructure for optimal data extraction, transformation, and loading (ETL) from various data sources using Azure Databricks and Azure Data Factory (ADF).
Develop a cloud-native, real-time stream processing and data lake platform that scales to massive data volumes.
Construct data pipelines that deliver actionable insights into customer acquisition, operational efficiency, and key business performance metrics.
Automate data pipelines using Logic Apps and DevOps practices.
Conduct root cause analysis on external and internal data processes to identify improvement opportunities.
Embrace continuous improvement, always seeking to enhance existing systems with better solutions.
Collaborate with stakeholders to design and code large-scale batch and real-time data pipelines on Azure.
Perform code reviews, emphasizing readability, extensibility, and testability.
Requirements:
Bachelor’s degree in Computer Science or a related field.
5+ years of experience as a Data Engineer.
Proficiency in Azure big data tools, including Azure Databricks, Azure Synapse, ADLS, and Azure Data Factory (ADF).
Strong knowledge of relational SQL and Azure Synapse.
Excellent problem-solving and analytical skills, especially in handling structured and unstructured datasets using Azure big data tools.
Hands-on experience with data pipeline and workflow management tools such as ADF and Logic Apps.
Familiarity with Azure cloud services, including Azure SQL DB, Azure Synapse, and Azure Databricks.
Proficiency in Python for scripting and automation.