
Overview
The primary purpose of this role is to translate business requirements and functional specifications into logical program designs and to deliver code modules, stable application systems, and software solutions using PySpark.
The core responsibility is to develop, configure, or modify the integrated business and/or enterprise application solutions within various computing environments. This role facilitates the implementation and maintenance of business and enterprise software solutions to ensure the successful deployment of released applications. Minimum Qualifications ¿ 2 years of experience in software development or a related field ¿ 1 year of experience in database technologies
1 year of experience working on project(s) involving the implementation of solutions applying development life cycles (SDLC) using PySpark PySpark Developers need to possess a range of technical skills to effectively work with PySpark and Spark clusters. Some key technical skills include: PySpark and Spark: Proficiency in PySpark, including the Spark DataFrame API and RDD (Resilient Distributed Datasets) programming model. Knowledge of Spark internals, data partitioning, and optimization techniques is advantageous. Data Manipulation and Analysis: Ability to manipulate and analyze large datasets using PySpark¿s DataFrame transformations and actions. This includes filtering, aggregating, joining, and performing complex data transformations. Distributed Computing: Understanding of distributed computing concepts, such as parallel processing, cluster management, and data partitioning. Experience with Spark cluster deployment, configuration, and optimization is valuable.
Data Serialization and Formats: Knowledge of different data serialization formats like JSON, Parquet, Avro, and CSV. Familiarity with handling unstructured data and working with NoSQL databases like Hadoop HBase or Apache Cassandra. Data Pipelines and ETL: Experience in building data pipelines and implementing Extract, Transform, Load (ETL) processes using PySpark. Understanding of data integration, data cleansing, and data quality techniques. Scope for Request: Scope for Request: To cover EVA Production support.
Job Type: Full-time
Pay: ₹681,345.20 - ₹2,084,799.84 per year
Schedule:
- Day shift
Work Location: In person