Esta oferta de empleo ya no se encuentra disponible. Continúa tu búsqueda de puestos aquí.
Data Engineer
Pune
Job No. atci-5535382-s2017551
Full-time
Descripción De La Posición
Project Role : Data Engineer
Project Role Description : Design, develop and maintain data solutions for data generation, collection, and processing. Create data pipelines, ensure data quality, and implement ETL (extract, transform and load) processes to migrate and deploy data across systems.
Must have skills : Microsoft Azure Databricks
Good to have skills : NA
Minimum 5 year(s) of experience is required
Educational Qualification : 15 years full time education
Summary:
As a Data Engineer, a typical day involves designing, developing, and maintaining comprehensive data solutions that support the generation, collection, and processing of data. This role requires creating efficient data pipelines and ensuring the integrity and quality of data throughout its lifecycle. The position also involves implementing processes to extract, transform, and load data, facilitating seamless migration and deployment across various systems. Collaboration with different teams to optimize data workflows and support organizational data needs is an integral part of daily activities, contributing to the overall data infrastructure and operational excellence.
Roles & Responsibilities:
- Expected to be an SME, collaborate and manage the team to perform.
- Responsible for team decisions.
- Engage with multiple teams and contribute on key decisions.
- Provide solutions to problems for their immediate team and across multiple teams.
- Lead efforts to identify and resolve data-related challenges, ensuring smooth project execution.
- Mentor junior team members to enhance their technical skills and professional growth.
- Coordinate cross-functional initiatives to align data engineering efforts with business objectives.
- Build and operate scalable Lakehouse pipelines on Databricks/Azure. Own ELT/streaming, Delta Lake optimization, Unity Catalog governance, and CI/CD. Integrate ADLS/ADF/Synapse, and deliver high-quality data sets for BI/ML/GenAI.
- Must-haves: PySpark, SQL, Databricks (Delta, DLT/Workflows), Azure data services, Unity Catalog, CI/CD.
- ML flow/Feature Store, Power BI, streaming/CDC, vector search/RAG, Terraform.
Professional & Technical Skills:
- Must To Have Skills: Proficiency in Microsoft Azure Databricks.
- Experience with building scalable and reliable data pipelines using cloud-based platforms.
- Strong knowledge of data integration techniques and ETL process design.
- Familiarity with data storage solutions and data warehousing concepts.
- Ability to optimize data workflows for performance and cost efficiency.
- Experience in troubleshooting and resolving complex data processing issues.
Additional Information:
- The candidate should have minimum 5 years of experience in Microsoft Azure Databricks.
- This position is based at our Pune office.
- A 15 years full time education is required.
Project Role Description : Design, develop and maintain data solutions for data generation, collection, and processing. Create data pipelines, ensure data quality, and implement ETL (extract, transform and load) processes to migrate and deploy data across systems.
Must have skills : Microsoft Azure Databricks
Good to have skills : NA
Minimum 5 year(s) of experience is required
Educational Qualification : 15 years full time education
Summary:
As a Data Engineer, a typical day involves designing, developing, and maintaining comprehensive data solutions that support the generation, collection, and processing of data. This role requires creating efficient data pipelines and ensuring the integrity and quality of data throughout its lifecycle. The position also involves implementing processes to extract, transform, and load data, facilitating seamless migration and deployment across various systems. Collaboration with different teams to optimize data workflows and support organizational data needs is an integral part of daily activities, contributing to the overall data infrastructure and operational excellence.
Roles & Responsibilities:
- Expected to be an SME, collaborate and manage the team to perform.
- Responsible for team decisions.
- Engage with multiple teams and contribute on key decisions.
- Provide solutions to problems for their immediate team and across multiple teams.
- Lead efforts to identify and resolve data-related challenges, ensuring smooth project execution.
- Mentor junior team members to enhance their technical skills and professional growth.
- Coordinate cross-functional initiatives to align data engineering efforts with business objectives.
- Build and operate scalable Lakehouse pipelines on Databricks/Azure. Own ELT/streaming, Delta Lake optimization, Unity Catalog governance, and CI/CD. Integrate ADLS/ADF/Synapse, and deliver high-quality data sets for BI/ML/GenAI.
- Must-haves: PySpark, SQL, Databricks (Delta, DLT/Workflows), Azure data services, Unity Catalog, CI/CD.
- ML flow/Feature Store, Power BI, streaming/CDC, vector search/RAG, Terraform.
Professional & Technical Skills:
- Must To Have Skills: Proficiency in Microsoft Azure Databricks.
- Experience with building scalable and reliable data pipelines using cloud-based platforms.
- Strong knowledge of data integration techniques and ETL process design.
- Familiarity with data storage solutions and data warehousing concepts.
- Ability to optimize data workflows for performance and cost efficiency.
- Experience in troubleshooting and resolving complex data processing issues.
Additional Information:
- The candidate should have minimum 5 years of experience in Microsoft Azure Databricks.
- This position is based at our Pune office.
- A 15 years full time education is required.
Requisitos
15 years full time education