Esta oferta de empleo ya no se encuentra disponible. Continúa tu búsqueda de puestos aquí.
Custom Software Engineer
Bengaluru
Job No. atci-5205801-s1910225
Full-time
Descripción De La Posición
Project Role : Custom Software Engineer
Project Role Description : Design, build and configure applications to meet business process and application requirements.
Must have skills : Data Architecture Principles
Good to have skills : NA
Minimum 12 year(s) of experience is required
Educational Qualification : 15 years full time education
Job Summary: Modern Data Architect to lead the design and implementation of next-generation data platforms. Responsible for architecting scalable, cloud-native, and real-time data ecosystems to support advanced analytics, AI/ML, and digital transformation initiatives. This role requires strong leadership, deep technical expertise, and experience across modern data architectures and platforms. Key Responsibilities: • Lead end-to-end data architecture for large-scale modernization and digital transformation programs. • Design and implement modern data platforms using cloud-native services (Azure, AWS, GCP) including lakehouses, streaming pipelines, and orchestration frameworks. • Define and drive adoption of modern architecture patterns such as data mesh, data fabric, data virtualization, and knowledge graphs. • Drive the adoption of modern data stacks: Spark, Databricks, Snowflake, Synapse, BigQuery, Kafka, dbt, Delta Lake, etc. • Develop logical and physical data models to support structured, semi-structured, and unstructured data. • Collaborate with data engineers, product owners, and business stakeholders to translate requirements into scalable data solutions. • Define and implement strategies for data governance, metadata management, master data management, Data security and access controls. • Evaluate and recommend tools, platforms, and frameworks aligned with enterprise data strategy. Required Skills and Experience: • Hands-on experience with at least one major cloud platform (Azure, AWS, GCP) or emerging platforms like IOMETE. • Expertise in modern data warehousing and lakehouse platforms: Snowflake, Databricks, Synapse, Redshift, BigQuery. • Deep knowledge of data integration, ETL/ELT tools (e.g., dbt, Azure Data Factory, Glue, Airflow). • Experience with streaming and real-time data platforms like Kafka, Spark Structured Streaming • Strong understanding of data modeling, data quality frameworks, and metadata management. • Familiarity with data mesh, data fabric, and data product concepts. • Experience with data cataloging and discovery tools: Collibra, Alation, Purview, etc. • Knowledge of DevOps, DataOps, and CI/CD for data pipelines. • Experience with AI/ML platform integration. • Excellent communication skills and ability to engage with technical and business stakeholders effectively. Additional Information: - The candidate should have minimum 12 years of experience in Data Architecture Principles. - This position is based at our Bengaluru office. - A 15 years full time education is required.
Project Role Description : Design, build and configure applications to meet business process and application requirements.
Must have skills : Data Architecture Principles
Good to have skills : NA
Minimum 12 year(s) of experience is required
Educational Qualification : 15 years full time education
Job Summary: Modern Data Architect to lead the design and implementation of next-generation data platforms. Responsible for architecting scalable, cloud-native, and real-time data ecosystems to support advanced analytics, AI/ML, and digital transformation initiatives. This role requires strong leadership, deep technical expertise, and experience across modern data architectures and platforms. Key Responsibilities: • Lead end-to-end data architecture for large-scale modernization and digital transformation programs. • Design and implement modern data platforms using cloud-native services (Azure, AWS, GCP) including lakehouses, streaming pipelines, and orchestration frameworks. • Define and drive adoption of modern architecture patterns such as data mesh, data fabric, data virtualization, and knowledge graphs. • Drive the adoption of modern data stacks: Spark, Databricks, Snowflake, Synapse, BigQuery, Kafka, dbt, Delta Lake, etc. • Develop logical and physical data models to support structured, semi-structured, and unstructured data. • Collaborate with data engineers, product owners, and business stakeholders to translate requirements into scalable data solutions. • Define and implement strategies for data governance, metadata management, master data management, Data security and access controls. • Evaluate and recommend tools, platforms, and frameworks aligned with enterprise data strategy. Required Skills and Experience: • Hands-on experience with at least one major cloud platform (Azure, AWS, GCP) or emerging platforms like IOMETE. • Expertise in modern data warehousing and lakehouse platforms: Snowflake, Databricks, Synapse, Redshift, BigQuery. • Deep knowledge of data integration, ETL/ELT tools (e.g., dbt, Azure Data Factory, Glue, Airflow). • Experience with streaming and real-time data platforms like Kafka, Spark Structured Streaming • Strong understanding of data modeling, data quality frameworks, and metadata management. • Familiarity with data mesh, data fabric, and data product concepts. • Experience with data cataloging and discovery tools: Collibra, Alation, Purview, etc. • Knowledge of DevOps, DataOps, and CI/CD for data pipelines. • Experience with AI/ML platform integration. • Excellent communication skills and ability to engage with technical and business stakeholders effectively. Additional Information: - The candidate should have minimum 12 years of experience in Data Architecture Principles. - This position is based at our Bengaluru office. - A 15 years full time education is required.
Requisitos
15 years full time education