Discover customer problems and frame AI-driven solutions.
Design and implement end-to-end cloud architectures and production apps on SAP BTP.
Troubleshoot, tune, and optimize jobs using Photon engine and serverless compute, ensuring cost efficiency and SLA reliability.
Package and deploy pipelines using Databricks Asset Bundles through CI/CD pipelines in GitHub or GitLab.
Ensure quality, security, observability, and CI/CD.
Collaborate closely with customer teams, product, UX, data, and architecture.
Prepare Gen AI-compatible datasets: manage vector embeddings, index with Databricks Vector Search, and use Feature Store with ML flow.
Build AI/LLM/Agentic features (RAG, hybrid retrieval) and integrate ML services.
Requirements
llms
data warehousing
sap btp
databricks
airflow
python
Hands on Experience in applying Performance optimization techniques
Hands-on with LLMs, RAG, agentic frameworks (e.g., LangChain, LangGraph, MCP).
Understanding data modeling and data warehousing principles is essential.
5–10 yrs software engineering for cloud/enterprise/SaaS.
SAP BTP & HANA experience; Databricks or similar data pipelines. Nice to Have: 1. Certifications: Data Engineering Certified Professional or similar certifications. 2. Machine Learning: Knowledge of machine learning concepts and experience with popular ML libraries. 3. Knowledge of big data processing (e.g., Spark, Hadoop, Hive, Kafka) 4. Data Orchestration: Apache Airflow. 5. Knowledge of CI/CD pipelines and DevOps practices in a cloud environment. 6. Experience with ETL tools like Informatica, Talend, Mati Llion, or Five Tran. 7. Familiarity with DBT (Data Build Tool) Additional Information: - The candidate should have minimum 7.5 years of experience in Data Engineering/Machine learning - This position is based at our Bengaluru/Pune office. Educational Qualification: - 15 years full time education is required.
Proven delivery of multi-tenant, cloud-native apps.
Experience with cloud-based services relevant to data engineering, data storage, data processing, data warehousing, real-time streaming, and serverless computing.