DES is a property of AIM Media House.

SCHEDULE

MAY. 15-16 2025
Hotel Radisson Blu, Bengaluru

MAY. 15-16 2025
Bengaluru

SCHEDULE

We are in the process of finalizing the schedule for 2025. Please check back this space again.

Expect more than 60 speakers to speak at DES 2025. To explore speaking opportunities with DES, write to info@aimmediahouse.com

Expand All +
  • Day 1 | HALL 2 - Practical Insights and Best Practices


  • Data engineers play a pivotal role in driving contextualized intelligence. They must ensure that data is enriched with the proper context to deliver meaningful insights. This tech talk will explore the practical steps and tools data engineers can use to build a contextual layer, ensuring AI models can deliver strategic insights based on relevant business contexts.

  • In this talk, I will dive into a real-time data pipeline solution I worked on for the Travel & Hospitality domain, powered by Apache Flink Stateful Functions. This solution efficiently handled over 800,000 streaming events per day, addressing critical use cases. I will share insights into the challenges faced, the strategies used to overcome them, and how we successfully scaled the pipeline to meet demanding business requirements. For the second part of the session, I will explore the hot topic: Will Low-Code ETL Replace Data Engineers? I firmly believe that Low-Code ETL is not here to replace data engineers but to accelerate the development process. However, for this to succeed, ETL platforms must empower data engineers with the freedom and control to customize and adapt the platform to their specific needs. I’ll discuss examples like Prophecy, a Low-Code ETL platform that strikes a balance by acting as an IDE, generating open-source code, and keeping implementation transparent. This approach ensures that while development is expedited, the power remains firmly in the hands of data engineers.

  • AI is reshaping the landscape of data engineering in transformative ways. One key trend is the rise of AI-powered data pipelines, where machine learning automates tasks like data cleaning, transformation, and feature engineering, significantly accelerating development cycles. Alongside this, AI-driven data observability tools are improving data reliability by detecting anomalies and quality issues in real time. Intelligent data catalogs are also evolving, using AI to automatically generate metadata, track lineage, and simplify data discovery. Natural language interfaces, such as Text-to-SQL systems powered by large language models, are making it easier for non-technical users to query databases using plain English. ETL and ELT processes are becoming smarter too, with AI automating schema mapping and transformation logic, reducing the need for manual coding. In parallel, synthetic data generation is gaining traction, enabling the creation of realistic datasets for testing, training, and privacy-sensitive use cases. Real-time data processing is becoming more intelligent as AI models integrate directly into streaming frameworks like Kafka and Spark, supporting use cases like fraud detection and personalization. Governance is also being augmented by AI, with tools now capable of automatically detecting sensitive data like PII and enforcing compliance with regulations.

  • The Unified Data platform is designed to address key limitations in data management for machine learning (ML) model development, training, and testing. Its primary focus is to improve turnaround time (TAT), enhance experimentation velocity, and provide a centralized solution for data needs.

  • Discover the engineering innovations behind KnowBe4's agentic platform, which utilizes Data and Generative AI to empower cybersecurity defense. Learn how KnowBe4 developed AIDA (Artificial Intelligence Defense Agents) to significantly reduce administrative overhead for customers by automating the configuration of KnowBe4 products. AIDA intelligently understands specific customer requirements based on user behavior, leading to automated risk reduction. This session will delve into the cutting-edge technologies that power this feature, including the latest advancements in Artificial Intelligence, Datamesh architecture, and big data engineering tools. We will explore how these elements are integrated to deliver current GenAI capabilities and lay the foundation for future intelligent features aimed at proactively mitigating human-related cyber risks.

  • Building a future-forward health insurance data platform focuses on leveraging emerging technologies like AI, machine learning, and blockchain to revolutionize data management in the insurance sector. By integrating diverse data sources such as medical records, wearable devices, and claims data, the platform enables personalized policies, faster claims processing, and enhanced risk management. This approach not only increases operational efficiency but also improves customer satisfaction by providing tailored coverage and real-time insights

  • This session explores how AI agents are transforming data engineering by automating complex workflows such as data ingestion, transformation, and pipeline orchestration. With real-time analytics and intelligent decision-making becoming critical, AI-driven automation is enabling greater efficiency, scalability, and accuracy in data processes. From automated anomaly detection to schema evolution and performance optimization, discover practical use cases that showcase the power of AI in simplifying and future-proofing data engineering strategies. Ideal for data engineers, architects, and AI enthusiasts, this talk offers insights into leveraging AI agents to reduce operational overhead and stay ahead in the era of intelligent automation.

  • Day 1 | Main Hall - Thought Leadership and Strategic Insights


  • In today's data-driven world, context is everything. Without context, raw data can often lead to inaccurate insights, misinterpretations, and missed opportunities. For data engineers, this means building platforms that enrich data with relevant business context and provide strategic insights. Join this session to explore how modern, context-centric platforms help businesses stay ahead in the data-driven landscape.

  • AI-driven automation is rapidly transforming data engineering workflows, from automated pipeline generation to self-healing data architectures. But does this mean traditional data engineering roles will become obsolete? This panel will debate whether AI is an enabler or a disruptor, discussing where human expertise remains irreplaceable and how engineers can adapt to stay relevant in the AI-powered future.

  • As organizations race to adopt Generative AI, a strong and scalable data foundation is essential. This session explores how building a FAIR-aligned Data Marketplace has transformed data access, quality, and governance- paving the way for AI-driven innovation. Learn how a persona-driven, self-service platform powered by Snowflake, Power BI, and advanced metadata cataloging accelerated decision-making, eliminated redundancy, and boosted adoption by 25%. Key innovations include a real-time Data Quality Index, seamless integration for both technical and non-technical users, and smart search capabilities using LLMs. Discover how strategic leadership, automation, and a culture of collaboration are driving enterprise-wide data democratization- preparing the organization to fully harness the power of Generative AI.

  • In modern systems, real-time insights aren't a luxury—they're a requirement. Whether you're debugging distributed systems, tracking financial transactions, or analyzing user behavior, sub-second query latency can be the difference between reacting and proactively optimizing. This talk dives into the technical foundations that make real-time analytics possible at scale, using ClickHouse as the case study. We'll explore the architectural underpinnings of its high-performance columnar engine, including vectorized execution, late materialization, and how it handles time series and semi-structured data like JSON with minimal overhead. Through real-world use cases—from high-throughput log ingestion in observability stacks, to petabyte-scale analytics for adtech, fintech, and user personalization—you’ll see how engineering teams are designing for low latency without sacrificing flexibility or scale. If you're a data engineer, backend developer, or platform architect working with high-velocity data, this session will give you a deeper understanding of how to build infrastructure that can keep up.

  • Day 2 | HALL 2 - Practical Insights and Best Practices


  • In today’s data-driven financial ecosystem, speed, scale, and accuracy are non-negotiable. This session explores the key principles and real-world practices behind building scalable data pipelines that power meaningful financial insights. From ingesting high-volume transactional data to ensuring real-time processing and analytics, we'll dive into architectural patterns, technology choices, and performance considerations that enable organizations to unlock value from their data. Whether you're modernizing legacy systems or designing from the ground up, this talk will equip you with practical takeaways to build resilient, future-ready data pipelines tailored for the dynamic demands of the financial domain.

  • This session is to outline strategic data engineering practices that drive innovation and business growth. Emphasizing scalability, data quality, and operational efficiency. The session highlights how modern data platforms, automation, and governance empower data-driven decision-making across the organization.

  • In today’s data-driven world, ensuring resilience in data architectures is crucial for maintaining seamless operations, minimizing downtime, and enabling real-time decision-making. This session explores the key principles of designing fault-tolerant, scalable, and adaptive data ecosystems that can withstand failures and disruptions. Key focus areas include strategies for high availability, redundancy, disaster recovery planning, and leveraging distributed computing and cloud-native technologies to enhance resilience. Real-world insights will showcase how resilient data architectures can optimize performance, improve reliability, and support business continuity in an evolving technological landscape

  • Real-time data pipelines enable organizations to ingest, process, and act on streaming data with minimal latency, ensuring they can make informed decisions instantly. This session will explore the core components of real-time data pipelines, including event-driven architectures, stream processing frameworks, and scalable data ingestion technologies. We’ll discuss real-world applications such as fraud detection, predictive analytics, IoT monitoring, and personalized customer experiences. By the end of the session, attendees will gain a deeper understanding of how real-time data pipelines drive agility, enhance operational efficiency, and unlock new business opportunities. Whether you're a data engineer, architect, or business leader, this talk will equip you with the knowledge to harness the power of real-time data for smarter decision-making.

  • As organizations increasingly rely on complex data ecosystems, maintaining data quality and trust has become paramount to business success. This discussion explores the critical components and best practices for implementing robust data quality frameworks within modern data architectures. We examine how organizations can establish automated testing, monitoring, and governance processes to ensure data reliability across the entire pipeline - from ingestion to consumption. Special attention is given to emerging tools and methodologies that enable continuous data quality assessment while scaling with growing data volumes and complexity.

  • In the competitive OTT landscape, delivering an exceptional video Quality of Experience (QoE) is crucial for user satisfaction and retention. This session will explore how data-driven approaches can transform the way QoE is monitored, analyzed, and optimized. The talk will delve into the key metrics that define QoE, including startup time, buffering frequency, bitrate quality, and playback interruptions. By leveraging real-time analytics, machine learning, and predictive modeling, OTT platforms can proactively identify and resolve issues before they impact the user experience. Attendees will gain actionable insights into: Designing scalable data pipelines for QoE monitoring. Using predictive analytics to anticipate and mitigate streaming disruptions. Optimizing encoding and delivery processes to reduce latency and improve video quality. This session is ideal for data engineers, OTT platform architects, and video delivery professionals looking to elevate their platform’s user experience through the power of data.

  • Day 2 | Main Hall - Thought Leadership and Strategic Insights


  • Developing reliable AI models requires high-quality data. Due to the shortcomings of traditional methods, there is a need to adopt data reliability engineering. This approach involves applying principles from manufacturing and considering systems as data factories. It entails significant adjustments to people, processes, and tools. - Data Testing: Validates data through migration testing, pipeline certification, and big data reconciliation. - Data Monitoring: Tracks data pipeline health using business rules, and handles exceptions in real time. - Data Observability: Uses AI/ML for anomaly detection, compliance measurement, and defect rate prediction. The talk will highlight these pillars—Data Testing, Data Monitoring, and Data Observability—showcasing their importance in preventing errors, ensuring compliance, and optimizing AI performance. Case studies and industry practices will demonstrate how to engineer quality data effectively.

  • The AI Hackathon on the Government e-Marketplace (GeM) brought together innovators to solve real-world public sector challenges using cutting-edge technologies—Generative AI, Deep Learning, and Machine Learning. This session showcases standout use cases that emerged from the hackathon, demonstrating how AI can enhance transparency, efficiency, and user experience on GeM. At the heart of these innovations lies a solid Data Engineering foundation—powering everything from data ingestion and transformation to model training and deployment. Without it, these AI solutions wouldn’t scale or succeed. This talk will highlight how robust data engineering practices made the leap from idea to impact possible.

  • The concept of data as a product is gaining traction, promising improved accessibility, ownership, and value creation. But is this approach practical for all organizations, or is it just another theoretical framework? This discussion will explore the real-world adoption of data product thinking, its impact on governance and analytics, and whether businesses are truly structured to make the most of it.


Our Pricing will change soon!

  • Early Bird Passes

    Expired
  • All access, 2 day passes
  • Group Discount available
  • Late Pass

    Available from 3 May 2025 onwards
  • All access, 2 day passes
  • No Group Discount available
  • Limited passes available
  • 19999

1000+

Attendees

50+

Speakers

4th

Edition

explore the frontiers of Data engineering.

Focused on data engineering innovation, this 2-day conference will give attendees direct access to top leaders & innovators from leading tech companies who will talk about the software deployment architecture of AI systems, how to produce the latest data frameworks and solutions for business use cases.

The Finkelstein Awards for Data Engineering Excellence 2025

Secure Your Seat at the Frontier of Data + engineering.

Early Bird Passes on sale now.

MAY. 15-16 2025
Bengaluru

Regular Passes to Expire Next Week!

Days
Hours
Minutes
Seconds