Loading…
In-person
11-12 December
Learn More and Register to Attend

The Sched app allows you to build your schedule but is not a substitute for your event registration. You must be registered for KubeCon + CloudNativeCon India 2024 to participate in the sessions. If you have not registered but would like to join us, please go to the event registration page to purchase a registration.

Please note: This schedule is automatically displayed in India Standard Time (UTC+5:30)To see the schedule in your preferred timezone, please select from the drop-down menu to the right, above "Filter by Date." The schedule is subject to change and session seating is available on a first-come, first-served basis. 
strong>AI_dev Sessions [clear filter]
Wednesday, December 11
 

3:45pm IST

From CPU to GPU: Progressive Delivery for Complex ML Deployments - Sumit Jain & Divyansh Saxena, Adobe
Wednesday December 11, 2024 3:45pm - 4:20pm IST
Applying modern deployment strategies, such as incremental releases, to complex ML architectures presents unique challenges, including managing resource constraints and ensuring cost efficiency. In the GPU environment, new replicas can take up to an hour to become operational due to extended image pull times. Despite these challenges, implementing advanced deployment strategies is crucial to minimise the impact of faulty versions, much like in traditional CPU workloads. In this session we will present on how to overcome these hurdles in GPU-based workloads. We will: - Start with traditional deployment approaches. - Progressively address advanced scenarios, such as implementing cost-efficient progressive delivery for high CPU-consuming workloads on GPU nodes and orchestrating sequential multi-resource deployments. - Reference our recent experience designing CI/CD for the GenAI applications. - Conclude by sharing cost comparison and best practices for GenAI model deployments.
Speakers
avatar for Divyansh Saxena

Divyansh Saxena

Software Engineer, Adobe
Divyansh is a CKAD ,Oracle and AWS certified Software Engineer currently working at Adobe with extensive experience in designing CI/CD pipelines, microservice architecture and advanced deployment strategies for smooth version releases. He previously worked in OLX and has a total of... Read More →
avatar for Sumit Jain

Sumit Jain

CI/CD Architect, Adobe
Sumit is a seasoned software engineer with 12 years of experience in the industry, currently serving as an architect at Adobe, where he focuses on CI/CD solutions, Developer productivity, Container orchestration, Observability & Diagnostics primarily using Argo tooling and GitOps... Read More →
Wednesday December 11, 2024 3:45pm - 4:20pm IST
Room 2
  AI_dev Sessions, MLOps + GenOps + DataOps
  • Content Experience Level Any

4:50pm IST

Scalable ML Inferencing Pipeline Using K8s - Smitha Jayaram & Vinod Eswaraprasad, NVIDIA
Wednesday December 11, 2024 4:50pm - 5:25pm IST
Inference engines are used to generate prediction or deduce new information based on certain rules and data. With rise in the number of applications that benefit from inferencing, there is a definitive need to build Inferencing Services that are robust, performant (latency optimized) and can seamlessly scale on demand. In this session, we will discuss a proven set of procedures and guidelines for building and managing inference pipelines on Kubernetes. We will discuss details of the underlying hardware (GPU/CPU/memory) and K8s configuration requirements of some of the well-known Inference engines. We will demonstrate how robust and fault tolerant pipelines for LLM and RAG can be built using basic K8s constructs like operators, statefulsets and persistent volumes; along with enabling automated monitoring, triaging and remediation of the failed hardware and software components.
Speakers
avatar for Smitha Jayaram

Smitha Jayaram

Principal Software Engineer, Nvidia
Smitha is principal software engineer at NVIDIA focusing on software solutions for building scalable, cloud native GPU computing infrastructure. She has been working in the area of building scalable storage solutions, and cloud native platform design for the last 24 years. In the... Read More →
avatar for Vinod Eswaraprasad

Vinod Eswaraprasad

Solution Engineering, NVIDIA
Vinod is principal software engineer at NVIDIA focusing on software solutions for building scalable GPU computing infrastructure. He has been working in the area of building fault-tolerant, scalable, and distributed platform architecture and design for the last 26 years. In the current... Read More →
Wednesday December 11, 2024 4:50pm - 5:25pm IST
Room 5

5:40pm IST

Search/Vector DBs as Microservices: Pros and Cons - Radu Gheorghe, Vespa.ai
Wednesday December 11, 2024 5:40pm - 6:15pm IST
Traditionally, search engines and vector DBs (the boundaries got blurry lately) are deployed as monoliths in a cluster, managed by a StatefulSet and a lot of technology-specific logic. Newer use-cases (e.g. RAG, semantic search in general) implicitly push for more stateless work: e.g. generate embeddings on the fly, re-rank, local LLMs. Decoupling this work and managing it separately makes more sense now, but is it worth the complexity? This session takes a deep dive into the architectures of search engines (mainly Vespa, but Elasticsearch, Solr and others will be in the picture, too) to highlight both pros and cons of separating concerns in information retrieval. And how much separation is good for which scenario. In the end, you should have a better idea about both the technology and the deployment type that fits your use-case best.
Speakers
avatar for Radu Gheorghe

Radu Gheorghe

Software Engineer, Vespa.ai
Radu is in the search space for 13 years, mainly on Elasticsearch, Solr, OpenSearch and more recently Vespa.ai. Helps users with both the relevance and the maintenance side of retrieval, and it's the latter that naturally leads to Kubernetes. Enjoys education in all its forms (training... Read More →
Wednesday December 11, 2024 5:40pm - 6:15pm IST
Room 4
 
Thursday, December 12
 

2:55pm IST

Bias-Busting AI with a Cloud Native Approach - Rajas Kakodkar, Broadcom
Thursday December 12, 2024 2:55pm - 3:30pm IST
As Large Language Models gain prominence, the discussion around AI fairness and ethics has never been more pressing. Cloud Native is emerging as the standard infrastructure for training, deploying, and serving AI workloads, but its role doesn't stop there—it can also play a crucial part in minimizing bias in AI models. This session will delve into: - Mapping Cloud Native principles like attestation and supply chain fundamentals to AI workloads, enhancing explainability and providing actionable recourse. - Leveraging Kubernetes' reconciliation loop to offer feedback that corrects false positives in AI output. - Utilizing the distributed nature of cloud computing to decentralize AI power, promoting more responsible automation. This session is designed for AI practitioners seeking insights from the Cloud Native tech stack, infrastructure beginners looking to understand its extensibility, and experts and maintainers aiming to extend the boundaries of current Cloud Native infrastructure.
Speakers
avatar for Rajas Kakodkar

Rajas Kakodkar

Staff Engineer at Broadcom | Tech Lead at CNCF TAG Runtime, Broadcom
Rajas is a staff engineer at Broadcom and a tech lead of the CNCF Technical Advisory Group, Runtime. He is actively involved in the AI working group in the CNCF. He is a Kubernetes contributor and has been a maintainer of the Kube Proxy Next Gen Project. He has also served as the... Read More →
Thursday December 12, 2024 2:55pm - 3:30pm IST
Room 4

4:50pm IST

Streamlining Machine Learning Operations with GitOps - Kunal Kushwaha, Civo
Thursday December 12, 2024 4:50pm - 5:25pm IST
As AI becomes more popular, it's bringing some new headaches when it comes to managing models, from building them to getting them into production. The usual DevOps tools don't always cut it, especially when you need everything to be consistent and handle large datasets without a hitch. As ML projects grow, trying to manage everything manually just gets messy and prone to mistakes. In this talk, I'll break down how GitOps can make life easier by automating and simplifying ML operations. With Git as the go-to source for everything, GitOps keeps your ML models consistent and reproducible across different environments. I'll share some real-world examples, talk about the challenges, and give you practical tips on how to bring GitOps into your ML workflow. The goal is to help you deploy models faster and with fewer headaches.
Speakers
avatar for Kunal Kushwaha

Kunal Kushwaha

DevRel manager, Civo
Kunal Kushwaha is a Developer Relations Manager at Civo, where he fosters a deep connection between developers and innovative cloud-native technologies. With a passion for empowering others, Kunal has taught thousands online and consulted numerous companies on their DevOps journey... Read More →
Thursday December 12, 2024 4:50pm - 5:25pm IST
Room 5

5:40pm IST

Unleashing Generative AI with Open Data Hub - Ritesh Shah & Prakhar Srivastava, Red Hat
Thursday December 12, 2024 5:40pm - 6:15pm IST
Explore AI innovation with the "Generative AI Demo using Open Data Hub," an open-source AI/ML platform tailored for Kubernetes. This demo features two Mistral-based LLMs, one utilizing Retrieval-Augmented Generation (RAG) with PostgreSQL vector extensions, and another without. Users interact through a Gradio chatbot, demonstrating how RAG improves accuracy. Two persona-based pipelines (Admin & Data Scientist) highlight Open Data Hub’s seamless integration with Tekton-powered OpenShift Pipelines, showcasing the Hybrid Cloud Strategy. Key components include the Kubernetes Platform, Node Feature Discovery, NVIDIA GPU Operators, Software Defined Storage, Tekton, Argo CD (for GitOps), Gitea Operator, Knative and Istio. Join us to see how GPU-accelerated AI and scalable, secure infrastructure power next-gen applications.
Speakers
avatar for Ritesh Shah

Ritesh Shah

Senior Principal Architect, AI Lead, Red Hat
Ritesh Shah is a Senior Principal Architect with Red Hat and focuses on creating and using next-generation platforms, including AI/ML workloads as well as application modernisation and deployment. Ritesh is an advocate for open source technologies and products, focusing on modern... Read More →
avatar for Prakhar Srivastava

Prakhar Srivastava

Team Lead, Global Content Development Portfolio & Technology, Red Hat
As a seasoned technology professional and Team Lead for the Global Content Development team at Red Hat, my role is integral to driving the optimization of our sales and marketing strategies, ensuring that our products stand out in a highly competitive market. My primary focus is on... Read More →
Thursday December 12, 2024 5:40pm - 6:15pm IST
Room 2
 
  • Filter By Date
  • Filter By Venue
  • Filter By Type
  • Content Experience Level
  • Timezone


Share Modal

Share this link via

Or copy link

Filter sessions
Apply filters to sessions.