Rohit Bhardwaj is a Director of Architecture working at Salesforce. Rohit has extensive experience architecting multi-tenant cloud-native solutions in Resilient Microservices Service-Oriented architectures using AWS Stack. In addition, Rohit has a proven ability in designing solutions and executing and delivering transformational programs that reduce costs and increase efficiencies.
As a trusted advisor, leader, and collaborator, Rohit applies problem resolution, analytical, and operational skills to all initiatives and develops strategic requirements and solution analysis through all stages of the project life cycle and product readiness to execution.
Rohit excels in designing scalable cloud microservice architectures using Spring Boot and Netflix OSS technologies using AWS and Google clouds. As a Security Ninja, Rohit looks for ways to resolve application security vulnerabilities using ethical hacking and threat modeling. Rohit is excited about architecting cloud technologies using Dockers, REDIS, NGINX, RightScale, RabbitMQ, Apigee, Azul Zing, Actuate BIRT reporting, Chef, Splunk, Rest-Assured, SoapUI, Dynatrace, and EnterpriseDB. In addition, Rohit has developed lambda architecture solutions using Apache Spark, Cassandra, and Camel for real-time analytics and integration projects.
Rohit has done MBA from Babson College in Corporate Entrepreneurship, Masters in Computer Science from Boston University and Harvard University. Rohit is a regular speaker at No Fluff Just Stuff, UberConf, RichWeb, GIDS, and other international conferences.
Rohit loves to connect on http://www.productivecloudinnovation.com.
http://linkedin.com/in/rohit-bhardwaj-cloud or using Twitter at rbhardwaj1.
AI agents don’t behave like humans. A single prompt can trigger thousands of parallel API calls, retries, and tool chains—creating bursty load, cache-miss storms, and runaway costs. This talk unpacks how to design and operate APIs that stay fast, reliable, and affordable under AI workloads. We’ll cover agent-aware rate limiting, backpressure & load shedding, deterministic-result caching, idempotency & deduplication, async/event-driven patterns, and autoscaling without bill shock. You’ll learn how to tag and trace agent traffic, set SLOs that survive tail latency, and build graceful-degradation playbooks that keep experiences usable when the graph goes wild.
Why scaling is different with AI
Failure modes to expect (and design for)
Traffic control & fairness
Resilience patterns
Caching that actually works for AI
Async & event-driven designs
Autoscaling without bill shock
Observability & cost governance
Testing & readiness
Runbooks & playbooks
Deliverables for attendees
Learning Objectives (Takeaways)
APIs designed for humans break when consumed by LLMs and autonomous agents. Documentation isn’t enough—endpoints must be machine-discoverable, deterministic, idempotent, and versioned with clear deprecation signals. This talk gives you a pragmatic lifecycle readiness framework: assess your current APIs, prioritize the ones that matter, and execute a phased roadmap (discovery → redesign → versioning → monitoring → deprecation). We’ll align with current best practices for function/tool calling, prompt-injection defenses, idempotency, and version sunset/deprecation headers, and show how to instrument agent traffic so you can govern cost and risk. You’ll leave with a scorecard, checklists, and KPIs to move from “works for humans” to agent-friendly, enterprise-grade APIs.
What “AI-Readiness” Means
Common Failure Modes Today
Assessment Framework (API Readiness Scorecard)
Prioritization Strategy
Roadmap Phases
Case Studies / Examples
Takeaways