Salesforce Inc.

09/22/2025 | Press release | Distributed by Public on 09/22/2025 07:12

5 AI Agent Pitfalls That Will Sink Your Agentic Enterprise Before it Launches

5 AI Agent Pitfalls That Will Sink Your Agentic Enterprise Before it Launches

An agentic enterprise means more than just deploying agents - it's a complete system.

Chasing quick wins? Don't miss the bigger picture.

Ajay Kumar

Vivek Mahapatra

September 22, 2025 9 min read

Share article

If you've ever deployed an AI agent at work, there's a good chance you followed this pattern: identify a problem, build an agent to solve it, and ship it fast. While this quick-win approach can be a valuable starting point, it can also turn into an unintended trap for organisations. Treating AI as a series of point solutions misses the bigger picture. The point isn't to build a few agents, but to become an agentic enterprise, an organization that uses AI to deepen customer satisfaction, increase retention, and create new revenue streams.

An agentic enterprise means more than just deploying agents - it's a complete system. It starts with trusted governance as a foundational layer, not an afterthought. It's powered by a scalable platform with workflow automations and integrations that allow agents to act reliably on data. And it treats data as a strategic asset, carefully curated and managed across internal and external sources to fuel AI applications and deliver enterprise-grade outcomes. Enterprise here doesn't mean "big company," but rather any organization that's ready to scale with trust, efficiency, and impact.

The vision may seem clear, but the devil is in the details. Getting from a great idea to a successful, scalable AI solution requires a disciplined process. Based on more than 80 live Agentforce deployments, below are the five most common pitfalls organizations should avoid.

1. The rush to build

It's understandable to want to hit the ground running with AI agents, but rushing to build without a clear purpose and business value is a trap. Teams are often eager to start configuring actions and prompts before conducting a formal strategic planning phase. Without a defined purpose, it's impossible to measure an agent's return on investment (ROI), justify future investments, or ensure a positive user experience. This leads to agents that try to do everything but excel at nothing, as their scope is constantly shifting.

Common gaps include vague success metrics like "improve CSAT" without a baseline and an undefined agent persona, leaving key questions about its role, data sources, and guardrails unanswered.

The solution: A strategic ideation phase should be considered essential and non-negotiable. This anchors the agent's development to a measurable business outcome. The first step is to conduct a discovery workshop to validate a specific business problem as a viable use case and define the agent's core "jobs to be done" (JTBD). Before you start building, establish specific, quantifiable success metrics. Examples include reducing average handle time (AHT) for warranty claims by 90 seconds, or achieving a 35% self-service resolution rate for return merchandise authorizations (RMAs). Finally, a formal governance structure involving business, technical, and operational stakeholders must be established to ensure alignment and make key decisions throughout the agent's lifecycle.

2. The over-privileged agent

While it might be convenient to simply grant your agent access to everything to avoid permission issues later, this can rapidly turn into a security and governance nightmare. In fact, the most significant security risk we've observed in agent development is granting permissions that exceed the agent's needs, creating a large and unnecessary attack surface. This is a critical implementation gap where uncontrolled record-level access allows the agent to view sensitive data it shouldn't be able to see. We've found that the agent user is often assigned a cloned System Administrator profile or an existing integration user profile with "Modify All Data" permissions. Furthermore, placing the agent user within the main role hierarchy grants it broad data access through inherited sharing rules.

The solution: Organizations should follow the principle of least privilege (PoLP) with surgical precision during the setup phase. Instead of cloning an admin profile, create a new user and a new profile from scratch for the agent. Permissions should be granted incrementally using permission sets, starting with zero access and adding only the specific object and field-level security (FLS) required for the agent's actions. For record-level access, keep the agent user out of the role hierarchy. Its data access should be managed primarily through restrictive Organization-Wide Defaults (OWD). If the agent needs access to specific records, use targeted sharing rules or, for more complex scenarios, programmatic Apex Sharing. This approach ensures the agent has access only to the data absolutely necessary for its function, significantly reducing security vulnerabilities.

3. The uncurated knowledge dump

More data automatically means smarter agents, right? Not quite. When you consider that the primary cause of hallucinations and irrelevant responses is poor data management, a "connect the firehose" approach probably isn't the best strategy.

Unstructured, unvalidated, or outdated information poisons the retrieval-augmented generation (RAG) process, making it impossible for agents to distinguish between current and irrelevant data. This leads to a useless agent that can't provide accurate information. We often see this when entire Salesforce Knowledge Bases are connected without curation, leading to an overwhelming amount of data for the agent to process. Additionally, teams frequently fail to validate the data ingestion process. They don't confirm that the necessary Data Cloud objects were created or that the search index is ready, leading to undiagnosed ingestion failures and a partially built or empty knowledge base.

The solution: The Agentforce Data Library (ADL) must be treated as a meticulously curated database, not a chaotic filing cabinet. Start by intentionally structuring your ADL architecture. For Salesforce Knowledge, strategically designate some fields as "identifying fields" to help the retriever find the correct article, while others are designated as "content fields" to provide the substance for the LLM's answer. After data ingestion, use the Data Cloud Query Editor to validate that the Unstructured Data Model Object (UDMO) and its associated Search Index Chunk/Vector Tables have been successfully created. Run a query to confirm that the search index status is "Ready." Finally, test the end-to-end RAG process to ensure that the correct retriever is being invoked by the "Answer Question with Knowledge" action and that it is fetching the most relevant chunks for the LLM to synthesize.

4. The ambiguity problem

A significant implementation gap we've observed is the use of vague and overlapping descriptions, which cripples an agent's ability to classify user intent and select the correct tool. Topic and action configurations aren't merely simple labeling exercises - these descriptions are direct instructions for the agent's reasoning engine.

In the absence of clear instructiolns, the agent is forced to guess, leading to misrouted conversations or incorrect action execution. We frequently see topic classifications that are not semantically distinct - for example, "Account Inquiry" versus "My Account" - which causes confusion. A major issue also arises in Flow-based actions where input and output parameters are given cryptic API names like "input1" or "caseId" without any descriptive text. The agent's reasoning engine relies heavily on these descriptions to understand how to use the Flow correctly.

The solution: To ensure your agent understands your intent, it's essential to write all descriptions and instructions with machine-level precision during the configure phase. Begin with minimal instructions and iterate, using advanced prompt engineering techniques like chain-of-thought processing for complex, multi-step tasks to guide the agent's logic. When configuring actions, choose the right tool for the job using a clear framework: use Apex for complex logic, Flow for declarative processes, and Prompt Templates for LLM-native tasks. Critically, provide clear and descriptive text for all Flow parameters. For an input variable named contactEmail, write a description like: "The primary email address of the customer to search for. Must be a valid email format." This provides the essential context the reasoning engine needs to function correctly. Finally, build input validation and error handling directly into your Apex or Flow to ensure the action is robust and secure.

5. The "launch and leave" mentality

The biggest mistake organizations tend to make is treating new agents like a one-and-done project, ignoring the need for rigorous testing and post-launch management. Without a continuous feedback loop, performance degrades as agents remain static while business workflows and processes continue to evolve. Even minor updates can introduce major regressions. Many companies limit testing to manual, one-off conversations in Agent Builder, completely ignoring the need for large-scale batch testing through Testing Center. Additionally, debugging is inefficient because teams aren't enabling features like "enrich event logs with conversation data," making it nearly impossible to perform proper root cause analysis. After deployment, there's often no formal process for monitoring agent performance, leaving a trove of valuable user interaction data from Utterance Analysis untouched.

The solution: The key to managing agents effectively is embracing an iterative and holistic lifecycle that spans the test, deploy, and monitor phases. Your testing strategy should combine speed and safety by using Agent Builder for quick development and Testing Center for automated regression testing. When debugging, always use the "Let Admin Debug Flow as other users" feature. This allows you to replicate the agent's exact permissions and data visibility, making troubleshooting much more efficient. For release management, never develop in your active production version. Instead, manage different versions of the agent for development, testing, and production, and use phased rollouts for major changes. Finally, remember that launch is just day one. Establish a formal operational process to regularly review metrics from Utterance Analysis. Use these insights to identify knowledge gaps, refine topic instructions, and prioritize the development of new agent capabilities, ensuring the agent continuously improves.

A virtuous cycle

Each of the obstacles we've discussed - from flawed strategy to weak security and poor data governance - are symptoms of a larger issue: a failure to think about the entire business ecosystem. The most successful AI implementations aren't isolated technical projects. They are deeply integrated with the organization's existing workflows, data infrastructure and, most importantly, its people and customers.

So, where should an agentic enterprise start?

Building an agentic enterprise requires a new model for AI product growth. The process isn't a linear sprint to launch, but a deliberate, cyclical journey:

AI product growth is a cyclical journey beginning with experimentation, then validation and scale.

At the top is a massive circle of experimentation, where you can explore a wide range of possibilities. This is where you work with teams across the business - marketing, sales, support, and beyond - to understand their challenges and identify potential AI use cases. The key is to run many small experiments, fail fast, and learn quickly.

Next, you'll narrow down to a few promising ideas for validation. This stage is about rigorously testing your hypotheses. Does this agent solve a real business problem? Does it drive a measurable outcome? This is where you move from a proof-of-concept to a working prototype with clear success metrics.

Finally, you select an even smaller subset of validated projects to scale. This is where you bridge the gap between a successful prototype and a production-ready application. This stage requires deep collaboration with the teams who will own and maintain the solution from this stage forward - engineering, operations, and IT. The goal is to build an experience that is robust, secure, and ready for prime time.

The five obstacles we explored offer a roadmap to navigate this journey, steering you away from common pitfalls and toward a continuous cycle of improvement. Ultimately, these applications must do more than just improve internal efficiency. They have to drive measurable customer value and lead to tangible outcomes like increased satisfaction, retention, and revenue. Building an agentic enterprise isn't merely about adding new features - it's about building an entirely new business model.

Share article

Ajay Kumar Senior Director, AI Product Management, Agentforce

Ajay Kumar is the Senior Director of AI Product Management for Agentforce, leveraging a background in scaling products and engineering solutions across multi-public clouds for both startups and enterprises. His expertise spans multiple Salesforce portfolios, ranging from MuleSoft, SF Platform,...Read More Marketing Cloud, and Commerce Clouds.

More by Ajay
Vivek Mahapatra Vice President, AI Product

Vivek Mahapatra is Vice President of AI Product at Salesforce and was part of the founding team of Agentforce. With more than 20 years of experience across sales, solution engineering, competitive strategy, and product management, he has built and scaled multiple businesses from 0→1 in both...Read More startups and global enterprises.

More by Vivek
Salesforce Inc. published this content on September 22, 2025, and is solely responsible for the information contained herein. Distributed via Public Technologies (PUBT), unedited and unaltered, on September 22, 2025 at 13:13 UTC. If you believe the information included in the content is inaccurate or outdated and requires editing or removal, please contact us at [email protected]