AI prompts and rules library for Agentforce Vibes development, content creation, and workflow automation.
This repository curates Salesforce-focused prompts and system rules from the wider developer community to accelerate Agentforce Vibes agentic workflows. Collections are organized by development disciplineβApex, LWC, flows, deployments, testing, investigation, spec-driven delivery, and moreβso contributors can share reusable prompts, scaffolds, and guardrails that other teams can adapt and extend.
Coming soon!
Browse the repository and copy/paste any prompt or rule directly into Agenforce Vibes.
You can register additional repos with the extension as long as they mirror this structure:
- Root folders named
prompts/andrules/, each containing category subfolders (e.g.,prompts/apex-development/). - Each prompt or rule stored in its own Markdown file with YAML frontmatter (
name,description,tags, optional setup metadata). - Category folders may include a
README.mddescribing their focus; empty folders are allowed for future content.
When you add a new library:
- Ensure the folder layout matches the table in
## ποΈ Structure. - Follow the naming conventions and prompt format outlined below.
- Register the repository with
Agentforce Vibes: Add Libraryin VS Code. - Refresh the extension to surface the new content instantly.
Every prompt begins with YAML frontmatter that surfaces key metadata to contributors and tooling:
---
name: Concise Prompt Title
description: One-sentence summary of the outcome you want
tags: category, use-case, tooling
requires_deploy: true # optional β include when pre-work is required
setup_summary: Deploy baseline trigger before refactor # optional helper text
---name,description, andtagsare required.- Use lowercase, comma-separated tags drawn from the category and focus area (e.g.,
apex, refactor, testing). - Add
requires_deploy(and an optionalsetup_summary) when the prompt depends on seed metadata or data.
After the frontmatter, organize the body with clear sections. A common pattern is:
## Setupβ only when pre-work, sample metadata, or environment configuration is needed.## Contextβ summarize the scenario, constraints, personas, or assets involved.## Instructionsβ detail the tasks in numbered steps, calling out decision points or checkpoints.## Testing & Reportingβ define verification steps, coverage expectations, or deliverables.## Follow-upsβ optional space for stretch goals, review questions, or iteration loops.
File: prompts/apex-development/trigger-refactoring.md
---
name: Trigger Refactor Helper
description: Refactor the Opportunity trigger into a handler pattern with tests
tags: apex, refactor, testing
requires_deploy: true
setup_summary: Deploy baseline trigger to target org before running instructions
---
## Setup
1. Deploy the baseline trigger shown below to your default or scratch org.
2. Confirm the trigger compiles successfully before continuing.
```apex
// ... baseline trigger omitted for brevity ...Refactor OpportunityTrigger into a handler class (or classes) that handles the same behavior using bulk-safe patterns. Ensure the trigger itself delegates and remains behaviorally identical.
- Create unit tests covering positive and negative paths for each handler method.
- Include a bulk test that updates 50
Opportunityrecords where only half qualify for theafter updatelogic. - Deploy the refactored code and run the tests, then report coverage and key observations.
These starter categories reflect the current repository layout. Contributors are welcome to propose new ones or reorganize as long as the structure stays consistent for the VS Code extension.
| Category | Purpose | Example Topics |
|---|---|---|
| apex-development | Build and optimize Apex codebases | Trigger frameworks, async patterns, governor limit tuning |
| lwc-development | Craft Lightning Web Components | Component architecture, reactive data, UI patterns |
| metadata-deployments | Plan and execute releases | Packaging, Git branching, rollback prep |
| vibe-coding | Agentforce Vibes coding workflows | Apex/LWC scaffolds, prompt-to-code translation |
| testing-automation | Validate platform behavior | Apex tests, Flow scenarios, regression suites |
| investigation-triage | Diagnose and resolve issues | Incident response, log analysis, performance forensics |
| data-operations | Manage data pipelines | ETL prompts, bulk operations, platform events |
| spec-driven-dev | Generate and refine specification-first workflows | Requirement capture, traceability matrices, auto-generated tasks |
| security-compliance | Enforce standards and controls | Permission audits, secure coding, compliance narratives |
| integration-fabric | Coordinate external services | API design, middleware coordination, error recovery |
| enablement-docs | Share knowledge and runbooks | Onboarding guides, release notes, changelog automation |
| Category | Focus | Example Assets |
|---|---|---|
| apex-development | Standards for Apex architecture and quality | Trigger guardrails, async execution policies |
| lwc-development | Front-end guardrails for Lightning Web Components | Accessibility checklists, component review templates |
| metadata-deployments | Release management discipline | Branching policies, deployment readiness reviews |
| vibe-coding | Coding quality for Agentforce Vibes assets | Code review criteria, secure pattern guides |
| testing-automation | Verification and validation expectations | Test coverage thresholds, regression playbooks |
| investigation-triage | Incident and root-cause response | Escalation runbooks, logging requirements |
| data-operations | Data stewardship and job governance | Data quality SLAs, bulk job safeguards |
| spec-driven-dev | Specification-first delivery standards | Definition-of-done templates, traceability requirements |
| security-compliance | Platform security and regulatory posture | Access reviews, compliance attestation steps |
| integration-fabric | External connection reliability | Retry policies, credential rotation standards |
| enablement-docs | Knowledge management and enablement | Release note templates, onboarding workflows |
| org-governance | Enterprise policy alignment | Org strategy playbooks, architecture review guidelines |
| support-operations | Production support excellence | Incident response SLAs, shift handover procedures |
| ai-safety | Responsible agent behavior | Ethical guidelines, bias detection checklists |
- Choose the right category based on use case (if nothing fits, propose a new category)
- Create a descriptive filename (use kebab-case:
my-prompt.md) - Add frontmatter with name, description, and tags
- Write clear instructions with placeholders for user input
- Test before committing
- Commit with message:
Add [name] for [use case]
- Use lowercase with hyphens:
code-review-helper.md - Be descriptive:
salesforce-apex-debug.mdnotdebug.md - Include context:
blog-post-outline.mdnotoutline.md
- β Be specific - Clear instructions yield better results
- β Use structure - Numbered lists and sections help
- β Add context - Explain what you want and why
- β Include examples - Show expected output format
- β Test thoroughly - Verify prompts work as intended
- β Clarify the objective β Capture the outcome, stakeholders, and success metrics directly in the frontmatter
- β Share context β Provide links, metadata, or sample records so Agentforce can ground its reasoning
- β Set guardrails β Define tone, compliance boundaries, what to avoid, and when to ask for confirmation
- β Guide the workflow β Break the request into staged checkpoints (ideate β propose β confirm β deliver)
- β Capture feedback loops β Invite GPT-5 to flag assumptions, pose questions, and suggest validation steps
- β Encourage adaptability β Note how the prompt or rule can flex across org types, industries, and data volumes
- Prime with examples: Include concise samples that illustrate the desired format or code pattern
- Model the format: Provide headings and numbered steps so Agentforce mirrors the final artifact
- Address ambiguity: Explicitly call out unknowns and ask Agentforce to gather missing inputs
- Control verbosity: Specify length limits, number of alternatives, or time horizons
- Request diagnostics: Ask Agentforce to share reasoning, risks, and verification plans when appropriate
---
name: Apex Service Hardening Plan
description: Audit and fortify an Apex service to stay within governor limits while preserving behavior
tags: apex-development, optimization, audit
requires_deploy: false
---
## Context
- Usage profile: [Invocation volume, entry points, data scale]
- Known issues: [Timeouts, limit exceptions, performance complaints]
- Stakeholders: [Product owners, support teams, compliance partners]
## Instructions
1. Summarize existing architecture, dependencies, and limit usage; list assumptions needing confirmation.
2. Propose at least two optimization strategies, including refactor scope, data implications, and rollback considerations.
3. Recommend a preferred strategy once assumptions are resolved, detailing implementation phases and change management steps.
## Testing & Reporting
- Define unit, integration, and bulk test coverage with pass criteria.
- Specify telemetry/observability updates (logging, metrics, alerts) to validate success.
- Produce an execution checklist with owners, timelines, and escalation contacts.- β One rule per file - Keep rules focused and modular
- β Use clear names - Describe what the rule enforces
- β Document purpose - Explain why the rule exists
- β Keep updated - Review and refine regularly
- β Version control - Track changes over time
See Contributing for complete details.
Contribution checklist
- Confirm the file lives in the correct category folder
- Complete the YAML frontmatter (
name,description,tags) - Include clear instructions and placeholders for user-specific details
- Add a short note on how others can adapt the prompt, especially for varying Salesforce environments
- Verify the content respects licensing and attribution requirements
- Provide any supporting references or context in the pull request description
Found an issue or have a suggestion?
- Open an issue in GitHub
- Suggest improvements via pull request
- Start a discussion in GitHub Discussions or the pull request thread
To update an existing prompt:
- Edit the
.mdfile - Update the description if behavior changed
- Test the updated prompt
- Commit with clear message:
Update [prompt]: [what changed]
To add a new category:
- Create a new folder in
prompts/orrules/ - Add a
README.mdexplaining the category - Add initial prompts/rules