Back to Blog
AI + Pentest2026-03-0810 min

How I use LLMs For Security Work: Part 2

From prompting to agents, workflows and assistants — a detailed guide on leveraging LLMs effectively for security work with advanced patterns and real-world examples building a threat enrichment pipeline.

D
DNA Research Team
Research Team, DNA Cyber Security

How I use LLMs For Security Work

In the previous post, author Josh Rickard shared fundamental LLM approaches. This follow-up dives into more specific examples, including more "advanced" patterns developed over approximately one year of practical application.

Terminology

Prompting

Prompting is direct interaction — providing statements with minimal context and general requirements, either as a question or problem description. The approach involves iterative cycles of question, answer, fix, validate, and test. The author distinguishes between casual "vibe coding" and the more rigorous requirement to deeply understand problems before articulating them to LLMs.

Agents

Agents involve very precise context given to an LLM — creating "guardrails for a request" through documentation, use cases, requirements specifications, and decision negations. These range from 1,000-word documents to 50,000+ character prompts with extensive contextual detail.

SKILLS.md

Markdown files capturing organizational knowledge about performing specific actions, feeding into agents for context and decision-making purposes.

Workflows

Workflows are direction documents for how an LLM should "analyze", "examine", or "perform" tasks, with defined steps including tool usage and result interpretation — similar to incident response playbooks but in markdown format.

Assistants

Assistants are the "newest evolution" — combining multiple workflows, triggering SKILLS.md files, and using tools (MCP servers) with discrete agents to perform tasks, with defined decision-making mechanisms controlled through API hooks.

Context Is Key

The more concrete outputs, decision points, paths to follow, descriptions of key terms, specifications, requirements and more in the context you provide to an LLM the better inference results.

The author demonstrates through example: a Chrome profile automation request improves significantly when adding context: "Ensure that the provided Google Chrome Profile name represents the display name shown within Google Chrome. For example, I have a profile named 'Thrunting - X organization name' and it should match the input value." Technical detail: profile display names reside in ./LocalState rather than parent directories.

Advanced example of LLM context

Back to the Basics

You must intimately know the problem before effectively utilizing LLMs for solutions. The author recommends several important categories of context to provide:

Personas / Identities / Personalities

Describe your preferred frameworks, technologies, and design patterns to shape LLM outputs toward familiar tools and approaches rather than unfamiliar alternatives.

Perspectives & Goals

Define viewpoints (SDET perspective, red team perspective) and specific objectives (threat discovery, novel threat identification).

Tools

List available resources including MCP servers, local infrastructure, and specific technical guidelines with examples of when to employ each tool.

Documentation

Provide API specifications, JSONSchemas, repositories, organizational terminology, similar solved problems, and critically, examples of what NOT to do with specific scenarios and their explanations.

Requirements

Specify technical and business requirements, framework/language guidelines, and identify potential complications.

Advanced Example: Threat Enrichment Pipeline

The author provides a hypothetical scenario (developed in approximately 30 minutes) for building a threat enrichment pipeline. This example demonstrates comprehensive context provision:

Description: A senior security researcher and senior software engineer task: building a highly scalable threat enrichment pipeline processing 100 to 1 billion IPs daily, extracting IPv4/IPv6 addresses with contextual information, performing geolocation enrichment with required 10ms latency.

Tools Specified

  • Golang
  • gRPC & protobuf for inter-service communication
  • PostgreSQL with ip4r extension
  • pgbouncer for scalable queries
  • Redis (rueidis client library) for caching
  • Kubernetes with lightweight images
  • Prometheus for monitoring and metrics

Requirements

  • 1 billion events daily processing capability
  • Complete IP extraction with contextual preservation (source vs. destination)
  • Extractable format easily modified for future support
  • 18-hour minimum cache retention with invalidation capability
  • Encrypted communications with secure authorization between services
  • ip4r-typed indexes
  • Straightforward geolocation data upgrade procedures

What Not to Do: The author emphasizes — do not create separate IPv4 and IPv6 processing paths. Specifying what NOT to do is just as important as specifying what to do.

Conclusion

Individuals capable of describing problems thoroughly and efficiently will "prevail as we move into this new era of technology advancement." Effective LLM utilization requires the discipline of complete problem comprehension — comparable to project management guidance but applied to AI interaction.

You must intimately know the problem before effectively utilizing LLMs for solutions. Context is everything — the more concrete, the better the results.

#LLM#Security#Prompting#Agents#Workflows#Threat Enrichment#AI

Ready for Human + AI Security?

Experience next-gen Penetration Testing — where 15+ year experts combine cutting-edge AI to protect your business.

Contact us now