Files
ollama-k80-lab/prompts/README.md
Shang Chieh Tseng d8563fcc51 Add organized prompts structure with templates, examples, and responses
- Create templates/ for reusable prompt templates
- Create examples/ for test context inputs
- Create responses/ for multi-model output comparison
- Move Professional Communication Assistant to templates/
- Rename Task Context to deadline-extension-request example
- Add comprehensive README with usage guidelines

🤖 Generated with [Claude Code](https://claude.ai/code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-07-20 23:57:39 +08:00

2.2 KiB

Prompts

This directory contains LLM prompt templates and example contexts for workflow automation in the Ollama K80 Lab environment.

Structure

prompts/
├── templates/          # Reusable prompt templates
│   └── professional-communication-assistant.md
├── examples/           # Context examples for testing prompts
│   └── deadline-extension-request.md
└── responses/          # Model responses organized by example
    └── deadline-extension-request/
        ├── qwen2.5-vl.md
        ├── gemma-3-12b.md
        ├── phi4-14b.md
        └── deepseek-r1-32b.md

Usage

Templates

Prompt templates are structured prompts designed for specific use cases. They include:

  • Clear instructions and guidelines
  • Configurable parameters (marked with placeholders)
  • Multiple tone/style variations where applicable

Examples

Example contexts provide sample inputs to test and demonstrate prompt templates:

  • Real-world scenarios
  • Edge cases
  • Different complexity levels

Responses

Model responses are organized by example scenario, with each model's output saved separately:

  • Compare different models on the same prompt/context
  • Track model performance over time
  • Analyze response quality and consistency
  • Use kebab-case filenames matching model names

Integration

These prompts integrate with:

  • Dify workflows - For automated LLM-powered QA tasks
  • VS Code Continue plugin - For development assistance
  • Ollama API - Running on K80-optimized containers

Adding New Prompts

  1. Create template in templates/ with descriptive kebab-case naming
  2. Add corresponding examples in examples/
  3. Test with your target LLM models
  4. Save model responses in responses/example-name/model-name.md
  5. Update this README if needed

Testing Workflow

  1. Use template + example to generate prompts
  2. Run against multiple models (Qwen2.5-VL, Gemma 3, Phi-4, DeepSeek-R1, etc.)
  3. Save each model's response in the appropriate response folder
  4. Compare outputs for quality, consistency, and usefulness
  • /dify/ - Workflow automation configurations
  • /ollama37/ - Docker runtime for LLM execution
  • CLAUDE.md - Project development guidelines