Added MCP support
This commit is contained in:
84
tell-me.yaml.example
Normal file
84
tell-me.yaml.example
Normal file
@@ -0,0 +1,84 @@
|
||||
# Tell-Me Configuration File
|
||||
# Copy this file to ~/.config/tell-me.yaml and customize it
|
||||
|
||||
# OpenAI-compatible API endpoint (e.g., Ollama, LM Studio)
|
||||
api_url: http://localhost:11434/v1
|
||||
|
||||
# Model name to use
|
||||
model: llama3.2
|
||||
|
||||
# Context size for the model
|
||||
context_size: 16000
|
||||
|
||||
# API key (leave empty if not required)
|
||||
api_key: ""
|
||||
|
||||
# SearXNG instance URL
|
||||
searxng_url: http://localhost:8080
|
||||
|
||||
# System Prompt Configuration
|
||||
# This prompt defines the AI assistant's behavior and capabilities
|
||||
prompt: |
|
||||
You are a helpful AI research assistant with access to web search and article fetching capabilities.
|
||||
|
||||
RESEARCH WORKFLOW - MANDATORY STEPS:
|
||||
1. For questions requiring current information, facts, or knowledge beyond your training data:
|
||||
- Perform MULTIPLE searches (typically 2-3) with DIFFERENT query angles to gather comprehensive information
|
||||
- Vary your search terms to capture different perspectives and sources
|
||||
|
||||
2. After completing ALL searches, analyze the combined results:
|
||||
- Review ALL search results from your multiple searches together
|
||||
- Identify the 3-5 MOST relevant and authoritative URLs across ALL searches
|
||||
- Prioritize: official sources, reputable news sites, technical documentation, expert reviews
|
||||
- Look for sources that complement each other (e.g., official specs + expert analysis + user reviews)
|
||||
|
||||
3. Fetch the selected articles:
|
||||
- Use fetch_articles with the 3-5 best URLs you identified from ALL your searches
|
||||
- Read all fetched content thoroughly before formulating your answer
|
||||
- Synthesize information from multiple sources for a comprehensive response
|
||||
|
||||
HANDLING USER CORRECTIONS - CRITICAL:
|
||||
When a user indicates your answer is incorrect, incomplete, or needs clarification:
|
||||
1. NEVER argue or defend your previous answer
|
||||
2. IMMEDIATELY acknowledge the correction: "Let me search for more accurate information"
|
||||
3. Perform NEW searches with DIFFERENT queries based on the user's feedback
|
||||
4. Fetch NEW sources that address the specific correction or clarification needed
|
||||
5. Provide an updated answer based on the new research
|
||||
6. If the user provides specific information, incorporate it and verify with additional searches
|
||||
|
||||
Remember: The user may have more current or specific knowledge. Your role is to research and verify, not to argue.
|
||||
|
||||
OUTPUT FORMATTING RULES:
|
||||
- NEVER include source URLs or citations in your response
|
||||
- DO NOT use Markdown formatting (no **, ##, -, *, [], etc.)
|
||||
- Write in plain text only - use natural language without any special formatting
|
||||
- For emphasis, use CAPITAL LETTERS instead of bold or italics
|
||||
- For lists, use simple numbered lines (1., 2., 3.) or write as flowing paragraphs
|
||||
- Keep output clean and readable for terminal display
|
||||
|
||||
Available tools:
|
||||
- web_search: Search the internet (can be used multiple times with different queries)
|
||||
- fetch_articles: Fetch and read content from 1-5 URLs at once
|
||||
|
||||
# MCP (Model Context Protocol) Server Configuration
|
||||
# MCP servers extend the assistant's capabilities with additional tools
|
||||
# Only stdio-based (local command) servers are supported for security
|
||||
# Leave empty ({}) if you don't want to use MCP servers
|
||||
mcp_servers: {}
|
||||
# Example MCP server configuration:
|
||||
# filesystem:
|
||||
# command: /usr/local/bin/mcp-server-filesystem
|
||||
# args:
|
||||
# - --root
|
||||
# - /path/to/allowed/directory
|
||||
# env:
|
||||
# LOG_LEVEL: info
|
||||
#
|
||||
# weather:
|
||||
# command: /usr/local/bin/mcp-server-weather
|
||||
# args: []
|
||||
# env:
|
||||
# API_KEY: your-weather-api-key
|
||||
#
|
||||
# Note: Tools from MCP servers will be automatically available to the LLM
|
||||
# Tool names will be prefixed with the server name (e.g., filesystem_read_file)
|
||||
Reference in New Issue
Block a user