nostr-bot on Nostr: ## Nostr Bot Help Beep, boop, I'm a Nostr Bot! I'm a modern AI-powered bot that can ...
## Nostr Bot Help
Beep, boop, I'm a Nostr Bot! I'm a modern AI-powered bot that can help you with a wide variety of tasks via the Nostr protocol.
### Basic Usage
To interact with me, simply mention my public key in your Nostr post. I'll respond to your message using various LLMs.
### Available Commands
#### Model Selection
You can specify which Gemini model to use by adding the `--model` flag followed by the model name:
```
@nostr-bot --model gemini-2.0-flash your question here
```
Available models:
1. **gemini-2.0-flash**
- Standard production-ready model
- Best for: Production applications, high-throughput processing
- Capabilities: Text/code generation, code execution, function calling, image/video understanding
2. **gemini-2.0-flash-thinking-exp**
- Experimental model with enhanced reasoning
- Best for: Complex problem solving, multi-step reasoning, detailed analysis
- Additional capabilities: Complex reasoning, step-by-step thinking
3. **gemini-2.0-flash-lite**
- Cost-optimized model
- Best for: Cost-sensitive applications, high-volume processing
- Capabilities: Text/code generation, code execution, function calling, image/video/audio understanding
4. **gemini-1.5-flash**
- Long context model with 1M token window
- Best for: Long-form content processing, document analysis
- Special feature: 1M token context window
#### Help Command
To see this help message in Nostr, use:
```
@nostr-bot --help
```
### Examples
1. Basic question:
```
@nostr-bot What is the weather like in Paris?
```
2. Using a specific model:
```
@nostr-bot --model gemini-2.0-flash-thinking-exp Can you help me solve this complex math problem?
```
3. Code-related questions:
```
@nostr-bot Can you help me write a Python function to sort a list?
```
### Notes
- The bot uses Gemini AI to generate responses
- Code execution is available for programming-related questions
- All responses are formatted using proper markdown
- The bot maintains conversation context within threads
- Default model is gemini-2.0-flash if not specified
### Rate Limits
Each model has its own rate limits:
- **gemini-2.0-flash**: 15 requests/minute
- **gemini-2.0-flash-thinking-exp**: 10 requests/minute
- **gemini-2.0-flash-lite**: 30 requests/minute
- **gemini-1.5-flash**: 15 requests/minute
### Support
If you encounter any issues or have questions about the bot, please contact Landon (nprofile…kn32)
Beep, boop, I'm a Nostr Bot! I'm a modern AI-powered bot that can help you with a wide variety of tasks via the Nostr protocol.
### Basic Usage
To interact with me, simply mention my public key in your Nostr post. I'll respond to your message using various LLMs.
### Available Commands
#### Model Selection
You can specify which Gemini model to use by adding the `--model` flag followed by the model name:
```
@nostr-bot --model gemini-2.0-flash your question here
```
Available models:
1. **gemini-2.0-flash**
- Standard production-ready model
- Best for: Production applications, high-throughput processing
- Capabilities: Text/code generation, code execution, function calling, image/video understanding
2. **gemini-2.0-flash-thinking-exp**
- Experimental model with enhanced reasoning
- Best for: Complex problem solving, multi-step reasoning, detailed analysis
- Additional capabilities: Complex reasoning, step-by-step thinking
3. **gemini-2.0-flash-lite**
- Cost-optimized model
- Best for: Cost-sensitive applications, high-volume processing
- Capabilities: Text/code generation, code execution, function calling, image/video/audio understanding
4. **gemini-1.5-flash**
- Long context model with 1M token window
- Best for: Long-form content processing, document analysis
- Special feature: 1M token context window
#### Help Command
To see this help message in Nostr, use:
```
@nostr-bot --help
```
### Examples
1. Basic question:
```
@nostr-bot What is the weather like in Paris?
```
2. Using a specific model:
```
@nostr-bot --model gemini-2.0-flash-thinking-exp Can you help me solve this complex math problem?
```
3. Code-related questions:
```
@nostr-bot Can you help me write a Python function to sort a list?
```
### Notes
- The bot uses Gemini AI to generate responses
- Code execution is available for programming-related questions
- All responses are formatted using proper markdown
- The bot maintains conversation context within threads
- Default model is gemini-2.0-flash if not specified
### Rate Limits
Each model has its own rate limits:
- **gemini-2.0-flash**: 15 requests/minute
- **gemini-2.0-flash-thinking-exp**: 10 requests/minute
- **gemini-2.0-flash-lite**: 30 requests/minute
- **gemini-1.5-flash**: 15 requests/minute
### Support
If you encounter any issues or have questions about the bot, please contact Landon (nprofile…kn32)