Welcome to Snow CLI! Agentic coding in your terminal.
-
Type
snowto start Snow CLI or click the launch icon in the IDE plugin -
Snow CLI's default language is
English. You can go toLanguage Settingsto modify your language preference
After setting your language preference, go to API and Model Settings
The configuration interface provides comprehensive AI service configuration capabilities, supporting multiple profile management and rich model parameter settings.
Purpose: Manage multiple configuration sets for quick switching between different scenarios
Operations:
- Press Enter to access the profile selection interface
- Use up/down arrow keys to select profiles
- The currently active profile will display a green ✓ mark
Quick Actions:
- Press
nkey: Create new profile (requires entering profile name) - Press
dkey: Delete current profile (the default profile cannot be deleted)
Important Notes:
- Each profile independently saves all settings
- Switching profiles immediately loads all settings from that profile
Purpose: Base address of the API service
Configuration Method:
- Press Enter to enter edit mode
- Input the complete API address
- Press Enter again to confirm
Standard Addresses:
-
OpenAI Chat Completion
https://api.openai.com/v1For OpenAI's standard chat completion API
-
OpenAI Responses
https://api.openai.com/v1For OpenAI's response API with reasoning capabilities
-
Gemini
https://generativelanguage.googleapis.com/v1betaGoogle Gemini API service address
-
Anthropic
https://api.anthropic.com/v1API service address for Claude models
Important Notes:
- Supports proxy or third-party relay service addresses
- Ensure the address format is correct, starting with
https:// - Address typically includes version number at the end (e.g.,
/v1)
Purpose: Access key for API service
Configuration Method:
- Press Enter to enter edit mode
- Input the complete API Key
- Input will be automatically hidden and displayed as
*characters - Press Enter again to confirm
Important Notes:
- API Keys typically start with specific prefixes (e.g.,
sk-for OpenAI) - Keep your API Key secure to prevent disclosure
- Will only display as asterisks, never in plain text
Purpose: Select API calling method; different methods support different features
Available Options:
- OpenAI Chat Completion: Standard OpenAI chat API
- OpenAI Responses: OpenAI API with reasoning mode support
- Gemini: Google's Gemini model
- Anthropic: Claude model
Configuration Method:
- Press Enter to open selection list
- Use up/down arrow keys to select
- Press Enter to confirm
Important Notes:
- Different request methods display different advanced configuration options
- When switching request methods, specific feature configurations will automatically adjust
Purpose: Select which system prompt to use for the current profile
Available Options:
- Follow Global (None): Use global settings, no system prompt currently activated
- Follow Global (Name): Use the system prompt activated in global settings
- Not Use: Explicitly disable system prompt, even if there's an activated global prompt
- Select Specific Prompt: Choose from the list of configured system prompts
Configuration Method:
- Press Enter to open selection list
- Use up/down arrow keys to select
- Press Enter to confirm
Description:
- System prompts can be created and managed in the "System Prompt Management" interface
- Profile-level settings override global settings
- Selecting "Not Use" allows you to temporarily disable system prompts in specific scenarios
Purpose: Select which custom headers scheme to use for the current profile
Available Options:
- Follow Global (None): Use global settings, no headers scheme currently activated
- Follow Global (Name): Use the headers scheme activated in global settings
- Not Use: Explicitly disable custom headers, even if there's an activated global scheme
- Select Specific Scheme: Choose from the list of configured headers schemes
Configuration Method:
- Press Enter to open selection list
- Use up/down arrow keys to select
- Press Enter to confirm
Description:
- Custom headers schemes can be created and managed in the "Custom Headers Management" interface
- Profile-level settings override global settings
- Selecting "Not Use" allows you to temporarily disable custom headers in specific scenarios
Purpose: Automatically compress long text content to reduce token consumption
Default: Enabled
Configuration Method:
- Press Enter or Space key to toggle Enabled/Disabled status
- Displays "Enabled" or "Disabled"
Recommendation: Enabling can reduce API call costs but may lose some context details
Purpose: Display AI's reasoning and thinking process in the interface
Default: Enabled
Configuration Method:
- Press Enter or Space key to toggle Enabled/Disabled status
- Displays "Enabled" or "Disabled"
Recommendation: Enabling helps understand AI's reasoning process, useful for debugging and understanding results
When selecting the Anthropic request method, the following configuration options will appear:
Purpose: Enable Anthropic's Beta version features
Default: Disabled
Configuration Method:
- Press Enter or Space key to toggle Enabled/Disabled status
Important Notes: Beta features may be unstable, use with caution
Purpose: Set the expiration time for prompt caching
Available Options:
5m: 5 minutes1h: 1 hour
Default: 5 minutes
Configuration Method:
- Press Enter to open selection list
- Select cache duration
- Press Enter to confirm
Description: Longer cache times can reduce token consumption for repeated content
Purpose: Enable Claude's extended thinking feature
Default: Disabled
Configuration Method:
- Press Enter or Space key to toggle Enabled/Disabled status
Description: When enabled, AI will perform deeper reasoning
Purpose: Set the maximum token count for extended thinking mode
Default: 10000
Range: Minimum value 1000
Configuration Method:
- Press Enter to enter edit mode
- Input number (supports backspace deletion)
- Press Enter to confirm
Important Notes:
- Larger thinking budget enables deeper AI reasoning but consumes more tokens
- If input value is below minimum, it will automatically adjust to minimum value when saved
When selecting the Gemini request method, the following configuration options will appear:
Purpose: Enable Gemini's thinking and reasoning feature
Default: Disabled
Configuration Method:
- Press Enter or Space key to toggle Enabled/Disabled status
Purpose: Set the budget value for Gemini thinking mode
Default: 1024
Range: Minimum value 1
Configuration Method:
- Press Enter to enter edit mode
- Input number (supports backspace deletion)
- Press Enter to confirm
When selecting the OpenAI Responses request method, the following configuration options will appear:
Purpose: Enable OpenAI's reasoning feature
Default: Disabled
Configuration Method:
- Press Enter or Space key to toggle Enabled/Disabled status
Purpose: Set the intensity level of reasoning mode
Available Options:
LOW: Low-intensity reasoningMEDIUM: Medium-intensity reasoningHIGH: High-intensity reasoningXHIGH: Ultra-high intensity reasoning (only supported in responses method)
Default: HIGH
Configuration Method:
- Press Enter to open selection list
- Use up/down arrow keys to select intensity
- Press Enter to confirm
Important Notes: Higher reasoning intensity provides deeper reasoning but increases time and token consumption
Purpose: Primary model for complex tasks
Configuration Method:
- Press Enter to automatically fetch available model list (requires correct Base URL and API Key configuration)
- If fetching fails, will automatically enter manual input mode
- Can use alphanumeric input for fuzzy search filtering
- Select "Manual Input" option to manually enter model name
- Press
mkey to quickly enter manual input mode
Common Model Examples:
- OpenAI:
gpt-4,gpt-4-turbo,gpt-4o - Claude:
claude-3-5-sonnet-20241022,claude-3-opus-20240229 - Gemini:
gemini-2.0-flash-exp,gemini-pro
Recommendation: Choose more powerful models for complex programming tasks
Purpose: Auxiliary model for simple tasks
Configuration Method: Same as Advanced Model
Common Model Examples:
- OpenAI:
gpt-3.5-turbo,gpt-4o-mini - Claude:
claude-3-haiku-20240307 - Gemini:
gemini-flash
Recommendation: Choose models with fast response speed and lower cost
Purpose: Maximum context window size supported by the model
Default: 4000
Range: Minimum value 4000
Configuration Method:
- Press Enter to enter edit mode
- Input number (supports backspace deletion)
- Press Enter to confirm
Common Model Context Capacities:
- Claude 3.5 Sonnet: 200000
- GPT-4 Turbo: 128000
- GPT-4: 8192
- Gemini 2.0 Flash: 1000000
- Gemini Pro: 32768
Important Notes:
- Must be set to the actual context size supported by the model
- Setting too high will cause API call failures
- Setting too low will limit conversation length
Purpose: Maximum token count allowed for single response generation
Default: 4096
Range: Minimum value 100
Configuration Method:
- Press Enter to enter edit mode
- Input number (supports backspace deletion)
- Press Enter to confirm
Common Model Output Capacities:
- Claude 3.5 Sonnet: 64000
- GPT-4 Turbo: 4096
- GPT-4: 8192
- Gemini 2.0 Flash: 8192
Important Notes:
- Different models support different maximum output token counts
- Setting too high will increase response time and cost
- Recommend setting reasonably based on actual needs
Purpose: Controls matching precision when AI modifies files, affecting code editing accuracy
Default: 0.75
Range: 0.1 - 1.0 (decimal)
Configuration Method:
- Press Enter to enter edit mode
- Input decimal number (e.g., 0.75, 0.8)
- Supports backspace deletion
- Press Enter to confirm
Explanation:
- Higher threshold: Stricter matching, only very similar code will be modified, reducing risk of incorrect modifications
- Lower threshold: Looser matching, allows more differences (like indentation, whitespace), increasing flexibility
- Default 0.75: Balances precision and flexibility, suitable for most scenarios
Recommendations:
- Keep default value unless encountering specific issues
- If AI frequently can't find code to modify, slightly decrease (e.g., 0.6-0.7)
- If AI incorrectly modifies similar but different code, slightly increase (e.g., 0.8-0.85)
- Not recommended to set below 0.6 or above 0.9
Important Notes:
- This configuration affects all code editing operations
- Changes take effect immediately without restart
- Incorrect settings may cause code editing failures or incorrect modifications
- Up/Down Arrow Keys: Move between configuration items
- Enter Key: Enter edit mode or confirm input
- Esc Key: Save configuration and exit
- Ctrl+S / Cmd+S: Quick save configuration
- Space Key: Toggle switch-type configuration items (e.g., Enable/Disable)
- Configuration interface displays current position at top:
(Current item/Total items) - When configuration items exceed 8, will automatically scroll
- Currently selected configuration item displays
❯marker
In the model selection interface:
- Alphanumeric Input: Real-time filtering of model list
- Backspace: Delete filter characters
- Esc Key: Exit selection interface
- m Key: Quick entry to manual input mode
When editing token-related configurations:
- Number Keys: Append digits
- Backspace/Delete: Delete last digit
- Enter Key: Confirm and automatically validate minimum value
The system will automatically validate when saving configuration:
- Required Field Check: Base URL and API Key must be filled
- Format Validation: Check if Base URL format is correct
- Value Range: Automatically adjust token configurations above minimum values
- Request Method Matching: Validate compatibility between selected model and request method
Error Messages:
- Red error information will display at bottom of interface when validation fails
- Can try saving again after fixing errors
- Main Configuration File:
~/.snowcli/config.json - Profile Directory:
~/.snowcli/profiles/ - Auto Save: Automatically saves to currently active profile when exiting configuration interface
Solution:
- Check if Base URL and API Key are correct
- Check network connection and proxy settings
- If it continues to fail, use manual input mode (press
mkey)
Solution:
- Confirm you have saved configuration by pressing Esc or Ctrl+S
- Restart Snow CLI to ensure configuration is loaded
- Check if the correct profile is selected
Solution:
- Check if Max Context Tokens is set correctly
- Confirm it doesn't exceed the model's actual supported context size
- Appropriately reduce Max Tokens setting
Explanation: Specific configuration items for different request methods (such as Anthropic's Thinking feature) will automatically show/hide based on the current method. Configuration values are still saved and will be restored when switching back.
- First-Time Configuration: First set Basic configuration (Base URL, API Key, Request Method), then configure advanced features
- Multi-Scenario Usage: Create different profiles for different projects
- Cost Optimization: Reasonably set Max Tokens, enable Auto Compress feature
- Performance Optimization: Choose appropriate models based on task complexity, use Basic Model for simple tasks
- Debugging Recommendation: Enable Show Thinking to view AI reasoning process, helpful for understanding and optimizing prompts


