diff --git a/api/core/model_runtime/model_providers/novita/_assets/icon_l_en.svg b/api/core/model_runtime/model_providers/novita/_assets/icon_l_en.svg index 5c92cdbc6d8466..c27089a2ac0737 100644 --- a/api/core/model_runtime/model_providers/novita/_assets/icon_l_en.svg +++ b/api/core/model_runtime/model_providers/novita/_assets/icon_l_en.svg @@ -1,19 +1,11 @@ - - - - - - - - - - - - + + + + + - - - - + + + diff --git a/api/core/model_runtime/model_providers/novita/_assets/icon_s_en.svg b/api/core/model_runtime/model_providers/novita/_assets/icon_s_en.svg index 798c1d63485221..ae2d4815c46ad3 100644 --- a/api/core/model_runtime/model_providers/novita/_assets/icon_s_en.svg +++ b/api/core/model_runtime/model_providers/novita/_assets/icon_s_en.svg @@ -1,10 +1,3 @@ - - - - - - - - - + + diff --git a/api/core/model_runtime/model_providers/novita/llm/L3-8B-Stheno-v3.2.yaml b/api/core/model_runtime/model_providers/novita/llm/L3-8B-Stheno-v3.2.yaml new file mode 100644 index 00000000000000..443cbef4b34356 --- /dev/null +++ b/api/core/model_runtime/model_providers/novita/llm/L3-8B-Stheno-v3.2.yaml @@ -0,0 +1,41 @@ +model: Sao10K/L3-8B-Stheno-v3.2 +label: + zh_Hans: Sao10K/L3-8B-Stheno-v3.2 + en_US: Sao10K/L3-8B-Stheno-v3.2 +model_type: llm +features: + - agent-thought +model_properties: + mode: chat + context_size: 8192 +parameter_rules: + - name: temperature + use_template: temperature + min: 0 + max: 2 + default: 1 + - name: top_p + use_template: top_p + min: 0 + max: 1 + default: 1 + - name: max_tokens + use_template: max_tokens + min: 1 + max: 2048 + default: 512 + - name: frequency_penalty + use_template: frequency_penalty + min: -2 + max: 2 + default: 0 + - name: presence_penalty + use_template: presence_penalty + min: -2 + max: 2 + default: 0 +pricing: + input: '0.0005' + output: '0.0005' + unit: '0.0001' + currency: USD diff --git a/api/core/model_runtime/model_providers/novita/llm/_position.yaml b/api/core/model_runtime/model_providers/novita/llm/_position.yaml new file mode 100644 index 00000000000000..4176fc19caf6cc --- /dev/null +++ b/api/core/model_runtime/model_providers/novita/llm/_position.yaml @@ -0,0 +1,40 @@ +# Deepseek Models +- deepseek/deepseek_v3 + +# LLaMA Models +- meta-llama/llama-3.3-70b-instruct +- meta-llama/llama-3.2-11b-vision-instruct +- meta-llama/llama-3.2-3b-instruct +- meta-llama/llama-3.2-1b-instruct +- meta-llama/llama-3.1-70b-instruct +- meta-llama/llama-3.1-8b-instruct +- meta-llama/llama-3.1-8b-instruct-max +- meta-llama/llama-3.1-8b-instruct-bf16 +- meta-llama/llama-3-70b-instruct +- meta-llama/llama-3-8b-instruct + +# Mistral Models +- mistralai/mistral-nemo +- mistralai/mistral-7b-instruct + +# Qwen Models +- qwen/qwen-2.5-72b-instruct +- qwen/qwen-2-72b-instruct +- qwen/qwen-2-vl-72b-instruct +- qwen/qwen-2-7b-instruct + +# Other Models +- sao10k/L3-8B-Stheno-v3.2 +- sao10k/l3-70b-euryale-v2.1 +- sao10k/l31-70b-euryale-v2.2 +- sao10k/l3-8b-lunaris +- jondurbin/airoboros-l2-70b +- cognitivecomputations/dolphin-mixtral-8x22b +- google/gemma-2-9b-it +- nousresearch/hermes-2-pro-llama-3-8b +- sophosympatheia/midnight-rose-70b +- gryphe/mythomax-l2-13b +- nousresearch/nous-hermes-llama2-13b +- openchat/openchat-7b +- teknium/openhermes-2.5-mistral-7b +- microsoft/wizardlm-2-8x22b diff --git a/api/core/model_runtime/model_providers/novita/llm/deepseek_v3.yaml b/api/core/model_runtime/model_providers/novita/llm/deepseek_v3.yaml new file mode 100644 index 00000000000000..b18746622cb376 --- /dev/null +++ b/api/core/model_runtime/model_providers/novita/llm/deepseek_v3.yaml @@ -0,0 +1,41 @@ +model: deepseek/deepseek_v3 +label: + zh_Hans: deepseek/deepseek_v3 + en_US: deepseek/deepseek_v3 +model_type: llm +features: + - agent-thought +model_properties: + mode: chat + context_size: 64000 +parameter_rules: + - name: temperature + use_template: temperature + min: 0 + max: 2 + default: 1 + - name: top_p + use_template: top_p + min: 0 + max: 1 + default: 1 + - name: max_tokens + use_template: max_tokens + min: 1 + max: 2048 + default: 512 + - name: frequency_penalty + use_template: frequency_penalty + min: -2 + max: 2 + default: 0 + - name: presence_penalty + use_template: presence_penalty + min: -2 + max: 2 + default: 0 +pricing: + input: '0.0089' + output: '0.0089' + unit: '0.0001' + currency: USD diff --git a/api/core/model_runtime/model_providers/novita/llm/l3-8b-lunaris.yaml b/api/core/model_runtime/model_providers/novita/llm/l3-8b-lunaris.yaml new file mode 100644 index 00000000000000..d28c84084ab926 --- /dev/null +++ b/api/core/model_runtime/model_providers/novita/llm/l3-8b-lunaris.yaml @@ -0,0 +1,41 @@ +model: sao10k/l3-8b-lunaris +label: + zh_Hans: sao10k/l3-8b-lunaris + en_US: sao10k/l3-8b-lunaris +model_type: llm +features: + - agent-thought +model_properties: + mode: chat + context_size: 8192 +parameter_rules: + - name: temperature + use_template: temperature + min: 0 + max: 2 + default: 1 + - name: top_p + use_template: top_p + min: 0 + max: 1 + default: 1 + - name: max_tokens + use_template: max_tokens + min: 1 + max: 2048 + default: 512 + - name: frequency_penalty + use_template: frequency_penalty + min: -2 + max: 2 + default: 0 + - name: presence_penalty + use_template: presence_penalty + min: -2 + max: 2 + default: 0 +pricing: + input: '0.0005' + output: '0.0005' + unit: '0.0001' + currency: USD diff --git a/api/core/model_runtime/model_providers/novita/llm/l31-70b-euryale-v2.2.yaml b/api/core/model_runtime/model_providers/novita/llm/l31-70b-euryale-v2.2.yaml new file mode 100644 index 00000000000000..9c39ce51c3904d --- /dev/null +++ b/api/core/model_runtime/model_providers/novita/llm/l31-70b-euryale-v2.2.yaml @@ -0,0 +1,41 @@ +model: sao10k/l31-70b-euryale-v2.2 +label: + zh_Hans: sao10k/l31-70b-euryale-v2.2 + en_US: sao10k/l31-70b-euryale-v2.2 +model_type: llm +features: + - agent-thought +model_properties: + mode: chat + context_size: 16000 +parameter_rules: + - name: temperature + use_template: temperature + min: 0 + max: 2 + default: 1 + - name: top_p + use_template: top_p + min: 0 + max: 1 + default: 1 + - name: max_tokens + use_template: max_tokens + min: 1 + max: 2048 + default: 512 + - name: frequency_penalty + use_template: frequency_penalty + min: -2 + max: 2 + default: 0 + - name: presence_penalty + use_template: presence_penalty + min: -2 + max: 2 + default: 0 +pricing: + input: '0.0148' + output: '0.0148' + unit: '0.0001' + currency: USD diff --git a/api/core/model_runtime/model_providers/novita/llm/llama-3-8b-instruct.yaml b/api/core/model_runtime/model_providers/novita/llm/llama-3-8b-instruct.yaml index 9b5e5df4d071fe..7a754dfc11462c 100644 --- a/api/core/model_runtime/model_providers/novita/llm/llama-3-8b-instruct.yaml +++ b/api/core/model_runtime/model_providers/novita/llm/llama-3-8b-instruct.yaml @@ -35,7 +35,7 @@ parameter_rules: max: 2 default: 0 pricing: - input: '0.00063' - output: '0.00063' + input: '0.0004' + output: '0.0004' unit: '0.0001' currency: USD diff --git a/api/core/model_runtime/model_providers/novita/llm/llama-3.1-70b-instruct.yaml b/api/core/model_runtime/model_providers/novita/llm/llama-3.1-70b-instruct.yaml index 3a5c29c40fd7c7..96ef12a41c5556 100644 --- a/api/core/model_runtime/model_providers/novita/llm/llama-3.1-70b-instruct.yaml +++ b/api/core/model_runtime/model_providers/novita/llm/llama-3.1-70b-instruct.yaml @@ -7,7 +7,7 @@ features: - agent-thought model_properties: mode: chat - context_size: 8192 + context_size: 32768 parameter_rules: - name: temperature use_template: temperature @@ -35,7 +35,7 @@ parameter_rules: max: 2 default: 0 pricing: - input: '0.0055' - output: '0.0076' + input: '0.0034' + output: '0.0039' unit: '0.0001' currency: USD diff --git a/api/core/model_runtime/model_providers/novita/llm/llama-3.1-8b-instruct-bf16.yaml b/api/core/model_runtime/model_providers/novita/llm/llama-3.1-8b-instruct-bf16.yaml new file mode 100644 index 00000000000000..5ef72c770fde0e --- /dev/null +++ b/api/core/model_runtime/model_providers/novita/llm/llama-3.1-8b-instruct-bf16.yaml @@ -0,0 +1,41 @@ +model: meta-llama/llama-3.1-8b-instruct-bf16 +label: + zh_Hans: meta-llama/llama-3.1-8b-instruct-bf16 + en_US: meta-llama/llama-3.1-8b-instruct-bf16 +model_type: llm +features: + - agent-thought +model_properties: + mode: chat + context_size: 8192 +parameter_rules: + - name: temperature + use_template: temperature + min: 0 + max: 2 + default: 1 + - name: top_p + use_template: top_p + min: 0 + max: 1 + default: 1 + - name: max_tokens + use_template: max_tokens + min: 1 + max: 2048 + default: 512 + - name: frequency_penalty + use_template: frequency_penalty + min: -2 + max: 2 + default: 0 + - name: presence_penalty + use_template: presence_penalty + min: -2 + max: 2 + default: 0 +pricing: + input: '0.0006' + output: '0.0006' + unit: '0.0001' + currency: USD diff --git a/api/core/model_runtime/model_providers/novita/llm/llama-3.1-8b-instruct-max.yaml b/api/core/model_runtime/model_providers/novita/llm/llama-3.1-8b-instruct-max.yaml new file mode 100644 index 00000000000000..d75b3e461fad5e --- /dev/null +++ b/api/core/model_runtime/model_providers/novita/llm/llama-3.1-8b-instruct-max.yaml @@ -0,0 +1,41 @@ +model: meta-llama/llama-3.1-8b-instruct-max +label: + zh_Hans: meta-llama/llama-3.1-8b-instruct-max + en_US: meta-llama/llama-3.1-8b-instruct-max +model_type: llm +features: + - agent-thought +model_properties: + mode: chat + context_size: 16384 +parameter_rules: + - name: temperature + use_template: temperature + min: 0 + max: 2 + default: 1 + - name: top_p + use_template: top_p + min: 0 + max: 1 + default: 1 + - name: max_tokens + use_template: max_tokens + min: 1 + max: 2048 + default: 512 + - name: frequency_penalty + use_template: frequency_penalty + min: -2 + max: 2 + default: 0 + - name: presence_penalty + use_template: presence_penalty + min: -2 + max: 2 + default: 0 +pricing: + input: '0.0005' + output: '0.0005' + unit: '0.0001' + currency: USD diff --git a/api/core/model_runtime/model_providers/novita/llm/llama-3.1-8b-instruct.yaml b/api/core/model_runtime/model_providers/novita/llm/llama-3.1-8b-instruct.yaml index e6ef772a3f0784..49d9393e4da68d 100644 --- a/api/core/model_runtime/model_providers/novita/llm/llama-3.1-8b-instruct.yaml +++ b/api/core/model_runtime/model_providers/novita/llm/llama-3.1-8b-instruct.yaml @@ -7,7 +7,7 @@ features: - agent-thought model_properties: mode: chat - context_size: 8192 + context_size: 16384 parameter_rules: - name: temperature use_template: temperature @@ -35,7 +35,7 @@ parameter_rules: max: 2 default: 0 pricing: - input: '0.001' - output: '0.001' + input: '0.0005' + output: '0.0005' unit: '0.0001' currency: USD diff --git a/api/core/model_runtime/model_providers/novita/llm/llama-3.2-11b-vision-instruct.yaml b/api/core/model_runtime/model_providers/novita/llm/llama-3.2-11b-vision-instruct.yaml new file mode 100644 index 00000000000000..3952a86cacc19a --- /dev/null +++ b/api/core/model_runtime/model_providers/novita/llm/llama-3.2-11b-vision-instruct.yaml @@ -0,0 +1,41 @@ +model: meta-llama/llama-3.2-11b-vision-instruct +label: + zh_Hans: meta-llama/llama-3.2-11b-vision-instruct + en_US: meta-llama/llama-3.2-11b-vision-instruct +model_type: llm +features: + - agent-thought +model_properties: + mode: chat + context_size: 32768 +parameter_rules: + - name: temperature + use_template: temperature + min: 0 + max: 2 + default: 1 + - name: top_p + use_template: top_p + min: 0 + max: 1 + default: 1 + - name: max_tokens + use_template: max_tokens + min: 1 + max: 2048 + default: 512 + - name: frequency_penalty + use_template: frequency_penalty + min: -2 + max: 2 + default: 0 + - name: presence_penalty + use_template: presence_penalty + min: -2 + max: 2 + default: 0 +pricing: + input: '0.0006' + output: '0.0006' + unit: '0.0001' + currency: USD diff --git a/api/core/model_runtime/model_providers/novita/llm/llama-3.2-1b-instruct.yaml b/api/core/model_runtime/model_providers/novita/llm/llama-3.2-1b-instruct.yaml new file mode 100644 index 00000000000000..bf73b5cfc2cd77 --- /dev/null +++ b/api/core/model_runtime/model_providers/novita/llm/llama-3.2-1b-instruct.yaml @@ -0,0 +1,41 @@ +model: meta-llama/llama-3.2-1b-instruct +label: + zh_Hans: meta-llama/llama-3.2-1b-instruct + en_US: meta-llama/llama-3.2-1b-instruct +model_type: llm +features: + - agent-thought +model_properties: + mode: chat + context_size: 131000 +parameter_rules: + - name: temperature + use_template: temperature + min: 0 + max: 2 + default: 1 + - name: top_p + use_template: top_p + min: 0 + max: 1 + default: 1 + - name: max_tokens + use_template: max_tokens + min: 1 + max: 2048 + default: 512 + - name: frequency_penalty + use_template: frequency_penalty + min: -2 + max: 2 + default: 0 + - name: presence_penalty + use_template: presence_penalty + min: -2 + max: 2 + default: 0 +pricing: + input: '0.0002' + output: '0.0002' + unit: '0.0001' + currency: USD diff --git a/api/core/model_runtime/model_providers/novita/llm/llama-3.1-405b-instruct.yaml b/api/core/model_runtime/model_providers/novita/llm/llama-3.2-3b-instruct.yaml similarity index 79% rename from api/core/model_runtime/model_providers/novita/llm/llama-3.1-405b-instruct.yaml rename to api/core/model_runtime/model_providers/novita/llm/llama-3.2-3b-instruct.yaml index c5a45271ae94aa..66b4842faf5d26 100644 --- a/api/core/model_runtime/model_providers/novita/llm/llama-3.1-405b-instruct.yaml +++ b/api/core/model_runtime/model_providers/novita/llm/llama-3.2-3b-instruct.yaml @@ -1,7 +1,7 @@ -model: meta-llama/llama-3.1-405b-instruct +model: meta-llama/llama-3.2-3b-instruct label: - zh_Hans: meta-llama/llama-3.1-405b-instruct - en_US: meta-llama/llama-3.1-405b-instruct + zh_Hans: meta-llama/llama-3.2-3b-instruct + en_US: meta-llama/llama-3.2-3b-instruct model_type: llm features: - agent-thought @@ -35,7 +35,7 @@ parameter_rules: max: 2 default: 0 pricing: - input: '0.03' - output: '0.05' + input: '0.0003' + output: '0.0005' unit: '0.0001' currency: USD diff --git a/api/core/model_runtime/model_providers/novita/llm/llama-3.3-70b-instruct.yaml b/api/core/model_runtime/model_providers/novita/llm/llama-3.3-70b-instruct.yaml new file mode 100644 index 00000000000000..0046dc92f1157e --- /dev/null +++ b/api/core/model_runtime/model_providers/novita/llm/llama-3.3-70b-instruct.yaml @@ -0,0 +1,41 @@ +model: meta-llama/llama-3.3-70b-instruct +label: + zh_Hans: meta-llama/llama-3.3-70b-instruct + en_US: meta-llama/llama-3.3-70b-instruct +model_type: llm +features: + - agent-thought +model_properties: + mode: chat + context_size: 131072 +parameter_rules: + - name: temperature + use_template: temperature + min: 0 + max: 2 + default: 1 + - name: top_p + use_template: top_p + min: 0 + max: 1 + default: 1 + - name: max_tokens + use_template: max_tokens + min: 1 + max: 2048 + default: 512 + - name: frequency_penalty + use_template: frequency_penalty + min: -2 + max: 2 + default: 0 + - name: presence_penalty + use_template: presence_penalty + min: -2 + max: 2 + default: 0 +pricing: + input: '0.0039' + output: '0.0039' + unit: '0.0001' + currency: USD diff --git a/api/core/model_runtime/model_providers/novita/llm/mistral-nemo.yaml b/api/core/model_runtime/model_providers/novita/llm/mistral-nemo.yaml new file mode 100644 index 00000000000000..cb11ebbf941ab1 --- /dev/null +++ b/api/core/model_runtime/model_providers/novita/llm/mistral-nemo.yaml @@ -0,0 +1,41 @@ +model: mistralai/mistral-nemo +label: + zh_Hans: mistralai/mistral-nemo + en_US: mistralai/mistral-nemo +model_type: llm +features: + - agent-thought +model_properties: + mode: chat + context_size: 131072 +parameter_rules: + - name: temperature + use_template: temperature + min: 0 + max: 2 + default: 1 + - name: top_p + use_template: top_p + min: 0 + max: 1 + default: 1 + - name: max_tokens + use_template: max_tokens + min: 1 + max: 2048 + default: 512 + - name: frequency_penalty + use_template: frequency_penalty + min: -2 + max: 2 + default: 0 + - name: presence_penalty + use_template: presence_penalty + min: -2 + max: 2 + default: 0 +pricing: + input: '0.0017' + output: '0.0017' + unit: '0.0001' + currency: USD diff --git a/api/core/model_runtime/model_providers/novita/llm/mythomax-l2-13b.yaml b/api/core/model_runtime/model_providers/novita/llm/mythomax-l2-13b.yaml index 7e4ac3ffe02d0f..fd859dcb39b560 100644 --- a/api/core/model_runtime/model_providers/novita/llm/mythomax-l2-13b.yaml +++ b/api/core/model_runtime/model_providers/novita/llm/mythomax-l2-13b.yaml @@ -35,7 +35,7 @@ parameter_rules: max: 2 default: 0 pricing: - input: '0.00119' - output: '0.00119' + input: '0.0009' + output: '0.0009' unit: '0.0001' currency: USD diff --git a/api/core/model_runtime/model_providers/novita/llm/lzlv_70b.yaml b/api/core/model_runtime/model_providers/novita/llm/openchat-7b.yaml similarity index 83% rename from api/core/model_runtime/model_providers/novita/llm/lzlv_70b.yaml rename to api/core/model_runtime/model_providers/novita/llm/openchat-7b.yaml index 0cc68a8c45082d..cad52a44084998 100644 --- a/api/core/model_runtime/model_providers/novita/llm/lzlv_70b.yaml +++ b/api/core/model_runtime/model_providers/novita/llm/openchat-7b.yaml @@ -1,7 +1,7 @@ -model: lzlv_70b +model: openchat/openchat-7b label: - zh_Hans: lzlv_70b - en_US: lzlv_70b + zh_Hans: openchat/openchat-7b + en_US: openchat/openchat-7b model_type: llm features: - agent-thought @@ -35,7 +35,7 @@ parameter_rules: max: 2 default: 0 pricing: - input: '0.0058' - output: '0.0078' + input: '0.0006' + output: '0.0006' unit: '0.0001' currency: USD diff --git a/api/core/model_runtime/model_providers/novita/llm/Nous-Hermes-2-Mixtral-8x7B-DPO.yaml b/api/core/model_runtime/model_providers/novita/llm/qwen-2-72b-instruct.yaml similarity index 79% rename from api/core/model_runtime/model_providers/novita/llm/Nous-Hermes-2-Mixtral-8x7B-DPO.yaml rename to api/core/model_runtime/model_providers/novita/llm/qwen-2-72b-instruct.yaml index 7ff30458e2ff0e..bb2e935c019bad 100644 --- a/api/core/model_runtime/model_providers/novita/llm/Nous-Hermes-2-Mixtral-8x7B-DPO.yaml +++ b/api/core/model_runtime/model_providers/novita/llm/qwen-2-72b-instruct.yaml @@ -1,7 +1,7 @@ -model: Nous-Hermes-2-Mixtral-8x7B-DPO +model: qwen/qwen-2-72b-instruct label: - zh_Hans: Nous-Hermes-2-Mixtral-8x7B-DPO - en_US: Nous-Hermes-2-Mixtral-8x7B-DPO + zh_Hans: qwen/qwen-2-72b-instruct + en_US: qwen/qwen-2-72b-instruct model_type: llm features: - agent-thought @@ -35,7 +35,7 @@ parameter_rules: max: 2 default: 0 pricing: - input: '0.0027' - output: '0.0027' + input: '0.0034' + output: '0.0039' unit: '0.0001' currency: USD diff --git a/api/core/model_runtime/model_providers/novita/llm/qwen-2-7b-instruct.yaml b/api/core/model_runtime/model_providers/novita/llm/qwen-2-7b-instruct.yaml new file mode 100644 index 00000000000000..ff93d2eb11c9e4 --- /dev/null +++ b/api/core/model_runtime/model_providers/novita/llm/qwen-2-7b-instruct.yaml @@ -0,0 +1,41 @@ +model: qwen/qwen-2-7b-instruct +label: + zh_Hans: qwen/qwen-2-7b-instruct + en_US: qwen/qwen-2-7b-instruct +model_type: llm +features: + - agent-thought +model_properties: + mode: chat + context_size: 32768 +parameter_rules: + - name: temperature + use_template: temperature + min: 0 + max: 2 + default: 1 + - name: top_p + use_template: top_p + min: 0 + max: 1 + default: 1 + - name: max_tokens + use_template: max_tokens + min: 1 + max: 2048 + default: 512 + - name: frequency_penalty + use_template: frequency_penalty + min: -2 + max: 2 + default: 0 + - name: presence_penalty + use_template: presence_penalty + min: -2 + max: 2 + default: 0 +pricing: + input: '0.00054' + output: '0.00054' + unit: '0.0001' + currency: USD diff --git a/api/core/model_runtime/model_providers/novita/llm/qwen-2-vl-72b-instruct.yaml b/api/core/model_runtime/model_providers/novita/llm/qwen-2-vl-72b-instruct.yaml new file mode 100644 index 00000000000000..97097778c3101e --- /dev/null +++ b/api/core/model_runtime/model_providers/novita/llm/qwen-2-vl-72b-instruct.yaml @@ -0,0 +1,41 @@ +model: qwen/qwen-2-vl-72b-instruct +label: + zh_Hans: qwen/qwen-2-vl-72b-instruct + en_US: qwen/qwen-2-vl-72b-instruct +model_type: llm +features: + - agent-thought +model_properties: + mode: chat + context_size: 32768 +parameter_rules: + - name: temperature + use_template: temperature + min: 0 + max: 2 + default: 1 + - name: top_p + use_template: top_p + min: 0 + max: 1 + default: 1 + - name: max_tokens + use_template: max_tokens + min: 1 + max: 2048 + default: 512 + - name: frequency_penalty + use_template: frequency_penalty + min: -2 + max: 2 + default: 0 + - name: presence_penalty + use_template: presence_penalty + min: -2 + max: 2 + default: 0 +pricing: + input: '0.0045' + output: '0.0045' + unit: '0.0001' + currency: USD diff --git a/api/core/model_runtime/model_providers/novita/llm/qwen-2.5-72b-instruct.yaml b/api/core/model_runtime/model_providers/novita/llm/qwen-2.5-72b-instruct.yaml new file mode 100644 index 00000000000000..729fac1da915c8 --- /dev/null +++ b/api/core/model_runtime/model_providers/novita/llm/qwen-2.5-72b-instruct.yaml @@ -0,0 +1,41 @@ +model: qwen/qwen-2.5-72b-instruct +label: + zh_Hans: qwen/qwen-2.5-72b-instruct + en_US: qwen/qwen-2.5-72b-instruct +model_type: llm +features: + - agent-thought +model_properties: + mode: chat + context_size: 32000 +parameter_rules: + - name: temperature + use_template: temperature + min: 0 + max: 2 + default: 1 + - name: top_p + use_template: top_p + min: 0 + max: 1 + default: 1 + - name: max_tokens + use_template: max_tokens + min: 1 + max: 2048 + default: 512 + - name: frequency_penalty + use_template: frequency_penalty + min: -2 + max: 2 + default: 0 + - name: presence_penalty + use_template: presence_penalty + min: -2 + max: 2 + default: 0 +pricing: + input: '0.0038' + output: '0.004' + unit: '0.0001' + currency: USD diff --git a/api/core/model_runtime/model_providers/novita/llm/wizardlm-2-8x22b.yaml b/api/core/model_runtime/model_providers/novita/llm/wizardlm-2-8x22b.yaml index ef42568e8f627c..6da4c7eca0e9c9 100644 --- a/api/core/model_runtime/model_providers/novita/llm/wizardlm-2-8x22b.yaml +++ b/api/core/model_runtime/model_providers/novita/llm/wizardlm-2-8x22b.yaml @@ -35,7 +35,7 @@ parameter_rules: max: 2 default: 0 pricing: - input: '0.0064' - output: '0.0064' + input: '0.0062' + output: '0.0062' unit: '0.0001' currency: USD diff --git a/api/core/model_runtime/model_providers/novita/novita.yaml b/api/core/model_runtime/model_providers/novita/novita.yaml index f6341979891e65..b90d64c951a355 100644 --- a/api/core/model_runtime/model_providers/novita/novita.yaml +++ b/api/core/model_runtime/model_providers/novita/novita.yaml @@ -1,6 +1,6 @@ provider: novita label: - en_US: novita.ai + en_US: Novita AI description: en_US: An LLM API that matches various application scenarios with high cost-effectiveness. zh_Hans: 适配多种海外应用场景的高性价比 LLM API @@ -11,10 +11,10 @@ icon_large: background: "#eadeff" help: title: - en_US: Get your API key from novita.ai - zh_Hans: 从 novita.ai 获取 API Key + en_US: Get your API key from Novita AI + zh_Hans: 从 Novita AI 获取 API Key url: - en_US: https://novita.ai/settings#key-management?utm_source=dify&utm_medium=ch&utm_campaign=api + en_US: https://novita.ai/settings/key-management?utm_source=dify&utm_medium=ch&utm_campaign=api supported_model_types: - llm configurate_methods: