Skip to content

Commit

Permalink
chore: update version & resource (#176)
Browse files Browse the repository at this point in the history
* chore: update version & resource

* chore: fix resource fields
  • Loading branch information
danielhjz authored Jan 5, 2024
1 parent 5fe3c11 commit da8871e
Show file tree
Hide file tree
Showing 3 changed files with 59 additions and 3 deletions.
2 changes: 1 addition & 1 deletion pyproject.toml
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
[tool.poetry]
name = "qianfan"
version = "0.2.5"
version = "0.2.6"
description = "文心千帆大模型平台 Python SDK"
authors = []
license = "Apache-2.0"
Expand Down
30 changes: 28 additions & 2 deletions src/qianfan/resources/llm/chat_completion.py
Original file line number Diff line number Diff line change
Expand Up @@ -57,8 +57,6 @@ def _supported_models(cls) -> Dict[str, QfLLMInfo]:
"tools",
"tool_choice",
"system",
"tools",
"tool_choice",
},
),
"ERNIE-Bot": QfLLMInfo(
Expand Down Expand Up @@ -137,6 +135,8 @@ def _supported_models(cls) -> Dict[str, QfLLMInfo]:
"penalty_score",
"system",
"user_id",
"tools",
"tool_choice",
},
),
"EB-turbo-AppBuilder": QfLLMInfo(
Expand All @@ -149,6 +149,8 @@ def _supported_models(cls) -> Dict[str, QfLLMInfo]:
"penalty_score",
"system",
"user_id",
"tools",
"tool_choice",
},
),
"BLOOMZ-7B": QfLLMInfo(
Expand All @@ -162,6 +164,8 @@ def _supported_models(cls) -> Dict[str, QfLLMInfo]:
"top_p",
"penalty_score",
"stop",
"tools",
"tool_choice",
},
),
"Llama-2-7b-chat": QfLLMInfo(
Expand All @@ -175,6 +179,8 @@ def _supported_models(cls) -> Dict[str, QfLLMInfo]:
"top_p",
"penalty_score",
"stop",
"tools",
"tool_choice",
},
),
"Llama-2-13b-chat": QfLLMInfo(
Expand All @@ -188,6 +194,8 @@ def _supported_models(cls) -> Dict[str, QfLLMInfo]:
"top_p",
"penalty_score",
"stop",
"tools",
"tool_choice",
},
),
"Llama-2-70b-chat": QfLLMInfo(
Expand All @@ -201,6 +209,8 @@ def _supported_models(cls) -> Dict[str, QfLLMInfo]:
"top_p",
"penalty_score",
"stop",
"tools",
"tool_choice",
},
),
"Qianfan-BLOOMZ-7B-compressed": QfLLMInfo(
Expand All @@ -214,6 +224,8 @@ def _supported_models(cls) -> Dict[str, QfLLMInfo]:
"top_p",
"penalty_score",
"stop",
"tools",
"tool_choice",
},
),
"Qianfan-Chinese-Llama-2-7B": QfLLMInfo(
Expand All @@ -227,6 +239,8 @@ def _supported_models(cls) -> Dict[str, QfLLMInfo]:
"top_p",
"penalty_score",
"stop",
"tools",
"tool_choice",
},
),
"ChatGLM2-6B-32K": QfLLMInfo(
Expand All @@ -240,6 +254,8 @@ def _supported_models(cls) -> Dict[str, QfLLMInfo]:
"top_p",
"penalty_score",
"stop",
"tools",
"tool_choice",
},
),
"AquilaChat-7B": QfLLMInfo(
Expand All @@ -253,6 +269,8 @@ def _supported_models(cls) -> Dict[str, QfLLMInfo]:
"top_p",
"penalty_score",
"stop",
"tools",
"tool_choice",
},
),
"XuanYuan-70B-Chat-4bit": QfLLMInfo(
Expand All @@ -266,6 +284,8 @@ def _supported_models(cls) -> Dict[str, QfLLMInfo]:
"top_p",
"penalty_score",
"stop",
"tools",
"tool_choice",
},
),
"Qianfan-Chinese-Llama-2-13B": QfLLMInfo(
Expand All @@ -279,6 +299,8 @@ def _supported_models(cls) -> Dict[str, QfLLMInfo]:
"top_p",
"penalty_score",
"stop",
"tools",
"tool_choice",
},
),
"ChatLaw": QfLLMInfo(
Expand All @@ -289,6 +311,8 @@ def _supported_models(cls) -> Dict[str, QfLLMInfo]:
"user_id",
"temperature",
"top_p",
"tools",
"tool_choice",
},
),
"Yi-34B-Chat": QfLLMInfo(
Expand All @@ -302,6 +326,8 @@ def _supported_models(cls) -> Dict[str, QfLLMInfo]:
"top_p",
"penalty_score",
"stop",
"tools",
"tool_choice",
},
),
UNSPECIFIED_MODEL: QfLLMInfo(
Expand Down
30 changes: 30 additions & 0 deletions src/qianfan/resources/llm/completion.py
Original file line number Diff line number Diff line change
Expand Up @@ -137,6 +137,8 @@ def _supported_models(cls) -> Dict[str, QfLLMInfo]:
"penalty_score",
"system",
"user_id",
"tools",
"tool_choice",
},
),
"BLOOMZ-7B": QfLLMInfo(
Expand All @@ -150,6 +152,8 @@ def _supported_models(cls) -> Dict[str, QfLLMInfo]:
"top_p",
"penalty_score",
"stop",
"tools",
"tool_choice",
},
),
"Llama-2-7b-chat": QfLLMInfo(
Expand All @@ -163,6 +167,8 @@ def _supported_models(cls) -> Dict[str, QfLLMInfo]:
"top_p",
"penalty_score",
"stop",
"tools",
"tool_choice",
},
),
"Llama-2-13b-chat": QfLLMInfo(
Expand All @@ -176,6 +182,8 @@ def _supported_models(cls) -> Dict[str, QfLLMInfo]:
"top_p",
"penalty_score",
"stop",
"tools",
"tool_choice",
},
),
"Llama-2-70b-chat": QfLLMInfo(
Expand All @@ -189,6 +197,8 @@ def _supported_models(cls) -> Dict[str, QfLLMInfo]:
"top_p",
"penalty_score",
"stop",
"tools",
"tool_choice",
},
),
"Qianfan-BLOOMZ-7B-compressed": QfLLMInfo(
Expand All @@ -202,6 +212,8 @@ def _supported_models(cls) -> Dict[str, QfLLMInfo]:
"top_p",
"penalty_score",
"stop",
"tools",
"tool_choice",
},
),
"Qianfan-Chinese-Llama-2-7B": QfLLMInfo(
Expand All @@ -215,6 +227,8 @@ def _supported_models(cls) -> Dict[str, QfLLMInfo]:
"top_p",
"penalty_score",
"stop",
"tools",
"tool_choice",
},
),
"ChatGLM2-6B-32K": QfLLMInfo(
Expand All @@ -228,6 +242,8 @@ def _supported_models(cls) -> Dict[str, QfLLMInfo]:
"top_p",
"penalty_score",
"stop",
"tools",
"tool_choice",
},
),
"AquilaChat-7B": QfLLMInfo(
Expand All @@ -241,6 +257,8 @@ def _supported_models(cls) -> Dict[str, QfLLMInfo]:
"top_p",
"penalty_score",
"stop",
"tools",
"tool_choice",
},
),
"XuanYuan-70B-Chat-4bit": QfLLMInfo(
Expand All @@ -254,6 +272,8 @@ def _supported_models(cls) -> Dict[str, QfLLMInfo]:
"top_p",
"penalty_score",
"stop",
"tools",
"tool_choice",
},
),
"Qianfan-Chinese-Llama-2-13B": QfLLMInfo(
Expand All @@ -267,6 +287,8 @@ def _supported_models(cls) -> Dict[str, QfLLMInfo]:
"top_p",
"penalty_score",
"stop",
"tools",
"tool_choice",
},
),
"ChatLaw": QfLLMInfo(
Expand All @@ -277,6 +299,8 @@ def _supported_models(cls) -> Dict[str, QfLLMInfo]:
"user_id",
"temperature",
"top_p",
"tools",
"tool_choice",
},
),
"SQLCoder-7B": QfLLMInfo(
Expand All @@ -290,6 +314,8 @@ def _supported_models(cls) -> Dict[str, QfLLMInfo]:
"top_p",
"penalty_score",
"stop",
"tools",
"tool_choice",
},
),
"CodeLlama-7b-Instruct": QfLLMInfo(
Expand All @@ -303,6 +329,8 @@ def _supported_models(cls) -> Dict[str, QfLLMInfo]:
"top_p",
"penalty_score",
"stop",
"tools",
"tool_choice",
},
),
"Yi-34B-Chat": QfLLMInfo(
Expand All @@ -316,6 +344,8 @@ def _supported_models(cls) -> Dict[str, QfLLMInfo]:
"top_p",
"penalty_score",
"stop",
"tools",
"tool_choice",
},
),
UNSPECIFIED_MODEL: QfLLMInfo(
Expand Down

0 comments on commit da8871e

Please sign in to comment.