refactor(models): enhance model management with official provider marking and extended metadata

- Add OFFICIAL_PROVIDERS set to mark first-party vendors in models.dev
- Implement official provider marking function with cache compatibility
- Extend model metadata with family, context_limit, output_limit fields
- Improve frontend model selection UI with wider panel and better search
- Add dark mode support for provider logos
- Optimize scrollbar styling for model lists
- Update deployment documentation with clearer migration steps
This commit is contained in:
fawney19
2025-12-16 17:28:40 +08:00
parent edce43d45f
commit 46ff5a1a50
8 changed files with 282 additions and 64 deletions

View File

@@ -60,11 +60,11 @@ python generate_keys.py # 生成密钥, 并将生成的密钥填入 .env
# 3. 部署 # 3. 部署
docker-compose up -d docker-compose up -d
# 4. 更新 # 4. 首次部署时, 初始化数据库
docker-compose pull && docker-compose up -d
# 5. 数据库迁移 - 更新后执行
./migrate.sh ./migrate.sh
# 5. 更新
docker-compose pull && docker-compose up -d && ./migrate.sh
``` ```
### Docker Compose本地构建镜像 ### Docker Compose本地构建镜像

View File

@@ -50,6 +50,7 @@ export interface ModelsDevProvider {
name: string name: string
doc?: string doc?: string
models: Record<string, ModelsDevModel> models: Record<string, ModelsDevModel>
official?: boolean // 是否为官方提供商
} }
export type ModelsDevData = Record<string, ModelsDevProvider> export type ModelsDevData = Record<string, ModelsDevProvider>
@@ -68,7 +69,12 @@ export interface ModelsDevModelItem {
supportsVision?: boolean supportsVision?: boolean
supportsToolCall?: boolean supportsToolCall?: boolean
supportsReasoning?: boolean supportsReasoning?: boolean
supportsStructuredOutput?: boolean
supportsTemperature?: boolean
supportsAttachment?: boolean
openWeights?: boolean
deprecated?: boolean deprecated?: boolean
official?: boolean // 是否来自官方提供商
// 用于 display_metadata 的额外字段 // 用于 display_metadata 的额外字段
knowledgeCutoff?: string knowledgeCutoff?: string
releaseDate?: string releaseDate?: string
@@ -84,13 +90,21 @@ interface CacheData {
// 内存缓存 // 内存缓存
let memoryCache: CacheData | null = null let memoryCache: CacheData | null = null
function hasOfficialFlag(data: ModelsDevData): boolean {
return Object.values(data).some(provider => typeof provider?.official === 'boolean')
}
/** /**
* 获取 models.dev 数据(带缓存) * 获取 models.dev 数据(带缓存)
*/ */
export async function getModelsDevData(): Promise<ModelsDevData> { export async function getModelsDevData(): Promise<ModelsDevData> {
// 1. 检查内存缓存 // 1. 检查内存缓存
if (memoryCache && Date.now() - memoryCache.timestamp < CACHE_DURATION) { if (memoryCache && Date.now() - memoryCache.timestamp < CACHE_DURATION) {
return memoryCache.data // 兼容旧缓存:没有 official 字段时丢弃,强制刷新一次
if (hasOfficialFlag(memoryCache.data)) {
return memoryCache.data
}
memoryCache = null
} }
// 2. 检查 localStorage 缓存 // 2. 检查 localStorage 缓存
@@ -99,8 +113,12 @@ export async function getModelsDevData(): Promise<ModelsDevData> {
if (cached) { if (cached) {
const cacheData: CacheData = JSON.parse(cached) const cacheData: CacheData = JSON.parse(cached)
if (Date.now() - cacheData.timestamp < CACHE_DURATION) { if (Date.now() - cacheData.timestamp < CACHE_DURATION) {
memoryCache = cacheData // 兼容旧缓存:没有 official 字段时丢弃,强制刷新一次
return cacheData.data if (hasOfficialFlag(cacheData.data)) {
memoryCache = cacheData
return cacheData.data
}
localStorage.removeItem(CACHE_KEY)
} }
} }
} catch { } catch {
@@ -126,52 +144,75 @@ export async function getModelsDevData(): Promise<ModelsDevData> {
return data return data
} }
// 模型列表缓存(避免重复转换)
let modelsListCache: ModelsDevModelItem[] | null = null
let modelsListCacheTimestamp: number | null = null
/** /**
* 获取扁平化的模型列表 * 获取扁平化的模型列表
* 数据只加载一次,通过参数过滤官方/全部
*/ */
export async function getModelsDevList(): Promise<ModelsDevModelItem[]> { export async function getModelsDevList(officialOnly: boolean = true): Promise<ModelsDevModelItem[]> {
const data = await getModelsDevData() const data = await getModelsDevData()
const items: ModelsDevModelItem[] = [] const currentTimestamp = memoryCache?.timestamp ?? 0
for (const [providerId, provider] of Object.entries(data)) { // 如果缓存为空或数据已刷新,构建一次
if (!provider.models) continue if (!modelsListCache || modelsListCacheTimestamp !== currentTimestamp) {
const items: ModelsDevModelItem[] = []
for (const [modelId, model] of Object.entries(provider.models)) { for (const [providerId, provider] of Object.entries(data)) {
items.push({ if (!provider.models) continue
providerId,
providerName: provider.name, for (const [modelId, model] of Object.entries(provider.models)) {
modelId, items.push({
modelName: model.name || modelId, providerId,
family: model.family, providerName: provider.name,
inputPrice: model.cost?.input, modelId,
outputPrice: model.cost?.output, modelName: model.name || modelId,
contextLimit: model.limit?.context, family: model.family,
outputLimit: model.limit?.output, inputPrice: model.cost?.input,
supportsVision: model.input?.includes('image'), outputPrice: model.cost?.output,
supportsToolCall: model.tool_call, contextLimit: model.limit?.context,
supportsReasoning: model.reasoning, outputLimit: model.limit?.output,
deprecated: model.deprecated, supportsVision: model.input?.includes('image'),
// display_metadata 相关字段 supportsToolCall: model.tool_call,
knowledgeCutoff: model.knowledge, supportsReasoning: model.reasoning,
releaseDate: model.release_date, supportsStructuredOutput: model.structured_output,
inputModalities: model.input, supportsTemperature: model.temperature,
outputModalities: model.output, supportsAttachment: model.attachment,
}) openWeights: model.open_weights,
deprecated: model.deprecated,
official: provider.official,
// display_metadata 相关字段
knowledgeCutoff: model.knowledge,
releaseDate: model.release_date,
inputModalities: model.input,
outputModalities: model.output,
})
}
} }
// 按 provider 名称和模型名称排序
items.sort((a, b) => {
const providerCompare = a.providerName.localeCompare(b.providerName)
if (providerCompare !== 0) return providerCompare
return a.modelName.localeCompare(b.modelName)
})
modelsListCache = items
modelsListCacheTimestamp = currentTimestamp
} }
// 按 provider 名称和模型名称排序 // 根据参数过滤
items.sort((a, b) => { if (officialOnly) {
const providerCompare = a.providerName.localeCompare(b.providerName) return modelsListCache.filter(m => m.official)
if (providerCompare !== 0) return providerCompare }
return a.modelName.localeCompare(b.modelName) return modelsListCache
})
return items
} }
/** /**
* 搜索模型 * 搜索模型
* 搜索时包含所有提供商(包括第三方)
*/ */
export async function searchModelsDevModels( export async function searchModelsDevModels(
query: string, query: string,
@@ -180,7 +221,8 @@ export async function searchModelsDevModels(
excludeDeprecated?: boolean excludeDeprecated?: boolean
} }
): Promise<ModelsDevModelItem[]> { ): Promise<ModelsDevModelItem[]> {
const allModels = await getModelsDevList() // 搜索时包含全部提供商
const allModels = await getModelsDevList(false)
const { limit = 50, excludeDeprecated = true } = options || {} const { limit = 50, excludeDeprecated = true } = options || {}
const queryLower = query.toLowerCase() const queryLower = query.toLowerCase()
@@ -236,6 +278,8 @@ export function getProviderLogoUrl(providerId: string): string {
*/ */
export function clearModelsDevCache(): void { export function clearModelsDevCache(): void {
memoryCache = null memoryCache = null
modelsListCache = null
modelsListCacheTimestamp = null
try { try {
localStorage.removeItem(CACHE_KEY) localStorage.removeItem(CACHE_KEY)
} catch { } catch {

View File

@@ -14,7 +14,7 @@
<!-- 左侧模型选择仅创建模式 --> <!-- 左侧模型选择仅创建模式 -->
<div <div
v-if="!isEditMode" v-if="!isEditMode"
class="w-[200px] shrink-0 flex flex-col h-full" class="w-[260px] shrink-0 flex flex-col h-full"
> >
<!-- 搜索框 --> <!-- 搜索框 -->
<div class="relative mb-3"> <div class="relative mb-3">
@@ -22,13 +22,13 @@
<Input <Input
v-model="searchQuery" v-model="searchQuery"
type="text" type="text"
placeholder="搜索模型..." placeholder="搜索模型、提供商..."
class="pl-8 h-8 text-sm" class="pl-8 h-8 text-sm"
/> />
</div> </div>
<!-- 模型列表两级结构 --> <!-- 模型列表两级结构 -->
<div class="flex-1 overflow-y-auto border rounded-lg min-h-0"> <div class="flex-1 overflow-y-auto border rounded-lg min-h-0 scrollbar-thin">
<div <div
v-if="loading" v-if="loading"
class="flex items-center justify-center h-32" class="flex items-center justify-center h-32"
@@ -54,7 +54,7 @@
<img <img
:src="getProviderLogoUrl(group.providerId)" :src="getProviderLogoUrl(group.providerId)"
:alt="group.providerName" :alt="group.providerName"
class="w-4 h-4 rounded shrink-0" class="w-4 h-4 rounded shrink-0 dark:invert dark:brightness-90"
@error="handleLogoError" @error="handleLogoError"
> >
<span class="truncate font-medium text-xs flex-1">{{ group.providerName }}</span> <span class="truncate font-medium text-xs flex-1">{{ group.providerName }}</span>
@@ -90,7 +90,7 @@
<!-- 右侧表单 --> <!-- 右侧表单 -->
<div <div
class="flex-1 overflow-y-auto h-full" class="flex-1 overflow-y-auto h-full scrollbar-thin"
:class="isEditMode ? 'max-h-[70vh]' : ''" :class="isEditMode ? 'max-h-[70vh]' : ''"
> >
<form <form
@@ -141,6 +141,46 @@
@update:model-value="(v) => setConfigField('description', v || undefined)" @update:model-value="(v) => setConfigField('description', v || undefined)"
/> />
</div> </div>
<div class="grid grid-cols-3 gap-3">
<div class="space-y-1.5">
<Label
for="model-family"
class="text-xs"
>模型系列</Label>
<Input
id="model-family"
:model-value="form.config?.family || ''"
placeholder=" GPT-4Claude 3"
@update:model-value="(v) => setConfigField('family', v || undefined)"
/>
</div>
<div class="space-y-1.5">
<Label
for="model-context-limit"
class="text-xs"
>上下文限制</Label>
<Input
id="model-context-limit"
type="number"
:model-value="form.config?.context_limit ?? ''"
placeholder=" 128000"
@update:model-value="(v) => setConfigField('context_limit', v ? Number(v) : undefined)"
/>
</div>
<div class="space-y-1.5">
<Label
for="model-output-limit"
class="text-xs"
>输出限制</Label>
<Input
id="model-output-limit"
type="number"
:model-value="form.config?.output_limit ?? ''"
placeholder=" 8192"
@update:model-value="(v) => setConfigField('output_limit', v ? Number(v) : undefined)"
/>
</div>
</div>
</section> </section>
<!-- 能力配置 --> <!-- 能力配置 -->
@@ -329,10 +369,18 @@ const tieredPricingEditorRef = ref<InstanceType<typeof TieredPricingEditor> | nu
// 模型列表相关 // 模型列表相关
const loading = ref(false) const loading = ref(false)
const searchQuery = ref('') const searchQuery = ref('')
const allModels = ref<ModelsDevModelItem[]>([]) const allModelsCache = ref<ModelsDevModelItem[]>([]) // 全部模型(缓存)
const selectedModel = ref<ModelsDevModelItem | null>(null) const selectedModel = ref<ModelsDevModelItem | null>(null)
const expandedProvider = ref<string | null>(null) const expandedProvider = ref<string | null>(null)
// 当前显示的模型列表:有搜索词时用全部,否则只用官方
const allModels = computed(() => {
if (searchQuery.value) {
return allModelsCache.value
}
return allModelsCache.value.filter(m => m.official)
})
// 按提供商分组的模型 // 按提供商分组的模型
interface ProviderGroup { interface ProviderGroup {
providerId: string providerId: string
@@ -440,10 +488,11 @@ const availableCapabilities = ref<CapabilityDefinition[]>([])
// 加载模型列表 // 加载模型列表
async function loadModels() { async function loadModels() {
if (allModels.value.length > 0) return if (allModelsCache.value.length > 0) return
loading.value = true loading.value = true
try { try {
allModels.value = await getModelsDevList() // 只加载一次全部模型,过滤在 computed 中完成
allModelsCache.value = await getModelsDevList(false)
} catch (err) { } catch (err) {
log.error('Failed to load models:', err) log.error('Failed to load models:', err)
} finally { } finally {
@@ -498,6 +547,10 @@ function selectModel(model: ModelsDevModelItem) {
if (model.supportsVision) config.vision = true if (model.supportsVision) config.vision = true
if (model.supportsToolCall) config.function_calling = true if (model.supportsToolCall) config.function_calling = true
if (model.supportsReasoning) config.extended_thinking = true if (model.supportsReasoning) config.extended_thinking = true
if (model.supportsStructuredOutput) config.structured_output = true
if (model.supportsTemperature !== false) config.temperature = model.supportsTemperature
if (model.supportsAttachment) config.attachment = true
if (model.openWeights) config.open_weights = true
if (model.contextLimit) config.context_limit = model.contextLimit if (model.contextLimit) config.context_limit = model.contextLimit
if (model.outputLimit) config.output_limit = model.outputLimit if (model.outputLimit) config.output_limit = model.outputLimit
if (model.knowledgeCutoff) config.knowledge_cutoff = model.knowledgeCutoff if (model.knowledgeCutoff) config.knowledge_cutoff = model.knowledgeCutoff

View File

@@ -1169,4 +1169,26 @@ body[theme-mode='dark'] .literary-annotation {
.scrollbar-hide::-webkit-scrollbar { .scrollbar-hide::-webkit-scrollbar {
display: none; display: none;
} }
.scrollbar-thin {
scrollbar-width: thin;
scrollbar-color: hsl(var(--border)) transparent;
}
.scrollbar-thin::-webkit-scrollbar {
width: 6px;
}
.scrollbar-thin::-webkit-scrollbar-track {
background: transparent;
}
.scrollbar-thin::-webkit-scrollbar-thumb {
background-color: hsl(var(--border));
border-radius: 3px;
}
.scrollbar-thin::-webkit-scrollbar-thumb:hover {
background-color: hsl(var(--muted-foreground) / 0.5);
}
} }

View File

@@ -20,6 +20,27 @@ router = APIRouter()
CACHE_KEY = "aether:external:models_dev" CACHE_KEY = "aether:external:models_dev"
CACHE_TTL = 15 * 60 # 15 分钟 CACHE_TTL = 15 * 60 # 15 分钟
# 标记官方/一手提供商,前端可据此过滤第三方转售商
OFFICIAL_PROVIDERS = {
"anthropic", # Claude 官方
"openai", # OpenAI 官方
"google", # Gemini 官方
"google-vertex", # Google Vertex AI
"azure", # Azure OpenAI
"amazon-bedrock", # AWS Bedrock
"xai", # Grok 官方
"meta", # Llama 官方
"deepseek", # DeepSeek 官方
"mistral", # Mistral 官方
"cohere", # Cohere 官方
"zhipuai", # 智谱 AI 官方
"alibaba", # 阿里云(通义千问)
"minimax", # MiniMax 官方
"moonshot", # 月之暗面Kimi
"baichuan", # 百川智能
"ai21", # AI21 Labs
}
async def _get_cached_data() -> Optional[dict[str, Any]]: async def _get_cached_data() -> Optional[dict[str, Any]]:
"""从 Redis 获取缓存数据""" """从 Redis 获取缓存数据"""
@@ -47,15 +68,40 @@ async def _set_cached_data(data: dict) -> None:
logger.warning(f"写入 models.dev 缓存失败: {e}") logger.warning(f"写入 models.dev 缓存失败: {e}")
def _mark_official_providers(data: dict[str, Any]) -> dict[str, Any]:
"""为每个提供商标记是否为官方"""
result = {}
for provider_id, provider_data in data.items():
result[provider_id] = {
**provider_data,
"official": provider_id in OFFICIAL_PROVIDERS,
}
return result
@router.get("/external") @router.get("/external")
async def get_external_models(_: User = Depends(require_admin)) -> JSONResponse: async def get_external_models(_: User = Depends(require_admin)) -> JSONResponse:
""" """
获取 models.dev 的模型数据(代理请求,解决跨域问题) 获取 models.dev 的模型数据(代理请求,解决跨域问题)
数据缓存 15 分钟(使用 Redis多 worker 共享) 数据缓存 15 分钟(使用 Redis多 worker 共享)
每个提供商会标记 official 字段,前端可据此过滤
""" """
# 检查缓存 # 检查缓存
cached = await _get_cached_data() cached = await _get_cached_data()
if cached is not None: if cached is not None:
# 兼容旧缓存:如果没有 official 字段则补全并回写
try:
needs_mark = False
for provider_data in cached.values():
if not isinstance(provider_data, dict) or "official" not in provider_data:
needs_mark = True
break
if needs_mark:
marked_cached = _mark_official_providers(cached)
await _set_cached_data(marked_cached)
return JSONResponse(content=marked_cached)
except Exception as e:
logger.warning(f"处理 models.dev 缓存数据失败,将直接返回原缓存: {e}")
return JSONResponse(content=cached) return JSONResponse(content=cached)
# 从 models.dev 获取数据 # 从 models.dev 获取数据
@@ -65,10 +111,13 @@ async def get_external_models(_: User = Depends(require_admin)) -> JSONResponse:
response.raise_for_status() response.raise_for_status()
data = response.json() data = response.json()
# 写入缓存 # 标记官方提供商
await _set_cached_data(data) marked_data = _mark_official_providers(data)
return JSONResponse(content=data) # 写入缓存
await _set_cached_data(marked_data)
return JSONResponse(content=marked_data)
except httpx.TimeoutException: except httpx.TimeoutException:
raise HTTPException(status_code=504, detail="请求 models.dev 超时") raise HTTPException(status_code=504, detail="请求 models.dev 超时")
except httpx.HTTPStatusError as e: except httpx.HTTPStatusError as e:
@@ -77,3 +126,16 @@ async def get_external_models(_: User = Depends(require_admin)) -> JSONResponse:
) )
except Exception as e: except Exception as e:
raise HTTPException(status_code=502, detail=f"获取外部模型数据失败: {str(e)}") raise HTTPException(status_code=502, detail=f"获取外部模型数据失败: {str(e)}")
@router.delete("/external/cache")
async def clear_external_models_cache(_: User = Depends(require_admin)) -> dict:
"""清除 models.dev 缓存"""
redis = await get_redis_client()
if redis is None:
return {"cleared": False, "message": "Redis 未启用"}
try:
await redis.delete(CACHE_KEY)
return {"cleared": True}
except Exception as e:
raise HTTPException(status_code=500, detail=f"清除缓存失败: {str(e)}")

View File

@@ -65,6 +65,21 @@ class ModelInfo:
created_at: Optional[str] # ISO 格式 created_at: Optional[str] # ISO 格式
created_timestamp: int # Unix 时间戳 created_timestamp: int # Unix 时间戳
provider_name: str provider_name: str
# 能力配置
streaming: bool = True
vision: bool = False
function_calling: bool = False
extended_thinking: bool = False
image_generation: bool = False
structured_output: bool = False
# 规格参数
context_limit: Optional[int] = None
output_limit: Optional[int] = None
# 元信息
family: Optional[str] = None
knowledge_cutoff: Optional[str] = None
input_modalities: Optional[list[str]] = None
output_modalities: Optional[list[str]] = None
def get_available_provider_ids(db: Session, api_formats: list[str]) -> set[str]: def get_available_provider_ids(db: Session, api_formats: list[str]) -> set[str]:
@@ -181,13 +196,19 @@ def _extract_model_info(model: Any) -> ModelInfo:
global_model = model.global_model global_model = model.global_model
model_id: str = global_model.name if global_model else model.provider_model_name model_id: str = global_model.name if global_model else model.provider_model_name
display_name: str = global_model.display_name if global_model else model.provider_model_name display_name: str = global_model.display_name if global_model else model.provider_model_name
description: Optional[str] = global_model.description if global_model else None
created_at: Optional[str] = ( created_at: Optional[str] = (
model.created_at.strftime("%Y-%m-%dT%H:%M:%SZ") if model.created_at else None model.created_at.strftime("%Y-%m-%dT%H:%M:%SZ") if model.created_at else None
) )
created_timestamp: int = int(model.created_at.timestamp()) if model.created_at else 0 created_timestamp: int = int(model.created_at.timestamp()) if model.created_at else 0
provider_name: str = model.provider.name if model.provider else "unknown" provider_name: str = model.provider.name if model.provider else "unknown"
# 从 GlobalModel.config 提取配置信息
config: dict = {}
description: Optional[str] = None
if global_model:
config = global_model.config or {}
description = config.get("description")
return ModelInfo( return ModelInfo(
id=model_id, id=model_id,
display_name=display_name, display_name=display_name,
@@ -195,6 +216,21 @@ def _extract_model_info(model: Any) -> ModelInfo:
created_at=created_at, created_at=created_at,
created_timestamp=created_timestamp, created_timestamp=created_timestamp,
provider_name=provider_name, provider_name=provider_name,
# 能力配置
streaming=config.get("streaming", True),
vision=config.get("vision", False),
function_calling=config.get("function_calling", False),
extended_thinking=config.get("extended_thinking", False),
image_generation=config.get("image_generation", False),
structured_output=config.get("structured_output", False),
# 规格参数
context_limit=config.get("context_limit"),
output_limit=config.get("output_limit"),
# 元信息
family=config.get("family"),
knowledge_cutoff=config.get("knowledge_cutoff"),
input_modalities=config.get("input_modalities"),
output_modalities=config.get("output_modalities"),
) )

View File

@@ -251,8 +251,8 @@ def _build_gemini_list_response(
"version": "001", "version": "001",
"displayName": m.display_name, "displayName": m.display_name,
"description": m.description or f"Model {m.id}", "description": m.description or f"Model {m.id}",
"inputTokenLimit": 128000, "inputTokenLimit": m.context_limit if m.context_limit is not None else 128000,
"outputTokenLimit": 8192, "outputTokenLimit": m.output_limit if m.output_limit is not None else 8192,
"supportedGenerationMethods": ["generateContent", "countTokens"], "supportedGenerationMethods": ["generateContent", "countTokens"],
"temperature": 1.0, "temperature": 1.0,
"maxTemperature": 2.0, "maxTemperature": 2.0,
@@ -297,8 +297,8 @@ def _build_gemini_model_response(model_info: ModelInfo) -> dict:
"version": "001", "version": "001",
"displayName": model_info.display_name, "displayName": model_info.display_name,
"description": model_info.description or f"Model {model_info.id}", "description": model_info.description or f"Model {model_info.id}",
"inputTokenLimit": 128000, "inputTokenLimit": model_info.context_limit if model_info.context_limit is not None else 128000,
"outputTokenLimit": 8192, "outputTokenLimit": model_info.output_limit if model_info.output_limit is not None else 8192,
"supportedGenerationMethods": ["generateContent", "countTokens"], "supportedGenerationMethods": ["generateContent", "countTokens"],
"temperature": 1.0, "temperature": 1.0,
"maxTemperature": 2.0, "maxTemperature": 2.0,

View File

@@ -273,16 +273,17 @@ def get_db_url() -> str:
def init_db(): def init_db():
"""初始化数据库""" """初始化数据库
注意:数据库表结构由 Alembic 管理,部署时请运行 ./migrate.sh
"""
logger.info("初始化数据库...") logger.info("初始化数据库...")
# 确保引擎已创建 # 确保引擎已创建
engine = _ensure_engine() _ensure_engine()
# 创建所有表 # 数据库表结构由 Alembic 迁移管理
Base.metadata.create_all(bind=engine) # 首次部署或更新后请运行: ./migrate.sh
# 数据库表已通过SQLAlchemy自动创建
db = _SessionLocal() db = _SessionLocal()
try: try: