chore: adjust the evaluation result statement

main
jialin 2 months ago
parent 3c7dc5a49e
commit 351b2e1417

@ -133,9 +133,11 @@ export default {
'models.form.check.params': 'Checking configuration...',
'models.form.check.passed': 'Compatibility Check Passed',
'models.form.check.claims':
'The model requires approximately {vram} VRAM and {ram} RAM.',
'models.form.check.claims2': 'The model requires approximately {vram} VRAM.',
'models.form.check.claims3': 'The model requires approximately {ram} RAM.',
'The model will consume approximately {vram} VRAM and {ram} RAM.',
'models.form.check.claims2':
'The model will consume approximately {vram} VRAM.',
'models.form.check.claims3':
'The model will consume approximately {ram} RAM.',
'models.form.update.tips':
'Changes will only apply after you delete and recreate the instance.',
'models.table.download.progress': 'Download Progress',

@ -126,9 +126,9 @@ export default {
'models.form.restart.onerror.tips': '当发生错误时,将自动尝试恢复',
'models.form.check.params': '正在校验配置...',
'models.form.check.passed': '兼容性检查通过',
'models.form.check.claims': '该模型大约需要分配 {vram} 显存和 {ram} 内存。',
'models.form.check.claims2': '该模型大约需要分配 {vram} 显存。',
'models.form.check.claims3': '该模型大约需要分配 {ram} 内存。',
'models.form.check.claims': '该模型大约需要消耗 {vram} 显存和 {ram} 内存。',
'models.form.check.claims2': '该模型大约需要消耗 {vram} 显存。',
'models.form.check.claims3': '该模型大约需要消耗 {ram} 内存。',
'models.form.update.tips': '更改仅在删除并重新创建实例后生效。',
'models.table.download.progress': '下载进度',
'models.table.button.apiAccessInfo': 'API 接入信息',

@ -29,7 +29,6 @@ import CompatibilityAlert from './compatible-alert';
import DataForm from './data-form';
import HFModelFile from './hf-model-file';
import ModelCard from './model-card';
import OllamaTips from './ollama-tips';
import SearchModel from './search-model';
import Separator from './separator';
import TitleWrapper from './title-wrapper';
@ -458,11 +457,7 @@ const AddModal: FC<AddModalProps> = (props) => {
? backendOptionsMap.ascendMindie
: backendOptionsMap.vllm;
if (source === modelSourceMap.ollama_library_value) {
backend = backendOptionsMap.llamaBox;
}
form.current?.setFieldValue?.('backend', backend);
setIsGGUF(source === modelSourceMap.ollama_library_value);
}
};
@ -567,7 +562,6 @@ const AddModal: FC<AddModalProps> = (props) => {
selectedModel={selectedModel}
modelSource={props.source}
onSelectFile={handleSelectModelFile}
updateEvaluteState={updateEvaluateState}
collapsed={collapsed}
gpuOptions={props.gpuOptions}
></HFModelFile>
@ -621,9 +615,6 @@ const AddModal: FC<AddModalProps> = (props) => {
}
>
<>
{source === modelSourceMap.ollama_library_value && (
<OllamaTips></OllamaTips>
)}
{SEARCH_SOURCE.includes(source) &&
deploymentType === 'modelList' && (
<TitleWrapper>

Loading…
Cancel
Save