diff --git a/core/config/yaml/loadYaml.ts b/core/config/yaml/loadYaml.ts index d81e3ddbc7..5e0161ad80 100644 --- a/core/config/yaml/loadYaml.ts +++ b/core/config/yaml/loadYaml.ts @@ -147,6 +147,7 @@ async function configYamlToContinueConfig( ideSettings, writeLog, platformConfigMetadata, + continueConfig.systemMessage, ); continueConfig.models.push(...llms); } @@ -160,6 +161,7 @@ async function configYamlToContinueConfig( ideSettings, writeLog, platformConfigMetadata, + continueConfig.systemMessage, ); continueConfig.tabAutocompleteModels?.push(...llms); } diff --git a/core/config/yaml/models.ts b/core/config/yaml/models.ts index f7bd717a15..309e061d19 100644 --- a/core/config/yaml/models.ts +++ b/core/config/yaml/models.ts @@ -39,6 +39,7 @@ async function modelConfigToBaseLLM( ideSettings: IdeSettings, writeLog: (log: string) => Promise, platformConfigMetadata: PlatformConfigMetadata | undefined, + systemMessage: string | undefined, ): Promise { const cls = getModelClass(model, platformConfigMetadata); @@ -68,6 +69,7 @@ async function modelConfigToBaseLLM( uniqueId, title: model.name, model: modelName, + systemMessage, }; const llm = new cls(options); @@ -82,6 +84,7 @@ async function autodetectModels( ideSettings: IdeSettings, writeLog: (log: string) => Promise, platformConfigMetadata: PlatformConfigMetadata | undefined, + systemMessage: string | undefined, ): Promise { try { const modelNames = await llm.listModels(); @@ -102,6 +105,7 @@ async function autodetectModels( ideSettings, writeLog, platformConfigMetadata, + systemMessage, ); }), ); @@ -119,6 +123,7 @@ export async function llmsFromModelConfig( ideSettings: IdeSettings, writeLog: (log: string) => Promise, platformConfigMetadata: PlatformConfigMetadata | undefined, + systemMessage: string | undefined, ): Promise { const baseLlm = await modelConfigToBaseLLM( model, @@ -126,6 +131,7 @@ export async function llmsFromModelConfig( ideSettings, writeLog, platformConfigMetadata, + systemMessage, ); if (!baseLlm) { return []; @@ -140,6 +146,7 @@ export async function llmsFromModelConfig( ideSettings, writeLog, platformConfigMetadata, + systemMessage, ); return models; } else {