@@ -67,7 +67,7 @@ describe('LLM Model Selection with Special Characters', () => {
6767 vi . mocked ( options . getOptionBool ) . mockReturnValue ( true ) ;
6868 } ) ;
6969
70- describe ( 'OpenAI Model Names' , ( ) => {
70+ describe . skip ( 'OpenAI Model Names' , ( ) => {
7171 it ( 'should correctly handle model names with periods' , async ( ) => {
7272 const modelName = 'gpt-4.1-turbo-preview' ;
7373 vi . mocked ( options . getOption ) . mockImplementation ( ( key : string ) => {
@@ -82,7 +82,7 @@ describe('LLM Model Selection with Special Characters', () => {
8282
8383 // Spy on getOpenAIOptions to verify model name is passed correctly
8484 const getOpenAIOptionsSpy = vi . spyOn ( providers , 'getOpenAIOptions' ) ;
85-
85+
8686 try {
8787 await service . generateChatCompletion ( [ { role : 'user' , content : 'test' } ] , opts ) ;
8888 } catch ( error ) {
@@ -108,7 +108,7 @@ describe('LLM Model Selection with Special Characters', () => {
108108 } ;
109109
110110 const getOpenAIOptionsSpy = vi . spyOn ( providers , 'getOpenAIOptions' ) ;
111-
111+
112112 try {
113113 await service . generateChatCompletion ( [ { role : 'user' , content : 'test' } ] , opts ) ;
114114 } catch ( error ) {
@@ -127,7 +127,7 @@ describe('LLM Model Selection with Special Characters', () => {
127127 } ;
128128
129129 const getOpenAIOptionsSpy = vi . spyOn ( providers , 'getOpenAIOptions' ) ;
130-
130+
131131 const openaiOptions = providers . getOpenAIOptions ( opts ) ;
132132 expect ( openaiOptions . model ) . toBe ( modelName ) ;
133133 } ) ;
@@ -153,7 +153,7 @@ describe('LLM Model Selection with Special Characters', () => {
153153 } ) ;
154154
155155 const service = new OpenAIService ( ) ;
156-
156+
157157 // Access the private openai client through the service
158158 const client = ( service as any ) . getClient ( 'test-key' ) ;
159159 const createSpy = vi . spyOn ( client . chat . completions , 'create' ) ;
@@ -213,7 +213,7 @@ describe('LLM Model Selection with Special Characters', () => {
213213 } ) ;
214214
215215 const service = new AnthropicService ( ) ;
216-
216+
217217 // Access the private anthropic client
218218 const client = ( service as any ) . getClient ( 'test-key' ) ;
219219 const createSpy = vi . spyOn ( client . messages , 'create' ) ;
@@ -278,7 +278,7 @@ describe('LLM Model Selection with Special Characters', () => {
278278
279279 const ollamaOptions = await providers . getOllamaOptions ( opts ) ;
280280 expect ( ollamaOptions . model ) . toBe ( modelName ) ;
281-
281+
282282 // Also test with model specified in options
283283 const optsWithModel : ChatCompletionOptions = {
284284 model : 'another/model:v2.0@beta' ,
@@ -370,7 +370,7 @@ describe('LLM Model Selection with Special Characters', () => {
370370 describe ( 'Integration with REST API' , ( ) => {
371371 it ( 'should pass model names correctly through REST chat service' , async ( ) => {
372372 const modelName = 'gpt-4.1-turbo-preview@latest' ;
373-
373+
374374 // Mock the configuration helpers
375375 vi . doMock ( '../config/configuration_helpers.js' , ( ) => ( {
376376 getSelectedModelConfig : vi . fn ( ) . mockResolvedValue ( {
@@ -382,8 +382,8 @@ describe('LLM Model Selection with Special Characters', () => {
382382
383383 const { getSelectedModelConfig } = await import ( '../config/configuration_helpers.js' ) ;
384384 const config = await getSelectedModelConfig ( ) ;
385-
385+
386386 expect ( config ?. model ) . toBe ( modelName ) ;
387387 } ) ;
388388 } ) ;
389- } ) ;
389+ } ) ;
0 commit comments