|
| 1 | +import { describe, it, expect, vi, beforeEach, afterEach } from "vitest" |
| 2 | +import { getModels, flushModels } from "../modelCache" |
| 3 | +import * as customModels from "../../../../services/custom-models" |
| 4 | +import * as openrouter from "../openrouter" |
| 5 | + |
| 6 | +// Mock file data storage |
| 7 | +const mockReadFileData: Record<string, any> = {} |
| 8 | + |
| 9 | +// Mock the custom models service |
| 10 | +vi.mock("../../../../services/custom-models", () => ({ |
| 11 | + getCustomModelsForProvider: vi.fn(), |
| 12 | +})) |
| 13 | + |
| 14 | +// Mock the openrouter fetcher |
| 15 | +vi.mock("../openrouter", () => ({ |
| 16 | + getOpenRouterModels: vi.fn(), |
| 17 | +})) |
| 18 | + |
| 19 | +// Mock other dependencies |
| 20 | +vi.mock("../../../../utils/path", () => ({ |
| 21 | + getWorkspacePath: vi.fn(() => "/test/workspace"), |
| 22 | +})) |
| 23 | + |
| 24 | +vi.mock("../../../../core/config/ContextProxy", () => ({ |
| 25 | + ContextProxy: { |
| 26 | + instance: { |
| 27 | + globalStorageUri: { |
| 28 | + fsPath: "/test/storage", |
| 29 | + }, |
| 30 | + }, |
| 31 | + }, |
| 32 | +})) |
| 33 | + |
| 34 | +vi.mock("../../../../utils/storage", () => ({ |
| 35 | + getCacheDirectoryPath: vi.fn(() => "/test/cache"), |
| 36 | +})) |
| 37 | + |
| 38 | +// Mock safeWriteJson to populate our mock file data |
| 39 | +vi.mock("../../../../utils/safeWriteJson", () => ({ |
| 40 | + safeWriteJson: vi.fn((filePath: string, data: any) => { |
| 41 | + mockReadFileData[filePath] = data |
| 42 | + return Promise.resolve() |
| 43 | + }), |
| 44 | +})) |
| 45 | + |
| 46 | +// Mock fs.readFile to return the models that were written |
| 47 | +vi.mock("fs/promises", () => ({ |
| 48 | + default: { |
| 49 | + readFile: vi.fn((filePath: string) => { |
| 50 | + const data = mockReadFileData[filePath] |
| 51 | + if (!data) throw new Error("File not found") |
| 52 | + return Promise.resolve(JSON.stringify(data)) |
| 53 | + }), |
| 54 | + }, |
| 55 | + readFile: vi.fn((filePath: string) => { |
| 56 | + const data = mockReadFileData[filePath] |
| 57 | + if (!data) throw new Error("File not found") |
| 58 | + return Promise.resolve(JSON.stringify(data)) |
| 59 | + }), |
| 60 | +})) |
| 61 | + |
| 62 | +vi.mock("../../../../utils/fs", () => ({ |
| 63 | + fileExistsAtPath: vi.fn((filePath: string) => { |
| 64 | + return Promise.resolve(filePath in mockReadFileData) |
| 65 | + }), |
| 66 | +})) |
| 67 | + |
| 68 | +describe("Model Cache with Custom Models", () => { |
| 69 | + beforeEach(async () => { |
| 70 | + vi.clearAllMocks() |
| 71 | + // Clear both memory cache and mock file cache before each test |
| 72 | + await flushModels("openrouter") |
| 73 | + // Clear the mock file cache |
| 74 | + Object.keys(mockReadFileData).forEach((key) => delete mockReadFileData[key]) |
| 75 | + }) |
| 76 | + |
| 77 | + afterEach(() => { |
| 78 | + vi.restoreAllMocks() |
| 79 | + }) |
| 80 | + |
| 81 | + it("should merge custom models with provider-fetched models", async () => { |
| 82 | + const providerModels = { |
| 83 | + "openai/gpt-4": { |
| 84 | + maxTokens: 8000, |
| 85 | + contextWindow: 128000, |
| 86 | + supportsImages: true, |
| 87 | + supportsPromptCache: false, |
| 88 | + }, |
| 89 | + } |
| 90 | + |
| 91 | + const customModelDefs = { |
| 92 | + "custom/my-model": { |
| 93 | + maxTokens: 4096, |
| 94 | + contextWindow: 32000, |
| 95 | + supportsPromptCache: false, |
| 96 | + description: "My custom model", |
| 97 | + }, |
| 98 | + } |
| 99 | + |
| 100 | + vi.mocked(openrouter.getOpenRouterModels).mockResolvedValueOnce(providerModels) |
| 101 | + vi.mocked(customModels.getCustomModelsForProvider).mockResolvedValueOnce(customModelDefs) |
| 102 | + |
| 103 | + const result = await getModels({ provider: "openrouter" }) |
| 104 | + |
| 105 | + expect(result).toEqual({ |
| 106 | + ...providerModels, |
| 107 | + ...customModelDefs, |
| 108 | + }) |
| 109 | + expect(openrouter.getOpenRouterModels).toHaveBeenCalledTimes(1) |
| 110 | + expect(customModels.getCustomModelsForProvider).toHaveBeenCalledWith("openrouter", "/test/workspace") |
| 111 | + }) |
| 112 | + |
| 113 | + it("should allow custom models to override provider models", async () => { |
| 114 | + const providerModels = { |
| 115 | + "openai/gpt-4": { |
| 116 | + maxTokens: 8000, |
| 117 | + contextWindow: 128000, |
| 118 | + supportsImages: true, |
| 119 | + supportsPromptCache: false, |
| 120 | + }, |
| 121 | + } |
| 122 | + |
| 123 | + const customModelDefs = { |
| 124 | + "openai/gpt-4": { |
| 125 | + maxTokens: 16000, // Override max tokens |
| 126 | + contextWindow: 128000, |
| 127 | + supportsImages: true, |
| 128 | + supportsPromptCache: false, |
| 129 | + description: "Custom GPT-4 with higher token limit", |
| 130 | + }, |
| 131 | + } |
| 132 | + |
| 133 | + vi.mocked(openrouter.getOpenRouterModels).mockResolvedValueOnce(providerModels) |
| 134 | + vi.mocked(customModels.getCustomModelsForProvider).mockResolvedValueOnce(customModelDefs) |
| 135 | + |
| 136 | + const result = await getModels({ provider: "openrouter" }) |
| 137 | + |
| 138 | + expect(result["openai/gpt-4"]).toEqual(customModelDefs["openai/gpt-4"]) |
| 139 | + expect(result["openai/gpt-4"].maxTokens).toBe(16000) |
| 140 | + }) |
| 141 | + |
| 142 | + it("should handle empty custom models gracefully", async () => { |
| 143 | + const providerModels = { |
| 144 | + "openai/gpt-4": { |
| 145 | + maxTokens: 8000, |
| 146 | + contextWindow: 128000, |
| 147 | + supportsPromptCache: false, |
| 148 | + }, |
| 149 | + } |
| 150 | + |
| 151 | + vi.mocked(openrouter.getOpenRouterModels).mockResolvedValueOnce(providerModels) |
| 152 | + vi.mocked(customModels.getCustomModelsForProvider).mockResolvedValueOnce({}) |
| 153 | + |
| 154 | + const result = await getModels({ provider: "openrouter" }) |
| 155 | + |
| 156 | + expect(result).toEqual(providerModels) |
| 157 | + }) |
| 158 | + |
| 159 | + it("should work when provider returns no models", async () => { |
| 160 | + const customModelDefs = { |
| 161 | + "custom/model-1": { |
| 162 | + maxTokens: 4096, |
| 163 | + contextWindow: 32000, |
| 164 | + supportsPromptCache: false, |
| 165 | + }, |
| 166 | + } |
| 167 | + |
| 168 | + vi.mocked(openrouter.getOpenRouterModels).mockResolvedValueOnce({}) |
| 169 | + vi.mocked(customModels.getCustomModelsForProvider).mockResolvedValueOnce(customModelDefs) |
| 170 | + |
| 171 | + const result = await getModels({ provider: "openrouter" }) |
| 172 | + |
| 173 | + expect(result).toEqual(customModelDefs) |
| 174 | + }) |
| 175 | + |
| 176 | + it("should handle errors in custom models loading gracefully", async () => { |
| 177 | + const providerModels = { |
| 178 | + "openai/gpt-4": { |
| 179 | + maxTokens: 8000, |
| 180 | + contextWindow: 128000, |
| 181 | + supportsPromptCache: false, |
| 182 | + }, |
| 183 | + } |
| 184 | + |
| 185 | + vi.mocked(openrouter.getOpenRouterModels).mockResolvedValueOnce(providerModels) |
| 186 | + vi.mocked(customModels.getCustomModelsForProvider).mockRejectedValueOnce( |
| 187 | + new Error("Failed to load custom models"), |
| 188 | + ) |
| 189 | + |
| 190 | + // The error in loading custom models should cause the overall fetch to fail |
| 191 | + await expect(getModels({ provider: "openrouter" })).rejects.toThrow("Failed to load custom models") |
| 192 | + }) |
| 193 | + |
| 194 | + it("should flush cache for specific provider", async () => { |
| 195 | + const providerModels = { |
| 196 | + "openai/gpt-4": { |
| 197 | + maxTokens: 8000, |
| 198 | + contextWindow: 128000, |
| 199 | + supportsPromptCache: false, |
| 200 | + }, |
| 201 | + } |
| 202 | + |
| 203 | + // First call - should fetch |
| 204 | + vi.mocked(openrouter.getOpenRouterModels).mockResolvedValueOnce(providerModels) |
| 205 | + vi.mocked(customModels.getCustomModelsForProvider).mockResolvedValueOnce({}) |
| 206 | + await getModels({ provider: "openrouter" }) |
| 207 | + expect(openrouter.getOpenRouterModels).toHaveBeenCalledTimes(1) |
| 208 | + |
| 209 | + // Second call - should use cache (no new mocks needed) |
| 210 | + await getModels({ provider: "openrouter" }) |
| 211 | + expect(openrouter.getOpenRouterModels).toHaveBeenCalledTimes(1) |
| 212 | + |
| 213 | + // Flush cache |
| 214 | + await flushModels("openrouter") |
| 215 | + |
| 216 | + // Third call - should fetch again (set up mock again) |
| 217 | + vi.mocked(openrouter.getOpenRouterModels).mockResolvedValueOnce(providerModels) |
| 218 | + vi.mocked(customModels.getCustomModelsForProvider).mockResolvedValueOnce({}) |
| 219 | + await getModels({ provider: "openrouter" }) |
| 220 | + expect(openrouter.getOpenRouterModels).toHaveBeenCalledTimes(2) |
| 221 | + }) |
| 222 | +}) |
0 commit comments