Skip to content

Commit 8af0443

Browse files
committed
Remove geminiModels from constants / ai sdk
1 parent f700ff1 commit 8af0443

File tree

9 files changed

+12
-63
lines changed

9 files changed

+12
-63
lines changed

.agents/prompts/ask-prompts.ts

Lines changed: 2 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -144,10 +144,8 @@ ${PLACEHOLDER.GIT_CHANGES_PROMPT}`
144144
}
145145

146146
export const askAgentUserInputPrompt = (model: Model) => {
147-
const isFlash =
148-
model === models.gemini2_5_flash ||
149-
model === models.gemini2_5_flash_thinking
150-
const isGeminiPro = model === models.gemini2_5_pro_preview
147+
const isFlash = model === models.openrouter_gemini2_5_flash
148+
const isGeminiPro = model === models.openrouter_gemini2_5_pro_preview
151149

152150
return (
153151
PLACEHOLDER.KNOWLEDGE_FILES_CONTENTS +

.agents/prompts/base-prompts.ts

Lines changed: 2 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -229,10 +229,8 @@ export const baseAgentUserInputPrompt = (
229229
model: Model,
230230
mode: 'lite' | 'normal' | 'max' | 'experimental',
231231
) => {
232-
const isFlash =
233-
model === models.gemini2_5_flash ||
234-
model === models.gemini2_5_flash_thinking
235-
const isGeminiPro = model === models.gemini2_5_pro_preview
232+
const isFlash = model === models.openrouter_gemini2_5_flash
233+
const isGeminiPro = model === models.openrouter_gemini2_5_pro_preview
236234
const isGPT5 =
237235
model === models.openrouter_gpt5 || model === models.openrouter_gpt5_chat
238236
const isLite = mode === 'lite'

backend/src/llm-apis/message-cost-tracker.ts

Lines changed: 0 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -42,9 +42,6 @@ const TOKENS_COST_PER_M = {
4242
[models.o4mini]: 1.1,
4343
[models.deepseekChat]: 0.14,
4444
[models.deepseekReasoner]: 0.55,
45-
[models.gemini2flash]: 0.1,
46-
[models.gemini2_5_flash]: 0.15,
47-
[models.gemini2_5_flash_thinking]: 0.15,
4845
[models.ft_filepicker_003]: 0.1,
4946
[models.ft_filepicker_005]: 0.1,
5047
[models.openrouter_claude_sonnet_4]: 3,
@@ -72,9 +69,6 @@ const TOKENS_COST_PER_M = {
7269
[models.o4mini]: 1.1,
7370
[models.deepseekChat]: 0.28,
7471
[models.deepseekReasoner]: 2.19,
75-
[models.gemini2flash]: 0.4,
76-
[models.gemini2_5_flash]: 0.6,
77-
[models.gemini2_5_flash_thinking]: 3.5,
7872
[models.ft_filepicker_003]: 0.4,
7973
[models.ft_filepicker_005]: 0.4,
8074
[models.openrouter_claude_sonnet_4]: 15,
@@ -107,9 +101,6 @@ const TOKENS_COST_PER_M = {
107101
[models.o3]: 0.5,
108102
[models.o3mini]: 0.55,
109103
[models.o4mini]: 0.275,
110-
[models.gemini2flash]: 0.025,
111-
[models.gemini2_5_flash]: 0.0375,
112-
[models.gemini2_5_flash_thinking]: 0.2625,
113104
[models.ft_filepicker_003]: 0.025,
114105
[models.ft_filepicker_005]: 0.025,
115106
},
@@ -204,13 +195,6 @@ const calcCost = (
204195
if (model === 'relace-fast-apply') {
205196
return RELACE_FAST_APPLY_COST
206197
}
207-
if (model === models.gemini2_5_pro_preview) {
208-
return (
209-
getGemini25ProPreviewCost(input_tokens, output_tokens) +
210-
cache_creation_input_tokens * getPerTokenCost(model, 'cache_creation') +
211-
cache_read_input_tokens * getPerTokenCost(model, 'cache_read')
212-
)
213-
}
214198
if (model === models.openrouter_grok_4) {
215199
return (
216200
getGrok4Cost(input_tokens, output_tokens) +

backend/src/llm-apis/vercel-ai-sdk/ai-sdk.ts

Lines changed: 1 addition & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,6 @@ import { google } from '@ai-sdk/google'
22
import { openai } from '@ai-sdk/openai'
33
import {
44
finetunedVertexModels,
5-
geminiModels,
65
openaiModels,
76
} from '@codebuff/common/old-constants'
87
import { buildArray } from '@codebuff/common/util/array'
@@ -19,11 +18,7 @@ import { saveMessage } from '../message-cost-tracker'
1918
import { openRouterLanguageModel } from '../openrouter'
2019
import { vertexFinetuned } from './vertex-finetuned'
2120

22-
import type {
23-
GeminiModel,
24-
Model,
25-
OpenAIModel,
26-
} from '@codebuff/common/old-constants'
21+
import type { Model, OpenAIModel } from '@codebuff/common/old-constants'
2722
import type { Message } from '@codebuff/common/types/messages/codebuff-message'
2823
import type {
2924
OpenRouterProviderOptions,
@@ -52,9 +47,6 @@ const modelToAiSDKModel = (model: Model): LanguageModel => {
5247
) {
5348
return vertexFinetuned(model)
5449
}
55-
if (Object.values(geminiModels).includes(model as GeminiModel)) {
56-
return google.languageModel(model)
57-
}
5850
if (model === openaiModels.o3pro || model === openaiModels.o3) {
5951
return openai.responses(model)
6052
}

backend/src/prompt-agent-stream.ts

Lines changed: 0 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -59,14 +59,6 @@ export const getAgentStreamFromTemplate = (params: {
5959
if (!options.providerOptions) {
6060
options.providerOptions = {}
6161
}
62-
if (provider === 'gemini') {
63-
if (!options.providerOptions.gemini) {
64-
options.providerOptions.gemini = {}
65-
}
66-
if (!options.providerOptions.gemini.thinkingConfig) {
67-
options.providerOptions.gemini.thinkingConfig = { thinkingBudget: 128 }
68-
}
69-
}
7062
if (!options.providerOptions.openrouter) {
7163
options.providerOptions.openrouter = {}
7264
}

backend/src/templates/agent-registry.ts

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -108,6 +108,7 @@ async function fetchAgentFromDatabase(parsedAgentId: {
108108
agentId,
109109
version: agentConfig.version,
110110
fullAgentId: agentTemplate.id,
111+
parsedAgentId,
111112
},
112113
'fetchAgentFromDatabase: Successfully loaded and validated agent from database',
113114
)

common/src/old-constants.ts

Lines changed: 1 addition & 18 deletions
Original file line numberDiff line numberDiff line change
@@ -185,14 +185,6 @@ export const openaiModels = {
185185
} as const
186186
export type OpenAIModel = (typeof openaiModels)[keyof typeof openaiModels]
187187

188-
export const geminiModels = {
189-
gemini2_5_flash: 'gemini-2.5-flash-preview-05-20',
190-
gemini2_5_flash_thinking: 'gemini-2.5-flash-preview-05-20:thinking',
191-
gemini2flash: 'gemini-2.0-flash-001',
192-
gemini2_5_pro_preview: 'gemini-2.5-pro-preview-06-05',
193-
} as const
194-
export type GeminiModel = (typeof geminiModels)[keyof typeof geminiModels]
195-
196188
export const openrouterModels = {
197189
openrouter_claude_sonnet_4_5: 'anthropic/claude-sonnet-4.5',
198190
openrouter_claude_sonnet_4: 'anthropic/claude-4-sonnet-20250522',
@@ -250,7 +242,6 @@ export type FinetunedVertexModel =
250242
export const models = {
251243
// ...claudeModels,
252244
...openaiModels,
253-
...geminiModels,
254245
...deepseekModels,
255246
...openrouterModels,
256247
...finetunedVertexModels,
@@ -273,12 +264,6 @@ export const shortModelNames = {
273264
}
274265

275266
export const providerModelNames = {
276-
...Object.fromEntries(
277-
Object.entries(geminiModels).map(([name, model]) => [
278-
model,
279-
'gemini' as const,
280-
]),
281-
),
282267
// ...Object.fromEntries(
283268
// Object.entries(openrouterModels).map(([name, model]) => [
284269
// model,
@@ -355,9 +340,7 @@ export const providerDomains = {
355340
export function getLogoForModel(modelName: string): string | undefined {
356341
let domain: string | undefined
357342

358-
if (Object.values(geminiModels).includes(modelName as GeminiModel))
359-
domain = providerDomains.google
360-
else if (Object.values(openaiModels).includes(modelName as OpenAIModel))
343+
if (Object.values(openaiModels).includes(modelName as OpenAIModel))
361344
domain = providerDomains.openai
362345
else if (Object.values(deepseekModels).includes(modelName as DeepseekModel))
363346
domain = providerDomains.deepseek

scripts/ft-file-selection/relabel-for-offline-scoring.ts

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@ import {
1111
import {
1212
finetunedVertexModelNames,
1313
finetunedVertexModels,
14-
geminiModels,
14+
models,
1515
TEST_USER_ID,
1616
} from '@codebuff/common/old-constants'
1717
import { generateCompactId } from '@codebuff/common/util/string'
@@ -35,7 +35,7 @@ const START_CURSOR = '2025-05-31T00:00:00.000Z' // User-provided start cursor or
3535
const GROUND_TRUTH_MODEL = 'claude-opus-4-20250514-with-full-file-context-new'
3636

3737
const MODELS = [
38-
geminiModels.gemini2flash,
38+
models.openrouter_gemini2_5_flash,
3939
finetunedVertexModels.ft_filepicker_005,
4040
finetunedVertexModels.ft_filepicker_007,
4141
finetunedVertexModels.ft_filepicker_008,
@@ -45,7 +45,7 @@ const MODELS = [
4545
] as const
4646

4747
const modelDescriptions = {
48-
[geminiModels.gemini2flash]:
48+
[models.openrouter_gemini2_5_flash]:
4949
'gemini-2.0-flash-001: Base model, used to tune the finetuned models',
5050
[finetunedVertexModels.ft_filepicker_005]:
5151
'ft_filepicker_005: 74.6M tokens, 2 epochs, trained on Claude 3.5 Sonnet outputs',

scripts/ft-file-selection/relabel-traces.ts

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -72,7 +72,8 @@ async function runTraces() {
7272
output = await promptFlashWithFallbacks(
7373
messagesWithSystem(messages as Message[], system as System),
7474
{
75-
model: model as typeof models.gemini2_5_pro_preview,
75+
model:
76+
model as typeof models.openrouter_gemini2_5_pro_preview,
7677
clientSessionId: 'relabel-trace-run',
7778
fingerprintId: 'relabel-trace-run',
7879
userInputId: 'relabel-trace-run',

0 commit comments

Comments
 (0)