From ea1d88ede468f87ea80d2c0ae6abd7fb2cd5463f Mon Sep 17 00:00:00 2001 From: gericdong Date: Thu, 15 Feb 2024 15:18:55 -0500 Subject: [PATCH 1/2] fix(aiplatform): update Gemini model names --- generative-ai/snippets/countTokens.js | 2 +- generative-ai/snippets/functionCallingStreamChat.js | 2 +- generative-ai/snippets/functionCallingStreamContent.js | 2 +- generative-ai/snippets/nonStreamingChat.js | 2 +- generative-ai/snippets/nonStreamingContent.js | 2 +- generative-ai/snippets/nonStreamingMultipartContent.js | 2 +- generative-ai/snippets/safetySettings.js | 2 +- generative-ai/snippets/sendMultiModalPromptWithImage.js | 2 +- generative-ai/snippets/sendMultiModalPromptWithVideo.js | 2 +- generative-ai/snippets/streamChat.js | 2 +- generative-ai/snippets/streamContent.js | 2 +- generative-ai/snippets/streamMultipartContent.js | 2 +- generative-ai/snippets/test/countTokens.test.js | 2 +- generative-ai/snippets/test/functionCallingStreamChat.test.js | 2 +- .../snippets/test/functionCallingStreamContent.test.js | 2 +- generative-ai/snippets/test/nonStreamingChat.test.js | 2 +- generative-ai/snippets/test/nonStreamingContent.test.js | 2 +- .../snippets/test/nonStreamingMultipartContent.test.js | 2 +- generative-ai/snippets/test/safetySettings.test.js | 2 +- .../snippets/test/sendMultiModalPromptWithImage.test.js | 2 +- .../snippets/test/sendMultiModalPromptWithVideo.test.js | 2 +- generative-ai/snippets/test/streamChat.test.js | 2 +- generative-ai/snippets/test/streamContent.test.js | 2 +- generative-ai/snippets/test/streamMultipartContent.test.js | 2 +- 24 files changed, 24 insertions(+), 24 deletions(-) diff --git a/generative-ai/snippets/countTokens.js b/generative-ai/snippets/countTokens.js index 40ce023064..eb806aa616 100644 --- a/generative-ai/snippets/countTokens.js +++ b/generative-ai/snippets/countTokens.js @@ -21,7 +21,7 @@ const {VertexAI} = require('@google-cloud/vertexai'); async function countTokens( projectId = 'PROJECT_ID', location = 'us-central1', - model = 'gemini-pro' + model = 'gemini-1.0-pro' ) { // Initialize Vertex with your Cloud project and location const vertexAI = new VertexAI({project: projectId, location: location}); diff --git a/generative-ai/snippets/functionCallingStreamChat.js b/generative-ai/snippets/functionCallingStreamChat.js index 62caf5d12c..9ba0396724 100644 --- a/generative-ai/snippets/functionCallingStreamChat.js +++ b/generative-ai/snippets/functionCallingStreamChat.js @@ -55,7 +55,7 @@ const functionResponseParts = [ async function functionCallingStreamChat( projectId = 'PROJECT_ID', location = 'us-central1', - model = 'gemini-pro' + model = 'gemini-1.0-pro' ) { // Initialize Vertex with your Cloud project and location const vertexAI = new VertexAI({project: projectId, location: location}); diff --git a/generative-ai/snippets/functionCallingStreamContent.js b/generative-ai/snippets/functionCallingStreamContent.js index e0b71e2a29..1b5331cc32 100644 --- a/generative-ai/snippets/functionCallingStreamContent.js +++ b/generative-ai/snippets/functionCallingStreamContent.js @@ -55,7 +55,7 @@ const functionResponseParts = [ async function functionCallingStreamChat( projectId = 'PROJECT_ID', location = 'us-central1', - model = 'gemini-pro' + model = 'gemini--1.0-pro' ) { // Initialize Vertex with your Cloud project and location const vertexAI = new VertexAI({project: projectId, location: location}); diff --git a/generative-ai/snippets/nonStreamingChat.js b/generative-ai/snippets/nonStreamingChat.js index 8d2af6609b..8b5a4f705b 100644 --- a/generative-ai/snippets/nonStreamingChat.js +++ b/generative-ai/snippets/nonStreamingChat.js @@ -21,7 +21,7 @@ const {VertexAI} = require('@google-cloud/vertexai'); async function createNonStreamingChat( projectId = 'PROJECT_ID', location = 'us-central1', - model = 'gemini-pro' + model = 'gemini-1.0-pro' ) { // Initialize Vertex with your Cloud project and location const vertexAI = new VertexAI({project: projectId, location: location}); diff --git a/generative-ai/snippets/nonStreamingContent.js b/generative-ai/snippets/nonStreamingContent.js index 65b4dc80ea..fc0d815dd9 100644 --- a/generative-ai/snippets/nonStreamingContent.js +++ b/generative-ai/snippets/nonStreamingContent.js @@ -21,7 +21,7 @@ const {VertexAI} = require('@google-cloud/vertexai'); async function createNonStreamingContent( projectId = 'PROJECT_ID', location = 'us-central1', - model = 'gemini-pro' + model = 'gemini-1.0-pro' ) { // Initialize Vertex with your Cloud project and location const vertexAI = new VertexAI({project: projectId, location: location}); diff --git a/generative-ai/snippets/nonStreamingMultipartContent.js b/generative-ai/snippets/nonStreamingMultipartContent.js index 962339fd8e..75ca596bd1 100644 --- a/generative-ai/snippets/nonStreamingMultipartContent.js +++ b/generative-ai/snippets/nonStreamingMultipartContent.js @@ -21,7 +21,7 @@ const {VertexAI} = require('@google-cloud/vertexai'); async function createNonStreamingMultipartContent( projectId = 'PROJECT_ID', location = 'us-central1', - model = 'gemini-pro-vision', + model = 'gemini-1.0-pro-vision', image = 'gs://generativeai-downloads/images/scones.jpg', mimeType = 'image/jpeg' ) { diff --git a/generative-ai/snippets/safetySettings.js b/generative-ai/snippets/safetySettings.js index 306fa54db5..9aa07b1106 100644 --- a/generative-ai/snippets/safetySettings.js +++ b/generative-ai/snippets/safetySettings.js @@ -25,7 +25,7 @@ const { async function setSafetySettings( projectId = 'PROJECT_ID', location = 'us-central1', - model = 'gemini-pro' + model = 'gemini-1.0-pro' ) { // Initialize Vertex with your Cloud project and location const vertexAI = new VertexAI({project: projectId, location: location}); diff --git a/generative-ai/snippets/sendMultiModalPromptWithImage.js b/generative-ai/snippets/sendMultiModalPromptWithImage.js index acc4514b28..e7a30462c7 100644 --- a/generative-ai/snippets/sendMultiModalPromptWithImage.js +++ b/generative-ai/snippets/sendMultiModalPromptWithImage.js @@ -27,7 +27,7 @@ async function getBase64(url) { async function sendMultiModalPromptWithImage( projectId = 'PROJECT_ID', location = 'us-central1', - model = 'gemini-pro-vision' + model = 'gemini-1.0-pro-vision' ) { // For images, the SDK supports base64 strings const landmarkImage1 = await getBase64( diff --git a/generative-ai/snippets/sendMultiModalPromptWithVideo.js b/generative-ai/snippets/sendMultiModalPromptWithVideo.js index a6c4dad1cc..0a8d72c93a 100644 --- a/generative-ai/snippets/sendMultiModalPromptWithVideo.js +++ b/generative-ai/snippets/sendMultiModalPromptWithVideo.js @@ -21,7 +21,7 @@ const {VertexAI} = require('@google-cloud/vertexai'); async function sendMultiModalPromptWithVideo( projectId = 'PROJECT_ID', location = 'us-central1', - model = 'gemini-pro-vision' + model = 'gemini-1.0-pro-vision' ) { // Initialize Vertex with your Cloud project and location const vertexAI = new VertexAI({project: projectId, location: location}); diff --git a/generative-ai/snippets/streamChat.js b/generative-ai/snippets/streamChat.js index 7400774dc8..73dcb8b654 100644 --- a/generative-ai/snippets/streamChat.js +++ b/generative-ai/snippets/streamChat.js @@ -21,7 +21,7 @@ const {VertexAI} = require('@google-cloud/vertexai'); async function createStreamChat( projectId = 'PROJECT_ID', location = 'us-central1', - model = 'gemini-pro' + model = 'gemini-1.0-pro' ) { // Initialize Vertex with your Cloud project and location const vertexAI = new VertexAI({project: projectId, location: location}); diff --git a/generative-ai/snippets/streamContent.js b/generative-ai/snippets/streamContent.js index cfaf156e5d..c153dd8668 100644 --- a/generative-ai/snippets/streamContent.js +++ b/generative-ai/snippets/streamContent.js @@ -21,7 +21,7 @@ const {VertexAI} = require('@google-cloud/vertexai'); async function createStreamContent( projectId = 'PROJECT_ID', location = 'us-central1', - model = 'gemini-pro' + model = 'gemini-1.0-pro' ) { // Initialize Vertex with your Cloud project and location const vertexAI = new VertexAI({project: projectId, location: location}); diff --git a/generative-ai/snippets/streamMultipartContent.js b/generative-ai/snippets/streamMultipartContent.js index 670a0b13c2..fe8e5d3e0e 100644 --- a/generative-ai/snippets/streamMultipartContent.js +++ b/generative-ai/snippets/streamMultipartContent.js @@ -21,7 +21,7 @@ const {VertexAI} = require('@google-cloud/vertexai'); async function createStreamMultipartContent( projectId = 'PROJECT_ID', location = 'us-central1', - model = 'gemini-pro-vision', + model = 'gemini-1.0-pro-vision', image = 'gs://generativeai-downloads/images/scones.jpg', mimeType = 'image/jpeg' ) { diff --git a/generative-ai/snippets/test/countTokens.test.js b/generative-ai/snippets/test/countTokens.test.js index d168b26fe5..9cae763dba 100644 --- a/generative-ai/snippets/test/countTokens.test.js +++ b/generative-ai/snippets/test/countTokens.test.js @@ -23,7 +23,7 @@ const execSync = cmd => cp.execSync(cmd, {encoding: 'utf-8'}); describe('Count tokens', async () => { const project = 'cloud-llm-preview1'; const location = 'us-central1'; - const model = 'gemini-pro'; + const model = 'gemini-1.0-pro'; it('should count tokens', async () => { const output = execSync( diff --git a/generative-ai/snippets/test/functionCallingStreamChat.test.js b/generative-ai/snippets/test/functionCallingStreamChat.test.js index 482d79bd40..00439a3d04 100644 --- a/generative-ai/snippets/test/functionCallingStreamChat.test.js +++ b/generative-ai/snippets/test/functionCallingStreamChat.test.js @@ -23,7 +23,7 @@ const execSync = cmd => cp.execSync(cmd, {encoding: 'utf-8'}); describe('Generative AI Function Calling Stream Chat', () => { const project = 'cloud-llm-preview1'; const location = 'us-central1'; - const model = 'gemini-pro'; + const model = 'gemini-1.0-pro'; it('should create stream chat and begin the conversation the same in each instance', async () => { const output = execSync( diff --git a/generative-ai/snippets/test/functionCallingStreamContent.test.js b/generative-ai/snippets/test/functionCallingStreamContent.test.js index f80ea029d8..7524a7493e 100644 --- a/generative-ai/snippets/test/functionCallingStreamContent.test.js +++ b/generative-ai/snippets/test/functionCallingStreamContent.test.js @@ -23,7 +23,7 @@ const execSync = cmd => cp.execSync(cmd, {encoding: 'utf-8'}); describe('Generative AI Function Calling Stream Content', () => { const project = 'cloud-llm-preview1'; const location = 'us-central1'; - const model = 'gemini-pro'; + const model = 'gemini-1.0-pro'; it('should create stream chat and begin the conversation the same in each instance', async () => { const output = execSync( diff --git a/generative-ai/snippets/test/nonStreamingChat.test.js b/generative-ai/snippets/test/nonStreamingChat.test.js index 24ffe5a58b..9a6ccdc50b 100644 --- a/generative-ai/snippets/test/nonStreamingChat.test.js +++ b/generative-ai/snippets/test/nonStreamingChat.test.js @@ -23,7 +23,7 @@ const execSync = cmd => cp.execSync(cmd, {encoding: 'utf-8'}); describe('Generative AI NonStreaming Chat', async () => { const project = 'cloud-llm-preview1'; const location = 'us-central1'; - const model = 'gemini-pro'; + const model = 'gemini-1.0-pro'; it('should create nonstreaming chat and begin the conversation the same in each instance', async () => { const output = execSync( diff --git a/generative-ai/snippets/test/nonStreamingContent.test.js b/generative-ai/snippets/test/nonStreamingContent.test.js index 51134de8b2..f0eab652cf 100644 --- a/generative-ai/snippets/test/nonStreamingContent.test.js +++ b/generative-ai/snippets/test/nonStreamingContent.test.js @@ -23,7 +23,7 @@ const execSync = cmd => cp.execSync(cmd, {encoding: 'utf-8'}); describe('Generative AI NonStreaming Content', () => { const project = 'cloud-llm-preview1'; const location = 'us-central1'; - const model = 'gemini-pro'; + const model = 'gemini-1.0-pro'; it('should create nonstreaming content and begin the conversation the same in each instance', async () => { const output = execSync( diff --git a/generative-ai/snippets/test/nonStreamingMultipartContent.test.js b/generative-ai/snippets/test/nonStreamingMultipartContent.test.js index 5e0888bb86..ecfeb2eae3 100644 --- a/generative-ai/snippets/test/nonStreamingMultipartContent.test.js +++ b/generative-ai/snippets/test/nonStreamingMultipartContent.test.js @@ -23,7 +23,7 @@ const execSync = cmd => cp.execSync(cmd, {encoding: 'utf-8'}); describe('Generative AI NonStreaming Multipart Content', () => { const project = 'cloud-llm-preview1'; const location = 'us-central1'; - const model = 'gemini-pro-vision'; + const model = 'gemini-1.0-pro-vision'; const image = 'gs://generativeai-downloads/images/scones.jpg'; it('should create nonstreaming multipart content and begin the conversation the same in each instance', async () => { diff --git a/generative-ai/snippets/test/safetySettings.test.js b/generative-ai/snippets/test/safetySettings.test.js index 265def1791..454ae29c58 100644 --- a/generative-ai/snippets/test/safetySettings.test.js +++ b/generative-ai/snippets/test/safetySettings.test.js @@ -23,7 +23,7 @@ const execSync = cmd => cp.execSync(cmd, {encoding: 'utf-8'}); describe('Safety settings', async () => { const project = 'cloud-llm-preview1'; const location = 'us-central1'; - const model = 'gemini-pro'; + const model = 'gemini-1.0-pro'; it('should reject a dangerous request', async () => { const output = execSync( diff --git a/generative-ai/snippets/test/sendMultiModalPromptWithImage.test.js b/generative-ai/snippets/test/sendMultiModalPromptWithImage.test.js index 414fff580f..91a951552b 100644 --- a/generative-ai/snippets/test/sendMultiModalPromptWithImage.test.js +++ b/generative-ai/snippets/test/sendMultiModalPromptWithImage.test.js @@ -23,7 +23,7 @@ const execSync = cmd => cp.execSync(cmd, {encoding: 'utf-8'}); describe('Generative AI Stream MultiModal with Image', () => { const project = 'cloud-llm-preview1'; const location = 'us-central1'; - const model = 'gemini-pro-vision'; + const model = 'gemini-1.0-pro-vision'; it('should create stream multimodal content', async () => { const output = execSync( diff --git a/generative-ai/snippets/test/sendMultiModalPromptWithVideo.test.js b/generative-ai/snippets/test/sendMultiModalPromptWithVideo.test.js index 17e3e07bef..3b7972890e 100644 --- a/generative-ai/snippets/test/sendMultiModalPromptWithVideo.test.js +++ b/generative-ai/snippets/test/sendMultiModalPromptWithVideo.test.js @@ -23,7 +23,7 @@ const execSync = cmd => cp.execSync(cmd, {encoding: 'utf-8'}); describe('Generative AI Stream MultiModal with Video', () => { const project = 'cloud-llm-preview1'; const location = 'us-central1'; - const model = 'gemini-pro-vision'; + const model = 'gemini-1.0-pro-vision'; it('should create stream multimodal content', async () => { const output = execSync( diff --git a/generative-ai/snippets/test/streamChat.test.js b/generative-ai/snippets/test/streamChat.test.js index 823fe6ccce..1c9a128d6b 100644 --- a/generative-ai/snippets/test/streamChat.test.js +++ b/generative-ai/snippets/test/streamChat.test.js @@ -23,7 +23,7 @@ const execSync = cmd => cp.execSync(cmd, {encoding: 'utf-8'}); describe('Generative AI Stream Chat', () => { const project = 'cloud-llm-preview1'; const location = 'us-central1'; - const model = 'gemini-pro'; + const model = 'gemini-1.0-pro'; it('should create stream chat and begin the conversation the same in each instance', async () => { const output = execSync( diff --git a/generative-ai/snippets/test/streamContent.test.js b/generative-ai/snippets/test/streamContent.test.js index 769b68372e..03a0f47676 100644 --- a/generative-ai/snippets/test/streamContent.test.js +++ b/generative-ai/snippets/test/streamContent.test.js @@ -23,7 +23,7 @@ const execSync = cmd => cp.execSync(cmd, {encoding: 'utf-8'}); describe('Generative AI Stream Content', () => { const project = 'cloud-llm-preview1'; const location = 'us-central1'; - const model = 'gemini-pro'; + const model = 'gemini-1.0-pro'; it('should create stream content', async () => { const output = execSync( diff --git a/generative-ai/snippets/test/streamMultipartContent.test.js b/generative-ai/snippets/test/streamMultipartContent.test.js index d9b4fb061a..06153467d8 100644 --- a/generative-ai/snippets/test/streamMultipartContent.test.js +++ b/generative-ai/snippets/test/streamMultipartContent.test.js @@ -23,7 +23,7 @@ const execSync = cmd => cp.execSync(cmd, {encoding: 'utf-8'}); describe('Generative AI Stream Multipart Content', () => { const project = 'cloud-llm-preview1'; const location = 'us-central1'; - const model = 'gemini-pro-vision'; + const model = 'gemini-1.0-pro-vision'; const image = 'gs://generativeai-downloads/images/scones.jpg'; it('should create stream multipart content', async () => { From a1dab968029dcdb251365979fda18503ea5369c4 Mon Sep 17 00:00:00 2001 From: gericdong Date: Thu, 15 Feb 2024 15:26:32 -0500 Subject: [PATCH 2/2] Fix the model name --- generative-ai/snippets/functionCallingStreamContent.js | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/generative-ai/snippets/functionCallingStreamContent.js b/generative-ai/snippets/functionCallingStreamContent.js index 1b5331cc32..18247c7f3a 100644 --- a/generative-ai/snippets/functionCallingStreamContent.js +++ b/generative-ai/snippets/functionCallingStreamContent.js @@ -55,7 +55,7 @@ const functionResponseParts = [ async function functionCallingStreamChat( projectId = 'PROJECT_ID', location = 'us-central1', - model = 'gemini--1.0-pro' + model = 'gemini-1.0-pro' ) { // Initialize Vertex with your Cloud project and location const vertexAI = new VertexAI({project: projectId, location: location});