Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
49 changes: 17 additions & 32 deletions packages/vertexai/src/methods/chrome-adapter.test.ts
Original file line number Diff line number Diff line change
Expand Up @@ -15,6 +15,7 @@
* limitations under the License.
*/

import { VertexAIError } from '../errors';
import { expect, use } from 'chai';
import sinonChai from 'sinon-chai';
import chaiAsPromised from 'chai-as-promised';
Expand All @@ -26,7 +27,7 @@ import {
LanguageModelMessageContent
} from '../types/language-model';
import { match, stub } from 'sinon';
import { GenerateContentRequest } from '../types';
import { GenerateContentRequest, VertexAIErrorCode } from '../types';

use(sinonChai);
use(chaiAsPromised);
Expand Down Expand Up @@ -363,17 +364,8 @@ describe('ChromeAdapter', () => {
});
});
describe('countTokens', () => {
it('counts tokens from a singular input', async () => {
it('counts tokens is not yet available', async () => {
const inputText = 'first';
const expectedCount = 10;
const onDeviceParams = {
systemPrompt: 'be yourself'
} as LanguageModelCreateOptions;
const expectedOnDeviceParams = {
systemPrompt: 'be yourself',
expectedInputs: [{ type: 'image' }]
} as LanguageModelCreateOptions;

// setting up stubs
const languageModelProvider = {
create: () => Promise.resolve({})
Expand All @@ -385,34 +377,27 @@ describe('ChromeAdapter', () => {
languageModel
);

// overrides impl with stub method
const measureInputUsageStub = stub(
languageModel,
'measureInputUsage'
).resolves(expectedCount);

const adapter = new ChromeAdapter(
languageModelProvider,
'prefer_on_device',
onDeviceParams
'prefer_on_device'
);

const countTokenRequest = {
contents: [{ role: 'user', parts: [{ text: inputText }] }]
} as GenerateContentRequest;
const response = await adapter.countTokens(countTokenRequest);
// Asserts initialization params are proxied.
expect(createStub).to.have.been.calledOnceWith(expectedOnDeviceParams);
// Asserts Vertex input type is mapped to Chrome type.
expect(measureInputUsageStub).to.have.been.calledOnceWith([
{
type: 'text',
content: inputText
}
]);
expect(await response.json()).to.deep.equal({
totalTokens: expectedCount
});

try {
await adapter.countTokens(countTokenRequest);
} catch (e) {
// the call to countToken should be rejected with Error
expect((e as VertexAIError).code).to.equal(
VertexAIErrorCode.REQUEST_ERROR
);
expect((e as VertexAIError).message).includes('not yet available');
}

// Asserts that no language model was initialized
expect(createStub).not.called;
});
});
describe('generateContentStream', () => {
Expand Down
24 changes: 7 additions & 17 deletions packages/vertexai/src/methods/chrome-adapter.ts
Original file line number Diff line number Diff line change
Expand Up @@ -15,11 +15,13 @@
* limitations under the License.
*/

import { VertexAIError } from '../errors';
import {
CountTokensRequest,
GenerateContentRequest,
InferenceMode,
Part
Part,
VertexAIErrorCode
} from '../types';
import {
Availability,
Expand Down Expand Up @@ -129,23 +131,11 @@ export class ChromeAdapter {
return ChromeAdapter.toStreamResponse(stream);
}

async countTokens(request: CountTokensRequest): Promise<Response> {
// TODO: Check if the request contains an image, and if so, throw.
const session = await this.createSession(
// TODO: normalize on-device params during construction.
this.onDeviceParams || {}
);
// TODO: support multiple content objects when Chrome supports
// sequence<LanguageModelMessage>
const contents = await Promise.all(
request.contents[0].parts.map(ChromeAdapter.toLanguageModelMessageContent)
async countTokens(_request: CountTokensRequest): Promise<Response> {
throw new VertexAIError(
VertexAIErrorCode.REQUEST_ERROR,
'Count Tokens is not yet available for on-device model.'
);
const tokenCount = await session.measureInputUsage(contents);
return {
json: async () => ({
totalTokens: tokenCount
})
} as Response;
}

/**
Expand Down
Loading