diff --git a/sdk/embedbase-js/src/types.ts b/sdk/embedbase-js/src/types.ts index 2d9c728..5f5e5db 100644 --- a/sdk/embedbase-js/src/types.ts +++ b/sdk/embedbase-js/src/types.ts @@ -68,7 +68,7 @@ export interface ClientDatasets { public: boolean } -export type LLM = 'openai/gpt-4' | 'openai/gpt-3.5-turbo-16k' | 'google/bison' // | 'tiiuae/falcon-7b' | 'bigscience/bloomz-7b1' +export type LLM = 'openai/gpt-4' | 'openai/gpt-3.5-turbo-16k' | 'google/bison' | 'bigscience/bloomz-7b1' //| 'tiiuae/falcon-7b' export interface LLMDescription { name: LLM diff --git a/sdk/embedbase-js/test/client.test.ts b/sdk/embedbase-js/test/client.test.ts index 63933ff..ea06e9b 100644 --- a/sdk/embedbase-js/test/client.test.ts +++ b/sdk/embedbase-js/test/client.test.ts @@ -219,7 +219,7 @@ test('should be able to filter by metadata using where', async () => { }, TIMEOUT) test('should be able to chat', async () => { - for await (const res of embedbase.useModel('openai/gpt-3.5-turbo-16k').streamText('tell a story', { + for await (const res of embedbase.useModel('bigscience/bloomz-7b1').streamText('1+1=', { // maxNewTokens: 10, })) { expect(res).toBeDefined() @@ -390,7 +390,7 @@ test('should provide clear error message', async () => { }, TIMEOUT) test('should be able to generate text sync', async () => { - const res = await embedbase.useModel('tiiuae/falcon-7b').generateText('1+1=', { + const res = await embedbase.useModel('bigscience/bloomz-7b1').generateText('1+1=', { maxNewTokens: 10, }) expect(res).toBeDefined()