ChatCloudflareWorkersAI
info
Workers AI is currently in Open Beta and is not recommended for production data and traffic, and limits + access are subject to change
Workers AI allows you to run machine learning models, on the Cloudflare network, from your own code.
Usage
import { ChatCloudflareWorkersAI } from "langchain/chat_models/cloudflare_workersai";
const model = new ChatCloudflareWorkersAI({
  model: "@cf/meta/llama-2-7b-chat-int8", // Default value
  cloudflareAccountId: process.env.CLOUDFLARE_ACCOUNT_ID,
  cloudflareApiToken: process.env.CLOUDFLARE_API_TOKEN,
  // Pass a custom base URL to use Cloudflare AI Gateway
  // baseUrl: `https://gateway.ai.cloudflare.com/v1/{YOUR_ACCOUNT_ID}/{GATEWAY_NAME}/workers-ai/`,
});
const response = await model.invoke([
  ["system", "You are a helpful assistant that translates English to German."],
  ["human", `Translate "I love programming".`],
]);
console.log(response);
/*
AIMessage {
  content: `Sure! Here's the translation of "I love programming" into German:\n` +
    '\n' +
    '"Ich liebe Programmieren."\n' +
    '\n' +
    'In this sentence, "Ich" means "I," "liebe" means "love," and "Programmieren" means "programming."',
  additional_kwargs: {}
}
*/
const stream = await model.stream([
  ["system", "You are a helpful assistant that translates English to German."],
  ["human", `Translate "I love programming".`],
]);
for await (const chunk of stream) {
  console.log(chunk);
}
/*
  AIMessageChunk {
    content: 'S',
    additional_kwargs: {}
  }
  AIMessageChunk {
    content: 'ure',
    additional_kwargs: {}
  }
  AIMessageChunk {
    content: '!',
    additional_kwargs: {}
  }
  AIMessageChunk {
    content: ' Here',
    additional_kwargs: {}
  }
  ...
*/
API Reference:
- ChatCloudflareWorkersAI from 
langchain/chat_models/cloudflare_workersai