A tiny client module for the openAI API
npm install openai-apiYou can also visit the Issue tracker for more information or open a new issue.
This project is not affiliated with OpenAI and was written purely out of interest.
npm i openai-api
js
const OpenAI = require('openai-api');// Load your key from an environment variable or secret management service
// (do not include your key directly in your code)
const OPENAI_API_KEY = process.env.OPENAI_API_KEY;
const openai = new OpenAI(OPENAI_API_KEY);
`$3
`js
(async () => {
const gptResponse = await openai.complete({
engine: 'davinci',
prompt: 'this is a test',
maxTokens: 5,
temperature: 0.9,
topP: 1,
presencePenalty: 0,
frequencyPenalty: 0,
bestOf: 1,
n: 1,
stream: false,
stop: ['\n', "testing"]
}); console.log(gptResponse.data);
})();
`#### Example of a successful completion response:
`js
{
id: 'some-long-id',
object: 'text_completion',
created: 1616791508,
model: 'davinci:2020-05-03',
choices: [
{
text: " predicted text...",
index: 0,
logprobs: null,
finish_reason: 'length'
}
]
}
`$3
`js
(async () => {
const gptResponse = await openai.search({
engine: 'davinci',
documents: ["White House", "hospital", "school"],
query: "the president"
}); console.log(gptResponse.data);
})();
`
$3
`js
(async () => {
const gptResponse = await openai.answers({
"documents": ["Puppy A is happy.", "Puppy B is sad."],
"question": "which puppy is happy?",
"search_model": "ada",
"model": "curie",
"examples_context": "In 2017, U.S. life expectancy was 78.6 years.",
"examples": [["What is human life expectancy in the United States?", "78 years."]],
"max_tokens": 5,
"stop": ["\n", "<|endoftext|>"],
}); console.log(gptResponse.data);
})();
`$3
`js
(async () => {
const gptResponse = await openai.classification({
"examples": [
["A happy moment", "Positive"],
["I am sad.", "Negative"],
["I am feeling awesome", "Positive"]
],
"labels": ["Positive", "Negative", "Neutral"],
"query": "It is a raining day :(",
"search_model": "ada",
"model": "curie"
}); console.log(gptResponse.data);
})();
`$3
`js
(async () => {
const gptResponse = await openai.engines(); console.log(gptResponse.data);
})();
`$3
Documentation: https://beta.openai.com/docs/api-reference/embeddings`js
(async () => {
const gptResponse = await openai.embeddings({
"engine": "test-similarity-babbage-001",
"input": [
"A happy moment",
"I am sad.",
"I am feeling awesome"
],
}); console.log(gptResponse.data); // see index.d.ts interface Embedding
})();
`
$3
#### Not supported as of 4/21. See issue #20
The token limit is 2048 for completions using the OpenAI API. This method allows you to get the number of tokens in your prompt. This is done offline (no API call is made).`js
openai.encode('This is an encoding test. Number of tokens is not necessarily the same as word count.').then((result) => {
console.log("Number of tokens for string:" + result.length);
});
``-------
Powered by Obelisk.ooo