Optional fields: anyOptional configuration: ClientOptions & LegacyOpenAIInputThis feature is deprecated and will be removed in the future.
It is not recommended for use.
Penalizes repeated tokens according to frequency
Model name to use
Model name to use
Alias for model
Number of completions to generate for each prompt
Penalizes repeated tokens
Whether to stream the results or not. Enabling disables tokenUsage reporting
Sampling temperature to use
Total probability mass of tokens to consider at each step
Optional apiAPI key to use when making requests to OpenAI. Defaults to the value of
OPENAI_API_KEY environment variable.
Optional azureADTokenA function that returns an access token for Microsoft Entra (formerly known as Azure Active Directory), which will be invoked on every request.
Optional azureAzure OpenAI API deployment name to use for completions when making requests to Azure OpenAI. This is the name of the deployment you created in the Azure portal. e.g. "my-openai-deployment" this will be used in the endpoint URL: https://{InstanceName}.openai.azure.com/openai/deployments/my-openai-deployment/
Optional azureAzure OpenAI API instance name to use when making requests to Azure OpenAI. this is the name of the instance you created in the Azure portal. e.g. "my-openai-instance" this will be used in the endpoint URL: https://my-openai-instance.openai.azure.com/openai/deployments/{DeploymentName}/
Optional azureAPI key to use when making requests to Azure OpenAI.
Optional azureAPI version to use when making requests to Azure OpenAI.
Optional azureCustom endpoint for Azure OpenAI API. This is useful in case you have a deployment in another region. e.g. setting this value to "https://westeurope.api.cognitive.microsoft.com/openai/deployments" will be result in the endpoint URL: https://westeurope.api.cognitive.microsoft.com/openai/deployments/{DeploymentName}/
Optional logitDictionary used to adjust the probability of specific tokens being generated
Optional logprobsWhether to return log probabilities of the output tokens or not. If true, returns the log probabilities of each output token returned in the content of message.
Optional maxMaximum number of tokens to generate in the completion. -1 returns as many tokens as possible given the prompt and the model's maximum context size.
Optional modelHolds any additional parameters that are valid to pass to openai.createCompletion that are not explicitly specified on this class.
Optional openAIApiAPI key to use when making requests to OpenAI. Defaults to the value of
OPENAI_API_KEY environment variable.
Alias for apiKey
Optional organizationOptional stopList of stop words to use when generating
Alias for stopSequences
Optional stopList of stop words to use when generating
Optional timeoutTimeout to use when making requests to OpenAI.
Optional topAn integer between 0 and 5 specifying the number of most likely tokens to return at each token position, each with an associated log probability. logprobs must be set to true if this parameter is used.
Optional userUnique string identifier representing your end-user, which can help OpenAI to monitor and detect abuse.
Protected clientProtected clientOptional kwargs: Partial<CallOptions>Calls the OpenAI API with retry logic in case of failures.
The request to send to the OpenAI API.
Optional options: OpenAICoreRequestOptionsOptional configuration for the API call.
The response from the OpenAI API.
Optional options: OpenAICoreRequestOptionsOptional config: anyOptional config: anyGenerated using TypeDoc
Wrapper around OpenAI large language models that use the Chat endpoint.
To use you should have the
OPENAI_API_KEYenvironment variable set.To use with Azure you should have the:
AZURE_OPENAI_API_KEY,AZURE_OPENAI_API_INSTANCE_NAME,AZURE_OPENAI_API_DEPLOYMENT_NAMEandAZURE_OPENAI_API_VERSIONenvironment variables set.AZURE_OPENAI_BASE_PATHis optional and will overrideAZURE_OPENAI_API_INSTANCE_NAMEif you need to use a custom endpoint.Remarks
Any parameters that are valid to be passed to
openai.createChatCompletioncan be passed through modelKwargs, even if not explicitly available on this class.Example