OpenAI.ChatGPT.EntityFrameworkCore 4.0.0-alpha

This is a prerelease version of OpenAI.ChatGPT.EntityFrameworkCore.
There is a newer version of this package available.
See the version list below for details.
dotnet add package OpenAI.ChatGPT.EntityFrameworkCore --version 4.0.0-alpha                
NuGet\Install-Package OpenAI.ChatGPT.EntityFrameworkCore -Version 4.0.0-alpha                
This command is intended to be used within the Package Manager Console in Visual Studio, as it uses the NuGet module's version of Install-Package.
<PackageReference Include="OpenAI.ChatGPT.EntityFrameworkCore" Version="4.0.0-alpha" />                
For projects that support PackageReference, copy this XML node into the project file to reference the package.
paket add OpenAI.ChatGPT.EntityFrameworkCore --version 4.0.0-alpha                
#r "nuget: OpenAI.ChatGPT.EntityFrameworkCore, 4.0.0-alpha"                
#r directive can be used in F# Interactive and Polyglot Notebooks. Copy this into the interactive tool or source code of the script to reference the package.
// Install OpenAI.ChatGPT.EntityFrameworkCore as a Cake Addin
#addin nuget:?package=OpenAI.ChatGPT.EntityFrameworkCore&version=4.0.0-alpha&prerelease

// Install OpenAI.ChatGPT.EntityFrameworkCore as a Cake Tool
#tool nuget:?package=OpenAI.ChatGPT.EntityFrameworkCore&version=4.0.0-alpha&prerelease                

example gif...

ChatGPT integration for .NET (+DI)

Nuget.NET
OpenAI Chat Completions API (ChatGPT) integration with DI and EF Core supporting. It allows you to use the API in your .NET applications. Also, the client supports streaming responses (like ChatGPT) via async streams.

2023.11 UPD: GPT4Turbo and JSON mode support

StructuredResponse module allows you to get structured responses from the API as C# object. See: StructuredResponse section.

Content

Preparation

First, you need to create an OpenAI account and get an API key. You can do this at https://platform.openai.com/account/api-keys.

Installation

The easiest way to use ChatGPT service in your .NET project with DI and persistence (EF Core) supporting is to install the NuGet package OpenAI.ChatGPT.EntityFrameworkCore:

Install-Package OpenAI.ChatGPT.EntityFrameworkCore

If you don't want to use EF Core, you can install the package OpenAI.ChatGPT.AspNetCore and implement your own storage for chat history, using IChatHistoryStorage interface.

Usage

  1. Set the OpenAI API key or even host (optional) in your project user secrets, or the appsettings.json file (not safe):
{
  "OpenAICredentials": {
    "ApiKey": "your-api-key-from-openai",
    "ApiHost": "https://api.openai.com/v1/"
  }
}

Also, you can specify OpenAI API key as environment variable ASPNETCORE_OpenAICredentials:ApiKey.

  1. Add ChatGPT integration with EF to your DI container:
builder.Services.AddChatGptEntityFrameworkIntegration(
    options => options.UseSqlite("Data Source=chats.db"));

Instead of options.UseSqlite("Data Source=chats.db") use your own db and connection string.

  1. Inject ChatGPTFactory to your service and use it to create ChatGPT instance:
public class YourService
{
    private readonly ChatGPTFactory _chatGptFactory;

    public YourService(ChatGPTFactory chatGptFactory)
    {
        _chatGptFactory = chatGptFactory;
    }

    public async Task<string> GetAnswer(string text)
    {
        ChatGPT chatGpt = await _chatGptFactory.Create(userId);
        var chatService = await chatGpt.ContinueOrStartNewTopic();
        response = await _chatService.GetNextMessageResponse(_prompt);
        return response;
    }
}

See Blazor Example.

If you want to configure request parameters, you can do it in appsettings.json configuration or in ChatGPTFactory.Create or in ChatGPT.CreateChat methods.

{
  "ChatGPTConfig": {
      "InitialSystemMessage": "You are a helpful and kind assistant.",
      "InitialUserMessage": null,
      "MaxTokens": null,
      "Model": null,
      "Temperature": null,
      "PassUserIdToOpenAiRequests": true
  }
}

See parameters description inside ChatGPTConfig.

Exceptions

If the server response is not a success status code, the client will throw a NotExpectedResponseException. The exception will contain the error message from the OpenAI API.
By default, requesting cancellation or ChatService.Stop() method calling will throw OperationCanceledException. If you don't want to throw it (relevant for streaming responses), you can set throwOnCancellation parameter to false:

await foreach (string chunk in chatService.StreamNextMessageResponse(text, throwOnCancellation: false))
{
    //...
}

Thread safety and async

ChatGPTFactory, ChatGPT classes thread-safety is depend on the IChatHistoryStorage implementation. If you use ChatGPTFactory with entity framework, it's NOT thread-safe. ChatService class is not thread-safe.
Anyways, these services are designed to be used safely with DI, so you don't need to worry about it.
All the methods from all the packages are designed to be used in async context and use ConfigureAwait(false) (thanks for the ConfigureAwait.Fody package).

Retries, timeouts and other policies

Since ChatGPTFactory depends on IHttpClientFactory, you can easily use any of the available policies for it, like Polly.

Modules

StructuredResponse

This module allows you to get structured responses from the API as C# object. It's useful if you want to use the API for something more than just chat.

record City(string Name, int YearOfFoundation, string Country);

var message = Dialog
    .StartAsSystem("Return requested data.")
    .ThenUser("I need info about Almaty city");
City almaty = await _client.GetStructuredResponse<City>(message, model: ChatCompletionModels.Gpt4Turbo);
Console.WriteLine(almaty); // Name: "Almaty", Country: "Kazakhstan", YearOfFoundation: 1854

Under the hood, it uses the new json mode of the API for GPT4Turbo and for the gpt-3.5-turbo-1106. Regular GPT4 and GPT3.5Turbo models are also supported, but GPT3.5 responses may be unstable (for GPT3.5 it's strictly recommended to provide examples parameter).

More complex examples with arrays, nested objects and enums are available in tests: https://github.com/rodion-m/ChatGPT_API_dotnet/blob/f50d386f0b65a4ba8c1041a28bab2a1a475c2296/tests/OpenAI.ChatGpt.IntegrationTests/OpenAiClientTests/OpenAiClient_GetStructuredResponse.cs#L1

NuGet: https://www.nuget.org/packages/OpenAI.ChatGPT.Modules.StructuredResponse

Translator

This module allows you to translate messages from one language to another.

string textToTranslate = "Hello, world!";
string translatedText = await _client.TranslateText(textToTranslate, "English", "Russian");
Console.WriteLine(translatedText); // "Привет, мир!"

Also, it's possible to translate entire object in pair with StructuredResponse module:

var objectToTranslate = new Order(
    new List<Order.Item> 
    {
        new(1,"Book", 5),
        new(2,"Pen", 10),
        new(3,"Notebook", 3)
    }
);
Order translatedObject = await _client.TranslateObject(objectToTranslate, "English", "Russian");

See full example in tests: https://github.com/rodion-m/ChatGPT_API_dotnet/blob/11658b76b497b9cc4ac74621c91c5e22cd724f2e/tests/OpenAI.ChatGpt.IntegrationTests/ChatGptTranslatorServiceTests.cs#L36

NuGet: https://www.nuget.org/packages/OpenAI.ChatGPT.Modules.Translator

Examples

API Parameters

Here is a list of the main parameters that can be used in the ChatCompletions (ChatGPT) API request (OpenAI.ChatGpt/Models/ChatCompletion/ChatCompletionRequest.cs). Some of them are taken from this article: https://towardsdatascience.com/gpt-3-parameters-and-prompt-design-1a595dc5b405
Below listed parameters for ChatCompletions API.

Model

The prediction-generating AI model is specified by the engine parameter. The available models are described below: https://platform.openai.com/docs/models

C# Model API Model
ChatCompletionModels.Gpt4Turbo gpt-4-1106-preview
ChatCompletionModels.Gpt4 gpt-4
ChatCompletionModels.Gpt4_0613 gpt-4-0613
ChatCompletionModels.Gpt4_32k gpt-4-32k
ChatCompletionModels.Gpt4_32k_0613 gpt-4-32k-0613
ChatCompletionModels.Gpt3_5_Turbo gpt-3.5-turbo
ChatCompletionModels.Gpt3_5_Turbo_1106 gpt-3.5-turbo-1106
ChatCompletionModels.Gpt3_5_Turbo_16k gpt-3.5-turbo-16k
ChatCompletionModels.Gpt3_5_Turbo_0613 gpt-3.5-turbo-0613
ChatCompletionModels.Gpt3_5_Turbo_16k_0613 gpt-3.5-turbo-16k-0613
ChatCompletionModels.Gpt4_0314 gpt-4-0314
ChatCompletionModels.Gpt4_32k_0314 gpt-4-32k-0314
ChatCompletionModels.Gpt3_5_Turbo_0301 gpt-3.5-turbo-0301

MaxTokens

The maximum number of tokens allowed for the generated answer. Defaults to ChatCompletionRequest.MaxTokensDefault (64).

  • This value is validated and limited with ChatCompletionModels.GetMaxTokensLimitForModel method.
  • It's possible to calculate approximately tokens count using ChatCompletionMessage.CalculateApproxTotalTokenCount method
  • The number of tokens can be retrieved from the API response: ChatCompletionResponse.Usage.TotalTokens. As a rule of thumb for English, 1 token is around 4 characters (so 100 tokens ≈ 75 words). See tokenizer from OpenAI: https://platform.openai.com/tokenizer
  • Encoding algorithm can be found here: https://github.com/latitudegames/GPT-3-Encoder

Temperature

What sampling temperature to use, between 0 and 2.

  • Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic.
  • Predefined values are available in ChatCompletionTemperatures.
  • Default value is: ChatCompletionTemperatures.Balanced (0.5).

Description: Before being mapped into probabilities, the model outputs unnormalized values (logits). The logits are typically used with a function such as softmax to convert them into probabilities.

But, before applying the softmax function, we can use a trick inspired by thermodynamics and scale the logits with the temperature parameter, i.e. softmax(logits/temperature).

A temperature parameter close to 1 would mean that the logits are passed through the softmax function without modification. If the temperature is close to zero, the highest probable tokens will become very likely compared to the other tokens, i.e. the model becomes more deterministic and will always output the same set of tokens after a given sequence of words.

More parameters description can be found here: Some of them are taken from this article: https://towardsdatascience.com/gpt-3-parameters-and-prompt-design-1a595dc5b405

Using raw client without DI

If you don't need DI and chat history, you can use only the NuGet package OpenAI.ChatGPT:

Install-Package OpenAI.ChatGPT

Then create an instance of OpenAIClient:

_client = new OpenAiClient("{YOUR_OPENAI_API_KEY}");

Simple usage of the Chat Completions API (raw client)

string text = "Who are you?";
string response = await _client.GetChatCompletions(new UserMessage(text), maxTokens: 80);
Console.WriteLine(response);

Streaming response with async streams (like ChatGPT)

var text = "Write the world top 3 songs of Soul genre";
await foreach (string chunk in _client.StreamChatCompletions(new UserMessage(text), maxTokens: 80))
{
    Console.Write(chunk);
}

Continue dialog with ChatGPT (message history)

Use ThenAssistant and ThenUser methods to create a dialog:

var dialog = Dialog.StartAsUser("How many meters are in a kilometer? Write just the number.") //the message from user
          .ThenAssistant("1000") // response from the assistant
          .ThenUser("Convert it to hex. Write just the number."); // the next message from user

await foreach (var chunk in _client.StreamChatCompletions(dialog, maxTokens: 80))
{
    Console.Write(chunk);
}

Or just send message history as a collection.

Product Compatible and additional computed target framework versions.
.NET net6.0 is compatible.  net6.0-android was computed.  net6.0-ios was computed.  net6.0-maccatalyst was computed.  net6.0-macos was computed.  net6.0-tvos was computed.  net6.0-windows was computed.  net7.0 is compatible.  net7.0-android was computed.  net7.0-ios was computed.  net7.0-maccatalyst was computed.  net7.0-macos was computed.  net7.0-tvos was computed.  net7.0-windows was computed.  net8.0 is compatible.  net8.0-android was computed.  net8.0-browser was computed.  net8.0-ios was computed.  net8.0-maccatalyst was computed.  net8.0-macos was computed.  net8.0-tvos was computed.  net8.0-windows was computed. 
Compatible target framework(s)
Included target framework(s) (in package)
Learn more about Target Frameworks and .NET Standard.

NuGet packages

This package is not used by any NuGet packages.

GitHub repositories

This package is not used by any popular GitHub repositories.

Version Downloads Last updated
4.1.0 250 7/28/2024
4.1.0-alpha 285 12/17/2023
4.0.2-alpha 346 12/5/2023
4.0.1-alpha 122 12/5/2023
4.0.0-alpha 119 12/5/2023
3.3.0 439 11/24/2023
3.2.0 670 11/17/2023
3.1.1 212 11/11/2023
3.1.0 140 11/10/2023
3.0.0 186 11/8/2023
2.9.3 143 11/8/2023
2.9.2 124 11/7/2023
2.9.1 176 11/3/2023
2.9.0 316 10/20/2023
2.8.0 579 7/20/2023
2.7.1 184 7/13/2023
2.7.0 212 7/2/2023
2.6.0 186 6/17/2023
2.5.0 347 4/28/2023
2.4.2 205 4/24/2023
2.4.1 312 4/24/2023
2.4.0 214 4/24/2023
2.3.0 198 4/20/2023
2.2.2 231 4/19/2023
2.2.0 199 4/19/2023
2.1.0 207 4/18/2023
2.0.3 196 4/18/2023
2.0.2 209 4/18/2023 2.0.2 is deprecated because it is no longer maintained.
2.0.1 202 4/18/2023 2.0.1 is deprecated because it is no longer maintained.
2.0.0 218 4/17/2023 2.0.0 is deprecated because it is no longer maintained.