ThrottlingTroll 2.0.0

There is a newer version of this package available.
See the version list below for details.
dotnet add package ThrottlingTroll --version 2.0.0
NuGet\Install-Package ThrottlingTroll -Version 2.0.0
This command is intended to be used within the Package Manager Console in Visual Studio, as it uses the NuGet module's version of Install-Package.
<PackageReference Include="ThrottlingTroll" Version="2.0.0" />
For projects that support PackageReference, copy this XML node into the project file to reference the package.
paket add ThrottlingTroll --version 2.0.0
#r "nuget: ThrottlingTroll, 2.0.0"
#r directive can be used in F# Interactive and Polyglot Notebooks. Copy this into the interactive tool or source code of the script to reference the package.
// Install ThrottlingTroll as a Cake Addin
#addin nuget:?package=ThrottlingTroll&version=2.0.0

// Install ThrottlingTroll as a Cake Tool
#tool nuget:?package=ThrottlingTroll&version=2.0.0

ThrottlingTroll

Yet another take on rate limiting/throttling in ASP.NET.

Install it from NuGet:

dotnet add package ThrottlingTroll

Features

  • Ingress throttling, aka let your service automatically respond with 429 TooManyRequests to some obtrusive clients. Implemented as an ASP.NET Core Middleware.
  • Egress throttling, aka limit the number of calls your code is making against some external endpoint. Implemented as an HttpClient DelegatingHandler, which produces 429 TooManyRequests response (without making the actual call) when a limit is exceeded.
  • Storing rate counters in a distributed cache, making your throttling policy consistent across all your computing instances. Both Microsoft.Extensions.Caching.Distributed.IDistributedCache and StackExchange.Redis are supported.
  • Propagating 429 TooManyRequests from egress to ingress, aka when your service internally makes an HTTP request which results in 429 TooManyRequests, your service can automatically respond with same 429 TooManyRequests to its calling client.
  • Dynamically configuring rate limits, so that those limits can be adjusted on-the-go, without restarting the service.
  • Custom response fabrics. For ingress it gives full control on what to return when a request is being throttled, and also allows to implement delayed responses (instead of just returning 429 TooManyRequests). For egress it also allows ThrottlingTroll to do automatic retries for you.

How to configure

Quick example of an ingress config setting:

  "ThrottlingTrollIngress": {
    "Rules": [
      {
        "UriPattern": "/api/values",
        "RateLimit": {
          "Algorithm": "FixedWindow",
          "PermitLimit": 5,
          "IntervalInSeconds": 10
        }
      }
    ]
  }

ThrottlingTroll's configuration (both for ingress and egress) is represented by ThrottlingTrollConfig class. It contains a list of rate limiting Rules and some other settings and it can be configured:

  • Statically, via appsettings.json.
  • Programmatically, at service startup.
  • Dynamically, by providing a callback, which ThrottlingTroll will periodically call to get updated values.

See more examples for all of these options below.

Configuring rules and limits

Each Rule defines a pattern that HTTP requests should match. A pattern can include the following properties (all are optional):

  • UriPattern - a Regex pattern to match request URI against. Empty string or null means any URI. Note that this value is treated as a Regex, so symbols that have special meaning in Regex language must be escaped (e.g. to match a query string specify \\?abc=123 instead of ?abc=123).
  • Method - request's HTTP method. E.g. POST. Empty string or null means any method.
  • HeaderName - request's HTTP header to check. If specified, the rule will only apply to requests with this header set to HeaderValue.
  • HeaderValue - value for HTTP header identified by HeaderName. The rule will only apply to requests with that header set to this value. If HeaderName is specified and HeaderValue is not - that matches requests with any value in that header.
  • IdentityId - request's custom Identity ID. If specified, the rule will only apply to requests with this Identity ID. Along with IdentityId you will also need to provide a custom identity extraction routine via IdentityIdExtractor setting.
  • IdentityIdExtractor - a routine that takes a request object and should return the caller's identityId. Typically what serves as identityId is some api-key (taken from e.g. a header), or client's IP address, or the oid claim of a JWT token, or something similar. When IdentityIdExtractor is specified and IdentityId is not - different identityIds will automatically get their own rate counters, one counter per each unique identityId.

If any of the above properties is empty or not specified, this means matching any request.

Then each Rule must specify the rate limiting algorithm to be applied for matched requests and its parameters.

The following algorithms are currently supported:

  • FixedWindow. No more than PermitLimit requests are allowed in IntervalInSeconds. Example:
  "ThrottlingTrollIngress": {
    "Rules": [
      {
        "RateLimit": {
          "Algorithm": "FixedWindow",
          "PermitLimit": 5,
          "IntervalInSeconds": 10
        }
      }
    ]
  }
  • SlidingWindow. No more than PermitLimit requests are allowed in IntervalInSeconds, but that interval is split into NumOfBuckets. The main benefit of this algorithm over FixedWindow is that if a client constantly exceedes PermitLimit, it will never get any valid response and will always get 429 TooManyRequests. Example:
  "ThrottlingTrollIngress": {
    "Rules": [
      {
        "RateLimit": {
          "Algorithm": "SlidingWindow",
          "PermitLimit": 5,
          "IntervalInSeconds": 15,
          "NumOfBuckets": 3
        }
      }
    ]
  }

Configuring whitelist

Requests that should be whitelisted (exempt from the above Rules) can be specified via WhiteList property:

  "ThrottlingTrollIngress": {

    "WhiteList": [
      "/api/healthcheck",
      "api-key=my-unlimited-api-key"
    ]
  },

Specifying UniqueName property when sharing a distributed cache instance

When using the same instance of a distributed cache for multiple different services you might also need to specify a value for UniqueName property:

  "ThrottlingTrollIngress": {

    "UniqueName": "MyThrottledService123"
  }

That value will be used as a prefix for cache keys, so that those multiple services do not corrupt each other's rate counters.

How to use for Ingress Throttling

To configure via appsettings.json

  1. Add the following ThrottlingTrollIngress section to your config file:
  "ThrottlingTrollIngress": {

    "Rules": [
        ... here go rate limiting rules and limits...
    ],

    "WhiteList": [
        ... here go whitelisted URIs...
    ]
  }

  1. Add the following call to your startup code:
app.UseThrottlingTroll();

To configure programmatically

Use the following configuration method at startup:

app.UseThrottlingTroll(options =>
{
    options.Config = new ThrottlingTrollConfig
    {
        Rules = new[]
        {
            new ThrottlingTrollRule
            {
                UriPattern = "/api/values",
                LimitMethod = new SlidingWindowRateLimitMethod
                {
                    PermitLimit = 5,
                    IntervalInSeconds = 10,
                    NumOfBuckets = 5
                }
            },

            // add more rules here...
        }
    };
});

To configure dynamically

Specify a callback that loads rate limits from some shared persistent storage and a time interval to periodically call it:

app.UseThrottlingTroll(options =>
{
    options.GetConfigFunc = async () =>
    {
        var myThrottlingRules = await LoadThrottlingRulesFromDatabase();

        return new ThrottlingTrollConfig
        {
            Rules = myThrottlingRules
        };
    };

    options.IntervalToReloadConfigInSeconds = 10;
});

NOTE: if your callback throws an exception, ThrottlingTroll will get suspended (will not apply any rules) until the callback succeeds again.

To limit clients based on their identityId

If you have a custom way to identify your clients and you want to limit their calls individually, specify a custom IdentityIdExtractor routine like this:

app.UseThrottlingTroll(options =>
{
    options.Config = new ThrottlingTrollConfig
    {
        Rules = new[]
        {
            new ThrottlingTrollRule
            {
                LimitMethod = new FixedWindowRateLimitMethod
                {
                    PermitLimit = 3,
                    IntervalInSeconds = 15
                },

                IdentityIdExtractor = (request) =>
                {
                    // Identifying clients e.g. by their api-key
                    return ((IIncomingHttpRequestProxy)request).Request.Query["api-key"]
                }
            }
        }                    
    };
});

Then ThrottlingTroll will count their requests on a per-identity basis.

To customize responses with a custom response fabric

Provide a response fabric implementation via ResponseFabric option:

app.UseThrottlingTroll(options =>
{
    // Custom response fabric, returns 400 BadRequest + some custom content
    options.ResponseFabric = async (limitExceededResult, requestProxy, responseProxy, requestAborted) => 
    {
        responseProxy.StatusCode = StatusCodes.Status400BadRequest;

        responseProxy.SetHttpHeader(HeaderNames.RetryAfter, limitExceededResult.RetryAfterHeaderValue);

        await responseProxy.WriteAsync("Too many requests. Try again later.");
    };
});

To delay responses instead of returning errors

Provide a response fabric implementation with a delay in it. Also set ShouldContinueAsNormal to true (this will make ThrottlingTroll do the normal request processing instead of shortcutting to a 429 status) :

app.UseThrottlingTroll(options =>
{
    // Custom response fabric, impedes the normal response for 3 seconds
    options.ResponseFabric = async (limitExceededResult, requestProxy, responseProxy, requestAborted) =>
    {
        await Task.Delay(TimeSpan.FromSeconds(3));

        var ingressResponse = (IIngressHttpResponseProxy)responseProxy;
        ingressResponse.ShouldContinueAsNormal = true;
    };
});

How to use for Egress Throttling

To configure via appsettings.json

  1. Add the following ThrottlingTrollEgress section to your config file:
  "ThrottlingTrollEgress": {

    "Rules": [
        ... here go rate limiting rules and limits...
    ],

    "WhiteList": [
        ... here go whitelisted URIs...
    ]
    
  },

  1. Use the following code to configure a named HttpClient at startup:
builder.Services.AddHttpClient("my-throttled-httpclient").AddThrottlingTrollMessageHandler();
  1. Get an instance of that HttpClient via IHttpClientFactory:
var throttledHttpClient = this._httpClientFactory.CreateClient("my-throttled-httpclient");

To configure programmatically

Create an HttpClient instance like this:

var myThrottledHttpClient = new HttpClient
(
    new ThrottlingTrollHandler
    (
        new ThrottlingTrollEgressConfig
        {
            Rules = new[]
            {
                new ThrottlingTrollRule
                {
                    UriPattern = "/some/external/url",
                    LimitMethod = new SlidingWindowRateLimitMethod
                    {
                        PermitLimit = 5,
                        IntervalInSeconds = 10,
                        NumOfBuckets = 5
                    }
                },
            }
        }
    )
);

NOTE: normally HttpClient instances should be created once and reused.

To configure dynamically

  1. Use the following code to configure a named HttpClient at startup:
builder.Services.AddHttpClient("my-throttled-httpclient").AddThrottlingTrollMessageHandler(options =>
{
    options.GetConfigFunc = async () =>
    {
        var myThrottlingRules = await LoadThrottlingRulesFromDatabase();

        return new ThrottlingTrollConfig
        {
            Rules = myThrottlingRules
        };
    };

    options.IntervalToReloadConfigInSeconds = 10;
});
  1. Get an instance of that HttpClient via IHttpClientFactory:
var throttledHttpClient = this._httpClientFactory.CreateClient("my-throttled-httpclient");

To propagate from Egress to Ingress

If your service internally makes HTTP requests and you want to automatically propagate 429 TooManyRequests responses up to your service's clients, configure your ThrottlingTroll-enabled HttpClient with PropagateToIngress property set to true:

  "ThrottlingTrollEgress": {

    "PropagateToIngress": true
  }

This will make ThrottlingTrollHandler throw a dedicated ThrottlingTrollTooManyRequestsException, which then will be handled by ThrottlingTrollMiddleware. The Retry-After header value (if present) will also be propagated.

To use with RestSharp

Configure RestClient as follows:

var restClientOptions = new RestClientOptions("https://contoso.com/api/values")
{
    ConfigureMessageHandler = unused =>

        new ThrottlingTrollHandler
        (
            new ThrottlingTrollEgressConfig
            {
                Rules = new []
                {
                    new ThrottlingTrollRule
                    {
                        LimitMethod = new FixedWindowRateLimitMethod
                        {
                            PermitLimit = 3,
                            IntervalInSeconds = 10,
                        }
                    }
                }

            }
        )
};

var restClient = new RestClient(restClientOptions);

and then use it as normal.

To make HttpClient do automatic retries when getting 429 TooManyRequests

Provide a response fabric implementation via ResponseFabric option and set ShouldRetry to true in it:

builder.Services.AddHttpClient("my-retrying-httpclient").AddThrottlingTrollMessageHandler(options =>
{
    options.ResponseFabric = async (limitExceededResult, requestProxy, responseProxy, cancelToken) =>
    {
        // Doing no more than 10 automatic retries
        var egressResponse = (IEgressHttpResponseProxy)httpResponseProxy;
        egressResponse.ShouldRetry = egressResponse.RetryCount < 10;
    };
});

HttpClient will then first wait for the amount of time suggested by Retry-After response header and then re-send the request. This will happen no matter whether 429 TooManyRequests status was returned by the external resource or it was the egress rate limit that got exceeded.

Supported Rate Counter Stores

By default ThrottlingTroll will store rate counters in memory, using MemoryCacheCounterStore (which internally uses System.Runtime.Caching.MemoryCache).

Other supported options are:

You can also create your custom Counter Store by implementing the ICounterStore interface.

How to specify a Rate Counter Store to be used

Either put a desired ICounterStore implementation into DI container:

builder.Services.AddSingleton<ICounterStore>(
    provider => new DistributedCacheCounterStore(provider.GetRequiredService<IDistributedCache>())
);

Or provide it via UseThrottlingTroll() method:

app.UseThrottlingTroll(options =>
{
    options.CounterStore = new RedisCounterStore(app.Services.GetRequiredService<IConnectionMultiplexer>());
});

Samples

Here is a sample project, that demonstrates all the above concepts.

Product Compatible and additional computed target framework versions.
.NET net6.0 is compatible.  net6.0-android was computed.  net6.0-ios was computed.  net6.0-maccatalyst was computed.  net6.0-macos was computed.  net6.0-tvos was computed.  net6.0-windows was computed.  net7.0 was computed.  net7.0-android was computed.  net7.0-ios was computed.  net7.0-maccatalyst was computed.  net7.0-macos was computed.  net7.0-tvos was computed.  net7.0-windows was computed.  net8.0 was computed.  net8.0-android was computed.  net8.0-browser was computed.  net8.0-ios was computed.  net8.0-maccatalyst was computed.  net8.0-macos was computed.  net8.0-tvos was computed.  net8.0-windows was computed. 
Compatible target framework(s)
Included target framework(s) (in package)
Learn more about Target Frameworks and .NET Standard.

NuGet packages

This package is not used by any NuGet packages.

GitHub repositories

This package is not used by any popular GitHub repositories.

Version Downloads Last updated
6.1.2 127 4/5/2024
6.1.0 448 1/28/2024
6.1.0-beta1 170 1/28/2024
6.0.0 538 12/4/2023
6.0.0-beta6 310 11/26/2023
6.0.0-beta5 278 11/26/2023
6.0.0-beta4 275 11/26/2023
6.0.0-beta3 270 11/26/2023
6.0.0-beta2 258 11/26/2023
6.0.0-beta1 284 11/25/2023
5.0.0 703 11/2/2023
4.0.5 832 8/27/2023
4.0.4 574 7/22/2023
4.0.2 483 7/22/2023
4.0.1 493 7/22/2023
4.0.0 483 7/22/2023
3.0.4 971 5/17/2023
3.0.3 465 5/15/2023
3.0.2 472 5/14/2023
3.0.1 483 5/14/2023
3.0.0 474 5/13/2023
2.0.0 920 4/1/2023
1.2.0 603 2/21/2023
1.1.0 4,781 1/22/2023
1.0.0 583 1/18/2023