// GoToSocial // Copyright (C) GoToSocial Authors admin@gotosocial.org // SPDX-License-Identifier: AGPL-3.0-or-later // // This program is free software: you can redistribute it and/or modify // it under the terms of the GNU Affero General Public License as published by // the Free Software Foundation, either version 3 of the License, or // (at your option) any later version. // // This program is distributed in the hope that it will be useful, // but WITHOUT ANY WARRANTY; without even the implied warranty of // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the // GNU Affero General Public License for more details. // // You should have received a copy of the GNU Affero General Public License // along with this program. If not, see . /* The code in this file is adapted from MIT-licensed code in github.com/go-chi/chi. Thanks chi (thi)! See: https://github.com/go-chi/chi/blob/e6baba61759b26ddf7b14d1e02d1da81a4d76c08/middleware/throttle.go And: https://github.com/sponsors/pkieltyka */ package middleware import ( "net/http" "runtime" "strconv" "time" "github.com/gin-gonic/gin" ) // token represents a request that is being processed. type token struct{} // Throttle returns a gin middleware that performs throttling of incoming requests, // ensuring that only a certain number of requests are handled concurrently, to reduce // congestion of the server. // // Limits are configured using available CPUs and the given cpuMultiplier value. // Open request limit is available CPUs * multiplier; backlog limit is limit * multiplier. // // Example values for multiplier 8: // // 1 cpu = 08 open, 064 backlog // 2 cpu = 16 open, 128 backlog // 4 cpu = 32 open, 256 backlog // // Example values for multiplier 4: // // 1 cpu = 04 open, 016 backlog // 2 cpu = 08 open, 032 backlog // 4 cpu = 16 open, 064 backlog // // Callers will first attempt to get a backlog token. Once they have that, they will // wait in the backlog queue until they can get a token to allow their request to be // processed. // // If the backlog queue is full, the request context is closed, or the caller has been // waiting in the backlog for too long, this function will abort the request chain, // write a JSON error into the response, set an appropriate Retry-After value, and set // the HTTP response code to 503: Service Unavailable. // // If the multiplier is <= 0, a noop middleware will be returned instead. // // RetryAfter determines the Retry-After header value to be sent to throttled requests. // // Useful links: // // - https://developer.mozilla.org/en-US/docs/Web/HTTP/Headers/Retry-After // - https://developer.mozilla.org/en-US/docs/Web/HTTP/Status/503 func Throttle(cpuMultiplier int, retryAfter time.Duration) gin.HandlerFunc { if cpuMultiplier <= 0 { // throttling is disabled, return a noop middleware return func(c *gin.Context) {} } var ( limit = runtime.GOMAXPROCS(0) * cpuMultiplier backlogLimit = limit * cpuMultiplier backlogChannelSize = limit + backlogLimit tokens = make(chan token, limit) backlogTokens = make(chan token, backlogChannelSize) retryAfterStr = strconv.FormatUint(uint64(retryAfter/time.Second), 10) ) // prefill token channels for i := 0; i < limit; i++ { tokens <- token{} } for i := 0; i < backlogChannelSize; i++ { backlogTokens <- token{} } return func(c *gin.Context) { // inside this select, the caller tries to get a backlog token select { case <-c.Request.Context().Done(): // request context has been canceled already return case btok := <-backlogTokens: defer func() { // when we're finished, return the backlog token to the bucket backlogTokens <- btok }() // inside *this* select, the caller has a backlog token, // and they're waiting for their turn to be processed select { case <-c.Request.Context().Done(): // the request context has been canceled already return case tok := <-tokens: // the caller gets a token, so their request can now be processed defer func() { // whatever happens to the request, put the // token back in the bucket when we're finished tokens <- tok }() c.Next() // <- finally process the caller's request } default: // we don't have space in the backlog queue c.Header("Retry-After", retryAfterStr) c.JSON(http.StatusTooManyRequests, gin.H{"error": "server capacity exceeded"}) c.Abort() } } }