mirror of
https://github.com/superseriousbusiness/gotosocial.git
synced 2024-11-27 19:01:01 +00:00
[performance] massively improved ActivityPub delivery worker efficiency (#2812)
* add delivery worker type that pulls from queue to httpclient package * finish up some code commenting, bodge a vendored activity library change, integrate the deliverypool changes into transportcontroller * hook up queue deletion logic * support deleting queued http requests by target ID * don't index APRequest by hostname in the queue * use gorun * use the original context's values when wrapping msg type as delivery{} * actually log in the AP delivery worker ... * add uncommitted changes * use errors.AsV2() * use errorsv2.AsV2() * finish adding some code comments, add bad host handling to delivery workers * slightly tweak deliveryworkerpool API, use advanced sender multiplier * remove PopCtx() method, let others instead rely on Wait() * shuffle things around to move delivery stuff into transport/ subpkg * remove dead code * formatting * validate request before queueing for delivery * finish adding code comments, fix up backoff code * finish adding more code comments * clamp minimum no. senders to 1 * add start/stop logging to delivery worker, some slight changes * remove double logging * use worker ptrs * expose the embedded log fields in httpclient.Request{} * ensure request context values are preserved when updating ctx * add delivery worker tests * fix linter issues * ensure delivery worker gets inited in testrig * fix tests to delivering messages to check worker delivery queue * update error type to use ptr instead of value receiver * fix test calling Workers{}.Start() instead of testrig.StartWorkers() * update docs for advanced-sender-multiplier * update to the latest activity library version * add comment about not using httptest.Server{}
This commit is contained in:
parent
15733cddb2
commit
a483bd9e38
32 changed files with 1285 additions and 374 deletions
|
@ -124,6 +124,9 @@ var Start action.GTSAction = func(ctx context.Context) error {
|
||||||
TLSInsecureSkipVerify: config.GetHTTPClientTLSInsecureSkipVerify(),
|
TLSInsecureSkipVerify: config.GetHTTPClientTLSInsecureSkipVerify(),
|
||||||
})
|
})
|
||||||
|
|
||||||
|
// Initialize delivery worker with http client.
|
||||||
|
state.Workers.Delivery.Init(client)
|
||||||
|
|
||||||
// Initialize workers.
|
// Initialize workers.
|
||||||
state.Workers.Start()
|
state.Workers.Start()
|
||||||
defer state.Workers.Stop()
|
defer state.Workers.Stop()
|
||||||
|
|
|
@ -119,15 +119,10 @@ advanced-throttling-multiplier: 8
|
||||||
# Default: "30s"
|
# Default: "30s"
|
||||||
advanced-throttling-retry-after: "30s"
|
advanced-throttling-retry-after: "30s"
|
||||||
|
|
||||||
# Int. CPU multiplier for the amount of goroutines to spawn in order to send messages via ActivityPub.
|
# Int. CPU multiplier for the fixed number of goroutines to spawn in order to send messages via ActivityPub.
|
||||||
# Messages will be batched so that at most multiplier * CPU count messages will be sent out at once.
|
# Messages will be batched and pushed to a singular queue, from which multiplier * CPU count goroutines will
|
||||||
# This can be tuned to limit concurrent POSTing to remote inboxes, preventing your instance CPU
|
# pull and attempt deliveries. This can be tuned to limit concurrent posting to remote inboxes, preventing
|
||||||
# usage from skyrocketing when an account with many followers posts a new status.
|
# your instance CPU usage skyrocketing when accounts with many followers post statuses.
|
||||||
#
|
|
||||||
# Messages are split among available senders, and each sender processes its assigned messages in serial.
|
|
||||||
# For example, say a user with 1000 followers is on an instance with 2 CPUs. With the default multiplier
|
|
||||||
# of 2, this means 4 senders would be in process at once on this instance. When the user creates a new post,
|
|
||||||
# each sender would end up iterating through about 250 Create messages + delivering them to remote instances.
|
|
||||||
#
|
#
|
||||||
# If you set this to 0 or less, only 1 sender will be used regardless of CPU count. This may be
|
# If you set this to 0 or less, only 1 sender will be used regardless of CPU count. This may be
|
||||||
# useful in cases where you are working with very tight network or CPU constraints.
|
# useful in cases where you are working with very tight network or CPU constraints.
|
||||||
|
|
|
@ -1042,15 +1042,10 @@ advanced-throttling-multiplier: 8
|
||||||
# Default: "30s"
|
# Default: "30s"
|
||||||
advanced-throttling-retry-after: "30s"
|
advanced-throttling-retry-after: "30s"
|
||||||
|
|
||||||
# Int. CPU multiplier for the amount of goroutines to spawn in order to send messages via ActivityPub.
|
# Int. CPU multiplier for the fixed number of goroutines to spawn in order to send messages via ActivityPub.
|
||||||
# Messages will be batched so that at most multiplier * CPU count messages will be sent out at once.
|
# Messages will be batched and pushed to a singular queue, from which multiplier * CPU count goroutines will
|
||||||
# This can be tuned to limit concurrent POSTing to remote inboxes, preventing your instance CPU
|
# pull and attempt deliveries. This can be tuned to limit concurrent posting to remote inboxes, preventing
|
||||||
# usage from skyrocketing when an account with many followers posts a new status.
|
# your instance CPU usage skyrocketing when accounts with many followers post statuses.
|
||||||
#
|
|
||||||
# Messages are split among available senders, and each sender processes its assigned messages in serial.
|
|
||||||
# For example, say a user with 1000 followers is on an instance with 2 CPUs. With the default multiplier
|
|
||||||
# of 2, this means 4 senders would be in process at once on this instance. When the user creates a new post,
|
|
||||||
# each sender would end up iterating through about 250 Create messages + delivering them to remote instances.
|
|
||||||
#
|
#
|
||||||
# If you set this to 0 or less, only 1 sender will be used regardless of CPU count. This may be
|
# If you set this to 0 or less, only 1 sender will be used regardless of CPU count. This may be
|
||||||
# useful in cases where you are working with very tight network or CPU constraints.
|
# useful in cases where you are working with very tight network or CPU constraints.
|
||||||
|
|
2
go.mod
2
go.mod
|
@ -49,7 +49,7 @@ require (
|
||||||
github.com/spf13/cobra v1.8.0
|
github.com/spf13/cobra v1.8.0
|
||||||
github.com/spf13/viper v1.18.2
|
github.com/spf13/viper v1.18.2
|
||||||
github.com/stretchr/testify v1.9.0
|
github.com/stretchr/testify v1.9.0
|
||||||
github.com/superseriousbusiness/activity v1.6.0-gts.0.20240221151241-5d56c04088d4
|
github.com/superseriousbusiness/activity v1.6.0-gts.0.20240408131430-247f7f7110f0
|
||||||
github.com/superseriousbusiness/httpsig v1.2.0-SSB
|
github.com/superseriousbusiness/httpsig v1.2.0-SSB
|
||||||
github.com/superseriousbusiness/oauth2/v4 v4.3.2-SSB.0.20230227143000-f4900831d6c8
|
github.com/superseriousbusiness/oauth2/v4 v4.3.2-SSB.0.20230227143000-f4900831d6c8
|
||||||
github.com/tdewolff/minify/v2 v2.20.19
|
github.com/tdewolff/minify/v2 v2.20.19
|
||||||
|
|
4
go.sum
4
go.sum
|
@ -622,8 +622,8 @@ github.com/stretchr/testify v1.9.0/go.mod h1:r2ic/lqez/lEtzL7wO/rwa5dbSLXVDPFyf8
|
||||||
github.com/subosito/gotenv v1.6.0 h1:9NlTDc1FTs4qu0DDq7AEtTPNw6SVm7uBMsUCUjABIf8=
|
github.com/subosito/gotenv v1.6.0 h1:9NlTDc1FTs4qu0DDq7AEtTPNw6SVm7uBMsUCUjABIf8=
|
||||||
github.com/subosito/gotenv v1.6.0/go.mod h1:Dk4QP5c2W3ibzajGcXpNraDfq2IrhjMIvMSWPKKo0FU=
|
github.com/subosito/gotenv v1.6.0/go.mod h1:Dk4QP5c2W3ibzajGcXpNraDfq2IrhjMIvMSWPKKo0FU=
|
||||||
github.com/sunfish-shogi/bufseekio v0.0.0-20210207115823-a4185644b365/go.mod h1:dEzdXgvImkQ3WLI+0KQpmEx8T/C/ma9KeS3AfmU899I=
|
github.com/sunfish-shogi/bufseekio v0.0.0-20210207115823-a4185644b365/go.mod h1:dEzdXgvImkQ3WLI+0KQpmEx8T/C/ma9KeS3AfmU899I=
|
||||||
github.com/superseriousbusiness/activity v1.6.0-gts.0.20240221151241-5d56c04088d4 h1:kPjQR/hVZtROTzkxptp/EIR7Wm58O8jppwpCFrZ7sVU=
|
github.com/superseriousbusiness/activity v1.6.0-gts.0.20240408131430-247f7f7110f0 h1:zPdbgwbjPxrJqme2sFTMQoML5ukNWRhChOnilR47rss=
|
||||||
github.com/superseriousbusiness/activity v1.6.0-gts.0.20240221151241-5d56c04088d4/go.mod h1:AZw0Xb4Oju8rmaJCZ21gc5CPg47MmNgyac+Hx5jo8VM=
|
github.com/superseriousbusiness/activity v1.6.0-gts.0.20240408131430-247f7f7110f0/go.mod h1:AZw0Xb4Oju8rmaJCZ21gc5CPg47MmNgyac+Hx5jo8VM=
|
||||||
github.com/superseriousbusiness/go-jpeg-image-structure/v2 v2.0.0-20220321154430-d89a106fdabe h1:ksl2oCx/Qo8sNDc3Grb8WGKBM9nkvhCm25uvlT86azE=
|
github.com/superseriousbusiness/go-jpeg-image-structure/v2 v2.0.0-20220321154430-d89a106fdabe h1:ksl2oCx/Qo8sNDc3Grb8WGKBM9nkvhCm25uvlT86azE=
|
||||||
github.com/superseriousbusiness/go-jpeg-image-structure/v2 v2.0.0-20220321154430-d89a106fdabe/go.mod h1:gH4P6gN1V+wmIw5o97KGaa1RgXB/tVpC2UNzijhg3E4=
|
github.com/superseriousbusiness/go-jpeg-image-structure/v2 v2.0.0-20220321154430-d89a106fdabe/go.mod h1:gH4P6gN1V+wmIw5o97KGaa1RgXB/tVpC2UNzijhg3E4=
|
||||||
github.com/superseriousbusiness/go-png-image-structure/v2 v2.0.1-SSB h1:8psprYSK1KdOSH7yQ4PbJq0YYaGQY+gzdW/B0ExDb/8=
|
github.com/superseriousbusiness/go-png-image-structure/v2 v2.0.1-SSB h1:8psprYSK1KdOSH7yQ4PbJq0YYaGQY+gzdW/B0ExDb/8=
|
||||||
|
|
|
@ -18,13 +18,14 @@
|
||||||
package users
|
package users
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"errors"
|
|
||||||
"net/http"
|
"net/http"
|
||||||
|
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
apiutil "github.com/superseriousbusiness/gotosocial/internal/api/util"
|
apiutil "github.com/superseriousbusiness/gotosocial/internal/api/util"
|
||||||
"github.com/superseriousbusiness/gotosocial/internal/gtserror"
|
"github.com/superseriousbusiness/gotosocial/internal/gtserror"
|
||||||
"github.com/superseriousbusiness/gotosocial/internal/log"
|
"github.com/superseriousbusiness/gotosocial/internal/log"
|
||||||
|
|
||||||
|
errorsv2 "codeberg.org/gruf/go-errors/v2"
|
||||||
)
|
)
|
||||||
|
|
||||||
// InboxPOSTHandler deals with incoming POST requests to an actor's inbox.
|
// InboxPOSTHandler deals with incoming POST requests to an actor's inbox.
|
||||||
|
@ -32,18 +33,18 @@ import (
|
||||||
func (m *Module) InboxPOSTHandler(c *gin.Context) {
|
func (m *Module) InboxPOSTHandler(c *gin.Context) {
|
||||||
_, err := m.processor.Fedi().InboxPost(c.Request.Context(), c.Writer, c.Request)
|
_, err := m.processor.Fedi().InboxPost(c.Request.Context(), c.Writer, c.Request)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
errWithCode := new(gtserror.WithCode)
|
errWithCode := errorsv2.AsV2[gtserror.WithCode](err)
|
||||||
|
|
||||||
if !errors.As(err, errWithCode) {
|
if errWithCode == nil {
|
||||||
// Something else went wrong, and someone forgot to return
|
// Something else went wrong, and someone forgot to return
|
||||||
// an errWithCode! It's chill though. Log the error but don't
|
// an errWithCode! It's chill though. Log the error but don't
|
||||||
// return it as-is to the caller, to avoid leaking internals.
|
// return it as-is to the caller, to avoid leaking internals.
|
||||||
log.Errorf(c.Request.Context(), "returning Bad Request to caller, err was: %q", err)
|
log.Errorf(c.Request.Context(), "returning Bad Request to caller, err was: %q", err)
|
||||||
*errWithCode = gtserror.NewErrorBadRequest(err)
|
errWithCode = gtserror.NewErrorBadRequest(err)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Pass along confirmed error with code to the main error handler
|
// Pass along confirmed error with code to the main error handler
|
||||||
apiutil.ErrorHandler(c, *errWithCode, m.processor.InstanceGetV1)
|
apiutil.ErrorHandler(c, errWithCode, m.processor.InstanceGetV1)
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
1
internal/cache/cache.go
vendored
1
internal/cache/cache.go
vendored
|
@ -25,6 +25,7 @@ import (
|
||||||
)
|
)
|
||||||
|
|
||||||
type Caches struct {
|
type Caches struct {
|
||||||
|
|
||||||
// GTS provides access to the collection of
|
// GTS provides access to the collection of
|
||||||
// gtsmodel object caches. (used by the database).
|
// gtsmodel object caches. (used by the database).
|
||||||
GTS GTSCaches
|
GTS GTSCaches
|
||||||
|
|
|
@ -89,7 +89,7 @@ func (f *federatingActor) PostInboxScheme(ctx context.Context, w http.ResponseWr
|
||||||
// so we specifically have to check for already wrapped with code.
|
// so we specifically have to check for already wrapped with code.
|
||||||
//
|
//
|
||||||
ctx, authenticated, err := f.sideEffectActor.AuthenticatePostInbox(ctx, w, r)
|
ctx, authenticated, err := f.sideEffectActor.AuthenticatePostInbox(ctx, w, r)
|
||||||
if errors.As(err, new(gtserror.WithCode)) {
|
if errorsv2.AsV2[gtserror.WithCode](err) != nil {
|
||||||
// If it was already wrapped with an
|
// If it was already wrapped with an
|
||||||
// HTTP code then don't bother rewrapping
|
// HTTP code then don't bother rewrapping
|
||||||
// it, just return it as-is for caller to
|
// it, just return it as-is for caller to
|
||||||
|
@ -131,7 +131,7 @@ func (f *federatingActor) PostInboxScheme(ctx context.Context, w http.ResponseWr
|
||||||
// Check authorization of the activity; this will include blocks.
|
// Check authorization of the activity; this will include blocks.
|
||||||
authorized, err := f.sideEffectActor.AuthorizePostInbox(ctx, w, activity)
|
authorized, err := f.sideEffectActor.AuthorizePostInbox(ctx, w, activity)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
if errors.As(err, new(errOtherIRIBlocked)) {
|
if errorsv2.AsV2[*errOtherIRIBlocked](err) != nil {
|
||||||
// There's no direct block between requester(s) and
|
// There's no direct block between requester(s) and
|
||||||
// receiver. However, one or more of the other IRIs
|
// receiver. However, one or more of the other IRIs
|
||||||
// involved in the request (account replied to, note
|
// involved in the request (account replied to, note
|
||||||
|
@ -139,7 +139,7 @@ func (f *federatingActor) PostInboxScheme(ctx context.Context, w http.ResponseWr
|
||||||
// by the receiver. We don't need to return 403 here,
|
// by the receiver. We don't need to return 403 here,
|
||||||
// instead, just return 202 accepted but don't do any
|
// instead, just return 202 accepted but don't do any
|
||||||
// further processing of the activity.
|
// further processing of the activity.
|
||||||
return true, nil
|
return true, nil //nolint
|
||||||
}
|
}
|
||||||
|
|
||||||
// Real error has occurred.
|
// Real error has occurred.
|
||||||
|
|
|
@ -21,6 +21,7 @@ import (
|
||||||
"bytes"
|
"bytes"
|
||||||
"context"
|
"context"
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
|
"io"
|
||||||
"net/url"
|
"net/url"
|
||||||
"testing"
|
"testing"
|
||||||
"time"
|
"time"
|
||||||
|
@ -129,23 +130,27 @@ func (suite *FederatingActorTestSuite) TestSendRemoteFollower() {
|
||||||
suite.NotNil(activity)
|
suite.NotNil(activity)
|
||||||
|
|
||||||
// because we added 1 remote follower for zork, there should be a url in sentMessage
|
// because we added 1 remote follower for zork, there should be a url in sentMessage
|
||||||
var sent [][]byte
|
var sent []byte
|
||||||
if !testrig.WaitFor(func() bool {
|
if !testrig.WaitFor(func() bool {
|
||||||
sentI, ok := httpClient.SentMessages.Load(*testRemoteAccount.SharedInboxURI)
|
delivery, ok := suite.state.Workers.Delivery.Queue.Pop()
|
||||||
if ok {
|
|
||||||
sent, ok = sentI.([][]byte)
|
|
||||||
if !ok {
|
if !ok {
|
||||||
panic("SentMessages entry was not []byte")
|
return false
|
||||||
|
}
|
||||||
|
if !testrig.EqualRequestURIs(delivery.Request.URL, *testRemoteAccount.SharedInboxURI) {
|
||||||
|
panic("differing request uris")
|
||||||
|
}
|
||||||
|
sent, err = io.ReadAll(delivery.Request.Body)
|
||||||
|
if err != nil {
|
||||||
|
panic("error reading body: " + err.Error())
|
||||||
}
|
}
|
||||||
return true
|
return true
|
||||||
}
|
|
||||||
return false
|
|
||||||
}) {
|
}) {
|
||||||
suite.FailNow("timed out waiting for message")
|
suite.FailNow("timed out waiting for message")
|
||||||
}
|
}
|
||||||
|
|
||||||
dst := new(bytes.Buffer)
|
dst := new(bytes.Buffer)
|
||||||
err = json.Indent(dst, sent[0], "", " ")
|
err = json.Indent(dst, sent, "", " ")
|
||||||
suite.NoError(err)
|
suite.NoError(err)
|
||||||
suite.Equal(`{
|
suite.Equal(`{
|
||||||
"@context": "https://www.w3.org/ns/activitystreams",
|
"@context": "https://www.w3.org/ns/activitystreams",
|
||||||
|
|
|
@ -51,7 +51,7 @@ func (f *federatingDB) Delete(ctx context.Context, id *url.URL) error {
|
||||||
// in a delete we only get the URI, we can't know if we have a status or a profile or something else,
|
// in a delete we only get the URI, we can't know if we have a status or a profile or something else,
|
||||||
// so we have to try a few different things...
|
// so we have to try a few different things...
|
||||||
if s, err := f.state.DB.GetStatusByURI(ctx, id.String()); err == nil && requestingAcct.ID == s.AccountID {
|
if s, err := f.state.DB.GetStatusByURI(ctx, id.String()); err == nil && requestingAcct.ID == s.AccountID {
|
||||||
l.Debugf("uri is for STATUS with id: %s", s.ID)
|
l.Debugf("deleting status: %s", s.ID)
|
||||||
f.state.Workers.EnqueueFediAPI(ctx, messages.FromFediAPI{
|
f.state.Workers.EnqueueFediAPI(ctx, messages.FromFediAPI{
|
||||||
APObjectType: ap.ObjectNote,
|
APObjectType: ap.ObjectNote,
|
||||||
APActivityType: ap.ActivityDelete,
|
APActivityType: ap.ActivityDelete,
|
||||||
|
@ -61,7 +61,7 @@ func (f *federatingDB) Delete(ctx context.Context, id *url.URL) error {
|
||||||
}
|
}
|
||||||
|
|
||||||
if a, err := f.state.DB.GetAccountByURI(ctx, id.String()); err == nil && requestingAcct.ID == a.ID {
|
if a, err := f.state.DB.GetAccountByURI(ctx, id.String()); err == nil && requestingAcct.ID == a.ID {
|
||||||
l.Debugf("uri is for ACCOUNT with id %s", a.ID)
|
l.Debugf("deleting account: %s", a.ID)
|
||||||
f.state.Workers.EnqueueFediAPI(ctx, messages.FromFediAPI{
|
f.state.Workers.EnqueueFediAPI(ctx, messages.FromFediAPI{
|
||||||
APObjectType: ap.ObjectProfile,
|
APObjectType: ap.ObjectProfile,
|
||||||
APActivityType: ap.ActivityDelete,
|
APActivityType: ap.ActivityDelete,
|
||||||
|
|
|
@ -44,7 +44,7 @@ type errOtherIRIBlocked struct {
|
||||||
iriStrs []string
|
iriStrs []string
|
||||||
}
|
}
|
||||||
|
|
||||||
func (e errOtherIRIBlocked) Error() string {
|
func (e *errOtherIRIBlocked) Error() string {
|
||||||
iriStrsNice := "[" + strings.Join(e.iriStrs, ", ") + "]"
|
iriStrsNice := "[" + strings.Join(e.iriStrs, ", ") + "]"
|
||||||
if e.domainBlock {
|
if e.domainBlock {
|
||||||
return "domain block exists for one or more of " + iriStrsNice
|
return "domain block exists for one or more of " + iriStrsNice
|
||||||
|
@ -67,7 +67,7 @@ func newErrOtherIRIBlocked(
|
||||||
e.iriStrs = append(e.iriStrs, iri.String())
|
e.iriStrs = append(e.iriStrs, iri.String())
|
||||||
}
|
}
|
||||||
|
|
||||||
return e
|
return &e
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
|
|
@ -21,13 +21,13 @@ import (
|
||||||
"bytes"
|
"bytes"
|
||||||
"context"
|
"context"
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"errors"
|
|
||||||
"io"
|
"io"
|
||||||
"net/http"
|
"net/http"
|
||||||
"net/http/httptest"
|
"net/http/httptest"
|
||||||
"net/url"
|
"net/url"
|
||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
|
errorsv2 "codeberg.org/gruf/go-errors/v2"
|
||||||
"github.com/stretchr/testify/suite"
|
"github.com/stretchr/testify/suite"
|
||||||
"github.com/superseriousbusiness/gotosocial/internal/ap"
|
"github.com/superseriousbusiness/gotosocial/internal/ap"
|
||||||
"github.com/superseriousbusiness/gotosocial/internal/gtscontext"
|
"github.com/superseriousbusiness/gotosocial/internal/gtscontext"
|
||||||
|
@ -101,8 +101,8 @@ func (suite *FederatingProtocolTestSuite) authenticatePostInbox(
|
||||||
|
|
||||||
recorder := httptest.NewRecorder()
|
recorder := httptest.NewRecorder()
|
||||||
newContext, authed, err := suite.federator.AuthenticatePostInbox(ctx, recorder, request)
|
newContext, authed, err := suite.federator.AuthenticatePostInbox(ctx, recorder, request)
|
||||||
if withCode := new(gtserror.WithCode); (errors.As(err, withCode) &&
|
if withCode := errorsv2.AsV2[gtserror.WithCode](err); // nocollapse
|
||||||
(*withCode).Code() >= 500) || (err != nil && (*withCode) == nil) {
|
(withCode != nil && withCode.Code() >= 500) || (err != nil && withCode == nil) {
|
||||||
// NOTE: the behaviour here is a little strange as we have
|
// NOTE: the behaviour here is a little strange as we have
|
||||||
// the competing code styles of the go-fed interface expecting
|
// the competing code styles of the go-fed interface expecting
|
||||||
// that any err is a no-go, but authed bool is intended to be
|
// that any err is a no-go, but authed bool is intended to be
|
||||||
|
|
|
@ -35,7 +35,6 @@ import (
|
||||||
"codeberg.org/gruf/go-cache/v3"
|
"codeberg.org/gruf/go-cache/v3"
|
||||||
errorsv2 "codeberg.org/gruf/go-errors/v2"
|
errorsv2 "codeberg.org/gruf/go-errors/v2"
|
||||||
"codeberg.org/gruf/go-iotools"
|
"codeberg.org/gruf/go-iotools"
|
||||||
"codeberg.org/gruf/go-kv"
|
|
||||||
"github.com/superseriousbusiness/gotosocial/internal/gtscontext"
|
"github.com/superseriousbusiness/gotosocial/internal/gtscontext"
|
||||||
"github.com/superseriousbusiness/gotosocial/internal/gtserror"
|
"github.com/superseriousbusiness/gotosocial/internal/gtserror"
|
||||||
"github.com/superseriousbusiness/gotosocial/internal/log"
|
"github.com/superseriousbusiness/gotosocial/internal/log"
|
||||||
|
@ -109,11 +108,13 @@ type Client struct {
|
||||||
client http.Client
|
client http.Client
|
||||||
badHosts cache.TTLCache[string, struct{}]
|
badHosts cache.TTLCache[string, struct{}]
|
||||||
bodyMax int64
|
bodyMax int64
|
||||||
|
retries uint
|
||||||
}
|
}
|
||||||
|
|
||||||
// New returns a new instance of Client initialized using configuration.
|
// New returns a new instance of Client initialized using configuration.
|
||||||
func New(cfg Config) *Client {
|
func New(cfg Config) *Client {
|
||||||
var c Client
|
var c Client
|
||||||
|
c.retries = 5
|
||||||
|
|
||||||
d := &net.Dialer{
|
d := &net.Dialer{
|
||||||
Timeout: 15 * time.Second,
|
Timeout: 15 * time.Second,
|
||||||
|
@ -177,7 +178,7 @@ func New(cfg Config) *Client {
|
||||||
}}
|
}}
|
||||||
|
|
||||||
// Initiate outgoing bad hosts lookup cache.
|
// Initiate outgoing bad hosts lookup cache.
|
||||||
c.badHosts = cache.NewTTL[string, struct{}](0, 1000, 0)
|
c.badHosts = cache.NewTTL[string, struct{}](0, 512, 0)
|
||||||
c.badHosts.SetTTL(time.Hour, false)
|
c.badHosts.SetTTL(time.Hour, false)
|
||||||
if !c.badHosts.Start(time.Minute) {
|
if !c.badHosts.Start(time.Minute) {
|
||||||
log.Panic(nil, "failed to start transport controller cache")
|
log.Panic(nil, "failed to start transport controller cache")
|
||||||
|
@ -187,154 +188,184 @@ func New(cfg Config) *Client {
|
||||||
}
|
}
|
||||||
|
|
||||||
// Do will essentially perform http.Client{}.Do() with retry-backoff functionality.
|
// Do will essentially perform http.Client{}.Do() with retry-backoff functionality.
|
||||||
func (c *Client) Do(r *http.Request) (*http.Response, error) {
|
func (c *Client) Do(r *http.Request) (rsp *http.Response, err error) {
|
||||||
return c.DoSigned(r, func(r *http.Request) error {
|
|
||||||
return nil // no request signing
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
// DoSigned will essentially perform http.Client{}.Do() with retry-backoff functionality and requesting signing..
|
|
||||||
func (c *Client) DoSigned(r *http.Request, sign SignFunc) (rsp *http.Response, err error) {
|
|
||||||
const (
|
|
||||||
// max no. attempts.
|
|
||||||
maxRetries = 5
|
|
||||||
|
|
||||||
// starting backoff duration.
|
|
||||||
baseBackoff = 2 * time.Second
|
|
||||||
)
|
|
||||||
|
|
||||||
// First validate incoming request.
|
// First validate incoming request.
|
||||||
if err := ValidateRequest(r); err != nil {
|
if err := ValidateRequest(r); err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
// Get request hostname.
|
// Wrap in our own request
|
||||||
host := r.URL.Hostname()
|
// type for retry-backoff.
|
||||||
|
req := WrapRequest(r)
|
||||||
|
|
||||||
// Check whether request should fast fail.
|
if gtscontext.IsFastfail(r.Context()) {
|
||||||
fastFail := gtscontext.IsFastfail(r.Context())
|
// If the fast-fail flag was set, just
|
||||||
if !fastFail {
|
// attempt a single iteration instead of
|
||||||
// Check if recently reached max retries for this host
|
// following the below retry-backoff loop.
|
||||||
// so we don't bother with a retry-backoff loop. The only
|
rsp, _, err = c.DoOnce(&req)
|
||||||
// errors that are retried upon are server failure, TLS
|
|
||||||
// and domain resolution type errors, so this cached result
|
|
||||||
// indicates this server is likely having issues.
|
|
||||||
fastFail = c.badHosts.Has(host)
|
|
||||||
defer func() {
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
// On error return mark as bad-host.
|
return nil, fmt.Errorf("%w (fast fail)", err)
|
||||||
c.badHosts.Set(host, struct{}{})
|
|
||||||
}
|
}
|
||||||
}()
|
|
||||||
}
|
|
||||||
|
|
||||||
// Start a log entry for this request
|
|
||||||
l := log.WithContext(r.Context()).
|
|
||||||
WithFields(kv.Fields{
|
|
||||||
{"method", r.Method},
|
|
||||||
{"url", r.URL.String()},
|
|
||||||
}...)
|
|
||||||
|
|
||||||
for i := 0; i < maxRetries; i++ {
|
|
||||||
var backoff time.Duration
|
|
||||||
|
|
||||||
l.Info("performing request")
|
|
||||||
|
|
||||||
// Perform the request.
|
|
||||||
rsp, err = c.do(r)
|
|
||||||
if err == nil { //nolint:gocritic
|
|
||||||
|
|
||||||
// TooManyRequest means we need to slow
|
|
||||||
// down and retry our request. Codes over
|
|
||||||
// 500 generally indicate temp. outages.
|
|
||||||
if code := rsp.StatusCode; code < 500 &&
|
|
||||||
code != http.StatusTooManyRequests {
|
|
||||||
return rsp, nil
|
return rsp, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Create loggable error from response status code.
|
for {
|
||||||
err = fmt.Errorf(`http response: %s`, rsp.Status)
|
var retry bool
|
||||||
|
|
||||||
// Search for a provided "Retry-After" header value.
|
// Perform the http request.
|
||||||
if after := rsp.Header.Get("Retry-After"); after != "" {
|
rsp, retry, err = c.DoOnce(&req)
|
||||||
|
if err == nil {
|
||||||
// Get current time.
|
return rsp, nil
|
||||||
now := time.Now()
|
|
||||||
|
|
||||||
if u, _ := strconv.ParseUint(after, 10, 32); u != 0 {
|
|
||||||
// An integer number of backoff seconds was provided.
|
|
||||||
backoff = time.Duration(u) * time.Second
|
|
||||||
} else if at, _ := http.ParseTime(after); !at.Before(now) {
|
|
||||||
// An HTTP formatted future date-time was provided.
|
|
||||||
backoff = at.Sub(now)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Don't let their provided backoff exceed our max.
|
if !retry {
|
||||||
if max := baseBackoff * maxRetries; backoff > max {
|
// reached max retries, don't further backoff
|
||||||
backoff = max
|
return nil, fmt.Errorf("%w (max retries)", err)
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Close + unset rsp.
|
// Start new backoff sleep timer.
|
||||||
_ = rsp.Body.Close()
|
backoff := time.NewTimer(req.BackOff())
|
||||||
rsp = nil
|
|
||||||
|
|
||||||
} else if errorsv2.IsV2(err,
|
select {
|
||||||
|
// Request ctx cancelled.
|
||||||
|
case <-r.Context().Done():
|
||||||
|
backoff.Stop()
|
||||||
|
|
||||||
|
// Return context error.
|
||||||
|
err = r.Context().Err()
|
||||||
|
return nil, err
|
||||||
|
|
||||||
|
// Backoff for time.
|
||||||
|
case <-backoff.C:
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// DoOnce wraps an underlying http.Client{}.Do() to perform our wrapped request type:
|
||||||
|
// rewinding response body to permit reuse, signing request data when SignFunc provided,
|
||||||
|
// marking erroring hosts, updating retry attempt counts and setting backoff from header.
|
||||||
|
func (c *Client) DoOnce(r *Request) (rsp *http.Response, retry bool, err error) {
|
||||||
|
if r.attempts > c.retries {
|
||||||
|
// Ensure request hasn't reached max number of attempts.
|
||||||
|
err = fmt.Errorf("httpclient: reached max retries (%d)", c.retries)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// Update no.
|
||||||
|
// attempts.
|
||||||
|
r.attempts++
|
||||||
|
|
||||||
|
// Reset backoff.
|
||||||
|
r.backoff = 0
|
||||||
|
|
||||||
|
// Perform main routine.
|
||||||
|
rsp, retry, err = c.do(r)
|
||||||
|
|
||||||
|
if rsp != nil {
|
||||||
|
// Log successful rsp.
|
||||||
|
r.Entry.Info(rsp.Status)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// Log any errors.
|
||||||
|
r.Entry.Error(err)
|
||||||
|
|
||||||
|
switch {
|
||||||
|
case !retry:
|
||||||
|
// If they were told not to
|
||||||
|
// retry, also set number of
|
||||||
|
// attempts to prevent retry.
|
||||||
|
r.attempts = c.retries + 1
|
||||||
|
|
||||||
|
case r.attempts > c.retries:
|
||||||
|
// On max retries, mark this as
|
||||||
|
// a "badhost", i.e. is erroring.
|
||||||
|
c.badHosts.Set(r.Host, struct{}{})
|
||||||
|
|
||||||
|
// Ensure retry flag is unset
|
||||||
|
// when reached max attempts.
|
||||||
|
retry = false
|
||||||
|
|
||||||
|
case c.badHosts.Has(r.Host):
|
||||||
|
// When retry is still permitted,
|
||||||
|
// check host hasn't been marked
|
||||||
|
// as a "badhost", i.e. erroring.
|
||||||
|
r.attempts = c.retries + 1
|
||||||
|
retry = false
|
||||||
|
}
|
||||||
|
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// do performs the "meat" of DoOnce(), but it's separated out to allow
|
||||||
|
// easier wrapping of the response, retry, error returns with further logic.
|
||||||
|
func (c *Client) do(r *Request) (rsp *http.Response, retry bool, err error) {
|
||||||
|
// Perform the HTTP request.
|
||||||
|
rsp, err = c.client.Do(r.Request)
|
||||||
|
if err != nil {
|
||||||
|
|
||||||
|
if errorsv2.IsV2(err,
|
||||||
context.DeadlineExceeded,
|
context.DeadlineExceeded,
|
||||||
context.Canceled,
|
context.Canceled,
|
||||||
ErrBodyTooLarge,
|
ErrBodyTooLarge,
|
||||||
ErrReservedAddr,
|
ErrReservedAddr,
|
||||||
) {
|
) {
|
||||||
// Non-retryable errors.
|
// Non-retryable errors.
|
||||||
return nil, err
|
return nil, false, err
|
||||||
} else if errstr := err.Error(); // nocollapse
|
}
|
||||||
|
|
||||||
|
if errstr := err.Error(); //
|
||||||
strings.Contains(errstr, "stopped after 10 redirects") ||
|
strings.Contains(errstr, "stopped after 10 redirects") ||
|
||||||
strings.Contains(errstr, "tls: ") ||
|
strings.Contains(errstr, "tls: ") ||
|
||||||
strings.Contains(errstr, "x509: ") {
|
strings.Contains(errstr, "x509: ") {
|
||||||
// These error types aren't wrapped
|
// These error types aren't wrapped
|
||||||
// so we have to check the error string.
|
// so we have to check the error string.
|
||||||
// All are unrecoverable!
|
// All are unrecoverable!
|
||||||
return nil, err
|
return nil, false, err
|
||||||
} else if dnserr := (*net.DNSError)(nil); // nocollapse
|
}
|
||||||
errors.As(err, &dnserr) && dnserr.IsNotFound {
|
|
||||||
|
if dnserr := errorsv2.AsV2[*net.DNSError](err); //
|
||||||
|
dnserr != nil && dnserr.IsNotFound {
|
||||||
// DNS lookup failure, this domain does not exist
|
// DNS lookup failure, this domain does not exist
|
||||||
return nil, gtserror.SetNotFound(err)
|
return nil, false, gtserror.SetNotFound(err)
|
||||||
}
|
}
|
||||||
|
|
||||||
if fastFail {
|
// A retryable error.
|
||||||
// on fast-fail, don't bother backoff/retry
|
return nil, true, err
|
||||||
return nil, fmt.Errorf("%w (fast fail)", err)
|
|
||||||
|
} else if rsp.StatusCode > 500 ||
|
||||||
|
rsp.StatusCode == http.StatusTooManyRequests {
|
||||||
|
|
||||||
|
// Codes over 500 (and 429: too many requests)
|
||||||
|
// are generally temporary errors. For these
|
||||||
|
// we replace the response with a loggable error.
|
||||||
|
err = fmt.Errorf(`http response: %s`, rsp.Status)
|
||||||
|
|
||||||
|
// Search for a provided "Retry-After" header value.
|
||||||
|
if after := rsp.Header.Get("Retry-After"); after != "" {
|
||||||
|
|
||||||
|
// Get cur time.
|
||||||
|
now := time.Now()
|
||||||
|
|
||||||
|
if u, _ := strconv.ParseUint(after, 10, 32); u != 0 {
|
||||||
|
// An integer no. of backoff seconds was provided.
|
||||||
|
r.backoff = time.Duration(u) * time.Second
|
||||||
|
} else if at, _ := http.ParseTime(after); !at.Before(now) {
|
||||||
|
// An HTTP formatted future date-time was provided.
|
||||||
|
r.backoff = at.Sub(now)
|
||||||
}
|
}
|
||||||
|
|
||||||
if backoff == 0 {
|
// Don't let their provided backoff exceed our max.
|
||||||
// No retry-after found, set our predefined
|
if max := baseBackoff * time.Duration(c.retries); //
|
||||||
// backoff according to a multiplier of 2^n.
|
r.backoff > max {
|
||||||
backoff = baseBackoff * 1 << (i + 1)
|
r.backoff = max
|
||||||
}
|
|
||||||
|
|
||||||
l.Errorf("backing off for %s after http request error: %v", backoff, err)
|
|
||||||
|
|
||||||
select {
|
|
||||||
// Request ctx cancelled
|
|
||||||
case <-r.Context().Done():
|
|
||||||
return nil, r.Context().Err()
|
|
||||||
|
|
||||||
// Backoff for some time
|
|
||||||
case <-time.After(backoff):
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Set error return to trigger setting "bad host".
|
// Unset + close rsp.
|
||||||
err = errors.New("transport reached max retries")
|
_ = rsp.Body.Close()
|
||||||
return
|
return nil, true, err
|
||||||
}
|
|
||||||
|
|
||||||
// do wraps http.Client{}.Do() to provide safely limited response bodies.
|
|
||||||
func (c *Client) do(req *http.Request) (*http.Response, error) {
|
|
||||||
// Perform the HTTP request.
|
|
||||||
rsp, err := c.client.Do(req)
|
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Seperate the body implementers.
|
// Seperate the body implementers.
|
||||||
|
@ -364,11 +395,10 @@ func (c *Client) do(req *http.Request) (*http.Response, error) {
|
||||||
|
|
||||||
// Check response body not too large.
|
// Check response body not too large.
|
||||||
if rsp.ContentLength > c.bodyMax {
|
if rsp.ContentLength > c.bodyMax {
|
||||||
_ = rsp.Body.Close()
|
return nil, false, ErrBodyTooLarge
|
||||||
return nil, ErrBodyTooLarge
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return rsp, nil
|
return rsp, true, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// cast discard writer to full interface it supports.
|
// cast discard writer to full interface it supports.
|
||||||
|
|
69
internal/httpclient/request.go
Normal file
69
internal/httpclient/request.go
Normal file
|
@ -0,0 +1,69 @@
|
||||||
|
// GoToSocial
|
||||||
|
// Copyright (C) GoToSocial Authors admin@gotosocial.org
|
||||||
|
// SPDX-License-Identifier: AGPL-3.0-or-later
|
||||||
|
//
|
||||||
|
// This program is free software: you can redistribute it and/or modify
|
||||||
|
// it under the terms of the GNU Affero General Public License as published by
|
||||||
|
// the Free Software Foundation, either version 3 of the License, or
|
||||||
|
// (at your option) any later version.
|
||||||
|
//
|
||||||
|
// This program is distributed in the hope that it will be useful,
|
||||||
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
// GNU Affero General Public License for more details.
|
||||||
|
//
|
||||||
|
// You should have received a copy of the GNU Affero General Public License
|
||||||
|
// along with this program. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
|
package httpclient
|
||||||
|
|
||||||
|
import (
|
||||||
|
"net/http"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"github.com/superseriousbusiness/gotosocial/internal/log"
|
||||||
|
)
|
||||||
|
|
||||||
|
const (
|
||||||
|
// starting backoff duration.
|
||||||
|
baseBackoff = 2 * time.Second
|
||||||
|
)
|
||||||
|
|
||||||
|
// Request wraps an HTTP request
|
||||||
|
// to add our own retry / backoff.
|
||||||
|
type Request struct {
|
||||||
|
// Current backoff dur.
|
||||||
|
backoff time.Duration
|
||||||
|
|
||||||
|
// Delivery attempts.
|
||||||
|
attempts uint
|
||||||
|
|
||||||
|
// log fields.
|
||||||
|
log.Entry
|
||||||
|
|
||||||
|
// underlying request.
|
||||||
|
*http.Request
|
||||||
|
}
|
||||||
|
|
||||||
|
// WrapRequest wraps an existing http.Request within
|
||||||
|
// our own httpclient.Request with retry / backoff tracking.
|
||||||
|
func WrapRequest(r *http.Request) Request {
|
||||||
|
var rr Request
|
||||||
|
rr.Request = r
|
||||||
|
rr.Entry = log.WithContext(r.Context()).
|
||||||
|
WithField("method", r.Method).
|
||||||
|
WithField("url", r.URL.String()).
|
||||||
|
WithField("contentType", r.Header.Get("Content-Type"))
|
||||||
|
return rr
|
||||||
|
}
|
||||||
|
|
||||||
|
// GetBackOff returns the currently set backoff duration,
|
||||||
|
// (using a default according to no. attempts if needed).
|
||||||
|
func (r *Request) BackOff() time.Duration {
|
||||||
|
if r.backoff <= 0 {
|
||||||
|
// No backoff dur found, set our predefined
|
||||||
|
// backoff according to a multiplier of 2^n.
|
||||||
|
r.backoff = baseBackoff * 1 << (r.attempts + 1)
|
||||||
|
}
|
||||||
|
return r.backoff
|
||||||
|
}
|
|
@ -32,9 +32,7 @@ type SignFunc func(r *http.Request) error
|
||||||
// (RoundTripper implementer) to check request
|
// (RoundTripper implementer) to check request
|
||||||
// context for a signing function and using for
|
// context for a signing function and using for
|
||||||
// all subsequent trips through RoundTrip().
|
// all subsequent trips through RoundTrip().
|
||||||
type signingtransport struct {
|
type signingtransport struct{ http.Transport }
|
||||||
http.Transport // underlying transport
|
|
||||||
}
|
|
||||||
|
|
||||||
func (t *signingtransport) RoundTrip(r *http.Request) (*http.Response, error) {
|
func (t *signingtransport) RoundTrip(r *http.Request) (*http.Response, error) {
|
||||||
// Ensure updated host always set.
|
// Ensure updated host always set.
|
||||||
|
|
|
@ -38,7 +38,7 @@ func ValidateRequest(r *http.Request) error {
|
||||||
return fmt.Errorf("%w: empty url host", ErrInvalidRequest)
|
return fmt.Errorf("%w: empty url host", ErrInvalidRequest)
|
||||||
case r.URL.Scheme != "http" && r.URL.Scheme != "https":
|
case r.URL.Scheme != "http" && r.URL.Scheme != "https":
|
||||||
return fmt.Errorf("%w: unsupported protocol %q", ErrInvalidRequest, r.URL.Scheme)
|
return fmt.Errorf("%w: unsupported protocol %q", ErrInvalidRequest, r.URL.Scheme)
|
||||||
case strings.IndexFunc(r.Method, func(r rune) bool { return !httpguts.IsTokenRune(r) }) != -1:
|
case strings.IndexFunc(r.Method, isNotTokenRune) != -1:
|
||||||
return fmt.Errorf("%w: invalid method %q", ErrInvalidRequest, r.Method)
|
return fmt.Errorf("%w: invalid method %q", ErrInvalidRequest, r.Method)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -60,3 +60,8 @@ func ValidateRequest(r *http.Request) error {
|
||||||
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// isNotTokenRune wraps IsTokenRune to inverse result.
|
||||||
|
func isNotTokenRune(r rune) bool {
|
||||||
|
return !httpguts.IsTokenRune(r)
|
||||||
|
}
|
||||||
|
|
|
@ -33,6 +33,7 @@ import (
|
||||||
"github.com/superseriousbusiness/gotosocial/internal/media"
|
"github.com/superseriousbusiness/gotosocial/internal/media"
|
||||||
"github.com/superseriousbusiness/gotosocial/internal/state"
|
"github.com/superseriousbusiness/gotosocial/internal/state"
|
||||||
gtsstorage "github.com/superseriousbusiness/gotosocial/internal/storage"
|
gtsstorage "github.com/superseriousbusiness/gotosocial/internal/storage"
|
||||||
|
"github.com/superseriousbusiness/gotosocial/testrig"
|
||||||
)
|
)
|
||||||
|
|
||||||
type ManagerTestSuite struct {
|
type ManagerTestSuite struct {
|
||||||
|
@ -1197,8 +1198,8 @@ func (suite *ManagerTestSuite) TestSimpleJpegProcessBlockingWithDiskStorage() {
|
||||||
|
|
||||||
var state state.State
|
var state state.State
|
||||||
|
|
||||||
state.Workers.Start()
|
testrig.StartNoopWorkers(&state)
|
||||||
defer state.Workers.Stop()
|
defer testrig.StopWorkers(&state)
|
||||||
|
|
||||||
storage := >sstorage.Driver{
|
storage := >sstorage.Driver{
|
||||||
Storage: disk,
|
Storage: disk,
|
||||||
|
|
|
@ -21,6 +21,7 @@ import (
|
||||||
"context"
|
"context"
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"io"
|
||||||
"testing"
|
"testing"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
|
@ -55,7 +56,7 @@ func (suite *AccountTestSuite) TestAccountDeleteLocal() {
|
||||||
suite.NoError(errWithCode)
|
suite.NoError(errWithCode)
|
||||||
|
|
||||||
// the delete should be federated outwards to the following account's inbox
|
// the delete should be federated outwards to the following account's inbox
|
||||||
var sent [][]byte
|
var sent []byte
|
||||||
delete := new(struct {
|
delete := new(struct {
|
||||||
Actor string `json:"actor"`
|
Actor string `json:"actor"`
|
||||||
ID string `json:"id"`
|
ID string `json:"id"`
|
||||||
|
@ -66,16 +67,22 @@ func (suite *AccountTestSuite) TestAccountDeleteLocal() {
|
||||||
})
|
})
|
||||||
|
|
||||||
if !testrig.WaitFor(func() bool {
|
if !testrig.WaitFor(func() bool {
|
||||||
sentI, ok := suite.httpClient.SentMessages.Load(*followingAccount.SharedInboxURI)
|
delivery, ok := suite.state.Workers.Delivery.Queue.Pop()
|
||||||
if ok {
|
|
||||||
sent, ok = sentI.([][]byte)
|
|
||||||
if !ok {
|
if !ok {
|
||||||
panic("SentMessages entry was not [][]byte")
|
|
||||||
}
|
|
||||||
err = json.Unmarshal(sent[0], delete)
|
|
||||||
return err == nil
|
|
||||||
}
|
|
||||||
return false
|
return false
|
||||||
|
}
|
||||||
|
if !testrig.EqualRequestURIs(delivery.Request.URL, *followingAccount.SharedInboxURI) {
|
||||||
|
panic("differing request uris")
|
||||||
|
}
|
||||||
|
sent, err = io.ReadAll(delivery.Request.Body)
|
||||||
|
if err != nil {
|
||||||
|
panic("error reading body: " + err.Error())
|
||||||
|
}
|
||||||
|
err = json.Unmarshal(sent, delete)
|
||||||
|
if err != nil {
|
||||||
|
panic("error unmarshaling json: " + err.Error())
|
||||||
|
}
|
||||||
|
return true
|
||||||
}) {
|
}) {
|
||||||
suite.FailNow("timed out waiting for message")
|
suite.FailNow("timed out waiting for message")
|
||||||
}
|
}
|
||||||
|
|
|
@ -21,6 +21,7 @@ import (
|
||||||
"context"
|
"context"
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"io"
|
||||||
"testing"
|
"testing"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
|
@ -77,22 +78,6 @@ func (suite *FollowRequestTestSuite) TestFollowRequestAccept() {
|
||||||
Note: "",
|
Note: "",
|
||||||
}, relationship)
|
}, relationship)
|
||||||
|
|
||||||
// accept should be sent to Some_User
|
|
||||||
var sent [][]byte
|
|
||||||
if !testrig.WaitFor(func() bool {
|
|
||||||
sentI, ok := suite.httpClient.SentMessages.Load(targetAccount.InboxURI)
|
|
||||||
if ok {
|
|
||||||
sent, ok = sentI.([][]byte)
|
|
||||||
if !ok {
|
|
||||||
panic("SentMessages entry was not []byte")
|
|
||||||
}
|
|
||||||
return true
|
|
||||||
}
|
|
||||||
return false
|
|
||||||
}) {
|
|
||||||
suite.FailNow("timed out waiting for message")
|
|
||||||
}
|
|
||||||
|
|
||||||
accept := &struct {
|
accept := &struct {
|
||||||
Actor string `json:"actor"`
|
Actor string `json:"actor"`
|
||||||
ID string `json:"id"`
|
ID string `json:"id"`
|
||||||
|
@ -106,8 +91,29 @@ func (suite *FollowRequestTestSuite) TestFollowRequestAccept() {
|
||||||
To string `json:"to"`
|
To string `json:"to"`
|
||||||
Type string `json:"type"`
|
Type string `json:"type"`
|
||||||
}{}
|
}{}
|
||||||
err = json.Unmarshal(sent[0], accept)
|
|
||||||
suite.NoError(err)
|
// accept should be sent to Some_User
|
||||||
|
var sent []byte
|
||||||
|
if !testrig.WaitFor(func() bool {
|
||||||
|
delivery, ok := suite.state.Workers.Delivery.Queue.Pop()
|
||||||
|
if !ok {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
if !testrig.EqualRequestURIs(delivery.Request.URL, targetAccount.InboxURI) {
|
||||||
|
panic("differing request uris")
|
||||||
|
}
|
||||||
|
sent, err = io.ReadAll(delivery.Request.Body)
|
||||||
|
if err != nil {
|
||||||
|
panic("error reading body: " + err.Error())
|
||||||
|
}
|
||||||
|
err = json.Unmarshal(sent, accept)
|
||||||
|
if err != nil {
|
||||||
|
panic("error unmarshaling json: " + err.Error())
|
||||||
|
}
|
||||||
|
return true
|
||||||
|
}) {
|
||||||
|
suite.FailNow("timed out waiting for message")
|
||||||
|
}
|
||||||
|
|
||||||
suite.Equal(requestingAccount.URI, accept.Actor)
|
suite.Equal(requestingAccount.URI, accept.Actor)
|
||||||
suite.Equal(targetAccount.URI, accept.Object.Actor)
|
suite.Equal(targetAccount.URI, accept.Object.Actor)
|
||||||
|
@ -144,22 +150,6 @@ func (suite *FollowRequestTestSuite) TestFollowRequestReject() {
|
||||||
suite.NoError(errWithCode)
|
suite.NoError(errWithCode)
|
||||||
suite.EqualValues(&apimodel.Relationship{ID: "01FHMQX3GAABWSM0S2VZEC2SWC", Following: false, ShowingReblogs: false, Notifying: false, FollowedBy: false, Blocking: false, BlockedBy: false, Muting: false, MutingNotifications: false, Requested: false, DomainBlocking: false, Endorsed: false, Note: ""}, relationship)
|
suite.EqualValues(&apimodel.Relationship{ID: "01FHMQX3GAABWSM0S2VZEC2SWC", Following: false, ShowingReblogs: false, Notifying: false, FollowedBy: false, Blocking: false, BlockedBy: false, Muting: false, MutingNotifications: false, Requested: false, DomainBlocking: false, Endorsed: false, Note: ""}, relationship)
|
||||||
|
|
||||||
// reject should be sent to Some_User
|
|
||||||
var sent [][]byte
|
|
||||||
if !testrig.WaitFor(func() bool {
|
|
||||||
sentI, ok := suite.httpClient.SentMessages.Load(targetAccount.InboxURI)
|
|
||||||
if ok {
|
|
||||||
sent, ok = sentI.([][]byte)
|
|
||||||
if !ok {
|
|
||||||
panic("SentMessages entry was not []byte")
|
|
||||||
}
|
|
||||||
return true
|
|
||||||
}
|
|
||||||
return false
|
|
||||||
}) {
|
|
||||||
suite.FailNow("timed out waiting for message")
|
|
||||||
}
|
|
||||||
|
|
||||||
reject := &struct {
|
reject := &struct {
|
||||||
Actor string `json:"actor"`
|
Actor string `json:"actor"`
|
||||||
ID string `json:"id"`
|
ID string `json:"id"`
|
||||||
|
@ -173,8 +163,29 @@ func (suite *FollowRequestTestSuite) TestFollowRequestReject() {
|
||||||
To string `json:"to"`
|
To string `json:"to"`
|
||||||
Type string `json:"type"`
|
Type string `json:"type"`
|
||||||
}{}
|
}{}
|
||||||
err = json.Unmarshal(sent[0], reject)
|
|
||||||
suite.NoError(err)
|
// reject should be sent to Some_User
|
||||||
|
var sent []byte
|
||||||
|
if !testrig.WaitFor(func() bool {
|
||||||
|
delivery, ok := suite.state.Workers.Delivery.Queue.Pop()
|
||||||
|
if !ok {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
if !testrig.EqualRequestURIs(delivery.Request.URL, targetAccount.InboxURI) {
|
||||||
|
panic("differing request uris")
|
||||||
|
}
|
||||||
|
sent, err = io.ReadAll(delivery.Request.Body)
|
||||||
|
if err != nil {
|
||||||
|
panic("error reading body: " + err.Error())
|
||||||
|
}
|
||||||
|
err = json.Unmarshal(sent, reject)
|
||||||
|
if err != nil {
|
||||||
|
panic("error unmarshaling json: " + err.Error())
|
||||||
|
}
|
||||||
|
return true
|
||||||
|
}) {
|
||||||
|
suite.FailNow("timed out waiting for message")
|
||||||
|
}
|
||||||
|
|
||||||
suite.Equal(requestingAccount.URI, reject.Actor)
|
suite.Equal(requestingAccount.URI, reject.Actor)
|
||||||
suite.Equal(targetAccount.URI, reject.Object.Actor)
|
suite.Equal(targetAccount.URI, reject.Object.Actor)
|
||||||
|
|
|
@ -75,6 +75,12 @@ func (f *federate) DeleteAccount(ctx context.Context, account *gtsmodel.Account)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Drop any queued outgoing AP requests to / from account,
|
||||||
|
// (this stops any queued likes, boosts, creates etc).
|
||||||
|
f.state.Workers.Delivery.Queue.Delete("ActorID", account.URI)
|
||||||
|
f.state.Workers.Delivery.Queue.Delete("ObjectID", account.URI)
|
||||||
|
f.state.Workers.Delivery.Queue.Delete("TargetID", account.URI)
|
||||||
|
|
||||||
// Parse relevant URI(s).
|
// Parse relevant URI(s).
|
||||||
outboxIRI, err := parseURI(account.OutboxURI)
|
outboxIRI, err := parseURI(account.OutboxURI)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
@ -222,6 +228,11 @@ func (f *federate) DeleteStatus(ctx context.Context, status *gtsmodel.Status) er
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Drop any queued outgoing http requests for status,
|
||||||
|
// (this stops any queued likes, boosts, creates etc).
|
||||||
|
f.state.Workers.Delivery.Queue.Delete("ObjectID", status.URI)
|
||||||
|
f.state.Workers.Delivery.Queue.Delete("TargetID", status.URI)
|
||||||
|
|
||||||
// Ensure the status model is fully populated.
|
// Ensure the status model is fully populated.
|
||||||
if err := f.state.DB.PopulateStatus(ctx, status); err != nil {
|
if err := f.state.DB.PopulateStatus(ctx, status); err != nil {
|
||||||
return gtserror.Newf("error populating status: %w", err)
|
return gtserror.Newf("error populating status: %w", err)
|
||||||
|
|
|
@ -21,6 +21,7 @@ import (
|
||||||
"context"
|
"context"
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"io"
|
||||||
"testing"
|
"testing"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
|
@ -457,22 +458,6 @@ func (suite *FromFediAPITestSuite) TestProcessFollowRequestUnlocked() {
|
||||||
})
|
})
|
||||||
suite.NoError(err)
|
suite.NoError(err)
|
||||||
|
|
||||||
// an accept message should be sent to satan's inbox
|
|
||||||
var sent [][]byte
|
|
||||||
if !testrig.WaitFor(func() bool {
|
|
||||||
sentI, ok := suite.httpClient.SentMessages.Load(*originAccount.SharedInboxURI)
|
|
||||||
if ok {
|
|
||||||
sent, ok = sentI.([][]byte)
|
|
||||||
if !ok {
|
|
||||||
panic("SentMessages entry was not []byte")
|
|
||||||
}
|
|
||||||
return true
|
|
||||||
}
|
|
||||||
return false
|
|
||||||
}) {
|
|
||||||
suite.FailNow("timed out waiting for message")
|
|
||||||
}
|
|
||||||
|
|
||||||
accept := &struct {
|
accept := &struct {
|
||||||
Actor string `json:"actor"`
|
Actor string `json:"actor"`
|
||||||
ID string `json:"id"`
|
ID string `json:"id"`
|
||||||
|
@ -486,8 +471,29 @@ func (suite *FromFediAPITestSuite) TestProcessFollowRequestUnlocked() {
|
||||||
To string `json:"to"`
|
To string `json:"to"`
|
||||||
Type string `json:"type"`
|
Type string `json:"type"`
|
||||||
}{}
|
}{}
|
||||||
err = json.Unmarshal(sent[0], accept)
|
|
||||||
suite.NoError(err)
|
// an accept message should be sent to satan's inbox
|
||||||
|
var sent []byte
|
||||||
|
if !testrig.WaitFor(func() bool {
|
||||||
|
delivery, ok := suite.state.Workers.Delivery.Queue.Pop()
|
||||||
|
if !ok {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
if !testrig.EqualRequestURIs(delivery.Request.URL, *originAccount.SharedInboxURI) {
|
||||||
|
panic("differing request uris")
|
||||||
|
}
|
||||||
|
sent, err = io.ReadAll(delivery.Request.Body)
|
||||||
|
if err != nil {
|
||||||
|
panic("error reading body: " + err.Error())
|
||||||
|
}
|
||||||
|
err = json.Unmarshal(sent, accept)
|
||||||
|
if err != nil {
|
||||||
|
panic("error unmarshaling json: " + err.Error())
|
||||||
|
}
|
||||||
|
return true
|
||||||
|
}) {
|
||||||
|
suite.FailNow("timed out waiting for message")
|
||||||
|
}
|
||||||
|
|
||||||
suite.Equal(targetAccount.URI, accept.Actor)
|
suite.Equal(targetAccount.URI, accept.Actor)
|
||||||
suite.Equal(originAccount.URI, accept.Object.Actor)
|
suite.Equal(originAccount.URI, accept.Object.Actor)
|
||||||
|
|
96
internal/queue/wrappers.go
Normal file
96
internal/queue/wrappers.go
Normal file
|
@ -0,0 +1,96 @@
|
||||||
|
// GoToSocial
|
||||||
|
// Copyright (C) GoToSocial Authors admin@gotosocial.org
|
||||||
|
// SPDX-License-Identifier: AGPL-3.0-or-later
|
||||||
|
//
|
||||||
|
// This program is free software: you can redistribute it and/or modify
|
||||||
|
// it under the terms of the GNU Affero General Public License as published by
|
||||||
|
// the Free Software Foundation, either version 3 of the License, or
|
||||||
|
// (at your option) any later version.
|
||||||
|
//
|
||||||
|
// This program is distributed in the hope that it will be useful,
|
||||||
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
// GNU Affero General Public License for more details.
|
||||||
|
//
|
||||||
|
// You should have received a copy of the GNU Affero General Public License
|
||||||
|
// along with this program. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
|
package queue
|
||||||
|
|
||||||
|
import (
|
||||||
|
"sync/atomic"
|
||||||
|
|
||||||
|
"codeberg.org/gruf/go-structr"
|
||||||
|
)
|
||||||
|
|
||||||
|
// StructQueue wraps a structr.Queue{} to
|
||||||
|
// provide simple index caching by name.
|
||||||
|
type StructQueue[StructType any] struct {
|
||||||
|
queue structr.Queue[StructType]
|
||||||
|
index map[string]*structr.Index
|
||||||
|
wait atomic.Pointer[chan struct{}]
|
||||||
|
}
|
||||||
|
|
||||||
|
// Init initializes queue with structr.QueueConfig{}.
|
||||||
|
func (q *StructQueue[T]) Init(config structr.QueueConfig[T]) {
|
||||||
|
q.index = make(map[string]*structr.Index, len(config.Indices))
|
||||||
|
q.queue = structr.Queue[T]{}
|
||||||
|
q.queue.Init(config)
|
||||||
|
for _, cfg := range config.Indices {
|
||||||
|
q.index[cfg.Fields] = q.queue.Index(cfg.Fields)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Pop: see structr.Queue{}.PopFront().
|
||||||
|
func (q *StructQueue[T]) Pop() (value T, ok bool) {
|
||||||
|
return q.queue.PopFront()
|
||||||
|
}
|
||||||
|
|
||||||
|
// Push wraps structr.Queue{}.PushBack() to awaken those blocking on <-.Wait().
|
||||||
|
func (q *StructQueue[T]) Push(values ...T) {
|
||||||
|
q.queue.PushBack(values...)
|
||||||
|
q.broadcast()
|
||||||
|
}
|
||||||
|
|
||||||
|
// Delete pops (and drops!) all queued entries under index with key.
|
||||||
|
func (q *StructQueue[T]) Delete(index string, key ...any) {
|
||||||
|
i := q.index[index]
|
||||||
|
_ = q.queue.Pop(i, i.Key(key...))
|
||||||
|
}
|
||||||
|
|
||||||
|
// Len: see structr.Queue{}.Len().
|
||||||
|
func (q *StructQueue[T]) Len() int {
|
||||||
|
return q.queue.Len()
|
||||||
|
}
|
||||||
|
|
||||||
|
// Wait returns current wait channel, which may be
|
||||||
|
// blocked on to awaken when new value pushed to queue.
|
||||||
|
func (q *StructQueue[T]) Wait() <-chan struct{} {
|
||||||
|
var ch chan struct{}
|
||||||
|
|
||||||
|
for {
|
||||||
|
// Get channel ptr.
|
||||||
|
ptr := q.wait.Load()
|
||||||
|
if ptr != nil {
|
||||||
|
return *ptr
|
||||||
|
}
|
||||||
|
|
||||||
|
if ch == nil {
|
||||||
|
// Allocate new channel.
|
||||||
|
ch = make(chan struct{})
|
||||||
|
}
|
||||||
|
|
||||||
|
// Try set the new wait channel ptr.
|
||||||
|
if q.wait.CompareAndSwap(ptr, &ch) {
|
||||||
|
return ch
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// broadcast safely closes wait channel if
|
||||||
|
// currently set, releasing waiting goroutines.
|
||||||
|
func (q *StructQueue[T]) broadcast() {
|
||||||
|
if ptr := q.wait.Swap(nil); ptr != nil {
|
||||||
|
close(*ptr)
|
||||||
|
}
|
||||||
|
}
|
|
@ -28,7 +28,6 @@ import (
|
||||||
"io"
|
"io"
|
||||||
"net/http"
|
"net/http"
|
||||||
"net/url"
|
"net/url"
|
||||||
"runtime"
|
|
||||||
|
|
||||||
"codeberg.org/gruf/go-byteutil"
|
"codeberg.org/gruf/go-byteutil"
|
||||||
"codeberg.org/gruf/go-cache/v3"
|
"codeberg.org/gruf/go-cache/v3"
|
||||||
|
@ -56,7 +55,6 @@ type controller struct {
|
||||||
client pub.HttpClient
|
client pub.HttpClient
|
||||||
trspCache cache.TTLCache[string, *transport]
|
trspCache cache.TTLCache[string, *transport]
|
||||||
userAgent string
|
userAgent string
|
||||||
senders int // no. concurrent batch delivery routines.
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewController returns an implementation of the Controller interface for creating new transports
|
// NewController returns an implementation of the Controller interface for creating new transports
|
||||||
|
@ -65,15 +63,8 @@ func NewController(state *state.State, federatingDB federatingdb.DB, clock pub.C
|
||||||
host = config.GetHost()
|
host = config.GetHost()
|
||||||
proto = config.GetProtocol()
|
proto = config.GetProtocol()
|
||||||
version = config.GetSoftwareVersion()
|
version = config.GetSoftwareVersion()
|
||||||
senderMultiplier = config.GetAdvancedSenderMultiplier()
|
|
||||||
)
|
)
|
||||||
|
|
||||||
senders := senderMultiplier * runtime.GOMAXPROCS(0)
|
|
||||||
if senders < 1 {
|
|
||||||
// Clamp senders to 1.
|
|
||||||
senders = 1
|
|
||||||
}
|
|
||||||
|
|
||||||
c := &controller{
|
c := &controller{
|
||||||
state: state,
|
state: state,
|
||||||
fedDB: federatingDB,
|
fedDB: federatingDB,
|
||||||
|
@ -81,7 +72,6 @@ func NewController(state *state.State, federatingDB federatingdb.DB, clock pub.C
|
||||||
client: client,
|
client: client,
|
||||||
trspCache: cache.NewTTL[string, *transport](0, 100, 0),
|
trspCache: cache.NewTTL[string, *transport](0, 100, 0),
|
||||||
userAgent: fmt.Sprintf("gotosocial/%s (+%s://%s)", version, proto, host),
|
userAgent: fmt.Sprintf("gotosocial/%s (+%s://%s)", version, proto, host),
|
||||||
senders: senders,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return c
|
return c
|
||||||
|
|
|
@ -19,118 +19,188 @@ package transport
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
|
"encoding/json"
|
||||||
"net/http"
|
"net/http"
|
||||||
"net/url"
|
"net/url"
|
||||||
"sync"
|
|
||||||
|
|
||||||
"codeberg.org/gruf/go-byteutil"
|
"codeberg.org/gruf/go-byteutil"
|
||||||
apiutil "github.com/superseriousbusiness/gotosocial/internal/api/util"
|
apiutil "github.com/superseriousbusiness/gotosocial/internal/api/util"
|
||||||
"github.com/superseriousbusiness/gotosocial/internal/config"
|
"github.com/superseriousbusiness/gotosocial/internal/config"
|
||||||
|
"github.com/superseriousbusiness/gotosocial/internal/gtscontext"
|
||||||
"github.com/superseriousbusiness/gotosocial/internal/gtserror"
|
"github.com/superseriousbusiness/gotosocial/internal/gtserror"
|
||||||
|
"github.com/superseriousbusiness/gotosocial/internal/httpclient"
|
||||||
|
"github.com/superseriousbusiness/gotosocial/internal/transport/delivery"
|
||||||
)
|
)
|
||||||
|
|
||||||
func (t *transport) BatchDeliver(ctx context.Context, b []byte, recipients []*url.URL) error {
|
func (t *transport) BatchDeliver(ctx context.Context, obj map[string]interface{}, recipients []*url.URL) error {
|
||||||
var (
|
var (
|
||||||
// errs accumulates errors received during
|
// accumulated delivery reqs.
|
||||||
// attempted delivery by deliverer routines.
|
reqs []*delivery.Delivery
|
||||||
|
|
||||||
|
// accumulated preparation errs.
|
||||||
errs gtserror.MultiError
|
errs gtserror.MultiError
|
||||||
|
|
||||||
// wait blocks until all sender
|
|
||||||
// routines have returned.
|
|
||||||
wait sync.WaitGroup
|
|
||||||
|
|
||||||
// mutex protects 'recipients' and
|
|
||||||
// 'errs' for concurrent access.
|
|
||||||
mutex sync.Mutex
|
|
||||||
|
|
||||||
// Get current instance host info.
|
// Get current instance host info.
|
||||||
domain = config.GetAccountDomain()
|
domain = config.GetAccountDomain()
|
||||||
host = config.GetHost()
|
host = config.GetHost()
|
||||||
)
|
)
|
||||||
|
|
||||||
// Block on expect no. senders.
|
// Marshal object as JSON.
|
||||||
wait.Add(t.controller.senders)
|
b, err := json.Marshal(obj)
|
||||||
|
if err != nil {
|
||||||
for i := 0; i < t.controller.senders; i++ {
|
return gtserror.Newf("error marshaling json: %w", err)
|
||||||
go func() {
|
|
||||||
// Mark returned.
|
|
||||||
defer wait.Done()
|
|
||||||
|
|
||||||
for {
|
|
||||||
// Acquire lock.
|
|
||||||
mutex.Lock()
|
|
||||||
|
|
||||||
if len(recipients) == 0 {
|
|
||||||
// Reached end.
|
|
||||||
mutex.Unlock()
|
|
||||||
return
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Pop next recipient.
|
// Extract object IDs.
|
||||||
i := len(recipients) - 1
|
actID := getActorID(obj)
|
||||||
to := recipients[i]
|
objID := getObjectID(obj)
|
||||||
recipients = recipients[:i]
|
tgtID := getTargetID(obj)
|
||||||
|
|
||||||
// Done with lock.
|
|
||||||
mutex.Unlock()
|
|
||||||
|
|
||||||
|
for _, to := range recipients {
|
||||||
// Skip delivery to recipient if it is "us".
|
// Skip delivery to recipient if it is "us".
|
||||||
if to.Host == host || to.Host == domain {
|
if to.Host == host || to.Host == domain {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
|
|
||||||
// Attempt to deliver data to recipient.
|
// Prepare http client request.
|
||||||
if err := t.deliver(ctx, b, to); err != nil {
|
req, err := t.prepare(ctx,
|
||||||
mutex.Lock() // safely append err to accumulator.
|
actID,
|
||||||
errs.Appendf("error delivering to %s: %w", to, err)
|
objID,
|
||||||
mutex.Unlock()
|
tgtID,
|
||||||
}
|
b,
|
||||||
}
|
to,
|
||||||
}()
|
)
|
||||||
|
if err != nil {
|
||||||
|
errs.Append(err)
|
||||||
|
continue
|
||||||
}
|
}
|
||||||
|
|
||||||
// Wait for finish.
|
// Append to request queue.
|
||||||
wait.Wait()
|
reqs = append(reqs, req)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Push prepared request list to the delivery queue.
|
||||||
|
t.controller.state.Workers.Delivery.Queue.Push(reqs...)
|
||||||
|
|
||||||
// Return combined err.
|
// Return combined err.
|
||||||
return errs.Combine()
|
return errs.Combine()
|
||||||
}
|
}
|
||||||
|
|
||||||
func (t *transport) Deliver(ctx context.Context, b []byte, to *url.URL) error {
|
func (t *transport) Deliver(ctx context.Context, obj map[string]interface{}, to *url.URL) error {
|
||||||
// if 'to' host is our own, skip as we don't need to deliver to ourselves...
|
// if 'to' host is our own, skip as we don't need to deliver to ourselves...
|
||||||
if to.Host == config.GetHost() || to.Host == config.GetAccountDomain() {
|
if to.Host == config.GetHost() || to.Host == config.GetAccountDomain() {
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Deliver data to recipient.
|
// Marshal object as JSON.
|
||||||
return t.deliver(ctx, b, to)
|
b, err := json.Marshal(obj)
|
||||||
}
|
if err != nil {
|
||||||
|
return gtserror.Newf("error marshaling json: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
func (t *transport) deliver(ctx context.Context, b []byte, to *url.URL) error {
|
// Prepare http client request.
|
||||||
url := to.String()
|
req, err := t.prepare(ctx,
|
||||||
|
getActorID(obj),
|
||||||
// Use rewindable bytes reader for body.
|
getObjectID(obj),
|
||||||
var body byteutil.ReadNopCloser
|
getTargetID(obj),
|
||||||
body.Reset(b)
|
b,
|
||||||
|
to,
|
||||||
req, err := http.NewRequestWithContext(ctx, "POST", url, &body)
|
)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
req.Header.Add("Content-Type", string(apiutil.AppActivityLDJSON))
|
// Push prepared request to the delivery queue.
|
||||||
req.Header.Add("Accept-Charset", "utf-8")
|
t.controller.state.Workers.Delivery.Queue.Push(req)
|
||||||
|
|
||||||
rsp, err := t.POST(req, b)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
defer rsp.Body.Close()
|
|
||||||
|
|
||||||
if code := rsp.StatusCode; code != http.StatusOK &&
|
|
||||||
code != http.StatusCreated && code != http.StatusAccepted {
|
|
||||||
return gtserror.NewFromResponse(rsp)
|
|
||||||
}
|
|
||||||
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// prepare will prepare a POST http.Request{}
|
||||||
|
// to recipient at 'to', wrapping in a queued
|
||||||
|
// request object with signing function.
|
||||||
|
func (t *transport) prepare(
|
||||||
|
ctx context.Context,
|
||||||
|
actorID string,
|
||||||
|
objectID string,
|
||||||
|
targetID string,
|
||||||
|
data []byte,
|
||||||
|
to *url.URL,
|
||||||
|
) (
|
||||||
|
*delivery.Delivery,
|
||||||
|
error,
|
||||||
|
) {
|
||||||
|
url := to.String()
|
||||||
|
|
||||||
|
// Use rewindable reader for body.
|
||||||
|
var body byteutil.ReadNopCloser
|
||||||
|
body.Reset(data)
|
||||||
|
|
||||||
|
// Prepare POST signer.
|
||||||
|
sign := t.signPOST(data)
|
||||||
|
|
||||||
|
// Update to-be-used request context with signing details.
|
||||||
|
ctx = gtscontext.SetOutgoingPublicKeyID(ctx, t.pubKeyID)
|
||||||
|
ctx = gtscontext.SetHTTPClientSignFunc(ctx, sign)
|
||||||
|
|
||||||
|
// Prepare a new request with data body directed at URL.
|
||||||
|
r, err := http.NewRequestWithContext(ctx, "POST", url, &body)
|
||||||
|
if err != nil {
|
||||||
|
return nil, gtserror.Newf("error preparing request: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Set the standard ActivityPub content-type + charset headers.
|
||||||
|
r.Header.Add("Content-Type", string(apiutil.AppActivityLDJSON))
|
||||||
|
r.Header.Add("Accept-Charset", "utf-8")
|
||||||
|
|
||||||
|
// Validate the request before queueing for delivery.
|
||||||
|
if err := httpclient.ValidateRequest(r); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
return &delivery.Delivery{
|
||||||
|
ActorID: actorID,
|
||||||
|
ObjectID: objectID,
|
||||||
|
TargetID: targetID,
|
||||||
|
Request: httpclient.WrapRequest(r),
|
||||||
|
}, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// getObjectID extracts an object ID from 'serialized' ActivityPub object map.
|
||||||
|
func getObjectID(obj map[string]interface{}) string {
|
||||||
|
switch t := obj["object"].(type) {
|
||||||
|
case string:
|
||||||
|
return t
|
||||||
|
case map[string]interface{}:
|
||||||
|
id, _ := t["id"].(string)
|
||||||
|
return id
|
||||||
|
default:
|
||||||
|
return ""
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// getActorID extracts an actor ID from 'serialized' ActivityPub object map.
|
||||||
|
func getActorID(obj map[string]interface{}) string {
|
||||||
|
switch t := obj["actor"].(type) {
|
||||||
|
case string:
|
||||||
|
return t
|
||||||
|
case map[string]interface{}:
|
||||||
|
id, _ := t["id"].(string)
|
||||||
|
return id
|
||||||
|
default:
|
||||||
|
return ""
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// getTargetID extracts a target ID from 'serialized' ActivityPub object map.
|
||||||
|
func getTargetID(obj map[string]interface{}) string {
|
||||||
|
switch t := obj["target"].(type) {
|
||||||
|
case string:
|
||||||
|
return t
|
||||||
|
case map[string]interface{}:
|
||||||
|
id, _ := t["id"].(string)
|
||||||
|
return id
|
||||||
|
default:
|
||||||
|
return ""
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
323
internal/transport/delivery/delivery.go
Normal file
323
internal/transport/delivery/delivery.go
Normal file
|
@ -0,0 +1,323 @@
|
||||||
|
// GoToSocial
|
||||||
|
// Copyright (C) GoToSocial Authors admin@gotosocial.org
|
||||||
|
// SPDX-License-Identifier: AGPL-3.0-or-later
|
||||||
|
//
|
||||||
|
// This program is free software: you can redistribute it and/or modify
|
||||||
|
// it under the terms of the GNU Affero General Public License as published by
|
||||||
|
// the Free Software Foundation, either version 3 of the License, or
|
||||||
|
// (at your option) any later version.
|
||||||
|
//
|
||||||
|
// This program is distributed in the hope that it will be useful,
|
||||||
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
// GNU Affero General Public License for more details.
|
||||||
|
//
|
||||||
|
// You should have received a copy of the GNU Affero General Public License
|
||||||
|
// along with this program. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
|
package delivery
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"slices"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"codeberg.org/gruf/go-runners"
|
||||||
|
"codeberg.org/gruf/go-structr"
|
||||||
|
"github.com/superseriousbusiness/gotosocial/internal/gtscontext"
|
||||||
|
"github.com/superseriousbusiness/gotosocial/internal/httpclient"
|
||||||
|
"github.com/superseriousbusiness/gotosocial/internal/log"
|
||||||
|
"github.com/superseriousbusiness/gotosocial/internal/queue"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Delivery wraps an httpclient.Request{}
|
||||||
|
// to add ActivityPub ID IRI fields of the
|
||||||
|
// outgoing activity, so that deliveries may
|
||||||
|
// be indexed (and so, dropped from queue)
|
||||||
|
// by any of these possible ID IRIs.
|
||||||
|
type Delivery struct {
|
||||||
|
|
||||||
|
// ActorID contains the ActivityPub
|
||||||
|
// actor ID IRI (if any) of the activity
|
||||||
|
// being sent out by this request.
|
||||||
|
ActorID string
|
||||||
|
|
||||||
|
// ObjectID contains the ActivityPub
|
||||||
|
// object ID IRI (if any) of the activity
|
||||||
|
// being sent out by this request.
|
||||||
|
ObjectID string
|
||||||
|
|
||||||
|
// TargetID contains the ActivityPub
|
||||||
|
// target ID IRI (if any) of the activity
|
||||||
|
// being sent out by this request.
|
||||||
|
TargetID string
|
||||||
|
|
||||||
|
// Request is the prepared (+ wrapped)
|
||||||
|
// httpclient.Client{} request that
|
||||||
|
// constitutes this ActivtyPub delivery.
|
||||||
|
Request httpclient.Request
|
||||||
|
|
||||||
|
// internal fields.
|
||||||
|
next time.Time
|
||||||
|
}
|
||||||
|
|
||||||
|
func (dlv *Delivery) backoff() time.Duration {
|
||||||
|
if dlv.next.IsZero() {
|
||||||
|
return 0
|
||||||
|
}
|
||||||
|
return time.Until(dlv.next)
|
||||||
|
}
|
||||||
|
|
||||||
|
// WorkerPool wraps multiple Worker{}s in
|
||||||
|
// a singular struct for easy multi start/stop.
|
||||||
|
type WorkerPool struct {
|
||||||
|
|
||||||
|
// Client defines httpclient.Client{}
|
||||||
|
// passed to each of delivery pool Worker{}s.
|
||||||
|
Client *httpclient.Client
|
||||||
|
|
||||||
|
// Queue is the embedded queue.StructQueue{}
|
||||||
|
// passed to each of delivery pool Worker{}s.
|
||||||
|
Queue queue.StructQueue[*Delivery]
|
||||||
|
|
||||||
|
// internal fields.
|
||||||
|
workers []*Worker
|
||||||
|
}
|
||||||
|
|
||||||
|
// Init will initialize the Worker{} pool
|
||||||
|
// with given http client, request queue to pull
|
||||||
|
// from and number of delivery workers to spawn.
|
||||||
|
func (p *WorkerPool) Init(client *httpclient.Client) {
|
||||||
|
p.Client = client
|
||||||
|
p.Queue.Init(structr.QueueConfig[*Delivery]{
|
||||||
|
Indices: []structr.IndexConfig{
|
||||||
|
{Fields: "ActorID", Multiple: true},
|
||||||
|
{Fields: "ObjectID", Multiple: true},
|
||||||
|
{Fields: "TargetID", Multiple: true},
|
||||||
|
},
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
// Start will attempt to start 'n' Worker{}s.
|
||||||
|
func (p *WorkerPool) Start(n int) (ok bool) {
|
||||||
|
if ok = (len(p.workers) == 0); ok {
|
||||||
|
p.workers = make([]*Worker, n)
|
||||||
|
for i := range p.workers {
|
||||||
|
p.workers[i] = new(Worker)
|
||||||
|
p.workers[i].Client = p.Client
|
||||||
|
p.workers[i].Queue = &p.Queue
|
||||||
|
ok = p.workers[i].Start() && ok
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// Stop will attempt to stop contained Worker{}s.
|
||||||
|
func (p *WorkerPool) Stop() (ok bool) {
|
||||||
|
if ok = (len(p.workers) > 0); ok {
|
||||||
|
for i := range p.workers {
|
||||||
|
ok = p.workers[i].Stop() && ok
|
||||||
|
p.workers[i] = nil
|
||||||
|
}
|
||||||
|
p.workers = p.workers[:0]
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// Worker wraps an httpclient.Client{} to feed
|
||||||
|
// from queue.StructQueue{} for ActivityPub reqs
|
||||||
|
// to deliver. It does so while prioritizing new
|
||||||
|
// queued requests over backlogged retries.
|
||||||
|
type Worker struct {
|
||||||
|
|
||||||
|
// Client is the httpclient.Client{} that
|
||||||
|
// delivery worker will use for requests.
|
||||||
|
Client *httpclient.Client
|
||||||
|
|
||||||
|
// Queue is the Delivery{} message queue
|
||||||
|
// that delivery worker will feed from.
|
||||||
|
Queue *queue.StructQueue[*Delivery]
|
||||||
|
|
||||||
|
// internal fields.
|
||||||
|
backlog []*Delivery
|
||||||
|
service runners.Service
|
||||||
|
}
|
||||||
|
|
||||||
|
// Start will attempt to start the Worker{}.
|
||||||
|
func (w *Worker) Start() bool {
|
||||||
|
return w.service.GoRun(w.run)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Stop will attempt to stop the Worker{}.
|
||||||
|
func (w *Worker) Stop() bool {
|
||||||
|
return w.service.Stop()
|
||||||
|
}
|
||||||
|
|
||||||
|
// run wraps process to restart on any panic.
|
||||||
|
func (w *Worker) run(ctx context.Context) {
|
||||||
|
if w.Client == nil || w.Queue == nil {
|
||||||
|
panic("not yet initialized")
|
||||||
|
}
|
||||||
|
log.Infof(ctx, "%p: started delivery worker", w)
|
||||||
|
defer log.Infof(ctx, "%p: stopped delivery worker", w)
|
||||||
|
for returned := false; !returned; {
|
||||||
|
func() {
|
||||||
|
defer func() {
|
||||||
|
if r := recover(); r != nil {
|
||||||
|
log.Errorf(ctx, "recovered panic: %v", r)
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
w.process(ctx)
|
||||||
|
returned = true
|
||||||
|
}()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// process is the main delivery worker processing routine.
|
||||||
|
func (w *Worker) process(ctx context.Context) {
|
||||||
|
if w.Client == nil || w.Queue == nil {
|
||||||
|
// we perform this check here just
|
||||||
|
// to ensure the compiler knows these
|
||||||
|
// variables aren't nil in the loop,
|
||||||
|
// even if already checked by caller.
|
||||||
|
panic("not yet initialized")
|
||||||
|
}
|
||||||
|
|
||||||
|
loop:
|
||||||
|
for {
|
||||||
|
// Get next delivery.
|
||||||
|
dlv, ok := w.next(ctx)
|
||||||
|
if !ok {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// Check whether backoff required.
|
||||||
|
const min = 100 * time.Millisecond
|
||||||
|
if d := dlv.backoff(); d > min {
|
||||||
|
|
||||||
|
// Start backoff sleep timer.
|
||||||
|
backoff := time.NewTimer(d)
|
||||||
|
|
||||||
|
select {
|
||||||
|
case <-ctx.Done():
|
||||||
|
// Main ctx
|
||||||
|
// cancelled.
|
||||||
|
backoff.Stop()
|
||||||
|
return
|
||||||
|
|
||||||
|
case <-w.Queue.Wait():
|
||||||
|
// A new message was
|
||||||
|
// queued, re-add this
|
||||||
|
// to backlog + retry.
|
||||||
|
w.pushBacklog(dlv)
|
||||||
|
backoff.Stop()
|
||||||
|
continue loop
|
||||||
|
|
||||||
|
case <-backoff.C:
|
||||||
|
// success!
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Attempt delivery of AP request.
|
||||||
|
rsp, retry, err := w.Client.DoOnce(
|
||||||
|
&dlv.Request,
|
||||||
|
)
|
||||||
|
|
||||||
|
if err == nil {
|
||||||
|
// Ensure body closed.
|
||||||
|
_ = rsp.Body.Close()
|
||||||
|
continue loop
|
||||||
|
}
|
||||||
|
|
||||||
|
if !retry {
|
||||||
|
// Drop deliveries when no
|
||||||
|
// retry requested, or they
|
||||||
|
// reached max (either).
|
||||||
|
continue loop
|
||||||
|
}
|
||||||
|
|
||||||
|
// Determine next delivery attempt.
|
||||||
|
backoff := dlv.Request.BackOff()
|
||||||
|
dlv.next = time.Now().Add(backoff)
|
||||||
|
|
||||||
|
// Push to backlog.
|
||||||
|
w.pushBacklog(dlv)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// next gets the next available delivery, blocking until available if necessary.
|
||||||
|
func (w *Worker) next(ctx context.Context) (*Delivery, bool) {
|
||||||
|
loop:
|
||||||
|
for {
|
||||||
|
// Try pop next queued.
|
||||||
|
dlv, ok := w.Queue.Pop()
|
||||||
|
|
||||||
|
if !ok {
|
||||||
|
// Check the backlog.
|
||||||
|
if len(w.backlog) > 0 {
|
||||||
|
|
||||||
|
// Sort by 'next' time.
|
||||||
|
sortDeliveries(w.backlog)
|
||||||
|
|
||||||
|
// Pop next delivery.
|
||||||
|
dlv := w.popBacklog()
|
||||||
|
|
||||||
|
return dlv, true
|
||||||
|
}
|
||||||
|
|
||||||
|
select {
|
||||||
|
// Backlog is empty, we MUST
|
||||||
|
// block until next enqueued.
|
||||||
|
case <-w.Queue.Wait():
|
||||||
|
continue loop
|
||||||
|
|
||||||
|
// Worker was stopped.
|
||||||
|
case <-ctx.Done():
|
||||||
|
return nil, false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Replace request context for worker state canceling.
|
||||||
|
ctx := gtscontext.WithValues(ctx, dlv.Request.Context())
|
||||||
|
dlv.Request.Request = dlv.Request.Request.WithContext(ctx)
|
||||||
|
|
||||||
|
return dlv, true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// popBacklog pops next available from the backlog.
|
||||||
|
func (w *Worker) popBacklog() *Delivery {
|
||||||
|
if len(w.backlog) == 0 {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Pop from backlog.
|
||||||
|
dlv := w.backlog[0]
|
||||||
|
|
||||||
|
// Shift backlog down by one.
|
||||||
|
copy(w.backlog, w.backlog[1:])
|
||||||
|
w.backlog = w.backlog[:len(w.backlog)-1]
|
||||||
|
|
||||||
|
return dlv
|
||||||
|
}
|
||||||
|
|
||||||
|
// pushBacklog pushes the given delivery to backlog.
|
||||||
|
func (w *Worker) pushBacklog(dlv *Delivery) {
|
||||||
|
w.backlog = append(w.backlog, dlv)
|
||||||
|
}
|
||||||
|
|
||||||
|
// sortDeliveries sorts deliveries according
|
||||||
|
// to when is the first requiring re-attempt.
|
||||||
|
func sortDeliveries(d []*Delivery) {
|
||||||
|
slices.SortFunc(d, func(a, b *Delivery) int {
|
||||||
|
const k = +1
|
||||||
|
switch {
|
||||||
|
case a.next.Before(b.next):
|
||||||
|
return +k
|
||||||
|
case b.next.Before(a.next):
|
||||||
|
return -k
|
||||||
|
default:
|
||||||
|
return 0
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
205
internal/transport/delivery/delivery_test.go
Normal file
205
internal/transport/delivery/delivery_test.go
Normal file
|
@ -0,0 +1,205 @@
|
||||||
|
package delivery_test
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"io"
|
||||||
|
"math/rand"
|
||||||
|
"net"
|
||||||
|
"net/http"
|
||||||
|
"strconv"
|
||||||
|
"strings"
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
"codeberg.org/gruf/go-byteutil"
|
||||||
|
"github.com/superseriousbusiness/gotosocial/internal/config"
|
||||||
|
"github.com/superseriousbusiness/gotosocial/internal/httpclient"
|
||||||
|
"github.com/superseriousbusiness/gotosocial/internal/queue"
|
||||||
|
"github.com/superseriousbusiness/gotosocial/internal/transport/delivery"
|
||||||
|
)
|
||||||
|
|
||||||
|
func TestDeliveryWorkerPool(t *testing.T) {
|
||||||
|
for _, i := range []int{1, 2, 4, 8, 16, 32} {
|
||||||
|
t.Run("size="+strconv.Itoa(i), func(t *testing.T) {
|
||||||
|
testDeliveryWorkerPool(t, i, generateInput(100*i))
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func testDeliveryWorkerPool(t *testing.T, sz int, input []*testrequest) {
|
||||||
|
wp := new(delivery.WorkerPool)
|
||||||
|
wp.Init(httpclient.New(httpclient.Config{
|
||||||
|
AllowRanges: config.MustParseIPPrefixes([]string{
|
||||||
|
"127.0.0.0/8",
|
||||||
|
}),
|
||||||
|
}))
|
||||||
|
if !wp.Start(sz) {
|
||||||
|
t.Fatal("failed starting pool")
|
||||||
|
}
|
||||||
|
defer wp.Stop()
|
||||||
|
test(t, &wp.Queue, input)
|
||||||
|
}
|
||||||
|
|
||||||
|
func test(
|
||||||
|
t *testing.T,
|
||||||
|
queue *queue.StructQueue[*delivery.Delivery],
|
||||||
|
input []*testrequest,
|
||||||
|
) {
|
||||||
|
expect := make(chan *testrequest)
|
||||||
|
errors := make(chan error)
|
||||||
|
|
||||||
|
// Prepare an HTTP test handler that ensures expected delivery is received.
|
||||||
|
handler := http.HandlerFunc(func(rw http.ResponseWriter, r *http.Request) {
|
||||||
|
errors <- (<-expect).Equal(r)
|
||||||
|
})
|
||||||
|
|
||||||
|
// Start new HTTP test server listener.
|
||||||
|
l, err := net.Listen("tcp", "127.0.0.1:0")
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
defer l.Close()
|
||||||
|
|
||||||
|
// Start the HTTP server.
|
||||||
|
//
|
||||||
|
// specifically not using httptest.Server{} here as httptest
|
||||||
|
// links that server with its own http.Client{}, whereas we're
|
||||||
|
// using an httpclient.Client{} (well, delivery routine is).
|
||||||
|
srv := new(http.Server)
|
||||||
|
srv.Addr = "http://" + l.Addr().String()
|
||||||
|
srv.Handler = handler
|
||||||
|
go srv.Serve(l)
|
||||||
|
defer srv.Close()
|
||||||
|
|
||||||
|
// Range over test input.
|
||||||
|
for _, test := range input {
|
||||||
|
|
||||||
|
// Generate req for input.
|
||||||
|
req := test.Generate(srv.Addr)
|
||||||
|
r := httpclient.WrapRequest(req)
|
||||||
|
|
||||||
|
// Wrap the request in delivery.
|
||||||
|
dlv := new(delivery.Delivery)
|
||||||
|
dlv.Request = r
|
||||||
|
|
||||||
|
// Enqueue delivery!
|
||||||
|
queue.Push(dlv)
|
||||||
|
expect <- test
|
||||||
|
|
||||||
|
// Wait for errors from handler.
|
||||||
|
if err := <-errors; err != nil {
|
||||||
|
t.Error(err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
type testrequest struct {
|
||||||
|
method string
|
||||||
|
uri string
|
||||||
|
body []byte
|
||||||
|
}
|
||||||
|
|
||||||
|
// generateInput generates 'n' many testrequest cases.
|
||||||
|
func generateInput(n int) []*testrequest {
|
||||||
|
tests := make([]*testrequest, n)
|
||||||
|
for i := range tests {
|
||||||
|
tests[i] = new(testrequest)
|
||||||
|
tests[i].method = randomMethod()
|
||||||
|
tests[i].uri = randomURI()
|
||||||
|
tests[i].body = randomBody(tests[i].method)
|
||||||
|
}
|
||||||
|
return tests
|
||||||
|
}
|
||||||
|
|
||||||
|
var methods = []string{
|
||||||
|
http.MethodConnect,
|
||||||
|
http.MethodDelete,
|
||||||
|
http.MethodGet,
|
||||||
|
http.MethodHead,
|
||||||
|
http.MethodOptions,
|
||||||
|
http.MethodPatch,
|
||||||
|
http.MethodPost,
|
||||||
|
http.MethodPut,
|
||||||
|
http.MethodTrace,
|
||||||
|
}
|
||||||
|
|
||||||
|
// randomMethod generates a random http method.
|
||||||
|
func randomMethod() string {
|
||||||
|
return methods[rand.Intn(len(methods))]
|
||||||
|
}
|
||||||
|
|
||||||
|
// randomURI generates a random http uri.
|
||||||
|
func randomURI() string {
|
||||||
|
n := rand.Intn(5)
|
||||||
|
p := make([]string, n)
|
||||||
|
for i := range p {
|
||||||
|
p[i] = strconv.Itoa(rand.Int())
|
||||||
|
}
|
||||||
|
return "/" + strings.Join(p, "/")
|
||||||
|
}
|
||||||
|
|
||||||
|
// randomBody generates a random http body DEPENDING on method.
|
||||||
|
func randomBody(method string) []byte {
|
||||||
|
if requiresBody(method) {
|
||||||
|
return []byte(method + " " + randomURI())
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// requiresBody returns whether method requires body.
|
||||||
|
func requiresBody(method string) bool {
|
||||||
|
switch method {
|
||||||
|
case http.MethodPatch,
|
||||||
|
http.MethodPost,
|
||||||
|
http.MethodPut:
|
||||||
|
return true
|
||||||
|
default:
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Generate will generate a real http.Request{} from test data.
|
||||||
|
func (t *testrequest) Generate(addr string) *http.Request {
|
||||||
|
var body io.ReadCloser
|
||||||
|
if t.body != nil {
|
||||||
|
var b byteutil.ReadNopCloser
|
||||||
|
b.Reset(t.body)
|
||||||
|
body = &b
|
||||||
|
}
|
||||||
|
req, err := http.NewRequest(t.method, addr+t.uri, body)
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
return req
|
||||||
|
}
|
||||||
|
|
||||||
|
// Equal checks if request matches receiving test request.
|
||||||
|
func (t *testrequest) Equal(r *http.Request) error {
|
||||||
|
// Ensure methods match.
|
||||||
|
if t.method != r.Method {
|
||||||
|
return fmt.Errorf("differing request methods: t=%q r=%q", t.method, r.Method)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Ensure request URIs match.
|
||||||
|
if t.uri != r.URL.RequestURI() {
|
||||||
|
return fmt.Errorf("differing request urls: t=%q r=%q", t.uri, r.URL.RequestURI())
|
||||||
|
}
|
||||||
|
|
||||||
|
// Ensure body cases match.
|
||||||
|
if requiresBody(t.method) {
|
||||||
|
|
||||||
|
// Read request into memory.
|
||||||
|
b, err := io.ReadAll(r.Body)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("error reading request body: %v", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Compare the request bodies.
|
||||||
|
st := strings.TrimSpace(string(t.body))
|
||||||
|
sr := strings.TrimSpace(string(b))
|
||||||
|
if st != sr {
|
||||||
|
return fmt.Errorf("differing request bodies: t=%q r=%q", st, sr)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
|
@ -51,10 +51,10 @@ type Transport interface {
|
||||||
POST(*http.Request, []byte) (*http.Response, error)
|
POST(*http.Request, []byte) (*http.Response, error)
|
||||||
|
|
||||||
// Deliver sends an ActivityStreams object.
|
// Deliver sends an ActivityStreams object.
|
||||||
Deliver(ctx context.Context, b []byte, to *url.URL) error
|
Deliver(ctx context.Context, obj map[string]interface{}, to *url.URL) error
|
||||||
|
|
||||||
// BatchDeliver sends an ActivityStreams object to multiple recipients.
|
// BatchDeliver sends an ActivityStreams object to multiple recipients.
|
||||||
BatchDeliver(ctx context.Context, b []byte, recipients []*url.URL) error
|
BatchDeliver(ctx context.Context, obj map[string]interface{}, recipients []*url.URL) error
|
||||||
|
|
||||||
/*
|
/*
|
||||||
GET functions
|
GET functions
|
||||||
|
@ -77,7 +77,8 @@ type Transport interface {
|
||||||
Finger(ctx context.Context, targetUsername string, targetDomain string) ([]byte, error)
|
Finger(ctx context.Context, targetUsername string, targetDomain string) ([]byte, error)
|
||||||
}
|
}
|
||||||
|
|
||||||
// transport implements the Transport interface.
|
// transport implements
|
||||||
|
// the Transport interface.
|
||||||
type transport struct {
|
type transport struct {
|
||||||
controller *controller
|
controller *controller
|
||||||
pubKeyID string
|
pubKeyID string
|
||||||
|
|
|
@ -23,14 +23,22 @@ import (
|
||||||
"runtime"
|
"runtime"
|
||||||
|
|
||||||
"codeberg.org/gruf/go-runners"
|
"codeberg.org/gruf/go-runners"
|
||||||
|
"github.com/superseriousbusiness/gotosocial/internal/config"
|
||||||
"github.com/superseriousbusiness/gotosocial/internal/messages"
|
"github.com/superseriousbusiness/gotosocial/internal/messages"
|
||||||
"github.com/superseriousbusiness/gotosocial/internal/scheduler"
|
"github.com/superseriousbusiness/gotosocial/internal/scheduler"
|
||||||
|
"github.com/superseriousbusiness/gotosocial/internal/transport/delivery"
|
||||||
)
|
)
|
||||||
|
|
||||||
type Workers struct {
|
type Workers struct {
|
||||||
// Main task scheduler instance.
|
// Main task scheduler instance.
|
||||||
Scheduler scheduler.Scheduler
|
Scheduler scheduler.Scheduler
|
||||||
|
|
||||||
|
// Delivery provides a worker pool that
|
||||||
|
// handles outgoing ActivityPub deliveries.
|
||||||
|
// It contains an embedded (but accessible)
|
||||||
|
// indexed queue of Delivery{} objects.
|
||||||
|
Delivery delivery.WorkerPool
|
||||||
|
|
||||||
// ClientAPI provides a worker pool that handles both
|
// ClientAPI provides a worker pool that handles both
|
||||||
// incoming client actions, and our own side-effects.
|
// incoming client actions, and our own side-effects.
|
||||||
ClientAPI runners.WorkerPool
|
ClientAPI runners.WorkerPool
|
||||||
|
@ -65,13 +73,23 @@ type Workers struct {
|
||||||
_ nocopy
|
_ nocopy
|
||||||
}
|
}
|
||||||
|
|
||||||
// Start will start all of the contained worker pools (and global scheduler).
|
// Start will start all of the contained
|
||||||
|
// worker pools (and global scheduler).
|
||||||
func (w *Workers) Start() {
|
func (w *Workers) Start() {
|
||||||
// Get currently set GOMAXPROCS.
|
// Get currently set GOMAXPROCS.
|
||||||
maxprocs := runtime.GOMAXPROCS(0)
|
maxprocs := runtime.GOMAXPROCS(0)
|
||||||
|
|
||||||
tryUntil("starting scheduler", 5, w.Scheduler.Start)
|
tryUntil("starting scheduler", 5, w.Scheduler.Start)
|
||||||
|
|
||||||
|
tryUntil("start delivery workerpool", 5, func() bool {
|
||||||
|
n := config.GetAdvancedSenderMultiplier()
|
||||||
|
if n < 1 {
|
||||||
|
// clamp min senders to 1.
|
||||||
|
return w.Delivery.Start(1)
|
||||||
|
}
|
||||||
|
return w.Delivery.Start(n * maxprocs)
|
||||||
|
})
|
||||||
|
|
||||||
tryUntil("starting client API workerpool", 5, func() bool {
|
tryUntil("starting client API workerpool", 5, func() bool {
|
||||||
return w.ClientAPI.Start(4*maxprocs, 400*maxprocs)
|
return w.ClientAPI.Start(4*maxprocs, 400*maxprocs)
|
||||||
})
|
})
|
||||||
|
@ -88,6 +106,7 @@ func (w *Workers) Start() {
|
||||||
// Stop will stop all of the contained worker pools (and global scheduler).
|
// Stop will stop all of the contained worker pools (and global scheduler).
|
||||||
func (w *Workers) Stop() {
|
func (w *Workers) Stop() {
|
||||||
tryUntil("stopping scheduler", 5, w.Scheduler.Stop)
|
tryUntil("stopping scheduler", 5, w.Scheduler.Stop)
|
||||||
|
tryUntil("stopping delivery workerpool", 5, w.Delivery.Stop)
|
||||||
tryUntil("stopping client API workerpool", 5, w.ClientAPI.Stop)
|
tryUntil("stopping client API workerpool", 5, w.ClientAPI.Stop)
|
||||||
tryUntil("stopping federator workerpool", 5, w.Federator.Stop)
|
tryUntil("stopping federator workerpool", 5, w.Federator.Stop)
|
||||||
tryUntil("stopping media workerpool", 5, w.Media.Stop)
|
tryUntil("stopping media workerpool", 5, w.Media.Stop)
|
||||||
|
|
|
@ -30,7 +30,7 @@ import (
|
||||||
"github.com/superseriousbusiness/gotosocial/internal/filter/visibility"
|
"github.com/superseriousbusiness/gotosocial/internal/filter/visibility"
|
||||||
"github.com/superseriousbusiness/gotosocial/internal/messages"
|
"github.com/superseriousbusiness/gotosocial/internal/messages"
|
||||||
tlprocessor "github.com/superseriousbusiness/gotosocial/internal/processing/timeline"
|
tlprocessor "github.com/superseriousbusiness/gotosocial/internal/processing/timeline"
|
||||||
wprocessor "github.com/superseriousbusiness/gotosocial/internal/processing/workers"
|
"github.com/superseriousbusiness/gotosocial/internal/processing/workers"
|
||||||
"github.com/superseriousbusiness/gotosocial/internal/state"
|
"github.com/superseriousbusiness/gotosocial/internal/state"
|
||||||
"github.com/superseriousbusiness/gotosocial/internal/timeline"
|
"github.com/superseriousbusiness/gotosocial/internal/timeline"
|
||||||
"github.com/superseriousbusiness/gotosocial/internal/typeutils"
|
"github.com/superseriousbusiness/gotosocial/internal/typeutils"
|
||||||
|
@ -44,6 +44,8 @@ func StartNoopWorkers(state *state.State) {
|
||||||
state.Workers.ProcessFromClientAPI = func(context.Context, messages.FromClientAPI) error { return nil }
|
state.Workers.ProcessFromClientAPI = func(context.Context, messages.FromClientAPI) error { return nil }
|
||||||
state.Workers.ProcessFromFediAPI = func(context.Context, messages.FromFediAPI) error { return nil }
|
state.Workers.ProcessFromFediAPI = func(context.Context, messages.FromFediAPI) error { return nil }
|
||||||
|
|
||||||
|
state.Workers.Delivery.Init(nil)
|
||||||
|
|
||||||
_ = state.Workers.Scheduler.Start()
|
_ = state.Workers.Scheduler.Start()
|
||||||
_ = state.Workers.ClientAPI.Start(1, 10)
|
_ = state.Workers.ClientAPI.Start(1, 10)
|
||||||
_ = state.Workers.Federator.Start(1, 10)
|
_ = state.Workers.Federator.Start(1, 10)
|
||||||
|
@ -52,11 +54,13 @@ func StartNoopWorkers(state *state.State) {
|
||||||
|
|
||||||
// Starts workers on the provided state using processing functions from the given
|
// Starts workers on the provided state using processing functions from the given
|
||||||
// workers processor. Useful when you *do* want to trigger side effects in a test.
|
// workers processor. Useful when you *do* want to trigger side effects in a test.
|
||||||
func StartWorkers(state *state.State, wProcessor *wprocessor.Processor) {
|
func StartWorkers(state *state.State, processor *workers.Processor) {
|
||||||
state.Workers.EnqueueClientAPI = wProcessor.EnqueueClientAPI
|
state.Workers.EnqueueClientAPI = processor.EnqueueClientAPI
|
||||||
state.Workers.EnqueueFediAPI = wProcessor.EnqueueFediAPI
|
state.Workers.EnqueueFediAPI = processor.EnqueueFediAPI
|
||||||
state.Workers.ProcessFromClientAPI = wProcessor.ProcessFromClientAPI
|
state.Workers.ProcessFromClientAPI = processor.ProcessFromClientAPI
|
||||||
state.Workers.ProcessFromFediAPI = wProcessor.ProcessFromFediAPI
|
state.Workers.ProcessFromFediAPI = processor.ProcessFromFediAPI
|
||||||
|
|
||||||
|
state.Workers.Delivery.Init(nil)
|
||||||
|
|
||||||
_ = state.Workers.Scheduler.Start()
|
_ = state.Workers.Scheduler.Start()
|
||||||
_ = state.Workers.ClientAPI.Start(1, 10)
|
_ = state.Workers.ClientAPI.Start(1, 10)
|
||||||
|
@ -93,6 +97,64 @@ func StartTimelines(state *state.State, filter *visibility.Filter, converter *ty
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// EqualRequestURIs checks whether inputs have equal request URIs,
|
||||||
|
// handling cases of url.URL{}, *url.URL{}, string, *string.
|
||||||
|
func EqualRequestURIs(u1, u2 any) bool {
|
||||||
|
var uri1, uri2 string
|
||||||
|
|
||||||
|
requestURI := func(in string) (string, error) {
|
||||||
|
u, err := url.Parse(in)
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
return u.RequestURI(), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
switch u1 := u1.(type) {
|
||||||
|
case url.URL:
|
||||||
|
uri1 = u1.RequestURI()
|
||||||
|
case *url.URL:
|
||||||
|
uri1 = u1.RequestURI()
|
||||||
|
case *string:
|
||||||
|
var err error
|
||||||
|
uri1, err = requestURI(*u1)
|
||||||
|
if err != nil {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
case string:
|
||||||
|
var err error
|
||||||
|
uri1, err = requestURI(u1)
|
||||||
|
if err != nil {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
default:
|
||||||
|
panic("unsupported type")
|
||||||
|
}
|
||||||
|
|
||||||
|
switch u2 := u2.(type) {
|
||||||
|
case url.URL:
|
||||||
|
uri2 = u2.RequestURI()
|
||||||
|
case *url.URL:
|
||||||
|
uri2 = u2.RequestURI()
|
||||||
|
case *string:
|
||||||
|
var err error
|
||||||
|
uri2, err = requestURI(*u2)
|
||||||
|
if err != nil {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
case string:
|
||||||
|
var err error
|
||||||
|
uri2, err = requestURI(u2)
|
||||||
|
if err != nil {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
default:
|
||||||
|
panic("unsupported type")
|
||||||
|
}
|
||||||
|
|
||||||
|
return uri1 == uri2
|
||||||
|
}
|
||||||
|
|
||||||
// CreateMultipartFormData is a handy function for taking a fieldname and a filename, and creating a multipart form bytes buffer
|
// CreateMultipartFormData is a handy function for taking a fieldname and a filename, and creating a multipart form bytes buffer
|
||||||
// with the file contents set in the given fieldname. The extraFields param can be used to add extra FormFields to the request, as necessary.
|
// with the file contents set in the given fieldname. The extraFields param can be used to add extra FormFields to the request, as necessary.
|
||||||
// The returned bytes.Buffer b can be used like so:
|
// The returned bytes.Buffer b can be used like so:
|
||||||
|
|
8
vendor/github.com/superseriousbusiness/activity/pub/side_effect_actor.go
generated
vendored
8
vendor/github.com/superseriousbusiness/activity/pub/side_effect_actor.go
generated
vendored
|
@ -2,7 +2,6 @@ package pub
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"encoding/json"
|
|
||||||
"fmt"
|
"fmt"
|
||||||
"net/http"
|
"net/http"
|
||||||
"net/url"
|
"net/url"
|
||||||
|
@ -477,17 +476,12 @@ func (a *SideEffectActor) deliverToRecipients(c context.Context, boxIRI *url.URL
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
b, err := json.Marshal(m)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
tp, err := a.common.NewTransport(c, boxIRI, goFedUserAgent())
|
tp, err := a.common.NewTransport(c, boxIRI, goFedUserAgent())
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
return tp.BatchDeliver(c, b, recipients)
|
return tp.BatchDeliver(c, m, recipients)
|
||||||
}
|
}
|
||||||
|
|
||||||
// addToOutbox adds the activity to the outbox and creates the activity in the
|
// addToOutbox adds the activity to the outbox and creates the activity in the
|
||||||
|
|
79
vendor/github.com/superseriousbusiness/activity/pub/transport.go
generated
vendored
79
vendor/github.com/superseriousbusiness/activity/pub/transport.go
generated
vendored
|
@ -4,6 +4,7 @@ import (
|
||||||
"bytes"
|
"bytes"
|
||||||
"context"
|
"context"
|
||||||
"crypto"
|
"crypto"
|
||||||
|
"encoding/json"
|
||||||
"fmt"
|
"fmt"
|
||||||
"net/http"
|
"net/http"
|
||||||
"net/url"
|
"net/url"
|
||||||
|
@ -44,10 +45,10 @@ type Transport interface {
|
||||||
Dereference(c context.Context, iri *url.URL) (*http.Response, error)
|
Dereference(c context.Context, iri *url.URL) (*http.Response, error)
|
||||||
|
|
||||||
// Deliver sends an ActivityStreams object.
|
// Deliver sends an ActivityStreams object.
|
||||||
Deliver(c context.Context, b []byte, to *url.URL) error
|
Deliver(c context.Context, obj map[string]interface{}, to *url.URL) error
|
||||||
|
|
||||||
// BatchDeliver sends an ActivityStreams object to multiple recipients.
|
// BatchDeliver sends an ActivityStreams object to multiple recipients.
|
||||||
BatchDeliver(c context.Context, b []byte, recipients []*url.URL) error
|
BatchDeliver(c context.Context, obj map[string]interface{}, recipients []*url.URL) error
|
||||||
}
|
}
|
||||||
|
|
||||||
// Transport must be implemented by HttpSigTransport.
|
// Transport must be implemented by HttpSigTransport.
|
||||||
|
@ -138,7 +139,49 @@ func (h HttpSigTransport) Dereference(c context.Context, iri *url.URL) (*http.Re
|
||||||
}
|
}
|
||||||
|
|
||||||
// Deliver sends a POST request with an HTTP Signature.
|
// Deliver sends a POST request with an HTTP Signature.
|
||||||
func (h HttpSigTransport) Deliver(c context.Context, b []byte, to *url.URL) error {
|
func (h HttpSigTransport) Deliver(c context.Context, data map[string]interface{}, to *url.URL) error {
|
||||||
|
b, err := json.Marshal(data)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
return h.deliver(c, b, to)
|
||||||
|
}
|
||||||
|
|
||||||
|
// BatchDeliver sends concurrent POST requests. Returns an error if any of the requests had an error.
|
||||||
|
func (h HttpSigTransport) BatchDeliver(c context.Context, data map[string]interface{}, recipients []*url.URL) error {
|
||||||
|
b, err := json.Marshal(data)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
var wg sync.WaitGroup
|
||||||
|
errCh := make(chan error, len(recipients))
|
||||||
|
for _, recipient := range recipients {
|
||||||
|
wg.Add(1)
|
||||||
|
go func(r *url.URL) {
|
||||||
|
defer wg.Done()
|
||||||
|
if err := h.deliver(c, b, r); err != nil {
|
||||||
|
errCh <- err
|
||||||
|
}
|
||||||
|
}(recipient)
|
||||||
|
}
|
||||||
|
wg.Wait()
|
||||||
|
errs := make([]string, 0, len(recipients))
|
||||||
|
outer:
|
||||||
|
for {
|
||||||
|
select {
|
||||||
|
case e := <-errCh:
|
||||||
|
errs = append(errs, e.Error())
|
||||||
|
default:
|
||||||
|
break outer
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if len(errs) > 0 {
|
||||||
|
return fmt.Errorf("batch deliver had at least one failure: %s", strings.Join(errs, "; "))
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (h HttpSigTransport) deliver(c context.Context, b []byte, to *url.URL) error {
|
||||||
req, err := http.NewRequest("POST", to.String(), bytes.NewReader(b))
|
req, err := http.NewRequest("POST", to.String(), bytes.NewReader(b))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
|
@ -166,36 +209,6 @@ func (h HttpSigTransport) Deliver(c context.Context, b []byte, to *url.URL) erro
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// BatchDeliver sends concurrent POST requests. Returns an error if any of the requests had an error.
|
|
||||||
func (h HttpSigTransport) BatchDeliver(c context.Context, b []byte, recipients []*url.URL) error {
|
|
||||||
var wg sync.WaitGroup
|
|
||||||
errCh := make(chan error, len(recipients))
|
|
||||||
for _, recipient := range recipients {
|
|
||||||
wg.Add(1)
|
|
||||||
go func(r *url.URL) {
|
|
||||||
defer wg.Done()
|
|
||||||
if err := h.Deliver(c, b, r); err != nil {
|
|
||||||
errCh <- err
|
|
||||||
}
|
|
||||||
}(recipient)
|
|
||||||
}
|
|
||||||
wg.Wait()
|
|
||||||
errs := make([]string, 0, len(recipients))
|
|
||||||
outer:
|
|
||||||
for {
|
|
||||||
select {
|
|
||||||
case e := <-errCh:
|
|
||||||
errs = append(errs, e.Error())
|
|
||||||
default:
|
|
||||||
break outer
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if len(errs) > 0 {
|
|
||||||
return fmt.Errorf("batch deliver had at least one failure: %s", strings.Join(errs, "; "))
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// HttpClient sends http requests, and is an abstraction only needed by the
|
// HttpClient sends http requests, and is an abstraction only needed by the
|
||||||
// HttpSigTransport. The standard library's Client satisfies this interface.
|
// HttpSigTransport. The standard library's Client satisfies this interface.
|
||||||
type HttpClient interface {
|
type HttpClient interface {
|
||||||
|
|
2
vendor/modules.txt
vendored
2
vendor/modules.txt
vendored
|
@ -628,7 +628,7 @@ github.com/stretchr/testify/suite
|
||||||
# github.com/subosito/gotenv v1.6.0
|
# github.com/subosito/gotenv v1.6.0
|
||||||
## explicit; go 1.18
|
## explicit; go 1.18
|
||||||
github.com/subosito/gotenv
|
github.com/subosito/gotenv
|
||||||
# github.com/superseriousbusiness/activity v1.6.0-gts.0.20240221151241-5d56c04088d4
|
# github.com/superseriousbusiness/activity v1.6.0-gts.0.20240408131430-247f7f7110f0
|
||||||
## explicit; go 1.18
|
## explicit; go 1.18
|
||||||
github.com/superseriousbusiness/activity/pub
|
github.com/superseriousbusiness/activity/pub
|
||||||
github.com/superseriousbusiness/activity/streams
|
github.com/superseriousbusiness/activity/streams
|
||||||
|
|
Loading…
Reference in a new issue