2023-03-12 15:00:57 +00:00
|
|
|
// GoToSocial
|
|
|
|
// Copyright (C) GoToSocial Authors admin@gotosocial.org
|
|
|
|
// SPDX-License-Identifier: AGPL-3.0-or-later
|
|
|
|
//
|
|
|
|
// This program is free software: you can redistribute it and/or modify
|
|
|
|
// it under the terms of the GNU Affero General Public License as published by
|
|
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
|
|
// (at your option) any later version.
|
|
|
|
//
|
|
|
|
// This program is distributed in the hope that it will be useful,
|
|
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
// GNU Affero General Public License for more details.
|
|
|
|
//
|
|
|
|
// You should have received a copy of the GNU Affero General Public License
|
|
|
|
// along with this program. If not, see <http://www.gnu.org/licenses/>.
|
2023-02-11 11:48:38 +00:00
|
|
|
|
|
|
|
package media
|
|
|
|
|
|
|
|
import (
|
|
|
|
"context"
|
|
|
|
"errors"
|
|
|
|
"fmt"
|
|
|
|
"time"
|
|
|
|
|
|
|
|
"codeberg.org/gruf/go-store/v2/storage"
|
|
|
|
"github.com/superseriousbusiness/gotosocial/internal/db"
|
|
|
|
"github.com/superseriousbusiness/gotosocial/internal/gtserror"
|
|
|
|
"github.com/superseriousbusiness/gotosocial/internal/gtsmodel"
|
|
|
|
"github.com/superseriousbusiness/gotosocial/internal/log"
|
|
|
|
"github.com/superseriousbusiness/gotosocial/internal/regexes"
|
|
|
|
"github.com/superseriousbusiness/gotosocial/internal/uris"
|
|
|
|
)
|
|
|
|
|
|
|
|
const (
|
|
|
|
selectPruneLimit = 50 // Amount of media entries to select at a time from the db when pruning.
|
|
|
|
unusedLocalAttachmentDays = 3 // Number of days to keep local media in storage if not attached to a status.
|
|
|
|
)
|
|
|
|
|
2023-05-28 12:08:35 +00:00
|
|
|
// PruneAll runs all of the below pruning/uncacheing functions, and then cleans up any resulting
|
|
|
|
// empty directories from the storage driver. It can be called as a shortcut for calling the below
|
|
|
|
// pruning functions one by one.
|
|
|
|
//
|
|
|
|
// If blocking is true, then any errors encountered during the prune will be combined + returned to
|
|
|
|
// the caller. If blocking is false, the prune is run in the background and errors are just logged
|
|
|
|
// instead.
|
|
|
|
func (m *Manager) PruneAll(ctx context.Context, mediaCacheRemoteDays int, blocking bool) error {
|
2023-02-11 11:48:38 +00:00
|
|
|
const dry = false
|
|
|
|
|
|
|
|
f := func(innerCtx context.Context) error {
|
|
|
|
errs := gtserror.MultiError{}
|
|
|
|
|
|
|
|
pruned, err := m.PruneUnusedLocal(innerCtx, dry)
|
|
|
|
if err != nil {
|
|
|
|
errs = append(errs, fmt.Sprintf("error pruning unused local media (%s)", err))
|
|
|
|
} else {
|
2023-02-17 11:02:29 +00:00
|
|
|
log.Infof(ctx, "pruned %d unused local media", pruned)
|
2023-02-11 11:48:38 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
pruned, err = m.PruneUnusedRemote(innerCtx, dry)
|
|
|
|
if err != nil {
|
|
|
|
errs = append(errs, fmt.Sprintf("error pruning unused remote media: (%s)", err))
|
|
|
|
} else {
|
2023-02-17 11:02:29 +00:00
|
|
|
log.Infof(ctx, "pruned %d unused remote media", pruned)
|
2023-02-11 11:48:38 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
pruned, err = m.UncacheRemote(innerCtx, mediaCacheRemoteDays, dry)
|
|
|
|
if err != nil {
|
|
|
|
errs = append(errs, fmt.Sprintf("error uncacheing remote media older than %d day(s): (%s)", mediaCacheRemoteDays, err))
|
|
|
|
} else {
|
2023-02-17 11:02:29 +00:00
|
|
|
log.Infof(ctx, "uncached %d remote media older than %d day(s)", pruned, mediaCacheRemoteDays)
|
2023-02-11 11:48:38 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
pruned, err = m.PruneOrphaned(innerCtx, dry)
|
|
|
|
if err != nil {
|
|
|
|
errs = append(errs, fmt.Sprintf("error pruning orphaned media: (%s)", err))
|
|
|
|
} else {
|
2023-02-17 11:02:29 +00:00
|
|
|
log.Infof(ctx, "pruned %d orphaned media", pruned)
|
2023-02-11 11:48:38 +00:00
|
|
|
}
|
|
|
|
|
2023-02-13 18:40:48 +00:00
|
|
|
if err := m.state.Storage.Storage.Clean(innerCtx); err != nil {
|
2023-02-11 11:48:38 +00:00
|
|
|
errs = append(errs, fmt.Sprintf("error cleaning storage: (%s)", err))
|
|
|
|
} else {
|
2023-02-17 11:02:29 +00:00
|
|
|
log.Info(ctx, "cleaned storage")
|
2023-02-11 11:48:38 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
return errs.Combine()
|
|
|
|
}
|
|
|
|
|
|
|
|
if blocking {
|
|
|
|
return f(ctx)
|
|
|
|
}
|
|
|
|
|
|
|
|
go func() {
|
|
|
|
if err := f(context.Background()); err != nil {
|
2023-02-17 11:02:29 +00:00
|
|
|
log.Error(ctx, err)
|
2023-02-11 11:48:38 +00:00
|
|
|
}
|
|
|
|
}()
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2023-05-28 12:08:35 +00:00
|
|
|
// PruneUnusedRemote prunes unused/out of date headers and avatars cached on this instance.
|
|
|
|
//
|
|
|
|
// The returned int is the amount of media that was pruned by this function.
|
|
|
|
func (m *Manager) PruneUnusedRemote(ctx context.Context, dry bool) (int, error) {
|
2023-02-11 11:48:38 +00:00
|
|
|
var (
|
|
|
|
totalPruned int
|
|
|
|
maxID string
|
|
|
|
attachments []*gtsmodel.MediaAttachment
|
|
|
|
err error
|
|
|
|
)
|
|
|
|
|
|
|
|
// We don't know in advance how many remote attachments will meet
|
|
|
|
// our criteria for being 'unused'. So a dry run in this case just
|
|
|
|
// means we iterate through as normal, but do nothing with each entry
|
|
|
|
// instead of removing it. Define this here so we don't do the 'if dry'
|
|
|
|
// check inside the loop a million times.
|
|
|
|
var f func(ctx context.Context, attachment *gtsmodel.MediaAttachment) error
|
|
|
|
if !dry {
|
|
|
|
f = m.deleteAttachment
|
|
|
|
} else {
|
|
|
|
f = func(_ context.Context, _ *gtsmodel.MediaAttachment) error {
|
|
|
|
return nil // noop
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-02-13 18:40:48 +00:00
|
|
|
for attachments, err = m.state.DB.GetAvatarsAndHeaders(ctx, maxID, selectPruneLimit); err == nil && len(attachments) != 0; attachments, err = m.state.DB.GetAvatarsAndHeaders(ctx, maxID, selectPruneLimit) {
|
2023-02-11 11:48:38 +00:00
|
|
|
maxID = attachments[len(attachments)-1].ID // use the id of the last attachment in the slice as the next 'maxID' value
|
|
|
|
|
|
|
|
for _, attachment := range attachments {
|
2023-02-13 20:19:51 +00:00
|
|
|
// Retrieve owning account if possible.
|
|
|
|
var account *gtsmodel.Account
|
|
|
|
if accountID := attachment.AccountID; accountID != "" {
|
|
|
|
account, err = m.state.DB.GetAccountByID(ctx, attachment.AccountID)
|
|
|
|
if err != nil && !errors.Is(err, db.ErrNoEntries) {
|
|
|
|
// Only return on a real error.
|
|
|
|
return 0, fmt.Errorf("PruneUnusedRemote: error fetching account with id %s: %w", accountID, err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Prune each attachment that meets one of the following criteria:
|
|
|
|
// - Has no owning account in the database.
|
|
|
|
// - Is a header but isn't the owning account's current header.
|
|
|
|
// - Is an avatar but isn't the owning account's current avatar.
|
|
|
|
if account == nil ||
|
|
|
|
(*attachment.Header && attachment.ID != account.HeaderMediaAttachmentID) ||
|
|
|
|
(*attachment.Avatar && attachment.ID != account.AvatarMediaAttachmentID) {
|
2023-02-11 11:48:38 +00:00
|
|
|
if err := f(ctx, attachment); err != nil {
|
|
|
|
return totalPruned, err
|
|
|
|
}
|
|
|
|
totalPruned++
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Make sure we don't have a real error when we leave the loop.
|
|
|
|
if err != nil && !errors.Is(err, db.ErrNoEntries) {
|
|
|
|
return totalPruned, err
|
|
|
|
}
|
|
|
|
|
|
|
|
return totalPruned, nil
|
|
|
|
}
|
|
|
|
|
2023-05-28 12:08:35 +00:00
|
|
|
// PruneOrphaned prunes files that exist in storage but which do not have a corresponding
|
|
|
|
// entry in the database.
|
|
|
|
//
|
|
|
|
// If dry is true, then nothing will be changed, only the amount that *would* be removed
|
|
|
|
// is returned to the caller.
|
|
|
|
func (m *Manager) PruneOrphaned(ctx context.Context, dry bool) (int, error) {
|
2023-03-01 09:44:54 +00:00
|
|
|
// Emojis are stored under the instance account, so we
|
|
|
|
// need the ID of the instance account for the next part.
|
2023-02-13 18:40:48 +00:00
|
|
|
instanceAccount, err := m.state.DB.GetInstanceAccount(ctx, "")
|
2023-02-11 11:48:38 +00:00
|
|
|
if err != nil {
|
|
|
|
return 0, fmt.Errorf("PruneOrphaned: error getting instance account: %w", err)
|
|
|
|
}
|
2023-03-01 09:44:54 +00:00
|
|
|
|
2023-02-11 11:48:38 +00:00
|
|
|
instanceAccountID := instanceAccount.ID
|
|
|
|
|
2023-03-01 09:44:54 +00:00
|
|
|
var orphanedKeys []string
|
|
|
|
|
|
|
|
// Keys in storage will look like the following format:
|
|
|
|
// `[ACCOUNT_ID]/[MEDIA_TYPE]/[MEDIA_SIZE]/[MEDIA_ID].[EXTENSION]`
|
|
|
|
// We can filter out keys we're not interested in by matching through a regex.
|
|
|
|
if err := m.state.Storage.WalkKeys(ctx, func(ctx context.Context, key string) error {
|
|
|
|
if !regexes.FilePath.MatchString(key) {
|
|
|
|
// This is not our expected key format.
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Check whether this storage entry is orphaned.
|
2023-02-11 11:48:38 +00:00
|
|
|
orphaned, err := m.orphaned(ctx, key, instanceAccountID)
|
|
|
|
if err != nil {
|
2023-03-01 09:44:54 +00:00
|
|
|
return fmt.Errorf("error checking orphaned status: %w", err)
|
2023-02-11 11:48:38 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
if orphaned {
|
2023-03-01 09:44:54 +00:00
|
|
|
// Add this orphaned entry to list of keys.
|
2023-02-11 11:48:38 +00:00
|
|
|
orphanedKeys = append(orphanedKeys, key)
|
|
|
|
}
|
2023-03-01 09:44:54 +00:00
|
|
|
|
|
|
|
return nil
|
|
|
|
}); err != nil {
|
|
|
|
return 0, fmt.Errorf("PruneOrphaned: error walking keys: %w", err)
|
2023-02-11 11:48:38 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
totalPruned := len(orphanedKeys)
|
|
|
|
|
|
|
|
if dry {
|
|
|
|
// Dry run: don't remove anything.
|
|
|
|
return totalPruned, nil
|
|
|
|
}
|
|
|
|
|
2023-03-01 09:44:54 +00:00
|
|
|
// This is not a drill! We have to delete stuff!
|
|
|
|
return m.removeFiles(ctx, orphanedKeys...)
|
2023-02-11 11:48:38 +00:00
|
|
|
}
|
|
|
|
|
2023-05-28 12:08:35 +00:00
|
|
|
func (m *Manager) orphaned(ctx context.Context, key string, instanceAccountID string) (bool, error) {
|
2023-02-11 11:48:38 +00:00
|
|
|
pathParts := regexes.FilePath.FindStringSubmatch(key)
|
|
|
|
if len(pathParts) != 6 {
|
|
|
|
// This doesn't match our expectations so
|
|
|
|
// it wasn't created by gts; ignore it.
|
|
|
|
return false, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
var (
|
|
|
|
mediaType = pathParts[2]
|
|
|
|
mediaID = pathParts[4]
|
|
|
|
orphaned = false
|
|
|
|
)
|
|
|
|
|
|
|
|
// Look for keys in storage that we don't have an attachment for.
|
|
|
|
switch Type(mediaType) {
|
|
|
|
case TypeAttachment, TypeHeader, TypeAvatar:
|
2023-02-13 18:40:48 +00:00
|
|
|
if _, err := m.state.DB.GetAttachmentByID(ctx, mediaID); err != nil {
|
2023-02-11 11:48:38 +00:00
|
|
|
if !errors.Is(err, db.ErrNoEntries) {
|
|
|
|
return false, fmt.Errorf("error calling GetAttachmentByID: %w", err)
|
|
|
|
}
|
|
|
|
orphaned = true
|
|
|
|
}
|
|
|
|
case TypeEmoji:
|
|
|
|
// Look using the static URL for the emoji. Emoji images can change, so
|
|
|
|
// the MEDIA_ID part of the key for emojis will not necessarily correspond
|
|
|
|
// to the file that's currently being used as the emoji image.
|
|
|
|
staticURL := uris.GenerateURIForAttachment(instanceAccountID, string(TypeEmoji), string(SizeStatic), mediaID, mimePng)
|
2023-02-13 18:40:48 +00:00
|
|
|
if _, err := m.state.DB.GetEmojiByStaticURL(ctx, staticURL); err != nil {
|
2023-02-11 11:48:38 +00:00
|
|
|
if !errors.Is(err, db.ErrNoEntries) {
|
|
|
|
return false, fmt.Errorf("error calling GetEmojiByStaticURL: %w", err)
|
|
|
|
}
|
|
|
|
orphaned = true
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return orphaned, nil
|
|
|
|
}
|
|
|
|
|
2023-05-28 12:08:35 +00:00
|
|
|
// UncacheRemote uncaches all remote media attachments older than the given amount of days.
|
|
|
|
//
|
|
|
|
// In this context, uncacheing means deleting media files from storage and marking the attachment
|
|
|
|
// as cached=false in the database.
|
|
|
|
//
|
|
|
|
// If 'dry' is true, then only a dry run will be performed: nothing will actually be changed.
|
|
|
|
//
|
|
|
|
// The returned int is the amount of media that was/would be uncached by this function.
|
|
|
|
func (m *Manager) UncacheRemote(ctx context.Context, olderThanDays int, dry bool) (int, error) {
|
2023-02-11 11:48:38 +00:00
|
|
|
if olderThanDays < 0 {
|
|
|
|
return 0, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
olderThan := time.Now().Add(-time.Hour * 24 * time.Duration(olderThanDays))
|
|
|
|
|
|
|
|
if dry {
|
|
|
|
// Dry run, just count eligible entries without removing them.
|
2023-02-13 18:40:48 +00:00
|
|
|
return m.state.DB.CountRemoteOlderThan(ctx, olderThan)
|
2023-02-11 11:48:38 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
var (
|
|
|
|
totalPruned int
|
|
|
|
attachments []*gtsmodel.MediaAttachment
|
|
|
|
err error
|
|
|
|
)
|
|
|
|
|
2023-02-13 18:40:48 +00:00
|
|
|
for attachments, err = m.state.DB.GetRemoteOlderThan(ctx, olderThan, selectPruneLimit); err == nil && len(attachments) != 0; attachments, err = m.state.DB.GetRemoteOlderThan(ctx, olderThan, selectPruneLimit) {
|
2023-02-11 11:48:38 +00:00
|
|
|
olderThan = attachments[len(attachments)-1].CreatedAt // use the created time of the last attachment in the slice as the next 'olderThan' value
|
|
|
|
|
|
|
|
for _, attachment := range attachments {
|
|
|
|
if err := m.uncacheAttachment(ctx, attachment); err != nil {
|
|
|
|
return totalPruned, err
|
|
|
|
}
|
|
|
|
totalPruned++
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Make sure we don't have a real error when we leave the loop.
|
|
|
|
if err != nil && !errors.Is(err, db.ErrNoEntries) {
|
|
|
|
return totalPruned, err
|
|
|
|
}
|
|
|
|
|
|
|
|
return totalPruned, nil
|
|
|
|
}
|
|
|
|
|
2023-05-28 12:08:35 +00:00
|
|
|
// PruneUnusedLocal prunes unused media attachments that were uploaded by
|
|
|
|
// a user on this instance, but never actually attached to a status, or attached but
|
|
|
|
// later detached.
|
|
|
|
//
|
|
|
|
// The returned int is the amount of media that was pruned by this function.
|
|
|
|
func (m *Manager) PruneUnusedLocal(ctx context.Context, dry bool) (int, error) {
|
2023-02-11 11:48:38 +00:00
|
|
|
olderThan := time.Now().Add(-time.Hour * 24 * time.Duration(unusedLocalAttachmentDays))
|
|
|
|
|
|
|
|
if dry {
|
|
|
|
// Dry run, just count eligible entries without removing them.
|
2023-02-13 18:40:48 +00:00
|
|
|
return m.state.DB.CountLocalUnattachedOlderThan(ctx, olderThan)
|
2023-02-11 11:48:38 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
var (
|
|
|
|
totalPruned int
|
|
|
|
attachments []*gtsmodel.MediaAttachment
|
|
|
|
err error
|
|
|
|
)
|
|
|
|
|
2023-02-13 18:40:48 +00:00
|
|
|
for attachments, err = m.state.DB.GetLocalUnattachedOlderThan(ctx, olderThan, selectPruneLimit); err == nil && len(attachments) != 0; attachments, err = m.state.DB.GetLocalUnattachedOlderThan(ctx, olderThan, selectPruneLimit) {
|
2023-02-11 11:48:38 +00:00
|
|
|
olderThan = attachments[len(attachments)-1].CreatedAt // use the created time of the last attachment in the slice as the next 'olderThan' value
|
|
|
|
|
|
|
|
for _, attachment := range attachments {
|
|
|
|
if err := m.deleteAttachment(ctx, attachment); err != nil {
|
|
|
|
return totalPruned, err
|
|
|
|
}
|
|
|
|
totalPruned++
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Make sure we don't have a real error when we leave the loop.
|
|
|
|
if err != nil && !errors.Is(err, db.ErrNoEntries) {
|
|
|
|
return totalPruned, err
|
|
|
|
}
|
|
|
|
|
|
|
|
return totalPruned, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
Handy little helpers
|
|
|
|
*/
|
|
|
|
|
2023-05-28 12:08:35 +00:00
|
|
|
func (m *Manager) deleteAttachment(ctx context.Context, attachment *gtsmodel.MediaAttachment) error {
|
2023-03-01 09:44:54 +00:00
|
|
|
if _, err := m.removeFiles(ctx, attachment.File.Path, attachment.Thumbnail.Path); err != nil {
|
2023-02-11 11:48:38 +00:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
// Delete attachment completely.
|
2023-03-03 23:02:23 +00:00
|
|
|
return m.state.DB.DeleteAttachment(ctx, attachment.ID)
|
2023-02-11 11:48:38 +00:00
|
|
|
}
|
|
|
|
|
2023-05-28 12:08:35 +00:00
|
|
|
func (m *Manager) uncacheAttachment(ctx context.Context, attachment *gtsmodel.MediaAttachment) error {
|
2023-03-01 09:44:54 +00:00
|
|
|
if _, err := m.removeFiles(ctx, attachment.File.Path, attachment.Thumbnail.Path); err != nil {
|
2023-02-11 11:48:38 +00:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
// Update attachment to reflect that we no longer have it cached.
|
2023-03-31 13:19:50 +00:00
|
|
|
attachment.Cached = func() *bool { i := false; return &i }()
|
|
|
|
return m.state.DB.UpdateAttachment(ctx, attachment, "cached")
|
2023-02-11 11:48:38 +00:00
|
|
|
}
|
|
|
|
|
2023-05-28 12:08:35 +00:00
|
|
|
func (m *Manager) removeFiles(ctx context.Context, keys ...string) (int, error) {
|
2023-02-11 11:48:38 +00:00
|
|
|
errs := make(gtserror.MultiError, 0, len(keys))
|
|
|
|
|
|
|
|
for _, key := range keys {
|
2023-02-13 18:40:48 +00:00
|
|
|
if err := m.state.Storage.Delete(ctx, key); err != nil && !errors.Is(err, storage.ErrNotFound) {
|
2023-02-11 11:48:38 +00:00
|
|
|
errs = append(errs, "storage error removing "+key+": "+err.Error())
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-03-01 09:44:54 +00:00
|
|
|
return len(keys) - len(errs), errs.Combine()
|
2023-02-11 11:48:38 +00:00
|
|
|
}
|