Compare commits

...

2 Commits

Author SHA1 Message Date
c5ab0156fd more cache issues 2021-11-22 01:34:22 +01:00
686630b2df removed s3 bucket from config 2021-11-04 20:40:00 +01:00
8 changed files with 51 additions and 40 deletions

1
.env
View File

@@ -2,7 +2,6 @@
S3_ENDPOINT=localhost:9000
S3_ACCESS_KEY=testo
S3_SECRET_KEY=testotesto
S3_BUCKET=dev
S3_DISABLE_SSL=true
ADDRESS=:8080
VERBOSE=true

View File

@@ -12,7 +12,6 @@ type args struct {
S3Endpoint string `arg:"--s3-endpoint,required,env:S3_ENDPOINT" help:"host[:port]" placeholder:"ENDPOINT"`
S3AccessKey string `arg:"--s3-access-key,required,env:S3_ACCESS_KEY" placeholder:"ACCESS_KEY"`
S3SecretKey string `arg:"--s3-secret-key,required,env:S3_SECRET_KEY" placeholder:"SECRET_KEY"`
S3Bucket string `arg:"--s3-bucket,required,env:S3_BUCKET" placeholder:"BUCKET"`
S3DisableSSL bool `arg:"--s3-disable-ssl,env:S3_DISABLE_SSL" default:"false"`
Address string `arg:"--address,env:ADDRESS" default:":3000" help:"what address to listen on" placeholder:"ADDRESS"`
CacheTTL int64 `arg:"--cache-ttl,env:CACHE_TTL" help:"Time in seconds" default:"30" placeholder:"TTL"`
@@ -35,7 +34,6 @@ func main() {
S3SSL: !args.S3DisableSSL,
S3AccessKey: args.S3AccessKey,
S3SecretKey: args.S3SecretKey,
S3Bucket: args.S3Bucket,
DSN: args.DBConnection,
CacheTTL: time.Duration(args.CacheTTL) * time.Second,
CacheCleanup: time.Duration(args.CacheCleanup) * time.Second,

9
internal/cache/cache.go vendored Normal file
View File

@@ -0,0 +1,9 @@
package cache
import (
"github.com/graph-gophers/dataloader"
)
type S3Cache interface {
dataloader.Cache
}

View File

@@ -30,7 +30,7 @@ func deleteMutation(ctx context.Context, id types.ID) error {
return err
}
ctx.Value("loader").(*loader.Loader).InvalidateCacheForFile(ctx, id)
ctx.Value("loader").(*loader.Loader).InvalidedCacheForId(ctx, id)
return nil
}
@@ -69,7 +69,7 @@ func copyMutation(ctx context.Context, src, dest types.ID) (*types.File, error)
newID.Normalize()
ctx.Value("loader").(*loader.Loader).InvalidateCacheForFile(ctx, newID)
ctx.Value("loader").(*loader.Loader).InvalidedCacheForId(ctx, newID)
return &types.File{
ID: newID,
@@ -121,11 +121,16 @@ func moveDirMutation(ctx context.Context, src, dest types.ID) ([]*types.File, er
deleteMutation(ctx, file.ID)
loader.InvalidedCacheForId(ctx, newID)
loader.InvalidedCacheForId(ctx, file.ID)
result = append(result, &types.File{
ID: newID,
})
}
loader.InvalidedCacheForId(ctx, src)
return result, nil
}
@@ -170,7 +175,7 @@ func moveFileMutation(ctx context.Context, src, dest types.ID) (*types.File, err
newId.Normalize()
ctx.Value("loader").(*loader.Loader).InvalidateCacheForFile(ctx, newId)
ctx.Value("loader").(*loader.Loader).InvalidedCacheForId(ctx, newId)
return &types.File{
ID: newId,
@@ -201,7 +206,7 @@ func createDirectory(ctx context.Context, id types.ID) (*types.Directory, error)
newID.Normalize()
ctx.Value("loader").(*loader.Loader).InvalidateCacheForDir(ctx, newID)
ctx.Value("loader").(*loader.Loader).InvalidedCacheForId(ctx, newID)
return &types.Directory{
ID: newID,
@@ -254,7 +259,7 @@ func deleteDirectory(ctx context.Context, id types.ID) error {
}
}
loader.InvalidateCacheForDir(ctx, id)
loader.InvalidedCacheForId(ctx, id)
return nil
}

View File

@@ -194,8 +194,7 @@ func httpPostFile(ctx context.Context, rw http.ResponseWriter, r *http.Request)
}
loader := ctx.Value("loader").(*loader.Loader)
loader.InvalidateCacheForFile(ctx, *id)
loader.InvalidateCacheForDir(ctx, *id.Parent())
loader.InvalidedCacheForId(ctx, *id)
rw.WriteHeader(http.StatusCreated)
}

View File

@@ -6,6 +6,7 @@ import (
"path/filepath"
"strings"
"git.kapelle.org/niklas/s3browser/internal/cache"
"git.kapelle.org/niklas/s3browser/internal/helper"
types "git.kapelle.org/niklas/s3browser/internal/types"
"github.com/graph-gophers/dataloader"
@@ -17,26 +18,43 @@ type Loader struct {
listObjectsRecursiveLoader *dataloader.Loader
statObjectLoader *dataloader.Loader
listBucketsLoader *dataloader.Loader
listObjectsLoaderCache cache.S3Cache
listObjectsRecursiveLoaderCache cache.S3Cache
statObjectLoaderCache cache.S3Cache
listBucketsLoaderCache cache.S3Cache
}
func NewLoader(config types.AppConfig) *Loader {
listObjectsLoaderCache := &dataloader.NoCache{}
listObjectsRecursiveLoaderCache := &dataloader.NoCache{}
statObjectLoaderCache := cache.NewTTLCache(config.CacheTTL, config.CacheCleanup)
listBucketsLoaderCache := cache.NewTTLCache(config.CacheTTL, config.CacheCleanup)
return &Loader{
listObjectsLoader: dataloader.NewBatchedLoader(
listObjectsBatch,
dataloader.WithCache(&dataloader.NoCache{}),
dataloader.WithCache(listObjectsLoaderCache),
),
listObjectsLoaderCache: listObjectsLoaderCache,
listObjectsRecursiveLoader: dataloader.NewBatchedLoader(
listObjectsRecursiveBatch,
dataloader.WithCache(&dataloader.NoCache{}),
dataloader.WithCache(listObjectsRecursiveLoaderCache),
),
listObjectsRecursiveLoaderCache: listObjectsRecursiveLoaderCache,
statObjectLoader: dataloader.NewBatchedLoader(
statObjectBatch,
dataloader.WithCache(&dataloader.NoCache{}),
dataloader.WithCache(statObjectLoaderCache),
),
statObjectLoaderCache: statObjectLoaderCache,
listBucketsLoader: dataloader.NewBatchedLoader(
listBucketsBatch,
dataloader.WithCache(&dataloader.NoCache{}),
dataloader.WithCache(listBucketsLoaderCache),
),
listBucketsLoaderCache: listBucketsLoaderCache,
}
}
@@ -143,16 +161,17 @@ func (l *Loader) GetFilesRecursive(ctx context.Context, path types.ID) ([]types.
return files, nil
}
func (l *Loader) InvalidateCacheForFile(ctx context.Context, id types.ID) {
func (l *Loader) InvalidedCacheForId(ctx context.Context, id types.ID) {
parent := id.Parent()
l.statObjectLoader.Clear(ctx, id)
// Code below is useless for now until we use a propper cache for "listObjectsLoader" and "listObjectsRecursiveLoader"
// TODO: implement cache invalidation for "listObjectsLoader" and "listObjectsRecursiveLoader"
l.listObjectsLoader.Clear(ctx, id).Clear(ctx, parent)
}
func (l *Loader) InvalidateCacheForDir(ctx context.Context, path types.ID) {
parent := helper.GetParentDir(path)
l.listObjectsLoader.Clear(ctx, path).Clear(ctx, parent)
l.listObjectsRecursiveLoader.Clear(ctx, path).Clear(ctx, parent)
// Remove up from recursive list
for rParent := parent; rParent != nil; rParent = rParent.Parent() {
l.listObjectsRecursiveLoader.Clear(ctx, rParent)
}
}

View File

@@ -2,7 +2,6 @@ package s3browser
import (
"context"
"fmt"
"github.com/minio/minio-go/v7"
"github.com/minio/minio-go/v7/pkg/credentials"
@@ -17,26 +16,10 @@ import (
// setupS3Client connect the s3Client
func setupS3Client(config types.AppConfig) (*minio.Client, error) {
minioClient, err := minio.New(config.S3Endoint, &minio.Options{
return minio.New(config.S3Endoint, &minio.Options{
Creds: credentials.NewStaticV4(config.S3AccessKey, config.S3SecretKey, ""),
Secure: config.S3SSL,
})
if err != nil {
return nil, err
}
exists, err := minioClient.BucketExists(context.Background(), config.S3Bucket)
if err != nil {
return nil, err
}
if !exists {
return nil, fmt.Errorf("Bucket '%s' does not exist", config.S3Bucket)
}
return minioClient, nil
}
// Start starts the app

View File

@@ -12,7 +12,6 @@ type AppConfig struct {
S3AccessKey string
S3SecretKey string
S3SSL bool
S3Bucket string
DSN string
CacheTTL time.Duration
CacheCleanup time.Duration