Files
History_Api/internal/repositories/mediaRepository.go
AzenKain 0896fd587e
All checks were successful
Build and Release / release (push) Successful in 1m31s
UPDATE: FIx bug
2026-04-10 10:22:45 +07:00

299 lines
8.2 KiB
Go

package repositories
import (
"context"
"crypto/md5"
"encoding/json"
"fmt"
"history-api/internal/gen/sqlc"
"history-api/internal/models"
"history-api/pkg/cache"
"history-api/pkg/constants"
"history-api/pkg/convert"
"github.com/jackc/pgx/v5/pgtype"
)
type MediaRepository interface {
GetByID(ctx context.Context, id pgtype.UUID) (*models.MediaEntity, error)
GetByIDs(ctx context.Context, ids []string) ([]*models.MediaEntity, error)
GetByUserID(ctx context.Context, userId pgtype.UUID) ([]*models.MediaEntity, error)
Search(ctx context.Context, params sqlc.SearchMediasParams) ([]*models.MediaEntity, error)
Count(ctx context.Context, params sqlc.CountMediasParams) (int64, error)
Delete(ctx context.Context, id pgtype.UUID) error
BulkDelete(ctx context.Context, ids []pgtype.UUID) error
Create(ctx context.Context, params sqlc.CreateMediaParams) (*models.MediaEntity, error)
}
type mediaRepository struct {
q *sqlc.Queries
c cache.Cache
}
func NewMediaRepository(db sqlc.DBTX, c cache.Cache) MediaRepository {
return &mediaRepository{
q: sqlc.New(db),
c: c,
}
}
func (r *mediaRepository) generateQueryKey(prefix string, params any) string {
b, _ := json.Marshal(params)
hash := fmt.Sprintf("%x", md5.Sum(b))
return fmt.Sprintf("%s:%s", prefix, hash)
}
func (r *mediaRepository) getByIDsWithFallback(ctx context.Context, ids []string) ([]*models.MediaEntity, error) {
if len(ids) == 0 {
return []*models.MediaEntity{}, nil
}
keys := make([]string, len(ids))
for i, id := range ids {
keys[i] = fmt.Sprintf("media:id:%s", id)
}
raws := r.c.MGet(ctx, keys...)
var medias []*models.MediaEntity
missingMediasToCache := make(map[string]any)
for i, b := range raws {
if len(b) > 0 {
var m models.MediaEntity
if err := json.Unmarshal(b, &m); err == nil {
medias = append(medias, &m)
}
} else {
pgId := pgtype.UUID{}
err := pgId.Scan(ids[i])
if err != nil {
continue
}
dbMedia, err := r.GetByID(ctx, pgId)
if err == nil && dbMedia != nil {
medias = append(medias, dbMedia)
missingMediasToCache[keys[i]] = dbMedia
}
}
}
if len(missingMediasToCache) > 0 {
_ = r.c.MSet(ctx, missingMediasToCache, constants.NormalCacheDuration)
}
return medias, nil
}
func (r *mediaRepository) GetByIDs(ctx context.Context, ids []string) ([]*models.MediaEntity, error) {
return r.getByIDsWithFallback(ctx, ids)
}
func (r *mediaRepository) GetByID(ctx context.Context, id pgtype.UUID) (*models.MediaEntity, error) {
cacheId := fmt.Sprintf("media:id:%s", convert.UUIDToString(id))
var media models.MediaEntity
err := r.c.Get(ctx, cacheId, &media)
if err == nil {
_ = r.c.Set(ctx, cacheId, media, constants.NormalCacheDuration)
return &media, nil
}
row, err := r.q.GetMediaByID(ctx, id)
if err != nil {
return nil, err
}
media = models.MediaEntity{
ID: convert.UUIDToString(row.ID),
UserID: convert.UUIDToString(row.UserID),
StorageKey: row.StorageKey,
OriginalName: row.OriginalName,
MimeType: row.MimeType,
Size: row.Size,
FileMetadata: row.FileMetadata,
CreatedAt: convert.TimeToPtr(row.CreatedAt),
UpdatedAt: convert.TimeToPtr(row.UpdatedAt),
}
_ = r.c.Set(ctx, cacheId, media, constants.NormalCacheDuration)
return &media, nil
}
func (r *mediaRepository) Create(ctx context.Context, params sqlc.CreateMediaParams) (*models.MediaEntity, error) {
row, err := r.q.CreateMedia(ctx, params)
if err != nil {
return nil, err
}
go func() {
bgCtx := context.Background()
_ = r.c.DelByPattern(bgCtx, "media:search*")
_ = r.c.DelByPattern(bgCtx, "media:count*")
}()
media := models.MediaEntity{
ID: convert.UUIDToString(row.ID),
UserID: convert.UUIDToString(row.UserID),
StorageKey: row.StorageKey,
OriginalName: row.OriginalName,
MimeType: row.MimeType,
Size: row.Size,
FileMetadata: row.FileMetadata,
CreatedAt: convert.TimeToPtr(row.CreatedAt),
UpdatedAt: convert.TimeToPtr(row.UpdatedAt),
}
cacheId := fmt.Sprintf("media:id:%s", media.ID)
_ = r.c.Set(ctx, cacheId, media, constants.NormalCacheDuration)
_ = r.c.Del(ctx, fmt.Sprintf("media:userId:%s", convert.UUIDToString(params.UserID)))
return &media, nil
}
func (r *mediaRepository) Delete(ctx context.Context, id pgtype.UUID) error {
err := r.q.DeleteMedia(ctx, id)
if err != nil {
return err
}
cacheId := fmt.Sprintf("media:id:%s", convert.UUIDToString(id))
_ = r.c.Del(ctx, cacheId)
return nil
}
func (r *mediaRepository) BulkDelete(ctx context.Context, ids []pgtype.UUID) error {
if len(ids) == 0 {
return nil
}
err := r.q.DeleteMedias(ctx, ids)
if err != nil {
return err
}
keys := make([]string, len(ids))
for i, id := range ids {
keys[i] = fmt.Sprintf("media:id:%s", convert.UUIDToString(id))
}
_ = r.c.Del(ctx, keys...)
return nil
}
func (r *mediaRepository) Search(ctx context.Context, params sqlc.SearchMediasParams) ([]*models.MediaEntity, error) {
queryKey := r.generateQueryKey("media:search", params)
var cachedIDs []string
if err := r.c.Get(ctx, queryKey, &cachedIDs); err == nil && len(cachedIDs) > 0 {
listItem, err := r.getByIDsWithFallback(ctx, cachedIDs)
if err != nil {
return nil, err
}
newCachedIDs := make([]string, len(listItem))
for i, media := range listItem {
newCachedIDs[i] = media.ID
}
_ = r.c.Set(ctx, queryKey, newCachedIDs, constants.ListCacheDuration)
return listItem, err
}
rows, err := r.q.SearchMedias(ctx, params)
if err != nil {
return nil, err
}
var medias []*models.MediaEntity
var ids []string
mediasToCache := make(map[string]any)
for _, row := range rows {
media := &models.MediaEntity{
ID: convert.UUIDToString(row.ID),
UserID: convert.UUIDToString(row.UserID),
StorageKey: row.StorageKey,
OriginalName: row.OriginalName,
MimeType: row.MimeType,
Size: row.Size,
FileMetadata: row.FileMetadata,
CreatedAt: convert.TimeToPtr(row.CreatedAt),
UpdatedAt: convert.TimeToPtr(row.UpdatedAt),
}
ids = append(ids, media.ID)
medias = append(medias, media)
mediasToCache[fmt.Sprintf("media:id:%s", media.ID)] = media
}
if len(mediasToCache) > 0 {
_ = r.c.MSet(ctx, mediasToCache, constants.NormalCacheDuration)
}
if len(ids) > 0 {
_ = r.c.Set(ctx, queryKey, ids, constants.ListCacheDuration)
}
return medias, nil
}
func (r *mediaRepository) Count(ctx context.Context, params sqlc.CountMediasParams) (int64, error) {
queryKey := r.generateQueryKey("media:count", params)
var count int64
if err := r.c.Get(ctx, queryKey, &count); err == nil {
_ = r.c.Set(ctx, queryKey, count, constants.ListCacheDuration)
return count, nil
}
count, err := r.q.CountMedias(ctx, params)
if err != nil {
return 0, err
}
_ = r.c.Set(ctx, queryKey, count, constants.ListCacheDuration)
return count, nil
}
func (r *mediaRepository) GetByUserID(ctx context.Context, userId pgtype.UUID) ([]*models.MediaEntity, error) {
queryKey := fmt.Sprintf("media:userId:%s", convert.UUIDToString(userId))
var cachedIDs []string
if err := r.c.Get(ctx, queryKey, &cachedIDs); err == nil && len(cachedIDs) > 0 {
listItem, err := r.getByIDsWithFallback(ctx, cachedIDs)
if err != nil {
return nil, err
}
newCachedIDs := make([]string, len(listItem))
for i, media := range listItem {
newCachedIDs[i] = media.ID
}
_ = r.c.Set(ctx, queryKey, newCachedIDs, constants.ListCacheDuration)
return listItem, nil
}
rows, err := r.q.GetMediasByUserID(ctx, userId)
if err != nil {
return nil, err
}
var medias []*models.MediaEntity
var ids []string
mediasToCache := make(map[string]any)
for _, row := range rows {
media := &models.MediaEntity{
ID: convert.UUIDToString(row.ID),
UserID: convert.UUIDToString(row.UserID),
StorageKey: row.StorageKey,
OriginalName: row.OriginalName,
MimeType: row.MimeType,
Size: row.Size,
FileMetadata: row.FileMetadata,
CreatedAt: convert.TimeToPtr(row.CreatedAt),
UpdatedAt: convert.TimeToPtr(row.UpdatedAt),
}
ids = append(ids, media.ID)
medias = append(medias, media)
mediasToCache[fmt.Sprintf("media:id:%s", media.ID)] = media
}
if len(mediasToCache) > 0 {
_ = r.c.MSet(ctx, mediasToCache, constants.NormalCacheDuration)
}
if len(ids) > 0 {
_ = r.c.Set(ctx, queryKey, ids, constants.ListCacheDuration)
}
return medias, nil
}