Security: - Replace all binding: tags with validate: + c.Validate() in admin handlers - Add rate limiting to auth endpoints (login, register, password reset) - Add security headers (HSTS, XSS protection, nosniff, frame options) - Wire Google Pub/Sub token verification into webhook handler - Replace ParseUnverified with proper OIDC/JWKS key verification - Verify inner Apple JWS signatures in webhook handler - Add io.LimitReader (1MB) to all webhook body reads - Add ownership verification to file deletion - Move hardcoded admin credentials to env vars - Add uniqueIndex to User.Email - Hide ConfirmationCode from JSON serialization - Mask confirmation codes in admin responses - Use http.DetectContentType for upload validation - Fix path traversal in storage service - Replace os.Getenv with Viper in stripe service - Sanitize Redis URLs before logging - Separate DEBUG_FIXED_CODES from DEBUG flag - Reject weak SECRET_KEY in production - Add host check on /_next/* proxy routes - Use explicit localhost CORS origins in debug mode - Replace err.Error() with generic messages in all admin error responses Critical fixes: - Rewrite FCM to HTTP v1 API with OAuth 2.0 service account auth - Fix user_customuser -> auth_user table names in raw SQL - Fix dashboard verified query to use UserProfile model - Add escapeLikeWildcards() to prevent SQL wildcard injection Bug fixes: - Add bounds checks for days/expiring_soon query params (1-3650) - Add receipt_data/transaction_id empty-check to RestoreSubscription - Change Active bool -> *bool in device handler - Check all unchecked GORM/FindByIDWithProfile errors - Add validation for notification hour fields (0-23) - Add max=10000 validation on task description updates Transactions & data integrity: - Wrap registration flow in transaction - Wrap QuickComplete in transaction - Move image creation inside completion transaction - Wrap SetSpecialties in transaction - Wrap GetOrCreateToken in transaction - Wrap completion+image deletion in transaction Performance: - Batch completion summaries (2 queries vs 2N) - Reuse single http.Client in IAP validation - Cache dashboard counts (30s TTL) - Batch COUNT queries in admin user list - Add Limit(500) to document queries - Add reminder_stage+due_date filters to reminder queries - Parse AllowedTypes once at init - In-memory user cache in auth middleware (30s TTL) - Timezone change detection cache - Optimize P95 with per-endpoint sorted buffers - Replace crypto/md5 with hash/fnv for ETags Code quality: - Add sync.Once to all monitoring Stop()/Close() methods - Replace 8 fmt.Printf with zerolog in auth service - Log previously discarded errors - Standardize delete response shapes - Route hardcoded English through i18n - Remove FileURL from DocumentResponse (keep MediaURL only) - Thread user timezone through kanban board responses - Initialize empty slices to prevent null JSON - Extract shared field map for task Update/UpdateTx - Delete unused SoftDeleteModel, min(), formatCron, legacy handlers Worker & jobs: - Wire Asynq email infrastructure into worker - Register HandleReminderLogCleanup with daily 3AM cron - Use per-user timezone in HandleSmartReminder - Replace direct DB queries with repository calls - Delete legacy reminder handlers (~200 lines) - Delete unused task type constants Dependencies: - Replace archived jung-kurt/gofpdf with go-pdf/fpdf - Replace unmaintained gomail.v2 with wneessen/go-mail - Add TODO for Echo jwt v3 transitive dep removal Test infrastructure: - Fix MakeRequest/SeedLookupData error handling - Replace os.Exit(0) with t.Skip() in scope/consistency tests - Add 11 new FCM v1 tests Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
215 lines
7.3 KiB
Go
215 lines
7.3 KiB
Go
package main
|
|
|
|
import (
|
|
"context"
|
|
"os"
|
|
"os/signal"
|
|
"syscall"
|
|
|
|
"github.com/hibiken/asynq"
|
|
"github.com/redis/go-redis/v9"
|
|
"github.com/rs/zerolog/log"
|
|
|
|
"github.com/treytartt/honeydue-api/internal/config"
|
|
"github.com/treytartt/honeydue-api/internal/database"
|
|
"github.com/treytartt/honeydue-api/internal/monitoring"
|
|
"github.com/treytartt/honeydue-api/internal/push"
|
|
"github.com/treytartt/honeydue-api/internal/repositories"
|
|
"github.com/treytartt/honeydue-api/internal/services"
|
|
"github.com/treytartt/honeydue-api/internal/worker/jobs"
|
|
"github.com/treytartt/honeydue-api/pkg/utils"
|
|
)
|
|
|
|
func main() {
|
|
// Initialize logger
|
|
utils.InitLogger(true)
|
|
|
|
// Load configuration
|
|
cfg, err := config.Load()
|
|
if err != nil {
|
|
log.Fatal().Err(err).Msg("Failed to load configuration")
|
|
}
|
|
|
|
// Check worker kill switch
|
|
if !cfg.Features.WorkerEnabled {
|
|
log.Warn().Msg("Worker disabled by FEATURE_WORKER_ENABLED=false — exiting")
|
|
os.Exit(0)
|
|
}
|
|
|
|
// Initialize database
|
|
db, err := database.Connect(&cfg.Database, cfg.Server.Debug)
|
|
if err != nil {
|
|
log.Fatal().Err(err).Msg("Failed to connect to database")
|
|
}
|
|
log.Info().Msg("Connected to database")
|
|
|
|
// Get underlying *sql.DB for cleanup
|
|
sqlDB, _ := db.DB()
|
|
defer sqlDB.Close()
|
|
|
|
// Initialize push client (APNs + FCM)
|
|
var pushClient *push.Client
|
|
pushClient, err = push.NewClient(&cfg.Push, cfg.Features.PushEnabled)
|
|
if err != nil {
|
|
log.Warn().Err(err).Msg("Failed to initialize push client - push notifications disabled")
|
|
} else {
|
|
log.Info().
|
|
Bool("ios_enabled", pushClient.IsIOSEnabled()).
|
|
Bool("android_enabled", pushClient.IsAndroidEnabled()).
|
|
Msg("Push notification client initialized")
|
|
}
|
|
|
|
// Initialize email service (optional)
|
|
var emailService *services.EmailService
|
|
if cfg.Email.Host != "" {
|
|
emailService = services.NewEmailService(&cfg.Email, cfg.Features.EmailEnabled)
|
|
log.Info().Str("host", cfg.Email.Host).Msg("Email service initialized")
|
|
}
|
|
|
|
// Initialize notification service for actionable push notifications
|
|
notificationRepo := repositories.NewNotificationRepository(db)
|
|
notificationService := services.NewNotificationService(notificationRepo, pushClient)
|
|
log.Info().Msg("Notification service initialized")
|
|
|
|
// Parse Redis URL for Asynq
|
|
redisOpt, err := asynq.ParseRedisURI(cfg.Redis.URL)
|
|
if err != nil {
|
|
log.Fatal().Err(err).Msg("Failed to parse Redis URL")
|
|
}
|
|
|
|
// Initialize monitoring service (if Redis is available)
|
|
var monitoringService *monitoring.Service
|
|
redisClientOpt, ok := redisOpt.(asynq.RedisClientOpt)
|
|
if ok {
|
|
redisClient := redis.NewClient(&redis.Options{
|
|
Addr: redisClientOpt.Addr,
|
|
Password: redisClientOpt.Password,
|
|
DB: redisClientOpt.DB,
|
|
})
|
|
|
|
// Verify Redis connection
|
|
if err := redisClient.Ping(context.Background()).Err(); err != nil {
|
|
log.Warn().Err(err).Msg("Failed to connect to Redis for monitoring - monitoring disabled")
|
|
} else {
|
|
monitoringService = monitoring.NewService(monitoring.Config{
|
|
Process: "worker",
|
|
RedisClient: redisClient,
|
|
DB: db, // Pass database for enable_monitoring setting sync
|
|
})
|
|
|
|
// Reinitialize logger with monitoring writer
|
|
utils.InitLoggerWithWriter(cfg.Server.Debug, monitoringService.LogWriter())
|
|
|
|
// Create Asynq inspector for queue statistics
|
|
inspector := asynq.NewInspector(redisOpt)
|
|
monitoringService.SetAsynqInspector(inspector)
|
|
|
|
// Start stats collection
|
|
monitoringService.Start()
|
|
defer monitoringService.Stop()
|
|
|
|
log.Info().
|
|
Bool("log_capture_enabled", monitoringService.IsEnabled()).
|
|
Msg("Monitoring service initialized")
|
|
}
|
|
}
|
|
|
|
// Create Asynq server
|
|
srv := asynq.NewServer(
|
|
redisOpt,
|
|
asynq.Config{
|
|
Concurrency: 10,
|
|
Queues: map[string]int{
|
|
"critical": 6,
|
|
"default": 3,
|
|
"low": 1,
|
|
},
|
|
ErrorHandler: asynq.ErrorHandlerFunc(func(ctx context.Context, task *asynq.Task, err error) {
|
|
log.Error().
|
|
Err(err).
|
|
Str("type", task.Type()).
|
|
Bytes("payload", task.Payload()).
|
|
Msg("Task processing failed")
|
|
}),
|
|
},
|
|
)
|
|
|
|
// Create job handler
|
|
jobHandler := jobs.NewHandler(db, pushClient, emailService, notificationService, cfg)
|
|
|
|
// Create Asynq mux and register handlers
|
|
mux := asynq.NewServeMux()
|
|
mux.HandleFunc(jobs.TypeSmartReminder, jobHandler.HandleSmartReminder)
|
|
mux.HandleFunc(jobs.TypeDailyDigest, jobHandler.HandleDailyDigest)
|
|
mux.HandleFunc(jobs.TypeSendEmail, jobHandler.HandleSendEmail)
|
|
mux.HandleFunc(jobs.TypeSendPush, jobHandler.HandleSendPush)
|
|
mux.HandleFunc(jobs.TypeOnboardingEmails, jobHandler.HandleOnboardingEmails)
|
|
mux.HandleFunc(jobs.TypeReminderLogCleanup, jobHandler.HandleReminderLogCleanup)
|
|
|
|
// Register email job handlers (welcome, verification, password reset, password changed)
|
|
if emailService != nil {
|
|
emailJobHandler := jobs.NewEmailJobHandler(emailService)
|
|
emailJobHandler.RegisterHandlers(mux)
|
|
}
|
|
|
|
// Start scheduler for periodic tasks
|
|
scheduler := asynq.NewScheduler(redisOpt, nil)
|
|
|
|
// Schedule smart reminder notifications (runs every hour to support per-user custom times)
|
|
// Replaces old task reminder and overdue reminder with frequency-aware system
|
|
// Uses TaskReminderLog to prevent duplicate notifications
|
|
if _, err := scheduler.Register("0 * * * *", asynq.NewTask(jobs.TypeSmartReminder, nil)); err != nil {
|
|
log.Fatal().Err(err).Msg("Failed to register smart reminder job")
|
|
}
|
|
log.Info().Str("cron", "0 * * * *").Int("default_hour", cfg.Worker.TaskReminderHour).Msg("Registered smart reminder job (runs hourly for per-user times)")
|
|
|
|
// Schedule daily digest (runs every hour to support per-user custom times)
|
|
// The job handler filters users based on their preferred notification hour
|
|
if _, err := scheduler.Register("0 * * * *", asynq.NewTask(jobs.TypeDailyDigest, nil)); err != nil {
|
|
log.Fatal().Err(err).Msg("Failed to register daily digest job")
|
|
}
|
|
log.Info().Str("cron", "0 * * * *").Int("default_hour", cfg.Worker.DailyNotifHour).Msg("Registered daily digest job (runs hourly for per-user times)")
|
|
|
|
// Schedule onboarding emails (runs daily at 10:00 AM UTC)
|
|
// Sends emails to users who haven't created residences or tasks after registration
|
|
if _, err := scheduler.Register("0 10 * * *", asynq.NewTask(jobs.TypeOnboardingEmails, nil)); err != nil {
|
|
log.Fatal().Err(err).Msg("Failed to register onboarding emails job")
|
|
}
|
|
log.Info().Str("cron", "0 10 * * *").Msg("Registered onboarding emails job (runs daily at 10:00 AM UTC)")
|
|
|
|
// Schedule reminder log cleanup (runs daily at 3:00 AM UTC)
|
|
// Removes reminder logs older than 90 days to prevent table bloat
|
|
if _, err := scheduler.Register("0 3 * * *", asynq.NewTask(jobs.TypeReminderLogCleanup, nil)); err != nil {
|
|
log.Fatal().Err(err).Msg("Failed to register reminder log cleanup job")
|
|
}
|
|
log.Info().Str("cron", "0 3 * * *").Msg("Registered reminder log cleanup job (runs daily at 3:00 AM UTC)")
|
|
|
|
// Handle graceful shutdown
|
|
quit := make(chan os.Signal, 1)
|
|
signal.Notify(quit, syscall.SIGINT, syscall.SIGTERM)
|
|
|
|
// Start scheduler in goroutine
|
|
go func() {
|
|
if err := scheduler.Run(); err != nil {
|
|
log.Fatal().Err(err).Msg("Failed to start scheduler")
|
|
}
|
|
}()
|
|
|
|
// Start worker server in goroutine
|
|
go func() {
|
|
log.Info().Msg("Starting worker server...")
|
|
if err := srv.Run(mux); err != nil {
|
|
log.Fatal().Err(err).Msg("Failed to start worker server")
|
|
}
|
|
}()
|
|
|
|
<-quit
|
|
log.Info().Msg("Shutting down worker...")
|
|
|
|
// Graceful shutdown
|
|
srv.Shutdown()
|
|
scheduler.Shutdown()
|
|
|
|
log.Info().Msg("Worker stopped")
|
|
}
|