diff --git a/enterprise/replay/setup.go b/enterprise/replay/setup.go index 42c283d9ee..7ed02632e5 100644 --- a/enterprise/replay/setup.go +++ b/enterprise/replay/setup.go @@ -25,7 +25,7 @@ func loadConfig() { func initFileManager() (filemanager.FileManager, string, error) { bucket := strings.TrimSpace(config.GetString("JOBS_REPLAY_BACKUP_BUCKET", "")) if bucket == "" { - pkgLogger.Error("[[ Replay ]] JOBS_BACKUP_BUCKET is not set") + pkgLogger.Error("[[ Replay ]] JOBS_REPLAY_BACKUP_BUCKET is not set") panic("Bucket is not configured.") } diff --git a/jobsdb/backup.go b/jobsdb/backup.go index 66d44deb10..e57edcf8c5 100644 --- a/jobsdb/backup.go +++ b/jobsdb/backup.go @@ -29,7 +29,7 @@ type backupSettings struct { } func (b *backupSettings) isBackupEnabled() bool { - return masterBackupEnabled && b.instanceBackupEnabled && config.GetString("JOBS_BACKUP_BUCKET", "") != "" + return masterBackupEnabled && b.instanceBackupEnabled } func IsMasterBackupEnabled() bool { diff --git a/services/archiver/archiver.go b/services/archiver/archiver.go index 8c2fda8dbf..77092b729d 100644 --- a/services/archiver/archiver.go +++ b/services/archiver/archiver.go @@ -29,12 +29,6 @@ func loadConfig() { config.RegisterIntConfigVariable(100, &backupRowsBatchSize, true, 1, "Archiver.backupRowsBatchSize") } -func IsArchiverObjectStorageConfigured() bool { - provider := config.GetString("JOBS_BACKUP_STORAGE_PROVIDER", "") - bucket := config.GetString("JOBS_BACKUP_BUCKET", "") - return provider != "" && bucket != "" -} - // ArchiveOldRecords archives records in the table with the name`tableName` and `tsColumn` provided is used as the timestamp column. func ArchiveOldRecords(tableName, tsColumn string, archivalTimeInDays int, dbHandle *sql.DB) { stmt := fmt.Sprintf(`SELECT count(*), COALESCE(MIN(id),0), COALESCE(MAX(id),0) FROM %s WHERE %s < NOW() -INTERVAL '%d DAY'`, tableName, tsColumn, archivalTimeInDays) @@ -51,18 +45,6 @@ func ArchiveOldRecords(tableName, tsColumn string, archivalTimeInDays int, dbHan return } - // TODO: Should we skip deletion if object storage provider not configured? - if !IsArchiverObjectStorageConfigured() { - stmt = fmt.Sprintf(`DELETE FROM %s WHERE id >= %d and id <= %d`, tableName, minID, maxID) - _, err = dbHandle.Exec(stmt) - if err != nil { - pkgLogger.Errorf(`[Archiver]: Error in deleting %s records: %v`, tableName, err) - return - } - pkgLogger.Infof(`[Archiver]: Deleted %s records %d to %d. No object storage was configured for archival`, tableName, minID, maxID) - return - } - tmpDirPath, err := misc.CreateTMPDIR() if err != nil { pkgLogger.Errorf("[Archiver]: Failed to create tmp DIR") diff --git a/services/filemanager/filemanager.go b/services/filemanager/filemanager.go index e311e9663e..62b43c36ae 100644 --- a/services/filemanager/filemanager.go +++ b/services/filemanager/filemanager.go @@ -96,7 +96,7 @@ func GetProviderConfigFromEnv(ctx context.Context, provider string) map[string]i switch provider { case "S3": - providerConfig["bucketName"] = config.GetString("JOBS_BACKUP_BUCKET", "") + providerConfig["bucketName"] = config.GetString("JOBS_BACKUP_BUCKET", "rudder-saas") providerConfig["prefix"] = config.GetString("JOBS_BACKUP_PREFIX", "") providerConfig["accessKeyID"] = config.GetString("AWS_ACCESS_KEY_ID", "") providerConfig["accessKey"] = config.GetString("AWS_SECRET_ACCESS_KEY", "") @@ -109,7 +109,7 @@ func GetProviderConfigFromEnv(ctx context.Context, provider string) map[string]i } case "GCS": - providerConfig["bucketName"] = config.GetString("JOBS_BACKUP_BUCKET", "") + providerConfig["bucketName"] = config.GetString("JOBS_BACKUP_BUCKET", "rudder-saas") providerConfig["prefix"] = config.GetString("JOBS_BACKUP_PREFIX", "") credentials, err := os.ReadFile(config.GetString("GOOGLE_APPLICATION_CREDENTIALS", "")) if err == nil { @@ -117,13 +117,13 @@ func GetProviderConfigFromEnv(ctx context.Context, provider string) map[string]i } case "AZURE_BLOB": - providerConfig["containerName"] = config.GetString("JOBS_BACKUP_BUCKET", "") + providerConfig["containerName"] = config.GetString("JOBS_BACKUP_BUCKET", "rudder-saas") providerConfig["prefix"] = config.GetString("JOBS_BACKUP_PREFIX", "") providerConfig["accountName"] = config.GetString("AZURE_STORAGE_ACCOUNT", "") providerConfig["accountKey"] = config.GetString("AZURE_STORAGE_ACCESS_KEY", "") case "MINIO": - providerConfig["bucketName"] = config.GetString("JOBS_BACKUP_BUCKET", "") + providerConfig["bucketName"] = config.GetString("JOBS_BACKUP_BUCKET", "rudder-saas") providerConfig["prefix"] = config.GetString("JOBS_BACKUP_PREFIX", "") providerConfig["endPoint"] = config.GetString("MINIO_ENDPOINT", "localhost:9000") providerConfig["accessKeyID"] = config.GetString("MINIO_ACCESS_KEY_ID", "minioadmin") @@ -131,7 +131,7 @@ func GetProviderConfigFromEnv(ctx context.Context, provider string) map[string]i providerConfig["useSSL"] = config.GetBool("MINIO_SSL", false) case "DIGITAL_OCEAN_SPACES": - providerConfig["bucketName"] = config.GetString("JOBS_BACKUP_BUCKET", "") + providerConfig["bucketName"] = config.GetString("JOBS_BACKUP_BUCKET", "rudder-saas") providerConfig["prefix"] = config.GetString("JOBS_BACKUP_PREFIX", "") providerConfig["endPoint"] = config.GetString("DO_SPACES_ENDPOINT", "") providerConfig["accessKeyID"] = config.GetString("DO_SPACES_ACCESS_KEY_ID", "") diff --git a/warehouse/archiver.go b/warehouse/archiver.go index a7947e1883..0912642842 100644 --- a/warehouse/archiver.go +++ b/warehouse/archiver.go @@ -13,7 +13,6 @@ import ( "github.com/iancoleman/strcase" "github.com/lib/pq" "github.com/rudderlabs/rudder-server/config" - "github.com/rudderlabs/rudder-server/services/archiver" "github.com/rudderlabs/rudder-server/services/archiver/tablearchiver" "github.com/rudderlabs/rudder-server/services/filemanager" "github.com/rudderlabs/rudder-server/utils/misc" @@ -272,7 +271,7 @@ func archiveUploads(dbHandle *sql.DB) { var storedStagingFilesLocation string if len(stagingFileIDs) > 0 { - if archiver.IsArchiverObjectStorageConfigured() && !hasUsedRudderStorage { + if !hasUsedRudderStorage { filterSQL := fmt.Sprintf(`id IN (%v)`, misc.IntArrayToString(stagingFileIDs, ",")) storedStagingFilesLocation, err = backupRecords(backupRecordsArgs{ tableName: warehouseutils.WarehouseStagingFilesTable,