| package message |
|
|
| import ( |
| "context" |
| "database/sql" |
| "fmt" |
| "sync" |
| "time" |
|
|
| "github.com/google/uuid" |
| "github.com/target/goalert/alert/alertlog" |
| "github.com/target/goalert/app/lifecycle" |
| "github.com/target/goalert/config" |
| "github.com/target/goalert/engine/processinglock" |
| "github.com/target/goalert/gadb" |
| "github.com/target/goalert/lock" |
| "github.com/target/goalert/notification" |
| "github.com/target/goalert/permission" |
| "github.com/target/goalert/retry" |
| "github.com/target/goalert/util" |
| "github.com/target/goalert/util/log" |
| "github.com/target/goalert/util/sqlutil" |
|
|
| "github.com/pkg/errors" |
| ) |
|
|
| |
| type DB struct { |
| lock *processinglock.Lock |
|
|
| pausable lifecycle.Pausable |
|
|
| stuckMessages *sql.Stmt |
|
|
| setSending *sql.Stmt |
|
|
| lockStmt *sql.Stmt |
| currentTime *sql.Stmt |
| retryReset *sql.Stmt |
| retryClear *sql.Stmt |
|
|
| sendDeadlineExpired *sql.Stmt |
|
|
| failDisabledCM *sql.Stmt |
| alertlogstore *alertlog.Store |
|
|
| failSMSVoice *sql.Stmt |
|
|
| sentByCMType *sql.Stmt |
|
|
| cleanupStatusUpdateOptOut *sql.Stmt |
|
|
| tempFail *sql.Stmt |
| permFail *sql.Stmt |
| updateStatus *sql.Stmt |
|
|
| advLock *sql.Stmt |
| advLockCleanup *sql.Stmt |
|
|
| createAlertBundle *sql.Stmt |
| bundleMessages *sql.Stmt |
|
|
| deleteAny *sql.Stmt |
|
|
| lastSent time.Time |
| sentMessages map[string]Message |
| } |
|
|
| |
| func NewDB(ctx context.Context, db *sql.DB, a *alertlog.Store, pausable lifecycle.Pausable) (*DB, error) { |
| lock, err := processinglock.NewLock(ctx, db, processinglock.Config{ |
| Type: processinglock.TypeMessage, |
| Version: 11, |
| }) |
| if err != nil { |
| return nil, err |
| } |
| p := &util.Prepare{DB: db, Ctx: ctx} |
|
|
| tempFail := p.P(` |
| update outgoing_messages |
| set |
| last_status = 'failed', |
| last_status_at = now(), |
| status_details = $3, |
| provider_msg_id = coalesce($2, provider_msg_id), |
| next_retry_at = CASE WHEN retry_count < 3 THEN now() + '15 seconds'::interval ELSE null END |
| where id = $1 or provider_msg_id = $2 |
| `) |
| permFail := p.P(` |
| update outgoing_messages |
| set |
| last_status = 'failed', |
| last_status_at = now(), |
| status_details = $3, |
| cycle_id = null, |
| provider_msg_id = coalesce($2, provider_msg_id), |
| next_retry_at = null |
| where id = $1 or provider_msg_id = $2 |
| `) |
| updateStatus := p.P(` |
| update outgoing_messages |
| set |
| last_status = cast($4 as enum_outgoing_messages_status), |
| last_status_at = now(), |
| status_details = $5, |
| cycle_id = null, |
| sending_deadline = null, |
| sent_at = coalesce(sent_at, fired_at, now()), |
| fired_at = null, |
| provider_msg_id = coalesce($2, provider_msg_id), |
| provider_seq = CASE WHEN $3 = -1 THEN provider_seq ELSE $3 END, |
| next_retry_at = null, |
| src_value = coalesce(src_value, $6) |
| where |
| (id = $1 or provider_msg_id = $2) and |
| (provider_seq <= $3 or $3 = -1) and |
| last_status not in ('failed', 'pending') |
| `) |
| if p.Err != nil { |
| return nil, p.Err |
| } |
| return &DB{ |
| lock: lock, |
| pausable: pausable, |
| alertlogstore: a, |
|
|
| updateStatus: updateStatus, |
| tempFail: tempFail, |
| permFail: permFail, |
|
|
| sentMessages: make(map[string]Message), |
|
|
| advLock: p.P(`select pg_try_advisory_lock($1)`), |
| advLockCleanup: p.P(` |
| select pg_terminate_backend(lock.pid) |
| from pg_locks lock |
| join pg_database pgdat on |
| datname = current_database() and |
| lock.database = pgdat.oid |
| join pg_stat_activity act on |
| act.datid = pgdat.oid and |
| act.pid = lock.pid and |
| act.state = 'idle' and |
| act.state_change < now() - '1 minute'::interval |
| where objid = $1 and locktype = 'advisory' and granted |
| `), |
|
|
| stuckMessages: p.P(` |
| with sel as ( |
| select id, provider_msg_id |
| from outgoing_messages msg |
| where |
| last_status = 'queued_remotely' and |
| last_status_at < now()-'1 minute'::interval and |
| provider_msg_id notnull |
| order by |
| last_status_at |
| limit 10 |
| for update |
| ) |
| update outgoing_messages msg |
| set last_status_at = now() |
| from sel |
| where msg.id = sel.id |
| returning msg.id, msg.provider_msg_id |
| `), |
|
|
| sentByCMType: p.P(` |
| select count(*) |
| from outgoing_messages msg |
| join user_contact_methods cm on cm.id = msg.contact_method_id |
| where msg.sent_at > $1 and cm.type = $2 |
| `), |
|
|
| cleanupStatusUpdateOptOut: p.P(` |
| delete from outgoing_messages msg |
| using user_contact_methods cm |
| where |
| msg.message_type = 'alert_status_update' and |
| ( |
| msg.last_status = 'pending' or |
| (msg.last_status = 'failed' and msg.next_retry_at notnull) |
| ) and |
| not cm.enable_status_updates and cm.id = msg.contact_method_id |
| `), |
| setSending: p.P(` |
| update outgoing_messages |
| set |
| last_status = 'sending', |
| last_status_at = now(), |
| status_details = '', |
| sending_deadline = now() + '10 seconds'::interval, |
| fired_at = now(), |
| provider_seq = 0, |
| provider_msg_id = null, |
| next_retry_at = null |
| where id = $1 |
| `), |
|
|
| sendDeadlineExpired: p.P(` |
| update outgoing_messages |
| set |
| last_status = 'failed', |
| last_status_at = now(), |
| status_details = 'send deadline expired', |
| cycle_id = null, |
| next_retry_at = null |
| where |
| last_status = 'sending' and |
| sending_deadline <= now() |
| `), |
| retryReset: p.P(` |
| update outgoing_messages |
| set |
| last_status = 'pending', |
| status_details = '', |
| next_retry_at = null, |
| retry_count = retry_count + 1, |
| fired_at = null, |
| sent_at = null, |
| provider_msg_id = null, |
| provider_seq = 0 |
| where |
| last_status = 'failed' and |
| now() > next_retry_at and |
| retry_count < 3 |
| `), |
| retryClear: p.P(` |
| update outgoing_messages |
| set |
| next_retry_at = null, |
| cycle_id = null |
| where |
| last_status = 'failed' and |
| retry_count >= 3 and |
| (cycle_id notnull or next_retry_at notnull) |
| `), |
|
|
| lockStmt: p.P(`lock outgoing_messages in exclusive mode`), |
| currentTime: p.P(`select now()`), |
|
|
| failDisabledCM: p.P(` |
| with disabled as ( |
| update outgoing_messages msg |
| set |
| last_status = 'failed', |
| last_status_at = now(), |
| status_details = 'contact method disabled', |
| cycle_id = null, |
| next_retry_at = null |
| from user_contact_methods cm |
| where |
| msg.last_status = 'pending' and |
| msg.message_type != 'verification_message' and |
| cm.id = msg.contact_method_id and |
| cm.disabled |
| returning msg.id as msg_id, alert_id, msg.user_id, cm.id as cm_id |
| ) select distinct msg_id, alert_id, user_id, cm_id from disabled where alert_id notnull |
| `), |
|
|
| failSMSVoice: p.P(` |
| update outgoing_messages msg |
| set |
| last_status = 'failed', |
| last_status_at = now(), |
| status_details = 'SMS/Voice support not enabled by administrator', |
| cycle_id = null, |
| next_retry_at = null |
| from user_contact_methods cm |
| where |
| msg.last_status = 'pending' and |
| cm.type in ('SMS', 'VOICE') and |
| cm.id = msg.contact_method_id |
| returning msg.id as msg_id, alert_id, msg.user_id, cm.id as cm_id |
| `), |
|
|
| createAlertBundle: p.P(` |
| insert into outgoing_messages ( |
| id, |
| created_at, |
| message_type, |
| contact_method_id, |
| channel_id, |
| user_id, |
| service_id |
| ) values ( |
| $1, $2, 'alert_notification_bundle', $3, $4, $5, $6 |
| ) |
| `), |
|
|
| bundleMessages: p.P(` |
| update outgoing_messages |
| set |
| last_status = 'bundled', |
| last_status_at = now(), |
| status_details = $1, |
| cycle_id = null |
| where id = any($2::uuid[]) |
| `), |
|
|
| deleteAny: p.P(`delete from outgoing_messages where id = any($1)`), |
| }, p.Err |
| } |
|
|
| func (db *DB) currentQueue(ctx context.Context, tx *sql.Tx, now time.Time) (*queue, error) { |
| cutoff := now.Add(-maxThrottleDuration(PerCMThrottle, GlobalCMThrottle)) |
| sentSince := db.lastSent |
| if sentSince.IsZero() { |
| sentSince = cutoff |
| } |
|
|
| rows, err := gadb.New(tx).MessageMgrGetPending(ctx, sql.NullTime{Time: sentSince, Valid: true}) |
| if err != nil { |
| return nil, errors.Wrap(err, "fetch outgoing messages") |
| } |
|
|
| result := make([]Message, 0, len(rows)) |
| for _, row := range rows { |
| var msg Message |
| msg.ID = row.ID.String() |
| msg.Type = row.MessageType |
|
|
| msg.AlertID = int(row.AlertID.Int64) |
| msg.AlertLogID = int(row.AlertLogID.Int64) |
| if row.UserVerificationCodeID.Valid { |
| msg.VerifyID = row.UserVerificationCodeID.UUID.String() |
| } |
| if row.UserID.Valid { |
| msg.UserID = row.UserID.UUID.String() |
| } |
| if row.ServiceID.Valid { |
| msg.ServiceID = row.ServiceID.UUID.String() |
| } |
| msg.CreatedAt = row.CreatedAt |
| msg.SentAt = row.SentAt.Time |
| msg.DestID.CMID = row.CmID |
| msg.DestID.NCID = row.ChanID |
| msg.Dest = row.Dest.DestV1 |
| msg.StatusAlertIDs = row.StatusAlertIds |
| if row.ScheduleID.Valid { |
| msg.ScheduleID = row.ScheduleID.UUID.String() |
| } |
| if msg.Dest.Type == "" { |
| log.Debugf(ctx, "unknown message type for message %s", msg.ID) |
| continue |
| } |
|
|
| if !msg.SentAt.IsZero() { |
| |
| db.sentMessages[msg.ID] = msg |
| continue |
| } |
|
|
| result = append(result, msg) |
| } |
|
|
| for id, msg := range db.sentMessages { |
| if msg.SentAt.Before(cutoff) { |
| delete(db.sentMessages, id) |
| continue |
| } |
| result = append(result, msg) |
| } |
| db.lastSent = now |
|
|
| result, toDelete := dedupOnCallNotifications(result) |
| if len(toDelete) > 0 { |
| _, err = tx.StmtContext(ctx, db.deleteAny).ExecContext(ctx, sqlutil.UUIDArray(toDelete)) |
| if err != nil { |
| return nil, fmt.Errorf("delete duplicate on-call notifications: %w", err) |
| } |
| } |
|
|
| cfg := config.FromContext(ctx) |
| result, toDelete = dedupStatusMessages(result) |
| if len(toDelete) > 0 { |
| _, err = tx.StmtContext(ctx, db.deleteAny).ExecContext(ctx, sqlutil.UUIDArray(toDelete)) |
| if err != nil { |
| return nil, fmt.Errorf("delete duplicate status updates: %w", err) |
| } |
| } |
|
|
| result, err = dedupAlerts(result, func(parentID string, duplicateIDs []string) error { |
| _, err = tx.StmtContext(ctx, db.bundleMessages).ExecContext(ctx, parentID, sqlutil.UUIDArray(duplicateIDs)) |
| if err != nil { |
| return fmt.Errorf("bundle '%v' by pointing to '%s': %w", duplicateIDs, parentID, err) |
| } |
|
|
| return nil |
| }) |
| if err != nil { |
| return nil, fmt.Errorf("dedup alerts: %w", err) |
| } |
|
|
| if cfg.General.DisableMessageBundles { |
| return newQueue(result, now), nil |
| } |
|
|
| result, err = bundleAlertMessages(result, func(msg Message) (string, error) { |
| var userID sql.NullString |
| if msg.UserID != "" { |
| userID.Valid = true |
| userID.String = msg.UserID |
| } |
|
|
| newID := uuid.NewString() |
| _, err := tx.StmtContext(ctx, db.createAlertBundle).ExecContext(ctx, newID, msg.CreatedAt, msg.DestID.CMID, msg.DestID.NCID, userID, msg.ServiceID) |
| if err != nil { |
| return "", err |
| } |
|
|
| return newID, nil |
| }, func(parentID string, ids []string) error { |
| _, err = tx.StmtContext(ctx, db.bundleMessages).ExecContext(ctx, parentID, sqlutil.UUIDArray(ids)) |
| return err |
| }) |
| if err != nil { |
| return nil, err |
| } |
|
|
| return newQueue(result, now), nil |
| } |
|
|
| |
| func (db *DB) UpdateMessageStatus(ctx context.Context, status *notification.SendResult) error { |
| return retry.DoTemporaryError(func(int) error { |
| return db._UpdateMessageStatus(ctx, status) |
| }, |
| retry.Log(ctx), |
| retry.Limit(10), |
| retry.FibBackoff(time.Millisecond*100), |
| ) |
| } |
|
|
| func (db *DB) _UpdateMessageStatus(ctx context.Context, status *notification.SendResult) error { |
| if status == nil { |
| |
| return nil |
| } |
| err := permission.LimitCheckAny(ctx, permission.System) |
| if err != nil { |
| return err |
| } |
| var cbID sql.NullString |
| if status.ID != "" { |
| cbID.Valid = true |
| cbID.String = status.ID |
| } |
|
|
| if status.State == notification.StateFailedTemp { |
| _, err = db.tempFail.ExecContext(ctx, cbID, status.ProviderMessageID, status.Details) |
| return err |
| } |
| if status.State == notification.StateFailedPerm { |
| _, err = db.permFail.ExecContext(ctx, cbID, status.ProviderMessageID, status.Details) |
| return err |
| } |
|
|
| var s Status |
| switch status.State { |
| case notification.StateSending: |
| s = StatusQueuedRemotely |
| case notification.StateSent: |
| s = StatusSent |
| case notification.StateDelivered: |
| s = StatusDelivered |
| case notification.StateRead: |
| s = StatusRead |
| } |
|
|
| var srcValue sql.NullString |
| if status.SrcValue != "" { |
| srcValue.Valid = true |
| srcValue.String = status.SrcValue |
| } |
|
|
| _, err = db.updateStatus.ExecContext(ctx, cbID, status.ProviderMessageID, status.Sequence, s, status.Details, srcValue) |
| return err |
| } |
|
|
| |
| type SendFunc func(context.Context, *Message) (*notification.SendResult, error) |
|
|
| |
| var ErrAbort = errors.New("aborted due to pause") |
|
|
| |
| type StatusFunc func(ctx context.Context, providerID notification.ProviderMessageID) (*notification.Status, error) |
|
|
| |
| func (db *DB) SendMessages(ctx context.Context, send SendFunc, status StatusFunc) error { |
| err := db._SendMessages(ctx, send, status) |
| if db.pausable.IsPausing() { |
| return ErrAbort |
| } |
| return err |
| } |
|
|
| func (db *DB) _SendMessages(ctx context.Context, send SendFunc, status StatusFunc) error { |
| err := permission.LimitCheckAny(ctx, permission.System) |
| if err != nil { |
| return err |
| } |
| log.Debugf(ctx, "Sending outgoing messages.") |
|
|
| execCtx, execCancel := context.WithCancel(ctx) |
| execDone := make(chan struct{}) |
| go func() { |
| select { |
| case <-ctx.Done(): |
| case <-db.pausable.PauseWait(): |
| case <-execDone: |
| } |
| execCancel() |
| }() |
|
|
| res, err := db.advLockCleanup.ExecContext(execCtx, lock.GlobalMessageSending) |
| if err != nil { |
| return errors.Wrap(err, "terminate stale backend locks") |
| } |
| rowsCount, _ := res.RowsAffected() |
| if rowsCount > 0 { |
| log.Log(execCtx, errors.Errorf("terminated %d stale backend instance(s) holding message sending lock", rowsCount)) |
| } |
|
|
| cLock, err := db.lock.Conn(execCtx) |
| if err != nil { |
| return errors.Wrap(err, "get DB conn") |
| } |
| defer cLock.Close() |
|
|
| var gotLock bool |
| err = cLock.WithTx(execCtx, func(tx *sql.Tx) error { |
| return tx.StmtContext(execCtx, db.advLock).QueryRowContext(execCtx, lock.GlobalMessageSending).Scan(&gotLock) |
| }) |
| if err != nil { |
| return errors.Wrap(err, "acquire global sending advisory lock") |
| } |
| if !gotLock { |
| return processinglock.ErrNoLock |
| } |
| defer func() { |
| _, _ = cLock.ExecWithoutLock(log.FromContext(execCtx).BackgroundContext(), `select pg_advisory_unlock(4912)`) |
| }() |
|
|
| tx, err := cLock.BeginTx(execCtx, nil) |
| if err != nil { |
| return errors.Wrap(err, "begin transaction") |
| } |
| defer sqlutil.Rollback(ctx, "engine: message: send", tx) |
|
|
| _, err = tx.Stmt(db.lockStmt).ExecContext(execCtx) |
| if err != nil { |
| return errors.Wrap(err, "acquire exclusive locks") |
| } |
|
|
| var t time.Time |
| err = tx.Stmt(db.currentTime).QueryRowContext(execCtx).Scan(&t) |
| if err != nil { |
| return errors.Wrap(err, "get current time") |
| } |
|
|
| _, err = tx.Stmt(db.cleanupStatusUpdateOptOut).ExecContext(execCtx) |
| if err != nil { |
| return errors.Wrap(err, "clear disabled status updates") |
| } |
|
|
| type msgMeta struct { |
| MessageID string |
| AlertID int |
| UserID string |
| CMID string |
| } |
|
|
| var msgs []msgMeta |
|
|
| |
| cfg := config.FromContext(ctx) |
| if !cfg.Twilio.Enable { |
| rows, err := tx.StmtContext(ctx, db.failSMSVoice).QueryContext(execCtx) |
| if err != nil { |
| return errors.Wrap(err, "check for failed message") |
| } |
| defer rows.Close() |
|
|
| for rows.Next() { |
| var alertID sql.NullInt64 |
| var msg msgMeta |
| err = rows.Scan(&msg.MessageID, &alertID, &msg.UserID, &msg.CMID) |
| if err != nil { |
| return errors.Wrap(err, "scan all failed messages") |
| } |
| if !alertID.Valid { |
| continue |
| } |
| msg.AlertID = int(alertID.Int64) |
| msgs = append(msgs, msg) |
| } |
| } |
|
|
| |
| rows, err := tx.Stmt(db.failDisabledCM).QueryContext(execCtx) |
| if err != nil { |
| return errors.Wrap(err, "check for disabled CMs") |
| } |
| defer rows.Close() |
|
|
| for rows.Next() { |
| var msg msgMeta |
| err = rows.Scan(&msg.MessageID, &msg.AlertID, &msg.UserID, &msg.CMID) |
| if err != nil { |
| return errors.Wrap(err, "scan all disabled CM messages") |
| } |
| msgs = append(msgs, msg) |
| } |
|
|
| for _, m := range msgs { |
| meta := alertlog.NotificationMetaData{ |
| MessageID: m.MessageID, |
| } |
|
|
| |
| db.alertlogstore.MustLogTx(permission.UserSourceContext(ctx, m.UserID, permission.RoleUser, &permission.SourceInfo{ |
| Type: permission.SourceTypeContactMethod, |
| ID: m.CMID, |
| }), tx, m.AlertID, alertlog.TypeNotificationSent, meta) |
| } |
|
|
| _, err = tx.Stmt(db.sendDeadlineExpired).ExecContext(ctx) |
| if err != nil { |
| return errors.Wrap(err, "fail expired messages") |
| } |
|
|
| _, err = tx.Stmt(db.retryClear).ExecContext(ctx) |
| if err != nil { |
| return errors.Wrap(err, "clear max retries") |
| } |
|
|
| _, err = tx.Stmt(db.retryReset).ExecContext(execCtx) |
| if err != nil { |
| return errors.Wrap(err, "reset retry messages") |
| } |
|
|
| q, err := db.currentQueue(ctx, tx, t) |
| if err != nil { |
| return errors.Wrap(err, "get pending messages") |
| } |
|
|
| err = tx.Commit() |
| if err != nil { |
| return errors.Wrap(err, "commit message updates") |
| } |
|
|
| var wg sync.WaitGroup |
| for _, t := range q.Types() { |
| wg.Add(1) |
| go func(typ string) { |
| defer wg.Done() |
| err := db.sendMessagesByType(ctx, cLock, send, q, typ) |
| if err != nil && !errors.Is(err, processinglock.ErrNoLock) { |
| log.Log(ctx, errors.Wrap(err, "send")) |
| } |
| }(t) |
| } |
| wg.Wait() |
|
|
| return db.updateStuckMessages(ctx, status) |
| } |
|
|
| func (db *DB) refreshMessageState(ctx context.Context, statusFn StatusFunc, id string, providerID notification.ProviderMessageID, res chan *notification.SendResult) { |
| ctx, cancel := context.WithTimeout(ctx, 5*time.Second) |
| defer cancel() |
|
|
| status, err := statusFn(ctx, providerID) |
| if errors.Is(err, notification.ErrStatusUnsupported) { |
| |
| res <- nil |
| return |
| } |
| if err != nil { |
| |
| log.Log(ctx, err) |
| res <- nil |
| return |
| } |
|
|
| stat := *status |
| if stat.State == notification.StateFailedTemp { |
| stat.State = notification.StateFailedPerm |
| } |
| stat.Sequence = -1 |
| res <- ¬ification.SendResult{Status: stat, ID: id, ProviderMessageID: providerID} |
| } |
|
|
| func (db *DB) updateStuckMessages(ctx context.Context, statusFn StatusFunc) error { |
| tx, err := db.lock.BeginTx(ctx, nil) |
| if err != nil { |
| return err |
| } |
| defer sqlutil.Rollback(ctx, "message: update stuck messages", tx) |
|
|
| rows, err := tx.Stmt(db.stuckMessages).QueryContext(ctx) |
| if err != nil { |
| return err |
| } |
| defer rows.Close() |
|
|
| type msg struct { |
| id string |
| providerID notification.ProviderMessageID |
| } |
| var toCheck []msg |
| for rows.Next() { |
| var m msg |
| err = rows.Scan(&m.id, &m.providerID) |
| if err != nil { |
| return err |
| } |
| toCheck = append(toCheck, m) |
| } |
|
|
| err = tx.Commit() |
| if err != nil { |
| return err |
| } |
|
|
| ch := make(chan *notification.SendResult, len(toCheck)) |
| for _, m := range toCheck { |
| go db.refreshMessageState(ctx, statusFn, m.id, m.providerID, ch) |
| } |
|
|
| for range toCheck { |
| err := db._UpdateMessageStatus(ctx, <-ch) |
| if err != nil { |
| log.Log(ctx, errors.Wrap(err, "update stale message status")) |
| } |
| } |
|
|
| return nil |
| } |
|
|
| func (db *DB) sendMessagesByType(ctx context.Context, cLock *processinglock.Conn, send SendFunc, q *queue, typ string) error { |
| ch := make(chan error) |
| var count int |
| for { |
| msg := q.NextByType(typ) |
| if msg == nil { |
| break |
| } |
| count++ |
| go func() { |
| _, err := db.sendMessage(ctx, cLock, send, msg) |
| ch <- err |
| }() |
| } |
|
|
| var failed bool |
| for i := 0; i < count; i++ { |
| select { |
| case err := <-ch: |
| if err != nil { |
| log.Log(ctx, fmt.Errorf("send message: %w", err)) |
| failed = true |
| continue |
| } |
| case <-ctx.Done(): |
| return ctx.Err() |
| } |
| } |
|
|
| if failed { |
| return errors.New("one or more failures when sending") |
| } |
|
|
| return nil |
| } |
|
|
| func (db *DB) sendMessage(ctx context.Context, cLock *processinglock.Conn, send SendFunc, m *Message) (bool, error) { |
| ctx = log.WithFields(ctx, log.Fields{ |
| "DestID": m.DestID, |
| "DestType": m.Dest.Type, |
| "CallbackID": m.ID, |
| }) |
|
|
| if m.AlertID != 0 { |
| ctx = log.WithField(ctx, "AlertID", m.AlertID) |
| } |
| _, err := cLock.Exec(ctx, db.setSending, m.ID) |
| if err != nil { |
| return false, err |
| } |
| sCtx, cancel := context.WithTimeout(ctx, 5*time.Second) |
| var status *notification.SendResult |
| err = retry.DoTemporaryError(func(int) error { |
| status, err = send(sCtx, m) |
| return err |
| }, |
| retry.Log(ctx), |
| retry.Limit(10), |
| retry.FibBackoff(65*time.Millisecond), |
| ) |
| cancel() |
|
|
| var pID notification.ProviderMessageID |
| if status != nil { |
| pID = status.ProviderMessageID |
| } |
|
|
| retryExec := func(s *sql.Stmt, args ...interface{}) error { |
| return retry.DoTemporaryError(func(int) error { |
| _, err := s.ExecContext(ctx, args...) |
| return err |
| }, |
| retry.Limit(15), |
| retry.FibBackoff(time.Millisecond*50), |
| ) |
| } |
| if err != nil { |
| log.Log(ctx, errors.Wrap(err, "send message")) |
|
|
| err = retryExec(db.tempFail, m.ID, pID, err.Error()) |
| return false, errors.Wrap(err, "mark failed message") |
| } |
|
|
| if status.State == notification.StateFailedTemp { |
| err = retryExec(db.tempFail, m.ID, pID, status.Details) |
| return false, errors.Wrap(err, "mark failed message (temp)") |
| } |
| if status.State == notification.StateFailedPerm { |
| err = retryExec(db.permFail, m.ID, pID, status.Details) |
| return false, errors.Wrap(err, "mark failed message (perm)") |
| } |
|
|
| return true, errors.Wrap(db.UpdateMessageStatus(ctx, status), "update message status") |
| } |
|
|