We use [AutomaticRetry(Attempts = 0)] and [DisableConcurrentExecution(timeoutInSeconds: 1800)] in our job because we don’t expect that a job will be requeued.
But, the job was retried 10 times and it might cause by unstable Redis connection (Azure).
Also, I am wondering why a job ID was processed by a WorkerId and hand over to another new WorkderId ?
Any ideas or thoughts will be appreciated.
An error from Application Insights
Timeout expired. The timeout elapsed prior to obtaining a distributed lock on the ‘Failed to acquire lock on hangfire:production:WriteJob.Perform within given timeout (00:30:00)’ resource.
at Hangfire.Redis.RedisLock.Acquire (Hangfire.Redis.StackExchange, Version=220.127.116.11, Culture=neutral, PublicKeyToken=null)