fix: migrate to logger.warning usage (#2456)

Signed-off-by: Emmanuel Ferdman <emmanuelferdman@gmail.com>
This commit is contained in:
Emmanuel Ferdman 2025-05-07 17:44:02 +03:00 committed by GitHub
parent a3d7732ed8
commit abc981fa08
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
7 changed files with 10 additions and 10 deletions

View file

@ -187,7 +187,7 @@ class MailSender:
TimeoutError,
) as e:
if retries > 0:
LOG.warn(
LOG.warning(
f"Retrying sending email due to error {e}. {retries} retries left. Will wait {0.3*retries} seconds."
)
time.sleep(0.3 * retries)

View file

@ -920,7 +920,7 @@ def check_custom_domain():
):
alias_count = Alias.filter(Alias.custom_domain_id == custom_domain.id).count()
if alias_count > 0:
LOG.warn(
LOG.warning(
f"Custom Domain {custom_domain} has {alias_count} aliases. Won't delete"
)
else:

View file

@ -34,7 +34,7 @@ class PostgresEventSource(EventSource):
try:
self.__listen(on_event)
except Exception as e:
LOG.warn(f"Error listening to events: {e}")
LOG.warning(f"Error listening to events: {e}")
sleep(_POSTGRES_RECONNECT_INTERVAL_SECONDS)
self.__connect()
@ -68,7 +68,7 @@ class PostgresEventSource(EventSource):
else:
LOG.info(f"Could not find event with id={notify.payload}")
except Exception as e:
LOG.warn(f"Error getting event: {e}")
LOG.warning(f"Error getting event: {e}")
Session.close() # Ensure we get a new connection and we don't leave a dangling tx
def __connect(self):
@ -112,5 +112,5 @@ class DeadLetterEventSource(EventSource):
LOG.debug("No dead letter events")
sleep(_DEAD_LETTER_INTERVAL_SECONDS)
except Exception as e:
LOG.warn(f"Error getting dead letter event: {e}")
LOG.warning(f"Error getting dead letter event: {e}")
sleep(_DEAD_LETTER_INTERVAL_SECONDS)

View file

@ -46,5 +46,5 @@ class Runner:
event.retry_count = event.retry_count + 1
Session.commit()
except Exception as e:
LOG.warn(f"Exception processing event [id={event.id}]: {e}")
LOG.warning(f"Exception processing event [id={event.id}]: {e}")
newrelic.agent.record_custom_metric("Custom/sync_event_failed", 1)

View file

@ -372,13 +372,13 @@ def execute():
jobs_done += 1
LOG.d("Processed job %s", job)
except Exception as e:
LOG.warn(f"Error processing job (id={job.id} name={job.name}): {e}")
LOG.warning(f"Error processing job (id={job.id} name={job.name}): {e}")
# Increment manually, as the attempts increment is done by the take_job but not
# updated in our instance
job_attempts = job.attempts + 1
if job_attempts >= config.JOB_MAX_ATTEMPTS:
LOG.warn(
LOG.warning(
f"Marking job (id={job.id} name={job.name} attempts={job_attempts}) as ERROR"
)
job.state = JobState.error.value

View file

@ -17,4 +17,4 @@ class MetricExporter:
self.__newrelic.send(metrics)
LOG.info("Upcloud metrics sent to NewRelic")
except Exception as e:
LOG.warn(f"Could not export metrics: {e}")
LOG.warning(f"Could not export metrics: {e}")

View file

@ -35,7 +35,7 @@ def get_metric(json: Any, metric: str) -> UpcloudMetric:
UpcloudRecord(time=time, db_role=db_role, label=label, value=value)
)
else:
LOG.warn(f"Could not get value for metric {metric}")
LOG.warning(f"Could not get value for metric {metric}")
return UpcloudMetric(metric_name=metric, records=records)