5c5d01915a
Porting 851/909/910/913/919/920/922/923/924/925/926/927 patches to debian, other patches are confirmed in debian version. Story: 2009221 Task: 43416 Signed-off-by: Yue Tao <yue.tao@windriver.com> Change-Id: I5af677c90342bae7c30991bf465e0db79c71667d
836 lines
32 KiB
Diff
836 lines
32 KiB
Diff
From 69266c451910d2b57313b2fe7561e07cd5400d27 Mon Sep 17 00:00:00 2001
|
|
From: Lennart Poettering <lennart@poettering.net>
|
|
Date: Mon, 23 Nov 2020 18:02:40 +0100
|
|
Subject: [PATCH 19/20] sd-event: add ability to ratelimit event sources
|
|
|
|
Let's a concept of "rate limiting" to event sources: if specific event
|
|
sources fire too often in some time interval temporarily take them
|
|
offline, and take them back online once the interval passed.
|
|
|
|
This is a simple scheme of avoiding starvation of event sources if some
|
|
event source fires too often.
|
|
|
|
This introduces the new conceptual states of "offline" and "online" for
|
|
event sources: an event source is "online" only when enabled *and* not
|
|
ratelimited, and offline in all other cases. An event source that is
|
|
online hence has its fds registered in the epoll, its signals in the
|
|
signalfd and so on.
|
|
|
|
(cherry picked from commit b6d5481b3d9f7c9b1198ab54b54326ec73e855bf)
|
|
|
|
Related: #1819868
|
|
|
|
[commit 395eb7753a9772f505102fbbe3ba3261b57abbe9 from
|
|
https://github.com/systemd-rhel/rhel-8/
|
|
|
|
LZ: Moved the changes in libsystemd.sym to libsystemd.sym.m4 from the
|
|
file changing history; patch ratelimit.h in its old path; dropped
|
|
SOURCE_INOTIFY related parts in sd-event.c because it hasn't been
|
|
added in this systemd version.]
|
|
|
|
Signed-off-by: Li Zhou <li.zhou@windriver.com>
|
|
---
|
|
src/libsystemd/libsystemd.sym | 7 +
|
|
src/libsystemd/sd-event/event-source.h | 6 +
|
|
src/libsystemd/sd-event/sd-event.c | 396 +++++++++++++++++++++----
|
|
src/systemd/sd-event.h | 3 +
|
|
4 files changed, 350 insertions(+), 62 deletions(-)
|
|
|
|
diff --git a/src/libsystemd/libsystemd.sym b/src/libsystemd/libsystemd.sym
|
|
index f83b364c96..b03bcd952f 100644
|
|
--- a/src/libsystemd/libsystemd.sym
|
|
+++ b/src/libsystemd/libsystemd.sym
|
|
@@ -736,3 +736,10 @@ global:
|
|
sd_device_has_current_tag;
|
|
sd_device_set_sysattr_valuef;
|
|
} LIBSYSTEMD_246;
|
|
+
|
|
+LIBSYSTEMD_248 {
|
|
+global:
|
|
+ sd_event_source_set_ratelimit;
|
|
+ sd_event_source_get_ratelimit;
|
|
+ sd_event_source_is_ratelimited;
|
|
+} LIBSYSTEMD_246;
|
|
diff --git a/src/libsystemd/sd-event/event-source.h b/src/libsystemd/sd-event/event-source.h
|
|
index 189d3b48df..f0d2a1b9e6 100644
|
|
--- a/src/libsystemd/sd-event/event-source.h
|
|
+++ b/src/libsystemd/sd-event/event-source.h
|
|
@@ -11,6 +11,7 @@
|
|
#include "hashmap.h"
|
|
#include "list.h"
|
|
#include "prioq.h"
|
|
+#include "ratelimit.h"
|
|
|
|
typedef enum EventSourceType {
|
|
SOURCE_IO,
|
|
@@ -61,6 +62,7 @@ struct sd_event_source {
|
|
bool dispatching:1;
|
|
bool floating:1;
|
|
bool exit_on_failure:1;
|
|
+ bool ratelimited:1;
|
|
|
|
int64_t priority;
|
|
unsigned pending_index;
|
|
@@ -72,6 +74,10 @@ struct sd_event_source {
|
|
|
|
LIST_FIELDS(sd_event_source, sources);
|
|
|
|
+ RateLimit rate_limit;
|
|
+
|
|
+ /* These are primarily fields relevant for time event sources, but since any event source can
|
|
+ * effectively become one when rate-limited, this is part of the common fields. */
|
|
unsigned earliest_index;
|
|
unsigned latest_index;
|
|
|
|
diff --git a/src/libsystemd/sd-event/sd-event.c b/src/libsystemd/sd-event/sd-event.c
|
|
index 39f13cb409..3f1a6776fe 100644
|
|
--- a/src/libsystemd/sd-event/sd-event.c
|
|
+++ b/src/libsystemd/sd-event/sd-event.c
|
|
@@ -37,6 +37,16 @@ static bool EVENT_SOURCE_WATCH_PIDFD(sd_event_source *s) {
|
|
s->child.options == WEXITED;
|
|
}
|
|
|
|
+static bool event_source_is_online(sd_event_source *s) {
|
|
+ assert(s);
|
|
+ return s->enabled != SD_EVENT_OFF && !s->ratelimited;
|
|
+}
|
|
+
|
|
+static bool event_source_is_offline(sd_event_source *s) {
|
|
+ assert(s);
|
|
+ return s->enabled == SD_EVENT_OFF || s->ratelimited;
|
|
+}
|
|
+
|
|
static const char* const event_source_type_table[_SOURCE_EVENT_SOURCE_TYPE_MAX] = {
|
|
[SOURCE_IO] = "io",
|
|
[SOURCE_TIME_REALTIME] = "realtime",
|
|
@@ -55,7 +65,25 @@ static const char* const event_source_type_table[_SOURCE_EVENT_SOURCE_TYPE_MAX]
|
|
|
|
DEFINE_PRIVATE_STRING_TABLE_LOOKUP_TO_STRING(event_source_type, int);
|
|
|
|
-#define EVENT_SOURCE_IS_TIME(t) IN_SET((t), SOURCE_TIME_REALTIME, SOURCE_TIME_BOOTTIME, SOURCE_TIME_MONOTONIC, SOURCE_TIME_REALTIME_ALARM, SOURCE_TIME_BOOTTIME_ALARM)
|
|
+#define EVENT_SOURCE_IS_TIME(t) \
|
|
+ IN_SET((t), \
|
|
+ SOURCE_TIME_REALTIME, \
|
|
+ SOURCE_TIME_BOOTTIME, \
|
|
+ SOURCE_TIME_MONOTONIC, \
|
|
+ SOURCE_TIME_REALTIME_ALARM, \
|
|
+ SOURCE_TIME_BOOTTIME_ALARM)
|
|
+
|
|
+#define EVENT_SOURCE_CAN_RATE_LIMIT(t) \
|
|
+ IN_SET((t), \
|
|
+ SOURCE_IO, \
|
|
+ SOURCE_TIME_REALTIME, \
|
|
+ SOURCE_TIME_BOOTTIME, \
|
|
+ SOURCE_TIME_MONOTONIC, \
|
|
+ SOURCE_TIME_REALTIME_ALARM, \
|
|
+ SOURCE_TIME_BOOTTIME_ALARM, \
|
|
+ SOURCE_SIGNAL, \
|
|
+ SOURCE_DEFER, \
|
|
+ SOURCE_INOTIFY)
|
|
|
|
struct sd_event {
|
|
unsigned n_ref;
|
|
@@ -81,7 +109,7 @@ struct sd_event {
|
|
Hashmap *signal_data; /* indexed by priority */
|
|
|
|
Hashmap *child_sources;
|
|
- unsigned n_enabled_child_sources;
|
|
+ unsigned n_online_child_sources;
|
|
|
|
Set *post_sources;
|
|
|
|
@@ -146,6 +174,11 @@ static int pending_prioq_compare(const void *a, const void *b) {
|
|
if (x->enabled == SD_EVENT_OFF && y->enabled != SD_EVENT_OFF)
|
|
return 1;
|
|
|
|
+ /* Non rate-limited ones first. */
|
|
+ r = CMP(!!x->ratelimited, !!y->ratelimited);
|
|
+ if (r != 0)
|
|
+ return r;
|
|
+
|
|
/* Lower priority values first */
|
|
r = CMP(x->priority, y->priority);
|
|
if (r != 0)
|
|
@@ -168,6 +201,11 @@ static int prepare_prioq_compare(const void *a, const void *b) {
|
|
if (x->enabled == SD_EVENT_OFF && y->enabled != SD_EVENT_OFF)
|
|
return 1;
|
|
|
|
+ /* Non rate-limited ones first. */
|
|
+ r = CMP(!!x->ratelimited, !!y->ratelimited);
|
|
+ if (r != 0)
|
|
+ return r;
|
|
+
|
|
/* Move most recently prepared ones last, so that we can stop
|
|
* preparing as soon as we hit one that has already been
|
|
* prepared in the current iteration */
|
|
@@ -179,12 +217,30 @@ static int prepare_prioq_compare(const void *a, const void *b) {
|
|
return CMP(x->priority, y->priority);
|
|
}
|
|
|
|
+static usec_t time_event_source_next(const sd_event_source *s) {
|
|
+ assert(s);
|
|
+
|
|
+ /* We have two kinds of event sources that have elapsation times associated with them: the actual
|
|
+ * time based ones and the ones for which a ratelimit can be in effect (where we want to be notified
|
|
+ * once the ratelimit time window ends). Let's return the next elapsing time depending on what we are
|
|
+ * looking at here. */
|
|
+
|
|
+ if (s->ratelimited) { /* If rate-limited the next elapsation is when the ratelimit time window ends */
|
|
+ assert(s->rate_limit.begin != 0);
|
|
+ assert(s->rate_limit.interval != 0);
|
|
+ return usec_add(s->rate_limit.begin, s->rate_limit.interval);
|
|
+ }
|
|
+
|
|
+ /* Otherwise this must be a time event source, if not ratelimited */
|
|
+ if (EVENT_SOURCE_IS_TIME(s->type))
|
|
+ return s->time.next;
|
|
+
|
|
+ return USEC_INFINITY;
|
|
+}
|
|
+
|
|
static int earliest_time_prioq_compare(const void *a, const void *b) {
|
|
const sd_event_source *x = a, *y = b;
|
|
|
|
- assert(EVENT_SOURCE_IS_TIME(x->type));
|
|
- assert(x->type == y->type);
|
|
-
|
|
/* Enabled ones first */
|
|
if (x->enabled != SD_EVENT_OFF && y->enabled == SD_EVENT_OFF)
|
|
return -1;
|
|
@@ -198,19 +254,30 @@ static int earliest_time_prioq_compare(const void *a, const void *b) {
|
|
return 1;
|
|
|
|
/* Order by time */
|
|
- return CMP(x->time.next, y->time.next);
|
|
+ return CMP(time_event_source_next(x), time_event_source_next(y));
|
|
}
|
|
|
|
static usec_t time_event_source_latest(const sd_event_source *s) {
|
|
- return usec_add(s->time.next, s->time.accuracy);
|
|
+ assert(s);
|
|
+
|
|
+ if (s->ratelimited) { /* For ratelimited stuff the earliest and the latest time shall actually be the
|
|
+ * same, as we should avoid adding additional inaccuracy on an inaccuracy time
|
|
+ * window */
|
|
+ assert(s->rate_limit.begin != 0);
|
|
+ assert(s->rate_limit.interval != 0);
|
|
+ return usec_add(s->rate_limit.begin, s->rate_limit.interval);
|
|
+ }
|
|
+
|
|
+ /* Must be a time event source, if not ratelimited */
|
|
+ if (EVENT_SOURCE_IS_TIME(s->type))
|
|
+ return usec_add(s->time.next, s->time.accuracy);
|
|
+
|
|
+ return USEC_INFINITY;
|
|
}
|
|
|
|
static int latest_time_prioq_compare(const void *a, const void *b) {
|
|
const sd_event_source *x = a, *y = b;
|
|
|
|
- assert(EVENT_SOURCE_IS_TIME(x->type));
|
|
- assert(x->type == y->type);
|
|
-
|
|
/* Enabled ones first */
|
|
if (x->enabled != SD_EVENT_OFF && y->enabled == SD_EVENT_OFF)
|
|
return -1;
|
|
@@ -661,12 +728,12 @@ static void event_gc_signal_data(sd_event *e, const int64_t *priority, int sig)
|
|
* and possibly drop the signalfd for it. */
|
|
|
|
if (sig == SIGCHLD &&
|
|
- e->n_enabled_child_sources > 0)
|
|
+ e->n_online_child_sources > 0)
|
|
return;
|
|
|
|
if (e->signal_sources &&
|
|
e->signal_sources[sig] &&
|
|
- e->signal_sources[sig]->enabled != SD_EVENT_OFF)
|
|
+ event_source_is_online(e->signal_sources[sig]))
|
|
return;
|
|
|
|
/*
|
|
@@ -713,11 +780,17 @@ static void event_source_time_prioq_reshuffle(sd_event_source *s) {
|
|
struct clock_data *d;
|
|
|
|
assert(s);
|
|
- assert(EVENT_SOURCE_IS_TIME(s->type));
|
|
|
|
/* Called whenever the event source's timer ordering properties changed, i.e. time, accuracy,
|
|
* pending, enable state. Makes sure the two prioq's are ordered properly again. */
|
|
- assert_se(d = event_get_clock_data(s->event, s->type));
|
|
+
|
|
+ if (s->ratelimited)
|
|
+ d = &s->event->monotonic;
|
|
+ else {
|
|
+ assert(EVENT_SOURCE_IS_TIME(s->type));
|
|
+ assert_se(d = event_get_clock_data(s->event, s->type));
|
|
+ }
|
|
+
|
|
prioq_reshuffle(d->earliest, s, &s->earliest_index);
|
|
prioq_reshuffle(d->latest, s, &s->latest_index);
|
|
d->needs_rearm = true;
|
|
@@ -758,12 +831,18 @@ static void source_disconnect(sd_event_source *s) {
|
|
case SOURCE_TIME_BOOTTIME:
|
|
case SOURCE_TIME_MONOTONIC:
|
|
case SOURCE_TIME_REALTIME_ALARM:
|
|
- case SOURCE_TIME_BOOTTIME_ALARM: {
|
|
- struct clock_data *d;
|
|
- assert_se(d = event_get_clock_data(s->event, s->type));
|
|
- event_source_time_prioq_remove(s, d);
|
|
+ case SOURCE_TIME_BOOTTIME_ALARM:
|
|
+ /* Only remove this event source from the time event source here if it is not ratelimited. If
|
|
+ * it is ratelimited, we'll remove it below, separately. Why? Because the clock used might
|
|
+ * differ: ratelimiting always uses CLOCK_MONOTONIC, but timer events might use any clock */
|
|
+
|
|
+ if (!s->ratelimited) {
|
|
+ struct clock_data *d;
|
|
+ assert_se(d = event_get_clock_data(s->event, s->type));
|
|
+ event_source_time_prioq_remove(s, d);
|
|
+ }
|
|
+
|
|
break;
|
|
- }
|
|
|
|
case SOURCE_SIGNAL:
|
|
if (s->signal.sig > 0) {
|
|
@@ -778,9 +857,9 @@ static void source_disconnect(sd_event_source *s) {
|
|
|
|
case SOURCE_CHILD:
|
|
if (s->child.pid > 0) {
|
|
- if (s->enabled != SD_EVENT_OFF) {
|
|
- assert(s->event->n_enabled_child_sources > 0);
|
|
- s->event->n_enabled_child_sources--;
|
|
+ if (event_source_is_online(s)) {
|
|
+ assert(s->event->n_online_child_sources > 0);
|
|
+ s->event->n_online_child_sources--;
|
|
}
|
|
|
|
(void) hashmap_remove(s->event->child_sources, PID_TO_PTR(s->child.pid));
|
|
@@ -850,6 +929,9 @@ static void source_disconnect(sd_event_source *s) {
|
|
if (s->prepare)
|
|
prioq_remove(s->event->prepare, s, &s->prepare_index);
|
|
|
|
+ if (s->ratelimited)
|
|
+ event_source_time_prioq_remove(s, &s->event->monotonic);
|
|
+
|
|
event = TAKE_PTR(s->event);
|
|
LIST_REMOVE(sources, event->sources, s);
|
|
event->n_sources--;
|
|
@@ -1322,7 +1404,7 @@ _public_ int sd_event_add_child(
|
|
if (!callback)
|
|
callback = child_exit_callback;
|
|
|
|
- if (e->n_enabled_child_sources == 0) {
|
|
+ if (e->n_online_child_sources == 0) {
|
|
/* Caller must block SIGCHLD before using us to watch children, even if pidfd is available,
|
|
* for compatibility with pre-pidfd and because we don't want the reap the child processes
|
|
* ourselves, i.e. call waitid(), and don't want Linux' default internal logic for that to
|
|
@@ -1387,7 +1469,7 @@ _public_ int sd_event_add_child(
|
|
e->need_process_child = true;
|
|
}
|
|
|
|
- e->n_enabled_child_sources++;
|
|
+ e->n_online_child_sources++;
|
|
|
|
if (ret)
|
|
*ret = s;
|
|
@@ -1419,7 +1501,7 @@ _public_ int sd_event_add_child_pidfd(
|
|
if (!callback)
|
|
callback = child_exit_callback;
|
|
|
|
- if (e->n_enabled_child_sources == 0) {
|
|
+ if (e->n_online_child_sources == 0) {
|
|
r = signal_is_blocked(SIGCHLD);
|
|
if (r < 0)
|
|
return r;
|
|
@@ -1469,7 +1551,7 @@ _public_ int sd_event_add_child_pidfd(
|
|
e->need_process_child = true;
|
|
}
|
|
|
|
- e->n_enabled_child_sources++;
|
|
+ e->n_online_child_sources++;
|
|
|
|
if (ret)
|
|
*ret = s;
|
|
@@ -2055,7 +2137,7 @@ _public_ int sd_event_source_set_io_fd(sd_event_source *s, int fd) {
|
|
if (s->io.fd == fd)
|
|
return 0;
|
|
|
|
- if (s->enabled == SD_EVENT_OFF) {
|
|
+ if (event_source_is_offline(s)) {
|
|
s->io.fd = fd;
|
|
s->io.registered = false;
|
|
} else {
|
|
@@ -2122,7 +2204,7 @@ _public_ int sd_event_source_set_io_events(sd_event_source *s, uint32_t events)
|
|
if (r < 0)
|
|
return r;
|
|
|
|
- if (s->enabled != SD_EVENT_OFF) {
|
|
+ if (event_source_is_online(s)) {
|
|
r = source_io_register(s, s->enabled, events);
|
|
if (r < 0)
|
|
return r;
|
|
@@ -2225,7 +2307,7 @@ _public_ int sd_event_source_set_priority(sd_event_source *s, int64_t priority)
|
|
|
|
event_gc_inode_data(s->event, old_inode_data);
|
|
|
|
- } else if (s->type == SOURCE_SIGNAL && s->enabled != SD_EVENT_OFF) {
|
|
+ } else if (s->type == SOURCE_SIGNAL && event_source_is_online(s)) {
|
|
struct signal_data *old, *d;
|
|
|
|
/* Move us from the signalfd belonging to the old
|
|
@@ -2272,20 +2354,29 @@ _public_ int sd_event_source_get_enabled(sd_event_source *s, int *ret) {
|
|
return s->enabled != SD_EVENT_OFF;
|
|
}
|
|
|
|
-static int event_source_disable(sd_event_source *s) {
|
|
+static int event_source_offline(
|
|
+ sd_event_source *s,
|
|
+ int enabled,
|
|
+ bool ratelimited) {
|
|
+
|
|
+ bool was_offline;
|
|
int r;
|
|
|
|
assert(s);
|
|
- assert(s->enabled != SD_EVENT_OFF);
|
|
+ assert(enabled == SD_EVENT_OFF || ratelimited);
|
|
|
|
/* Unset the pending flag when this event source is disabled */
|
|
- if (!IN_SET(s->type, SOURCE_DEFER, SOURCE_EXIT)) {
|
|
+ if (s->enabled != SD_EVENT_OFF &&
|
|
+ enabled == SD_EVENT_OFF &&
|
|
+ !IN_SET(s->type, SOURCE_DEFER, SOURCE_EXIT)) {
|
|
r = source_set_pending(s, false);
|
|
if (r < 0)
|
|
return r;
|
|
}
|
|
|
|
- s->enabled = SD_EVENT_OFF;
|
|
+ was_offline = event_source_is_offline(s);
|
|
+ s->enabled = enabled;
|
|
+ s->ratelimited = ratelimited;
|
|
|
|
switch (s->type) {
|
|
|
|
@@ -2306,8 +2397,10 @@ static int event_source_disable(sd_event_source *s) {
|
|
break;
|
|
|
|
case SOURCE_CHILD:
|
|
- assert(s->event->n_enabled_child_sources > 0);
|
|
- s->event->n_enabled_child_sources--;
|
|
+ if (!was_offline) {
|
|
+ assert(s->event->n_online_child_sources > 0);
|
|
+ s->event->n_online_child_sources--;
|
|
+ }
|
|
|
|
if (EVENT_SOURCE_WATCH_PIDFD(s))
|
|
source_child_pidfd_unregister(s);
|
|
@@ -2328,26 +2421,42 @@ static int event_source_disable(sd_event_source *s) {
|
|
assert_not_reached("Wut? I shouldn't exist.");
|
|
}
|
|
|
|
- return 0;
|
|
+ return 1;
|
|
}
|
|
|
|
-static int event_source_enable(sd_event_source *s, int enable) {
|
|
+static int event_source_online(
|
|
+ sd_event_source *s,
|
|
+ int enabled,
|
|
+ bool ratelimited) {
|
|
+
|
|
+ bool was_online;
|
|
int r;
|
|
|
|
assert(s);
|
|
- assert(IN_SET(enable, SD_EVENT_ON, SD_EVENT_ONESHOT));
|
|
- assert(s->enabled == SD_EVENT_OFF);
|
|
+ assert(enabled != SD_EVENT_OFF || !ratelimited);
|
|
|
|
/* Unset the pending flag when this event source is enabled */
|
|
- if (!IN_SET(s->type, SOURCE_DEFER, SOURCE_EXIT)) {
|
|
+ if (s->enabled == SD_EVENT_OFF &&
|
|
+ enabled != SD_EVENT_OFF &&
|
|
+ !IN_SET(s->type, SOURCE_DEFER, SOURCE_EXIT)) {
|
|
r = source_set_pending(s, false);
|
|
if (r < 0)
|
|
return r;
|
|
}
|
|
|
|
+ /* Are we really ready for onlining? */
|
|
+ if (enabled == SD_EVENT_OFF || ratelimited) {
|
|
+ /* Nope, we are not ready for onlining, then just update the precise state and exit */
|
|
+ s->enabled = enabled;
|
|
+ s->ratelimited = ratelimited;
|
|
+ return 0;
|
|
+ }
|
|
+
|
|
+ was_online = event_source_is_online(s);
|
|
+
|
|
switch (s->type) {
|
|
case SOURCE_IO:
|
|
- r = source_io_register(s, enable, s->io.events);
|
|
+ r = source_io_register(s, enabled, s->io.events);
|
|
if (r < 0)
|
|
return r;
|
|
break;
|
|
@@ -2365,7 +2474,7 @@ static int event_source_enable(sd_event_source *s, int enable) {
|
|
if (EVENT_SOURCE_WATCH_PIDFD(s)) {
|
|
/* yes, we have pidfd */
|
|
|
|
- r = source_child_pidfd_register(s, enable);
|
|
+ r = source_child_pidfd_register(s, enabled);
|
|
if (r < 0)
|
|
return r;
|
|
} else {
|
|
@@ -2378,8 +2487,8 @@ static int event_source_enable(sd_event_source *s, int enable) {
|
|
}
|
|
}
|
|
|
|
- s->event->n_enabled_child_sources++;
|
|
-
|
|
+ if (!was_online)
|
|
+ s->event->n_online_child_sources++;
|
|
break;
|
|
|
|
case SOURCE_TIME_REALTIME:
|
|
@@ -2397,7 +2506,8 @@ static int event_source_enable(sd_event_source *s, int enable) {
|
|
assert_not_reached("Wut? I shouldn't exist.");
|
|
}
|
|
|
|
- s->enabled = enable;
|
|
+ s->enabled = enabled;
|
|
+ s->ratelimited = ratelimited;
|
|
|
|
/* Non-failing operations below */
|
|
switch (s->type) {
|
|
@@ -2417,7 +2527,7 @@ static int event_source_enable(sd_event_source *s, int enable) {
|
|
break;
|
|
}
|
|
|
|
- return 0;
|
|
+ return 1;
|
|
}
|
|
|
|
_public_ int sd_event_source_set_enabled(sd_event_source *s, int m) {
|
|
@@ -2435,7 +2545,7 @@ _public_ int sd_event_source_set_enabled(sd_event_source *s, int m) {
|
|
return 0;
|
|
|
|
if (m == SD_EVENT_OFF)
|
|
- r = event_source_disable(s);
|
|
+ r = event_source_offline(s, m, s->ratelimited);
|
|
else {
|
|
if (s->enabled != SD_EVENT_OFF) {
|
|
/* Switching from "on" to "oneshot" or back? If that's the case, we can take a shortcut, the
|
|
@@ -2444,7 +2554,7 @@ _public_ int sd_event_source_set_enabled(sd_event_source *s, int m) {
|
|
return 0;
|
|
}
|
|
|
|
- r = event_source_enable(s, m);
|
|
+ r = event_source_online(s, m, s->ratelimited);
|
|
}
|
|
if (r < 0)
|
|
return r;
|
|
@@ -2701,6 +2811,96 @@ _public_ void *sd_event_source_set_userdata(sd_event_source *s, void *userdata)
|
|
return ret;
|
|
}
|
|
|
|
+static int event_source_enter_ratelimited(sd_event_source *s) {
|
|
+ int r;
|
|
+
|
|
+ assert(s);
|
|
+
|
|
+ /* When an event source becomes ratelimited, we place it in the CLOCK_MONOTONIC priority queue, with
|
|
+ * the end of the rate limit time window, much as if it was a timer event source. */
|
|
+
|
|
+ if (s->ratelimited)
|
|
+ return 0; /* Already ratelimited, this is a NOP hence */
|
|
+
|
|
+ /* Make sure we can install a CLOCK_MONOTONIC event further down. */
|
|
+ r = setup_clock_data(s->event, &s->event->monotonic, CLOCK_MONOTONIC);
|
|
+ if (r < 0)
|
|
+ return r;
|
|
+
|
|
+ /* Timer event sources are already using the earliest/latest queues for the timer scheduling. Let's
|
|
+ * first remove them from the prioq appropriate for their own clock, so that we can use the prioq
|
|
+ * fields of the event source then for adding it to the CLOCK_MONOTONIC prioq instead. */
|
|
+ if (EVENT_SOURCE_IS_TIME(s->type))
|
|
+ event_source_time_prioq_remove(s, event_get_clock_data(s->event, s->type));
|
|
+
|
|
+ /* Now, let's add the event source to the monotonic clock instead */
|
|
+ r = event_source_time_prioq_put(s, &s->event->monotonic);
|
|
+ if (r < 0)
|
|
+ goto fail;
|
|
+
|
|
+ /* And let's take the event source officially offline */
|
|
+ r = event_source_offline(s, s->enabled, /* ratelimited= */ true);
|
|
+ if (r < 0) {
|
|
+ event_source_time_prioq_remove(s, &s->event->monotonic);
|
|
+ goto fail;
|
|
+ }
|
|
+
|
|
+ event_source_pp_prioq_reshuffle(s);
|
|
+
|
|
+ log_debug("Event source %p (%s) entered rate limit state.", s, strna(s->description));
|
|
+ return 0;
|
|
+
|
|
+fail:
|
|
+ /* Reinstall time event sources in the priority queue as before. This shouldn't fail, since the queue
|
|
+ * space for it should already be allocated. */
|
|
+ if (EVENT_SOURCE_IS_TIME(s->type))
|
|
+ assert_se(event_source_time_prioq_put(s, event_get_clock_data(s->event, s->type)) >= 0);
|
|
+
|
|
+ return r;
|
|
+}
|
|
+
|
|
+static int event_source_leave_ratelimit(sd_event_source *s) {
|
|
+ int r;
|
|
+
|
|
+ assert(s);
|
|
+
|
|
+ if (!s->ratelimited)
|
|
+ return 0;
|
|
+
|
|
+ /* Let's take the event source out of the monotonic prioq first. */
|
|
+ event_source_time_prioq_remove(s, &s->event->monotonic);
|
|
+
|
|
+ /* Let's then add the event source to its native clock prioq again — if this is a timer event source */
|
|
+ if (EVENT_SOURCE_IS_TIME(s->type)) {
|
|
+ r = event_source_time_prioq_put(s, event_get_clock_data(s->event, s->type));
|
|
+ if (r < 0)
|
|
+ goto fail;
|
|
+ }
|
|
+
|
|
+ /* Let's try to take it online again. */
|
|
+ r = event_source_online(s, s->enabled, /* ratelimited= */ false);
|
|
+ if (r < 0) {
|
|
+ /* Do something roughly sensible when this failed: undo the two prioq ops above */
|
|
+ if (EVENT_SOURCE_IS_TIME(s->type))
|
|
+ event_source_time_prioq_remove(s, event_get_clock_data(s->event, s->type));
|
|
+
|
|
+ goto fail;
|
|
+ }
|
|
+
|
|
+ event_source_pp_prioq_reshuffle(s);
|
|
+ ratelimit_reset(&s->rate_limit);
|
|
+
|
|
+ log_debug("Event source %p (%s) left rate limit state.", s, strna(s->description));
|
|
+ return 0;
|
|
+
|
|
+fail:
|
|
+ /* Do something somewhat reasonable when we cannot move an event sources out of ratelimited mode:
|
|
+ * simply put it back in it, maybe we can then process it more successfully next iteration. */
|
|
+ assert_se(event_source_time_prioq_put(s, &s->event->monotonic) >= 0);
|
|
+
|
|
+ return r;
|
|
+}
|
|
+
|
|
static usec_t sleep_between(sd_event *e, usec_t a, usec_t b) {
|
|
usec_t c;
|
|
assert(e);
|
|
@@ -2798,7 +2998,7 @@ static int event_arm_timer(
|
|
d->needs_rearm = false;
|
|
|
|
a = prioq_peek(d->earliest);
|
|
- if (!a || a->enabled == SD_EVENT_OFF || a->time.next == USEC_INFINITY) {
|
|
+ if (!a || a->enabled == SD_EVENT_OFF || time_event_source_next(a) == USEC_INFINITY) {
|
|
|
|
if (d->fd < 0)
|
|
return 0;
|
|
@@ -2817,7 +3017,7 @@ static int event_arm_timer(
|
|
b = prioq_peek(d->latest);
|
|
assert_se(b && b->enabled != SD_EVENT_OFF);
|
|
|
|
- t = sleep_between(e, a->time.next, time_event_source_latest(b));
|
|
+ t = sleep_between(e, time_event_source_next(a), time_event_source_latest(b));
|
|
if (d->next == t)
|
|
return 0;
|
|
|
|
@@ -2895,10 +3095,22 @@ static int process_timer(
|
|
|
|
for (;;) {
|
|
s = prioq_peek(d->earliest);
|
|
- if (!s ||
|
|
- s->time.next > n ||
|
|
- s->enabled == SD_EVENT_OFF ||
|
|
- s->pending)
|
|
+ if (!s || time_event_source_next(s) > n)
|
|
+ break;
|
|
+
|
|
+ if (s->ratelimited) {
|
|
+ /* This is an event sources whose ratelimit window has ended. Let's turn it on
|
|
+ * again. */
|
|
+ assert(s->ratelimited);
|
|
+
|
|
+ r = event_source_leave_ratelimit(s);
|
|
+ if (r < 0)
|
|
+ return r;
|
|
+
|
|
+ continue;
|
|
+ }
|
|
+
|
|
+ if (s->enabled == SD_EVENT_OFF || s->pending)
|
|
break;
|
|
|
|
r = source_set_pending(s, true);
|
|
@@ -2943,7 +3155,7 @@ static int process_child(sd_event *e) {
|
|
if (s->pending)
|
|
continue;
|
|
|
|
- if (s->enabled == SD_EVENT_OFF)
|
|
+ if (event_source_is_offline(s))
|
|
continue;
|
|
|
|
if (s->child.exited)
|
|
@@ -2990,7 +3202,7 @@ static int process_pidfd(sd_event *e, sd_event_source *s, uint32_t revents) {
|
|
if (s->pending)
|
|
return 0;
|
|
|
|
- if (s->enabled == SD_EVENT_OFF)
|
|
+ if (event_source_is_offline(s))
|
|
return 0;
|
|
|
|
if (!EVENT_SOURCE_WATCH_PIDFD(s))
|
|
@@ -3150,7 +3362,7 @@ static int event_inotify_data_process(sd_event *e, struct inotify_data *d) {
|
|
|
|
LIST_FOREACH(inotify.by_inode_data, s, inode_data->event_sources) {
|
|
|
|
- if (s->enabled == SD_EVENT_OFF)
|
|
+ if (event_source_is_offline(s))
|
|
continue;
|
|
|
|
r = source_set_pending(s, true);
|
|
@@ -3186,7 +3398,7 @@ static int event_inotify_data_process(sd_event *e, struct inotify_data *d) {
|
|
* sources if IN_IGNORED or IN_UNMOUNT is set. */
|
|
LIST_FOREACH(inotify.by_inode_data, s, inode_data->event_sources) {
|
|
|
|
- if (s->enabled == SD_EVENT_OFF)
|
|
+ if (event_source_is_offline(s))
|
|
continue;
|
|
|
|
if ((d->buffer.ev.mask & (IN_IGNORED|IN_UNMOUNT)) == 0 &&
|
|
@@ -3240,6 +3452,16 @@ static int source_dispatch(sd_event_source *s) {
|
|
* callback might have invalidated/disconnected the event source. */
|
|
saved_event = sd_event_ref(s->event);
|
|
|
|
+ /* Check if we hit the ratelimit for this event source, if so, let's disable it. */
|
|
+ assert(!s->ratelimited);
|
|
+ if (!ratelimit_below(&s->rate_limit)) {
|
|
+ r = event_source_enter_ratelimited(s);
|
|
+ if (r < 0)
|
|
+ return r;
|
|
+
|
|
+ return 1;
|
|
+ }
|
|
+
|
|
if (!IN_SET(s->type, SOURCE_DEFER, SOURCE_EXIT)) {
|
|
r = source_set_pending(s, false);
|
|
if (r < 0)
|
|
@@ -3253,7 +3475,7 @@ static int source_dispatch(sd_event_source *s) {
|
|
* post sources as pending */
|
|
|
|
SET_FOREACH(z, s->event->post_sources) {
|
|
- if (z->enabled == SD_EVENT_OFF)
|
|
+ if (event_source_is_offline(z))
|
|
continue;
|
|
|
|
r = source_set_pending(z, true);
|
|
@@ -3373,7 +3595,7 @@ static int event_prepare(sd_event *e) {
|
|
sd_event_source *s;
|
|
|
|
s = prioq_peek(e->prepare);
|
|
- if (!s || s->prepare_iteration == e->iteration || s->enabled == SD_EVENT_OFF)
|
|
+ if (!s || s->prepare_iteration == e->iteration || event_source_is_offline(s))
|
|
break;
|
|
|
|
s->prepare_iteration = e->iteration;
|
|
@@ -3413,7 +3635,7 @@ static int dispatch_exit(sd_event *e) {
|
|
assert(e);
|
|
|
|
p = prioq_peek(e->exit);
|
|
- if (!p || p->enabled == SD_EVENT_OFF) {
|
|
+ if (!p || event_source_is_offline(p)) {
|
|
e->state = SD_EVENT_FINISHED;
|
|
return 0;
|
|
}
|
|
@@ -3435,7 +3657,7 @@ static sd_event_source* event_next_pending(sd_event *e) {
|
|
if (!p)
|
|
return NULL;
|
|
|
|
- if (p->enabled == SD_EVENT_OFF)
|
|
+ if (event_source_is_offline(p))
|
|
return NULL;
|
|
|
|
return p;
|
|
@@ -4049,3 +4271,53 @@ _public_ int sd_event_source_set_exit_on_failure(sd_event_source *s, int b) {
|
|
s->exit_on_failure = b;
|
|
return 1;
|
|
}
|
|
+
|
|
+_public_ int sd_event_source_set_ratelimit(sd_event_source *s, uint64_t interval, unsigned burst) {
|
|
+ int r;
|
|
+
|
|
+ assert_return(s, -EINVAL);
|
|
+
|
|
+ /* Turning on ratelimiting on event source types that don't support it, is a loggable offense. Doing
|
|
+ * so is a programming error. */
|
|
+ assert_return(EVENT_SOURCE_CAN_RATE_LIMIT(s->type), -EDOM);
|
|
+
|
|
+ /* When ratelimiting is configured we'll always reset the rate limit state first and start fresh,
|
|
+ * non-ratelimited. */
|
|
+ r = event_source_leave_ratelimit(s);
|
|
+ if (r < 0)
|
|
+ return r;
|
|
+
|
|
+ s->rate_limit = (RateLimit) { interval, burst };
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+_public_ int sd_event_source_get_ratelimit(sd_event_source *s, uint64_t *ret_interval, unsigned *ret_burst) {
|
|
+ assert_return(s, -EINVAL);
|
|
+
|
|
+ /* Querying whether an event source has ratelimiting configured is not a loggable offsense, hence
|
|
+ * don't use assert_return(). Unlike turning on ratelimiting it's not really a programming error */
|
|
+ if (!EVENT_SOURCE_CAN_RATE_LIMIT(s->type))
|
|
+ return -EDOM;
|
|
+
|
|
+ if (!ratelimit_configured(&s->rate_limit))
|
|
+ return -ENOEXEC;
|
|
+
|
|
+ if (ret_interval)
|
|
+ *ret_interval = s->rate_limit.interval;
|
|
+ if (ret_burst)
|
|
+ *ret_burst = s->rate_limit.burst;
|
|
+
|
|
+ return 0;
|
|
+}
|
|
+
|
|
+_public_ int sd_event_source_is_ratelimited(sd_event_source *s) {
|
|
+ assert_return(s, -EINVAL);
|
|
+
|
|
+ if (!EVENT_SOURCE_CAN_RATE_LIMIT(s->type))
|
|
+ return false;
|
|
+
|
|
+ if (!ratelimit_configured(&s->rate_limit))
|
|
+ return false;
|
|
+
|
|
+ return s->ratelimited;
|
|
+}
|
|
diff --git a/src/systemd/sd-event.h b/src/systemd/sd-event.h
|
|
index 937c9bd460..2ae2a0da48 100644
|
|
--- a/src/systemd/sd-event.h
|
|
+++ b/src/systemd/sd-event.h
|
|
@@ -162,6 +162,9 @@ int sd_event_source_get_floating(sd_event_source *s);
|
|
int sd_event_source_set_floating(sd_event_source *s, int b);
|
|
int sd_event_source_get_exit_on_failure(sd_event_source *s);
|
|
int sd_event_source_set_exit_on_failure(sd_event_source *s, int b);
|
|
+int sd_event_source_set_ratelimit(sd_event_source *s, uint64_t interval_usec, unsigned burst);
|
|
+int sd_event_source_get_ratelimit(sd_event_source *s, uint64_t *ret_interval_usec, unsigned *ret_burst);
|
|
+int sd_event_source_is_ratelimited(sd_event_source *s);
|
|
|
|
/* Define helpers so that __attribute__((cleanup(sd_event_unrefp))) and similar may be used. */
|
|
_SD_DEFINE_POINTER_CLEANUP_FUNC(sd_event, sd_event_unref);
|
|
--
|
|
2.17.1
|
|
|