#include #ifdef G_OS_UNIX #include #endif static GMainLoop *loop; static void stop_waiting (gpointer data) { g_main_loop_quit (loop); } static gboolean unreachable_callback (gpointer data) { g_assert_not_reached (); return G_SOURCE_REMOVE; } static void unreachable_void_callback (gpointer data) { g_assert_not_reached (); } static void test_seconds (void) { guint id; /* Bug 642052 mentions that g_timeout_add_seconds(21475) schedules a * job that runs once per second. * * Test that that isn't true anymore by scheduling two jobs: * - one, as above * - another that runs in 2100ms * * If everything is working properly, the 2100ms one should run first * (and exit the mainloop). If we ever see the 21475 second job run * then we have trouble (since it ran in less than 2 seconds). * * We need a timeout of at least 2 seconds because * g_timeout_add_seconds() can add as much as an additional second of * latency. */ g_test_bug ("https://bugzilla.gnome.org/show_bug.cgi?id=642052"); loop = g_main_loop_new (NULL, FALSE); g_timeout_add_once (2100, stop_waiting, NULL); id = g_timeout_add_seconds (21475, unreachable_callback, NULL); g_main_loop_run (loop); g_main_loop_unref (loop); g_source_remove (id); } static void test_seconds_once (void) { /* Use the same principle as in test_seconds() */ loop = g_main_loop_new (NULL, FALSE); g_timeout_add_once (2100, stop_waiting, NULL); g_timeout_add_seconds_once (21475, unreachable_void_callback, NULL); g_main_loop_run (loop); g_main_loop_unref (loop); } static void test_weeks_overflow (void) { guint id; guint interval_seconds; /* Internally, the guint interval (in seconds) was converted to milliseconds * then stored in a guint variable. This meant that any interval larger than * G_MAXUINT / 1000 would overflow. * * On a system with 32-bit guint, the interval (G_MAXUINT / 1000) + 1 seconds * (49.7 days) would end wrapping to 704 milliseconds. * * Test that that isn't true anymore by scheduling two jobs: * - one, as above * - another that runs in 2100ms * * If everything is working properly, the 2100ms one should run first * (and exit the mainloop). If we ever see the other job run * then we have trouble (since it ran in less than 2 seconds). * * We need a timeout of at least 2 seconds because * g_timeout_add_seconds() can add as much as an additional second of * latency. */ g_test_bug ("https://gitlab.gnome.org/GNOME/glib/issues/1600"); loop = g_main_loop_new (NULL, FALSE); g_timeout_add_once (2100, stop_waiting, NULL); interval_seconds = 1 + G_MAXUINT / 1000; id = g_timeout_add_seconds (interval_seconds, unreachable_callback, NULL); g_main_loop_run (loop); g_main_loop_unref (loop); g_source_remove (id); } /* The ready_time for a GSource is stored as a gint64, as an absolute monotonic * time in microseconds. To call poll(), this must be converted to a relative * timeout, in milliseconds, as a gint. If the ready_time is sufficiently far * in the future, the timeout will not fit. Previously, it would be narrowed in * an implementation-defined way; if this gave a negative result, poll() would * block forever. * * This test creates a GSource with the largest possible ready_time (a little * over 292 millennia, assuming g_get_monotonic_time() starts from near 0 when * the system boots), adds it to a GMainContext, queries it for the parameters * to pass to poll() -- essentially the first half of * g_main_context_iteration() -- and checks that the timeout is a large * positive number. */ static void test_far_future_ready_time (void) { GSourceFuncs source_funcs = { 0 }; GMainContext *context = g_main_context_new (); GSource *source = g_source_new (&source_funcs, sizeof (GSource)); gboolean acquired, ready; gint priority, timeout_, n_fds; g_source_set_ready_time (source, G_MAXINT64); g_source_attach (source, context); acquired = g_main_context_acquire (context); g_assert_true (acquired); ready = g_main_context_prepare (context, &priority); g_assert_false (ready); n_fds = 0; n_fds = g_main_context_query (context, priority, &timeout_, NULL, n_fds); g_assert_cmpint (n_fds, >=, 0); /* The true timeout in milliseconds doesn't fit into a gint. We definitely * don't want poll() to block forever: */ g_assert_cmpint (timeout_, >=, 0); /* Instead, we want it to block for as long as possible: */ g_assert_cmpint (timeout_, ==, G_MAXINT); g_main_context_release (context); g_main_context_unref (context); g_source_unref (source); } static gint64 last_time; static gint count; static gboolean test_func (gpointer data) { gint64 current_time; current_time = g_get_monotonic_time (); /* We accept 2 on the first iteration because _add_seconds() can * have an initial latency of 1 second, see its documentation. * * Allow up to 500ms leeway for rounding and scheduling. */ if (count == 0) g_assert_cmpint (current_time / 1000 - last_time / 1000, <=, 2500); else g_assert_cmpint (current_time / 1000 - last_time / 1000, <=, 1500); last_time = current_time; count++; /* Make the timeout take up to 0.1 seconds. * We should still get scheduled for the next second. */ g_usleep (count * 10000); if (count < 10) return TRUE; g_main_loop_quit (loop); return FALSE; } static void test_rounding (void) { loop = g_main_loop_new (NULL, FALSE); last_time = g_get_monotonic_time (); g_timeout_add_seconds (1, test_func, NULL); g_main_loop_run (loop); g_main_loop_unref (loop); } int main (int argc, char *argv[]) { g_test_init (&argc, &argv, NULL); g_test_add_func ("/timeout/seconds", test_seconds); g_test_add_func ("/timeout/seconds-once", test_seconds_once); g_test_add_func ("/timeout/weeks-overflow", test_weeks_overflow); g_test_add_func ("/timeout/far-future-ready-time", test_far_future_ready_time); g_test_add_func ("/timeout/rounding", test_rounding); return g_test_run (); }