apr_atomic_read32(&lingering_count),
apr_atomic_read32(&clogged_count),
apr_atomic_read32(&suspended_count),
- ap_queue_info_get_idlers(worker_queue_info));
+ ap_queue_info_num_idlers(worker_queue_info));
for (i = 0; i < num_listensocks; i++)
apr_pollset_add(event_pollset, &listener_pollfd[i]);
/*
static APR_INLINE int connections_above_limit(void)
{
- apr_uint32_t i_count = ap_queue_info_get_idlers(worker_queue_info);
+ apr_uint32_t i_count = ap_queue_info_num_idlers(worker_queue_info);
if (i_count > 0) {
apr_uint32_t c_count = apr_atomic_read32(&connection_count);
apr_uint32_t l_count = apr_atomic_read32(&lingering_count);
apr_atomic_read32(&connection_count));
ap_log_error(APLOG_MARK, APLOG_TRACE1, 0, ap_server_conf,
"Idle workers: %u",
- ap_queue_info_get_idlers(worker_queue_info));
+ ap_queue_info_num_idlers(worker_queue_info));
workers_were_busy = 1;
}
else if (!listener_may_exit) {
}
}
-apr_uint32_t ap_queue_info_get_idlers(fd_queue_info_t *queue_info)
+apr_uint32_t ap_queue_info_num_idlers(fd_queue_info_t *queue_info)
{
apr_uint32_t val;
val = apr_atomic_read32(&queue_info->idlers);
apr_status_t ap_queue_info_try_get_idler(fd_queue_info_t *queue_info);
apr_status_t ap_queue_info_wait_for_idler(fd_queue_info_t *queue_info,
int *had_to_block);
+apr_uint32_t ap_queue_info_num_idlers(fd_queue_info_t *queue_info);
apr_status_t ap_queue_info_term(fd_queue_info_t *queue_info);
-apr_uint32_t ap_queue_info_get_idlers(fd_queue_info_t *queue_info);
-void ap_free_idle_pools(fd_queue_info_t *queue_info);
typedef struct timer_event_t timer_event_t;
void ap_pop_pool(apr_pool_t **recycled_pool, fd_queue_info_t *queue_info);
void ap_push_pool(fd_queue_info_t *queue_info, apr_pool_t *pool_to_recycle);
+void ap_free_idle_pools(fd_queue_info_t *queue_info);
apr_status_t ap_queue_init(fd_queue_t *queue, int queue_capacity,
apr_pool_t *a);