include/boost/corosio/native/detail/select/select_scheduler.hpp

87.9% Lines (145/165) 100.0% List of functions (10/10)
select_scheduler.hpp
f(x) Functions (10)
Line TLA Hits Source Code
1 //
2 // Copyright (c) 2026 Steve Gerbino
3 //
4 // Distributed under the Boost Software License, Version 1.0. (See accompanying
5 // file LICENSE_1_0.txt or copy at http://www.boost.org/LICENSE_1_0.txt)
6 //
7 // Official repository: https://github.com/cppalliance/corosio
8 //
9
10 #ifndef BOOST_COROSIO_NATIVE_DETAIL_SELECT_SELECT_SCHEDULER_HPP
11 #define BOOST_COROSIO_NATIVE_DETAIL_SELECT_SELECT_SCHEDULER_HPP
12
13 #include <boost/corosio/detail/platform.hpp>
14
15 #if BOOST_COROSIO_HAS_SELECT
16
17 #include <boost/corosio/detail/config.hpp>
18 #include <boost/capy/ex/execution_context.hpp>
19
20 #include <boost/corosio/native/detail/reactor/reactor_scheduler.hpp>
21
22 #include <boost/corosio/native/detail/select/select_traits.hpp>
23 #include <boost/corosio/detail/timer_service.hpp>
24 #include <boost/corosio/native/detail/make_err.hpp>
25 #include <boost/corosio/native/detail/posix/posix_resolver_service.hpp>
26 #include <boost/corosio/native/detail/posix/posix_signal_service.hpp>
27 #include <boost/corosio/native/detail/posix/posix_stream_file_service.hpp>
28 #include <boost/corosio/native/detail/posix/posix_random_access_file_service.hpp>
29
30 #include <boost/corosio/detail/except.hpp>
31
32 #include <sys/select.h>
33 #include <unistd.h>
34 #include <errno.h>
35 #include <fcntl.h>
36
37 #include <atomic>
38 #include <chrono>
39 #include <cstdint>
40 #include <limits>
41 #include <mutex>
42 #include <unordered_map>
43
44 namespace boost::corosio::detail {
45
46 struct select_op;
47
48 /** POSIX scheduler using select() for I/O multiplexing.
49
50 This scheduler implements the scheduler interface using the POSIX select()
51 call for I/O event notification. It inherits the shared reactor threading
52 model from reactor_scheduler: signal state machine, inline completion
53 budget, work counting, and the do_one event loop.
54
55 The design mirrors epoll_scheduler for behavioral consistency:
56 - Same single-reactor thread coordination model
57 - Same deferred I/O pattern (reactor marks ready; workers do I/O)
58 - Same timer integration pattern
59
60 Known Limitations:
61 - FD_SETSIZE (~1024) limits maximum concurrent connections
62 - O(n) scanning: rebuilds fd_sets each iteration
63 - Level-triggered only (no edge-triggered mode)
64
65 @par Thread Safety
66 All public member functions are thread-safe.
67 */
68 class BOOST_COROSIO_DECL select_scheduler final : public reactor_scheduler
69 {
70 public:
71 /** Construct the scheduler.
72
73 Creates a self-pipe for reactor interruption.
74
75 @param ctx Reference to the owning execution_context.
76 @param concurrency_hint Hint for expected thread count (unused).
77 */
78 select_scheduler(capy::execution_context& ctx, int concurrency_hint = -1);
79
80 /// Destroy the scheduler.
81 ~select_scheduler() override;
82
83 select_scheduler(select_scheduler const&) = delete;
84 select_scheduler& operator=(select_scheduler const&) = delete;
85
86 /// Shut down the scheduler, draining pending operations.
87 void shutdown() override;
88
89 /** Return the maximum file descriptor value supported.
90
91 Returns FD_SETSIZE - 1, the maximum fd value that can be
92 monitored by select(). Operations with fd >= FD_SETSIZE
93 will fail with EINVAL.
94
95 @return The maximum supported file descriptor value.
96 */
97 static constexpr int max_fd() noexcept
98 {
99 return FD_SETSIZE - 1;
100 }
101
102 /** Register a descriptor for persistent monitoring.
103
104 The fd is added to the registered_descs_ map and will be
105 included in subsequent select() calls. The reactor is
106 interrupted so a blocked select() rebuilds its fd_sets.
107
108 @param fd The file descriptor to register.
109 @param desc Pointer to descriptor state for this fd.
110 */
111 void register_descriptor(int fd, reactor_descriptor_state* desc) const;
112
113 /** Deregister a persistently registered descriptor.
114
115 @param fd The file descriptor to deregister.
116 */
117 void deregister_descriptor(int fd) const;
118
119 /** Interrupt the reactor so it rebuilds its fd_sets.
120
121 Called when a write or connect op is registered after
122 the reactor's snapshot was taken. Without this, select()
123 may block not watching for writability on the fd.
124 */
125 void notify_reactor() const;
126
127 private:
128 void
129 run_task(lock_type& lock, context_type* ctx,
130 long timeout_us) override;
131 void interrupt_reactor() const override;
132 long calculate_timeout(long requested_timeout_us) const;
133
134 // Self-pipe for interrupting select()
135 int pipe_fds_[2]; // [0]=read, [1]=write
136
137 // Per-fd tracking for fd_set building
138 mutable std::unordered_map<int, reactor_descriptor_state*> registered_descs_;
139 mutable int max_fd_ = -1;
140 };
141
142 239x inline select_scheduler::select_scheduler(capy::execution_context& ctx, int)
143 239x : pipe_fds_{-1, -1}
144 239x , max_fd_(-1)
145 {
146 239x if (::pipe(pipe_fds_) < 0)
147 detail::throw_system_error(make_err(errno), "pipe");
148
149 717x for (int i = 0; i < 2; ++i)
150 {
151 478x int flags = ::fcntl(pipe_fds_[i], F_GETFL, 0);
152 478x if (flags == -1)
153 {
154 int errn = errno;
155 ::close(pipe_fds_[0]);
156 ::close(pipe_fds_[1]);
157 detail::throw_system_error(make_err(errn), "fcntl F_GETFL");
158 }
159 478x if (::fcntl(pipe_fds_[i], F_SETFL, flags | O_NONBLOCK) == -1)
160 {
161 int errn = errno;
162 ::close(pipe_fds_[0]);
163 ::close(pipe_fds_[1]);
164 detail::throw_system_error(make_err(errn), "fcntl F_SETFL");
165 }
166 478x if (::fcntl(pipe_fds_[i], F_SETFD, FD_CLOEXEC) == -1)
167 {
168 int errn = errno;
169 ::close(pipe_fds_[0]);
170 ::close(pipe_fds_[1]);
171 detail::throw_system_error(make_err(errn), "fcntl F_SETFD");
172 }
173 }
174
175 239x timer_svc_ = &get_timer_service(ctx, *this);
176 239x timer_svc_->set_on_earliest_changed(
177 2952x timer_service::callback(this, [](void* p) {
178 2713x static_cast<select_scheduler*>(p)->interrupt_reactor();
179 2713x }));
180
181 239x get_resolver_service(ctx, *this);
182 239x get_signal_service(ctx, *this);
183 239x get_stream_file_service(ctx, *this);
184 239x get_random_access_file_service(ctx, *this);
185
186 239x completed_ops_.push(&task_op_);
187 239x }
188
189 478x inline select_scheduler::~select_scheduler()
190 {
191 239x if (pipe_fds_[0] >= 0)
192 239x ::close(pipe_fds_[0]);
193 239x if (pipe_fds_[1] >= 0)
194 239x ::close(pipe_fds_[1]);
195 478x }
196
197 inline void
198 239x select_scheduler::shutdown()
199 {
200 239x shutdown_drain();
201
202 239x if (pipe_fds_[1] >= 0)
203 239x interrupt_reactor();
204 239x }
205
206 inline void
207 5147x select_scheduler::register_descriptor(
208 int fd, reactor_descriptor_state* desc) const
209 {
210 5147x if (fd < 0 || fd >= FD_SETSIZE)
211 detail::throw_system_error(make_err(EINVAL), "select: fd out of range");
212
213 5147x desc->registered_events = reactor_event_read | reactor_event_write;
214 5147x desc->fd = fd;
215 5147x desc->scheduler_ = this;
216 5147x desc->mutex.set_enabled(!single_threaded_);
217 5147x desc->ready_events_.store(0, std::memory_order_relaxed);
218
219 {
220 5147x conditionally_enabled_mutex::scoped_lock lock(desc->mutex);
221 5147x desc->impl_ref_.reset();
222 5147x desc->read_ready = false;
223 5147x desc->write_ready = false;
224 5147x }
225
226 {
227 5147x mutex_type::scoped_lock lock(mutex_);
228 5147x registered_descs_[fd] = desc;
229 5147x if (fd > max_fd_)
230 5143x max_fd_ = fd;
231 5147x }
232
233 5147x interrupt_reactor();
234 5147x }
235
236 inline void
237 5147x select_scheduler::deregister_descriptor(int fd) const
238 {
239 5147x mutex_type::scoped_lock lock(mutex_);
240
241 5147x auto it = registered_descs_.find(fd);
242 5147x if (it == registered_descs_.end())
243 return;
244
245 5147x registered_descs_.erase(it);
246
247 5147x if (fd == max_fd_)
248 {
249 5084x max_fd_ = pipe_fds_[0];
250 10054x for (auto& [registered_fd, state] : registered_descs_)
251 {
252 4970x if (registered_fd > max_fd_)
253 4955x max_fd_ = registered_fd;
254 }
255 }
256 5147x }
257
258 inline void
259 2500x select_scheduler::notify_reactor() const
260 {
261 2500x interrupt_reactor();
262 2500x }
263
264 inline void
265 10764x select_scheduler::interrupt_reactor() const
266 {
267 10764x char byte = 1;
268 10764x [[maybe_unused]] auto r = ::write(pipe_fds_[1], &byte, 1);
269 10764x }
270
271 inline long
272 110691x select_scheduler::calculate_timeout(long requested_timeout_us) const
273 {
274 110691x if (requested_timeout_us == 0)
275 return 0;
276
277 110691x auto nearest = timer_svc_->nearest_expiry();
278 110691x if (nearest == timer_service::time_point::max())
279 56x return requested_timeout_us;
280
281 110635x auto now = std::chrono::steady_clock::now();
282 110635x if (nearest <= now)
283 376x return 0;
284
285 auto timer_timeout_us =
286 110259x std::chrono::duration_cast<std::chrono::microseconds>(nearest - now)
287 110259x .count();
288
289 110259x constexpr auto long_max =
290 static_cast<long long>((std::numeric_limits<long>::max)());
291 auto capped_timer_us =
292 110259x (std::min)((std::max)(static_cast<long long>(timer_timeout_us),
293 110259x static_cast<long long>(0)),
294 110259x long_max);
295
296 110259x if (requested_timeout_us < 0)
297 110253x return static_cast<long>(capped_timer_us);
298
299 return static_cast<long>(
300 6x (std::min)(static_cast<long long>(requested_timeout_us),
301 6x capped_timer_us));
302 }
303
304 inline void
305 131032x select_scheduler::run_task(
306 lock_type& lock, context_type* ctx, long timeout_us)
307 {
308 long effective_timeout_us =
309 131032x task_interrupted_ ? 0 : calculate_timeout(timeout_us);
310
311 // Snapshot registered descriptors while holding lock.
312 // Record which fds need write monitoring to avoid a hot loop:
313 // select is level-triggered so writable sockets (nearly always
314 // writable) would cause select() to return immediately every
315 // iteration if unconditionally added to write_fds.
316 struct fd_entry
317 {
318 int fd;
319 reactor_descriptor_state* desc;
320 bool needs_write;
321 };
322 fd_entry snapshot[FD_SETSIZE];
323 131032x int snapshot_count = 0;
324
325 393696x for (auto& [fd, desc] : registered_descs_)
326 {
327 262664x if (snapshot_count < FD_SETSIZE)
328 {
329 262664x conditionally_enabled_mutex::scoped_lock desc_lock(desc->mutex);
330 262664x snapshot[snapshot_count].fd = fd;
331 262664x snapshot[snapshot_count].desc = desc;
332 262664x snapshot[snapshot_count].needs_write =
333 262664x (desc->write_op || desc->connect_op);
334 262664x ++snapshot_count;
335 262664x }
336 }
337
338 131032x if (lock.owns_lock())
339 110691x lock.unlock();
340
341 131032x task_cleanup on_exit{this, &lock, ctx};
342
343 fd_set read_fds, write_fds, except_fds;
344 2227544x FD_ZERO(&read_fds);
345 2227544x FD_ZERO(&write_fds);
346 2227544x FD_ZERO(&except_fds);
347
348 131032x FD_SET(pipe_fds_[0], &read_fds);
349 131032x int nfds = pipe_fds_[0];
350
351 393696x for (int i = 0; i < snapshot_count; ++i)
352 {
353 262664x int fd = snapshot[i].fd;
354 262664x FD_SET(fd, &read_fds);
355 262664x if (snapshot[i].needs_write)
356 2502x FD_SET(fd, &write_fds);
357 262664x FD_SET(fd, &except_fds);
358 262664x if (fd > nfds)
359 130761x nfds = fd;
360 }
361
362 struct timeval tv;
363 131032x struct timeval* tv_ptr = nullptr;
364 131032x if (effective_timeout_us >= 0)
365 {
366 130976x tv.tv_sec = effective_timeout_us / 1000000;
367 130976x tv.tv_usec = effective_timeout_us % 1000000;
368 130976x tv_ptr = &tv;
369 }
370
371 131032x int ready = ::select(nfds + 1, &read_fds, &write_fds, &except_fds, tv_ptr);
372
373 // EINTR: signal interrupted select(), just retry.
374 // EBADF: an fd was closed between snapshot and select(); retry
375 // with a fresh snapshot from registered_descs_.
376 131032x if (ready < 0)
377 {
378 if (errno == EINTR || errno == EBADF)
379 return;
380 detail::throw_system_error(make_err(errno), "select");
381 }
382
383 // Process timers outside the lock
384 131032x timer_svc_->process_expired();
385
386 131032x op_queue local_ops;
387
388 131032x if (ready > 0)
389 {
390 115890x if (FD_ISSET(pipe_fds_[0], &read_fds))
391 {
392 char buf[256];
393 10616x while (::read(pipe_fds_[0], buf, sizeof(buf)) > 0)
394 {
395 }
396 }
397
398 335139x for (int i = 0; i < snapshot_count; ++i)
399 {
400 219249x int fd = snapshot[i].fd;
401 219249x reactor_descriptor_state* desc = snapshot[i].desc;
402
403 219249x std::uint32_t flags = 0;
404 219249x if (FD_ISSET(fd, &read_fds))
405 113260x flags |= reactor_event_read;
406 219249x if (FD_ISSET(fd, &write_fds))
407 2499x flags |= reactor_event_write;
408 219249x if (FD_ISSET(fd, &except_fds))
409 flags |= reactor_event_error;
410
411 219249x if (flags == 0)
412 103498x continue;
413
414 115751x desc->add_ready_events(flags);
415
416 115751x bool expected = false;
417 115751x if (desc->is_enqueued_.compare_exchange_strong(
418 expected, true, std::memory_order_release,
419 std::memory_order_relaxed))
420 {
421 115751x local_ops.push(desc);
422 }
423 }
424 }
425
426 131032x lock.lock();
427
428 131032x if (!local_ops.empty())
429 113261x completed_ops_.splice(local_ops);
430 131032x }
431
432 } // namespace boost::corosio::detail
433
434 #endif // BOOST_COROSIO_HAS_SELECT
435
436 #endif // BOOST_COROSIO_NATIVE_DETAIL_SELECT_SELECT_SCHEDULER_HPP
437