Skip to content

Latest commit

 

History

History
308 lines (228 loc) · 15.9 KB

0007-event_loop-refactor.md

File metadata and controls

308 lines (228 loc) · 15.9 KB

Summary

Refactor the event loop interface to be decoupled from libevent and adapts to the needs of multithreaded execution contexts (RFC #0002).

Motivation

The event loop is the central piece for enabling non-blocking operations and other event-based activities.

The original implementation of Crystal::EventLoop etc. is heavily based on libevent which was the first backing driver. Now we have an event loop based on IOCP for Windows and a version for the stripped-down abilities on WASI. Another implementation based on io_uring has already been proposed (#10768).

All these implementations have quite different designs and requirements. Shoehorning them into an API intended for libevent isn’t ideal. Hence the interface should be refactored to be more generic.

Another concern is that io_uring is generally available on Linux, but requires a modern kernel. A Crystal program should be able to fall back to an epoll based selector if io_uring is unavailable.

There’s an ongoing refactoring effort for enhanced multi-threading features, which includes a refactoring of the scheduler (RFC #0002). This context also asks for some changes to the event loop interface to allow more capable and efficient scheduling implementations. This has also brought up the idea to use system event loops directly instead of libevent (which is essentially a cross-platform wrapper around them). This would also allow to take control of the abstraction and open possibilities for further optimization in Crystal resulting in more performance.

Guide-level explanation

Crystal::EventLoop exposes a generic and compact interface for actions that are to be performed on Crystals event loop. This interface is used in the implementation of asynchronous IO and other evented operations.

Compared to the previous event loop implementation, actions are entirely self-contained instead of scattered between the event loop and the system implementations of Socket and Crystal::System::FileDescriptor.

For example, the previous, platform-specific implementation of Socket#unbuffered_read for Unix systems:

private def unbuffered_read(slice : Bytes) : Int32
  evented_read(slice, "Error reading socket") do
    LibC.recv(fd, slice, slice.size, 0).to_i32
  end
end

The new, portable implementation of Socket#unbuffered_read:

private def unbuffered_read(slice : Bytes) : Int32
  event_loop.read(self, slice)
end

The new API initially supports adaptations of the existing backends libevent and IOCP. The next steps are additional implementations based on io_uring, as well system selectors (kqueue and epoll), skipping libevent to improve performance.

Further extensions can widen the scope of the event loop.

Crystal::Eventloop is an internal API, thus changing it does not break backwards compatibility.

Terminology

  • Event loop: an abstraction to wait on specific events, such as timers (e.g. wait until a certain amount of time has passed) or IO (e.g. wait for an socket to be readable or writable).

  • Crystal event loop: Component of the Crystal runtime that subscribes to and dispatches events from the OS and other sources to enqueue a waiting fiber when a desired event (such as IO availability) has occured. It’s an interface between Crystal’s scheduler and the system’s event loop backend.

  • System selector: The system implementation of an event loop which the runtime component depends on to receive events from the operating system. Example implementations: epoll, kqueue, IOCP, libevent (a cross-platform wrapper of the former), io_uring

  • Scheduler: manages fibers’ executions inside the program (controlled by Crystal), unlike threads that are scheduled by the operating system (outside of the accessibility of the program).

General Design Principles

  • Generic API: Independent of a specific event loop design
  • Black Box: No interference with internals of the event loop
  • Pluggable: It must be possible to compile multiple event loop implementations and choose at runtime. Only one type of implementation will typically be active (it would be feasible to have different event loop implementations in different execution contexts, if there's a use case for that).

Reference-level explanation

The new EventLoop interface consits of individual module interfaces which defines operations on the event loop related to a specific concept. The run and interrupt methods are shown for completenes, but they are part of RFC 0002 (#14568).

abstract class Crystal::System::EventLoop
  # Runs the loop.
  #
  # Returns immediately if events are activable. Set `blocking` to false to
  # return immediately if there are no activable events; set it to true to wait
  # for activable events, which will block the current thread until then.
  #
  # Returns `true` on normal returns (e.g. has activated events, has pending
  # events but blocking was false) and `false` when there are no registered
  # events.
  abstract def run(blocking : Bool) : Bool

  # Tells a blocking run loop to no longer wait for events to activate. It may
  # for example enqueue a NOOP event with an immediate (or past) timeout. Having
  # activated an event, the loop shall return, allowing the blocked thread to
  # continue.
  #
  # Should be a NOOP when the loop isn't running or is running in a nonblocking
  # mode.
  #
  # NOTE: we assume that multiple threads won't run the event loop at the same
  #       time in parallel, but this assumption may change in the future!
  abstract def interrupt : Nil

  include FileDescriptor
  include Socket

  module FileDescriptor
    # Reads at least one byte from the file descriptor into *slice*.
    #
    # Blocks the current fiber if no data is available for reading, continuing
    # when available. Otherwise returns immediately.
    #
    # Returns the number of bytes read (up to `slice.size`).
    # Returns 0 when EOF is reached.
    abstract def read(file_descriptor : Crystal::System::FileDescriptor, slice : Bytes) : Int32

    # Writes at least one byte from *slice* to the file descriptor.
    #
    # Blocks the current fiber if the file descriptor isn't ready for writing,
    # continuing when ready. Otherwise returns immediately.
    #
    # Returns the number of bytes written (up to `slice.size`).
    abstract def write(file_descriptor : Crystal::System::FileDescriptor, slice : Bytes) : Int32

    # Closes the file descriptor resource.
    abstract def close(file_descriptor : Crystal::System::FileDescriptor) : Nil

    # Removes the file descriptor from the event loop. Can be used to free up
    # memory resources associated with the file descriptor, as well as removing
    # the file descriptor from kernel data structures.
    #
    # Called by `::IO::FileDescriptor#finalize` before closing the file
    # descriptor. Errors shall be silently ignored.
    abstract def remove(file_descriptor : Crystal::System::FileDescriptor) : Nil
  end

  module Socket
    # Reads at least one byte from the socket into *slice*.
    #
    # Blocks the current fiber if no data is available for reading, continuing
    # when available. Otherwise returns immediately.
    #
    # Returns the number of bytes read (up to `slice.size`).
    # Returns 0 when the socket is closed and no data available.
    #
    # Use `#receive_from` for capturing the source address of a message.
    abstract def read(socket : ::Socket, slice : Bytes) : Int32

    # Writes at least one byte from *slice* to the socket.
    #
    # Blocks the current fiber if the socket is not ready for writing,
    # continuing when ready. Otherwise returns immediately.
    #
    # Returns the number of bytes written (up to `slice.size`).
    #
    # Use `#send_to` for sending a message to a specific target address.
    abstract def write(socket : ::Socket, slice : Bytes) : Int32

    # Accepts an incoming TCP connection on the socket.
    #
    # Blocks the current fiber if no connection is waiting, continuing when one
    # becomes available. Otherwise returns immediately.
    #
    # Returns a handle to the socket for the new connection.
    abstract def accept(socket : ::Socket) : ::Socket::Handle?

    # Opens a connection on *socket* to the target *address*.
    #
    # Blocks the current fiber and continues when the connection is established.
    #
    # Returns `IO::Error` in case of an error. The caller is responsible for
    # raising it as an exception if necessary.
    abstract def connect(socket : ::Socket, address : ::Socket::Addrinfo | ::Socket::Address, timeout : ::Time::Span?) : IO::Error?

    # Sends at least one byte from *slice* to the socket with a target address
    # *address*.
    #
    # Blocks the current fiber if the socket is not ready for writing,
    # continuing when ready. Otherwise returns immediately.
    #
    # Returns the number of bytes sent (up to `slice.size`).
    abstract def send_to(socket : ::Socket, slice : Bytes, address : ::Socket::Address) : Int32

    # Receives at least one byte from the socket into *slice*, capturing the
    # source address.
    #
    # Blocks the current fiber if no data is available for reading, continuing
    # when available. Otherwise returns immediately.
    #
    # Returns a tuple containing the number of bytes received (up to `slice.size`)
    # and the source address.
    abstract def receive_from(socket : ::Socket, slice : Bytes) : Tuple(Int32, ::Socket::Address)

    # Closes the socket.
    abstract def close(socket : ::Socket) : Nil

    # Removes the socket from the event loop. Can be used to free up memory
    # resources associated with the socket, as well as removing the socket from
    # kernel data structures.
    #
    # Called by `::Socket#finalize` before closing the socket. Errors shall be
    # silently ignored.
    abstract def remove(socket : ::Socket) : Nil
  end
end

Notable differences from the previous API:

  • Timeout and resume events are not on the event loop. The scheduler is supposed to handle them directly. Crystal::System::Event gets removed as this was its only use case.
  • The behaviour of event loop actions read and write has been unified: they both read/write at least one byte and return the number of bytes read/written. This keeps the event loop implementation minimal and more versatile. Previously, write was expected to write all bytes of the given slice.
  • Socket is not part of the core library, so references to its types (e.g. in type restrictions) would not resolve. The EventLoop interface is split into individual modules and the Socket module is always defined, but only filled with abstract defs when socket.cr is explicitly required.

Drawbacks

None.

Rationale and alternatives

The rationale is to create our own system independent event loop, and to keep it as abstract as possible to support any kinds of models (asynchronous, polling) over the actual operations.

Alternatives, as they exist today, leak system specifics into the stdlib types (see IO::Evented).

The proposed model completely eliminates that by creating a standardized interface.

Prior art

References

Unresolved questions

What’s the scope of the Crystal event loop?

This is a list of operations which we expect to go through the event loop:

  • File descriptor: read, write
  • Socket: read, write, accept, connect, send, receive
  • Process: wait (Windows)

Potential extensions:

Should events from the Crystal runtime be part of the event loop as well?

  • Fiber: sleep
  • Select actions: send, receive, timeout

Optional event loop features

Some activities are managed on the event loop on one platform but not on others. Example would be Process#wait which goes through IOCP on Windows but on Unix it’s part of signal handling. (Note: Perhaps we could try to get that on the event loop on Unix as well? 🤔 But there are other examples of system differences)

Do we require these optional methods to be present in all event loop implementations, i.e. they’re part of the global interface? Some impls would then just raise “Not implemented”. Alternatively, we could keep them out of the main interface and check for availability via event_loop.responds_to?. or a sub interface (is_a?(EventLoop::Process)). Or…?

Future possibilities

  • Alternative event loop implementations based directly on the system selectors instead of libevent (RFC 0009)

Type for sizes

Currently, the return type of unbuffered_read is unspecified and there’s a bit of a mess. Technically, it can only be Int32 because that’s the size of Slice. We could use the same in the event loop API. However, in order to be future proof for bigger sizes (crystal-lang/crystal#4011), we could design the API with SizeT instead. Considering it’s a low-level system API, this should be fine and makes a lot of sense.
Currently, the only possible values would still be the positive range of Int32, so there would be no conversion risk.

Blocking event loop

There should be a basic implementation with blocking syscalls when non-blocking operation is unavailable. This would currently be used for WASI, for example, and allows IO to work, although not as efficiently.

When evented operations are not available (for example file system operations or DNS resolution), the scheduler could automatically fall back to execute the operation in a separate thread.

This works well with the libevent implementation because if the lib calls never return EAGAIN, the code works entirely without libevent. The WASI implementation raises when trying to create a resume event, but that never happens when IO operations are all blocking (~O_NONBLOCK). For that to happen, we’d need to explicitly set the IO to blocking.

Alternative idea: If polling is unavailable, we could consider sleeping the fiber for some time and let it retry again 🤷

#connect timeout

Socket#connect is the only method with a timeout parameter. This seems weird. All other timeouts are configured on the IO instance. It probably makes sense to standardize, introduce a new property Socket#connect_timeout, and deprecate the def parameter.

Bulk events without fibers

For some applications it might be useful to interact with the event loop directly, being able to push operations in bulk without having to spawn (and wait) a fiber for each one.

Integration with select

select actions are implemented completely independent of the event loop, yet they operate in a similar domain (wait for something to happen). These actions are special in that they ensure atomicity. When waiting on multiple actions simultaneously, it’s guaranteed that only one of them executes. This probably won’t be exactly possible with most event loop features.

It would certainly be nice if we could use more actions with select and maybe putting everything on the event loop is actually the best path forward with that?

select
when data = @socket.gets
end