/* * Copyright 2022-2025 The OpenSSL Project Authors. All Rights Reserved. * * Licensed under the Apache License 2.0 (the "License"). You may not use * this file except in compliance with the License. You can obtain a copy * in the file LICENSE in the source distribution or at * https://www.openssl.org/source/license.html */ #include #include #include #include #include "quic_local.h" #include "internal/hashfunc.h" #include "internal/ssl_unwrap.h" #include "internal/quic_tls.h" #include "internal/quic_rx_depack.h" #include "internal/quic_error.h" #include "internal/quic_engine.h" #include "internal/quic_port.h" #include "internal/quic_reactor_wait_ctx.h" #include "internal/time.h" typedef struct qctx_st QCTX; static void qc_cleanup(QUIC_CONNECTION *qc, int have_lock); static void aon_write_finish(QUIC_XSO *xso); static int create_channel(QUIC_CONNECTION *qc, SSL_CTX *ctx); static QUIC_XSO *create_xso_from_stream(QUIC_CONNECTION *qc, QUIC_STREAM *qs); static QUIC_CONNECTION *create_qc_from_incoming_conn(QUIC_LISTENER *ql, QUIC_CHANNEL *ch); static int qc_try_create_default_xso_for_write(QCTX *ctx); static int qc_wait_for_default_xso_for_read(QCTX *ctx, int peek); static void qctx_lock(QCTX *qctx); static void qctx_unlock(QCTX *qctx); static void qctx_lock_for_io(QCTX *ctx); static int quic_do_handshake(QCTX *ctx); static void qc_update_reject_policy(QUIC_CONNECTION *qc); static void qc_touch_default_xso(QUIC_CONNECTION *qc); static void qc_set_default_xso(QUIC_CONNECTION *qc, QUIC_XSO *xso, int touch); static void qc_set_default_xso_keep_ref(QUIC_CONNECTION *qc, QUIC_XSO *xso, int touch, QUIC_XSO **old_xso); static SSL *quic_conn_stream_new(QCTX *ctx, uint64_t flags, int need_lock); static int quic_validate_for_write(QUIC_XSO *xso, int *err); static int quic_mutation_allowed(QUIC_CONNECTION *qc, int req_active); static void qctx_maybe_autotick(QCTX *ctx); static int qctx_should_autotick(QCTX *ctx); /* * QCTX is a utility structure which provides information we commonly wish to * unwrap upon an API call being dispatched to us, namely: * * - a pointer to the QUIC_CONNECTION (regardless of whether a QCSO or QSSO * was passed); * - a pointer to any applicable QUIC_XSO (e.g. if a QSSO was passed, or if * a QCSO with a default stream was passed); * - whether a QSSO was passed (xso == NULL must not be used to determine this * because it may be non-NULL when a QCSO is passed if that QCSO has a * default stream); * - a pointer to a QUIC_LISTENER object, if one is relevant; * - whether we are in "I/O context", meaning that non-normal errors can * be reported via SSL_get_error() as well as via ERR. Functions such as * SSL_read(), SSL_write() and SSL_do_handshake() are "I/O context" * functions which are allowed to change the value returned by * SSL_get_error. However, other functions (including functions which call * SSL_do_handshake() implicitly) are not allowed to change the return value * of SSL_get_error. */ struct qctx_st { QUIC_OBJ *obj; QUIC_DOMAIN *qd; QUIC_LISTENER *ql; QUIC_CONNECTION *qc; QUIC_XSO *xso; int is_stream, is_listener, is_domain, in_io; }; QUIC_NEEDS_LOCK static void quic_set_last_error(QCTX *ctx, int last_error) { if (!ctx->in_io) return; if (ctx->is_stream && ctx->xso != NULL) ctx->xso->last_error = last_error; else if (!ctx->is_stream && ctx->qc != NULL) ctx->qc->last_error = last_error; } /* * Raise a 'normal' error, meaning one that can be reported via SSL_get_error() * rather than via ERR. Note that normal errors must always be raised while * holding a lock. */ QUIC_NEEDS_LOCK static int quic_raise_normal_error(QCTX *ctx, int err) { assert(ctx->in_io); quic_set_last_error(ctx, err); return 0; } /* * Raise a 'non-normal' error, meaning any error that is not reported via * SSL_get_error() and must be reported via ERR. * * qc should be provided if available. In exceptional circumstances when qc is * not known NULL may be passed. This should generally only happen when an * expect_...() function defined below fails, which generally indicates a * dispatch error or caller error. * * ctx should be NULL if the connection lock is not held. */ static int quic_raise_non_normal_error(QCTX *ctx, const char *file, int line, const char *func, int reason, const char *fmt, ...) { va_list args; if (ctx != NULL) { quic_set_last_error(ctx, SSL_ERROR_SSL); if (reason == SSL_R_PROTOCOL_IS_SHUTDOWN && ctx->qc != NULL) ossl_quic_channel_restore_err_state(ctx->qc->ch); } ERR_new(); ERR_set_debug(file, line, func); va_start(args, fmt); ERR_vset_error(ERR_LIB_SSL, reason, fmt, args); va_end(args); return 0; } #define QUIC_RAISE_NORMAL_ERROR(ctx, err) \ quic_raise_normal_error((ctx), (err)) #define QUIC_RAISE_NON_NORMAL_ERROR(ctx, reason, msg) \ quic_raise_non_normal_error((ctx), \ OPENSSL_FILE, OPENSSL_LINE, \ OPENSSL_FUNC, \ (reason), \ (msg)) /* * Flags for expect_quic_as: * * QCTX_C * The input SSL object may be a QCSO. * * QCTX_S * The input SSL object may be a QSSO or a QCSO with a default stream * attached. * * (Note this means there is no current way to require an SSL object with a * QUIC stream which is not a QCSO; a QCSO with a default stream attached * is always considered to satisfy QCTX_S.) * * QCTX_AUTO_S * The input SSL object may be a QSSO or a QCSO with a default stream * attached. If no default stream is currently attached to a QCSO, * one may be auto-created if possible. * * If QCTX_REMOTE_INIT is set, an auto-created default XSO is * initiated by the remote party (i.e., local party reads first). * * If it is not set, an auto-created default XSO is * initiated by the local party (i.e., local party writes first). * * QCTX_L * The input SSL object may be a QLSO. * * QCTX_LOCK * If and only if the function returns successfully, the ctx * is guaranteed to be locked. * * QCTX_IO * Begin an I/O context. If not set, begins a non-I/O context. * This determines whether SSL_get_error() is updated; the value it returns * is modified only by an I/O call. * * QCTX_NO_ERROR * Don't raise an error if the object type is wrong. Should not be used in * conjunction with any flags that may raise errors not related to a wrong * object type. */ #define QCTX_C (1U << 0) #define QCTX_S (1U << 1) #define QCTX_L (1U << 2) #define QCTX_AUTO_S (1U << 3) #define QCTX_REMOTE_INIT (1U << 4) #define QCTX_LOCK (1U << 5) #define QCTX_IO (1U << 6) #define QCTX_D (1U << 7) #define QCTX_NO_ERROR (1U << 8) /* * Called when expect_quic failed. Used to diagnose why such a call failed and * raise a reasonable error code based on the configured preconditions in flags. */ static int wrong_type(const SSL *s, uint32_t flags) { const uint32_t mask = QCTX_C | QCTX_S | QCTX_L | QCTX_D; int code = ERR_R_UNSUPPORTED; if ((flags & QCTX_NO_ERROR) != 0) return 1; else if ((flags & mask) == QCTX_D) code = SSL_R_DOMAIN_USE_ONLY; else if ((flags & mask) == QCTX_L) code = SSL_R_LISTENER_USE_ONLY; else if ((flags & mask) == QCTX_C) code = SSL_R_CONN_USE_ONLY; else if ((flags & mask) == QCTX_S || (flags & mask) == (QCTX_C | QCTX_S)) code = SSL_R_NO_STREAM; return QUIC_RAISE_NON_NORMAL_ERROR(NULL, code, NULL); } /* * Given a QDSO, QCSO, QSSO or QLSO, initialises a QCTX, determining the * contextually applicable QUIC_LISTENER, QUIC_CONNECTION and QUIC_XSO * pointers. * * After this returns 1, all fields of the passed QCTX are initialised. * Returns 0 on failure. This function is intended to be used to provide API * semantics and as such, it invokes QUIC_RAISE_NON_NORMAL_ERROR() on failure * unless the QCTX_NO_ERROR flag is set. * * The flags argument controls the preconditions and postconditions of this * function. See above for the different flags. * * The fields of a QCTX are initialised as follows depending on the identity of * the SSL object, and assuming the preconditions demanded by the flags field as * described above are met: * * QDSO QLSO QCSO QSSO * qd non-NULL maybe maybe maybe * ql NULL non-NULL maybe maybe * qc NULL NULL non-NULL non-NULL * xso NULL NULL maybe non-NULL * is_stream 0 0 0 1 * is_listener 0 1 0 0 * is_domain 1 0 0 0 * */ static int expect_quic_as(const SSL *s, QCTX *ctx, uint32_t flags) { int ok = 0, locked = 0, lock_requested = ((flags & QCTX_LOCK) != 0); QUIC_DOMAIN *qd; QUIC_LISTENER *ql; QUIC_CONNECTION *qc; QUIC_XSO *xso; if ((flags & QCTX_AUTO_S) != 0) flags |= QCTX_S; ctx->obj = NULL; ctx->qd = NULL; ctx->ql = NULL; ctx->qc = NULL; ctx->xso = NULL; ctx->is_stream = 0; ctx->is_listener = 0; ctx->is_domain = 0; ctx->in_io = ((flags & QCTX_IO) != 0); if (s == NULL) { QUIC_RAISE_NON_NORMAL_ERROR(NULL, ERR_R_PASSED_NULL_PARAMETER, NULL); goto err; } switch (s->type) { case SSL_TYPE_QUIC_DOMAIN: if ((flags & QCTX_D) == 0) { wrong_type(s, flags); goto err; } qd = (QUIC_DOMAIN *)s; ctx->obj = &qd->obj; ctx->qd = qd; ctx->is_domain = 1; break; case SSL_TYPE_QUIC_LISTENER: if ((flags & QCTX_L) == 0) { wrong_type(s, flags); goto err; } ql = (QUIC_LISTENER *)s; ctx->obj = &ql->obj; ctx->qd = ql->domain; ctx->ql = ql; ctx->is_listener = 1; break; case SSL_TYPE_QUIC_CONNECTION: qc = (QUIC_CONNECTION *)s; ctx->obj = &qc->obj; ctx->qd = qc->domain; ctx->ql = qc->listener; /* never changes, so can be read without lock */ ctx->qc = qc; if ((flags & QCTX_AUTO_S) != 0) { if ((flags & QCTX_IO) != 0) qctx_lock_for_io(ctx); else qctx_lock(ctx); locked = 1; } if ((flags & QCTX_AUTO_S) != 0 && qc->default_xso == NULL) { if (!quic_mutation_allowed(qc, /*req_active=*/0)) { QUIC_RAISE_NON_NORMAL_ERROR(ctx, SSL_R_PROTOCOL_IS_SHUTDOWN, NULL); goto err; } /* If we haven't finished the handshake, try to advance it. */ if (quic_do_handshake(ctx) < 1) /* ossl_quic_do_handshake raised error here */ goto err; if ((flags & QCTX_REMOTE_INIT) != 0) { if (!qc_wait_for_default_xso_for_read(ctx, /*peek=*/0)) goto err; } else { if (!qc_try_create_default_xso_for_write(ctx)) goto err; } } if ((flags & QCTX_C) == 0 && (qc->default_xso == NULL || (flags & QCTX_S) == 0)) { wrong_type(s, flags); goto err; } ctx->xso = qc->default_xso; break; case SSL_TYPE_QUIC_XSO: if ((flags & QCTX_S) == 0) { wrong_type(s, flags); goto err; } xso = (QUIC_XSO *)s; ctx->obj = &xso->obj; ctx->qd = xso->conn->domain; ctx->ql = xso->conn->listener; ctx->qc = xso->conn; ctx->xso = xso; ctx->is_stream = 1; break; default: QUIC_RAISE_NON_NORMAL_ERROR(NULL, ERR_R_INTERNAL_ERROR, NULL); goto err; } if (lock_requested && !locked) { if ((flags & QCTX_IO) != 0) qctx_lock_for_io(ctx); else qctx_lock(ctx); locked = 1; } ok = 1; err: if (locked && (!ok || !lock_requested)) qctx_unlock(ctx); return ok; } static int is_quic_c(const SSL *s, QCTX *ctx, int raiseerrs) { uint32_t flags = QCTX_C; if (!raiseerrs) flags |= QCTX_NO_ERROR; return expect_quic_as(s, ctx, flags); } /* Same as expect_quic_cs except that errors are not raised if raiseerrs == 0 */ static int is_quic_cs(const SSL *s, QCTX *ctx, int raiseerrs) { uint32_t flags = QCTX_C | QCTX_S; if (!raiseerrs) flags |= QCTX_NO_ERROR; return expect_quic_as(s, ctx, flags); } static int expect_quic_cs(const SSL *s, QCTX *ctx) { return expect_quic_as(s, ctx, QCTX_C | QCTX_S); } static int expect_quic_csl(const SSL *s, QCTX *ctx) { return expect_quic_as(s, ctx, QCTX_C | QCTX_S | QCTX_L); } static int expect_quic_csld(const SSL *s, QCTX *ctx) { return expect_quic_as(s, ctx, QCTX_C | QCTX_S | QCTX_L | QCTX_D); } #define expect_quic_any expect_quic_csld static int expect_quic_listener(const SSL *s, QCTX *ctx) { return expect_quic_as(s, ctx, QCTX_L); } static int expect_quic_domain(const SSL *s, QCTX *ctx) { return expect_quic_as(s, ctx, QCTX_D); } /* * Like expect_quic_cs(), but requires a QUIC_XSO be contextually available. In * other words, requires that the passed QSO be a QSSO or a QCSO with a default * stream. * * remote_init determines if we expect the default XSO to be remotely created or * not. If it is -1, do not instantiate a default XSO if one does not yet exist. * * Channel mutex is acquired and retained on success. */ QUIC_ACQUIRES_LOCK static int ossl_unused expect_quic_with_stream_lock(const SSL *s, int remote_init, int in_io, QCTX *ctx) { uint32_t flags = QCTX_S | QCTX_LOCK; if (remote_init >= 0) flags |= QCTX_AUTO_S; if (remote_init > 0) flags |= QCTX_REMOTE_INIT; if (in_io) flags |= QCTX_IO; return expect_quic_as(s, ctx, flags); } /* * Like expect_quic_cs(), but fails if called on a QUIC_XSO. ctx->xso may still * be non-NULL if the QCSO has a default stream. */ static int ossl_unused expect_quic_conn_only(const SSL *s, QCTX *ctx) { return expect_quic_as(s, ctx, QCTX_C); } /* * Ensures that the domain mutex is held for a method which touches channel * state. * * Precondition: Domain mutex is not held (unchecked) */ static void qctx_lock(QCTX *ctx) { #if defined(OPENSSL_THREADS) assert(ctx->obj != NULL); ossl_crypto_mutex_lock(ossl_quic_obj_get0_mutex(ctx->obj)); #endif } /* Precondition: Channel mutex is held (unchecked) */ QUIC_NEEDS_LOCK static void qctx_unlock(QCTX *ctx) { #if defined(OPENSSL_THREADS) assert(ctx->obj != NULL); ossl_crypto_mutex_unlock(ossl_quic_obj_get0_mutex(ctx->obj)); #endif } static void qctx_lock_for_io(QCTX *ctx) { qctx_lock(ctx); ctx->in_io = 1; /* * We are entering an I/O function so we must update the values returned by * SSL_get_error and SSL_want. Set no error. This will be overridden later * if a call to QUIC_RAISE_NORMAL_ERROR or QUIC_RAISE_NON_NORMAL_ERROR * occurs during the API call. */ quic_set_last_error(ctx, SSL_ERROR_NONE); } /* * This predicate is the criterion which should determine API call rejection for * *most* mutating API calls, particularly stream-related operations for send * parts. * * A call is rejected (this function returns 0) if shutdown is in progress * (stream flushing), or we are in a TERMINATING or TERMINATED state. If * req_active=1, the connection must be active (i.e., the IDLE state is also * rejected). */ static int quic_mutation_allowed(QUIC_CONNECTION *qc, int req_active) { if (qc->shutting_down || ossl_quic_channel_is_term_any(qc->ch)) return 0; if (req_active && !ossl_quic_channel_is_active(qc->ch)) return 0; return 1; } static int qctx_is_top_level(QCTX *ctx) { return ctx->obj->parent_obj == NULL; } static int qctx_blocking(QCTX *ctx) { return ossl_quic_obj_blocking(ctx->obj); } /* * Block until a predicate is met. * * Precondition: Must have a channel. * Precondition: Must hold channel lock (unchecked). */ QUIC_NEEDS_LOCK static int block_until_pred(QCTX *ctx, int (*pred)(void *arg), void *pred_arg, uint32_t flags) { QUIC_ENGINE *qeng; QUIC_REACTOR *rtor; qeng = ossl_quic_obj_get0_engine(ctx->obj); assert(qeng != NULL); /* * Any attempt to block auto-disables tick inhibition as otherwise we will * hang around forever. */ ossl_quic_engine_set_inhibit_tick(qeng, 0); rtor = ossl_quic_engine_get0_reactor(qeng); return ossl_quic_reactor_block_until_pred(rtor, pred, pred_arg, flags); } /* * QUIC Front-End I/O API: Initialization * ====================================== * * SSL_new => ossl_quic_new * ossl_quic_init * SSL_reset => ossl_quic_reset * SSL_clear => ossl_quic_clear * ossl_quic_deinit * SSL_free => ossl_quic_free * * SSL_set_options => ossl_quic_set_options * SSL_get_options => ossl_quic_get_options * SSL_clear_options => ossl_quic_clear_options * */ /* SSL_new */ SSL *ossl_quic_new(SSL_CTX *ctx) { QUIC_CONNECTION *qc = NULL; SSL_CONNECTION *sc = NULL; /* * QUIC_server_method should not be used with SSL_new. * It should only be used with SSL_new_listener. */ if (ctx->method == OSSL_QUIC_server_method()) { QUIC_RAISE_NON_NORMAL_ERROR(NULL, ERR_R_SHOULD_NOT_HAVE_BEEN_CALLED, NULL); return NULL; } qc = OPENSSL_zalloc(sizeof(*qc)); if (qc == NULL) { QUIC_RAISE_NON_NORMAL_ERROR(NULL, ERR_R_CRYPTO_LIB, NULL); return NULL; } /* Create the QUIC domain mutex. */ #if defined(OPENSSL_THREADS) if ((qc->mutex = ossl_crypto_mutex_new()) == NULL) { QUIC_RAISE_NON_NORMAL_ERROR(NULL, ERR_R_CRYPTO_LIB, NULL); goto err; } #endif /* Create the handshake layer. */ qc->tls = ossl_ssl_connection_new_int(ctx, &qc->obj.ssl, TLS_method()); if (qc->tls == NULL || (sc = SSL_CONNECTION_FROM_SSL(qc->tls)) == NULL) { QUIC_RAISE_NON_NORMAL_ERROR(NULL, ERR_R_INTERNAL_ERROR, NULL); goto err; } /* override the user_ssl of the inner connection */ sc->s3.flags |= TLS1_FLAGS_QUIC | TLS1_FLAGS_QUIC_INTERNAL; /* Restrict options derived from the SSL_CTX. */ sc->options &= OSSL_QUIC_PERMITTED_OPTIONS_CONN; sc->pha_enabled = 0; /* Determine mode of operation. */ #if !defined(OPENSSL_NO_QUIC_THREAD_ASSIST) qc->is_thread_assisted = ((ctx->domain_flags & SSL_DOMAIN_FLAG_THREAD_ASSISTED) != 0); #endif qc->as_server = 0; qc->as_server_state = qc->as_server; if (!create_channel(qc, ctx)) goto err; ossl_quic_channel_set_msg_callback(qc->ch, ctx->msg_callback, &qc->obj.ssl); ossl_quic_channel_set_msg_callback_arg(qc->ch, ctx->msg_callback_arg); /* Initialise the QUIC_CONNECTION's QUIC_OBJ base. */ if (!ossl_quic_obj_init(&qc->obj, ctx, SSL_TYPE_QUIC_CONNECTION, NULL, qc->engine, qc->port)) { QUIC_RAISE_NON_NORMAL_ERROR(NULL, ERR_R_INTERNAL_ERROR, NULL); goto err; } /* Initialise libssl APL-related state. */ qc->default_stream_mode = SSL_DEFAULT_STREAM_MODE_AUTO_BIDI; qc->default_ssl_mode = qc->obj.ssl.ctx->mode; qc->default_ssl_options = qc->obj.ssl.ctx->options & OSSL_QUIC_PERMITTED_OPTIONS; qc->incoming_stream_policy = SSL_INCOMING_STREAM_POLICY_AUTO; qc->last_error = SSL_ERROR_NONE; qc_update_reject_policy(qc); /* * We do not create the default XSO yet. The reason for this is that the * stream ID of the default XSO will depend on whether the stream is client * or server-initiated, which depends on who transmits first. Since we do * not know whether the application will be using a client-transmits-first * or server-transmits-first protocol, we defer default XSO creation until * the client calls SSL_read() or SSL_write(). If it calls SSL_read() first, * we take that as a cue that the client is expecting a server-initiated * stream, and vice versa if SSL_write() is called first. */ return &qc->obj.ssl; err: if (qc != NULL) { qc_cleanup(qc, /*have_lock=*/0); OPENSSL_free(qc); } return NULL; } QUIC_NEEDS_LOCK static void quic_unref_port_bios(QUIC_PORT *port) { BIO *b; b = ossl_quic_port_get_net_rbio(port); BIO_free_all(b); b = ossl_quic_port_get_net_wbio(port); BIO_free_all(b); } QUIC_NEEDS_LOCK static void qc_cleanup(QUIC_CONNECTION *qc, int have_lock) { SSL_free(qc->tls); qc->tls = NULL; ossl_quic_channel_free(qc->ch); qc->ch = NULL; if (qc->port != NULL && qc->listener == NULL && qc->pending == 0) { /* TODO */ quic_unref_port_bios(qc->port); ossl_quic_port_free(qc->port); qc->port = NULL; ossl_quic_engine_free(qc->engine); qc->engine = NULL; } #if defined(OPENSSL_THREADS) if (have_lock) /* tsan doesn't like freeing locked mutexes */ ossl_crypto_mutex_unlock(qc->mutex); if (qc->listener == NULL && qc->pending == 0) ossl_crypto_mutex_free(&qc->mutex); #endif } /* SSL_free */ QUIC_TAKES_LOCK static void quic_free_listener(QCTX *ctx) { quic_unref_port_bios(ctx->ql->port); ossl_quic_port_drop_incoming(ctx->ql->port); ossl_quic_port_free(ctx->ql->port); if (ctx->ql->domain == NULL) { ossl_quic_engine_free(ctx->ql->engine); #if defined(OPENSSL_THREADS) ossl_crypto_mutex_free(&ctx->ql->mutex); #endif } else { SSL_free(&ctx->ql->domain->obj.ssl); } } /* SSL_free */ QUIC_TAKES_LOCK static void quic_free_domain(QCTX *ctx) { ossl_quic_engine_free(ctx->qd->engine); #if defined(OPENSSL_THREADS) ossl_crypto_mutex_free(&ctx->qd->mutex); #endif } QUIC_TAKES_LOCK void ossl_quic_free(SSL *s) { QCTX ctx; int is_default; /* We should never be called on anything but a QSO. */ if (!expect_quic_any(s, &ctx)) return; if (ctx.is_domain) { quic_free_domain(&ctx); return; } if (ctx.is_listener) { quic_free_listener(&ctx); return; } qctx_lock(&ctx); if (ctx.is_stream) { /* * When a QSSO is freed, the XSO is freed immediately, because the XSO * itself only contains API personality layer data. However the * underlying QUIC_STREAM is not freed immediately but is instead marked * as deleted for later collection. */ assert(ctx.qc->num_xso > 0); --ctx.qc->num_xso; /* If a stream's send part has not been finished, auto-reset it. */ if (( ctx.xso->stream->send_state == QUIC_SSTREAM_STATE_READY || ctx.xso->stream->send_state == QUIC_SSTREAM_STATE_SEND) && !ossl_quic_sstream_get_final_size(ctx.xso->stream->sstream, NULL)) ossl_quic_stream_map_reset_stream_send_part(ossl_quic_channel_get_qsm(ctx.qc->ch), ctx.xso->stream, 0); /* Do STOP_SENDING for the receive part, if applicable. */ if ( ctx.xso->stream->recv_state == QUIC_RSTREAM_STATE_RECV || ctx.xso->stream->recv_state == QUIC_RSTREAM_STATE_SIZE_KNOWN) ossl_quic_stream_map_stop_sending_recv_part(ossl_quic_channel_get_qsm(ctx.qc->ch), ctx.xso->stream, 0); /* Update stream state. */ ctx.xso->stream->deleted = 1; ossl_quic_stream_map_update_state(ossl_quic_channel_get_qsm(ctx.qc->ch), ctx.xso->stream); is_default = (ctx.xso == ctx.qc->default_xso); qctx_unlock(&ctx); /* * Unref the connection in most cases; the XSO has a ref to the QC and * not vice versa. But for a default XSO, to avoid circular references, * the QC refs the XSO but the XSO does not ref the QC. If we are the * default XSO, we only get here when the QC is being torn down anyway, * so don't call SSL_free(qc) as we are already in it. */ if (!is_default) SSL_free(&ctx.qc->obj.ssl); /* Note: SSL_free calls OPENSSL_free(xso) for us */ return; } /* * Free the default XSO, if any. The QUIC_STREAM is not deleted at this * stage, but is freed during the channel free when the whole QSM is freed. */ if (ctx.qc->default_xso != NULL) { QUIC_XSO *xso = ctx.qc->default_xso; qctx_unlock(&ctx); SSL_free(&xso->obj.ssl); qctx_lock(&ctx); ctx.qc->default_xso = NULL; } /* Ensure we have no remaining XSOs. */ assert(ctx.qc->num_xso == 0); #if !defined(OPENSSL_NO_QUIC_THREAD_ASSIST) if (ctx.qc->is_thread_assisted && ctx.qc->started) { ossl_quic_thread_assist_wait_stopped(&ctx.qc->thread_assist); ossl_quic_thread_assist_cleanup(&ctx.qc->thread_assist); } #endif /* * Note: SSL_free (that called this function) calls OPENSSL_free(ctx.qc) for * us */ qc_cleanup(ctx.qc, /*have_lock=*/1); /* Note: SSL_free calls OPENSSL_free(qc) for us */ if (ctx.qc->listener != NULL) SSL_free(&ctx.qc->listener->obj.ssl); if (ctx.qc->domain != NULL) SSL_free(&ctx.qc->domain->obj.ssl); } /* SSL method init */ int ossl_quic_init(SSL *s) { /* Same op as SSL_clear, forward the call. */ return ossl_quic_clear(s); } /* SSL method deinit */ void ossl_quic_deinit(SSL *s) { /* No-op. */ } /* SSL_clear (ssl_reset method) */ int ossl_quic_reset(SSL *s) { QCTX ctx; if (!expect_quic_any(s, &ctx)) return 0; ERR_raise(ERR_LIB_SSL, ERR_R_UNSUPPORTED); return 0; } /* ssl_clear method (unused) */ int ossl_quic_clear(SSL *s) { QCTX ctx; if (!expect_quic_any(s, &ctx)) return 0; ERR_raise(ERR_LIB_SSL, ERR_R_UNSUPPORTED); return 0; } int ossl_quic_set_override_now_cb(SSL *s, OSSL_TIME (*now_cb)(void *arg), void *now_cb_arg) { QCTX ctx; if (!expect_quic_any(s, &ctx)) return 0; qctx_lock(&ctx); ossl_quic_engine_set_time_cb(ctx.obj->engine, now_cb, now_cb_arg); qctx_unlock(&ctx); return 1; } void ossl_quic_conn_force_assist_thread_wake(SSL *s) { QCTX ctx; if (!expect_quic_conn_only(s, &ctx)) return; #if !defined(OPENSSL_NO_QUIC_THREAD_ASSIST) if (ctx.qc->is_thread_assisted && ctx.qc->started) ossl_quic_thread_assist_notify_deadline_changed(&ctx.qc->thread_assist); #endif } QUIC_NEEDS_LOCK static void qc_touch_default_xso(QUIC_CONNECTION *qc) { qc->default_xso_created = 1; qc_update_reject_policy(qc); } /* * Changes default XSO. Allows caller to keep reference to the old default XSO * (if any). Reference to new XSO is transferred from caller. */ QUIC_NEEDS_LOCK static void qc_set_default_xso_keep_ref(QUIC_CONNECTION *qc, QUIC_XSO *xso, int touch, QUIC_XSO **old_xso) { int refs; *old_xso = NULL; if (qc->default_xso != xso) { *old_xso = qc->default_xso; /* transfer old XSO ref to caller */ qc->default_xso = xso; if (xso == NULL) { /* * Changing to not having a default XSO. XSO becomes standalone and * now has a ref to the QC. */ if (!ossl_assert(SSL_up_ref(&qc->obj.ssl))) return; } else { /* * Changing from not having a default XSO to having one. The new XSO * will have had a reference to the QC we need to drop to avoid a * circular reference. * * Currently we never change directly from one default XSO to * another, though this function would also still be correct if this * weren't the case. */ assert(*old_xso == NULL); CRYPTO_DOWN_REF(&qc->obj.ssl.references, &refs); assert(refs > 0); } } if (touch) qc_touch_default_xso(qc); } /* * Changes default XSO, releasing the reference to any previous default XSO. * Reference to new XSO is transferred from caller. */ QUIC_NEEDS_LOCK static void qc_set_default_xso(QUIC_CONNECTION *qc, QUIC_XSO *xso, int touch) { QUIC_XSO *old_xso = NULL; qc_set_default_xso_keep_ref(qc, xso, touch, &old_xso); if (old_xso != NULL) SSL_free(&old_xso->obj.ssl); } QUIC_NEEDS_LOCK static void xso_update_options(QUIC_XSO *xso) { int cleanse = ((xso->ssl_options & SSL_OP_CLEANSE_PLAINTEXT) != 0); if (xso->stream->rstream != NULL) ossl_quic_rstream_set_cleanse(xso->stream->rstream, cleanse); if (xso->stream->sstream != NULL) ossl_quic_sstream_set_cleanse(xso->stream->sstream, cleanse); } /* * SSL_set_options * --------------- * * Setting options on a QCSO * - configures the handshake-layer options; * - configures the default data-plane options for new streams; * - configures the data-plane options on the default XSO, if there is one. * * Setting options on a QSSO * - configures data-plane options for that stream only. */ QUIC_TAKES_LOCK static uint64_t quic_mask_or_options(SSL *ssl, uint64_t mask_value, uint64_t or_value) { QCTX ctx; uint64_t hs_mask_value, hs_or_value, ret; if (!expect_quic_cs(ssl, &ctx)) return 0; qctx_lock(&ctx); if (!ctx.is_stream) { /* * If we were called on the connection, we apply any handshake option * changes. */ hs_mask_value = (mask_value & OSSL_QUIC_PERMITTED_OPTIONS_CONN); hs_or_value = (or_value & OSSL_QUIC_PERMITTED_OPTIONS_CONN); SSL_clear_options(ctx.qc->tls, hs_mask_value); SSL_set_options(ctx.qc->tls, hs_or_value); /* Update defaults for new streams. */ ctx.qc->default_ssl_options = ((ctx.qc->default_ssl_options & ~mask_value) | or_value) & OSSL_QUIC_PERMITTED_OPTIONS; } ret = ctx.qc->default_ssl_options; if (ctx.xso != NULL) { ctx.xso->ssl_options = ((ctx.xso->ssl_options & ~mask_value) | or_value) & OSSL_QUIC_PERMITTED_OPTIONS_STREAM; xso_update_options(ctx.xso); if (ctx.is_stream) ret = ctx.xso->ssl_options; } qctx_unlock(&ctx); return ret; } uint64_t ossl_quic_set_options(SSL *ssl, uint64_t options) { return quic_mask_or_options(ssl, 0, options); } /* SSL_clear_options */ uint64_t ossl_quic_clear_options(SSL *ssl, uint64_t options) { return quic_mask_or_options(ssl, options, 0); } /* SSL_get_options */ uint64_t ossl_quic_get_options(const SSL *ssl) { return quic_mask_or_options((SSL *)ssl, 0, 0); } /* * QUIC Front-End I/O API: Network BIO Configuration * ================================================= * * Handling the different BIOs is difficult: * * - It is more or less a requirement that we use non-blocking network I/O; * we need to be able to have timeouts on recv() calls, and make best effort * (non blocking) send() and recv() calls. * * The only sensible way to do this is to configure the socket into * non-blocking mode. We could try to do select() before calling send() or * recv() to get a guarantee that the call will not block, but this will * probably run into issues with buggy OSes which generate spurious socket * readiness events. In any case, relying on this to work reliably does not * seem sane. * * Timeouts could be handled via setsockopt() socket timeout options, but * this depends on OS support and adds another syscall to every network I/O * operation. It also has obvious thread safety concerns if we want to move * to concurrent use of a single socket at some later date. * * Some OSes support a MSG_DONTWAIT flag which allows a single I/O option to * be made non-blocking. However some OSes (e.g. Windows) do not support * this, so we cannot rely on this. * * As such, we need to configure any FD in non-blocking mode. This may * confound users who pass a blocking socket to libssl. However, in practice * it would be extremely strange for a user of QUIC to pass an FD to us, * then also try and send receive traffic on the same socket(!). Thus the * impact of this should be limited, and can be documented. * * - We support both blocking and non-blocking operation in terms of the API * presented to the user. One prospect is to set the blocking mode based on * whether the socket passed to us was already in blocking mode. However, * Windows has no API for determining if a socket is in blocking mode (!), * therefore this cannot be done portably. Currently therefore we expose an * explicit API call to set this, and default to blocking mode. * * - We need to determine our initial destination UDP address. The "natural" * way for a user to do this is to set the peer variable on a BIO_dgram. * However, this has problems because BIO_dgram's peer variable is used for * both transmission and reception. This means it can be constantly being * changed to a malicious value (e.g. if some random unrelated entity on the * network starts sending traffic to us) on every read call. This is not a * direct issue because we use the 'stateless' BIO_sendmmsg and BIO_recvmmsg * calls only, which do not use this variable. However, we do need to let * the user specify the peer in a 'normal' manner. The compromise here is * that we grab the current peer value set at the time the write BIO is set * and do not read the value again. * * - We also need to support memory BIOs (e.g. BIO_dgram_pair) or custom BIOs. * Currently we do this by only supporting non-blocking mode. * */ /* * Determines what initial destination UDP address we should use, if possible. * If this fails the client must set the destination address manually, or use a * BIO which does not need a destination address. */ static int csm_analyse_init_peer_addr(BIO *net_wbio, BIO_ADDR *peer) { if (BIO_dgram_detect_peer_addr(net_wbio, peer) <= 0) return 0; return 1; } static int quic_set0_net_rbio(QUIC_OBJ *obj, BIO *net_rbio) { QUIC_PORT *port; BIO *old_rbio = NULL; port = ossl_quic_obj_get0_port(obj); old_rbio = ossl_quic_port_get_net_rbio(port); if (old_rbio == net_rbio) return 0; if (!ossl_quic_port_set_net_rbio(port, net_rbio)) return 0; BIO_free_all(old_rbio); if (net_rbio != NULL) BIO_set_nbio(net_rbio, 1); /* best effort autoconfig */ return 1; } static int quic_set0_net_wbio(QUIC_OBJ *obj, BIO *net_wbio) { QUIC_PORT *port; BIO *old_wbio = NULL; port = ossl_quic_obj_get0_port(obj); old_wbio = ossl_quic_port_get_net_wbio(port); if (old_wbio == net_wbio) return 0; if (!ossl_quic_port_set_net_wbio(port, net_wbio)) return 0; BIO_free_all(old_wbio); if (net_wbio != NULL) BIO_set_nbio(net_wbio, 1); /* best effort autoconfig */ return 1; } void ossl_quic_conn_set0_net_rbio(SSL *s, BIO *net_rbio) { QCTX ctx; if (!expect_quic_csl(s, &ctx)) return; /* Returns 0 if no change. */ if (!quic_set0_net_rbio(ctx.obj, net_rbio)) return; } void ossl_quic_conn_set0_net_wbio(SSL *s, BIO *net_wbio) { QCTX ctx; if (!expect_quic_csl(s, &ctx)) return; /* Returns 0 if no change. */ if (!quic_set0_net_wbio(ctx.obj, net_wbio)) return; } BIO *ossl_quic_conn_get_net_rbio(const SSL *s) { QCTX ctx; QUIC_PORT *port; if (!expect_quic_csl(s, &ctx)) return NULL; port = ossl_quic_obj_get0_port(ctx.obj); assert(port != NULL); return ossl_quic_port_get_net_rbio(port); } BIO *ossl_quic_conn_get_net_wbio(const SSL *s) { QCTX ctx; QUIC_PORT *port; if (!expect_quic_csl(s, &ctx)) return NULL; port = ossl_quic_obj_get0_port(ctx.obj); assert(port != NULL); return ossl_quic_port_get_net_wbio(port); } int ossl_quic_conn_get_blocking_mode(const SSL *s) { QCTX ctx; if (!expect_quic_csl(s, &ctx)) return 0; return qctx_blocking(&ctx); } QUIC_TAKES_LOCK int ossl_quic_conn_set_blocking_mode(SSL *s, int blocking) { int ret = 0; unsigned int mode; QCTX ctx; if (!expect_quic_csl(s, &ctx)) return 0; qctx_lock(&ctx); /* Sanity check - can we support the request given the current network BIO? */ if (blocking) { /* * If called directly on a top-level object (QCSO or QLSO), update our * information on network BIO capabilities. */ if (qctx_is_top_level(&ctx)) ossl_quic_engine_update_poll_descriptors(ctx.obj->engine, /*force=*/1); /* Cannot enable blocking mode if we do not have pollable FDs. */ if (!ossl_quic_obj_can_support_blocking(ctx.obj)) { ret = QUIC_RAISE_NON_NORMAL_ERROR(&ctx, ERR_R_UNSUPPORTED, NULL); goto out; } } mode = (blocking != 0) ? QUIC_BLOCKING_MODE_BLOCKING : QUIC_BLOCKING_MODE_NONBLOCKING; ossl_quic_obj_set_blocking_mode(ctx.obj, mode); ret = 1; out: qctx_unlock(&ctx); return ret; } int ossl_quic_conn_set_initial_peer_addr(SSL *s, const BIO_ADDR *peer_addr) { QCTX ctx; if (!expect_quic_cs(s, &ctx)) return 0; if (ctx.qc->started) return QUIC_RAISE_NON_NORMAL_ERROR(&ctx, ERR_R_SHOULD_NOT_HAVE_BEEN_CALLED, NULL); if (peer_addr == NULL) { BIO_ADDR_clear(&ctx.qc->init_peer_addr); return 1; } return BIO_ADDR_copy(&ctx.qc->init_peer_addr, peer_addr); } /* * QUIC Front-End I/O API: Asynchronous I/O Management * =================================================== * * (BIO/)SSL_handle_events => ossl_quic_handle_events * (BIO/)SSL_get_event_timeout => ossl_quic_get_event_timeout * (BIO/)SSL_get_poll_fd => ossl_quic_get_poll_fd * */ /* SSL_handle_events; performs QUIC I/O and timeout processing. */ QUIC_TAKES_LOCK int ossl_quic_handle_events(SSL *s) { QCTX ctx; if (!expect_quic_any(s, &ctx)) return 0; qctx_lock(&ctx); ossl_quic_reactor_tick(ossl_quic_obj_get0_reactor(ctx.obj), 0); qctx_unlock(&ctx); return 1; } /* * SSL_get_event_timeout. Get the time in milliseconds until the SSL object * should next have events handled by the application by calling * SSL_handle_events(). tv is set to 0 if the object should have events handled * immediately. If no timeout is currently active, *is_infinite is set to 1 and * the value of *tv is undefined. */ QUIC_TAKES_LOCK int ossl_quic_get_event_timeout(SSL *s, struct timeval *tv, int *is_infinite) { QCTX ctx; QUIC_REACTOR *reactor; OSSL_TIME deadline; OSSL_TIME basetime; if (!expect_quic_any(s, &ctx)) return 0; qctx_lock(&ctx); reactor = ossl_quic_obj_get0_reactor(ctx.obj); deadline = ossl_quic_reactor_get_tick_deadline(reactor); if (ossl_time_is_infinite(deadline)) { qctx_unlock(&ctx); *is_infinite = 1; /* * Robustness against faulty applications that don't check *is_infinite; * harmless long timeout. */ tv->tv_sec = 1000000; tv->tv_usec = 0; return 1; } basetime = ossl_quic_engine_get_time(ctx.obj->engine); qctx_unlock(&ctx); *tv = ossl_time_to_timeval(ossl_time_subtract(deadline, basetime)); *is_infinite = 0; return 1; } /* SSL_get_rpoll_descriptor */ int ossl_quic_get_rpoll_descriptor(SSL *s, BIO_POLL_DESCRIPTOR *desc) { QCTX ctx; QUIC_PORT *port = NULL; BIO *net_rbio; if (!expect_quic_csl(s, &ctx)) return 0; port = ossl_quic_obj_get0_port(ctx.obj); net_rbio = ossl_quic_port_get_net_rbio(port); if (desc == NULL || net_rbio == NULL) return QUIC_RAISE_NON_NORMAL_ERROR(&ctx, ERR_R_PASSED_INVALID_ARGUMENT, NULL); return BIO_get_rpoll_descriptor(net_rbio, desc); } /* SSL_get_wpoll_descriptor */ int ossl_quic_get_wpoll_descriptor(SSL *s, BIO_POLL_DESCRIPTOR *desc) { QCTX ctx; QUIC_PORT *port = NULL; BIO *net_wbio; if (!expect_quic_csl(s, &ctx)) return 0; port = ossl_quic_obj_get0_port(ctx.obj); net_wbio = ossl_quic_port_get_net_wbio(port); if (desc == NULL || net_wbio == NULL) return QUIC_RAISE_NON_NORMAL_ERROR(&ctx, ERR_R_PASSED_INVALID_ARGUMENT, NULL); return BIO_get_wpoll_descriptor(net_wbio, desc); } /* SSL_net_read_desired */ QUIC_TAKES_LOCK int ossl_quic_get_net_read_desired(SSL *s) { QCTX ctx; int ret; if (!expect_quic_csl(s, &ctx)) return 0; qctx_lock(&ctx); ret = ossl_quic_reactor_net_read_desired(ossl_quic_obj_get0_reactor(ctx.obj)); qctx_unlock(&ctx); return ret; } /* SSL_net_write_desired */ QUIC_TAKES_LOCK int ossl_quic_get_net_write_desired(SSL *s) { int ret; QCTX ctx; if (!expect_quic_csl(s, &ctx)) return 0; qctx_lock(&ctx); ret = ossl_quic_reactor_net_write_desired(ossl_quic_obj_get0_reactor(ctx.obj)); qctx_unlock(&ctx); return ret; } /* * QUIC Front-End I/O API: Connection Lifecycle Operations * ======================================================= * * SSL_do_handshake => ossl_quic_do_handshake * SSL_set_connect_state => ossl_quic_set_connect_state * SSL_set_accept_state => ossl_quic_set_accept_state * SSL_shutdown => ossl_quic_shutdown * SSL_ctrl => ossl_quic_ctrl * (BIO/)SSL_connect => ossl_quic_connect * (BIO/)SSL_accept => ossl_quic_accept * */ QUIC_NEEDS_LOCK static void qc_shutdown_flush_init(QUIC_CONNECTION *qc) { QUIC_STREAM_MAP *qsm; if (qc->shutting_down) return; qsm = ossl_quic_channel_get_qsm(qc->ch); ossl_quic_stream_map_begin_shutdown_flush(qsm); qc->shutting_down = 1; } /* Returns 1 if all shutdown-flush streams have been done with. */ QUIC_NEEDS_LOCK static int qc_shutdown_flush_finished(QUIC_CONNECTION *qc) { QUIC_STREAM_MAP *qsm = ossl_quic_channel_get_qsm(qc->ch); return qc->shutting_down && ossl_quic_stream_map_is_shutdown_flush_finished(qsm); } /* SSL_shutdown */ static int quic_shutdown_wait(void *arg) { QUIC_CONNECTION *qc = arg; return ossl_quic_channel_is_terminated(qc->ch); } /* Returns 1 if shutdown flush process has finished or is inapplicable. */ static int quic_shutdown_flush_wait(void *arg) { QUIC_CONNECTION *qc = arg; return ossl_quic_channel_is_term_any(qc->ch) || qc_shutdown_flush_finished(qc); } static int quic_shutdown_peer_wait(void *arg) { QUIC_CONNECTION *qc = arg; return ossl_quic_channel_is_term_any(qc->ch); } QUIC_TAKES_LOCK int ossl_quic_conn_shutdown(SSL *s, uint64_t flags, const SSL_SHUTDOWN_EX_ARGS *args, size_t args_len) { int ret; QCTX ctx; int stream_flush = ((flags & SSL_SHUTDOWN_FLAG_NO_STREAM_FLUSH) == 0); int no_block = ((flags & SSL_SHUTDOWN_FLAG_NO_BLOCK) != 0); int wait_peer = ((flags & SSL_SHUTDOWN_FLAG_WAIT_PEER) != 0); if (!expect_quic_cs(s, &ctx)) return -1; if (ctx.is_stream) { QUIC_RAISE_NON_NORMAL_ERROR(&ctx, SSL_R_CONN_USE_ONLY, NULL); return -1; } qctx_lock(&ctx); if (ossl_quic_channel_is_terminated(ctx.qc->ch)) { qctx_unlock(&ctx); return 1; } /* Phase 1: Stream Flushing */ if (!wait_peer && stream_flush) { qc_shutdown_flush_init(ctx.qc); if (!qc_shutdown_flush_finished(ctx.qc)) { if (!no_block && qctx_blocking(&ctx)) { ret = block_until_pred(&ctx, quic_shutdown_flush_wait, ctx.qc, 0); if (ret < 1) { ret = 0; goto err; } } else { qctx_maybe_autotick(&ctx); } } if (!qc_shutdown_flush_finished(ctx.qc)) { qctx_unlock(&ctx); return 0; /* ongoing */ } } /* Phase 2: Connection Closure */ if (wait_peer && !ossl_quic_channel_is_term_any(ctx.qc->ch)) { if (!no_block && qctx_blocking(&ctx)) { ret = block_until_pred(&ctx, quic_shutdown_peer_wait, ctx.qc, 0); if (ret < 1) { ret = 0; goto err; } } else { qctx_maybe_autotick(&ctx); } if (!ossl_quic_channel_is_term_any(ctx.qc->ch)) { ret = 0; /* peer hasn't closed yet - still not done */ goto err; } /* * We are at least terminating - go through the normal process of * waiting until we are in the TERMINATED state. */ } /* Block mutation ops regardless of if we did stream flush. */ ctx.qc->shutting_down = 1; /* * This call is a no-op if we are already terminating, so it doesn't * affect the wait_peer case. */ ossl_quic_channel_local_close(ctx.qc->ch, args != NULL ? args->quic_error_code : 0, args != NULL ? args->quic_reason : NULL); SSL_set_shutdown(ctx.qc->tls, SSL_SENT_SHUTDOWN); if (ossl_quic_channel_is_terminated(ctx.qc->ch)) { qctx_unlock(&ctx); return 1; } /* Phase 3: Terminating Wait Time */ if (!no_block && qctx_blocking(&ctx) && (flags & SSL_SHUTDOWN_FLAG_RAPID) == 0) { ret = block_until_pred(&ctx, quic_shutdown_wait, ctx.qc, 0); if (ret < 1) { ret = 0; goto err; } } else { qctx_maybe_autotick(&ctx); } ret = ossl_quic_channel_is_terminated(ctx.qc->ch); err: qctx_unlock(&ctx); return ret; } /* SSL_ctrl */ long ossl_quic_ctrl(SSL *s, int cmd, long larg, void *parg) { QCTX ctx; if (!expect_quic_csl(s, &ctx)) return 0; switch (cmd) { case SSL_CTRL_MODE: if (ctx.is_listener) return QUIC_RAISE_NON_NORMAL_ERROR(&ctx, ERR_R_UNSUPPORTED, NULL); /* If called on a QCSO, update the default mode. */ if (!ctx.is_stream) ctx.qc->default_ssl_mode |= (uint32_t)larg; /* * If we were called on a QSSO or have a default stream, we also update * that. */ if (ctx.xso != NULL) { /* Cannot enable EPW while AON write in progress. */ if (ctx.xso->aon_write_in_progress) larg &= ~SSL_MODE_ENABLE_PARTIAL_WRITE; ctx.xso->ssl_mode |= (uint32_t)larg; return ctx.xso->ssl_mode; } return ctx.qc->default_ssl_mode; case SSL_CTRL_CLEAR_MODE: if (ctx.is_listener) return QUIC_RAISE_NON_NORMAL_ERROR(&ctx, ERR_R_UNSUPPORTED, NULL); if (!ctx.is_stream) ctx.qc->default_ssl_mode &= ~(uint32_t)larg; if (ctx.xso != NULL) { ctx.xso->ssl_mode &= ~(uint32_t)larg; return ctx.xso->ssl_mode; } return ctx.qc->default_ssl_mode; case SSL_CTRL_SET_MSG_CALLBACK_ARG: if (ctx.is_listener) return QUIC_RAISE_NON_NORMAL_ERROR(&ctx, ERR_R_UNSUPPORTED, NULL); ossl_quic_channel_set_msg_callback_arg(ctx.qc->ch, parg); /* This ctrl also needs to be passed to the internal SSL object */ return SSL_ctrl(ctx.qc->tls, cmd, larg, parg); case DTLS_CTRL_GET_TIMEOUT: /* DTLSv1_get_timeout */ { int is_infinite; if (!ossl_quic_get_event_timeout(s, parg, &is_infinite)) return 0; return !is_infinite; } case DTLS_CTRL_HANDLE_TIMEOUT: /* DTLSv1_handle_timeout */ /* For legacy compatibility with DTLS calls. */ return ossl_quic_handle_events(s) == 1 ? 1 : -1; /* Mask ctrls we shouldn't support for QUIC. */ case SSL_CTRL_GET_READ_AHEAD: case SSL_CTRL_SET_READ_AHEAD: case SSL_CTRL_SET_MAX_SEND_FRAGMENT: case SSL_CTRL_SET_SPLIT_SEND_FRAGMENT: case SSL_CTRL_SET_MAX_PIPELINES: return 0; default: /* * Probably a TLS related ctrl. Send back to the frontend SSL_ctrl * implementation. Either SSL_ctrl will handle it itself by direct * access into handshake layer state, or failing that, it will be passed * to the handshake layer via the SSL_METHOD vtable. If the ctrl is not * supported by anything, the handshake layer's ctrl method will finally * return 0. */ if (ctx.is_listener) return QUIC_RAISE_NON_NORMAL_ERROR(&ctx, ERR_R_UNSUPPORTED, NULL); return ossl_ctrl_internal(&ctx.qc->obj.ssl, cmd, larg, parg, /*no_quic=*/1); } } /* SSL_set_connect_state */ int ossl_quic_set_connect_state(SSL *s, int raiseerrs) { QCTX ctx; if (!is_quic_c(s, &ctx, raiseerrs)) return 0; if (ctx.qc->as_server_state == 0) return 1; /* Cannot be changed after handshake started */ if (ctx.qc->started) { if (raiseerrs) QUIC_RAISE_NON_NORMAL_ERROR(NULL, SSL_R_INVALID_COMMAND, NULL); return 0; } ctx.qc->as_server_state = 0; return 1; } /* SSL_set_accept_state */ int ossl_quic_set_accept_state(SSL *s, int raiseerrs) { QCTX ctx; if (!is_quic_c(s, &ctx, raiseerrs)) return 0; if (ctx.qc->as_server_state == 1) return 1; /* Cannot be changed after handshake started */ if (ctx.qc->started) { if (raiseerrs) QUIC_RAISE_NON_NORMAL_ERROR(NULL, SSL_R_INVALID_COMMAND, NULL); return 0; } ctx.qc->as_server_state = 1; return 1; } /* SSL_do_handshake */ struct quic_handshake_wait_args { QUIC_CONNECTION *qc; }; static int tls_wants_non_io_retry(QUIC_CONNECTION *qc) { int want = SSL_want(qc->tls); if (want == SSL_X509_LOOKUP || want == SSL_CLIENT_HELLO_CB || want == SSL_RETRY_VERIFY) return 1; return 0; } static int quic_handshake_wait(void *arg) { struct quic_handshake_wait_args *args = arg; if (!quic_mutation_allowed(args->qc, /*req_active=*/1)) return -1; if (ossl_quic_channel_is_handshake_complete(args->qc->ch)) return 1; if (tls_wants_non_io_retry(args->qc)) return 1; return 0; } static int configure_channel(QUIC_CONNECTION *qc) { assert(qc->ch != NULL); if (!ossl_quic_channel_set_peer_addr(qc->ch, &qc->init_peer_addr)) return 0; return 1; } static int need_notifier_for_domain_flags(uint64_t domain_flags) { return (domain_flags & SSL_DOMAIN_FLAG_THREAD_ASSISTED) != 0 || ((domain_flags & SSL_DOMAIN_FLAG_MULTI_THREAD) != 0 && (domain_flags & SSL_DOMAIN_FLAG_BLOCKING) != 0); } QUIC_NEEDS_LOCK static int create_channel(QUIC_CONNECTION *qc, SSL_CTX *ctx) { QUIC_ENGINE_ARGS engine_args = {0}; QUIC_PORT_ARGS port_args = {0}; engine_args.libctx = ctx->libctx; engine_args.propq = ctx->propq; #if defined(OPENSSL_THREADS) engine_args.mutex = qc->mutex; #endif if (need_notifier_for_domain_flags(ctx->domain_flags)) engine_args.reactor_flags |= QUIC_REACTOR_FLAG_USE_NOTIFIER; qc->engine = ossl_quic_engine_new(&engine_args); if (qc->engine == NULL) { QUIC_RAISE_NON_NORMAL_ERROR(NULL, ERR_R_INTERNAL_ERROR, NULL); return 0; } port_args.channel_ctx = ctx; qc->port = ossl_quic_engine_create_port(qc->engine, &port_args); if (qc->port == NULL) { QUIC_RAISE_NON_NORMAL_ERROR(NULL, ERR_R_INTERNAL_ERROR, NULL); ossl_quic_engine_free(qc->engine); return 0; } qc->ch = ossl_quic_port_create_outgoing(qc->port, qc->tls); if (qc->ch == NULL) { QUIC_RAISE_NON_NORMAL_ERROR(NULL, ERR_R_INTERNAL_ERROR, NULL); ossl_quic_port_free(qc->port); ossl_quic_engine_free(qc->engine); return 0; } return 1; } /* * Configures a channel with the information we have accumulated via calls made * to us from the application prior to starting a handshake attempt. */ QUIC_NEEDS_LOCK static int ensure_channel_started(QCTX *ctx) { QUIC_CONNECTION *qc = ctx->qc; if (!qc->started) { if (!configure_channel(qc)) { QUIC_RAISE_NON_NORMAL_ERROR(ctx, ERR_R_INTERNAL_ERROR, "failed to configure channel"); return 0; } if (!ossl_quic_channel_start(qc->ch)) { ossl_quic_channel_restore_err_state(qc->ch); QUIC_RAISE_NON_NORMAL_ERROR(ctx, ERR_R_INTERNAL_ERROR, "failed to start channel"); return 0; } #if !defined(OPENSSL_NO_QUIC_THREAD_ASSIST) if (qc->is_thread_assisted) if (!ossl_quic_thread_assist_init_start(&qc->thread_assist, qc->ch)) { QUIC_RAISE_NON_NORMAL_ERROR(ctx, ERR_R_INTERNAL_ERROR, "failed to start assist thread"); return 0; } #endif } qc->started = 1; return 1; } QUIC_NEEDS_LOCK static int quic_do_handshake(QCTX *ctx) { int ret; QUIC_CONNECTION *qc = ctx->qc; QUIC_PORT *port; BIO *net_rbio, *net_wbio; if (ossl_quic_channel_is_handshake_complete(qc->ch)) /* Handshake already completed. */ return 1; if (!quic_mutation_allowed(qc, /*req_active=*/0)) return QUIC_RAISE_NON_NORMAL_ERROR(ctx, SSL_R_PROTOCOL_IS_SHUTDOWN, NULL); if (qc->as_server != qc->as_server_state) { QUIC_RAISE_NON_NORMAL_ERROR(ctx, ERR_R_PASSED_INVALID_ARGUMENT, NULL); return -1; /* Non-protocol error */ } port = ossl_quic_obj_get0_port(ctx->obj); net_rbio = ossl_quic_port_get_net_rbio(port); net_wbio = ossl_quic_port_get_net_wbio(port); if (net_rbio == NULL || net_wbio == NULL) { /* Need read and write BIOs. */ QUIC_RAISE_NON_NORMAL_ERROR(ctx, SSL_R_BIO_NOT_SET, NULL); return -1; /* Non-protocol error */ } if (!qc->started && ossl_quic_port_is_addressed_w(port) && BIO_ADDR_family(&qc->init_peer_addr) == AF_UNSPEC) { /* * We are trying to connect and are using addressed mode, which means we * need an initial peer address; if we do not have a peer address yet, * we should try to autodetect one. * * We do this as late as possible because some BIOs (e.g. BIO_s_connect) * may not be able to provide us with a peer address until they have * finished their own processing. They may not be able to perform this * processing until an application has finished configuring that BIO * (e.g. with setter calls), which might happen after SSL_set_bio is * called. */ if (!csm_analyse_init_peer_addr(net_wbio, &qc->init_peer_addr)) /* best effort */ BIO_ADDR_clear(&qc->init_peer_addr); else ossl_quic_channel_set_peer_addr(qc->ch, &qc->init_peer_addr); } if (!qc->started && ossl_quic_port_is_addressed_w(port) && BIO_ADDR_family(&qc->init_peer_addr) == AF_UNSPEC) { /* * If we still don't have a peer address in addressed mode, we can't do * anything. */ QUIC_RAISE_NON_NORMAL_ERROR(ctx, SSL_R_REMOTE_PEER_ADDRESS_NOT_SET, NULL); return -1; /* Non-protocol error */ } /* * Start connection process. Note we may come here multiple times in * non-blocking mode, which is fine. */ if (!ensure_channel_started(ctx)) /* raises on failure */ return -1; /* Non-protocol error */ if (ossl_quic_channel_is_handshake_complete(qc->ch)) /* The handshake is now done. */ return 1; if (!qctx_blocking(ctx)) { /* Try to advance the reactor. */ qctx_maybe_autotick(ctx); if (ossl_quic_channel_is_handshake_complete(qc->ch)) /* The handshake is now done. */ return 1; if (ossl_quic_channel_is_term_any(qc->ch)) { QUIC_RAISE_NON_NORMAL_ERROR(ctx, SSL_R_PROTOCOL_IS_SHUTDOWN, NULL); return 0; } else if (ossl_quic_obj_desires_blocking(&qc->obj)) { /* * As a special case when doing a handshake when blocking mode is * desired yet not available, see if the network BIOs have become * poll descriptor-enabled. This supports BIOs such as BIO_s_connect * which do late creation of socket FDs and therefore cannot expose * a poll descriptor until after a network BIO is set on the QCSO. */ ossl_quic_engine_update_poll_descriptors(qc->obj.engine, /*force=*/1); } } /* * We are either in blocking mode or just entered it due to the code above. */ if (qctx_blocking(ctx)) { /* In blocking mode, wait for the handshake to complete. */ struct quic_handshake_wait_args args; args.qc = qc; ret = block_until_pred(ctx, quic_handshake_wait, &args, 0); if (!quic_mutation_allowed(qc, /*req_active=*/1)) { QUIC_RAISE_NON_NORMAL_ERROR(ctx, SSL_R_PROTOCOL_IS_SHUTDOWN, NULL); return 0; /* Shutdown before completion */ } else if (ret <= 0) { QUIC_RAISE_NON_NORMAL_ERROR(ctx, ERR_R_INTERNAL_ERROR, NULL); return -1; /* Non-protocol error */ } if (tls_wants_non_io_retry(qc)) { QUIC_RAISE_NORMAL_ERROR(ctx, SSL_get_error(qc->tls, 0)); return -1; } assert(ossl_quic_channel_is_handshake_complete(qc->ch)); return 1; } if (tls_wants_non_io_retry(qc)) { QUIC_RAISE_NORMAL_ERROR(ctx, SSL_get_error(qc->tls, 0)); return -1; } /* * Otherwise, indicate that the handshake isn't done yet. * We can only get here in non-blocking mode. */ QUIC_RAISE_NORMAL_ERROR(ctx, SSL_ERROR_WANT_READ); return -1; /* Non-protocol error */ } QUIC_TAKES_LOCK int ossl_quic_do_handshake(SSL *s) { int ret; QCTX ctx; if (!expect_quic_cs(s, &ctx)) return 0; qctx_lock_for_io(&ctx); ret = quic_do_handshake(&ctx); qctx_unlock(&ctx); return ret; } /* SSL_connect */ int ossl_quic_connect(SSL *s) { /* Ensure we are in connect state (no-op if non-idle). */ if (!ossl_quic_set_connect_state(s, 1)) return -1; /* Begin or continue the handshake */ return ossl_quic_do_handshake(s); } /* SSL_accept */ int ossl_quic_accept(SSL *s) { /* Ensure we are in accept state (no-op if non-idle). */ if (!ossl_quic_set_accept_state(s, 1)) return -1; /* Begin or continue the handshake */ return ossl_quic_do_handshake(s); } /* * QUIC Front-End I/O API: Stream Lifecycle Operations * =================================================== * * SSL_stream_new => ossl_quic_conn_stream_new * */ /* * Try to create the default XSO if it doesn't already exist. Returns 1 if the * default XSO was created. Returns 0 if it was not (e.g. because it already * exists). Note that this is NOT an error condition. */ QUIC_NEEDS_LOCK static int qc_try_create_default_xso_for_write(QCTX *ctx) { uint64_t flags = 0; QUIC_CONNECTION *qc = ctx->qc; if (qc->default_xso_created || qc->default_stream_mode == SSL_DEFAULT_STREAM_MODE_NONE) /* * We only do this once. If the user detaches a previously created * default XSO we don't auto-create another one. */ return QUIC_RAISE_NON_NORMAL_ERROR(ctx, SSL_R_NO_STREAM, NULL); /* Create a locally-initiated stream. */ if (qc->default_stream_mode == SSL_DEFAULT_STREAM_MODE_AUTO_UNI) flags |= SSL_STREAM_FLAG_UNI; qc_set_default_xso(qc, (QUIC_XSO *)quic_conn_stream_new(ctx, flags, /*needs_lock=*/0), /*touch=*/0); if (qc->default_xso == NULL) return QUIC_RAISE_NON_NORMAL_ERROR(ctx, ERR_R_INTERNAL_ERROR, NULL); qc_touch_default_xso(qc); return 1; } struct quic_wait_for_stream_args { QUIC_CONNECTION *qc; QUIC_STREAM *qs; QCTX *ctx; uint64_t expect_id; }; QUIC_NEEDS_LOCK static int quic_wait_for_stream(void *arg) { struct quic_wait_for_stream_args *args = arg; if (!quic_mutation_allowed(args->qc, /*req_active=*/1)) { /* If connection is torn down due to an error while blocking, stop. */ QUIC_RAISE_NON_NORMAL_ERROR(args->ctx, SSL_R_PROTOCOL_IS_SHUTDOWN, NULL); return -1; } args->qs = ossl_quic_stream_map_get_by_id(ossl_quic_channel_get_qsm(args->qc->ch), args->expect_id | QUIC_STREAM_DIR_BIDI); if (args->qs == NULL) args->qs = ossl_quic_stream_map_get_by_id(ossl_quic_channel_get_qsm(args->qc->ch), args->expect_id | QUIC_STREAM_DIR_UNI); if (args->qs != NULL) return 1; /* stream now exists */ return 0; /* did not get a stream, keep trying */ } QUIC_NEEDS_LOCK static int qc_wait_for_default_xso_for_read(QCTX *ctx, int peek) { /* Called on a QCSO and we don't currently have a default stream. */ uint64_t expect_id; QUIC_CONNECTION *qc = ctx->qc; QUIC_STREAM *qs; int res; struct quic_wait_for_stream_args wargs; OSSL_RTT_INFO rtt_info; /* * If default stream functionality is disabled or we already detached * one, don't make another default stream and just fail. */ if (qc->default_xso_created || qc->default_stream_mode == SSL_DEFAULT_STREAM_MODE_NONE) return QUIC_RAISE_NON_NORMAL_ERROR(ctx, SSL_R_NO_STREAM, NULL); /* * The peer may have opened a stream since we last ticked. So tick and * see if the stream with ordinal 0 (remote, bidi/uni based on stream * mode) exists yet. QUIC stream IDs must be allocated in order, so the * first stream created by a peer must have an ordinal of 0. */ expect_id = qc->as_server ? QUIC_STREAM_INITIATOR_CLIENT : QUIC_STREAM_INITIATOR_SERVER; qs = ossl_quic_stream_map_get_by_id(ossl_quic_channel_get_qsm(qc->ch), expect_id | QUIC_STREAM_DIR_BIDI); if (qs == NULL) qs = ossl_quic_stream_map_get_by_id(ossl_quic_channel_get_qsm(qc->ch), expect_id | QUIC_STREAM_DIR_UNI); if (qs == NULL) { qctx_maybe_autotick(ctx); qs = ossl_quic_stream_map_get_by_id(ossl_quic_channel_get_qsm(qc->ch), expect_id); } if (qs == NULL) { if (peek) return 0; if (ossl_quic_channel_is_term_any(qc->ch)) { return QUIC_RAISE_NON_NORMAL_ERROR(ctx, SSL_R_PROTOCOL_IS_SHUTDOWN, NULL); } else if (!qctx_blocking(ctx)) { /* Non-blocking mode, so just bail immediately. */ return QUIC_RAISE_NORMAL_ERROR(ctx, SSL_ERROR_WANT_READ); } /* Block until we have a stream. */ wargs.qc = qc; wargs.qs = NULL; wargs.ctx = ctx; wargs.expect_id = expect_id; res = block_until_pred(ctx, quic_wait_for_stream, &wargs, 0); if (res == 0) return QUIC_RAISE_NON_NORMAL_ERROR(ctx, ERR_R_INTERNAL_ERROR, NULL); else if (res < 0 || wargs.qs == NULL) /* quic_wait_for_stream raised error here */ return 0; qs = wargs.qs; } /* * We now have qs != NULL. Remove it from the incoming stream queue so that * it isn't also returned by any future SSL_accept_stream calls. */ ossl_statm_get_rtt_info(ossl_quic_channel_get_statm(qc->ch), &rtt_info); ossl_quic_stream_map_remove_from_accept_queue(ossl_quic_channel_get_qsm(qc->ch), qs, rtt_info.smoothed_rtt); /* * Now make qs the default stream, creating the necessary XSO. */ qc_set_default_xso(qc, create_xso_from_stream(qc, qs), /*touch=*/0); if (qc->default_xso == NULL) return QUIC_RAISE_NON_NORMAL_ERROR(ctx, ERR_R_INTERNAL_ERROR, NULL); qc_touch_default_xso(qc); /* inhibits default XSO */ return 1; } QUIC_NEEDS_LOCK static QUIC_XSO *create_xso_from_stream(QUIC_CONNECTION *qc, QUIC_STREAM *qs) { QUIC_XSO *xso = NULL; if ((xso = OPENSSL_zalloc(sizeof(*xso))) == NULL) { QUIC_RAISE_NON_NORMAL_ERROR(NULL, ERR_R_CRYPTO_LIB, NULL); goto err; } if (!ossl_quic_obj_init(&xso->obj, qc->obj.ssl.ctx, SSL_TYPE_QUIC_XSO, &qc->obj.ssl, NULL, NULL)) { QUIC_RAISE_NON_NORMAL_ERROR(NULL, ERR_R_INTERNAL_ERROR, NULL); goto err; } /* XSO refs QC */ if (!SSL_up_ref(&qc->obj.ssl)) { QUIC_RAISE_NON_NORMAL_ERROR(NULL, ERR_R_SSL_LIB, NULL); goto err; } xso->conn = qc; xso->ssl_mode = qc->default_ssl_mode; xso->ssl_options = qc->default_ssl_options & OSSL_QUIC_PERMITTED_OPTIONS_STREAM; xso->last_error = SSL_ERROR_NONE; xso->stream = qs; ++qc->num_xso; xso_update_options(xso); return xso; err: OPENSSL_free(xso); return NULL; } struct quic_new_stream_wait_args { QUIC_CONNECTION *qc; int is_uni; }; static int quic_new_stream_wait(void *arg) { struct quic_new_stream_wait_args *args = arg; QUIC_CONNECTION *qc = args->qc; if (!quic_mutation_allowed(qc, /*req_active=*/1)) return -1; if (ossl_quic_channel_is_new_local_stream_admissible(qc->ch, args->is_uni)) return 1; return 0; } /* locking depends on need_lock */ static SSL *quic_conn_stream_new(QCTX *ctx, uint64_t flags, int need_lock) { int ret; QUIC_CONNECTION *qc = ctx->qc; QUIC_XSO *xso = NULL; QUIC_STREAM *qs = NULL; int is_uni = ((flags & SSL_STREAM_FLAG_UNI) != 0); int no_blocking = ((flags & SSL_STREAM_FLAG_NO_BLOCK) != 0); int advance = ((flags & SSL_STREAM_FLAG_ADVANCE) != 0); if (need_lock) qctx_lock(ctx); if (!quic_mutation_allowed(qc, /*req_active=*/0)) { QUIC_RAISE_NON_NORMAL_ERROR(ctx, SSL_R_PROTOCOL_IS_SHUTDOWN, NULL); goto err; } if (!advance && !ossl_quic_channel_is_new_local_stream_admissible(qc->ch, is_uni)) { struct quic_new_stream_wait_args args; /* * Stream count flow control currently doesn't permit this stream to be * opened. */ if (no_blocking || !qctx_blocking(ctx)) { QUIC_RAISE_NON_NORMAL_ERROR(ctx, SSL_R_STREAM_COUNT_LIMITED, NULL); goto err; } args.qc = qc; args.is_uni = is_uni; /* Blocking mode - wait until we can get a stream. */ ret = block_until_pred(ctx, quic_new_stream_wait, &args, 0); if (!quic_mutation_allowed(qc, /*req_active=*/1)) { QUIC_RAISE_NON_NORMAL_ERROR(ctx, SSL_R_PROTOCOL_IS_SHUTDOWN, NULL); goto err; /* Shutdown before completion */ } else if (ret <= 0) { QUIC_RAISE_NON_NORMAL_ERROR(ctx, ERR_R_INTERNAL_ERROR, NULL); goto err; /* Non-protocol error */ } } qs = ossl_quic_channel_new_stream_local(qc->ch, is_uni); if (qs == NULL) { QUIC_RAISE_NON_NORMAL_ERROR(ctx, ERR_R_INTERNAL_ERROR, NULL); goto err; } xso = create_xso_from_stream(qc, qs); if (xso == NULL) goto err; qc_touch_default_xso(qc); /* inhibits default XSO */ if (need_lock) qctx_unlock(ctx); return &xso->obj.ssl; err: OPENSSL_free(xso); ossl_quic_stream_map_release(ossl_quic_channel_get_qsm(qc->ch), qs); if (need_lock) qctx_unlock(ctx); return NULL; } QUIC_TAKES_LOCK SSL *ossl_quic_conn_stream_new(SSL *s, uint64_t flags) { QCTX ctx; if (!expect_quic_conn_only(s, &ctx)) return NULL; return quic_conn_stream_new(&ctx, flags, /*need_lock=*/1); } /* * QUIC Front-End I/O API: Steady-State Operations * =============================================== * * Here we dispatch calls to the steady-state front-end I/O API functions; that * is, the functions used during the established phase of a QUIC connection * (e.g. SSL_read, SSL_write). * * Each function must handle both blocking and non-blocking modes. As discussed * above, all QUIC I/O is implemented using non-blocking mode internally. * * SSL_get_error => partially implemented by ossl_quic_get_error * SSL_want => ossl_quic_want * (BIO/)SSL_read => ossl_quic_read * (BIO/)SSL_write => ossl_quic_write * SSL_pending => ossl_quic_pending * SSL_stream_conclude => ossl_quic_conn_stream_conclude * SSL_key_update => ossl_quic_key_update */ /* SSL_get_error */ int ossl_quic_get_error(const SSL *s, int i) { QCTX ctx; int net_error, last_error; /* SSL_get_errors() should not raise new errors */ if (!is_quic_cs(s, &ctx, 0 /* suppress errors */)) return SSL_ERROR_SSL; qctx_lock(&ctx); net_error = ossl_quic_channel_net_error(ctx.qc->ch); last_error = ctx.is_stream ? ctx.xso->last_error : ctx.qc->last_error; qctx_unlock(&ctx); if (net_error) return SSL_ERROR_SYSCALL; return last_error; } /* Converts a code returned by SSL_get_error to a code returned by SSL_want. */ static int error_to_want(int error) { switch (error) { case SSL_ERROR_WANT_CONNECT: /* never used - UDP is connectionless */ case SSL_ERROR_WANT_ACCEPT: /* never used - UDP is connectionless */ case SSL_ERROR_ZERO_RETURN: default: return SSL_NOTHING; case SSL_ERROR_WANT_READ: return SSL_READING; case SSL_ERROR_WANT_WRITE: return SSL_WRITING; case SSL_ERROR_WANT_RETRY_VERIFY: return SSL_RETRY_VERIFY; case SSL_ERROR_WANT_CLIENT_HELLO_CB: return SSL_CLIENT_HELLO_CB; case SSL_ERROR_WANT_X509_LOOKUP: return SSL_X509_LOOKUP; } } /* SSL_want */ int ossl_quic_want(const SSL *s) { QCTX ctx; int w; if (!expect_quic_cs(s, &ctx)) return SSL_NOTHING; qctx_lock(&ctx); w = error_to_want(ctx.is_stream ? ctx.xso->last_error : ctx.qc->last_error); qctx_unlock(&ctx); return w; } /* * SSL_write * --------- * * The set of functions below provide the implementation of the public SSL_write * function. We must handle: * * - both blocking and non-blocking operation at the application level, * depending on how we are configured; * * - SSL_MODE_ENABLE_PARTIAL_WRITE being on or off; * * - SSL_MODE_ACCEPT_MOVING_WRITE_BUFFER. * */ QUIC_NEEDS_LOCK static void quic_post_write(QUIC_XSO *xso, int did_append, int did_append_all, uint64_t flags, int do_tick) { /* * We have appended at least one byte to the stream. * Potentially mark stream as active, depending on FC. */ if (did_append) ossl_quic_stream_map_update_state(ossl_quic_channel_get_qsm(xso->conn->ch), xso->stream); if (did_append_all && (flags & SSL_WRITE_FLAG_CONCLUDE) != 0) ossl_quic_sstream_fin(xso->stream->sstream); /* * Try and send. * * TODO(QUIC FUTURE): It is probably inefficient to try and do this * immediately, plus we should eventually consider Nagle's algorithm. */ if (do_tick) ossl_quic_reactor_tick(ossl_quic_channel_get_reactor(xso->conn->ch), 0); } struct quic_write_again_args { QUIC_XSO *xso; const unsigned char *buf; size_t len; size_t total_written; int err; uint64_t flags; }; /* * Absolute maximum write buffer size, enforced to prevent a rogue peer from * deliberately inducing DoS. This has been chosen based on the optimal buffer * size for an RTT of 500ms and a bandwidth of 100 Mb/s. */ #define MAX_WRITE_BUF_SIZE (6 * 1024 * 1024) /* * Ensure spare buffer space available (up until a limit, at least). */ QUIC_NEEDS_LOCK static int sstream_ensure_spare(QUIC_SSTREAM *sstream, uint64_t spare) { size_t cur_sz = ossl_quic_sstream_get_buffer_size(sstream); size_t avail = ossl_quic_sstream_get_buffer_avail(sstream); size_t spare_ = (spare > SIZE_MAX) ? SIZE_MAX : (size_t)spare; size_t new_sz, growth; if (spare_ <= avail || cur_sz == MAX_WRITE_BUF_SIZE) return 1; growth = spare_ - avail; if (cur_sz + growth > MAX_WRITE_BUF_SIZE) new_sz = MAX_WRITE_BUF_SIZE; else new_sz = cur_sz + growth; return ossl_quic_sstream_set_buffer_size(sstream, new_sz); } /* * Append to a QUIC_STREAM's QUIC_SSTREAM, ensuring buffer space is expanded * as needed according to flow control. */ QUIC_NEEDS_LOCK static int xso_sstream_append(QUIC_XSO *xso, const unsigned char *buf, size_t len, size_t *actual_written) { QUIC_SSTREAM *sstream = xso->stream->sstream; uint64_t cur = ossl_quic_sstream_get_cur_size(sstream); uint64_t cwm = ossl_quic_txfc_get_cwm(&xso->stream->txfc); uint64_t permitted = (cwm >= cur ? cwm - cur : 0); if (len > permitted) len = (size_t)permitted; if (!sstream_ensure_spare(sstream, len)) return 0; return ossl_quic_sstream_append(sstream, buf, len, actual_written); } QUIC_NEEDS_LOCK static int quic_write_again(void *arg) { struct quic_write_again_args *args = arg; size_t actual_written = 0; if (!quic_mutation_allowed(args->xso->conn, /*req_active=*/1)) /* If connection is torn down due to an error while blocking, stop. */ return -2; if (!quic_validate_for_write(args->xso, &args->err)) /* * Stream may have become invalid for write due to connection events * while we blocked. */ return -2; args->err = ERR_R_INTERNAL_ERROR; if (!xso_sstream_append(args->xso, args->buf, args->len, &actual_written)) return -2; quic_post_write(args->xso, actual_written > 0, args->len == actual_written, args->flags, 0); args->buf += actual_written; args->len -= actual_written; args->total_written += actual_written; if (args->len == 0) /* Written everything, done. */ return 1; /* Not written everything yet, keep trying. */ return 0; } QUIC_NEEDS_LOCK static int quic_write_blocking(QCTX *ctx, const void *buf, size_t len, uint64_t flags, size_t *written) { int res; QUIC_XSO *xso = ctx->xso; struct quic_write_again_args args; size_t actual_written = 0; /* First make a best effort to append as much of the data as possible. */ if (!xso_sstream_append(xso, buf, len, &actual_written)) { /* Stream already finished or allocation error. */ *written = 0; return QUIC_RAISE_NON_NORMAL_ERROR(ctx, ERR_R_INTERNAL_ERROR, NULL); } quic_post_write(xso, actual_written > 0, actual_written == len, flags, 1); /* * Record however much data we wrote */ *written = actual_written; if (actual_written == len) { /* Managed to append everything on the first try. */ return 1; } /* * We did not manage to append all of the data immediately, so the stream * buffer has probably filled up. This means we need to block until some of * it is freed up. */ args.xso = xso; args.buf = (const unsigned char *)buf + actual_written; args.len = len - actual_written; args.total_written = 0; args.err = ERR_R_INTERNAL_ERROR; args.flags = flags; res = block_until_pred(ctx, quic_write_again, &args, 0); if (res <= 0) { if (!quic_mutation_allowed(xso->conn, /*req_active=*/1)) return QUIC_RAISE_NON_NORMAL_ERROR(ctx, SSL_R_PROTOCOL_IS_SHUTDOWN, NULL); else return QUIC_RAISE_NON_NORMAL_ERROR(ctx, args.err, NULL); } /* * When waiting on extra buffer space to be available, args.total_written * holds the amount of remaining data we requested to write, which will be * something less than the len parameter passed in, however much we wrote * here, add it to the value that we wrote when we initially called * xso_sstream_append */ *written += args.total_written; return 1; } /* * Functions to manage All-or-Nothing (AON) (that is, non-ENABLE_PARTIAL_WRITE) * write semantics. */ static void aon_write_begin(QUIC_XSO *xso, const unsigned char *buf, size_t buf_len, size_t already_sent) { assert(!xso->aon_write_in_progress); xso->aon_write_in_progress = 1; xso->aon_buf_base = buf; xso->aon_buf_pos = already_sent; xso->aon_buf_len = buf_len; } static void aon_write_finish(QUIC_XSO *xso) { xso->aon_write_in_progress = 0; xso->aon_buf_base = NULL; xso->aon_buf_pos = 0; xso->aon_buf_len = 0; } QUIC_NEEDS_LOCK static int quic_write_nonblocking_aon(QCTX *ctx, const void *buf, size_t len, uint64_t flags, size_t *written) { QUIC_XSO *xso = ctx->xso; const void *actual_buf; size_t actual_len, actual_written = 0; int accept_moving_buffer = ((xso->ssl_mode & SSL_MODE_ACCEPT_MOVING_WRITE_BUFFER) != 0); if (xso->aon_write_in_progress) { /* * We are in the middle of an AON write (i.e., a previous write did not * manage to append all data to the SSTREAM and we have Enable Partial * Write (EPW) mode disabled.) */ if ((!accept_moving_buffer && xso->aon_buf_base != buf) || len != xso->aon_buf_len) /* * Pointer must not have changed if we are not in accept moving * buffer mode. Length must never change. */ return QUIC_RAISE_NON_NORMAL_ERROR(ctx, SSL_R_BAD_WRITE_RETRY, NULL); actual_buf = (unsigned char *)buf + xso->aon_buf_pos; actual_len = len - xso->aon_buf_pos; assert(actual_len > 0); } else { actual_buf = buf; actual_len = len; } /* First make a best effort to append as much of the data as possible. */ if (!xso_sstream_append(xso, actual_buf, actual_len, &actual_written)) { /* Stream already finished or allocation error. */ *written = 0; return QUIC_RAISE_NON_NORMAL_ERROR(ctx, ERR_R_INTERNAL_ERROR, NULL); } quic_post_write(xso, actual_written > 0, actual_written == actual_len, flags, qctx_should_autotick(ctx)); if (actual_written == actual_len) { /* We have sent everything. */ if (xso->aon_write_in_progress) { /* * We have sent everything, and we were in the middle of an AON * write. The output write length is the total length of the AON * buffer, not however many bytes we managed to write to the stream * in this call. */ *written = xso->aon_buf_len; aon_write_finish(xso); } else { *written = actual_written; } return 1; } if (xso->aon_write_in_progress) { /* * AON write is in progress but we have not written everything yet. We * may have managed to send zero bytes, or some number of bytes less * than the total remaining which need to be appended during this * AON operation. */ xso->aon_buf_pos += actual_written; assert(xso->aon_buf_pos < xso->aon_buf_len); return QUIC_RAISE_NORMAL_ERROR(ctx, SSL_ERROR_WANT_WRITE); } /* * Not in an existing AON operation but partial write is not enabled, so we * need to begin a new AON operation. However we needn't bother if we didn't * actually append anything. */ if (actual_written > 0) aon_write_begin(xso, buf, len, actual_written); /* * AON - We do not publicly admit to having appended anything until AON * completes. */ *written = 0; return QUIC_RAISE_NORMAL_ERROR(ctx, SSL_ERROR_WANT_WRITE); } QUIC_NEEDS_LOCK static int quic_write_nonblocking_epw(QCTX *ctx, const void *buf, size_t len, uint64_t flags, size_t *written) { QUIC_XSO *xso = ctx->xso; /* Simple best effort operation. */ if (!xso_sstream_append(xso, buf, len, written)) { /* Stream already finished or allocation error. */ *written = 0; return QUIC_RAISE_NON_NORMAL_ERROR(ctx, ERR_R_INTERNAL_ERROR, NULL); } quic_post_write(xso, *written > 0, *written == len, flags, qctx_should_autotick(ctx)); if (*written == 0) /* SSL_write_ex returns 0 if it didn't write anything. */ return QUIC_RAISE_NORMAL_ERROR(ctx, SSL_ERROR_WANT_WRITE); return 1; } QUIC_NEEDS_LOCK static int quic_validate_for_write(QUIC_XSO *xso, int *err) { QUIC_STREAM_MAP *qsm; if (xso == NULL || xso->stream == NULL) { *err = ERR_R_INTERNAL_ERROR; return 0; } switch (xso->stream->send_state) { default: case QUIC_SSTREAM_STATE_NONE: *err = SSL_R_STREAM_RECV_ONLY; return 0; case QUIC_SSTREAM_STATE_READY: qsm = ossl_quic_channel_get_qsm(xso->conn->ch); if (!ossl_quic_stream_map_ensure_send_part_id(qsm, xso->stream)) { *err = ERR_R_INTERNAL_ERROR; return 0; } /* FALLTHROUGH */ case QUIC_SSTREAM_STATE_SEND: case QUIC_SSTREAM_STATE_DATA_SENT: if (ossl_quic_sstream_get_final_size(xso->stream->sstream, NULL)) { *err = SSL_R_STREAM_FINISHED; return 0; } return 1; case QUIC_SSTREAM_STATE_DATA_RECVD: *err = SSL_R_STREAM_FINISHED; return 0; case QUIC_SSTREAM_STATE_RESET_SENT: case QUIC_SSTREAM_STATE_RESET_RECVD: *err = SSL_R_STREAM_RESET; return 0; } } QUIC_TAKES_LOCK int ossl_quic_write_flags(SSL *s, const void *buf, size_t len, uint64_t flags, size_t *written) { int ret; QCTX ctx; int partial_write, err; *written = 0; if (len == 0) { /* Do not autocreate default XSO for zero-length writes. */ if (!expect_quic_cs(s, &ctx)) return 0; qctx_lock_for_io(&ctx); } else { if (!expect_quic_with_stream_lock(s, /*remote_init=*/0, /*io=*/1, &ctx)) return 0; } partial_write = ((ctx.xso != NULL) ? ((ctx.xso->ssl_mode & SSL_MODE_ENABLE_PARTIAL_WRITE) != 0) : 0); if ((flags & ~SSL_WRITE_FLAG_CONCLUDE) != 0) { ret = QUIC_RAISE_NON_NORMAL_ERROR(&ctx, SSL_R_UNSUPPORTED_WRITE_FLAG, NULL); goto out; } if (!quic_mutation_allowed(ctx.qc, /*req_active=*/0)) { ret = QUIC_RAISE_NON_NORMAL_ERROR(&ctx, SSL_R_PROTOCOL_IS_SHUTDOWN, NULL); goto out; } /* * If we haven't finished the handshake, try to advance it. * We don't accept writes until the handshake is completed. */ if (quic_do_handshake(&ctx) < 1) { ret = 0; goto out; } /* Ensure correct stream state, stream send part not concluded, etc. */ if (len > 0 && !quic_validate_for_write(ctx.xso, &err)) { ret = QUIC_RAISE_NON_NORMAL_ERROR(&ctx, err, NULL); goto out; } if (len == 0) { if ((flags & SSL_WRITE_FLAG_CONCLUDE) != 0) quic_post_write(ctx.xso, 0, 1, flags, qctx_should_autotick(&ctx)); ret = 1; goto out; } if (qctx_blocking(&ctx)) ret = quic_write_blocking(&ctx, buf, len, flags, written); else if (partial_write) ret = quic_write_nonblocking_epw(&ctx, buf, len, flags, written); else ret = quic_write_nonblocking_aon(&ctx, buf, len, flags, written); out: qctx_unlock(&ctx); return ret; } QUIC_TAKES_LOCK int ossl_quic_write(SSL *s, const void *buf, size_t len, size_t *written) { return ossl_quic_write_flags(s, buf, len, 0, written); } /* * SSL_read * -------- */ struct quic_read_again_args { QCTX *ctx; QUIC_STREAM *stream; void *buf; size_t len; size_t *bytes_read; int peek; }; QUIC_NEEDS_LOCK static int quic_validate_for_read(QUIC_XSO *xso, int *err, int *eos) { QUIC_STREAM_MAP *qsm; *eos = 0; if (xso == NULL || xso->stream == NULL) { *err = ERR_R_INTERNAL_ERROR; return 0; } switch (xso->stream->recv_state) { default: case QUIC_RSTREAM_STATE_NONE: *err = SSL_R_STREAM_SEND_ONLY; return 0; case QUIC_RSTREAM_STATE_RECV: case QUIC_RSTREAM_STATE_SIZE_KNOWN: case QUIC_RSTREAM_STATE_DATA_RECVD: return 1; case QUIC_RSTREAM_STATE_DATA_READ: *eos = 1; return 0; case QUIC_RSTREAM_STATE_RESET_RECVD: qsm = ossl_quic_channel_get_qsm(xso->conn->ch); ossl_quic_stream_map_notify_app_read_reset_recv_part(qsm, xso->stream); /* FALLTHROUGH */ case QUIC_RSTREAM_STATE_RESET_READ: *err = SSL_R_STREAM_RESET; return 0; } } QUIC_NEEDS_LOCK static int quic_read_actual(QCTX *ctx, QUIC_STREAM *stream, void *buf, size_t buf_len, size_t *bytes_read, int peek) { int is_fin = 0, err, eos; QUIC_CONNECTION *qc = ctx->qc; if (!quic_validate_for_read(ctx->xso, &err, &eos)) { if (eos) { ctx->xso->retired_fin = 1; return QUIC_RAISE_NORMAL_ERROR(ctx, SSL_ERROR_ZERO_RETURN); } else { return QUIC_RAISE_NON_NORMAL_ERROR(ctx, err, NULL); } } if (peek) { if (!ossl_quic_rstream_peek(stream->rstream, buf, buf_len, bytes_read, &is_fin)) return QUIC_RAISE_NON_NORMAL_ERROR(ctx, ERR_R_INTERNAL_ERROR, NULL); } else { if (!ossl_quic_rstream_read(stream->rstream, buf, buf_len, bytes_read, &is_fin)) return QUIC_RAISE_NON_NORMAL_ERROR(ctx, ERR_R_INTERNAL_ERROR, NULL); } if (!peek) { if (*bytes_read > 0) { /* * We have read at least one byte from the stream. Inform stream-level * RXFC of the retirement of controlled bytes. Update the active stream * status (the RXFC may now want to emit a frame granting more credit to * the peer). */ OSSL_RTT_INFO rtt_info; ossl_statm_get_rtt_info(ossl_quic_channel_get_statm(qc->ch), &rtt_info); if (!ossl_quic_rxfc_on_retire(&stream->rxfc, *bytes_read, rtt_info.smoothed_rtt)) return QUIC_RAISE_NON_NORMAL_ERROR(ctx, ERR_R_INTERNAL_ERROR, NULL); } if (is_fin && !peek) { QUIC_STREAM_MAP *qsm = ossl_quic_channel_get_qsm(ctx->qc->ch); ossl_quic_stream_map_notify_totally_read(qsm, ctx->xso->stream); } if (*bytes_read > 0) ossl_quic_stream_map_update_state(ossl_quic_channel_get_qsm(qc->ch), stream); } if (*bytes_read == 0 && is_fin) { ctx->xso->retired_fin = 1; return QUIC_RAISE_NORMAL_ERROR(ctx, SSL_ERROR_ZERO_RETURN); } return 1; } QUIC_NEEDS_LOCK static int quic_read_again(void *arg) { struct quic_read_again_args *args = arg; if (!quic_mutation_allowed(args->ctx->qc, /*req_active=*/1)) { /* If connection is torn down due to an error while blocking, stop. */ QUIC_RAISE_NON_NORMAL_ERROR(args->ctx, SSL_R_PROTOCOL_IS_SHUTDOWN, NULL); return -1; } if (!quic_read_actual(args->ctx, args->stream, args->buf, args->len, args->bytes_read, args->peek)) return -1; if (*args->bytes_read > 0) /* got at least one byte, the SSL_read op can finish now */ return 1; return 0; /* did not read anything, keep trying */ } QUIC_TAKES_LOCK static int quic_read(SSL *s, void *buf, size_t len, size_t *bytes_read, int peek) { int ret, res; QCTX ctx; struct quic_read_again_args args; *bytes_read = 0; if (!expect_quic_cs(s, &ctx)) return 0; qctx_lock_for_io(&ctx); /* If we haven't finished the handshake, try to advance it. */ if (quic_do_handshake(&ctx) < 1) { ret = 0; /* ossl_quic_do_handshake raised error here */ goto out; } if (ctx.xso == NULL) { /* * Called on a QCSO and we don't currently have a default stream. * * Wait until we get a stream initiated by the peer (blocking mode) or * fail if we don't have one yet (non-blocking mode). */ if (!qc_wait_for_default_xso_for_read(&ctx, /*peek=*/0)) { ret = 0; /* error already raised here */ goto out; } ctx.xso = ctx.qc->default_xso; } if (!quic_read_actual(&ctx, ctx.xso->stream, buf, len, bytes_read, peek)) { ret = 0; /* quic_read_actual raised error here */ goto out; } if (*bytes_read > 0) { /* * Even though we succeeded, tick the reactor here to ensure we are * handling other aspects of the QUIC connection. */ if (quic_mutation_allowed(ctx.qc, /*req_active=*/0)) qctx_maybe_autotick(&ctx); ret = 1; } else if (!quic_mutation_allowed(ctx.qc, /*req_active=*/0)) { ret = QUIC_RAISE_NON_NORMAL_ERROR(&ctx, SSL_R_PROTOCOL_IS_SHUTDOWN, NULL); goto out; } else if (qctx_blocking(&ctx)) { /* * We were not able to read anything immediately, so our stream * buffer is empty. This means we need to block until we get * at least one byte. */ args.ctx = &ctx; args.stream = ctx.xso->stream; args.buf = buf; args.len = len; args.bytes_read = bytes_read; args.peek = peek; res = block_until_pred(&ctx, quic_read_again, &args, 0); if (res == 0) { ret = QUIC_RAISE_NON_NORMAL_ERROR(&ctx, ERR_R_INTERNAL_ERROR, NULL); goto out; } else if (res < 0) { ret = 0; /* quic_read_again raised error here */ goto out; } ret = 1; } else { /* * We did not get any bytes and are not in blocking mode. * Tick to see if this delivers any more. */ qctx_maybe_autotick(&ctx); /* Try the read again. */ if (!quic_read_actual(&ctx, ctx.xso->stream, buf, len, bytes_read, peek)) { ret = 0; /* quic_read_actual raised error here */ goto out; } if (*bytes_read > 0) ret = 1; /* Succeeded this time. */ else ret = QUIC_RAISE_NORMAL_ERROR(&ctx, SSL_ERROR_WANT_READ); } out: qctx_unlock(&ctx); return ret; } int ossl_quic_read(SSL *s, void *buf, size_t len, size_t *bytes_read) { return quic_read(s, buf, len, bytes_read, 0); } int ossl_quic_peek(SSL *s, void *buf, size_t len, size_t *bytes_read) { return quic_read(s, buf, len, bytes_read, 1); } /* * SSL_pending * ----------- */ QUIC_TAKES_LOCK static size_t ossl_quic_pending_int(const SSL *s, int check_channel) { QCTX ctx; size_t avail = 0; if (!expect_quic_cs(s, &ctx)) return 0; qctx_lock(&ctx); if (!ctx.qc->started) goto out; if (ctx.xso == NULL) { /* No XSO yet, but there might be a default XSO eligible to be created. */ if (qc_wait_for_default_xso_for_read(&ctx, /*peek=*/1)) { ctx.xso = ctx.qc->default_xso; } else { QUIC_RAISE_NON_NORMAL_ERROR(&ctx, SSL_R_NO_STREAM, NULL); goto out; } } if (ctx.xso->stream == NULL) { QUIC_RAISE_NON_NORMAL_ERROR(&ctx, ERR_R_INTERNAL_ERROR, NULL); goto out; } if (check_channel) avail = ossl_quic_stream_recv_pending(ctx.xso->stream, /*include_fin=*/1) || ossl_quic_channel_has_pending(ctx.qc->ch) || ossl_quic_channel_is_term_any(ctx.qc->ch); else avail = ossl_quic_stream_recv_pending(ctx.xso->stream, /*include_fin=*/0); out: qctx_unlock(&ctx); return avail; } size_t ossl_quic_pending(const SSL *s) { return ossl_quic_pending_int(s, /*check_channel=*/0); } int ossl_quic_has_pending(const SSL *s) { /* Do we have app-side pending data or pending URXEs or RXEs? */ return ossl_quic_pending_int(s, /*check_channel=*/1) > 0; } /* * SSL_stream_conclude * ------------------- */ QUIC_TAKES_LOCK int ossl_quic_conn_stream_conclude(SSL *s) { QCTX ctx; QUIC_STREAM *qs; int err; if (!expect_quic_with_stream_lock(s, /*remote_init=*/0, /*io=*/0, &ctx)) return 0; qs = ctx.xso->stream; if (!quic_mutation_allowed(ctx.qc, /*req_active=*/1)) { qctx_unlock(&ctx); return QUIC_RAISE_NON_NORMAL_ERROR(&ctx, SSL_R_PROTOCOL_IS_SHUTDOWN, NULL); } if (!quic_validate_for_write(ctx.xso, &err)) { qctx_unlock(&ctx); return QUIC_RAISE_NON_NORMAL_ERROR(&ctx, err, NULL); } if (ossl_quic_sstream_get_final_size(qs->sstream, NULL)) { qctx_unlock(&ctx); return 1; } ossl_quic_sstream_fin(qs->sstream); quic_post_write(ctx.xso, 1, 0, 0, qctx_should_autotick(&ctx)); qctx_unlock(&ctx); return 1; } /* * SSL_inject_net_dgram * -------------------- */ QUIC_TAKES_LOCK int SSL_inject_net_dgram(SSL *s, const unsigned char *buf, size_t buf_len, const BIO_ADDR *peer, const BIO_ADDR *local) { int ret = 0; QCTX ctx; QUIC_DEMUX *demux; QUIC_PORT *port; if (!expect_quic_csl(s, &ctx)) return 0; qctx_lock(&ctx); port = ossl_quic_obj_get0_port(ctx.obj); if (port == NULL) { QUIC_RAISE_NON_NORMAL_ERROR(&ctx, ERR_R_UNSUPPORTED, NULL); goto err; } demux = ossl_quic_port_get0_demux(port); ret = ossl_quic_demux_inject(demux, buf, buf_len, peer, local); err: qctx_unlock(&ctx); return ret; } /* * SSL_get0_connection * ------------------- */ SSL *ossl_quic_get0_connection(SSL *s) { QCTX ctx; if (!expect_quic_cs(s, &ctx)) return NULL; return &ctx.qc->obj.ssl; } /* * SSL_get0_listener * ----------------- */ SSL *ossl_quic_get0_listener(SSL *s) { QCTX ctx; if (!expect_quic_csl(s, &ctx)) return NULL; return ctx.ql != NULL ? &ctx.ql->obj.ssl : NULL; } /* * SSL_get0_domain * --------------- */ SSL *ossl_quic_get0_domain(SSL *s) { QCTX ctx; if (!expect_quic_any(s, &ctx)) return NULL; return ctx.qd != NULL ? &ctx.qd->obj.ssl : NULL; } /* * SSL_get_domain_flags * -------------------- */ int ossl_quic_get_domain_flags(const SSL *ssl, uint64_t *domain_flags) { QCTX ctx; if (!expect_quic_any(ssl, &ctx)) return 0; if (domain_flags != NULL) *domain_flags = ctx.obj->domain_flags; return 1; } /* * SSL_get_stream_type * ------------------- */ int ossl_quic_get_stream_type(SSL *s) { QCTX ctx; if (!expect_quic_cs(s, &ctx)) return SSL_STREAM_TYPE_BIDI; if (ctx.xso == NULL) { /* * If deferred XSO creation has yet to occur, proceed according to the * default stream mode. If AUTO_BIDI or AUTO_UNI is set, we cannot know * what kind of stream will be created yet, so return BIDI on the basis * that at this time, the client still has the option of calling * SSL_read() or SSL_write() first. */ if (ctx.qc->default_xso_created || ctx.qc->default_stream_mode == SSL_DEFAULT_STREAM_MODE_NONE) return SSL_STREAM_TYPE_NONE; else return SSL_STREAM_TYPE_BIDI; } if (ossl_quic_stream_is_bidi(ctx.xso->stream)) return SSL_STREAM_TYPE_BIDI; if (ossl_quic_stream_is_server_init(ctx.xso->stream) != ctx.qc->as_server) return SSL_STREAM_TYPE_READ; else return SSL_STREAM_TYPE_WRITE; } /* * SSL_get_stream_id * ----------------- */ QUIC_TAKES_LOCK uint64_t ossl_quic_get_stream_id(SSL *s) { QCTX ctx; uint64_t id; if (!expect_quic_with_stream_lock(s, /*remote_init=*/-1, /*io=*/0, &ctx)) return UINT64_MAX; id = ctx.xso->stream->id; qctx_unlock(&ctx); return id; } /* * SSL_is_stream_local * ------------------- */ QUIC_TAKES_LOCK int ossl_quic_is_stream_local(SSL *s) { QCTX ctx; int is_local; if (!expect_quic_with_stream_lock(s, /*remote_init=*/-1, /*io=*/0, &ctx)) return -1; is_local = ossl_quic_stream_is_local_init(ctx.xso->stream); qctx_unlock(&ctx); return is_local; } /* * SSL_set_default_stream_mode * --------------------------- */ QUIC_TAKES_LOCK int ossl_quic_set_default_stream_mode(SSL *s, uint32_t mode) { QCTX ctx; if (!expect_quic_conn_only(s, &ctx)) return 0; qctx_lock(&ctx); if (ctx.qc->default_xso_created) { qctx_unlock(&ctx); return QUIC_RAISE_NON_NORMAL_ERROR(&ctx, ERR_R_SHOULD_NOT_HAVE_BEEN_CALLED, "too late to change default stream mode"); } switch (mode) { case SSL_DEFAULT_STREAM_MODE_NONE: case SSL_DEFAULT_STREAM_MODE_AUTO_BIDI: case SSL_DEFAULT_STREAM_MODE_AUTO_UNI: ctx.qc->default_stream_mode = mode; break; default: qctx_unlock(&ctx); return QUIC_RAISE_NON_NORMAL_ERROR(&ctx, ERR_R_PASSED_INVALID_ARGUMENT, "bad default stream type"); } qctx_unlock(&ctx); return 1; } /* * SSL_detach_stream * ----------------- */ QUIC_TAKES_LOCK SSL *ossl_quic_detach_stream(SSL *s) { QCTX ctx; QUIC_XSO *xso = NULL; if (!expect_quic_conn_only(s, &ctx)) return NULL; qctx_lock(&ctx); /* Calling this function inhibits default XSO autocreation. */ /* QC ref to any default XSO is transferred to us and to caller. */ qc_set_default_xso_keep_ref(ctx.qc, NULL, /*touch=*/1, &xso); qctx_unlock(&ctx); return xso != NULL ? &xso->obj.ssl : NULL; } /* * SSL_attach_stream * ----------------- */ QUIC_TAKES_LOCK int ossl_quic_attach_stream(SSL *conn, SSL *stream) { QCTX ctx; QUIC_XSO *xso; int nref; if (!expect_quic_conn_only(conn, &ctx)) return 0; if (stream == NULL || stream->type != SSL_TYPE_QUIC_XSO) return QUIC_RAISE_NON_NORMAL_ERROR(&ctx, ERR_R_PASSED_NULL_PARAMETER, "stream to attach must be a valid QUIC stream"); xso = (QUIC_XSO *)stream; qctx_lock(&ctx); if (ctx.qc->default_xso != NULL) { qctx_unlock(&ctx); return QUIC_RAISE_NON_NORMAL_ERROR(&ctx, ERR_R_SHOULD_NOT_HAVE_BEEN_CALLED, "connection already has a default stream"); } /* * It is a caller error for the XSO being attached as a default XSO to have * more than one ref. */ if (!CRYPTO_GET_REF(&xso->obj.ssl.references, &nref)) { qctx_unlock(&ctx); return QUIC_RAISE_NON_NORMAL_ERROR(&ctx, ERR_R_INTERNAL_ERROR, "ref"); } if (nref != 1) { qctx_unlock(&ctx); return QUIC_RAISE_NON_NORMAL_ERROR(&ctx, ERR_R_PASSED_INVALID_ARGUMENT, "stream being attached must have " "only 1 reference"); } /* Caller's reference to the XSO is transferred to us. */ /* Calling this function inhibits default XSO autocreation. */ qc_set_default_xso(ctx.qc, xso, /*touch=*/1); qctx_unlock(&ctx); return 1; } /* * SSL_set_incoming_stream_policy * ------------------------------ */ QUIC_NEEDS_LOCK static int qc_get_effective_incoming_stream_policy(QUIC_CONNECTION *qc) { switch (qc->incoming_stream_policy) { case SSL_INCOMING_STREAM_POLICY_AUTO: if ((qc->default_xso == NULL && !qc->default_xso_created) || qc->default_stream_mode == SSL_DEFAULT_STREAM_MODE_NONE) return SSL_INCOMING_STREAM_POLICY_ACCEPT; else return SSL_INCOMING_STREAM_POLICY_REJECT; default: return qc->incoming_stream_policy; } } QUIC_NEEDS_LOCK static void qc_update_reject_policy(QUIC_CONNECTION *qc) { int policy = qc_get_effective_incoming_stream_policy(qc); int enable_reject = (policy == SSL_INCOMING_STREAM_POLICY_REJECT); ossl_quic_channel_set_incoming_stream_auto_reject(qc->ch, enable_reject, qc->incoming_stream_aec); } QUIC_TAKES_LOCK int ossl_quic_set_incoming_stream_policy(SSL *s, int policy, uint64_t aec) { int ret = 1; QCTX ctx; if (!expect_quic_conn_only(s, &ctx)) return 0; qctx_lock(&ctx); switch (policy) { case SSL_INCOMING_STREAM_POLICY_AUTO: case SSL_INCOMING_STREAM_POLICY_ACCEPT: case SSL_INCOMING_STREAM_POLICY_REJECT: ctx.qc->incoming_stream_policy = policy; ctx.qc->incoming_stream_aec = aec; break; default: QUIC_RAISE_NON_NORMAL_ERROR(&ctx, ERR_R_PASSED_INVALID_ARGUMENT, NULL); ret = 0; break; } qc_update_reject_policy(ctx.qc); qctx_unlock(&ctx); return ret; } /* * SSL_get_value, SSL_set_value * ---------------------------- */ QUIC_TAKES_LOCK static int qc_getset_idle_timeout(QCTX *ctx, uint32_t class_, uint64_t *p_value_out, uint64_t *p_value_in) { int ret = 0; uint64_t value_out = 0, value_in; qctx_lock(ctx); switch (class_) { case SSL_VALUE_CLASS_FEATURE_REQUEST: value_out = ossl_quic_channel_get_max_idle_timeout_request(ctx->qc->ch); if (p_value_in != NULL) { value_in = *p_value_in; if (value_in > OSSL_QUIC_VLINT_MAX) { QUIC_RAISE_NON_NORMAL_ERROR(ctx, ERR_R_PASSED_INVALID_ARGUMENT, NULL); goto err; } if (ossl_quic_channel_have_generated_transport_params(ctx->qc->ch)) { QUIC_RAISE_NON_NORMAL_ERROR(ctx, SSL_R_FEATURE_NOT_RENEGOTIABLE, NULL); goto err; } ossl_quic_channel_set_max_idle_timeout_request(ctx->qc->ch, value_in); } break; case SSL_VALUE_CLASS_FEATURE_PEER_REQUEST: case SSL_VALUE_CLASS_FEATURE_NEGOTIATED: if (p_value_in != NULL) { QUIC_RAISE_NON_NORMAL_ERROR(ctx, SSL_R_UNSUPPORTED_CONFIG_VALUE_OP, NULL); goto err; } if (!ossl_quic_channel_is_handshake_complete(ctx->qc->ch)) { QUIC_RAISE_NON_NORMAL_ERROR(ctx, SSL_R_FEATURE_NEGOTIATION_NOT_COMPLETE, NULL); goto err; } value_out = (class_ == SSL_VALUE_CLASS_FEATURE_NEGOTIATED) ? ossl_quic_channel_get_max_idle_timeout_actual(ctx->qc->ch) : ossl_quic_channel_get_max_idle_timeout_peer_request(ctx->qc->ch); break; default: QUIC_RAISE_NON_NORMAL_ERROR(ctx, SSL_R_UNSUPPORTED_CONFIG_VALUE_CLASS, NULL); goto err; } ret = 1; err: qctx_unlock(ctx); if (ret && p_value_out != NULL) *p_value_out = value_out; return ret; } QUIC_TAKES_LOCK static int qc_get_stream_avail(QCTX *ctx, uint32_t class_, int is_uni, int is_remote, uint64_t *value) { int ret = 0; if (class_ != SSL_VALUE_CLASS_GENERIC) { QUIC_RAISE_NON_NORMAL_ERROR(ctx, SSL_R_UNSUPPORTED_CONFIG_VALUE_CLASS, NULL); return 0; } qctx_lock(ctx); *value = is_remote ? ossl_quic_channel_get_remote_stream_count_avail(ctx->qc->ch, is_uni) : ossl_quic_channel_get_local_stream_count_avail(ctx->qc->ch, is_uni); ret = 1; qctx_unlock(ctx); return ret; } QUIC_NEEDS_LOCK static int qctx_should_autotick(QCTX *ctx) { int event_handling_mode; QUIC_OBJ *obj = ctx->obj; for (; (event_handling_mode = obj->event_handling_mode) == SSL_VALUE_EVENT_HANDLING_MODE_INHERIT && obj->parent_obj != NULL; obj = obj->parent_obj); return event_handling_mode != SSL_VALUE_EVENT_HANDLING_MODE_EXPLICIT; } QUIC_NEEDS_LOCK static void qctx_maybe_autotick(QCTX *ctx) { if (!qctx_should_autotick(ctx)) return; ossl_quic_reactor_tick(ossl_quic_obj_get0_reactor(ctx->obj), 0); } QUIC_TAKES_LOCK static int qc_getset_event_handling(QCTX *ctx, uint32_t class_, uint64_t *p_value_out, uint64_t *p_value_in) { int ret = 0; uint64_t value_out = 0; qctx_lock(ctx); if (class_ != SSL_VALUE_CLASS_GENERIC) { QUIC_RAISE_NON_NORMAL_ERROR(ctx, SSL_R_UNSUPPORTED_CONFIG_VALUE_CLASS, NULL); goto err; } if (p_value_in != NULL) { switch (*p_value_in) { case SSL_VALUE_EVENT_HANDLING_MODE_INHERIT: case SSL_VALUE_EVENT_HANDLING_MODE_IMPLICIT: case SSL_VALUE_EVENT_HANDLING_MODE_EXPLICIT: break; default: QUIC_RAISE_NON_NORMAL_ERROR(ctx, ERR_R_PASSED_INVALID_ARGUMENT, NULL); goto err; } value_out = *p_value_in; ctx->obj->event_handling_mode = (int)value_out; } else { value_out = ctx->obj->event_handling_mode; } ret = 1; err: qctx_unlock(ctx); if (ret && p_value_out != NULL) *p_value_out = value_out; return ret; } QUIC_TAKES_LOCK static int qc_get_stream_write_buf_stat(QCTX *ctx, uint32_t class_, uint64_t *p_value_out, size_t (*getter)(QUIC_SSTREAM *sstream)) { int ret = 0; size_t value = 0; qctx_lock(ctx); if (class_ != SSL_VALUE_CLASS_GENERIC) { QUIC_RAISE_NON_NORMAL_ERROR(ctx, SSL_R_UNSUPPORTED_CONFIG_VALUE_CLASS, NULL); goto err; } if (ctx->xso == NULL) { QUIC_RAISE_NON_NORMAL_ERROR(ctx, SSL_R_NO_STREAM, NULL); goto err; } if (!ossl_quic_stream_has_send(ctx->xso->stream)) { QUIC_RAISE_NON_NORMAL_ERROR(ctx, SSL_R_STREAM_RECV_ONLY, NULL); goto err; } if (ossl_quic_stream_has_send_buffer(ctx->xso->stream)) value = getter(ctx->xso->stream->sstream); ret = 1; err: qctx_unlock(ctx); *p_value_out = (uint64_t)value; return ret; } QUIC_NEEDS_LOCK static int expect_quic_for_value(SSL *s, QCTX *ctx, uint32_t id) { switch (id) { case SSL_VALUE_EVENT_HANDLING_MODE: case SSL_VALUE_STREAM_WRITE_BUF_SIZE: case SSL_VALUE_STREAM_WRITE_BUF_USED: case SSL_VALUE_STREAM_WRITE_BUF_AVAIL: return expect_quic_cs(s, ctx); default: return expect_quic_conn_only(s, ctx); } } QUIC_TAKES_LOCK int ossl_quic_get_value_uint(SSL *s, uint32_t class_, uint32_t id, uint64_t *value) { QCTX ctx; if (!expect_quic_for_value(s, &ctx, id)) return 0; if (value == NULL) return QUIC_RAISE_NON_NORMAL_ERROR(&ctx, ERR_R_PASSED_INVALID_ARGUMENT, NULL); switch (id) { case SSL_VALUE_QUIC_IDLE_TIMEOUT: return qc_getset_idle_timeout(&ctx, class_, value, NULL); case SSL_VALUE_QUIC_STREAM_BIDI_LOCAL_AVAIL: return qc_get_stream_avail(&ctx, class_, /*uni=*/0, /*remote=*/0, value); case SSL_VALUE_QUIC_STREAM_BIDI_REMOTE_AVAIL: return qc_get_stream_avail(&ctx, class_, /*uni=*/0, /*remote=*/1, value); case SSL_VALUE_QUIC_STREAM_UNI_LOCAL_AVAIL: return qc_get_stream_avail(&ctx, class_, /*uni=*/1, /*remote=*/0, value); case SSL_VALUE_QUIC_STREAM_UNI_REMOTE_AVAIL: return qc_get_stream_avail(&ctx, class_, /*uni=*/1, /*remote=*/1, value); case SSL_VALUE_EVENT_HANDLING_MODE: return qc_getset_event_handling(&ctx, class_, value, NULL); case SSL_VALUE_STREAM_WRITE_BUF_SIZE: return qc_get_stream_write_buf_stat(&ctx, class_, value, ossl_quic_sstream_get_buffer_size); case SSL_VALUE_STREAM_WRITE_BUF_USED: return qc_get_stream_write_buf_stat(&ctx, class_, value, ossl_quic_sstream_get_buffer_used); case SSL_VALUE_STREAM_WRITE_BUF_AVAIL: return qc_get_stream_write_buf_stat(&ctx, class_, value, ossl_quic_sstream_get_buffer_avail); default: return QUIC_RAISE_NON_NORMAL_ERROR(&ctx, SSL_R_UNSUPPORTED_CONFIG_VALUE, NULL); } return 1; } QUIC_TAKES_LOCK int ossl_quic_set_value_uint(SSL *s, uint32_t class_, uint32_t id, uint64_t value) { QCTX ctx; if (!expect_quic_for_value(s, &ctx, id)) return 0; switch (id) { case SSL_VALUE_QUIC_IDLE_TIMEOUT: return qc_getset_idle_timeout(&ctx, class_, NULL, &value); case SSL_VALUE_EVENT_HANDLING_MODE: return qc_getset_event_handling(&ctx, class_, NULL, &value); default: return QUIC_RAISE_NON_NORMAL_ERROR(&ctx, SSL_R_UNSUPPORTED_CONFIG_VALUE, NULL); } return 1; } /* * SSL_accept_stream * ----------------- */ struct wait_for_incoming_stream_args { QCTX *ctx; QUIC_STREAM *qs; }; QUIC_NEEDS_LOCK static int wait_for_incoming_stream(void *arg) { struct wait_for_incoming_stream_args *args = arg; QUIC_CONNECTION *qc = args->ctx->qc; QUIC_STREAM_MAP *qsm = ossl_quic_channel_get_qsm(qc->ch); if (!quic_mutation_allowed(qc, /*req_active=*/1)) { /* If connection is torn down due to an error while blocking, stop. */ QUIC_RAISE_NON_NORMAL_ERROR(args->ctx, SSL_R_PROTOCOL_IS_SHUTDOWN, NULL); return -1; } args->qs = ossl_quic_stream_map_peek_accept_queue(qsm); if (args->qs != NULL) return 1; /* got a stream */ return 0; /* did not get a stream, keep trying */ } QUIC_TAKES_LOCK SSL *ossl_quic_accept_stream(SSL *s, uint64_t flags) { QCTX ctx; int ret; SSL *new_s = NULL; QUIC_STREAM_MAP *qsm; QUIC_STREAM *qs; QUIC_XSO *xso; OSSL_RTT_INFO rtt_info; if (!expect_quic_conn_only(s, &ctx)) return NULL; qctx_lock(&ctx); if (qc_get_effective_incoming_stream_policy(ctx.qc) == SSL_INCOMING_STREAM_POLICY_REJECT) { QUIC_RAISE_NON_NORMAL_ERROR(&ctx, ERR_R_SHOULD_NOT_HAVE_BEEN_CALLED, NULL); goto out; } qsm = ossl_quic_channel_get_qsm(ctx.qc->ch); qs = ossl_quic_stream_map_peek_accept_queue(qsm); if (qs == NULL) { if (qctx_blocking(&ctx) && (flags & SSL_ACCEPT_STREAM_NO_BLOCK) == 0) { struct wait_for_incoming_stream_args args; args.ctx = &ctx; args.qs = NULL; ret = block_until_pred(&ctx, wait_for_incoming_stream, &args, 0); if (ret == 0) { QUIC_RAISE_NON_NORMAL_ERROR(&ctx, ERR_R_INTERNAL_ERROR, NULL); goto out; } else if (ret < 0 || args.qs == NULL) { goto out; } qs = args.qs; } else { goto out; } } xso = create_xso_from_stream(ctx.qc, qs); if (xso == NULL) goto out; ossl_statm_get_rtt_info(ossl_quic_channel_get_statm(ctx.qc->ch), &rtt_info); ossl_quic_stream_map_remove_from_accept_queue(qsm, qs, rtt_info.smoothed_rtt); new_s = &xso->obj.ssl; /* Calling this function inhibits default XSO autocreation. */ qc_touch_default_xso(ctx.qc); /* inhibits default XSO */ out: qctx_unlock(&ctx); return new_s; } /* * SSL_get_accept_stream_queue_len * ------------------------------- */ QUIC_TAKES_LOCK size_t ossl_quic_get_accept_stream_queue_len(SSL *s) { QCTX ctx; size_t v; if (!expect_quic_conn_only(s, &ctx)) return 0; qctx_lock(&ctx); v = ossl_quic_stream_map_get_total_accept_queue_len(ossl_quic_channel_get_qsm(ctx.qc->ch)); qctx_unlock(&ctx); return v; } /* * SSL_stream_reset * ---------------- */ int ossl_quic_stream_reset(SSL *ssl, const SSL_STREAM_RESET_ARGS *args, size_t args_len) { QCTX ctx; QUIC_STREAM_MAP *qsm; QUIC_STREAM *qs; uint64_t error_code; int ok, err; if (!expect_quic_with_stream_lock(ssl, /*remote_init=*/0, /*io=*/0, &ctx)) return 0; qsm = ossl_quic_channel_get_qsm(ctx.qc->ch); qs = ctx.xso->stream; error_code = (args != NULL ? args->quic_error_code : 0); if (!quic_validate_for_write(ctx.xso, &err)) { ok = QUIC_RAISE_NON_NORMAL_ERROR(&ctx, err, NULL); goto err; } ok = ossl_quic_stream_map_reset_stream_send_part(qsm, qs, error_code); if (ok) ctx.xso->requested_reset = 1; err: qctx_unlock(&ctx); return ok; } /* * SSL_get_stream_read_state * ------------------------- */ static void quic_classify_stream(QUIC_CONNECTION *qc, QUIC_STREAM *qs, int is_write, int *state, uint64_t *app_error_code) { int local_init; uint64_t final_size; local_init = (ossl_quic_stream_is_server_init(qs) == qc->as_server); if (app_error_code != NULL) *app_error_code = UINT64_MAX; else app_error_code = &final_size; /* throw away value */ if (!ossl_quic_stream_is_bidi(qs) && local_init != is_write) { /* * Unidirectional stream and this direction of transmission doesn't * exist. */ *state = SSL_STREAM_STATE_WRONG_DIR; } else if (ossl_quic_channel_is_term_any(qc->ch)) { /* Connection already closed. */ *state = SSL_STREAM_STATE_CONN_CLOSED; } else if (!is_write && qs->recv_state == QUIC_RSTREAM_STATE_DATA_READ) { /* Application has read a FIN. */ *state = SSL_STREAM_STATE_FINISHED; } else if ((!is_write && qs->stop_sending) || (is_write && ossl_quic_stream_send_is_reset(qs))) { /* * Stream has been reset locally. FIN takes precedence over this for the * read case as the application need not care if the stream is reset * after a FIN has been successfully processed. */ *state = SSL_STREAM_STATE_RESET_LOCAL; *app_error_code = !is_write ? qs->stop_sending_aec : qs->reset_stream_aec; } else if ((!is_write && ossl_quic_stream_recv_is_reset(qs)) || (is_write && qs->peer_stop_sending)) { /* * Stream has been reset remotely. */ *state = SSL_STREAM_STATE_RESET_REMOTE; *app_error_code = !is_write ? qs->peer_reset_stream_aec : qs->peer_stop_sending_aec; } else if (is_write && ossl_quic_sstream_get_final_size(qs->sstream, &final_size)) { /* * Stream has been finished. Stream reset takes precedence over this for * the write case as peer may not have received all data. */ *state = SSL_STREAM_STATE_FINISHED; } else { /* Stream still healthy. */ *state = SSL_STREAM_STATE_OK; } } static int quic_get_stream_state(SSL *ssl, int is_write) { QCTX ctx; int state; if (!expect_quic_with_stream_lock(ssl, /*remote_init=*/-1, /*io=*/0, &ctx)) return SSL_STREAM_STATE_NONE; quic_classify_stream(ctx.qc, ctx.xso->stream, is_write, &state, NULL); qctx_unlock(&ctx); return state; } int ossl_quic_get_stream_read_state(SSL *ssl) { return quic_get_stream_state(ssl, /*is_write=*/0); } /* * SSL_get_stream_write_state * -------------------------- */ int ossl_quic_get_stream_write_state(SSL *ssl) { return quic_get_stream_state(ssl, /*is_write=*/1); } /* * SSL_get_stream_read_error_code * ------------------------------ */ static int quic_get_stream_error_code(SSL *ssl, int is_write, uint64_t *app_error_code) { QCTX ctx; int state; if (!expect_quic_with_stream_lock(ssl, /*remote_init=*/-1, /*io=*/0, &ctx)) return -1; quic_classify_stream(ctx.qc, ctx.xso->stream, /*is_write=*/0, &state, app_error_code); qctx_unlock(&ctx); switch (state) { case SSL_STREAM_STATE_FINISHED: return 0; case SSL_STREAM_STATE_RESET_LOCAL: case SSL_STREAM_STATE_RESET_REMOTE: return 1; default: return -1; } } int ossl_quic_get_stream_read_error_code(SSL *ssl, uint64_t *app_error_code) { return quic_get_stream_error_code(ssl, /*is_write=*/0, app_error_code); } /* * SSL_get_stream_write_error_code * ------------------------------- */ int ossl_quic_get_stream_write_error_code(SSL *ssl, uint64_t *app_error_code) { return quic_get_stream_error_code(ssl, /*is_write=*/1, app_error_code); } /* * Write buffer size mutation * -------------------------- */ int ossl_quic_set_write_buffer_size(SSL *ssl, size_t size) { int ret = 0; QCTX ctx; if (!expect_quic_with_stream_lock(ssl, /*remote_init=*/-1, /*io=*/0, &ctx)) return 0; if (!ossl_quic_stream_has_send(ctx.xso->stream)) { /* Called on a unidirectional receive-only stream - error. */ QUIC_RAISE_NON_NORMAL_ERROR(&ctx, ERR_R_SHOULD_NOT_HAVE_BEEN_CALLED, NULL); goto out; } if (!ossl_quic_stream_has_send_buffer(ctx.xso->stream)) { /* * If the stream has a send part but we have disposed of it because we * no longer need it, this is a no-op. */ ret = 1; goto out; } if (!ossl_quic_sstream_set_buffer_size(ctx.xso->stream->sstream, size)) { QUIC_RAISE_NON_NORMAL_ERROR(&ctx, ERR_R_INTERNAL_ERROR, NULL); goto out; } ret = 1; out: qctx_unlock(&ctx); return ret; } /* * SSL_get_conn_close_info * ----------------------- */ int ossl_quic_get_conn_close_info(SSL *ssl, SSL_CONN_CLOSE_INFO *info, size_t info_len) { QCTX ctx; const QUIC_TERMINATE_CAUSE *tc; if (!expect_quic_conn_only(ssl, &ctx)) return -1; tc = ossl_quic_channel_get_terminate_cause(ctx.qc->ch); if (tc == NULL) return 0; info->error_code = tc->error_code; info->frame_type = tc->frame_type; info->reason = tc->reason; info->reason_len = tc->reason_len; info->flags = 0; if (!tc->remote) info->flags |= SSL_CONN_CLOSE_FLAG_LOCAL; if (!tc->app) info->flags |= SSL_CONN_CLOSE_FLAG_TRANSPORT; return 1; } /* * SSL_key_update * -------------- */ int ossl_quic_key_update(SSL *ssl, int update_type) { QCTX ctx; if (!expect_quic_conn_only(ssl, &ctx)) return 0; switch (update_type) { case SSL_KEY_UPDATE_NOT_REQUESTED: /* * QUIC signals peer key update implicily by triggering a local * spontaneous TXKU. Silently upgrade this to SSL_KEY_UPDATE_REQUESTED. */ case SSL_KEY_UPDATE_REQUESTED: break; default: QUIC_RAISE_NON_NORMAL_ERROR(&ctx, ERR_R_PASSED_INVALID_ARGUMENT, NULL); return 0; } qctx_lock(&ctx); /* Attempt to perform a TXKU. */ if (!ossl_quic_channel_trigger_txku(ctx.qc->ch)) { QUIC_RAISE_NON_NORMAL_ERROR(&ctx, SSL_R_TOO_MANY_KEY_UPDATES, NULL); qctx_unlock(&ctx); return 0; } qctx_unlock(&ctx); return 1; } /* * SSL_get_key_update_type * ----------------------- */ int ossl_quic_get_key_update_type(const SSL *s) { /* * We always handle key updates immediately so a key update is never * pending. */ return SSL_KEY_UPDATE_NONE; } /** * @brief Allocates an SSL object for a user from a QUIC channel. * * This function creates a new QUIC_CONNECTION object based on an incoming * connection associated with the provided QUIC_LISTENER. If the connection * creation fails, the function returns NULL. Otherwise, it returns a pointer * to the SSL object associated with the newly created connection. * * Note: This function is a registered port callback made from * ossl_quic_new_listener and ossl_quic_new_listener_from, and allows for * pre-allocation of the user_ssl object when a channel is created, rather than * when it is accepted * * @param ch Pointer to the QUIC_CHANNEL representing the incoming connection. * @param arg Pointer to a QUIC_LISTENER used to create the connection. * * @return Pointer to the SSL object on success, or NULL on failure. */ static SSL *alloc_port_user_ssl(QUIC_CHANNEL *ch, void *arg) { QUIC_LISTENER *ql = arg; QUIC_CONNECTION *qc = create_qc_from_incoming_conn(ql, ch); return (qc == NULL) ? NULL : &qc->obj.ssl; } /* * QUIC Front-End I/O API: Listeners * ================================= */ /* * SSL_new_listener * ---------------- */ SSL *ossl_quic_new_listener(SSL_CTX *ctx, uint64_t flags) { QUIC_LISTENER *ql = NULL; QUIC_ENGINE_ARGS engine_args = {0}; QUIC_PORT_ARGS port_args = {0}; if ((ql = OPENSSL_zalloc(sizeof(*ql))) == NULL) { QUIC_RAISE_NON_NORMAL_ERROR(NULL, ERR_R_CRYPTO_LIB, NULL); goto err; } #if defined(OPENSSL_THREADS) if ((ql->mutex = ossl_crypto_mutex_new()) == NULL) { QUIC_RAISE_NON_NORMAL_ERROR(NULL, ERR_R_CRYPTO_LIB, NULL); goto err; } #endif engine_args.libctx = ctx->libctx; engine_args.propq = ctx->propq; #if defined(OPENSSL_THREADS) engine_args.mutex = ql->mutex; #endif if (need_notifier_for_domain_flags(ctx->domain_flags)) engine_args.reactor_flags |= QUIC_REACTOR_FLAG_USE_NOTIFIER; if ((ql->engine = ossl_quic_engine_new(&engine_args)) == NULL) { QUIC_RAISE_NON_NORMAL_ERROR(NULL, ERR_R_INTERNAL_ERROR, NULL); goto err; } port_args.channel_ctx = ctx; port_args.is_multi_conn = 1; port_args.get_conn_user_ssl = alloc_port_user_ssl; port_args.user_ssl_arg = ql; if ((flags & SSL_LISTENER_FLAG_NO_VALIDATE) == 0) port_args.do_addr_validation = 1; ql->port = ossl_quic_engine_create_port(ql->engine, &port_args); if (ql->port == NULL) { QUIC_RAISE_NON_NORMAL_ERROR(NULL, ERR_R_INTERNAL_ERROR, NULL); goto err; } /* TODO(QUIC FUTURE): Implement SSL_LISTENER_FLAG_NO_ACCEPT */ ossl_quic_port_set_allow_incoming(ql->port, 1); /* Initialise the QUIC_LISTENER's object header. */ if (!ossl_quic_obj_init(&ql->obj, ctx, SSL_TYPE_QUIC_LISTENER, NULL, ql->engine, ql->port)) goto err; return &ql->obj.ssl; err: if (ql != NULL) ossl_quic_engine_free(ql->engine); #if defined(OPENSSL_THREADS) ossl_crypto_mutex_free(&ql->mutex); #endif OPENSSL_free(ql); return NULL; } /* * SSL_new_listener_from * --------------------- */ SSL *ossl_quic_new_listener_from(SSL *ssl, uint64_t flags) { QCTX ctx; QUIC_LISTENER *ql = NULL; QUIC_PORT_ARGS port_args = {0}; if (!expect_quic_domain(ssl, &ctx)) return NULL; if (!SSL_up_ref(&ctx.qd->obj.ssl)) return NULL; qctx_lock(&ctx); if ((ql = OPENSSL_zalloc(sizeof(*ql))) == NULL) { QUIC_RAISE_NON_NORMAL_ERROR(NULL, ERR_R_CRYPTO_LIB, NULL); goto err; } port_args.channel_ctx = ssl->ctx; port_args.is_multi_conn = 1; port_args.get_conn_user_ssl = alloc_port_user_ssl; port_args.user_ssl_arg = ql; if ((flags & SSL_LISTENER_FLAG_NO_VALIDATE) == 0) port_args.do_addr_validation = 1; ql->port = ossl_quic_engine_create_port(ctx.qd->engine, &port_args); if (ql->port == NULL) { QUIC_RAISE_NON_NORMAL_ERROR(NULL, ERR_R_INTERNAL_ERROR, NULL); goto err; } ql->domain = ctx.qd; ql->engine = ctx.qd->engine; #if defined(OPENSSL_THREADS) ql->mutex = ctx.qd->mutex; #endif /* * TODO(QUIC FUTURE): Implement SSL_LISTENER_FLAG_NO_ACCEPT * Given that we have apis to create client SSL objects from * server SSL objects (see SSL_new_from_listener), we have aspirations * to enable a flag that allows for the creation of the latter, but not * be used to do accept any connections. This is a placeholder for the * implementation of that flag */ ossl_quic_port_set_allow_incoming(ql->port, 1); /* Initialise the QUIC_LISTENER's object header. */ if (!ossl_quic_obj_init(&ql->obj, ssl->ctx, SSL_TYPE_QUIC_LISTENER, &ctx.qd->obj.ssl, NULL, ql->port)) goto err; qctx_unlock(&ctx); return &ql->obj.ssl; err: if (ql != NULL) ossl_quic_port_free(ql->port); OPENSSL_free(ql); qctx_unlock(&ctx); SSL_free(&ctx.qd->obj.ssl); return NULL; } /* * SSL_new_from_listener * --------------------- * code here is derived from ossl_quic_new(). The `ssl` argument is * a listener object which already comes with QUIC port/engine. The newly * created QUIC connection object (QCSO) is going to share the port/engine * with listener (`ssl`). The `ssl` also becomes a parent of QCSO created * by this function. The caller uses QCSO instance to connect to * remote QUIC server. * * The QCSO created here requires us to also create a channel so we * can connect to remote server. */ SSL *ossl_quic_new_from_listener(SSL *ssl, uint64_t flags) { QCTX ctx; QUIC_CONNECTION *qc = NULL; QUIC_LISTENER *ql; SSL_CONNECTION *sc = NULL; if (flags != 0) return NULL; if (!expect_quic_listener(ssl, &ctx)) return NULL; if (!SSL_up_ref(&ctx.ql->obj.ssl)) return NULL; qctx_lock(&ctx); ql = ctx.ql; /* * listeners (server) contexts don't typically * allocate a token cache because they don't need * to store them, but here we are using a server side * ctx as a client, so we should allocate one now */ if (ssl->ctx->tokencache == NULL) if ((ssl->ctx->tokencache = ossl_quic_new_token_store()) == NULL) goto err; if ((qc = OPENSSL_zalloc(sizeof(*qc))) == NULL) { QUIC_RAISE_NON_NORMAL_ERROR(NULL, ERR_R_CRYPTO_LIB, NULL); goto err; } /* * NOTE: setting a listener here is needed so `qc_cleanup()` does the right * thing. Setting listener to ql avoids premature destruction of port in * qc_cleanup() */ qc->listener = ql; qc->engine = ql->engine; qc->port = ql->port; /* create channel */ #if defined(OPENSSL_THREADS) /* this is the engine mutex */ qc->mutex = ql->mutex; #endif #if !defined(OPENSSL_NO_QUIC_THREAD_ASSIST) qc->is_thread_assisted = ((ql->obj.domain_flags & SSL_DOMAIN_FLAG_THREAD_ASSISTED) != 0); #endif /* Create the handshake layer. */ qc->tls = ossl_ssl_connection_new_int(ql->obj.ssl.ctx, NULL, TLS_method()); if (qc->tls == NULL || (sc = SSL_CONNECTION_FROM_SSL(qc->tls)) == NULL) { QUIC_RAISE_NON_NORMAL_ERROR(NULL, ERR_R_INTERNAL_ERROR, NULL); goto err; } sc->s3.flags |= TLS1_FLAGS_QUIC | TLS1_FLAGS_QUIC_INTERNAL; qc->default_ssl_options = OSSL_QUIC_PERMITTED_OPTIONS; qc->last_error = SSL_ERROR_NONE; /* * This is QCSO, we don't expect to accept connections * on success the channel assumes ownership of tls, we need * to grab reference for qc. */ qc->ch = ossl_quic_port_create_outgoing(qc->port, qc->tls); ossl_quic_channel_set_msg_callback(qc->ch, ql->obj.ssl.ctx->msg_callback, &qc->obj.ssl); ossl_quic_channel_set_msg_callback_arg(qc->ch, ql->obj.ssl.ctx->msg_callback_arg); /* * We deliberately pass NULL for engine and port, because we don't want to * to turn QCSO we create here into an event leader, nor port leader. * Both those roles are occupied already by listener (`ssl`) we use * to create a new QCSO here. */ if (!ossl_quic_obj_init(&qc->obj, ql->obj.ssl.ctx, SSL_TYPE_QUIC_CONNECTION, &ql->obj.ssl, NULL, NULL)) { QUIC_RAISE_NON_NORMAL_ERROR(NULL, ERR_R_INTERNAL_ERROR, NULL); goto err; } /* Initialise libssl APL-related state. */ qc->default_stream_mode = SSL_DEFAULT_STREAM_MODE_AUTO_BIDI; qc->default_ssl_mode = qc->obj.ssl.ctx->mode; qc->default_ssl_options = qc->obj.ssl.ctx->options & OSSL_QUIC_PERMITTED_OPTIONS; qc->incoming_stream_policy = SSL_INCOMING_STREAM_POLICY_AUTO; qc->last_error = SSL_ERROR_NONE; qc_update_reject_policy(qc); qctx_unlock(&ctx); return &qc->obj.ssl; err: if (qc != NULL) { qc_cleanup(qc, /* have_lock= */ 0); OPENSSL_free(qc); } qctx_unlock(&ctx); SSL_free(&ctx.ql->obj.ssl); return NULL; } /* * SSL_listen * ---------- */ QUIC_NEEDS_LOCK static int ql_listen(QUIC_LISTENER *ql) { if (ql->listening) return 1; ossl_quic_port_set_allow_incoming(ql->port, 1); ql->listening = 1; return 1; } QUIC_TAKES_LOCK int ossl_quic_listen(SSL *ssl) { QCTX ctx; int ret; if (!expect_quic_listener(ssl, &ctx)) return 0; qctx_lock_for_io(&ctx); ret = ql_listen(ctx.ql); qctx_unlock(&ctx); return ret; } /* * SSL_accept_connection * --------------------- */ static int quic_accept_connection_wait(void *arg) { QUIC_PORT *port = arg; if (!ossl_quic_port_is_running(port)) return -1; if (ossl_quic_port_have_incoming(port)) return 1; return 0; } QUIC_TAKES_LOCK SSL *ossl_quic_accept_connection(SSL *ssl, uint64_t flags) { int ret; QCTX ctx; SSL *conn_ssl = NULL; SSL_CONNECTION *conn = NULL; QUIC_CHANNEL *new_ch = NULL; QUIC_CONNECTION *qc; int no_block = ((flags & SSL_ACCEPT_CONNECTION_NO_BLOCK) != 0); if (!expect_quic_listener(ssl, &ctx)) return NULL; qctx_lock_for_io(&ctx); if (!ql_listen(ctx.ql)) goto out; /* Wait for an incoming connection if needed. */ new_ch = ossl_quic_port_pop_incoming(ctx.ql->port); if (new_ch == NULL && ossl_quic_port_is_running(ctx.ql->port)) { if (!no_block && qctx_blocking(&ctx)) { ret = block_until_pred(&ctx, quic_accept_connection_wait, ctx.ql->port, 0); if (ret < 1) goto out; } else { qctx_maybe_autotick(&ctx); } if (!ossl_quic_port_is_running(ctx.ql->port)) goto out; new_ch = ossl_quic_port_pop_incoming(ctx.ql->port); } if (new_ch == NULL && ossl_quic_port_is_running(ctx.ql->port)) { /* No connections already queued. */ ossl_quic_reactor_tick(ossl_quic_engine_get0_reactor(ctx.ql->engine), 0); new_ch = ossl_quic_port_pop_incoming(ctx.ql->port); } /* * port_make_channel pre-allocates our user_ssl for us for each newly * created channel, so once we pop the new channel from the port above * we just need to extract it */ if (new_ch == NULL || (conn_ssl = ossl_quic_channel_get0_tls(new_ch)) == NULL || (conn = SSL_CONNECTION_FROM_SSL(conn_ssl)) == NULL || (conn_ssl = SSL_CONNECTION_GET_USER_SSL(conn)) == NULL) goto out; qc = (QUIC_CONNECTION *)conn_ssl; qc->listener = ctx.ql; qc->pending = 0; if (!SSL_up_ref(&ctx.ql->obj.ssl)) { SSL_free(conn_ssl); SSL_free(ossl_quic_channel_get0_tls(new_ch)); conn_ssl = NULL; } out: qctx_unlock(&ctx); return conn_ssl; } static QUIC_CONNECTION *create_qc_from_incoming_conn(QUIC_LISTENER *ql, QUIC_CHANNEL *ch) { QUIC_CONNECTION *qc = NULL; if ((qc = OPENSSL_zalloc(sizeof(*qc))) == NULL) { QUIC_RAISE_NON_NORMAL_ERROR(NULL, ERR_R_CRYPTO_LIB, NULL); goto err; } if (!ossl_quic_obj_init(&qc->obj, ql->obj.ssl.ctx, SSL_TYPE_QUIC_CONNECTION, &ql->obj.ssl, NULL, NULL)) { QUIC_RAISE_NON_NORMAL_ERROR(NULL, ERR_R_INTERNAL_ERROR, NULL); goto err; } ossl_quic_channel_get_peer_addr(ch, &qc->init_peer_addr); /* best effort */ qc->pending = 1; qc->engine = ql->engine; qc->port = ql->port; qc->ch = ch; #if defined(OPENSSL_THREADS) qc->mutex = ql->mutex; #endif qc->tls = ossl_quic_channel_get0_tls(ch); qc->started = 1; qc->as_server = 1; qc->as_server_state = 1; qc->default_stream_mode = SSL_DEFAULT_STREAM_MODE_AUTO_BIDI; qc->default_ssl_options = ql->obj.ssl.ctx->options & OSSL_QUIC_PERMITTED_OPTIONS; qc->incoming_stream_policy = SSL_INCOMING_STREAM_POLICY_AUTO; qc->last_error = SSL_ERROR_NONE; qc_update_reject_policy(qc); return qc; err: OPENSSL_free(qc); return NULL; } DEFINE_LHASH_OF_EX(QUIC_TOKEN); struct ssl_token_store_st { LHASH_OF(QUIC_TOKEN) *cache; CRYPTO_REF_COUNT references; CRYPTO_MUTEX *mutex; }; static unsigned long quic_token_hash(const QUIC_TOKEN *item) { return (unsigned long)ossl_fnv1a_hash(item->hashkey, item->hashkey_len); } static int quic_token_cmp(const QUIC_TOKEN *a, const QUIC_TOKEN *b) { if (a->hashkey_len != b->hashkey_len) return 1; return memcmp(a->hashkey, b->hashkey, a->hashkey_len); } SSL_TOKEN_STORE *ossl_quic_new_token_store(void) { int ok = 0; SSL_TOKEN_STORE *newcache = OPENSSL_zalloc(sizeof(SSL_TOKEN_STORE)); if (newcache == NULL) goto out; newcache->cache = lh_QUIC_TOKEN_new(quic_token_hash, quic_token_cmp); if (newcache->cache == NULL) goto out; #if defined(OPENSSL_THREADS) if ((newcache->mutex = ossl_crypto_mutex_new()) == NULL) goto out; #endif if (!CRYPTO_NEW_REF(&newcache->references, 1)) goto out; ok = 1; out: if (!ok) { ossl_quic_free_token_store(newcache); newcache = NULL; } return newcache; } static void free_this_token(QUIC_TOKEN *tok) { ossl_quic_free_peer_token(tok); } void ossl_quic_free_token_store(SSL_TOKEN_STORE *hdl) { int refs; if (hdl == NULL) return; if (!CRYPTO_DOWN_REF(&hdl->references, &refs)) return; if (refs > 0) return; /* last reference, we can clean up */ ossl_crypto_mutex_free(&hdl->mutex); lh_QUIC_TOKEN_doall(hdl->cache, free_this_token); lh_QUIC_TOKEN_free(hdl->cache); OPENSSL_free(hdl); return; } /** * @brief build a new QUIC_TOKEN * * This function creates a new token storage structure for saving in our * tokencache * * In an effort to make allocation and freeing of these tokens a bit faster * We do them in a single allocation in this format * +---------------+ --\ * | hashkey * |---| | * | hashkey_len | | | QUIC_TOKEN * | token * |---|--| | * | token_len | | | | * +---------------+<--| | --/ * | hashkey buf | | * | | | * |---------------|<-----| * | token buf | * | | * +---------------+ * * @param peer - the peer address that sent the token * @param token - the buffer holding the token * @param token_len - the size of token * * @returns a QUIC_TOKEN pointer or NULL on error */ static QUIC_TOKEN *ossl_quic_build_new_token(BIO_ADDR *peer, uint8_t *token, size_t token_len) { QUIC_TOKEN *new_token; size_t hashkey_len = 0; size_t addr_len = 0; int family; unsigned short port; int *famptr; unsigned short *portptr; uint8_t *addrptr; if ((token != NULL && token_len == 0) || (token == NULL && token_len != 0)) return NULL; if (!BIO_ADDR_rawaddress(peer, NULL, &addr_len)) return NULL; family = BIO_ADDR_family(peer); port = BIO_ADDR_rawport(peer); hashkey_len += sizeof(int); /* hashkey(family) */ hashkey_len += sizeof(unsigned short); /* hashkey(port) */ hashkey_len += addr_len; /* hashkey(address) */ new_token = OPENSSL_zalloc(sizeof(QUIC_TOKEN) + hashkey_len + token_len); if (new_token == NULL) return NULL; if (!CRYPTO_NEW_REF(&new_token->references, 1)) { OPENSSL_free(new_token); return NULL; } new_token->hashkey_len = hashkey_len; /* hashkey is allocated inline, immediately after the QUIC_TOKEN struct */ new_token->hashkey = (uint8_t *)(new_token + 1); /* token buffer follows the hashkey in the inline allocation */ new_token->token = new_token->hashkey + hashkey_len; new_token->token_len = token_len; famptr = (int *)new_token->hashkey; portptr = (unsigned short *)(famptr + 1); addrptr = (uint8_t *)(portptr + 1); *famptr = family; *portptr = port; if (!BIO_ADDR_rawaddress(peer, addrptr, NULL)) { ossl_quic_free_peer_token(new_token); return NULL; } if (token != NULL) memcpy(new_token->token, token, token_len); return new_token; } int ossl_quic_set_peer_token(SSL_CTX *ctx, BIO_ADDR *peer, const uint8_t *token, size_t token_len) { SSL_TOKEN_STORE *c = ctx->tokencache; QUIC_TOKEN *tok, *old = NULL; if (ctx->tokencache == NULL) return 0; tok = ossl_quic_build_new_token(peer, (uint8_t *)token, token_len); if (tok == NULL) return 0; /* we might be sharing this cache, lock it */ ossl_crypto_mutex_lock(c->mutex); old = lh_QUIC_TOKEN_retrieve(c->cache, tok); if (old != NULL) { lh_QUIC_TOKEN_delete(c->cache, old); ossl_quic_free_peer_token(old); } lh_QUIC_TOKEN_insert(c->cache, tok); ossl_crypto_mutex_unlock(c->mutex); return 1; } int ossl_quic_get_peer_token(SSL_CTX *ctx, BIO_ADDR *peer, QUIC_TOKEN **token) { SSL_TOKEN_STORE *c = ctx->tokencache; QUIC_TOKEN *key = NULL; QUIC_TOKEN *tok = NULL; int ret; int rc = 0; if (c == NULL) return 0; key = ossl_quic_build_new_token(peer, NULL, 0); if (key == NULL) return 0; ossl_crypto_mutex_lock(c->mutex); tok = lh_QUIC_TOKEN_retrieve(c->cache, key); if (tok != NULL) { *token = tok; CRYPTO_UP_REF(&tok->references, &ret); rc = 1; } ossl_crypto_mutex_unlock(c->mutex); ossl_quic_free_peer_token(key); return rc; } void ossl_quic_free_peer_token(QUIC_TOKEN *token) { int refs = 0; if (!CRYPTO_DOWN_REF(&token->references, &refs)) return; if (refs > 0) return; CRYPTO_FREE_REF(&token->references); OPENSSL_free(token); } /* * SSL_get_accept_connection_queue_len * ----------------------------------- */ QUIC_TAKES_LOCK size_t ossl_quic_get_accept_connection_queue_len(SSL *ssl) { QCTX ctx; int ret; if (!expect_quic_listener(ssl, &ctx)) return 0; qctx_lock(&ctx); ret = ossl_quic_port_get_num_incoming_channels(ctx.ql->port); qctx_unlock(&ctx); return ret; } /* * QUIC Front-End I/O API: Domains * =============================== */ /* * SSL_new_domain * -------------- */ SSL *ossl_quic_new_domain(SSL_CTX *ctx, uint64_t flags) { QUIC_DOMAIN *qd = NULL; QUIC_ENGINE_ARGS engine_args = {0}; uint64_t domain_flags; domain_flags = ctx->domain_flags; if ((flags & (SSL_DOMAIN_FLAG_SINGLE_THREAD | SSL_DOMAIN_FLAG_MULTI_THREAD | SSL_DOMAIN_FLAG_THREAD_ASSISTED)) != 0) domain_flags = flags; else domain_flags = ctx->domain_flags | flags; if (!ossl_adjust_domain_flags(domain_flags, &domain_flags)) return NULL; if ((qd = OPENSSL_zalloc(sizeof(*qd))) == NULL) { QUIC_RAISE_NON_NORMAL_ERROR(NULL, ERR_R_CRYPTO_LIB, NULL); return NULL; } #if defined(OPENSSL_THREADS) if ((qd->mutex = ossl_crypto_mutex_new()) == NULL) { QUIC_RAISE_NON_NORMAL_ERROR(NULL, ERR_R_CRYPTO_LIB, NULL); goto err; } #endif engine_args.libctx = ctx->libctx; engine_args.propq = ctx->propq; #if defined(OPENSSL_THREADS) engine_args.mutex = qd->mutex; #endif if (need_notifier_for_domain_flags(domain_flags)) engine_args.reactor_flags |= QUIC_REACTOR_FLAG_USE_NOTIFIER; if ((qd->engine = ossl_quic_engine_new(&engine_args)) == NULL) { QUIC_RAISE_NON_NORMAL_ERROR(NULL, ERR_R_INTERNAL_ERROR, NULL); goto err; } /* Initialise the QUIC_DOMAIN's object header. */ if (!ossl_quic_obj_init(&qd->obj, ctx, SSL_TYPE_QUIC_DOMAIN, NULL, qd->engine, NULL)) goto err; ossl_quic_obj_set_domain_flags(&qd->obj, domain_flags); return &qd->obj.ssl; err: ossl_quic_engine_free(qd->engine); #if defined(OPENSSL_THREADS) ossl_crypto_mutex_free(&qd->mutex); #endif OPENSSL_free(qd); return NULL; } /* * QUIC Front-End I/O API: SSL_CTX Management * ========================================== */ long ossl_quic_ctx_ctrl(SSL_CTX *ctx, int cmd, long larg, void *parg) { switch (cmd) { default: return ssl3_ctx_ctrl(ctx, cmd, larg, parg); } } long ossl_quic_callback_ctrl(SSL *s, int cmd, void (*fp) (void)) { QCTX ctx; if (!expect_quic_conn_only(s, &ctx)) return 0; switch (cmd) { case SSL_CTRL_SET_MSG_CALLBACK: ossl_quic_channel_set_msg_callback(ctx.qc->ch, (ossl_msg_cb)fp, &ctx.qc->obj.ssl); /* This callback also needs to be set on the internal SSL object */ return ssl3_callback_ctrl(ctx.qc->tls, cmd, fp);; default: /* Probably a TLS related ctrl. Defer to our internal SSL object */ return ssl3_callback_ctrl(ctx.qc->tls, cmd, fp); } } long ossl_quic_ctx_callback_ctrl(SSL_CTX *ctx, int cmd, void (*fp) (void)) { return ssl3_ctx_callback_ctrl(ctx, cmd, fp); } int ossl_quic_renegotiate_check(SSL *ssl, int initok) { /* We never do renegotiation. */ return 0; } const SSL_CIPHER *ossl_quic_get_cipher_by_char(const unsigned char *p) { const SSL_CIPHER *ciph = ssl3_get_cipher_by_char(p); if ((ciph->algorithm2 & SSL_QUIC) == 0) return NULL; return ciph; } /* * These functions define the TLSv1.2 (and below) ciphers that are supported by * the SSL_METHOD. Since QUIC only supports TLSv1.3 we don't support any. */ int ossl_quic_num_ciphers(void) { return 0; } const SSL_CIPHER *ossl_quic_get_cipher(unsigned int u) { return NULL; } /* * SSL_get_shutdown() * ------------------ */ int ossl_quic_get_shutdown(const SSL *s) { QCTX ctx; int shut = 0; if (!expect_quic_conn_only(s, &ctx)) return 0; if (ossl_quic_channel_is_term_any(ctx.qc->ch)) { shut |= SSL_SENT_SHUTDOWN; if (!ossl_quic_channel_is_closing(ctx.qc->ch)) shut |= SSL_RECEIVED_SHUTDOWN; } return shut; } /* * QUIC Polling Support APIs * ========================= */ /* Do we have the R (read) condition? */ QUIC_NEEDS_LOCK static int test_poll_event_r(QUIC_XSO *xso) { int fin = 0; size_t avail = 0; /* * If a stream has had the fin bit set on the last packet * received, then we need to return a 1 here to raise * SSL_POLL_EVENT_R, so that the stream can have its completion * detected and closed gracefully by an application. * However, if the client reads the data via SSL_read[_ex], that api * provides no stream status, and as a result the stream state moves to * QUIC_RSTREAM_STATE_DATA_READ, and the receive buffer is freed, which * stored the fin state, so its not directly know-able here. Instead * check for the stream state being QUIC_RSTREAM_STATE_DATA_READ, which * is only set if the last stream frame received had the fin bit set, and * the client read the data. This catches our poll/read/poll case */ if (xso->stream->recv_state == QUIC_RSTREAM_STATE_DATA_READ) return 1; return ossl_quic_stream_has_recv_buffer(xso->stream) && ossl_quic_rstream_available(xso->stream->rstream, &avail, &fin) && (avail > 0 || (fin && !xso->retired_fin)); } /* Do we have the ER (exception: read) condition? */ QUIC_NEEDS_LOCK static int test_poll_event_er(QUIC_XSO *xso) { return ossl_quic_stream_has_recv(xso->stream) && ossl_quic_stream_recv_is_reset(xso->stream) && !xso->retired_fin; } /* Do we have the W (write) condition? */ QUIC_NEEDS_LOCK static int test_poll_event_w(QUIC_XSO *xso) { return !xso->conn->shutting_down && ossl_quic_stream_has_send_buffer(xso->stream) && ossl_quic_sstream_get_buffer_avail(xso->stream->sstream) && !ossl_quic_sstream_get_final_size(xso->stream->sstream, NULL) && ossl_quic_txfc_get_cwm(&xso->stream->txfc) > ossl_quic_sstream_get_cur_size(xso->stream->sstream) && quic_mutation_allowed(xso->conn, /*req_active=*/1); } /* Do we have the EW (exception: write) condition? */ QUIC_NEEDS_LOCK static int test_poll_event_ew(QUIC_XSO *xso) { return ossl_quic_stream_has_send(xso->stream) && xso->stream->peer_stop_sending && !xso->requested_reset && !xso->conn->shutting_down; } /* Do we have the EC (exception: connection) condition? */ QUIC_NEEDS_LOCK static int test_poll_event_ec(QUIC_CONNECTION *qc) { return ossl_quic_channel_is_term_any(qc->ch); } /* Do we have the ECD (exception: connection drained) condition? */ QUIC_NEEDS_LOCK static int test_poll_event_ecd(QUIC_CONNECTION *qc) { return ossl_quic_channel_is_terminated(qc->ch); } /* Do we have the IS (incoming: stream) condition? */ QUIC_NEEDS_LOCK static int test_poll_event_is(QUIC_CONNECTION *qc, int is_uni) { return ossl_quic_stream_map_get_accept_queue_len(ossl_quic_channel_get_qsm(qc->ch), is_uni); } /* Do we have the OS (outgoing: stream) condition? */ QUIC_NEEDS_LOCK static int test_poll_event_os(QUIC_CONNECTION *qc, int is_uni) { /* Is it currently possible for us to make an outgoing stream? */ return quic_mutation_allowed(qc, /*req_active=*/1) && ossl_quic_channel_get_local_stream_count_avail(qc->ch, is_uni) > 0; } /* Do we have the EL (exception: listener) condition? */ QUIC_NEEDS_LOCK static int test_poll_event_el(QUIC_LISTENER *ql) { return !ossl_quic_port_is_running(ql->port); } /* Do we have the IC (incoming: connection) condition? */ QUIC_NEEDS_LOCK static int test_poll_event_ic(QUIC_LISTENER *ql) { return ossl_quic_port_get_num_incoming_channels(ql->port) > 0; } QUIC_TAKES_LOCK int ossl_quic_conn_poll_events(SSL *ssl, uint64_t events, int do_tick, uint64_t *p_revents) { QCTX ctx; uint64_t revents = 0; if (!expect_quic_csl(ssl, &ctx)) return 0; qctx_lock(&ctx); if (ctx.qc != NULL && !ctx.qc->started) { /* We can only try to write on non-started connection. */ if ((events & SSL_POLL_EVENT_W) != 0) revents |= SSL_POLL_EVENT_W; goto end; } if (do_tick) ossl_quic_reactor_tick(ossl_quic_obj_get0_reactor(ctx.obj), 0); if (ctx.xso != NULL) { /* SSL object has a stream component. */ if ((events & SSL_POLL_EVENT_R) != 0 && test_poll_event_r(ctx.xso)) revents |= SSL_POLL_EVENT_R; if ((events & SSL_POLL_EVENT_ER) != 0 && test_poll_event_er(ctx.xso)) revents |= SSL_POLL_EVENT_ER; if ((events & SSL_POLL_EVENT_W) != 0 && test_poll_event_w(ctx.xso)) revents |= SSL_POLL_EVENT_W; if ((events & SSL_POLL_EVENT_EW) != 0 && test_poll_event_ew(ctx.xso)) revents |= SSL_POLL_EVENT_EW; } if (ctx.qc != NULL && !ctx.is_stream) { if ((events & SSL_POLL_EVENT_EC) != 0 && test_poll_event_ec(ctx.qc)) revents |= SSL_POLL_EVENT_EC; if ((events & SSL_POLL_EVENT_ECD) != 0 && test_poll_event_ecd(ctx.qc)) revents |= SSL_POLL_EVENT_ECD; if ((events & SSL_POLL_EVENT_ISB) != 0 && test_poll_event_is(ctx.qc, /*uni=*/0)) revents |= SSL_POLL_EVENT_ISB; if ((events & SSL_POLL_EVENT_ISU) != 0 && test_poll_event_is(ctx.qc, /*uni=*/1)) revents |= SSL_POLL_EVENT_ISU; if ((events & SSL_POLL_EVENT_OSB) != 0 && test_poll_event_os(ctx.qc, /*uni=*/0)) revents |= SSL_POLL_EVENT_OSB; if ((events & SSL_POLL_EVENT_OSU) != 0 && test_poll_event_os(ctx.qc, /*uni=*/1)) revents |= SSL_POLL_EVENT_OSU; } if (ctx.is_listener) { if ((events & SSL_POLL_EVENT_EL) != 0 && test_poll_event_el(ctx.ql)) revents |= SSL_POLL_EVENT_EL; if ((events & SSL_POLL_EVENT_IC) != 0 && test_poll_event_ic(ctx.ql)) revents |= SSL_POLL_EVENT_IC; } end: qctx_unlock(&ctx); *p_revents = revents; return 1; } QUIC_TAKES_LOCK int ossl_quic_get_notifier_fd(SSL *ssl) { QCTX ctx; QUIC_REACTOR *rtor; RIO_NOTIFIER *nfy; int nfd = -1; if (!expect_quic_any(ssl, &ctx)) return -1; qctx_lock(&ctx); rtor = ossl_quic_obj_get0_reactor(ctx.obj); nfy = ossl_quic_reactor_get0_notifier(rtor); if (nfy == NULL) goto end; nfd = ossl_rio_notifier_as_fd(nfy); end: qctx_unlock(&ctx); return nfd; } QUIC_TAKES_LOCK void ossl_quic_enter_blocking_section(SSL *ssl, QUIC_REACTOR_WAIT_CTX *wctx) { QCTX ctx; QUIC_REACTOR *rtor; if (!expect_quic_any(ssl, &ctx)) return; qctx_lock(&ctx); rtor = ossl_quic_obj_get0_reactor(ctx.obj); ossl_quic_reactor_wait_ctx_enter(wctx, rtor); qctx_unlock(&ctx); } QUIC_TAKES_LOCK void ossl_quic_leave_blocking_section(SSL *ssl, QUIC_REACTOR_WAIT_CTX *wctx) { QCTX ctx; QUIC_REACTOR *rtor; if (!expect_quic_any(ssl, &ctx)) return; qctx_lock(&ctx); rtor = ossl_quic_obj_get0_reactor(ctx.obj); ossl_quic_reactor_wait_ctx_leave(wctx, rtor); qctx_unlock(&ctx); } /* * Internal Testing APIs * ===================== */ QUIC_CHANNEL *ossl_quic_conn_get_channel(SSL *s) { QCTX ctx; if (!expect_quic_conn_only(s, &ctx)) return NULL; return ctx.qc->ch; } int ossl_quic_set_diag_title(SSL_CTX *ctx, const char *title) { #ifndef OPENSSL_NO_QLOG OPENSSL_free(ctx->qlog_title); ctx->qlog_title = NULL; if (title == NULL) return 1; if ((ctx->qlog_title = OPENSSL_strdup(title)) == NULL) return 0; #endif return 1; }