]> git.ipfire.org Git - thirdparty/haproxy.git/commitdiff
MEDIUM: mux-spop: Introduce the SPOP multiplexer
authorChristopher Faulet <cfaulet@haproxy.com>
Thu, 4 Jul 2024 09:37:23 +0000 (11:37 +0200)
committerChristopher Faulet <cfaulet@haproxy.com>
Fri, 12 Jul 2024 13:27:04 +0000 (15:27 +0200)
It is no possible yet to use it. Idles connections and pipelining mode are
not supported for now. But it should be possible to open a SPOP connection,
perform the HELLO handshake, send a NOTIFY frame based on data produced by
the client side and receive the corresponding ACK frame to transfer its
content to the client side.

The related issue is #2502.

Makefile
dev/flags/flags.c
include/haproxy/mux_spop-t.h [new file with mode: 0644]
include/haproxy/spoe.h
include/haproxy/stconn-t.h
src/flt_spoe.c
src/mux_spop.c [new file with mode: 0644]

index 1178229eb78fc936028e2bf2d1fa5c12705f1e2a..905481281a8d50daba9ea5274331172ead89cecf 100644 (file)
--- a/Makefile
+++ b/Makefile
@@ -983,7 +983,7 @@ OBJS += src/mux_h2.o src/mux_h1.o src/mux_fcgi.o src/stream.o               \
         src/hpack-huff.o src/freq_ctr.o src/dict.o src/wdt.o           \
         src/pipe.o src/init.o src/http_acl.o src/hpack-enc.o           \
         src/ebtree.o src/dgram.o src/hash.o src/version.o              \
-        src/limits.o
+        src/limits.o src/mux_spop.o
 
 ifneq ($(TRACE),)
   OBJS += src/calltrace.o
index 0ad04656b74be1ba42ffdb81ab5955c518aa94b6..4d1bc2c4f7d05e9a9c4df34f17dbbad26746b660 100644 (file)
@@ -12,6 +12,7 @@
 #include <haproxy/mux_fcgi-t.h>
 #include <haproxy/mux_h2-t.h>
 #include <haproxy/mux_h1-t.h>
+#include <haproxy/mux_spop-t.h>
 #include <haproxy/peers-t.h>
 #include <haproxy/quic_conn-t.h>
 #include <haproxy/stconn-t.h>
 #define SHOW_AS_PEERS 0x00080000
 #define SHOW_AS_PEER  0x00100000
 #define SHOW_AS_QC    0x00200000
+#define SHOW_AS_SPOPC 0x00400000
+#define SHOW_AS_SPOPS 0x00800000
 
 // command line names, must be in exact same order as the SHOW_AS_* flags above
 // so that show_as_words[i] matches flag 1U<<i.
 const char *show_as_words[] = { "ana", "chn", "conn", "sc", "stet", "strm", "task", "txn", "sd", "hsl", "htx", "hmsg", "fd", "h2c", "h2s",  "h1c", "h1s", "fconn", "fstrm",
-                               "peers", "peer", "qc"};
+                               "peers", "peer", "qc", "spopc", "spops"};
 
 /* will be sufficient for even largest flag names */
 static char buf[4096];
@@ -161,6 +164,8 @@ int main(int argc, char **argv)
                if (show_as & SHOW_AS_PEERS) printf("peers->flags = %s\n",(peers_show_flags  (buf, bsz, " | ", flags), buf));
                if (show_as & SHOW_AS_PEER)  printf("peer->flags = %s\n", (peer_show_flags   (buf, bsz, " | ", flags), buf));
                if (show_as & SHOW_AS_QC)    printf("qc->flags = %s\n",   (qc_show_flags     (buf, bsz, " | ", flags), buf));
+               if (show_as & SHOW_AS_SPOPC) printf("spopc->flags = %s\n",(spop_conn_show_flags(buf, bsz, " | ", flags), buf));
+               if (show_as & SHOW_AS_SPOPS) printf("spops->flags = %s\n",(spop_strm_show_flags(buf, bsz, " | ", flags), buf));
        }
        return 0;
 }
diff --git a/include/haproxy/mux_spop-t.h b/include/haproxy/mux_spop-t.h
new file mode 100644 (file)
index 0000000..b914b74
--- /dev/null
@@ -0,0 +1,147 @@
+/* SPDX-License-Identifier: LGPL-2.0-or-later */
+
+#ifndef _HAPROXY_MUX_SPOP_T_H
+#define _HAPROXY_MUX_SPOP_T_H
+
+#include <haproxy/api-t.h>
+#include <haproxy/show_flags-t.h>
+
+/**** SPOP connection flags (32 bit), in spop_conn->flags ****/
+#define SPOP_CF_NONE           0x00000000
+
+/* Flags indicating why writing to the mux is blocked */
+#define SPOP_CF_MUX_MALLOC      0x00000001 /* mux is blocked on lack connection's mux buffer */
+#define SPOP_CF_MUX_MFULL       0x00000002 /* mux is blocked on connection's mux buffer full */
+#define SPOP_CF_MUX_BLOCK_ANY   0x00000003 /* mux is blocked on connection's mux buffer full */
+
+#define SPOP_CF_WAIT_INLIST     0x00000010  // there is at least one stream blocked by another stream in send_list
+#define SPOP_CF_DEM_DALLOC      0x00000020  /* demux blocked on lack of connection's demux buffer */
+#define SPOP_CF_DEM_DFULL       0x00000040  /* demux blocked on connection's demux buffer full */
+#define SPOP_CF_DEM_MROOM       0x00000080  /* demux blocked on lack of room in mux buffer */
+#define SPOP_CF_DEM_SALLOC      0x00000100  /* demux blocked on lack of stream's rx buffer */
+#define SPOP_CF_DEM_SFULL       0x00000200  /* demux blocked on stream request buffer full */
+#define SPOP_CF_DEM_TOOMANY     0x00000400  /* demux blocked waiting for some stream connectors to leave */
+#define SPOP_CF_DEM_BLOCK_ANY   0x000007E0  /* aggregate of the demux flags above except DALLOC/DFULL */
+
+/* Other flags */
+#define SPOP_CF_DISCO_SENT      0x00001000  /* a frame DISCONNECT was successfully sent */
+#define SPOP_CF_DISCO_FAILED    0x00002000  /* failed to disconnect */
+#define SPOP_CF_WAIT_FOR_HS     0x00004000  /* We did check that at least a stream was waiting for handshake */
+#define SPOP_CF_DEM_SHORT_READ  0x00008000  // demux blocked on incomplete frame
+/* unused  0x00010000 */
+#define SPOP_CF_RCVD_SHUT       0x00020000  // a recv() attempt already failed on a shutdown
+#define SPOP_CF_END_REACHED     0x00040000  // pending data too short with RCVD_SHUT present
+#define SPOP_CF_ERR_PENDING     0x00080000  /* A write error was detected (block sends but not reads) */
+#define SPOP_CF_ERROR           0x00100000  /* A read error was detected (handled has an abort) */
+
+
+/* This function is used to report flags in debugging tools. Please reflect
+ * below any single-bit flag addition above in the same order via the
+ * __APPEND_FLAG macro. The new end of the buffer is returned.
+ */
+static forceinline char *spop_conn_show_flags(char *buf, size_t len, const char *delim, uint flg)
+{
+#define _(f, ...) __APPEND_FLAG(buf, len, delim, flg, f, #f, __VA_ARGS__)
+       /* prologue */
+       _(0);
+       /* flags */
+       _(SPOP_CF_MUX_MALLOC, _(SPOP_CF_MUX_MFULL, _(SPOP_CF_WAIT_INLIST,
+       _(SPOP_CF_DEM_DALLOC, _(SPOP_CF_DEM_DFULL, _(SPOP_CF_DEM_MROOM,
+       _(SPOP_CF_DEM_SALLOC, _(SPOP_CF_DEM_SFULL, _(SPOP_CF_DEM_TOOMANY,
+       _(SPOP_CF_DISCO_SENT, _(SPOP_CF_DISCO_FAILED, _(SPOP_CF_WAIT_FOR_HS,
+       _(SPOP_CF_DEM_SHORT_READ, _(SPOP_CF_RCVD_SHUT, _(SPOP_CF_END_REACHED,
+       _(SPOP_CF_ERR_PENDING, _(SPOP_CF_ERROR)))))))))))))))));
+       /* epilogue */
+       _(~0U);
+       return buf;
+#undef _
+}
+
+/**** SPOP stream flags (32 bit), in spop_strm->flags ****/
+#define SPOP_SF_NONE           0x00000000
+// #define SPOP_SF_ACK_RCVD       0x00000001 /* ACK freme received */
+//#define SPOP_SF_ES_SENT        0x00000002 /* end-of-stream sent */
+//#define SPOP_SF_EP_SENT        0x00000004 /* end-of-param sent */
+//#define SPOP_SF_DISCON_SENT      0x00000008 /* disconnect sent */
+
+/* Stream flags indicating the reason the stream is blocked */
+#define SPOP_SF_BLK_MBUSY      0x00000010 /* blocked waiting for mux access (transient) */
+#define SPOP_SF_BLK_MROOM      0x00000020 /* blocked waiting for room in the mux */
+#define SPOP_SF_BLK_ANY        0x00000030 /* any of the reasons above */
+
+//#define SPOP_SF_BEGIN_SENT     0x00000100  /* a BEGIN_REQUEST record was sent for this stream */
+//#define SPOP_SF_OUTGOING_DATA  0x00000200  /* set whenever we've seen outgoing data */
+#define SPOP_SF_NOTIFIED       0x00000400  /* a paused stream was notified to try to send again */
+
+/* This function is used to report flags in debugging tools. Please reflect
+ * below any single-bit flag addition above in the same order via the
+ * __APPEND_FLAG macro. The new end of the buffer is returned.
+ */
+static forceinline char *spop_strm_show_flags(char *buf, size_t len, const char *delim, uint flg)
+{
+#define _(f, ...) __APPEND_FLAG(buf, len, delim, flg, f, #f, __VA_ARGS__)
+       /* prologue */
+       _(0);
+       /* flags */
+       _(SPOP_SF_BLK_MBUSY, _(SPOP_SF_BLK_MROOM, _(SPOP_SF_NOTIFIED)));
+       /* epilogue */
+       _(~0U);
+       return buf;
+#undef _
+}
+
+/* SPOP connection state (spop_conn->state) */
+enum spop_conn_st {
+       SPOP_CS_HA_HELLO = 0, /* init done, waiting for sending HELLO frame */
+       SPOP_CS_AGENT_HELLO,     /* HELLO frame sent, waiting for agent HELLO frame to define the connection settings */
+       SPOP_CS_FRAME_H,       /* HELLO handshake finished, waiting for a frame header */
+       SPOP_CS_FRAME_P,       /* Frame header received, waiting for a frame data */
+       SPOP_CS_ERROR,         /* send DISCONNECT frame to be able ti close the conneciton */
+       SPOP_CS_CLOSING,       /* DISCONNECT frame sent, waiting for the agent DISCONNECT frame before closing */
+       SPOP_CS_CLOSED,        /* Agent DISCONNECT frame received and  close the connection ASAP */
+       SPOP_CS_ENTRIES
+} __attribute__((packed));
+
+/* returns a spop_conn state as an abbreviated 3-letter string, or "???" if unknown */
+static inline const char *spop_conn_st_to_str(enum spop_conn_st st)
+{
+       switch (st) {
+               case SPOP_CS_HA_HELLO   : return "HHL";
+               case SPOP_CS_AGENT_HELLO: return "AHL";
+               case SPOP_CS_FRAME_H    : return "FRH";
+               case SPOP_CS_FRAME_P    : return "FRP";
+               case SPOP_CS_ERROR      : return "ERR";
+               case SPOP_CS_CLOSING    : return "CLI";
+               case SPOP_CS_CLOSED     : return "CLO";
+               default                 : return "???";
+       }
+}
+
+/* SPOP stream state, in spop_strm->state */
+enum spop_strm_st {
+       SPOP_SS_IDLE = 0,
+       SPOP_SS_OPEN,
+       SPOP_SS_HREM,     // half-closed(remote)
+       SPOP_SS_HLOC,     // half-closed(local)
+       SPOP_SS_ERROR,
+       SPOP_SS_CLOSED,
+       SPOP_SS_ENTRIES
+} __attribute__((packed));
+
+
+/* returns a spop_strm state as an abbreviated 3-letter string, or "???" if unknown */
+static inline const char *spop_strm_st_to_str(enum spop_strm_st st)
+{
+       switch (st) {
+               case SPOP_SS_IDLE   : return "IDL";
+               case SPOP_SS_OPEN   : return "OPN";
+               case SPOP_SS_HREM   : return "RCL";
+               case SPOP_SS_HLOC   : return "HCL";
+               case SPOP_SS_ERROR  : return "ERR";
+               case SPOP_SS_CLOSED : return "CLO";
+               default             : return "???";
+       }
+}
+
+
+#endif /* _HAPROXY_MUX_SPOP_T_H */
index 55b2c0d50fa9711c1ca08a4861ca6e228fdab67f..16999a6fca716fee136169ec618e15f815a9c149 100644 (file)
@@ -27,6 +27,9 @@
 #include <haproxy/sample-t.h>
 #include <haproxy/spoe-t.h>
 
+struct appctx;
+
+struct spoe_agent *spoe_appctx_agent(struct appctx *appctx);
 
 /* Encode a buffer. Its length <len> is encoded as a varint, followed by a copy
  * of <str>. It must have enough space in <*buf> to encode the buffer, else an
index 244343dc14e78b2698143d09028e833736b0186b..ffd9c4461648be238b6f31ded755a639e6914128 100644 (file)
@@ -280,6 +280,7 @@ struct stconn;
 #define SE_ABRT_SRC_MUX_H2    0x03 /* Code set by the H2 mux */
 #define SE_ABRT_SRC_MUX_QUIC  0x04 /* Code set by the QUIC/H3 mux */
 #define SE_ABRT_SRC_MUX_FCGI  0x05 /* Code set by the FCGI mux */
+#define SE_ABRT_SRC_MUX_SPOP  0x06 /* Code set by the SPOP mux */
 
 struct se_abort_info {
        uint32_t info;
index 0970cc842952037ab9665ead653252e31c49efd1..edc3169b00f1e6beef557948828b7fa8f96a6d6e 100644 (file)
@@ -1099,6 +1099,20 @@ spoe_recv_frame(struct appctx *appctx, char *buf, size_t framesz)
 /********************************************************************
  * Functions that manage the SPOE applet
  ********************************************************************/
+struct spoe_agent *spoe_appctx_agent(struct appctx *appctx)
+{
+       struct spoe_appctx  *spoe_appctx;
+
+       if (!appctx)
+               return NULL;
+
+       spoe_appctx = SPOE_APPCTX(appctx);
+       if (!spoe_appctx)
+               return NULL;
+
+       return spoe_appctx->agent;
+}
+
 static int
 spoe_wakeup_appctx(struct appctx *appctx)
 {
diff --git a/src/mux_spop.c b/src/mux_spop.c
new file mode 100644 (file)
index 0000000..83616d9
--- /dev/null
@@ -0,0 +1,3398 @@
+/* SPDX-License-Identifier: GPL-2.0-or-later */
+
+#include <import/ist.h>
+#include <import/eb32tree.h>
+#include <import/ebmbtree.h>
+
+#include <haproxy/api.h>
+#include <haproxy/cfgparse.h>
+#include <haproxy/connection.h>
+#include <haproxy/dynbuf.h>
+#include <haproxy/list.h>
+#include <haproxy/mux_spop-t.h>
+#include <haproxy/net_helper.h>
+#include <haproxy/proxy.h>
+#include <haproxy/spoe.h>
+#include <haproxy/stconn.h>
+#include <haproxy/task.h>
+#include <haproxy/trace.h>
+
+/* 32 buffers: one for the ring's root, rest for the mbuf itself */
+#define SPOP_C_MBUF_CNT 32
+
+
+/* SPOP connection descriptor */
+struct spop_conn {
+       struct connection *conn;
+
+       enum spop_conn_st state;             /* SPOP connection state */
+       enum spop_error errcode;             /* SPOP error code (SPOP_ERR_*) */
+       uint32_t streams_limit;              /* maximum number of concurrent streams the peer supports */
+       uint32_t max_id;                     /* highest ID known on this connection, <0 before HELLO handshake */
+       uint32_t flags;                      /* Connection flags: SPOP_CF_* */
+
+       uint32_t dsi;                        /* dmux stream ID (<0 = idle ) */
+       uint32_t dfi;                        /* dmux frame ID (if dsi >= 0) */
+       uint32_t dfl;                        /* demux frame length (if dsi >= 0) */
+       uint32_t dff;                        /* demux frame flags */
+       uint8_t  dft;                        /* demux frame type (if dsi >= 0) */
+
+       struct buffer dbuf;                  /* demux buffer */
+       struct buffer mbuf[SPOP_C_MBUF_CNT]; /* mux buffers (ring) */
+
+       int timeout;                         /* idle timeout duration in ticks */
+       int shut_timeout;                    /* idle timeout duration in ticks after shutdown */
+
+       unsigned int nb_streams;             /* number of streams in the tree */
+       unsigned int nb_sc;                  /* number of attached stream connectors */
+       unsigned int nb_reserved;            /* number of reserved streams */
+       unsigned int stream_cnt;             /* total number of streams seen */
+
+       struct proxy *proxy;                 /* the proxy this connection was created for */
+       struct spoe_agent *agent;            /* SPOE agent used by this mux */
+       struct task *task;                   /* timeout management task */
+       struct eb_root streams_by_id;        /* all active streams by their ID */
+
+       struct list send_list;               /* list of blocked streams requesting to send */
+
+       struct buffer_wait buf_wait;         /* Wait list for buffer allocation */
+       struct wait_event wait_event;        /* To be used if we're waiting for I/Os */
+};
+
+/* SPOP stream descriptor */
+struct spop_strm {
+       struct sedesc *sd;
+       struct session *sess;
+       struct spop_conn *spop_conn;
+
+       int32_t id;                   /* stream ID */
+       uint32_t fid;                 /* frame ID */
+
+       uint32_t flags;               /* Connection flags: SPOP_SF_* */
+       enum spop_error errcode;      /* SPOP error code (SPOP_ERR_*) */
+       enum spop_strm_st state;      /* SPOP stream state */
+
+       struct buffer rxbuf;          /* receive buffer, always valid (buf_empty or real buffer) */
+
+       struct eb32_node by_id;       /* place in spop_conn's streams_by_id */
+       struct wait_event *subs;      /* Address of the wait_event the stream connector associated is waiting on */
+       struct list list;        /* To be used when adding in spop_conn->send_list */
+};
+
+/* descriptor for an SPOP frame header */
+struct spop_frame_header {
+        uint32_t len;       /* length, host order */
+        uint32_t flags;     /* frame flags */
+        uint64_t sid;       /* stream id, host order */
+        uint64_t fid;       /* frame id, host order */
+        uint8_t  type;      /* frame type */
+};
+
+/* trace source and events */
+static void spop_trace(enum trace_level level, uint64_t mask,
+                     const struct trace_source *src,
+                     const struct ist where, const struct ist func,
+                     const void *a1, const void *a2, const void *a3, const void *a4);
+
+/* The event representation is split like this :
+ *   spop_conn - internal SPOP connection
+ *   spop_strm - internal SPOP stream
+ *   strm  - application layer
+ *   rx    - data receipt
+ *   tx    - data transmission
+ */
+static const struct trace_event spop_trace_events[] = {
+#define           SPOP_EV_SPOP_CONN_NEW     (1ULL <<  0)
+       { .mask = SPOP_EV_SPOP_CONN_NEW,    .name = "spop_conn_new",  .desc = "new SPOP connection" },
+#define           SPOP_EV_SPOP_CONN_RECV    (1ULL <<  1)
+       { .mask = SPOP_EV_SPOP_CONN_RECV,   .name = "spop_conn_recv", .desc = "Rx on SPOP connection" },
+#define           SPOP_EV_SPOP_CONN_SEND    (1ULL <<  2)
+       { .mask = SPOP_EV_SPOP_CONN_SEND,   .name = "spop_conn_send", .desc = "Tx on SPOP connection" },
+#define           SPOP_EV_SPOP_CONN_BLK     (1ULL <<  3)
+       { .mask = SPOP_EV_SPOP_CONN_BLK,    .name = "spop_conn_blk",  .desc = "SPOP connection blocked" },
+#define           SPOP_EV_SPOP_CONN_WAKE    (1ULL <<  4)
+       { .mask = SPOP_EV_SPOP_CONN_WAKE,   .name = "spop_conn_wake", .desc = "SPOP connection woken up" },
+#define           SPOP_EV_SPOP_CONN_END     (1ULL <<  5)
+       { .mask = SPOP_EV_SPOP_CONN_END,    .name = "spop_conn_end",  .desc = "SPOP connection terminated" },
+#define           SPOP_EV_SPOP_CONN_ERR     (1ULL <<  6)
+       { .mask = SPOP_EV_SPOP_CONN_ERR,    .name = "spop_conn_err",  .desc = "error on SPOP connection" },
+
+#define           SPOP_EV_RX_FHDR           (1ULL <<  7)
+       { .mask = SPOP_EV_RX_FHDR,          .name = "rx_fhdr",        .desc = "SPOP frame header received" },
+#define           SPOP_EV_RX_FRAME          (1ULL <<  8)
+       { .mask = SPOP_EV_RX_FRAME,         .name = "rx_frame",       .desc = "receipt of any SPOP frame" },
+#define           SPOP_EV_RX_EOI            (1ULL <<  9)
+       { .mask = SPOP_EV_RX_EOI,           .name = "rx_eoi",         .desc = "receipt of end of SPOP input" },
+#define           SPOP_EV_RX_HELLO          (1ULL << 10)
+       { .mask = SPOP_EV_RX_HELLO,         .name = "rx_hello",       .desc = "receipt of SPOP AGENT HELLO frame" },
+#define           SPOP_EV_RX_ACK            (1ULL << 11)
+       { .mask = SPOP_EV_RX_ACK,           .name = "rx_ack",         .desc = "receipt of SPOP AGENT ACK frame" },
+#define           SPOP_EV_RX_DISCO          (1ULL << 12)
+       { .mask = SPOP_EV_RX_DISCO,         .name = "rx_disconnect",  .desc = "receipt of SPOP AGENT DISCONNECT frame" },
+
+#define           SPOP_EV_TX_FRAME          (1ULL << 13)
+       { .mask = SPOP_EV_TX_FRAME,         .name = "tx_frame",       .desc = "transmission of any SPOP frame" },
+#define           SPOP_EV_TX_EOI            (1ULL << 14)
+       { .mask = SPOP_EV_TX_EOI,           .name = "tx_eoi",         .desc = "transmission of SPOP end of input" },
+#define           SPOP_EV_TX_HELLO          (1ULL << 15)
+       { .mask = SPOP_EV_TX_HELLO,         .name = "tx_hello",       .desc = "transmission of SPOP HAPROXY HELLO frame" },
+#define           SPOP_EV_TX_NOTIFY         (1ULL << 16)
+       { .mask = SPOP_EV_TX_NOTIFY,        .name = "tx_notify",      .desc = "transmission of SPOP HAPROXY NOTIFY frame" },
+#define           SPOP_EV_TX_DISCO          (1ULL << 17)
+       { .mask = SPOP_EV_TX_DISCO,         .name = "tx_disconnect",  .desc = "transmission of SPOP HAPROXY DISCONNECT frame" },
+
+#define           SPOP_EV_SPOP_STRM_NEW     (1ULL << 18)
+       { .mask = SPOP_EV_SPOP_STRM_NEW,    .name = "spop_strm_new",  .desc = "new SPOP stream" },
+#define           SPOP_EV_SPOP_STRM_BLK     (1ULL << 19)
+       { .mask = SPOP_EV_SPOP_STRM_BLK,    .name = "spop_strm_blk",  .desc = "SPOP stream blocked" },
+#define           SPOP_EV_SPOP_STRM_END     (1ULL << 20)
+       { .mask = SPOP_EV_SPOP_STRM_END,    .name = "spop_strm_end",  .desc = "SPOP stream terminated" },
+#define           SPOP_EV_SPOP_STRM_ERR     (1ULL << 21)
+       { .mask = SPOP_EV_SPOP_STRM_ERR,    .name = "spop_strm_err",  .desc = "error on SPOP stream" },
+
+#define           SPOP_EV_STRM_NEW      (1ULL << 22)
+       { .mask = SPOP_EV_STRM_NEW,     .name = "strm_new",           .desc = "app-layer stream creation" },
+#define           SPOP_EV_STRM_RECV     (1ULL << 23)
+       { .mask = SPOP_EV_STRM_RECV,    .name = "strm_recv",          .desc = "receiving data for stream" },
+#define           SPOP_EV_STRM_SEND     (1ULL << 24)
+       { .mask = SPOP_EV_STRM_SEND,    .name = "strm_send",          .desc = "sending data for stream" },
+#define           SPOP_EV_STRM_FULL     (1ULL << 25)
+       { .mask = SPOP_EV_STRM_FULL,    .name = "strm_full",          .desc = "stream buffer full" },
+#define           SPOP_EV_STRM_WAKE     (1ULL << 26)
+       { .mask = SPOP_EV_STRM_WAKE,    .name = "strm_wake",          .desc = "stream woken up" },
+#define           SPOP_EV_STRM_SHUT     (1ULL << 27)
+       { .mask = SPOP_EV_STRM_SHUT,    .name = "strm_shut",          .desc = "stream shutdown" },
+#define           SPOP_EV_STRM_END      (1ULL << 28)
+       { .mask = SPOP_EV_STRM_END,     .name = "strm_end",           .desc = "detaching app-layer stream" },
+#define           SPOP_EV_STRM_ERR      (1ULL << 29)
+       { .mask = SPOP_EV_STRM_ERR,     .name = "strm_err",           .desc = "stream error" },
+
+       { }
+};
+
+static const struct name_desc spop_trace_lockon_args[4] = {
+       /* arg1 */ { /* already used by the connection */ },
+       /* arg2 */ { .name="spop_strm", .desc="SPOP stream" },
+       /* arg3 */ { },
+       /* arg4 */ { }
+};
+
+
+static const struct name_desc spop_trace_decoding[] = {
+#define SPOP_VERB_CLEAN    1
+       { .name="clean",    .desc="only user-friendly stuff, generally suitable for level \"user\"" },
+#define SPOP_VERB_MINIMAL  2
+       { .name="minimal",  .desc="report only spop_conn/spop_strm state and flags, no real decoding" },
+#define SPOP_VERB_SIMPLE   3
+       { .name="simple",   .desc="add request/response status line or htx info when available" },
+#define SPOP_VERB_ADVANCED 4
+       { .name="advanced", .desc="add header fields or record decoding when available" },
+#define SPOP_VERB_COMPLETE 5
+       { .name="complete", .desc="add full data dump when available" },
+       { /* end */ }
+};
+
+static struct trace_source trace_spop __read_mostly = {
+       .name = IST("spop"),
+       .desc = "SPOP multiplexer",
+       .arg_def = TRC_ARG1_CONN,  // TRACE()'s first argument is always a connection
+       .default_cb = spop_trace,
+       .known_events = spop_trace_events,
+       .lockon_args = spop_trace_lockon_args,
+       .decoding = spop_trace_decoding,
+       .report_events = ~0,  // report everything by default
+};
+
+#define TRACE_SOURCE &trace_spop
+INITCALL1(STG_REGISTER, trace_register_source, TRACE_SOURCE);
+
+/* SPOP connection and stream pools */
+DECLARE_STATIC_POOL(pool_head_spop_conn, "spop_conn", sizeof(struct spop_conn));
+DECLARE_STATIC_POOL(pool_head_spop_strm, "spop_strm", sizeof(struct spop_strm));
+
+
+const char *spop_err_reasons[SPOP_ERR_ENTRIES] = {
+       [SPOP_ERR_NONE]               = "normal",
+       [SPOP_ERR_IO]                 = "I/O error",
+       [SPOP_ERR_TOUT]               = "a timeout occurred",
+       [SPOP_ERR_TOO_BIG]            = "frame is too big",
+       [SPOP_ERR_INVALID]            = "invalid frame received",
+       [SPOP_ERR_NO_VSN]             = "version value not found",
+       [SPOP_ERR_NO_FRAME_SIZE]      = "max-frame-size value not found",
+       [SPOP_ERR_NO_CAP]             = "capabilities value not found",
+       [SPOP_ERR_BAD_VSN]            = "unsupported version",
+       [SPOP_ERR_BAD_FRAME_SIZE]     = "max-frame-size too big or too small",
+       [SPOP_ERR_FRAG_NOT_SUPPORTED] = "fragmentation not supported",
+       [SPOP_ERR_INTERLACED_FRAMES]  = "invalid interlaced frames",
+       [SPOP_ERR_FRAMEID_NOTFOUND]   = "frame-id not found",
+       [SPOP_ERR_RES]                = "resource allocation error",
+       [SPOP_ERR_UNKNOWN]            = "an unknown error occurred",
+};
+
+
+/* Helper to get static string length, excluding the terminating null byte */
+#define SPOP_SLEN(str) (sizeof(str)-1)
+
+/* Predefined key used in HELLO/DISCONNECT frames */
+#define SPOP_SUPPORTED_VERSIONS_KEY     "supported-versions"
+#define SPOP_VERSION_KEY                "version"
+#define SPOP_MAX_FRAME_SIZE_KEY         "max-frame-size"
+#define SPOP_CAPABILITIES_KEY           "capabilities"
+#define SPOP_ENGINE_ID_KEY              "engine-id"
+#define SPOP_HEALTHCHECK_KEY            "healthcheck"
+#define SPOP_STATUS_CODE_KEY            "status-code"
+#define SPOP_MSG_KEY                    "message"
+
+struct spop_version {
+       char *str;
+       int   min;
+       int   max;
+};
+
+/* All supported versions */
+static struct spop_version spop_supported_versions[] = {
+       /* 1.0 is now unsupported because of a bug about frame's flags*/
+       {"2.0", 2000, 2000},
+       {NULL,  0, 0}
+};
+
+/* Comma-separated list of supported versions */
+#define SPOP_SUPPORTED_VERSIONS_VAL  "2.0"
+
+static struct task *spop_timeout_task(struct task *t, void *context, unsigned int state);
+static int spop_process(struct spop_conn *spop_conn);
+static struct task *spop_io_cb(struct task *t, void *ctx, unsigned int state);
+static inline struct spop_strm *spop_conn_st_by_id(struct spop_conn *spop_conn, int id);
+static struct spop_strm *spop_stconn_new(struct spop_conn *spop_conn, struct stconn *sc, struct session *sess);
+static void spop_strm_notify_recv(struct spop_strm *spop_strm);
+static void spop_strm_notify_send(struct spop_strm *spop_strm);
+static void spop_strm_alert(struct spop_strm *spop_strm);
+static inline void spop_remove_from_list(struct spop_strm *spop_strm);
+static inline void spop_conn_restart_reading(const struct spop_conn *spop_conn, int consider_buffer);
+
+/* a dummy closed endpoint */
+static const struct sedesc closed_ep = {
+       .sc        = NULL,
+       .flags     = SE_FL_DETACHED,
+};
+
+/* a dmumy closed stream */
+static const struct spop_strm *spop_closed_stream = &(const struct spop_strm){
+        .sd        = (struct sedesc *)&closed_ep,
+        .spop_conn = NULL,
+        .state     = SPOP_SS_CLOSED,
+       .flags     = SPOP_SF_NONE, // TODO ?
+        .id        = 0,
+};
+
+/* a dummy idle stream for use with any unknown stream */
+static const struct spop_strm *spop_unknown_stream = &(const struct spop_strm){
+       .sd        = (struct sedesc*)&closed_ep,
+       .spop_conn = NULL,
+       .state     = SPOP_SS_IDLE,
+       .flags     = SPOP_SF_NONE,
+       .id        = 0,
+};
+
+/* returns the stconn associated to the stream */
+static forceinline struct stconn *spop_strm_sc(const struct spop_strm *spop_strm)
+{
+       return spop_strm->sd->sc;
+}
+
+static inline struct sedesc *spop_strm_opposite_sd(struct spop_strm *spop_strm)
+{
+       return se_opposite(spop_strm->sd);
+}
+
+static inline void spop_trace_buf(const struct buffer *buf, size_t ofs, size_t len)
+{
+       size_t block1, block2;
+       int line, ptr, newptr;
+
+       block1 = b_contig_data(buf, ofs);
+       block2 = 0;
+       if (block1 > len)
+               block1 = len;
+       block2 = len - block1;
+
+       ofs = b_peek_ofs(buf, ofs);
+
+       line = 0;
+       ptr = ofs;
+       while (ptr < ofs + block1) {
+               newptr = dump_text_line(&trace_buf, b_orig(buf), b_size(buf), ofs + block1, &line, ptr);
+               if (newptr == ptr)
+                       break;
+               ptr = newptr;
+       }
+
+       line = ptr = 0;
+       while (ptr < block2) {
+               newptr = dump_text_line(&trace_buf, b_orig(buf), b_size(buf), block2, &line, ptr);
+               if (newptr == ptr)
+                       break;
+               ptr = newptr;
+       }
+}
+
+/* the SPOP traces always expect that arg1, if non-null, is of type connection
+ * (from which we can derive spop_conn), that arg2, if non-null, is of type spop_strm,
+ * and that arg3, if non-null, is a buffer for rx/tx headers.
+ */
+static void spop_trace(enum trace_level level, uint64_t mask, const struct trace_source *src,
+                      const struct ist where, const struct ist func,
+                      const void *a1, const void *a2, const void *a3, const void *a4)
+{
+       const struct connection *conn = a1;
+       struct spop_conn *spop_conn = conn ? conn->ctx : NULL;
+       const struct spop_strm *spop_strm = a2;
+       const struct buffer *buf = a3;
+       const size_t *val = a4;
+
+       if (!spop_conn)
+               spop_conn = (spop_strm ? spop_strm->spop_conn : NULL);
+
+       if (!spop_conn || src->verbosity < SPOP_VERB_CLEAN)
+               return;
+
+       if (src->verbosity == SPOP_VERB_CLEAN)
+               return;
+
+       /* Display the value to the 4th argument (level > STATE) */
+       if (src->level > TRACE_LEVEL_STATE && val)
+               chunk_appendf(&trace_buf, " - VAL=%lu", (long)*val);
+
+       /* Display spop_conn info and, if defined, spop_strm info */
+       chunk_appendf(&trace_buf, " - spop_conn=%p(%s,0x%08x)", spop_conn, spop_conn_st_to_str(spop_conn->state), spop_conn->flags);
+       if (spop_conn->conn)
+               chunk_appendf(&trace_buf, " conn=%p(0x%08x)", spop_conn->conn, spop_conn->conn->flags);
+       if (spop_conn->errcode)
+               chunk_appendf(&trace_buf, " err=%02x", spop_conn->errcode);
+
+       if (spop_strm) {
+               if (spop_strm->id <= 0)
+                       chunk_appendf(&trace_buf, " dsi=%d dfi=%d", spop_conn->dsi, spop_conn->dfi);
+               if (spop_strm != spop_closed_stream && spop_strm != spop_unknown_stream) {
+                       chunk_appendf(&trace_buf, " spop_strm=%p(%d,%s,0x%08x)", spop_strm, spop_strm->id, spop_strm_st_to_str(spop_strm->state), spop_strm->flags);
+                       if (spop_strm->sd) {
+                               chunk_appendf(&trace_buf, " sd=%p(0x%08x)", spop_strm->sd, se_fl_get(spop_strm->sd));
+                               if (spop_strm_sc(spop_strm))
+                                       chunk_appendf(&trace_buf, " sc=%p(0x%08x)", spop_strm_sc(spop_strm), spop_strm_sc(spop_strm)->flags);
+                       }
+                       if (spop_strm->errcode)
+                               chunk_appendf(&trace_buf, " err=%02x", spop_strm->errcode);
+               }
+       }
+       if (src->verbosity == SPOP_VERB_MINIMAL)
+               return;
+
+       /* Display mbuf and dbuf info (level > USER & verbosity > SIMPLE) */
+       if (src->level > TRACE_LEVEL_USER) {
+               if (src->verbosity == SPOP_VERB_COMPLETE ||
+                   (src->verbosity == SPOP_VERB_ADVANCED && (mask & (SPOP_EV_SPOP_CONN_RECV|SPOP_EV_RX_FRAME))))
+                       chunk_appendf(&trace_buf, " dbuf=%u@%p+%u/%u",
+                                     (unsigned int)b_data(&spop_conn->dbuf), b_orig(&spop_conn->dbuf),
+                                     (unsigned int)b_head_ofs(&spop_conn->dbuf), (unsigned int)b_size(&spop_conn->dbuf));
+               if (src->verbosity == SPOP_VERB_COMPLETE ||
+                   (src->verbosity == SPOP_VERB_ADVANCED && (mask & (SPOP_EV_SPOP_CONN_SEND|SPOP_EV_TX_FRAME)))) {
+                       struct buffer *hmbuf = br_head(spop_conn->mbuf);
+                       struct buffer *tmbuf = br_tail(spop_conn->mbuf);
+
+                       chunk_appendf(&trace_buf, " .mbuf=[%u..%u|%u],h=[%u@%p+%u/%u],t=[%u@%p+%u/%u]",
+                                     br_head_idx(spop_conn->mbuf), br_tail_idx(spop_conn->mbuf), br_size(spop_conn->mbuf),
+                                     (unsigned int)b_data(hmbuf), b_orig(hmbuf),
+                                     (unsigned int)b_head_ofs(hmbuf), (unsigned int)b_size(hmbuf),
+                                     (unsigned int)b_data(tmbuf), b_orig(tmbuf),
+                                     (unsigned int)b_head_ofs(tmbuf), (unsigned int)b_size(tmbuf));
+               }
+
+               if (spop_strm && (src->verbosity == SPOP_VERB_COMPLETE ||
+                             (src->verbosity == SPOP_VERB_ADVANCED && (mask & (SPOP_EV_STRM_RECV)))))
+                       chunk_appendf(&trace_buf, " rxbuf=%u@%p+%u/%u",
+                                     (unsigned int)b_data(&spop_strm->rxbuf), b_orig(&spop_strm->rxbuf),
+                                     (unsigned int)b_head_ofs(&spop_strm->rxbuf), (unsigned int)b_size(&spop_strm->rxbuf));
+       }
+
+       /* Display htx info if defined (level > USER) */
+       if (src->level > TRACE_LEVEL_USER && buf) {
+               int full = 0, max = 3000, chunk = 1024;
+
+               /* Full info (level > STATE && verbosity > SIMPLE) */
+               if (src->level > TRACE_LEVEL_STATE) {
+                       if (src->verbosity == SPOP_VERB_COMPLETE)
+                               full = 1;
+                       else if (src->verbosity == SPOP_VERB_ADVANCED) {
+                               full = 1;
+                               max = 256;
+                               chunk = 64;
+                       }
+               }
+
+               chunk_appendf(&trace_buf, " buf=%u@%p+%u/%u",
+                             (unsigned int)b_data(buf), b_orig(buf),
+                             (unsigned int)b_head_ofs(buf), (unsigned int)b_size(buf));
+
+               if (b_data(buf) && full) {
+                       chunk_memcat(&trace_buf, "\n", 1);
+                       if (b_data(buf) < max)
+                               spop_trace_buf(buf, 0, b_data(buf));
+                       else {
+                               spop_trace_buf(buf, 0, chunk);
+                               chunk_memcat(&trace_buf, "  ...\n", 6);
+                               spop_trace_buf(buf, b_data(buf) - chunk, chunk);
+                       }
+               }
+       }
+}
+
+
+/*****************************************************/
+/* functions below are for dynamic buffer management */
+/*****************************************************/
+/* Tries to grab a buffer and to re-enable processing on mux <target>. The
+ * spop_conn flags are used to figure what buffer was requested. It returns 1 if
+ * the allocation succeeds, in which case the connection is woken up, or 0 if
+ * it's impossible to wake up and we prefer to be woken up later.
+ */
+static int spop_buf_available(void *target)
+{
+       struct spop_conn *spop_conn = target;
+       struct spop_strm *spop_strm;
+
+       if ((spop_conn->flags & SPOP_CF_DEM_DALLOC) && b_alloc(&spop_conn->dbuf, DB_MUX_RX)) {
+               TRACE_STATE("unblocking spop_conn, dbuf allocated", SPOP_EV_SPOP_CONN_RECV|SPOP_EV_SPOP_CONN_BLK|SPOP_EV_SPOP_CONN_WAKE, spop_conn->conn);
+               spop_conn->flags &= ~SPOP_CF_DEM_DALLOC;
+               spop_conn_restart_reading(spop_conn, 1);
+               return 1;
+       }
+
+       if ((spop_conn->flags & SPOP_CF_MUX_MALLOC) && b_alloc(br_tail(spop_conn->mbuf), DB_MUX_TX)) {
+               TRACE_STATE("unblocking spop_conn, mbuf allocated", SPOP_EV_SPOP_CONN_SEND|SPOP_EV_SPOP_CONN_BLK|SPOP_EV_SPOP_CONN_WAKE, spop_conn->conn);
+               spop_conn->flags &= ~SPOP_CF_MUX_MALLOC;
+               if (spop_conn->flags & SPOP_CF_DEM_MROOM) {
+                       spop_conn->flags &= ~SPOP_CF_DEM_MROOM;
+                       spop_conn_restart_reading(spop_conn, 1);
+               }
+               return 1;
+       }
+
+       if ((spop_conn->flags & SPOP_CF_DEM_SALLOC) &&
+           (spop_strm = spop_conn_st_by_id(spop_conn, spop_conn->dsi)) && spop_strm_sc(spop_strm) &&
+           b_alloc(&spop_strm->rxbuf, DB_SE_RX)) {
+               TRACE_STATE("unblocking spop_strm, rxbuf allocated", SPOP_EV_STRM_RECV|SPOP_EV_SPOP_STRM_BLK|SPOP_EV_STRM_WAKE, spop_conn->conn, spop_strm);
+               spop_conn->flags &= ~SPOP_CF_DEM_SALLOC;
+               spop_conn_restart_reading(spop_conn, 1);
+               return 1;
+       }
+
+       return 0;
+}
+
+static inline struct buffer *spop_get_buf(struct spop_conn *spop_conn, struct buffer *bptr)
+{
+       struct buffer *buf = NULL;
+
+       if (likely(!LIST_INLIST(&spop_conn->buf_wait.list)) &&
+           unlikely((buf = b_alloc(bptr, DB_MUX_RX)) == NULL)) {
+               b_queue(DB_MUX_RX, &spop_conn->buf_wait, spop_conn, spop_buf_available);
+       }
+       return buf;
+}
+
+static inline void spop_release_buf(struct spop_conn *spop_conn, struct buffer *bptr)
+{
+       if (bptr->size) {
+               b_free(bptr);
+               offer_buffers(NULL, 1);
+       }
+}
+
+static inline void spop_release_mbuf(struct spop_conn *spop_conn)
+{
+       struct buffer *buf;
+       unsigned int count = 0;
+
+       while (b_size(buf = br_head_pick(spop_conn->mbuf))) {
+               b_free(buf);
+               count++;
+       }
+       if (count)
+               offer_buffers(NULL, count);
+}
+
+/*****************************************************************/
+/* functions below are dedicated to the mux setup and management */
+/*****************************************************************/
+/* Indicates whether or not the we may call the spop_recv() function to attempt
+ * to receive data into the buffer and/or demux pending data. The condition is
+ * a bit complex due to some API limits for now. The rules are the following :
+ *   - if an error or a shutdown was detected on the connection and the buffer
+ *     is empty, we must not attempt to receive
+ *   - if the demux buf failed to be allocated, we must not try to receive and
+ *     we know there is nothing pending
+ *   - if no flag indicates a blocking condition, we may attempt to receive,
+ *     regardless of whether the demux buffer is full or not, so that only
+ *     de demux part decides whether or not to block. This is needed because
+ *     the connection API indeed prevents us from re-enabling receipt that is
+ *     already enabled in a polled state, so we must always immediately stop
+ *     as soon as the demux can't proceed so as never to hit an end of read
+ *     with data pending in the buffers.
+ *   - otherwise must may not attempt
+ */
+static inline int spop_recv_allowed(const struct spop_conn *spop_conn)
+{
+       if (b_data(&spop_conn->dbuf) == 0 &&
+           ((spop_conn->flags & (SPOP_CF_RCVD_SHUT|SPOP_CF_ERROR)) || spop_conn->state == SPOP_CS_CLOSED))
+               return 0;
+
+       if (!(spop_conn->flags & SPOP_CF_DEM_DALLOC) &&
+           !(spop_conn->flags & SPOP_CF_DEM_BLOCK_ANY))
+               return 1;
+
+       return 0;
+}
+
+/* Restarts reading on the connection if it was not enabled */
+static inline void spop_conn_restart_reading(const struct spop_conn *spop_conn, int consider_buffer)
+{
+       if (!spop_recv_allowed(spop_conn))
+               return;
+       if ((!consider_buffer || !b_data(&spop_conn->dbuf)) &&
+           (spop_conn->wait_event.events & SUB_RETRY_RECV))
+               return;
+       tasklet_wakeup(spop_conn->wait_event.tasklet);
+}
+
+/* Returns the number of allocatable outgoing streams for the connection taking
+ * the number reserved streams into account.
+ */
+static inline int spop_streams_left(const struct spop_conn *spop_conn)
+{
+       int ret;
+
+       ret = (unsigned int)(0x7FFFFFFF - spop_conn->max_id) - spop_conn->nb_reserved - 1;
+       if (ret < 0)
+               ret = 0;
+       return ret;
+}
+
+/* Returns the number of streams in use on a connection to figure if it's
+ * idle or not. We check nb_sc and not nb_streams as the caller will want
+ * to know if it was the last one after a detach().
+ */
+static int spop_used_streams(struct connection *conn)
+{
+       struct spop_conn *spop_conn = conn->ctx;
+
+       return spop_conn->nb_sc;
+}
+
+/* Returns the number of concurrent streams available on the connection */
+static int spop_avail_streams(struct connection *conn)
+{
+       struct server *srv = objt_server(conn->target);
+       struct spop_conn *spop_conn = conn->ctx;
+       int ret1, ret2;
+
+       /* Don't open new stream if the connection is closed */
+       if (spop_conn->state == SPOP_CS_CLOSED)
+               return 0;
+
+       /* May be negative if this setting has changed */
+       ret1 = (spop_conn->streams_limit - spop_conn->nb_streams);
+
+       /* we must also consider the limit imposed by stream IDs */
+       ret2 = spop_streams_left(spop_conn);
+       ret1 = MIN(ret1, ret2);
+       if (ret1 > 0 && srv && srv->max_reuse >= 0) {
+               ret2 = ((spop_conn->stream_cnt <= srv->max_reuse) ? srv->max_reuse - spop_conn->stream_cnt + 1: 0);
+               ret1 = MIN(ret1, ret2);
+       }
+       return ret1;
+}
+
+/* Initializes the mux once it's attached. Only outgoing connections are
+ * supported. So the context is already initialized before installing the
+ * mux. <input> is always used as Input buffer and may contain data. It is the
+ * caller responsibility to not reuse it anymore. Returns < 0 on error.
+ */
+static int spop_init(struct connection *conn, struct proxy *px, struct session *sess,
+                    struct buffer *input)
+{
+       struct spop_conn *spop_conn;
+       struct spop_strm *spop_strm;
+       struct sedesc *sdo = NULL;
+       struct task *t = NULL;
+       void *conn_ctx = conn->ctx;
+
+       TRACE_ENTER(SPOP_EV_SPOP_CONN_NEW);
+
+       spop_conn = pool_alloc(pool_head_spop_conn);
+       if (!spop_conn) {
+               TRACE_ERROR("spop_conn allocation failure", SPOP_EV_SPOP_CONN_NEW|SPOP_EV_SPOP_CONN_END|SPOP_EV_SPOP_CONN_ERR);
+               goto fail_conn;
+       }
+
+       spop_conn->shut_timeout = spop_conn->timeout = px->timeout.server;
+       if (tick_isset(px->timeout.serverfin))
+               spop_conn->shut_timeout = px->timeout.serverfin;
+       spop_conn->flags = SPOP_CF_NONE;
+
+       spop_conn->proxy = px;
+       spop_conn->agent = NULL;
+       spop_conn->task = NULL;
+       if (tick_isset(spop_conn->timeout)) {
+               t = task_new_here();
+               if (!t) {
+                       TRACE_ERROR("spop_conn task allocation failure", SPOP_EV_SPOP_CONN_NEW|SPOP_EV_SPOP_CONN_END|SPOP_EV_SPOP_CONN_ERR);
+                       goto fail;
+               }
+
+               spop_conn->task = t;
+               t->process = spop_timeout_task;
+               t->context = spop_conn;
+               t->expire = tick_add(now_ms, spop_conn->timeout);
+       }
+
+       spop_conn->wait_event.tasklet = tasklet_new();
+       if (!spop_conn->wait_event.tasklet)
+               goto fail;
+       spop_conn->wait_event.tasklet->process = spop_io_cb;
+       spop_conn->wait_event.tasklet->context = spop_conn;
+       spop_conn->wait_event.events = 0;
+
+       /* Initialise the context. */
+       spop_conn->state = SPOP_CS_HA_HELLO;
+       spop_conn->errcode = SPOP_ERR_NONE;
+       spop_conn->conn = conn;
+       spop_conn->streams_limit = 1;
+       spop_conn->max_id = -1;
+       spop_conn->nb_streams = 0;
+       spop_conn->nb_sc = 0;
+       spop_conn->nb_reserved = 0;
+       spop_conn->stream_cnt = 0;
+
+       spop_conn->dbuf = *input;
+       spop_conn->dsi = -1;
+
+       br_init(spop_conn->mbuf, sizeof(spop_conn->mbuf) / sizeof(spop_conn->mbuf[0]));
+       spop_conn->streams_by_id = EB_ROOT;
+       LIST_INIT(&spop_conn->send_list);
+       LIST_INIT(&spop_conn->buf_wait.list);
+
+       conn->ctx = spop_conn;
+
+       if (t)
+               task_queue(t);
+
+       /* FIXME: this is temporary, for outgoing connections we need to
+        * immediately allocate a stream until the code is modified so that the
+        * caller calls ->attach(). For now the outgoing sc is stored as
+        * conn->ctx by the caller and saved in conn_ctx.
+        */
+       spop_strm = spop_stconn_new(spop_conn, conn_ctx, sess);
+       if (!spop_strm)
+               goto fail;
+
+       /* Retrieve the SPOE agent attached to the opposite endpoint. Only
+        * undefined when there is no opposite endpoint (healthcheck)
+        */
+       sdo = spop_strm_opposite_sd(spop_strm);
+       if (sdo) {
+               spop_conn->agent = spoe_appctx_agent(sc_appctx(sdo->sc));
+               BUG_ON(!spop_conn->agent);
+       }
+
+       /* Repare to read something */
+       spop_conn_restart_reading(spop_conn, 1);
+       TRACE_LEAVE(SPOP_EV_SPOP_CONN_NEW, conn);
+       return 0;
+
+  fail:
+       task_destroy(t);
+       tasklet_free(spop_conn->wait_event.tasklet);
+       pool_free(pool_head_spop_conn, spop_conn);
+  fail_conn:
+       conn->ctx = conn_ctx; // restore saved ctx
+       TRACE_DEVEL("leaving in error", SPOP_EV_SPOP_CONN_NEW|SPOP_EV_SPOP_CONN_END|SPOP_EV_SPOP_CONN_ERR);
+       return -1;
+}
+
+/* Release function. This one should be called to free all resources allocated
+ * to the mux.
+ */
+static void spop_release(struct spop_conn *spop_conn)
+{
+       struct connection *conn = spop_conn->conn;
+
+       TRACE_POINT(SPOP_EV_SPOP_CONN_END);
+
+       b_dequeue(&spop_conn->buf_wait);
+
+       spop_release_buf(spop_conn, &spop_conn->dbuf);
+       spop_release_mbuf(spop_conn);
+
+       if (spop_conn->task) {
+               spop_conn->task->context = NULL;
+               task_wakeup(spop_conn->task, TASK_WOKEN_OTHER);
+               spop_conn->task = NULL;
+       }
+       tasklet_free(spop_conn->wait_event.tasklet);
+       if (conn && spop_conn->wait_event.events != 0)
+               conn->xprt->unsubscribe(conn, conn->xprt_ctx, spop_conn->wait_event.events,
+                                       &spop_conn->wait_event);
+
+       pool_free(pool_head_spop_conn, spop_conn);
+
+       if (conn) {
+               conn->mux = NULL;
+               conn->ctx = NULL;
+               TRACE_DEVEL("freeing conn", SPOP_EV_SPOP_CONN_END, conn);
+
+               conn_stop_tracking(conn);
+               conn_full_close(conn);
+               if (conn->destroy_cb)
+                       conn->destroy_cb(conn);
+               conn_free(conn);
+       }
+}
+
+
+/* Returns the next allocatable outgoing stream ID for the SPOP connection, or
+ * -1 if no more is allocatable.
+ */
+static inline int32_t spop_conn_get_next_sid(const struct spop_conn *spop_conn)
+{
+       int32_t id = (spop_conn->max_id + 1) | 1;
+
+       if ((id & 0x80000000U))
+               id = -1;
+       return id;
+}
+
+/* Returns the stream associated with id <id> or NULL if not found */
+static inline struct spop_strm *spop_conn_st_by_id(struct spop_conn *spop_conn, int id)
+{
+       struct eb32_node *node;
+
+       if (id == 0)
+               return (struct spop_strm *)spop_closed_stream;
+
+       if (id > spop_conn->max_id)
+               return (struct spop_strm *)spop_unknown_stream;
+
+       node = eb32_lookup(&spop_conn->streams_by_id, id);
+       if (!node)
+               return (struct spop_strm *)spop_unknown_stream;
+       return container_of(node, struct spop_strm, by_id);
+}
+
+/* Detect a pending read0 for a SPOP connection. It happens if a read0 was
+ * already reported on a previous xprt->rcvbuf() AND a frame parser failed
+ * to parse pending data, confirming no more progress is possible because
+ * we're facing a truncated frame. The function returns 1 to report a read0
+ * or 0 otherwise.
+ */
+static int spop_conn_read0_pending(struct spop_conn *spop_conn)
+{
+       return !!(spop_conn->flags & SPOP_CF_END_REACHED);
+}
+
+/* returns true if the connection is allowed to expire, false otherwise. A
+ * connection may expire when it has no attached streams. As long as streams
+ * are attached, the application layer is responsible for timeout management,
+ * and each layer will detach when it doesn't want to wait anymore. When the
+ * last one leaves, the connection must take over timeout management.
+ */
+static inline int spop_conn_may_expire(const struct spop_conn *spop_conn)
+{
+       return !spop_conn->nb_sc;
+}
+
+// TODO: spop_conn_max_concurrent_streams() : default 20 but add a agent parameter !
+
+/* Returns true if the SPOP connection must be release */
+static inline int spop_conn_is_dead(struct spop_conn *spop_conn)
+{
+       if (eb_is_empty(&spop_conn->streams_by_id) &&                   /* don't close if streams exist */
+           ((spop_conn->flags & SPOP_CF_ERROR) ||                      /* errors close immediately */
+            (spop_conn->flags & SPOP_CF_ERR_PENDING && spop_conn->state < SPOP_CS_FRAME_H) || /* early error during connect */
+            (spop_conn->state == SPOP_CS_CLOSED && !spop_conn->task) ||/* a timeout stroke earlier */
+            (!(spop_conn->conn->owner)) ||                             /* Nobody's left to take care of the connection, drop it now */
+            (!br_data(spop_conn->mbuf) &&                              /* mux buffer empty, also process clean events below */
+             (spop_conn->flags & SPOP_CF_RCVD_SHUT))))
+             return 1;
+       return 0;
+}
+
+/* update spop_conn timeout if needed */
+static __maybe_unused void spop_conn_update_timeout(struct spop_conn *spop_conn)
+{
+       TRACE_ENTER(SPOP_EV_SPOP_CONN_WAKE, spop_conn->conn);
+
+       if (!spop_conn->task)
+               goto leave;
+
+       if (spop_conn_may_expire(spop_conn)) {
+               /* no more streams attached */
+               if (br_data(spop_conn->mbuf)) {
+                       /* pending output data: always the regular data timeout */
+                       spop_conn->task->expire = tick_add_ifset(now_ms, spop_conn->timeout);
+               }
+               else {
+                       if (spop_conn->flags & (SPOP_CF_DISCO_SENT|SPOP_CF_DISCO_FAILED)) {
+                               /* DISCONNECT sent (or failed), closing in progress */
+                               int exp = tick_add_ifset(now_ms, spop_conn->shut_timeout);
+
+                               spop_conn->task->expire = tick_first(spop_conn->task->expire, exp);
+                       }
+
+                       /* if a timeout above was not set, fall back to the default one */
+                       if (!tick_isset(spop_conn->task->expire))
+                               spop_conn->task->expire = tick_add_ifset(now_ms, spop_conn->timeout);
+               }
+
+               // TODO: handle close_spread for now, there is no idle connection
+       }
+       else
+               spop_conn->task->expire = TICK_ETERNITY;
+       task_queue(spop_conn->task);
+ leave:
+       TRACE_LEAVE(SPOP_EV_SPOP_CONN_WAKE);
+}
+
+/********************************************************/
+/* functions below are for the SPOP protocol processing */
+/********************************************************/
+/* Marks an error on the connection. Before HELLO frame are sent, we must not send
+ * a DISCONNECT frame, so we set SPOP_CF_DISCO_FAILED to make sure it will not
+ * even try.. */
+static inline void spop_conn_error(struct spop_conn *spop_conn, enum spop_error err)
+{
+       TRACE_POINT(SPOP_EV_SPOP_CONN_ERR, spop_conn->conn, 0, 0, (size_t[]){err});
+       spop_conn->errcode = err;
+       if (spop_conn->state == SPOP_CS_HA_HELLO)
+               spop_conn->flags |= SPOP_CF_DISCO_FAILED;
+       spop_conn->state = SPOP_CS_ERROR;
+}
+
+/* Marks an error on the stream */
+static inline void spop_strm_error(struct spop_strm *spop_strm, enum spop_error err)
+{
+       if (spop_strm->id && spop_strm->state != SPOP_SS_ERROR) {
+               TRACE_POINT(SPOP_EV_SPOP_STRM_ERR, spop_strm->spop_conn->conn, spop_strm);
+               spop_strm->errcode = err;
+               if (spop_strm->state < SPOP_SS_ERROR) {
+                       spop_strm->state = SPOP_SS_ERROR;
+                       TRACE_STATE("switching to ERROR", SPOP_EV_SPOP_STRM_ERR, spop_strm->spop_conn->conn, spop_strm);
+               }
+               se_fl_set_error(spop_strm->sd);
+               if (!spop_strm->sd->abort_info.info) {
+                        spop_strm->sd->abort_info.info = (SE_ABRT_SRC_MUX_SPOP << SE_ABRT_SRC_SHIFT);
+                        spop_strm->sd->abort_info.code = spop_strm->errcode;
+                }
+       }
+}
+
+/* Attempts to notify the data layer of recv availability */
+static void spop_strm_notify_recv(struct spop_strm *spop_strm)
+{
+       if (spop_strm->subs && (spop_strm->subs->events & SUB_RETRY_RECV)) {
+               TRACE_POINT(SPOP_EV_STRM_WAKE, spop_strm->spop_conn->conn, spop_strm);
+               tasklet_wakeup(spop_strm->subs->tasklet);
+               spop_strm->subs->events &= ~SUB_RETRY_RECV;
+               if (!spop_strm->subs->events)
+                       spop_strm->subs = NULL;
+       }
+}
+
+/* Attempts to notify the data layer of send availability */
+static void spop_strm_notify_send(struct spop_strm *spop_strm)
+{
+       if (spop_strm->subs && (spop_strm->subs->events & SUB_RETRY_SEND)) {
+               TRACE_POINT(SPOP_EV_STRM_WAKE, spop_strm->spop_conn->conn, spop_strm);
+               spop_strm->flags |= SPOP_SF_NOTIFIED;
+               tasklet_wakeup(spop_strm->subs->tasklet);
+               spop_strm->subs->events &= ~SUB_RETRY_SEND;
+               if (!spop_strm->subs->events)
+                       spop_strm->subs = NULL;
+       }
+}
+
+/* Alerts the data layer, trying to wake it up by all means, following
+ * this sequence :
+ *   - if the spop stream' data layer is subscribed to recv, then it's woken up
+ *     for recv
+ *   - if its subscribed to send, then it's woken up for send
+ *   - if it was subscribed to neither, its ->wake() callback is called
+ * It is safe to call this function with a closed stream which doesn't have a
+ * stream connector anymore.
+ */
+static void spop_strm_alert(struct spop_strm *spop_strm)
+{
+       TRACE_POINT(SPOP_EV_STRM_WAKE, spop_strm->spop_conn->conn, spop_strm);
+       if (spop_strm->subs) {
+               spop_strm_notify_recv(spop_strm);
+               spop_strm_notify_send(spop_strm);
+       }
+       else if (spop_strm_sc(spop_strm) && spop_strm_sc(spop_strm)->app_ops->wake != NULL) {
+               TRACE_POINT(SPOP_EV_STRM_WAKE, spop_strm->spop_conn->conn, spop_strm);
+               spop_strm_sc(spop_strm)->app_ops->wake(spop_strm_sc(spop_strm));
+       }
+}
+
+/* Writes the 32-bit frame size <len> at address <frame> */
+static inline void spop_set_frame_size(void *frame, uint32_t len)
+{
+        write_n32(frame, len);
+}
+
+/* reads <bytes> bytes from buffer <b> starting at relative offset <o> from the
+ * current pointer, dealing with wrapping, and stores the result in <dst>. It's
+ * the caller's responsibility to verify that there are at least <bytes> bytes
+ * available in the buffer's input prior to calling this function. The buffer
+ * is assumed not to hold any output data.
+ */
+static inline __maybe_unused void spop_get_buf_bytes(void *dst, size_t bytes,
+                                    const struct buffer *b, int o)
+{
+        readv_bytes(dst, bytes, b_peek(b, o), b_wrap(b) - b_peek(b, o), b_orig(b));
+}
+
+static inline __maybe_unused uint16_t spop_get_n16(const struct buffer *b, int o)
+{
+        return readv_n16(b_peek(b, o), b_wrap(b) - b_peek(b, o), b_orig(b));
+}
+
+static inline __maybe_unused uint32_t spop_get_n32(const struct buffer *b, int o)
+{
+        return readv_n32(b_peek(b, o), b_wrap(b) - b_peek(b, o), b_orig(b));
+}
+
+static inline __maybe_unused uint64_t spop_get_n64(const struct buffer *b, int o)
+{
+        return readv_n64(b_peek(b, o), b_wrap(b) - b_peek(b, o), b_orig(b));
+}
+
+static __maybe_unused int spop_get_varint(const struct buffer *b, int o, uint64_t *i)
+{
+       unsigned char *p;
+       size_t idx = o;
+       int r;
+
+       if (idx > b_data(b))
+               return -1;
+
+       p = (unsigned char *)b_peek(b, idx++);
+       *i = (uint64_t)*p;
+       if (*i < 240)
+               return 1;
+
+       r = 4;
+       do {
+               if (idx > b_data(b))
+                       return -1;
+               p = (unsigned char *)b_peek(b, idx++);
+               *i += (uint64_t)*p << r;
+               r += 7;
+       } while (*p >= 128);
+
+       return (idx - o);
+}
+
+/* Peeks an SPOP frame header from offset <o> of buffer <b> into descriptor <h>.
+ * Returns zero if some bytes are missing, otherwise the number of read bytes is
+ * returned on success. The buffer is assumed not to contain any output data.
+ */
+static int spop_peek_frame_hdr(const struct buffer *b, int o, struct spop_frame_header *hdr)
+{
+       int o1, o2;
+
+        if (b_data(b) < o + 11)
+                return 0;
+
+       o1 = o;
+       hdr->len = spop_get_n32(b, o1);
+
+       hdr->type = *(uint8_t*)b_peek(b, o1+4);
+       hdr->flags = spop_get_n32(b, o1+5);
+       o1 += 9;
+
+       /* Get the stream-id and the frame-id */
+       o2 = spop_get_varint(b, o1, &hdr->sid);
+       if (o2 == -1)
+               return 0;
+       o1 += o2;
+       o2 = spop_get_varint(b, o1, &hdr->fid);
+       if (o2 == -1)
+               return 0;
+       o1 += o2;
+
+       /* Remove the header length from the frame length */
+       hdr->len -= o1 - (o + 4);
+       return (o1 - o);
+}
+
+/* skip the <o> bytes corresponding to the frame header possibly parsed by
+ * spop_peek_frame_hdr() above.
+ */
+static inline void spop_skip_frame_hdr(struct buffer *b, size_t o)
+{
+        b_del(b, o);
+}
+
+/* same as above, automatically advances the buffer on success */
+static inline int spop_get_frame_hdr(struct buffer *b, struct spop_frame_header *hdr)
+{
+        int ret;
+
+        ret = spop_peek_frame_hdr(b, 0, hdr);
+        if (ret > 0)
+                spop_skip_frame_hdr(b, ret);
+        return ret;
+}
+
+
+/* Marks a SPOP stream as CLOSED and decrement the number of active streams for
+ * its connection if the stream was not yet closed. Please use this exclusively
+ * before closing a stream to ensure stream count is well maintained.
+ */
+static inline void spop_strm_close(struct spop_strm *spop_strm)
+{
+       if (spop_strm->state != SPOP_SS_CLOSED) {
+               TRACE_ENTER(SPOP_EV_SPOP_STRM_END, spop_strm->spop_conn->conn, spop_strm);
+               spop_strm->spop_conn->nb_streams--;
+               if (!spop_strm->id)
+                       spop_strm->spop_conn->nb_reserved--;
+               if (spop_strm_sc(spop_strm)) {
+                       if (!se_fl_test(spop_strm->sd, SE_FL_EOS) && !b_data(&spop_strm->rxbuf))
+                               spop_strm_notify_recv(spop_strm);
+               }
+               spop_strm->state = SPOP_SS_CLOSED;
+               TRACE_STATE("switching to CLOSED", SPOP_EV_SPOP_STRM_END, spop_strm->spop_conn->conn, spop_strm);
+               TRACE_LEAVE(SPOP_EV_SPOP_STRM_END, spop_strm->spop_conn->conn, spop_strm);
+       }
+}
+
+/* Check spop_conn and spop_strm flags to evaluate if EOI/EOS/ERR_PENDING/ERROR flags must
+ * be set on the SE.
+ */
+static inline void spop_strm_propagate_term_flags(struct spop_conn *spop_conn, struct spop_strm *spop_strm)
+{
+       if (spop_conn_read0_pending(spop_conn) || spop_strm->state == SPOP_SS_CLOSED) {
+               se_fl_set(spop_strm->sd, SE_FL_EOS);
+               if (spop_conn->errcode)
+                       se_fl_set(spop_strm->sd, SE_FL_ERROR);
+       }
+       if (se_fl_test(spop_strm->sd, SE_FL_ERR_PENDING))
+               se_fl_set(spop_strm->sd, SE_FL_ERROR);
+}
+
+
+/* Detaches a SPOP stream from its SPOP connection and releases it to the
+ * spop_strm pool.
+ */
+static void spop_strm_destroy(struct spop_strm *spop_strm)
+{
+       struct connection *conn = spop_strm->spop_conn->conn;
+
+       TRACE_ENTER(SPOP_EV_SPOP_STRM_END, conn, spop_strm);
+
+       spop_strm_close(spop_strm);
+       eb32_delete(&spop_strm->by_id);
+       if (b_size(&spop_strm->rxbuf)) {
+               b_free(&spop_strm->rxbuf);
+               offer_buffers(NULL, 1);
+       }
+       if (spop_strm->subs)
+               spop_strm->subs->events = 0;
+
+       /* There's no need to explicitly call unsubscribe here, the only
+        * reference left would be in the spop_conn send_list/fctl_list, and if
+        * we're in it, we're getting out anyway
+        */
+       spop_remove_from_list(spop_strm);
+
+       BUG_ON(spop_strm->sd && !se_fl_test(spop_strm->sd, SE_FL_ORPHAN));
+       sedesc_free(spop_strm->sd);
+       pool_free(pool_head_spop_strm, spop_strm);
+
+       TRACE_LEAVE(SPOP_EV_SPOP_STRM_END, conn);
+}
+
+/* Allocates a new stream <id> for connection <spop_conn> and adds it into spop_conn's
+ * stream tree. In case of error, nothing is added and NULL is returned. The
+ * causes of errors can be any failed memory allocation. The caller is
+ * responsible for checking if the connection may support an extra stream prior
+ * to calling this function.
+ */
+static struct spop_strm *spop_strm_new(struct spop_conn *spop_conn, int id)
+{
+       struct spop_strm *spop_strm;
+
+       TRACE_ENTER(SPOP_EV_SPOP_STRM_NEW, spop_conn->conn);
+
+       spop_strm = pool_alloc(pool_head_spop_strm);
+       if (!spop_strm) {
+               TRACE_ERROR("spop_strm allocation failure", SPOP_EV_SPOP_STRM_NEW|SPOP_EV_SPOP_STRM_ERR|SPOP_EV_SPOP_STRM_END, spop_conn->conn);
+               goto out;
+       }
+
+       spop_strm->subs = NULL;
+       LIST_INIT(&spop_strm->list);
+       spop_strm->spop_conn = spop_conn;
+       spop_strm->sd = NULL;
+       spop_strm->flags = SPOP_SF_NONE;
+       spop_strm->state = SPOP_SS_IDLE;
+       spop_strm->rxbuf = BUF_NULL;
+
+       spop_strm->by_id.key = spop_strm->id = id;
+       spop_strm->fid = 0;
+       if (id > 0) {
+               spop_conn->max_id = id;
+               spop_strm->state = SPOP_SS_OPEN;
+       }
+       else
+               spop_conn->nb_reserved++;
+
+       eb32_insert(&spop_conn->streams_by_id, &spop_strm->by_id);
+       spop_conn->nb_streams++;
+       spop_conn->stream_cnt++;
+
+       TRACE_LEAVE(SPOP_EV_SPOP_STRM_NEW, spop_conn->conn, spop_strm);
+       return spop_strm;
+
+  out:
+       TRACE_DEVEL("leaving in error", SPOP_EV_SPOP_STRM_NEW|SPOP_EV_SPOP_STRM_ERR|SPOP_EV_SPOP_STRM_END, spop_conn->conn);
+       return NULL;
+}
+
+/* Allocates a new stream associated to stream connector <sc> on the SPOP connection
+ * <spop_conn> and returns it, or NULL in case of memory allocation error or if the
+ * highest possible stream ID was reached.
+ */
+static struct spop_strm *spop_stconn_new(struct spop_conn *spop_conn, struct stconn *sc,
+                                        struct session *sess)
+{
+       struct spop_strm *spop_strm = NULL;
+
+       TRACE_ENTER(SPOP_EV_SPOP_STRM_NEW, spop_conn->conn);
+       if (spop_conn->nb_streams >= spop_conn->streams_limit) {
+               TRACE_ERROR("streams_limit reached", SPOP_EV_SPOP_STRM_NEW|SPOP_EV_SPOP_STRM_END|SPOP_EV_SPOP_STRM_ERR, spop_conn->conn);
+               goto out;
+       }
+
+       if (spop_streams_left(spop_conn) < 1) {
+               TRACE_ERROR("!streams_left", SPOP_EV_SPOP_STRM_NEW|SPOP_EV_SPOP_STRM_END|SPOP_EV_SPOP_STRM_ERR, spop_conn->conn);
+               goto out;
+       }
+
+       /* Defer choosing the ID until we send the first message to create the stream */
+       spop_strm = spop_strm_new(spop_conn, 0);
+       if (!spop_strm) {
+               TRACE_ERROR("fstream allocation failure", SPOP_EV_SPOP_STRM_NEW|SPOP_EV_SPOP_STRM_END|SPOP_EV_SPOP_STRM_ERR, spop_conn->conn);
+               goto out;
+       }
+       if (sc_attach_mux(sc, spop_strm, spop_conn->conn) < 0)
+               goto out;
+       spop_strm->sd = sc->sedesc;
+       spop_strm->sess = sess;
+       spop_conn->nb_sc++;
+       TRACE_LEAVE(SPOP_EV_SPOP_STRM_NEW, spop_conn->conn, spop_strm);
+       return spop_strm;
+
+  out:
+       TRACE_DEVEL("leaving on error", SPOP_EV_SPOP_STRM_NEW|SPOP_EV_SPOP_STRM_END|SPOP_EV_SPOP_STRM_ERR, spop_conn->conn);
+       spop_strm_destroy(spop_strm);
+       return NULL;
+}
+
+/* Wakes a specific stream and assign its stream connector some SE_FL_* flags among
+ * SE_FL_ERR_PENDING and SE_FL_ERROR if needed. The stream's state is
+ * automatically updated accordingly. If the stream is orphaned, it is
+ * destroyed.
+ */
+static void spop_strm_wake_one_stream(struct spop_strm *spop_strm)
+{
+       struct spop_conn *spop_conn = spop_strm->spop_conn;
+
+       TRACE_ENTER(SPOP_EV_STRM_WAKE, spop_conn->conn, spop_strm);
+
+       if (!spop_strm_sc(spop_strm)) {
+               /* this stream was already orphaned */
+               spop_strm_destroy(spop_strm);
+               TRACE_DEVEL("leaving with no spop_strm", SPOP_EV_STRM_WAKE, spop_conn->conn);
+               return;
+       }
+
+       if (spop_conn_read0_pending(spop_conn)) {
+               if (spop_strm->state == SPOP_SS_OPEN) {
+                       spop_strm->state = SPOP_SS_HREM;
+                       TRACE_STATE("switching to HREM", SPOP_EV_STRM_WAKE|SPOP_EV_SPOP_STRM_END, spop_conn->conn, spop_strm);
+               }
+               else if (spop_strm->state == SPOP_SS_HLOC)
+                       spop_strm_close(spop_strm);
+       }
+
+       if (spop_conn->state == SPOP_CS_CLOSED || (spop_conn->flags & (SPOP_CF_ERR_PENDING|SPOP_CF_ERROR))) {
+               if (spop_conn->state == SPOP_CS_CLOSED || (spop_conn->flags & SPOP_CF_ERROR))
+                       se_fl_set(spop_strm->sd, SE_FL_EOS);
+               se_fl_set_error(spop_strm->sd);
+               if (!spop_strm->sd->abort_info.info) {
+                       spop_strm->sd->abort_info.info = (SE_ABRT_SRC_MUX_SPOP << SE_ABRT_SRC_SHIFT);
+                       spop_strm->sd->abort_info.code = spop_conn->errcode;
+               }
+
+               if (spop_strm->state < SPOP_SS_ERROR) {
+                       spop_strm->state = SPOP_SS_ERROR;
+                       TRACE_STATE("switching to ERROR", SPOP_EV_STRM_WAKE|SPOP_EV_SPOP_STRM_END, spop_conn->conn, spop_strm);
+               }
+       }
+
+       spop_strm_alert(spop_strm);
+
+       TRACE_LEAVE(SPOP_EV_STRM_WAKE, spop_conn->conn, spop_strm);
+}
+
+/* Wakes unassigned streams (ID == 0) attached to the connection. */
+static void spop_wake_unassigned_streams(struct spop_conn *spop_conn)
+{
+       struct eb32_node *node;
+       struct spop_strm *spop_strm;
+
+       node = eb32_lookup(&spop_conn->streams_by_id, 0);
+       while (node) {
+               spop_strm = container_of(node, struct spop_strm, by_id);
+               if (spop_strm->id > 0)
+                       break;
+               node = eb32_next(node);
+               spop_strm_wake_one_stream(spop_strm);
+       }
+}
+
+/* Wakes the streams attached to the connection, whose id is greater than <last>
+ * or unassigned.
+ */
+static void spop_wake_some_streams(struct spop_conn *spop_conn, int last)
+{
+       struct eb32_node *node;
+       struct spop_strm *spop_strm;
+
+       TRACE_ENTER(SPOP_EV_STRM_WAKE, spop_conn->conn);
+
+       /* Wake all streams with ID > last */
+       node = eb32_lookup_ge(&spop_conn->streams_by_id, last + 1);
+       while (node) {
+               spop_strm = container_of(node, struct spop_strm, by_id);
+               node = eb32_next(node);
+               spop_strm_wake_one_stream(spop_strm);
+       }
+       spop_wake_unassigned_streams(spop_conn);
+
+       TRACE_LEAVE(SPOP_EV_STRM_WAKE, spop_conn->conn);
+}
+
+/* Sends a HAPROXY HELLO frame. Returns > 0 on success, 0 if it couldn't do
+ * anything. It is highly unexpected, but if the frame is larger than a buffer
+ * and cannot be encoded in one time, an error is triggered and the connection is
+ * closed. HEELO frame cannot be split.
+ */
+static int spop_conn_send_hello(struct spop_conn *spop_conn)
+{
+       struct buffer outbuf;
+       struct buffer *mbuf;
+       struct buffer *chk;
+       char *p, *end;
+       size_t sz;
+       int ret = 0;
+
+       TRACE_ENTER(SPOP_EV_TX_FRAME|SPOP_EV_TX_HELLO, spop_conn->conn);
+
+       mbuf = br_tail(spop_conn->mbuf);
+  retry:
+       if (!spop_get_buf(spop_conn, mbuf)) {
+               spop_conn->flags |= SPOP_CF_MUX_MALLOC;
+               spop_conn->flags |= SPOP_CF_DEM_MROOM;
+               TRACE_STATE("waiting for fconn mbuf ring allocation", SPOP_EV_TX_FRAME|SPOP_EV_SPOP_CONN_BLK, spop_conn->conn);
+               ret = 0;
+               goto end;
+       }
+
+       while (1) {
+               outbuf = b_make(b_tail(mbuf), b_contig_space(mbuf), 0, 0);
+               if (outbuf.size >= 11 || !b_space_wraps(mbuf))
+                       break;
+         realign_again:
+               b_slow_realign(mbuf, trash.area, b_data(mbuf));
+       }
+
+       if (outbuf.size < 11)
+               goto full;
+
+       /* len: 4-bytes (fill later) type: (1)HAPROXY-HELLO, flags: 4-bytes (FIN=1)
+        * stream-id: 0x00, frame-id: 0x00 */
+       memcpy(outbuf.area, "\x00\x00\x00\x00\x01\x00\x00\x00\x01\x00\x00", 11);
+       outbuf.data = 11;
+
+       p = b_tail(&outbuf);
+       end = b_orig(&outbuf) + b_size(&outbuf);
+
+       /* "supported-versions" K/V item */
+       sz = SPOP_SLEN(SPOP_SUPPORTED_VERSIONS_KEY);
+       if (spoe_encode_buffer(SPOP_SUPPORTED_VERSIONS_KEY, sz, &p, end) == -1)
+               goto full;
+
+       *p++ = SPOP_DATA_T_STR;
+       sz = SPOP_SLEN(SPOP_SUPPORTED_VERSIONS_VAL);
+       if (spoe_encode_buffer(SPOP_SUPPORTED_VERSIONS_VAL, sz, &p, end) == -1)
+               goto full;
+
+       /* "max-fram-size" K/V item */
+       sz = SPOP_SLEN(SPOP_MAX_FRAME_SIZE_KEY);
+       if (spoe_encode_buffer(SPOP_MAX_FRAME_SIZE_KEY, sz, &p, end) == -1)
+               goto full;
+
+       *p++ = SPOP_DATA_T_UINT32;
+       if (encode_varint(spop_conn->agent ? spop_conn->agent->max_frame_size : SPOP_MAX_FRAME_SIZE, &p, end) == -1)
+               goto full;
+
+       /* "capabilities" K/V item */
+       sz = SPOP_SLEN(SPOP_CAPABILITIES_KEY);
+       if (spoe_encode_buffer(SPOP_CAPABILITIES_KEY, sz, &p, end) == -1)
+               goto full;
+
+       *p++ = SPOP_DATA_T_STR;
+       chk = get_trash_chunk();
+       if (spop_conn->agent && (spop_conn->agent->flags & SPOE_FL_PIPELINING)) {
+               memcpy(chk->area, "pipelining", 10);
+               chk->data += 10;
+       }
+       if (spoe_encode_buffer(chk->area, chk->data, &p, end) == -1)
+               goto full;
+
+       /* (optional) "engine-id" K/V item, if present */
+       if (spop_conn->agent && spop_conn->agent->engine_id != NULL) {
+               sz = SPOP_SLEN(SPOP_ENGINE_ID_KEY);
+               if (spoe_encode_buffer(SPOP_ENGINE_ID_KEY, sz, &p, end) == -1)
+                       goto full;
+
+               *p++ = SPOP_DATA_T_STR;
+               sz = strlen(spop_conn->agent->engine_id);
+               if (spoe_encode_buffer(spop_conn->agent->engine_id, sz, &p, end) == -1)
+                       goto full;
+       }
+       /* If there is no agent attached to the opposite endpoint, it means it is a standalone connection.
+        * This only happens for healthchecks
+        */
+       if (!spop_conn->agent) {
+               /* Add "healthcheck" K/V item */
+               sz = SPOP_SLEN(SPOP_HEALTHCHECK_KEY);
+               if (spoe_encode_buffer(SPOP_HEALTHCHECK_KEY, sz, &p, end) == -1)
+                       goto full;
+               *p++ = (SPOP_DATA_T_BOOL | SPOP_DATA_FL_TRUE);
+       }
+
+       outbuf.data += p - b_tail(&outbuf);
+
+       /* update the frame's size now */
+       TRACE_PROTO("SPOP HAPROXY HELLO frame xferred", SPOP_EV_TX_FRAME|SPOP_EV_TX_HELLO, spop_conn->conn, 0, 0, (size_t[]){outbuf.data});
+       spop_set_frame_size(outbuf.area, outbuf.data - 4);
+       b_add(mbuf, outbuf.data);
+       ret = 1;
+
+  end:
+       TRACE_LEAVE(SPOP_EV_TX_FRAME|SPOP_EV_TX_HELLO, spop_conn->conn);
+       return ret;
+  full:
+       /* Too large to be encoded. For HELLO frame, it is an error */
+       if (!b_data(mbuf)) {
+               TRACE_ERROR("HAPROXY HELLO frame too large", SPOP_EV_TX_FRAME|SPOP_EV_TX_HELLO|SPOP_EV_SPOP_CONN_ERR, spop_conn->conn);
+               goto fail;
+       }
+
+       if ((mbuf = br_tail_add(spop_conn->mbuf)) != NULL)
+               goto retry;
+       spop_conn->flags |= SPOP_CF_MUX_MFULL;
+       spop_conn->flags |= SPOP_CF_DEM_MROOM;
+       TRACE_STATE("mbuf ring full", SPOP_EV_TX_FRAME|SPOP_EV_SPOP_CONN_BLK, spop_conn->conn);
+       ret = 0;
+       goto end;
+  fail:
+       spop_conn->state = SPOP_CS_CLOSED;
+       TRACE_STATE("switching to CLOSED", SPOP_EV_TX_FRAME|SPOP_EV_TX_HELLO|SPOP_EV_SPOP_CONN_END, spop_conn->conn);
+       TRACE_DEVEL("leaving on error", SPOP_EV_TX_FRAME|SPOP_EV_TX_HELLO|SPOP_EV_SPOP_CONN_ERR, spop_conn->conn);
+       return 0;
+}
+
+/* Sends an DISCONNECT frame for each active streams. Closed streams are
+ * excluded, as the streams which already received the end-of-stream. It returns
+ * > 0 if the record was sent tp all streams. Otherwise it returns 0.
+ */
+static int spop_conn_send_disconnect(struct spop_conn *spop_conn)
+{
+       const char *reason;
+       struct buffer outbuf;
+       struct buffer *mbuf;
+       char *p, *end;
+       size_t sz;
+       int ret = 0;
+
+       TRACE_ENTER(SPOP_EV_TX_FRAME|SPOP_EV_TX_DISCO, spop_conn->conn);
+
+       mbuf = br_tail(spop_conn->mbuf);
+  retry:
+       if (!spop_get_buf(spop_conn, mbuf)) {
+               spop_conn->flags |= SPOP_CF_MUX_MALLOC;
+               spop_conn->flags |= SPOP_CF_DEM_MROOM;
+               TRACE_STATE("waiting for fconn mbuf ring allocation", SPOP_EV_TX_FRAME|SPOP_EV_SPOP_CONN_BLK, spop_conn->conn);
+               ret = 0;
+               goto end;
+       }
+
+       while (1) {
+               outbuf = b_make(b_tail(mbuf), b_contig_space(mbuf), 0, 0);
+               if (outbuf.size >= 11 || !b_space_wraps(mbuf))
+                       break;
+         realign_again:
+               b_slow_realign(mbuf, trash.area, b_data(mbuf));
+       }
+
+       if (outbuf.size < 11)
+               goto full;
+
+       /* len: 4-bytes (fill later) type: (2)HAPROXY-DISCONNECT, flags: 4-bytes (FIN=1)
+        * stream-id: 0x00, frame-id: 0x00 */
+       memcpy(outbuf.area, "\x00\x00\x00\x00\x02\x00\x00\x00\x01\x00\x00", 11);
+       outbuf.data = 11;
+
+       p = b_tail(&outbuf);
+       end = b_orig(&outbuf) + b_size(&outbuf);
+
+       /* "status-code" K/V item */
+       sz = SPOP_SLEN(SPOP_STATUS_CODE_KEY);
+       if (spoe_encode_buffer(SPOP_STATUS_CODE_KEY, sz, &p, end) == -1)
+               goto full;
+
+       *p++ = SPOP_DATA_T_UINT32;
+       if (encode_varint(spop_conn->errcode, &p, end) == -1)
+               goto full;
+
+       /* "message" K/V item */
+       sz = SPOP_SLEN(SPOP_MSG_KEY);
+       if (spoe_encode_buffer(SPOP_MSG_KEY, sz, &p, end) == -1)
+               goto full;
+
+       /*Get the message corresponding to the status code */
+       reason = spop_err_reasons[spop_conn->errcode];
+
+       *p++ = SPOP_DATA_T_STR;
+       sz = strlen(reason);
+       if (spoe_encode_buffer(reason, sz, &p, end) == -1)
+               goto full;
+
+       outbuf.data += p - b_tail(&outbuf);
+
+       /* update the frame's size now */
+       TRACE_PROTO("SPOP HAPROXY DISCONNECT frame xferred", SPOP_EV_TX_FRAME|SPOP_EV_TX_DISCO, spop_conn->conn, 0, 0, (size_t[]){outbuf.data});
+       spop_set_frame_size(outbuf.area, outbuf.data - 4);
+       b_add(mbuf, outbuf.data);
+       spop_conn->flags |= SPOP_CF_DISCO_SENT;
+       ret = 1;
+
+  end:
+       TRACE_LEAVE(SPOP_EV_TX_FRAME|SPOP_EV_TX_DISCO, spop_conn->conn);
+       return ret;
+  full:
+       /* Too large to be encoded. For DISCONNECT frame, it is an error */
+       if (!b_data(mbuf)) {
+               TRACE_ERROR("SPOP HAPROXY DISCO frame too large", SPOP_EV_TX_FRAME|SPOP_EV_TX_DISCO|SPOP_EV_SPOP_CONN_ERR, spop_conn->conn);
+               goto fail;
+       }
+
+       if ((mbuf = br_tail_add(spop_conn->mbuf)) != NULL)
+               goto retry;
+       spop_conn->flags |= SPOP_CF_MUX_MFULL;
+       spop_conn->flags |= SPOP_CF_DEM_MROOM;
+       TRACE_STATE("mbuf ring full", SPOP_EV_TX_FRAME|SPOP_EV_SPOP_CONN_BLK, spop_conn->conn);
+       ret = 0;
+       goto end;
+  fail:
+       spop_conn->state = SPOP_CS_CLOSED;
+       TRACE_STATE("switching to CLOSED", SPOP_EV_TX_FRAME|SPOP_EV_TX_DISCO|SPOP_EV_SPOP_CONN_END, spop_conn->conn);
+       TRACE_DEVEL("leaving on error", SPOP_EV_TX_FRAME|SPOP_EV_TX_DISCO|SPOP_EV_SPOP_CONN_ERR, spop_conn->conn);
+       return 0;
+}
+
+/* Processes an AGENT HELLO frame. Returns > 0 on success, 0 if it couldn't do
+ * anything. It is highly unexpected, but if the frame is larger than a buffer
+ * and cannot be decoded in one time, an error is triggered and the connection
+ * is closed. HELLO frame cannot be split.
+ */
+static int spop_conn_handle_hello(struct spop_conn *spop_conn)
+{
+       struct buffer *dbuf;
+       char *p, *end;
+       int vsn, max_frame_size;
+       unsigned int flags;
+
+       TRACE_ENTER(SPOP_EV_RX_FRAME|SPOP_EV_RX_HELLO, spop_conn->conn);
+
+       dbuf = &spop_conn->dbuf;
+
+       /* Record too large to be fully decoded */
+       if (b_size(dbuf) < (spop_conn->dfl))
+               goto fail;
+
+       /* process full record only */
+       if (b_data(dbuf) < (spop_conn->dfl)) {
+               TRACE_DEVEL("leaving on missing data", SPOP_EV_RX_FRAME|SPOP_EV_RX_HELLO, spop_conn->conn);
+               return 0;
+       }
+
+       if (unlikely(b_contig_data(dbuf, b_head_ofs(dbuf)) < spop_conn->dfl)) {
+               /* Realign the dmux buffer if the frame wraps. It is unexpected
+                * at this stage because it should be the first record received
+                * from the FCGI application.
+                */
+               b_slow_realign_ofs(dbuf, trash.area, 0);
+       }
+
+       p = b_head(dbuf);
+       end = p  + spop_conn->dfl;
+
+       /* There are 3 mandatory items: "version", "max-frame-size" and
+        * "capabilities" */
+
+       /* Loop on K/V items */
+       vsn = max_frame_size = flags = 0;
+       while (p < end) {
+               char  *str;
+               uint64_t sz;
+               int    ret;
+
+               /* Decode the item key */
+               ret = spoe_decode_buffer(&p, end, &str, &sz);
+               if (ret == -1 || !sz) {
+                       spop_conn_error(spop_conn, SPOP_ERR_INVALID);
+                       goto fail;
+               }
+
+               /* Check "version" K/V item */
+               if (sz >= strlen(SPOP_VERSION_KEY) && !memcmp(str, SPOP_VERSION_KEY, strlen(SPOP_VERSION_KEY))) {
+                       int i, type = *p++;
+
+                       /* The value must be a string */
+                       if ((type & SPOP_DATA_T_MASK) != SPOP_DATA_T_STR) {
+                               spop_conn_error(spop_conn, SPOP_ERR_INVALID);
+                               goto fail;
+                       }
+                       if (spoe_decode_buffer(&p, end, &str, &sz) == -1) {
+                               spop_conn_error(spop_conn, SPOP_ERR_INVALID);
+                               goto fail;
+                       }
+
+                       vsn = spoe_str_to_vsn(str, sz);
+                       if (vsn == -1) {
+                               spop_conn_error(spop_conn, SPOP_ERR_BAD_VSN);
+                               goto fail;
+                       }
+                       for (i = 0; spop_supported_versions[i].str != NULL; ++i) {
+                               if (vsn >= spop_supported_versions[i].min &&
+                                   vsn <= spop_supported_versions[i].max)
+                                       break;
+                       }
+                       if (spop_supported_versions[i].str == NULL) {
+                               spop_conn_error(spop_conn, SPOP_ERR_BAD_VSN);
+                               goto fail;
+                       }
+               }
+               /* Check "max-frame-size" K/V item */
+               else if (sz >= strlen(SPOP_MAX_FRAME_SIZE_KEY) && !memcmp(str, SPOP_MAX_FRAME_SIZE_KEY, SPOP_SLEN(SPOP_MAX_FRAME_SIZE_KEY))) {
+                       int type = *p++;
+
+                       /* The value must be integer */
+                       if ((type & SPOP_DATA_T_MASK) != SPOP_DATA_T_INT32 &&
+                           (type & SPOP_DATA_T_MASK) != SPOP_DATA_T_INT64 &&
+                           (type & SPOP_DATA_T_MASK) != SPOP_DATA_T_UINT32 &&
+                           (type & SPOP_DATA_T_MASK) != SPOP_DATA_T_UINT64) {
+                               spop_conn_error(spop_conn, SPOP_ERR_INVALID);
+                               goto fail;
+                       }
+                       if (decode_varint(&p, end, &sz) == -1) {
+                               spop_conn_error(spop_conn, SPOP_ERR_INVALID);
+                               goto fail;
+                       }
+                       if (sz < SPOP_MIN_FRAME_SIZE ||
+                           sz > SPOP_MAX_FRAME_SIZE/* agent->max_frame_size */) {
+                               spop_conn_error(spop_conn, SPOP_ERR_BAD_FRAME_SIZE);
+                               goto fail;
+                       }
+                       max_frame_size = sz;
+               }
+               /* Check "capabilities" K/V item */
+               else if (sz >= strlen(SPOP_CAPABILITIES_KEY) && !memcmp(str, SPOP_CAPABILITIES_KEY, SPOP_SLEN(SPOP_CAPABILITIES_KEY))) {
+                       int type = *p++;
+
+                       /* The value must be a string */
+                       if ((type & SPOP_DATA_T_MASK) != SPOP_DATA_T_STR) {
+                               spop_conn_error(spop_conn, SPOP_ERR_INVALID);
+                               goto fail;
+                       }
+                       if (spoe_decode_buffer(&p, end, &str, &sz) == -1) {
+                               spop_conn_error(spop_conn, SPOP_ERR_INVALID);
+                               goto fail;
+                       }
+
+                       while (sz) {
+                               char *delim;
+
+                               /* Skip leading spaces */
+                               for (; isspace((unsigned char)*str) && sz; str++, sz--);
+
+                               /* if (sz >= 10 && !strncmp(str, "pipelining", 10)) { */
+                               /*      str += 10; sz -= 10; */
+                               /*      if (!sz || isspace((unsigned char)*str) || *str == ',') */
+                               /*              flags |= SPOE_APPCTX_FL_PIPELINING; */
+                               /* } */
+
+                               /* Get the next comma or break */
+                               if (!sz || (delim = memchr(str, ',', sz)) == NULL)
+                                       break;
+                               delim++;
+                               sz -= (delim - str);
+                               str = delim;
+                       }
+               }
+               else {
+                       /* Silently ignore unknown item */
+                       if (spoe_skip_data(&p, end) == -1) {
+                               spop_conn_error(spop_conn, SPOP_ERR_INVALID);
+                               goto fail;
+                       }
+               }
+       }
+
+       /* Final checks */
+       if (!vsn) {
+               spop_conn_error(spop_conn, SPOP_ERR_NO_VSN);
+               goto fail;
+       }
+       if (!max_frame_size) {
+               spop_conn_error(spop_conn, SPOP_ERR_NO_FRAME_SIZE);
+               goto fail;
+       }
+       /* if (!agent) */
+       /*      flags &= ~SPOE_APPCTX_FL_PIPELINING; */
+       /* else { */
+       /*      if ((flags & SPOE_APPCTX_FL_PIPELINING) && !(agent->flags & SPOE_FL_PIPELINING)) */
+       /*              flags &= ~SPOE_APPCTX_FL_PIPELINING; */
+       /* } */
+
+       /* SPOE_APPCTX(appctx)->version        = (unsigned int)vsn; */
+       /* SPOE_APPCTX(appctx)->max_frame_size = (unsigned int)max_frame_size; */
+       /* SPOE_APPCTX(appctx)->flags         |= flags; */
+
+
+       TRACE_PROTO("SPOP AGENT HELLO frame rcvd", SPOP_EV_RX_FRAME|SPOP_EV_RX_HELLO, spop_conn->conn, 0, 0, (size_t[]){spop_conn->dfl});
+       b_del(&spop_conn->dbuf, spop_conn->dfl);
+       spop_conn->dfl = 0;
+       spop_wake_unassigned_streams(spop_conn);
+       TRACE_LEAVE(SPOP_EV_RX_FRAME|SPOP_EV_RX_HELLO, spop_conn->conn);
+       return 1;
+  fail:
+       spop_conn->state = SPOP_CS_CLOSED;
+       spop_conn->flags |= SPOP_CF_ERROR;
+       TRACE_STATE("switching to CLOSED", SPOP_EV_RX_FRAME|SPOP_EV_RX_HELLO, spop_conn->conn);
+       TRACE_DEVEL("leaving on error", SPOP_EV_RX_FRAME|SPOP_EV_RX_HELLO|SPOP_EV_SPOP_CONN_ERR, spop_conn->conn);
+       return 0;
+}
+
+/* Processes an AGENT DISCONNECT frame. Returns > 0 on success, 0 if it couldn't
+ * do anything. It is highly unexpected, but if the frame is larger than a
+ * buffer and cannot be decoded in one time, an error is triggered and the
+ * connection is closed. DISCONNECT frame cannot be split.
+ */
+static int spop_conn_handle_disconnect(struct spop_conn *spop_conn)
+{
+       struct buffer *dbuf;
+       char *p, *end;
+       //struct ist reason;
+       unsigned int status_code, flags;
+
+       TRACE_ENTER(SPOP_EV_RX_FRAME|SPOP_EV_RX_DISCO, spop_conn->conn);
+
+       dbuf = &spop_conn->dbuf;
+
+       /* Record too large to be fully decoded */
+       if (b_size(dbuf) < (spop_conn->dfl))
+               goto fail;
+
+       /* process full record only */
+       if (b_data(dbuf) < (spop_conn->dfl)) {
+               TRACE_DEVEL("leaving on missing data", SPOP_EV_RX_FRAME|SPOP_EV_RX_DISCO, spop_conn->conn);
+               return 0;
+       }
+
+       if (unlikely(b_contig_data(dbuf, b_head_ofs(dbuf)) < spop_conn->dfl)) {
+               /* Realign the dmux buffer if the frame wraps. It is unexpected
+                * at this stage because it should be the first record received
+                * from the FCGI application.
+                */
+               b_slow_realign_ofs(dbuf, trash.area, 0);
+       }
+
+       p = b_head(dbuf);
+       end = p  + spop_conn->dfl;
+
+        /* There are 2 mandatory items: "status-code" and "message" */
+
+       /* Loop on K/V items */
+       status_code = flags = 0;
+       while (p < end) {
+               char  *str;
+               uint64_t sz;
+               int    ret;
+
+               /* Decode the item key */
+               ret = spoe_decode_buffer(&p, end, &str, &sz);
+               if (ret == -1 || !sz) {
+                       spop_conn_error(spop_conn, SPOP_ERR_INVALID);
+                       goto fail;
+               }
+
+                /* Check "status-code" K/V item */
+                if (sz >= strlen(SPOP_STATUS_CODE_KEY) && !memcmp(str, SPOP_STATUS_CODE_KEY, strlen(SPOP_STATUS_CODE_KEY))) {
+                        int type = *p++;
+
+                        /* The value must be an integer */
+                        if ((type & SPOP_DATA_T_MASK) != SPOP_DATA_T_INT32 &&
+                            (type & SPOP_DATA_T_MASK) != SPOP_DATA_T_INT64 &&
+                            (type & SPOP_DATA_T_MASK) != SPOP_DATA_T_UINT32 &&
+                            (type & SPOP_DATA_T_MASK) != SPOP_DATA_T_UINT64) {
+                               spop_conn_error(spop_conn, SPOP_ERR_INVALID);
+                               goto fail;
+                        }
+                        if (decode_varint(&p, end, &sz) == -1) {
+                               spop_conn_error(spop_conn, SPOP_ERR_INVALID);
+                               goto fail;
+                        }
+                        status_code = sz;
+                }
+               // TODO: for now skip the reason
+                /* /\* Check "message" K/V item *\/ */
+                /* else if (sz >= strlen(SPOP_MSG_KEY) && !memcmp(str, SPOP_MSG_KEY, strlen(SPOP_MSG_KEY))) { */
+                /*         int type = *p++; */
+
+                /*         /\* The value must be a string *\/ */
+                /*         if ((type & SPOP_DATA_T_MASK) != SPOP_DATA_T_STR) { */
+               /*              spop_conn_error(spop_conn, SPOP_ERR_INVALID); */
+               /*              goto fail; */
+                /*         } */
+                /*         ret = spoe_decode_buffer(&p, end, &str, &sz); */
+                /*         if (ret == -1 || sz > 255) { */
+               /*              spop_conn_error(spop_conn, SPOP_ERR_INVALID); */
+               /*              goto fail; */
+                /*         } */
+               /*      reason = ist2(str, sz); */
+                /* } */
+                else {
+                        /* Silently ignore unknown item */
+                        if (spoe_skip_data(&p, end) == -1) {
+                               spop_conn_error(spop_conn, SPOP_ERR_INVALID);
+                               goto fail;
+                        }
+                }
+       }
+
+       TRACE_PROTO("SPOP AGENT DISCONNECT frame rcvd", SPOP_EV_RX_FRAME|SPOP_EV_RX_DISCO, spop_conn->conn, 0, 0, (size_t[]){spop_conn->dfl});
+       b_del(&spop_conn->dbuf, spop_conn->dfl);
+       spop_conn->dfl = 0;
+       spop_conn_error(spop_conn, status_code);
+       spop_conn->state = SPOP_CS_CLOSED;
+       spop_wake_some_streams(spop_conn, 0/*last*/);
+       TRACE_LEAVE(SPOP_EV_RX_FRAME|SPOP_EV_RX_DISCO, spop_conn->conn);
+       return 1;
+  fail:
+       spop_conn->state = SPOP_CS_CLOSED;
+       spop_conn->flags |= SPOP_CF_ERROR;
+       TRACE_STATE("switching to CLOSED", SPOP_EV_RX_FRAME|SPOP_EV_RX_DISCO, spop_conn->conn);
+       TRACE_DEVEL("leaving on error", SPOP_EV_RX_FRAME|SPOP_EV_RX_DISCO|SPOP_EV_SPOP_CONN_ERR, spop_conn->conn);
+       return 0;
+}
+
+/* Processes an AGENT ACK frame. Returns > 0 on success, 0 if it couldn't
+ * do anything.
+ *
+ * TODO: for now cannot be fragmented.
+ */
+static int spop_conn_handle_ack(struct spop_conn *spop_conn, struct spop_strm *spop_strm)
+{
+       struct buffer *dbuf, *rxbuf;
+       unsigned int flen, sent = 0;
+
+       TRACE_ENTER(SPOP_EV_RX_FRAME|SPOP_EV_RX_ACK, spop_conn->conn, spop_strm);
+
+       dbuf = &spop_conn->dbuf;
+
+       /* Record too large to be fully decoded */
+       if (b_size(dbuf) < (spop_conn->dfl))
+               goto fail;
+
+       /* process full record only */
+       if (b_data(dbuf) < (spop_conn->dfl)) {
+               TRACE_DEVEL("leaving on missing data", SPOP_EV_RX_FRAME|SPOP_EV_RX_DISCO, spop_conn->conn);
+               return 0;
+       }
+
+       if (unlikely(b_contig_data(dbuf, b_head_ofs(dbuf)) < spop_conn->dfl)) {
+               /* Realign the dmux buffer if the frame wraps. It is unexpected
+                * at this stage because it should be the first record received
+                * from the FCGI application.
+                */
+               b_slow_realign_ofs(dbuf, trash.area, 0);
+       }
+
+       spop_conn->flags &= ~SPOP_CF_DEM_SFULL;
+       rxbuf = spop_get_buf(spop_conn, &spop_strm->rxbuf);
+       if (!rxbuf) {
+               spop_conn->flags |= SPOP_CF_DEM_SALLOC;
+               TRACE_STATE("waiting for an spop_strm rxbuf", SPOP_EV_RX_FRAME|SPOP_EV_RX_ACK|SPOP_EV_SPOP_STRM_BLK, spop_conn->conn, spop_strm);
+               goto fail;
+       }
+
+       flen = spop_conn->dfl;
+       if (!flen)
+               goto end;
+
+       // TODO: For now we know all data were received
+       /* if (flen > b_data(&h2c->dbuf)) { */
+       /*      flen = b_data(&h2c->dbuf); */
+       /*      if (!flen) */
+       /*              goto fail; */
+       /* } */
+
+       // TODO: for now, we take care to xfer all data at once !
+       if (flen > b_room(rxbuf)) {
+               spop_conn->flags |= SPOP_CF_DEM_SFULL;
+               TRACE_STATE("spop_strm rxbuf is full", SPOP_EV_RX_FRAME|SPOP_EV_RX_ACK|SPOP_EV_SPOP_STRM_BLK, spop_conn->conn, spop_strm);
+               goto fail;
+       }
+
+       sent = b_xfer(rxbuf, dbuf, flen);
+       BUG_ON(sent != flen);
+       /* b_del(&spop_conn->dbuf, sent); */
+       spop_conn->dfl -= sent;
+
+       // TODO: may happen or not ?
+       /* /\* call the upper layers to process the frame, then let the upper layer */
+       /*  * notify the stream about any change. */
+       /*  *\/ */
+       /* if (!spop_strm_sc(spop_strm)) { */
+       /*      /\* The upper layer has already closed *\/ */
+
+       /* } */
+       if (spop_strm->state == SPOP_SS_OPEN)
+               spop_strm->state = SPOP_SS_HREM;
+       else
+               spop_strm_close(spop_strm);
+
+  end:
+       TRACE_PROTO("SPOP AGENT ACK frame rcvd", SPOP_EV_RX_FRAME|SPOP_EV_RX_ACK, spop_conn->conn, spop_strm, 0, (size_t[]){sent});
+       spop_conn->state = SPOP_CS_FRAME_H;
+       TRACE_LEAVE(SPOP_EV_RX_FRAME|SPOP_EV_RX_ACK, spop_conn->conn, spop_strm);
+       return 1;
+
+ fail:
+       TRACE_DEVEL("leaving on error", SPOP_EV_RX_FRAME|SPOP_EV_RX_ACK|SPOP_EV_SPOP_CONN_ERR, spop_conn->conn, spop_strm);
+       return 0;
+}
+
+/* resume each spop_strm eligible for sending in list head <head> */
+static void spop_resume_each_sending_spop_strm(struct spop_conn *spop_conn, struct list *head)
+{
+       struct spop_strm *spop_strm, *spop_strm_back;
+
+       TRACE_ENTER(SPOP_EV_SPOP_CONN_SEND|SPOP_EV_STRM_WAKE, spop_conn->conn);
+
+       list_for_each_entry_safe(spop_strm, spop_strm_back, head, list) {
+               if (spop_conn->flags & SPOP_CF_MUX_BLOCK_ANY ||
+                   spop_conn->state >= SPOP_CS_ERROR)
+                       break;
+
+               spop_strm->flags &= ~SPOP_SF_BLK_ANY;
+
+               if (spop_strm->flags & SPOP_SF_NOTIFIED)
+                       continue;
+
+               /* If the sender changed his mind and unsubscribed, let's just
+                * remove the stream from the send_list.
+                */
+               if (!spop_strm->subs || !(spop_strm->subs->events & SUB_RETRY_SEND)) {
+                       LIST_DEL_INIT(&spop_strm->list);
+                       continue;
+               }
+
+               if (spop_strm->subs && spop_strm->subs->events & SUB_RETRY_SEND) {
+                       spop_strm->flags |= SPOP_SF_NOTIFIED;
+                       tasklet_wakeup(spop_strm->subs->tasklet);
+                       spop_strm->subs->events &= ~SUB_RETRY_SEND;
+                       if (!spop_strm->subs->events)
+                               spop_strm->subs = NULL;
+               }
+       }
+
+       TRACE_LEAVE(SPOP_EV_SPOP_CONN_SEND|SPOP_EV_STRM_WAKE, spop_conn->conn);
+}
+
+/* removes a stream from the list it may be in. If a stream has recently been
+ * appended to the send_list, it might have been waiting on this one when
+ * entering spop_snd_buf() and expecting it to complete before starting to send
+ * in turn. For this reason we check (and clear) SPOP_CF_WAIT_INLIST to detect
+ * this condition, and we try to resume sending streams if it happens. Note
+ * that we don't need to do it for fctl_list as this list is relevant before
+ * (only consulted after) a window update on the connection, and not because
+ * of any competition with other streams.
+ */
+static inline void spop_remove_from_list(struct spop_strm *spop_strm)
+{
+       struct spop_conn *spop_conn = spop_strm->spop_conn;
+
+       if (!LIST_INLIST(&spop_strm->list))
+               return;
+
+       LIST_DEL_INIT(&spop_strm->list);
+       if (spop_conn->flags & SPOP_CF_WAIT_INLIST) {
+               spop_conn->flags &= ~SPOP_CF_WAIT_INLIST;
+               spop_resume_each_sending_spop_strm(spop_conn, &spop_conn->send_list);
+       }
+}
+
+
+/* process Rx records to be demultiplexed */
+static void spop_process_demux(struct spop_conn *spop_conn)
+{
+       struct spop_strm *spop_strm = NULL, *tmp_spop_strm;
+       struct spop_frame_header hdr;
+
+       TRACE_ENTER(SPOP_EV_SPOP_CONN_WAKE, spop_conn->conn);
+
+       if (spop_conn->state >= SPOP_CS_ERROR)
+               goto out;
+
+       if (unlikely(spop_conn->state < SPOP_CS_FRAME_H)) {
+               if (spop_conn->state == SPOP_CS_HA_HELLO) {
+                       TRACE_STATE("waiting AGENT HELLO frame to be sent", SPOP_EV_RX_FRAME|SPOP_EV_RX_FHDR|SPOP_EV_RX_HELLO, spop_conn->conn);
+                       goto out;
+               }
+               if (spop_conn->state == SPOP_CS_AGENT_HELLO) {
+                       /* ensure that what is pending is a valid AGENT HELLO frame. */
+                       TRACE_STATE("receiving AGENT HELLO frame header", SPOP_EV_RX_FRAME|SPOP_EV_RX_FHDR, spop_conn->conn);
+                       if (!spop_get_frame_hdr(&spop_conn->dbuf, &hdr)) {
+                               spop_conn->flags |= SPOP_CF_DEM_SHORT_READ;
+                               TRACE_ERROR("header frame not available yet", SPOP_EV_RX_FRAME|SPOP_EV_RX_FHDR, spop_conn->conn);
+                               goto done;
+                       }
+
+                       if (hdr.sid || hdr.fid || hdr.type != SPOP_FRM_T_AGENT_HELLO || !(hdr.flags & SPOP_FRM_FL_FIN)) {
+                               spop_conn_error(spop_conn, SPOP_ERR_INVALID);
+                               spop_conn->state = SPOP_CS_CLOSED;
+                               TRACE_ERROR("unexpected frame type or flags", SPOP_EV_RX_FRAME|SPOP_EV_RX_FHDR|SPOP_EV_RX_HELLO|SPOP_EV_SPOP_CONN_ERR, spop_conn->conn);
+                               TRACE_STATE("switching to CLOSED", SPOP_EV_RX_FRAME|SPOP_EV_RX_FHDR|SPOP_EV_RX_HELLO|SPOP_EV_SPOP_CONN_ERR, spop_conn->conn);
+                               goto done;
+                       }
+
+                       if ((int)hdr.len < 0 || (int)hdr.len > SPOP_MAX_FRAME_SIZE) {
+                               spop_conn_error(spop_conn, SPOP_ERR_BAD_FRAME_SIZE);
+                               spop_conn->state = SPOP_CS_CLOSED;
+                               TRACE_ERROR("invalid AGENT HELLO frame length", SPOP_EV_RX_FRAME|SPOP_EV_RX_FHDR|SPOP_EV_RX_HELLO|SPOP_EV_SPOP_CONN_ERR, spop_conn->conn);
+                               TRACE_STATE("switching to CLOSED", SPOP_EV_RX_FRAME|SPOP_EV_RX_FHDR|SPOP_EV_RX_HELLO|SPOP_EV_SPOP_CONN_ERR, spop_conn->conn);
+                               goto done;
+                       }
+
+                       goto new_frame;
+               }
+       }
+
+       /* process as many incoming frames as possible below */
+       while (1) {
+               int ret = 0;
+
+               if (!b_data(&spop_conn->dbuf)) {
+                       TRACE_DEVEL("no more Rx data", SPOP_EV_RX_FRAME, spop_conn->conn);
+                       spop_conn->flags |= SPOP_CF_DEM_SHORT_READ;
+                       break;
+               }
+
+               if (spop_conn->state >= SPOP_CS_ERROR) {
+                       TRACE_STATE("end of connection reported", SPOP_EV_RX_FRAME|SPOP_EV_RX_EOI, spop_conn->conn);
+                       break;
+               }
+
+               if (spop_conn->state == SPOP_CS_FRAME_H) {
+                       TRACE_PROTO("receiving SPOP frame header", SPOP_EV_RX_FRAME|SPOP_EV_RX_FHDR, spop_conn->conn);
+                       if (!spop_get_frame_hdr(&spop_conn->dbuf, &hdr)) {
+                               spop_conn->flags |= SPOP_CF_DEM_SHORT_READ;
+                               break;
+                       }
+
+                       if ((int)hdr.len < 0 || (int)hdr.len > spop_conn->agent->max_frame_size) {
+                               spop_conn_error(spop_conn, SPOP_ERR_BAD_FRAME_SIZE);
+                               TRACE_ERROR("invalid SPOP frame length", SPOP_EV_RX_FRAME|SPOP_EV_RX_FHDR|SPOP_EV_SPOP_CONN_ERR, spop_conn->conn);
+                               TRACE_STATE("switching to CLOSED", SPOP_EV_RX_FRAME|SPOP_EV_RX_FHDR|SPOP_EV_SPOP_CONN_ERR, spop_conn->conn);
+                               break;
+                       }
+
+                 new_frame:
+                       spop_conn->dsi = hdr.sid;
+                       spop_conn->dfi = hdr.fid;
+                       spop_conn->dft = hdr.type;
+                       spop_conn->dfl = hdr.len;
+                       spop_conn->dff = hdr.flags;
+                       spop_conn->state = SPOP_CS_FRAME_P;
+                       TRACE_STATE("SPOP frame header rcvd, switching to FRAME_P", SPOP_EV_RX_FRAME|SPOP_EV_RX_FHDR, spop_conn->conn);
+
+                       // TODO: Perform some check (flags & FIN)
+               }
+
+
+               tmp_spop_strm = spop_conn_st_by_id(spop_conn, spop_conn->dsi);
+               if (tmp_spop_strm != spop_strm && spop_strm && spop_strm_sc(spop_strm) &&
+                   (b_data(&spop_strm->rxbuf) ||
+                    spop_conn_read0_pending(spop_conn) ||
+                    spop_strm->state == SPOP_SS_CLOSED ||
+                    se_fl_test(spop_strm->sd, SE_FL_ERROR | SE_FL_ERR_PENDING | SE_FL_EOS))) {
+                       /* we may have to signal the upper layers */
+                       TRACE_DEVEL("notifying stream before switching SID", SPOP_EV_RX_FRAME|SPOP_EV_STRM_WAKE, spop_conn->conn, spop_strm);
+                       se_fl_set(spop_strm->sd, SE_FL_RCV_MORE);
+                       spop_strm_notify_recv(spop_strm);
+               }
+               spop_strm = tmp_spop_strm;
+
+               // check stream for frame error ?
+
+               switch (spop_conn->dft) {
+               case SPOP_FRM_T_AGENT_HELLO:
+                       if (spop_conn->state == SPOP_CS_FRAME_P)
+                               ret = spop_conn_handle_hello(spop_conn);
+                       break;
+               case SPOP_FRM_T_AGENT_DISCON:
+                       if (spop_conn->state == SPOP_CS_FRAME_P)
+                               ret = spop_conn_handle_disconnect(spop_conn);
+                       break;
+               case SPOP_FRM_T_AGENT_ACK:
+                       if (spop_conn->state == SPOP_CS_FRAME_P)
+                               ret = spop_conn_handle_ack(spop_conn, spop_strm);
+                       break;
+               default:
+                       TRACE_PROTO("receiving SPOP ignored frame", SPOP_EV_RX_FRAME, spop_conn->conn, spop_strm);
+                       /* drop frames that we ignore. */
+                       ret = MIN(b_data(&spop_conn->dbuf), spop_conn->dfl);
+                       b_del(&spop_conn->dbuf, ret);
+                       spop_conn->dfl -= ret;
+                       ret = (spop_conn->dfl == 0);
+                       break;
+               }
+
+               // TODO: SS_ERROR to CS_ERROR ?
+
+               if (spop_conn->state == SPOP_CS_ERROR) {
+                       TRACE_PROTO("sending SPOP HAPROXY DISCONNECT frame", SPOP_EV_TX_FRAME|SPOP_EV_TX_DISCO|SPOP_EV_TX_EOI, spop_conn->conn, spop_strm);
+                       ret = spop_conn_send_disconnect(spop_conn);
+               }
+
+               /* error or missing data condition met above ? */
+               if (ret <= 0)
+                       break;
+
+               if (spop_conn->state != SPOP_CS_FRAME_H) {
+                       if (spop_conn->dfl) {
+                               TRACE_DEVEL("skipping remaining frame payload", SPOP_EV_RX_FRAME, spop_conn->conn, spop_strm);
+                               ret = MIN(b_data(&spop_conn->dbuf), spop_conn->dfl);
+                               b_del(&spop_conn->dbuf, ret);
+                               spop_conn->dfl -= ret;
+                       }
+                       if (!spop_conn->dfl) {
+                               TRACE_STATE("switching to FRAME_H", SPOP_EV_RX_FRAME|SPOP_EV_RX_FHDR, spop_conn->conn);
+                               spop_conn->state = SPOP_CS_FRAME_H;
+                       }
+               }
+       }
+
+  done:
+       if (spop_conn->state >= SPOP_CS_ERROR || (spop_conn->flags & SPOP_CF_DEM_SHORT_READ)) {
+               if (spop_conn->flags & SPOP_CF_RCVD_SHUT)
+                       spop_conn->flags |= SPOP_CF_END_REACHED;
+       }
+
+       if (spop_strm && spop_strm_sc(spop_strm) &&
+           (b_data(&spop_strm->rxbuf) ||
+            spop_conn_read0_pending(spop_conn) ||
+            spop_strm->state == SPOP_SS_CLOSED ||
+            se_fl_test(spop_strm->sd, SE_FL_ERROR | SE_FL_ERR_PENDING | SE_FL_EOS))) {
+               /* we may have to signal the upper layers */
+               TRACE_DEVEL("notifying stream before switching SID", SPOP_EV_RX_FRAME|SPOP_EV_STRM_WAKE, spop_conn->conn, spop_strm);
+               se_fl_set(spop_strm->sd, SE_FL_RCV_MORE);
+               spop_strm_notify_recv(spop_strm);
+       }
+
+       spop_conn_restart_reading(spop_conn, 0);
+  out:
+       TRACE_LEAVE(SPOP_EV_SPOP_CONN_WAKE, spop_conn->conn);
+       return;
+}
+
+
+/* process Tx records from streams to be multiplexed. Returns > 0 if it reached
+ * the end.
+ */
+static int spop_process_mux(struct spop_conn *spop_conn)
+{
+       TRACE_ENTER(SPOP_EV_SPOP_CONN_WAKE, spop_conn->conn);
+
+       if (unlikely(spop_conn->state < SPOP_CS_FRAME_H)) {
+               if (unlikely(spop_conn->state == SPOP_CS_HA_HELLO)) {
+                       TRACE_PROTO("sending SPOP HAPROXY HELLO fraame", SPOP_EV_TX_FRAME, spop_conn->conn);
+                       if (unlikely(!spop_conn_send_hello(spop_conn)))
+                               goto fail;
+                       spop_conn->state = SPOP_CS_AGENT_HELLO;
+                       TRACE_STATE("waiting for SPOP AGENT HELLO reply", SPOP_EV_TX_FRAME|SPOP_EV_RX_FRAME, spop_conn->conn);
+               }
+               /* need to wait for the other side */
+               if (spop_conn->state < SPOP_CS_FRAME_H)
+                       goto done;
+       }
+
+       spop_conn->flags &= ~SPOP_CF_WAIT_INLIST;
+       spop_resume_each_sending_spop_strm(spop_conn, &spop_conn->send_list);
+
+  fail:
+       /* Nothing to do */
+
+  done:
+       TRACE_LEAVE(SPOP_EV_SPOP_CONN_WAKE, spop_conn->conn);
+       return 1;
+}
+
+/* Attempt to read data, and subscribe if none available.
+ * The function returns 1 if data has been received, otherwise zero.
+ */
+static int spop_recv(struct spop_conn *spop_conn)
+{
+       struct connection *conn = spop_conn->conn;
+       struct buffer *buf;
+       int max;
+       size_t ret;
+
+       TRACE_ENTER(SPOP_EV_SPOP_CONN_RECV, conn);
+
+       if (spop_conn->wait_event.events & SUB_RETRY_RECV) {
+               TRACE_DEVEL("leaving on sub_recv", SPOP_EV_SPOP_CONN_RECV, conn);
+               return (b_data(&spop_conn->dbuf));
+       }
+
+       if (!spop_recv_allowed(spop_conn)) {
+               TRACE_DEVEL("leaving on !recv_allowed", SPOP_EV_SPOP_CONN_RECV, conn);
+               return 1;
+       }
+
+       buf = spop_get_buf(spop_conn, &spop_conn->dbuf);
+       if (!buf) {
+               TRACE_DEVEL("waiting for spop_conn dbuf allocation", SPOP_EV_SPOP_CONN_RECV|SPOP_EV_SPOP_CONN_BLK, conn);
+               spop_conn->flags |= SPOP_CF_DEM_DALLOC;
+               return 0;
+       }
+
+       max = b_room(buf);
+       ret = max ? conn->xprt->rcv_buf(conn, conn->xprt_ctx, buf, max, 0) : 0;
+
+       if (max && !ret && spop_recv_allowed(spop_conn)) {
+               TRACE_DATA("failed to receive data, subscribing", SPOP_EV_SPOP_CONN_RECV, conn);
+               conn->xprt->subscribe(conn, conn->xprt_ctx, SUB_RETRY_RECV, &spop_conn->wait_event);
+       }
+       else
+               TRACE_DATA("recv data", SPOP_EV_SPOP_CONN_RECV, conn, 0, 0, (size_t[]){ret});
+
+       if (conn_xprt_read0_pending(conn)) {
+               TRACE_DATA("received read0", SPOP_EV_SPOP_CONN_RECV, conn);
+               spop_conn->flags |= SPOP_CF_RCVD_SHUT;
+       }
+       if (conn->flags & CO_FL_ERROR) {
+               TRACE_DATA("connection error", SPOP_EV_SPOP_CONN_RECV, conn);
+               spop_conn->flags |= SPOP_CF_ERROR;
+       }
+
+       if (!b_data(buf)) {
+               spop_release_buf(spop_conn, &spop_conn->dbuf);
+               goto end;
+       }
+
+       if (ret == max) {
+               TRACE_DEVEL("spop_conn dbuf full", SPOP_EV_SPOP_CONN_RECV|SPOP_EV_SPOP_CONN_BLK, conn);
+               spop_conn->flags |= SPOP_CF_DEM_DFULL;
+       }
+
+end:
+       TRACE_LEAVE(SPOP_EV_SPOP_CONN_RECV, conn);
+       return !!ret || (spop_conn->flags & (SPOP_CF_RCVD_SHUT|SPOP_CF_ERROR));
+}
+
+/* Try to send data if possible.
+ * The function returns 1 if data have been sent, otherwise zero.
+ */
+static int spop_send(struct spop_conn *spop_conn)
+{
+       struct connection *conn = spop_conn->conn;
+       int done;
+       int sent = 0;
+
+       TRACE_ENTER(SPOP_EV_SPOP_CONN_SEND, conn);
+
+       if (spop_conn->flags & (SPOP_CF_ERROR|SPOP_CF_ERR_PENDING)) {
+               TRACE_DEVEL("leaving on connection error", SPOP_EV_SPOP_CONN_SEND, conn);
+               if (spop_conn->flags & SPOP_CF_END_REACHED)
+                       spop_conn->flags |= SPOP_CF_ERROR;
+               b_reset(br_tail(spop_conn->mbuf));
+               return 1;
+       }
+
+
+       if (conn->flags & CO_FL_WAIT_XPRT) {
+               /* a handshake was requested */
+               goto schedule;
+       }
+
+       /* This loop is quite simple : it tries to fill as much as it can from
+        * pending streams into the existing buffer until it's reportedly full
+        * or the end of send requests is reached. Then it tries to send this
+        * buffer's contents out, marks it not full if at least one byte could
+        * be sent, and tries again.
+        *
+        * The snd_buf() function normally takes a "flags" argument which may
+        * be made of a combination of CO_SFL_MSG_MORE to indicate that more
+        * data immediately comes and CO_SFL_STREAMER to indicate that the
+        * connection is streaming lots of data (used to increase TLS record
+        * size at the expense of latency). The former can be sent any time
+        * there's a buffer full flag, as it indicates at least one stream
+        * attempted to send and failed so there are pending data. An
+        * alternative would be to set it as long as there's an active stream
+        * but that would be problematic for ACKs until we have an absolute
+        * guarantee that all waiters have at least one byte to send. The
+        * latter should possibly not be set for now.
+        */
+
+       done = 0;
+       while (!done) {
+               unsigned int flags = 0;
+               unsigned int released = 0;
+               struct buffer *buf;
+
+               /* fill as much as we can into the current buffer */
+               while (((spop_conn->flags & (SPOP_CF_MUX_MFULL|SPOP_CF_MUX_MALLOC)) == 0) && !done)
+                       done = spop_process_mux(spop_conn);
+
+               if (spop_conn->flags & SPOP_CF_MUX_MALLOC)
+                       done = 1; // we won't go further without extra buffers
+
+               if (conn->flags & CO_FL_ERROR)
+                       break;
+
+               if (spop_conn->flags & (SPOP_CF_MUX_MFULL | SPOP_CF_DEM_MROOM))
+                       flags |= CO_SFL_MSG_MORE;
+
+               for (buf = br_head(spop_conn->mbuf); b_size(buf); buf = br_del_head(spop_conn->mbuf)) {
+                       if (b_data(buf)) {
+                               int ret;
+
+                               ret = conn->xprt->snd_buf(conn, conn->xprt_ctx, buf, b_data(buf), flags);
+                               if (!ret) {
+                                       done = 1;
+                                       break;
+                               }
+                               sent = 1;
+                               TRACE_DATA("send data", SPOP_EV_SPOP_CONN_SEND, conn, 0, 0, (size_t[]){ret});
+                               b_del(buf, ret);
+                               if (b_data(buf)) {
+                                       done = 1;
+                                       break;
+                               }
+                       }
+                       b_free(buf);
+                       released++;
+               }
+
+               if (released)
+                       offer_buffers(NULL, released);
+
+               /* wrote at least one byte, the buffer is not full anymore */
+               if (spop_conn->flags & (SPOP_CF_MUX_MFULL | SPOP_CF_DEM_MROOM))
+                       TRACE_STATE("spop_conn mbuf ring not fill anymore", SPOP_EV_SPOP_CONN_SEND|SPOP_EV_SPOP_CONN_BLK, conn);
+               spop_conn->flags &= ~(SPOP_CF_MUX_MFULL | SPOP_CF_DEM_MROOM);
+       }
+
+       if (conn->flags & CO_FL_ERROR) {
+               spop_conn->flags |= SPOP_CF_ERR_PENDING;
+               if (spop_conn->flags & SPOP_CF_END_REACHED)
+                       spop_conn->flags |= SPOP_CF_ERROR;
+               b_reset(br_tail(spop_conn->mbuf));
+       }
+
+       /* We're not full anymore, so we can wake any task that are waiting
+        * for us.
+        */
+       if (!(spop_conn->flags & (SPOP_CF_MUX_MFULL | SPOP_CF_DEM_MROOM)) && spop_conn->state >= SPOP_CS_FRAME_H) {
+               spop_conn->flags &= ~SPOP_CF_WAIT_INLIST;
+               spop_resume_each_sending_spop_strm(spop_conn, &spop_conn->send_list);
+       }
+
+       /* We're done, no more to send */
+       if (!br_data(spop_conn->mbuf)) {
+               TRACE_DEVEL("leaving with everything sent", SPOP_EV_SPOP_CONN_SEND, conn);
+               goto end;
+       }
+schedule:
+       if (!(conn->flags & CO_FL_ERROR) && !(spop_conn->wait_event.events & SUB_RETRY_SEND)) {
+               TRACE_STATE("more data to send, subscribing", SPOP_EV_SPOP_CONN_SEND, conn);
+               conn->xprt->subscribe(conn, conn->xprt_ctx, SUB_RETRY_SEND, &spop_conn->wait_event);
+       }
+
+       TRACE_DEVEL("leaving with some data left to send", SPOP_EV_SPOP_CONN_SEND, conn);
+end:
+       return sent || (spop_conn->flags & (SPOP_CF_ERR_PENDING|SPOP_CF_ERROR));
+}
+
+
+/* this is the tasklet referenced in spop_conn->wait_event.tasklet */
+static struct task *spop_io_cb(struct task *t, void *ctx, unsigned int state)
+{
+       struct connection *conn;
+       struct spop_conn *spop_conn = ctx;
+       struct tasklet *tl = (struct tasklet *)t;
+       int conn_in_list;
+       int ret = 0;
+
+       if (state & TASK_F_USR1) {
+               /* the tasklet was idling on an idle connection, it might have
+                * been stolen, let's be careful!
+                */
+               HA_SPIN_LOCK(IDLE_CONNS_LOCK, &idle_conns[tid].idle_conns_lock);
+               if (tl->context == NULL) {
+                       /* The connection has been taken over by another thread,
+                        * we're no longer responsible for it, so just free the
+                        * tasklet, and do nothing.
+                        */
+                       HA_SPIN_UNLOCK(IDLE_CONNS_LOCK, &idle_conns[tid].idle_conns_lock);
+                       tasklet_free(tl);
+                       return NULL;
+               }
+               conn = spop_conn->conn;
+               TRACE_POINT(SPOP_EV_SPOP_CONN_WAKE, conn);
+
+               conn_in_list = conn->flags & CO_FL_LIST_MASK;
+               if (conn_in_list)
+                       conn_delete_from_tree(conn);
+
+               HA_SPIN_UNLOCK(IDLE_CONNS_LOCK, &idle_conns[tid].idle_conns_lock);
+       } else {
+               /* we're certain the connection was not in an idle list */
+               conn = spop_conn->conn;
+               TRACE_ENTER(SPOP_EV_SPOP_CONN_WAKE, conn);
+               conn_in_list = 0;
+       }
+
+       if (!(spop_conn->wait_event.events & SUB_RETRY_SEND))
+               ret = spop_send(spop_conn);
+       if (!(spop_conn->wait_event.events & SUB_RETRY_RECV))
+               ret |= spop_recv(spop_conn);
+       if (ret || b_data(&spop_conn->dbuf))
+               ret = spop_process(spop_conn);
+
+       /* If we were in an idle list, we want to add it back into it,
+        * unless spop_process() returned -1, which mean it has destroyed
+        * the connection (testing !ret is enough, if spop_process() wasn't
+        * called then ret will be 0 anyway.
+        */
+       if (ret < 0)
+               t = NULL;
+
+       if (!ret && conn_in_list) {
+               struct server *srv = objt_server(conn->target);
+
+               HA_SPIN_LOCK(IDLE_CONNS_LOCK, &idle_conns[tid].idle_conns_lock);
+               _srv_add_idle(srv, conn, conn_in_list == CO_FL_SAFE_LIST);
+               HA_SPIN_UNLOCK(IDLE_CONNS_LOCK, &idle_conns[tid].idle_conns_lock);
+       }
+       return t;
+}
+
+/* callback called on any event by the connection handler.
+ * It applies changes and returns zero, or < 0 if it wants immediate
+ * destruction of the connection (which normally doesn not happen in SPOP).
+ */
+static int spop_process(struct spop_conn *spop_conn)
+{
+       struct connection *conn = spop_conn->conn;
+
+       TRACE_POINT(SPOP_EV_SPOP_CONN_WAKE, conn);
+
+       if (!(spop_conn->flags & SPOP_CF_DEM_BLOCK_ANY) &&
+           (b_data(&spop_conn->dbuf) || (spop_conn->flags & SPOP_CF_RCVD_SHUT))) {
+               spop_process_demux(spop_conn);
+
+               if (spop_conn->state >= SPOP_CS_ERROR || (spop_conn->flags & SPOP_CF_ERROR))
+                       b_reset(&spop_conn->dbuf);
+
+               if (b_room(&spop_conn->dbuf))
+                       spop_conn->flags &= ~SPOP_CF_DEM_DFULL;
+       }
+       spop_send(spop_conn);
+
+       /*
+        * If we received early data, and the handshake is done, wake
+        * any stream that was waiting for it.
+        */
+       if (!(spop_conn->flags & SPOP_CF_WAIT_FOR_HS) &&
+           (conn->flags & (CO_FL_EARLY_SSL_HS | CO_FL_WAIT_XPRT | CO_FL_EARLY_DATA)) == CO_FL_EARLY_DATA) {
+               struct eb32_node *node;
+               struct spop_strm *spop_strm;
+
+               spop_conn->flags |= SPOP_CF_WAIT_FOR_HS;
+               node = eb32_lookup_ge(&spop_conn->streams_by_id, 1);
+
+               while (node) {
+                       spop_strm = container_of(node, struct spop_strm, by_id);
+                       if (spop_strm_sc(spop_strm) && se_fl_test(spop_strm->sd, SE_FL_WAIT_FOR_HS))
+                               spop_strm_notify_recv(spop_strm);
+                       node = eb32_next(node);
+               }
+       }
+
+       if ((spop_conn->flags & SPOP_CF_ERROR) || spop_conn_read0_pending(spop_conn) ||
+           spop_conn->state == SPOP_CS_CLOSED || (spop_conn->flags & SPOP_CF_DISCO_FAILED) /* || */
+           /* TODO: no sure ? eb_is_empty(&spop_conn->streams_by_id) */) {
+               spop_wake_some_streams(spop_conn, 0);
+
+               if (eb_is_empty(&spop_conn->streams_by_id)) {
+                       /* no more stream, kill the connection now */
+                       spop_release(spop_conn);
+                       TRACE_DEVEL("leaving after releasing the connection", SPOP_EV_SPOP_CONN_WAKE);
+                       return -1;
+               }
+
+               // TODO:
+               /* /\* connections in error must be removed from the idle lists *\/ */
+               /* if (conn->flags & CO_FL_LIST_MASK) { */
+               /*      HA_SPIN_LOCK(IDLE_CONNS_LOCK, &idle_conns[tid].idle_conns_lock); */
+               /*      conn_delete_from_tree(conn); */
+               /*      HA_SPIN_UNLOCK(IDLE_CONNS_LOCK, &idle_conns[tid].idle_conns_lock); */
+               /* } */
+       }
+
+       if (!b_data(&spop_conn->dbuf))
+               spop_release_buf(spop_conn, &spop_conn->dbuf);
+
+       if (spop_conn->state == SPOP_CS_CLOSED  || (spop_conn->flags & SPOP_CF_DISCO_FAILED) ||
+           (!br_data(spop_conn->mbuf) && ((spop_conn->flags & SPOP_CF_MUX_BLOCK_ANY) || LIST_ISEMPTY(&spop_conn->send_list))))
+               spop_release_mbuf(spop_conn);
+
+       if (spop_conn->task) {
+               spop_conn->task->expire = tick_add(now_ms, (spop_conn->state == SPOP_CS_CLOSED ? spop_conn->shut_timeout : spop_conn->timeout));
+               task_queue(spop_conn->task);
+       }
+
+       spop_send(spop_conn);
+       TRACE_LEAVE(SPOP_EV_SPOP_CONN_WAKE, conn);
+       return 0;
+}
+
+/* wake-up function called by the connection layer (mux_ops.wake) */
+static int spop_wake(struct connection *conn)
+{
+       struct spop_conn *spop_conn = conn->ctx;
+
+       TRACE_POINT(SPOP_EV_SPOP_CONN_WAKE, conn);
+       return (spop_process(spop_conn));
+}
+
+static int spop_ctl(struct connection *conn, enum mux_ctl_type mux_ctl, void *output)
+{
+       struct spop_conn *spop_conn = conn->ctx;
+       int ret = 0;
+
+       switch (mux_ctl) {
+       case MUX_CTL_STATUS:
+               if (!(conn->flags & CO_FL_WAIT_XPRT))
+                       ret |= MUX_STATUS_READY;
+               return ret;
+       case MUX_CTL_EXIT_STATUS:
+               return MUX_ES_UNKNOWN;
+       case MUX_CTL_GET_NBSTRM:
+               return spop_conn->nb_streams;
+       case MUX_CTL_GET_MAXSTRM:
+               return spop_conn->streams_limit;
+       default:
+               return -1;
+       }
+}
+
+static int spop_sctl(struct stconn *sc, enum mux_sctl_type mux_sctl, void *output)
+{
+       int ret = 0;
+       struct spop_strm *spop_strm = __sc_mux_strm(sc);
+
+       switch (mux_sctl) {
+       case MUX_SCTL_SID:
+               if (output)
+                       *((int64_t *)output) = spop_strm->id;
+               return ret;
+
+       default:
+               return -1;
+       }
+}
+
+/* Connection timeout management. The principle is that if there's no receipt
+ * nor sending for a certain amount of time, the connection is closed. If the
+ * MUX buffer still has lying data or is not allocatable, the connection is
+ * immediately killed. If it's allocatable and empty, we attempt to send a
+ * ABORT records.
+ */
+static struct task *spop_timeout_task(struct task *t, void *context, unsigned int state)
+{
+       struct spop_conn *spop_conn = context;
+       int expired = tick_is_expired(t->expire, now_ms);
+
+       TRACE_ENTER(SPOP_EV_SPOP_CONN_WAKE, (spop_conn ? spop_conn->conn : NULL));
+
+       if (spop_conn) {
+               HA_SPIN_LOCK(IDLE_CONNS_LOCK, &idle_conns[tid].idle_conns_lock);
+
+               /* Somebody already stole the connection from us, so we should not
+                * free it, we just have to free the task.
+                */
+               if (!t->context) {
+                       HA_SPIN_UNLOCK(IDLE_CONNS_LOCK, &idle_conns[tid].idle_conns_lock);
+                       spop_conn = NULL;
+                       goto do_leave;
+               }
+
+               if (!expired) {
+                       HA_SPIN_UNLOCK(IDLE_CONNS_LOCK, &idle_conns[tid].idle_conns_lock);
+                       TRACE_DEVEL("leaving (not expired)", SPOP_EV_SPOP_CONN_WAKE, spop_conn->conn);
+                       return t;
+               }
+
+               /* We're about to destroy the connection, so make sure nobody attempts
+                * to steal it from us.
+                */
+               if (spop_conn->conn->flags & CO_FL_LIST_MASK)
+                       conn_delete_from_tree(spop_conn->conn);
+
+               HA_SPIN_UNLOCK(IDLE_CONNS_LOCK, &idle_conns[tid].idle_conns_lock);
+       }
+
+do_leave:
+       task_destroy(t);
+
+       if (!spop_conn) {
+               /* resources were already deleted */
+               TRACE_DEVEL("leaving (not more spop_conn)", SPOP_EV_SPOP_CONN_WAKE);
+               return NULL;
+       }
+
+       spop_conn->task = NULL;
+       spop_conn->state = SPOP_CS_CLOSED;
+       spop_wake_some_streams(spop_conn, 0);
+
+       if (br_data(spop_conn->mbuf)) {
+               /* don't even try to send aborts, the buffer is stuck */
+               spop_conn->flags |= SPOP_CF_DISCO_FAILED;
+               goto end;
+       }
+
+       /* try to send but no need to insist */
+       if (!spop_conn_send_disconnect(spop_conn))
+               spop_conn->flags |= SPOP_CF_DISCO_FAILED;
+
+       if (br_data(spop_conn->mbuf) && !(spop_conn->flags & SPOP_CF_DISCO_FAILED) &&
+           conn_xprt_ready(spop_conn->conn)) {
+               unsigned int released = 0;
+               struct buffer *buf;
+
+               for (buf = br_head(spop_conn->mbuf); b_size(buf); buf = br_del_head(spop_conn->mbuf)) {
+                       if (b_data(buf)) {
+                               int ret = spop_conn->conn->xprt->snd_buf(spop_conn->conn, spop_conn->conn->xprt_ctx,
+                                                                        buf, b_data(buf), 0);
+                               if (!ret)
+                                       break;
+                               b_del(buf, ret);
+                               if (b_data(buf))
+                                       break;
+                               b_free(buf);
+                               released++;
+                       }
+               }
+
+               if (released)
+                       offer_buffers(NULL, released);
+       }
+
+  end:
+       /* either we can release everything now or it will be done later once
+        * the last stream closes.
+        */
+       if (eb_is_empty(&spop_conn->streams_by_id))
+               spop_release(spop_conn);
+
+       TRACE_LEAVE(SPOP_EV_SPOP_CONN_WAKE);
+       return NULL;
+}
+
+
+/*******************************************/
+/* functions below are used by the streams */
+/*******************************************/
+
+/* Append the description of what is present in error snapshot <es> into <out>.
+ * The description must be small enough to always fit in a buffer. The output
+ * buffer may be the trash so the trash must not be used inside this function.
+ */
+static __maybe_unused void spop_show_error_snapshot(struct buffer *out, const struct error_snapshot *es)
+{
+
+       // TODO !!!
+}
+/*
+ * Capture a bad response and archive it in the proxy's structure.  By default
+ * it tries to report the error position as h1m->err_pos. However if this one is
+ * not set, it will then report h1m->next, which is the last known parsing
+ * point. The function is able to deal with wrapping buffers. It always displays
+ * buffers as a contiguous area starting at buf->p. The direction is determined
+ * thanks to the h1m's flags.
+ */
+static __maybe_unused void spop_strm_capture_bad_message(struct spop_conn *spop_conn, struct spop_strm *spop_strm,
+                                                        struct buffer *buf)
+{
+       // TODO !!!
+}
+
+// PARSING FUNCTIONS
+
+
+/*
+ * Attach a new stream to a connection
+ * (Used for outgoing connections)
+ */
+static int spop_attach(struct connection *conn, struct sedesc *sd, struct session *sess)
+{
+       struct spop_strm *spop_strm;
+       struct spop_conn *spop_conn = conn->ctx;
+
+       TRACE_ENTER(SPOP_EV_SPOP_STRM_NEW, conn);
+       spop_strm = spop_stconn_new(spop_conn, sd->sc, sess);
+       if (!spop_strm)
+               goto err;
+
+       /* the connection is not idle anymore, let's mark this */
+       HA_ATOMIC_AND(&spop_conn->wait_event.tasklet->state, ~TASK_F_USR1);
+       xprt_set_used(conn, conn->xprt, conn->xprt_ctx);
+
+       TRACE_LEAVE(SPOP_EV_SPOP_STRM_NEW, conn, spop_strm);
+       return 0;
+
+  err:
+       TRACE_DEVEL("leaving on error", SPOP_EV_SPOP_STRM_NEW|SPOP_EV_SPOP_STRM_ERR, conn);
+       return -1;
+}
+
+/* Retrieves the first valid stream connector from this connection, or returns NULL.
+ * We have to scan because we may have some orphan streams. It might be
+ * beneficial to scan backwards from the end to reduce the likeliness to find
+ * orphans.
+ */
+static struct stconn *spop_get_first_sc(const struct connection *conn)
+{
+       struct spop_conn *spop_conn = conn->ctx;
+       struct spop_strm *spop_strm;
+       struct eb32_node *node;
+
+       node = eb32_first(&spop_conn->streams_by_id);
+       while (node) {
+               spop_strm = container_of(node, struct spop_strm, by_id);
+               if (spop_strm_sc(spop_strm))
+                       return spop_strm_sc(spop_strm);
+               node = eb32_next(node);
+       }
+       return NULL;
+}
+
+/*
+ * Destroy the mux and the associated connection, if it is no longer used
+ */
+static void spop_destroy(void *ctx)
+{
+       struct spop_conn *spop_conn = ctx;
+
+       TRACE_POINT(SPOP_EV_SPOP_CONN_END, spop_conn->conn);
+       if (eb_is_empty(&spop_conn->streams_by_id)) {
+               BUG_ON(spop_conn->conn->ctx != spop_conn);
+               spop_release(spop_conn);
+       }
+}
+
+/*
+ * Detach the stream from the connection and possibly release the connection.
+ */
+static void spop_detach(struct sedesc *sd)
+{
+       struct spop_strm *spop_strm = sd->se;
+       struct spop_conn *spop_conn;
+       //struct session *sess;
+
+       TRACE_ENTER(SPOP_EV_STRM_END, (spop_strm ? spop_strm->spop_conn->conn : NULL), spop_strm);
+
+       if (!spop_strm) {
+               TRACE_LEAVE(SPOP_EV_STRM_END);
+               return;
+       }
+
+       /* there's no txbuf so we're certain no to be able to send anything */
+       spop_strm->flags &= ~SPOP_SF_NOTIFIED;
+
+       //sess = spop_strm->sess;
+       spop_conn = spop_strm->spop_conn;
+       spop_conn->nb_sc--;
+
+       /* this stream may be blocked waiting for some data to leave, so orphan
+        * it in this case.
+        */
+       if (!(spop_conn->flags & (SPOP_CF_ERR_PENDING|SPOP_CF_ERROR)) && // FIXME: Be sure for ERR_PENDING
+           (spop_conn->state != SPOP_CS_CLOSED) &&
+           (spop_strm->flags & (SPOP_SF_BLK_MBUSY|SPOP_SF_BLK_MROOM)) &&
+           spop_strm->subs) {
+               TRACE_DEVEL("leaving on stream blocked", SPOP_EV_STRM_END|SPOP_EV_SPOP_STRM_BLK, spop_conn->conn, spop_strm);
+               /* refresh the timeout if none was active, so that the last
+                * leaving stream may arm it.
+                */
+               // TODO: Add spop_update_timeout  function
+                if (spop_conn->task && !tick_isset(spop_conn->task->expire))
+                        spop_conn->task->expire = tick_add(now_ms, (spop_conn->state == SPOP_CS_CLOSED ? spop_conn->shut_timeout : spop_conn->timeout));
+               /*      spop_update_timeout(spop_conn); */
+               return;
+       }
+
+       if ((spop_conn->flags & SPOP_CF_DEM_BLOCK_ANY && spop_strm->id == spop_conn->dsi)) {
+               /* unblock the connection if it was blocked on this stream. */
+               spop_conn->flags &= ~SPOP_CF_DEM_BLOCK_ANY;
+               spop_conn_restart_reading(spop_conn, 1);
+       }
+
+       spop_strm_destroy(spop_strm);
+
+       if (!(spop_conn->flags & (SPOP_CF_RCVD_SHUT|SPOP_CF_ERR_PENDING|SPOP_CF_ERROR))) {
+               // TODO: for now, cannection cannot be resued. Just close it
+               if (spop_conn->state < SPOP_CS_ERROR) {
+                       /* spop_conn_error(spop_conn, SPOP_ERR_NONE); */
+                       TRACE_PROTO("sending SPOP HAPROXY DISCONNECT frame", SPOP_EV_STRM_END|SPOP_EV_TX_DISCO, spop_conn->conn);
+                       if (!spop_conn_send_disconnect(spop_conn))
+                               spop_conn->flags |= SPOP_CF_DISCO_FAILED;
+                       if (!(spop_conn->wait_event.events & SUB_RETRY_SEND))
+                               tasklet_wakeup(spop_conn->wait_event.tasklet);
+               }
+       }
+
+       /* We don't want to close right now unless we're removing the last
+        * stream and the connection is in error.
+        */
+       if (spop_conn_is_dead(spop_conn)) {
+               /* no more stream will come, kill it now */
+               TRACE_DEVEL("leaving, killing dead connection", SPOP_EV_STRM_END, spop_conn->conn);
+               spop_release(spop_conn);
+       }
+       else if (spop_conn->task) {
+               spop_conn->task->expire = tick_add(now_ms, (spop_conn->state == SPOP_CS_CLOSED ? spop_conn->shut_timeout : spop_conn->timeout));
+               task_queue(spop_conn->task);
+               TRACE_DEVEL("leaving, refreshing connection's timeout", SPOP_EV_STRM_END, spop_conn->conn);
+       }
+       else
+               TRACE_DEVEL("leaving", SPOP_EV_STRM_END, spop_conn->conn);
+}
+
+/* Performs a synchronous or asynchronous shutr(). */
+static void spop_do_shutr(struct spop_strm *spop_strm)
+{
+       struct spop_conn *spop_conn = spop_strm->spop_conn;
+
+       TRACE_ENTER(SPOP_EV_STRM_SHUT, spop_conn->conn, spop_strm);
+
+       if (spop_strm->state == SPOP_SS_CLOSED)
+               goto done;
+
+       /* a connstream may require us to immediately kill the whole connection
+        * for example because of a "tcp-request content reject" rule that is
+        * normally used to limit abuse.
+        */
+       if (se_fl_test(spop_strm->sd, SE_FL_KILL_CONN) &&
+           !(spop_conn->flags & (SPOP_CF_DISCO_SENT|SPOP_CF_DISCO_FAILED))) {
+               TRACE_STATE("stream wants to kill the connection", SPOP_EV_STRM_SHUT, spop_conn->conn, spop_strm);
+               spop_conn->state = SPOP_CS_CLOSED;
+       }
+
+       spop_strm_close(spop_strm);
+
+       if (!(spop_conn->wait_event.events & SUB_RETRY_SEND))
+               tasklet_wakeup(spop_conn->wait_event.tasklet);
+  done:
+       TRACE_LEAVE(SPOP_EV_STRM_SHUT, spop_conn->conn, spop_strm);
+       return;
+}
+
+/* Performs a synchronous or asynchronous shutw(). */
+static void spop_do_shutw(struct spop_strm *spop_strm)
+{
+       struct spop_conn *spop_conn = spop_strm->spop_conn;
+
+       TRACE_ENTER(SPOP_EV_STRM_SHUT, spop_conn->conn, spop_strm);
+
+       if (spop_strm->state == SPOP_SS_HLOC || spop_strm->state == SPOP_SS_CLOSED)
+               goto done;
+
+       if (spop_strm->state == SPOP_SS_OPEN)
+               spop_strm->state = SPOP_SS_HLOC;
+       else if (spop_strm->state == SPOP_SS_HREM)
+               spop_strm_close(spop_strm);
+       else {
+               /* a connstream may require us to immediately kill the whole connection
+                * for example because of a "tcp-request content reject" rule that is
+                * normally used to limit abuse.
+                */
+               if (se_fl_test(spop_strm->sd, SE_FL_KILL_CONN)) {
+                       TRACE_STATE("stream wants to kill the connection", SPOP_EV_STRM_SHUT, spop_conn->conn, spop_strm);
+                       spop_conn->state = SPOP_CS_CLOSED;
+               }
+       }
+
+       if (!(spop_conn->wait_event.events & SUB_RETRY_SEND))
+               tasklet_wakeup(spop_conn->wait_event.tasklet);
+  done:
+       TRACE_LEAVE(SPOP_EV_STRM_SHUT, spop_conn->conn, spop_strm);
+       return;
+}
+
+
+static void spop_shut(struct stconn *sc, enum se_shut_mode mode, struct se_abort_info *reason)
+{
+       struct spop_strm *spop_strm = __sc_mux_strm(sc);
+
+       TRACE_ENTER(SPOP_EV_STRM_SHUT, spop_strm->spop_conn->conn, spop_strm);
+       if (mode & (SE_SHW_SILENT|SE_SHW_NORMAL))
+               spop_do_shutw(spop_strm);
+       if (mode & SE_SHR_RESET)
+               spop_do_shutr(spop_strm);
+       TRACE_LEAVE(SPOP_EV_STRM_SHUT, spop_strm->spop_conn->conn, spop_strm);
+}
+
+/* Called from the upper layer, to subscribe <es> to events <event_type>. The
+ * event subscriber <es> is not allowed to change from a previous call as long
+ * as at least one event is still subscribed. The <event_type> must only be a
+ * combination of SUB_RETRY_RECV and SUB_RETRY_SEND. It always returns 0.
+ */
+static int spop_subscribe(struct stconn *sc, int event_type, struct wait_event *es)
+{
+       struct spop_strm *spop_strm = __sc_mux_strm(sc);
+       struct spop_conn *spop_conn = spop_strm->spop_conn;
+
+       BUG_ON(event_type & ~(SUB_RETRY_SEND|SUB_RETRY_RECV));
+       BUG_ON(spop_strm->subs && spop_strm->subs != es);
+
+       es->events |= event_type;
+       spop_strm->subs = es;
+
+       if (event_type & SUB_RETRY_RECV)
+               TRACE_DEVEL("unsubscribe(recv)", SPOP_EV_STRM_RECV, spop_conn->conn, spop_strm);
+
+       if (event_type & SUB_RETRY_SEND) {
+               TRACE_DEVEL("unsubscribe(send)", SPOP_EV_STRM_SEND, spop_conn->conn, spop_strm);
+               if (!LIST_INLIST(&spop_strm->list))
+                       LIST_APPEND(&spop_conn->send_list, &spop_strm->list);
+       }
+       return 0;
+}
+
+/* Called from the upper layer, to unsubscribe <es> from events <event_type>
+ * (undo spop_subscribe). The <es> pointer is not allowed to differ from the one
+ * passed to the subscribe() call. It always returns zero.
+ */
+static int spop_unsubscribe(struct stconn *sc, int event_type, struct wait_event *es)
+{
+       struct spop_strm *spop_strm = __sc_mux_strm(sc);
+       struct spop_conn *spop_conn = spop_strm->spop_conn;
+
+       BUG_ON(event_type & ~(SUB_RETRY_SEND|SUB_RETRY_RECV));
+       BUG_ON(spop_strm->subs && spop_strm->subs != es);
+
+       es->events &= ~event_type;
+       if (!es->events)
+               spop_strm->subs = NULL;
+
+       if (event_type & SUB_RETRY_RECV)
+               TRACE_DEVEL("subscribe(recv)", SPOP_EV_STRM_RECV, spop_conn->conn, spop_strm);
+
+       if (event_type & SUB_RETRY_SEND) {
+               TRACE_DEVEL("subscribe(send)", SPOP_EV_STRM_SEND, spop_conn->conn, spop_strm);
+               spop_strm->flags &= ~SPOP_SF_NOTIFIED;
+               LIST_DEL_INIT(&spop_strm->list);
+       }
+       return 0;
+}
+
+/* Try to send a NOTIFY frame. Returns the number of bytes sent. The caller
+ * must check the stream's status to detect any error which might have happened
+ * subsequently to a successful send. Data are automatically removed from the
+ * buffer. Content of the frame is assumed to be valid since produced from
+ * the internal code.
+ */
+static size_t spop_strm_send_notify(struct spop_strm *spop_strm, struct buffer *buf, size_t count)
+{
+       struct spop_conn *spop_conn = spop_strm->spop_conn;
+       struct buffer outbuf;
+       struct buffer *mbuf;
+       char *p, *end;
+       size_t sz;
+       int ret = 0;
+
+       TRACE_ENTER(SPOP_EV_TX_FRAME|SPOP_EV_TX_NOTIFY, spop_conn->conn, spop_strm);
+
+       mbuf = br_tail(spop_conn->mbuf);
+  retry:
+       if (!spop_get_buf(spop_conn, mbuf)) {
+               spop_conn->flags |= SPOP_CF_MUX_MALLOC;
+               spop_strm->flags |= SPOP_SF_BLK_MROOM;
+               TRACE_STATE("waiting for fconn mbuf ring allocation", SPOP_EV_TX_FRAME|SPOP_EV_TX_NOTIFY|SPOP_EV_SPOP_STRM_BLK, spop_conn->conn, spop_strm);
+               ret = 0;
+               goto end;
+       }
+
+       while (1) {
+               outbuf = b_make(b_tail(mbuf), b_contig_space(mbuf), 0, 0);
+               if (outbuf.size >= 11 || !b_space_wraps(mbuf))
+                       break;
+         realign_again:
+               b_slow_realign(mbuf, trash.area, b_data(mbuf));
+       }
+
+       if (outbuf.size < 11)
+               goto full;
+
+       /* len: 4-bytes (fill later) type: (2)HAPROXY-NOTIFY, flags: 4-bytes (FIN=1) */
+       memcpy(outbuf.area, "\x00\x00\x00\x00\x03\x00\x00\x00\x01", 9);
+       outbuf.data = 9;
+
+       p = b_tail(&outbuf);
+       end = b_orig(&outbuf) + b_size(&outbuf);
+
+       if (encode_varint(spop_strm->id, &p, end) == -1)
+               goto full;
+       if (encode_varint(spop_strm->fid+1, &p, end) == -1)
+               goto full;
+
+       sz = count - 1; /* Skip the frame type */
+       if (p + sz > end)
+               goto full;
+       memcpy(p, b_peek(buf,1), sz);
+       p += sz;
+
+       outbuf.data += p - b_tail(&outbuf);
+
+       /* update the frame's size now */
+       TRACE_PROTO("SPOP HAPROXY NOTIFY frame xferred", SPOP_EV_TX_FRAME|SPOP_EV_TX_NOTIFY, spop_conn->conn, spop_strm, 0, (size_t[]){outbuf.data});
+       spop_set_frame_size(outbuf.area, outbuf.data - 4);
+       spop_strm->fid++;
+       b_add(mbuf, outbuf.data);
+       b_del(buf, count);
+       ret = count;
+
+  end:
+       TRACE_LEAVE(SPOP_EV_TX_FRAME|SPOP_EV_TX_NOTIFY, spop_conn->conn, spop_strm);
+       return ret;
+  full:
+       /* Too large to be encoded. For NOTIFY frame, it is an error */
+       if (!b_data(mbuf)) {
+               TRACE_ERROR("SPOP HAPROXY NOTIFY frame too large", SPOP_EV_TX_FRAME|SPOP_EV_TX_NOTIFY|SPOP_EV_SPOP_STRM_ERR, spop_conn->conn, spop_strm);
+               spop_strm_error(spop_strm, SPOP_ERR_TOO_BIG);
+               goto fail;
+       }
+
+       if ((mbuf = br_tail_add(spop_conn->mbuf)) != NULL)
+               goto retry;
+       spop_conn->flags |= SPOP_CF_MUX_MFULL;
+       spop_strm->flags |= SPOP_SF_BLK_MROOM;
+       TRACE_STATE("mbuf ring full", SPOP_EV_TX_FRAME|SPOP_EV_TX_NOTIFY|SPOP_EV_SPOP_STRM_BLK, spop_conn->conn, spop_strm);
+       ret = 0;
+       goto end;
+  fail:
+       TRACE_DEVEL("leaving on error", SPOP_EV_TX_FRAME|SPOP_EV_TX_NOTIFY|SPOP_EV_SPOP_STRM_ERR, spop_conn->conn, spop_strm);
+       return 0;
+}
+
+/* Called from the upper layer, to receive data
+ *
+ * The caller is responsible for defragmenting <buf> if necessary. But <flags>
+ * must be tested to know the calling context. If CO_RFL_BUF_FLUSH is set, it
+ * means the caller wants to flush input data (from the mux buffer and the
+ * channel buffer) to be able to use kernel splicing or any kind of mux-to-mux
+ * xfer. If CO_RFL_KEEP_RECV is set, the mux must always subscribe for read
+ * events before giving back. CO_RFL_BUF_WET is set if <buf> is congested with
+ * data scheduled for leaving soon. CO_RFL_BUF_NOT_STUCK is set to instruct the
+ * mux it may optimize the data copy to <buf> if necessary. Otherwise, it should
+ * copy as much data as possible.
+ */
+static size_t spop_rcv_buf(struct stconn *sc, struct buffer *buf, size_t count, int flags)
+{
+       struct spop_strm *spop_strm = __sc_mux_strm(sc);
+       struct spop_conn *spop_conn = spop_strm->spop_conn;
+       size_t ret = 0;
+
+       TRACE_ENTER(SPOP_EV_STRM_RECV, spop_conn->conn, spop_strm, 0, (size_t[]){count});
+
+       ret = b_xfer(buf, &spop_strm->rxbuf, MIN(b_data(&spop_strm->rxbuf), count));
+       if (b_data(&spop_strm->rxbuf))
+               se_fl_set(spop_strm->sd, SE_FL_RCV_MORE | SE_FL_WANT_ROOM);
+       else {
+               se_fl_clr(spop_strm->sd, SE_FL_RCV_MORE | SE_FL_WANT_ROOM);
+               spop_strm_propagate_term_flags(spop_conn, spop_strm);
+               if (b_size(&spop_strm->rxbuf)) {
+                       b_free(&spop_strm->rxbuf);
+                       offer_buffers(NULL, 1);
+               }
+       }
+
+       if (ret && spop_conn->dsi == spop_strm->id) { // TODO must match the frame id too !!!!
+               /* demux is blocking on this stream's buffer */
+               spop_conn->flags &= ~SPOP_CF_DEM_SFULL;
+               spop_conn_restart_reading(spop_conn, 1);
+       }
+
+       TRACE_LEAVE(SPOP_EV_STRM_RECV, spop_conn->conn, spop_strm, 0, (size_t[]){ret});
+       return ret;
+}
+
+
+/* Called from the upper layer, to send data from buffer <buf> for no more than
+ * <count> bytes. Returns the number of bytes effectively sent. Some status
+ * flags may be updated on the stream connector.
+ */
+static size_t spop_snd_buf(struct stconn *sc, struct buffer *buf, size_t count, int flags)
+{
+       struct spop_strm *spop_strm = __sc_mux_strm(sc);
+       struct spop_conn *spop_conn = spop_strm->spop_conn;
+       size_t total = 0;
+       size_t ret;
+
+       TRACE_ENTER(SPOP_EV_STRM_SEND, spop_conn->conn, spop_strm, 0, (size_t[]){count});
+
+       /* If we were not just woken because we wanted to send but couldn't,
+        * and there's somebody else that is waiting to send, do nothing,
+        * we will subscribe later and be put at the end of the list
+        */
+       if (!(spop_strm->flags & SPOP_SF_NOTIFIED) && !LIST_ISEMPTY(&spop_conn->send_list)) {
+               if (LIST_INLIST(&spop_strm->list))
+                       TRACE_DEVEL("stream already waiting, leaving", SPOP_EV_STRM_SEND|SPOP_EV_SPOP_STRM_BLK, spop_conn->conn, spop_strm);
+               else {
+                       TRACE_DEVEL("other streams already waiting, going to the queue and leaving", SPOP_EV_STRM_SEND|SPOP_EV_SPOP_STRM_BLK,
+                                   spop_conn->conn, spop_strm);
+                       spop_conn->flags |= SPOP_CF_WAIT_INLIST;
+               }
+               return 0;
+       }
+       spop_strm->flags &= ~SPOP_SF_NOTIFIED;
+
+       if (spop_conn->state < SPOP_CS_FRAME_H) {
+               TRACE_DEVEL("connection not ready, leaving", SPOP_EV_STRM_SEND|SPOP_EV_SPOP_STRM_BLK, spop_conn->conn, spop_strm);
+               return 0;
+       }
+
+       if (spop_conn->state >= SPOP_CS_ERROR) {
+               se_fl_set(spop_strm->sd, SE_FL_ERROR);
+               TRACE_DEVEL("connection is in error, leaving in error", SPOP_EV_STRM_SEND|SPOP_EV_SPOP_STRM_ERR|SPOP_EV_STRM_ERR,
+                           spop_conn->conn, spop_strm);
+               return 0;
+       }
+
+       if (spop_strm->id == 0) {
+               int32_t id = spop_conn_get_next_sid(spop_conn);
+
+               if (id < 0) {
+                       se_fl_set(spop_strm->sd, SE_FL_ERROR);
+                       TRACE_DEVEL("couldn't get a stream ID, leaving in error", SPOP_EV_STRM_SEND|SPOP_EV_SPOP_STRM_ERR|SPOP_EV_STRM_ERR,
+                                   spop_conn->conn, spop_strm);
+                       return 0;
+               }
+
+               eb32_delete(&spop_strm->by_id);
+               spop_strm->by_id.key = spop_strm->id = id;
+               spop_strm->state = SPOP_SS_OPEN;
+               spop_conn->max_id = id;
+               spop_conn->nb_reserved--;
+               eb32_insert(&spop_conn->streams_by_id, &spop_strm->by_id);
+       }
+
+       while (spop_strm->state < SPOP_SS_HLOC && !(spop_strm->flags & SPOP_SF_BLK_ANY) && count) {
+               enum spop_frame_type type = *b_peek(buf, 0);
+
+               switch (type) {
+               case SPOP_FRM_T_HAPROXY_NOTIFY:
+                       ret = spop_strm_send_notify(spop_strm, buf, count);
+                       if (ret > 0) {
+                               total += ret;
+                               count -= ret;
+                       }
+                       break;
+               default:
+                       TRACE_DEVEL("Unsupported frame type", SPOP_EV_STRM_SEND|SPOP_EV_STRM_ERR|SPOP_EV_SPOP_STRM_ERR,
+                                   spop_conn->conn, spop_strm, 0, (size_t[]){type});
+                       spop_strm_error(spop_strm, SPOP_ERR_INVALID);
+                       break;
+               }
+       }
+
+       if (spop_strm->state >= SPOP_SS_HLOC) {
+               /* trim any possibly pending data after we close */
+               total += count;
+               count = 0;
+       }
+       /* RST are sent similarly to frame acks */
+       if (spop_strm->state == SPOP_SS_ERROR) {
+               TRACE_DEVEL("reporting error to the app-layer stream", SPOP_EV_STRM_SEND|SPOP_EV_STRM_ERR|SPOP_EV_SPOP_STRM_ERR, spop_conn->conn, spop_strm);
+               se_fl_set_error(spop_strm->sd);
+               spop_strm_close(spop_strm);
+       }
+
+       if (total > 0) {
+               if (!(spop_conn->wait_event.events & SUB_RETRY_SEND)) {
+                       TRACE_DEVEL("data queued, waking up spop_conn sender", SPOP_EV_STRM_SEND|SPOP_EV_SPOP_CONN_SEND, spop_conn->conn, spop_strm);
+                       if (spop_send(spop_conn))
+                               tasklet_wakeup(spop_conn->wait_event.tasklet);
+               }
+
+       }
+
+       if (total > 0) {
+               /* Ok we managed to send something, leave the send_list if we were still there */
+               spop_remove_from_list(spop_strm);
+               TRACE_DEVEL("Removed from spop_strm list", SPOP_EV_STRM_SEND|SPOP_EV_SPOP_CONN_SEND, spop_conn->conn, spop_strm);
+       }
+
+       TRACE_LEAVE(SPOP_EV_STRM_SEND, spop_conn->conn, spop_strm, 0, (size_t[]){total});
+       return total;
+}
+
+
+/* for debugging with CLI's "show fd" command */
+static int spop_show_fd(struct buffer *msg, struct connection *conn)
+{
+       struct spop_conn *spop_conn = conn->ctx;
+       struct spop_strm *spop_strm = NULL;
+       struct eb32_node *node;
+       int send_cnt = 0;
+       int tree_cnt = 0;
+       int orph_cnt = 0;
+       struct buffer *hmbuf, *tmbuf;
+
+       if (!spop_conn)
+               return 0;
+
+       list_for_each_entry(spop_strm, &spop_conn->send_list, list)
+               send_cnt++;
+
+       spop_strm = NULL;
+       node = eb32_first(&spop_conn->streams_by_id);
+       while (node) {
+               spop_strm = container_of(node, struct spop_strm, by_id);
+               tree_cnt++;
+               if (!spop_strm_sc(spop_strm))
+                       orph_cnt++;
+               node = eb32_next(node);
+       }
+
+       hmbuf = br_head(spop_conn->mbuf);
+       tmbuf = br_tail(spop_conn->mbuf);
+       chunk_appendf(msg, " spop_conn.st0=%d .maxid=%d .flg=0x%04x .nbst=%u"
+                     " .nbcs=%u .send_cnt=%d .tree_cnt=%d .orph_cnt=%d .sub=%d "
+                     ".dsi=%d .dbuf=%u@%p+%u/%u .mbuf=[%u..%u|%u],h=[%u@%p+%u/%u],t=[%u@%p+%u/%u]",
+                     spop_conn->state, spop_conn->max_id, spop_conn->flags,
+                     spop_conn->nb_streams, spop_conn->nb_sc, send_cnt, tree_cnt, orph_cnt,
+                     spop_conn->wait_event.events, spop_conn->dsi,
+                     (unsigned int)b_data(&spop_conn->dbuf), b_orig(&spop_conn->dbuf),
+                     (unsigned int)b_head_ofs(&spop_conn->dbuf), (unsigned int)b_size(&spop_conn->dbuf),
+                     br_head_idx(spop_conn->mbuf), br_tail_idx(spop_conn->mbuf), br_size(spop_conn->mbuf),
+                     (unsigned int)b_data(hmbuf), b_orig(hmbuf),
+                     (unsigned int)b_head_ofs(hmbuf), (unsigned int)b_size(hmbuf),
+                     (unsigned int)b_data(tmbuf), b_orig(tmbuf),
+                     (unsigned int)b_head_ofs(tmbuf), (unsigned int)b_size(tmbuf));
+
+       if (spop_strm) {
+               chunk_appendf(msg, " last_spop_strm=%p .id=%d .flg=0x%04x .rxbuf=%u@%p+%u/%u .sc=%p",
+                             spop_strm, spop_strm->id, spop_strm->flags,
+                             (unsigned int)b_data(&spop_strm->rxbuf), b_orig(&spop_strm->rxbuf),
+                             (unsigned int)b_head_ofs(&spop_strm->rxbuf), (unsigned int)b_size(&spop_strm->rxbuf),
+                             spop_strm_sc(spop_strm));
+
+               chunk_appendf(msg, " .sd.flg=0x%08x", se_fl_get(spop_strm->sd));
+               if (!se_fl_test(spop_strm->sd, SE_FL_ORPHAN))
+                       chunk_appendf(msg, " .sc.flg=0x%08x .sc.app=%p",
+                                     spop_strm_sc(spop_strm)->flags, spop_strm_sc(spop_strm)->app);
+
+               chunk_appendf(msg, " .subs=%p", spop_strm->subs);
+               if (spop_strm->subs) {
+                       chunk_appendf(msg, "(ev=%d tl=%p", spop_strm->subs->events, spop_strm->subs->tasklet);
+                       chunk_appendf(msg, " tl.calls=%d tl.ctx=%p tl.fct=",
+                                     spop_strm->subs->tasklet->calls,
+                                     spop_strm->subs->tasklet->context);
+                       resolve_sym_name(msg, NULL, spop_strm->subs->tasklet->process);
+                       chunk_appendf(msg, ")");
+               }
+       }
+       return 0;
+}
+
+/* Migrate the the connection to the current thread.
+ * Return 0 if successful, non-zero otherwise.
+ * Expected to be called with the old thread lock held.
+ */
+static int spop_takeover(struct connection *conn, int orig_tid, int release)
+{
+       struct spop_conn *spop_conn = conn->ctx;
+       struct task *task;
+       struct task *new_task = NULL;
+       struct tasklet *new_tasklet = NULL;
+
+       /* Pre-allocate tasks so that we don't have to roll back after the xprt
+        * has been migrated.
+        */
+       if (!release) {
+               /* If the connection is attached to a buffer_wait (extremely
+                * rare), it will be woken up at any instant by its own thread
+                * and we can't undo it anyway, so let's give up on this one.
+                * It's not interesting anyway since it's not usable right now.
+                */
+               if (LIST_INLIST(&spop_conn->buf_wait.list))
+                       goto fail;
+
+               new_task = task_new_here();
+               new_tasklet = tasklet_new();
+               if (!new_task || !new_tasklet)
+                       goto fail;
+       }
+
+       if (fd_takeover(conn->handle.fd, conn) != 0)
+               goto fail;
+
+       if (conn->xprt->takeover && conn->xprt->takeover(conn, conn->xprt_ctx, orig_tid, release) != 0) {
+               /* We failed to takeover the xprt, even if the connection may
+                * still be valid, flag it as error'd, as we have already
+                * taken over the fd, and wake the tasklet, so that it will
+                * destroy it.
+                */
+               conn->flags |= CO_FL_ERROR;
+               tasklet_wakeup_on(spop_conn->wait_event.tasklet, orig_tid);
+               goto fail;
+       }
+
+       if (spop_conn->wait_event.events)
+               spop_conn->conn->xprt->unsubscribe(spop_conn->conn, spop_conn->conn->xprt_ctx,
+                   spop_conn->wait_event.events, &spop_conn->wait_event);
+
+       task = spop_conn->task;
+       if (task) {
+               /* only assign a task if there was already one, otherwise
+                * the preallocated new task will be released.
+                */
+               task->context = NULL;
+               spop_conn->task = NULL;
+               __ha_barrier_store();
+               task_kill(task);
+
+               spop_conn->task = new_task;
+               new_task = NULL;
+               if (!release) {
+                       spop_conn->task->process = spop_timeout_task;
+                       spop_conn->task->context = spop_conn;
+               }
+       }
+
+       /* To let the tasklet know it should free itself, and do nothing else,
+        * set its context to NULL;
+        */
+       spop_conn->wait_event.tasklet->context = NULL;
+       tasklet_wakeup_on(spop_conn->wait_event.tasklet, orig_tid);
+
+       spop_conn->wait_event.tasklet = new_tasklet;
+       if (!release) {
+               spop_conn->wait_event.tasklet->process = spop_io_cb;
+               spop_conn->wait_event.tasklet->context = spop_conn;
+               spop_conn->conn->xprt->subscribe(spop_conn->conn, spop_conn->conn->xprt_ctx,
+                                                SUB_RETRY_RECV, &spop_conn->wait_event);
+       }
+
+       if (release) {
+               /* we're being called for a server deletion and are running
+                * under thread isolation. That's the only way we can
+                * unregister a possible subscription of the original
+                * connection from its owner thread's queue, as this involves
+                * manipulating thread-unsafe areas. Note that it is not
+                * possible to just call b_dequeue() here as it would update
+                * the current thread's bufq_map and not the original one.
+                */
+               BUG_ON(!thread_isolated());
+               if (LIST_INLIST(&spop_conn->buf_wait.list))
+                       _b_dequeue(&spop_conn->buf_wait, orig_tid);
+       }
+
+       if (new_task)
+               __task_free(new_task);
+       return 0;
+ fail:
+       if (new_task)
+               __task_free(new_task);
+       tasklet_free(new_tasklet);
+       return -1;
+}
+
+static const struct mux_ops mux_spop_ops = {
+       .init          = spop_init,
+       .wake          = spop_wake,
+       .attach        = spop_attach,
+       .get_first_sc  = spop_get_first_sc,
+       .detach        = spop_detach,
+       .destroy       = spop_destroy,
+       .avail_streams = spop_avail_streams,
+       .used_streams  = spop_used_streams,
+       .rcv_buf       = spop_rcv_buf,
+       .snd_buf       = spop_snd_buf,
+       .subscribe     = spop_subscribe,
+       .unsubscribe   = spop_unsubscribe,
+       .shut          = spop_shut,
+       .ctl           = spop_ctl,
+       .sctl          = spop_sctl,
+       .show_fd       = spop_show_fd,
+       .takeover      = spop_takeover,
+       .flags         = MX_FL_HOL_RISK|MX_FL_NO_UPG,
+       .name          = "SPOP",
+};
+
+static struct mux_proto_list mux_proto_spop =
+       { .token = IST("spop"), .mode = PROTO_MODE_SPOP, .side = PROTO_SIDE_BE, .mux = &mux_spop_ops };
+
+static struct mux_proto_list mux_proto_default_spop =
+       { .token = IST(""), .mode = PROTO_MODE_SPOP, .side = PROTO_SIDE_BE, .mux = &mux_spop_ops };
+
+INITCALL1(STG_REGISTER, register_mux_proto, &mux_proto_spop);
+INITCALL1(STG_REGISTER, register_mux_proto, &mux_proto_default_spop);
+
+/*
+ * Local variables:
+ *  c-indent-level: 8
+ *  c-basic-offset: 8
+ * End:
+ */