Re: [RFC PATCH net-next] ipvs: enable the scheduling of icmp packets

To: Alex Gartrell <agartrell@xxxxxx>
Subject: Re: [RFC PATCH net-next] ipvs: enable the scheduling of icmp packets
Cc: horms@xxxxxxxxxxxx, lvs-devel@xxxxxxxxxxxxxxx, kernel-team@xxxxxx
From: Julian Anastasov <ja@xxxxxx>
Date: Tue, 11 Aug 2015 11:25:41 +0300 (EEST)

        More notes for today:

- ip_vs_in_icmp: lets add a bool flag (local var) if cp was
created from the ip_vs_try_to_schedule call, I prefer if cp
was present to call __ip_vs_conn_put:

        bool old = true;

                if (!ip_vs_try_to_schedule(AF_INET, skb, pd, &v, &cp,
                                           &ciph, !ipip))
                        return v;
                old = false;

        if (old)

- same for ip_vs_in_icmp_v6, we can use single place where to
call [__]ip_vs_conn_put, now they are 2 places

- ensure_mtu_is_adequate:
        - ipvsh->fragoffs can be for embedded header

- ip_vs_bypass_xmit_v6: uses ipvsh->daddr.in6 but it should use
        struct ipv6hdr *iph = ipv6_hdr(skb);

- we need a new flag in struct ip_vs_iphdr:
        - embedded:
                - we are working with embedded header
                - this is needed by:
                        - ensure_mtu_is_adequate to avoid sending
                        ICMP error as response to ICMP error
                - schedule path: to avoid many checks for inverse flag

- ip_vs_in_icmp* can call ip_vs_fill_* with new arg 'inverse', so
that the scheduling path see in ipvsh->saddr always the client address.
The SH scheduler can know that addresses and ports are reversed by
checking the embedded flag which implies 'inverse' mode. Then only
ip_vs_sh_get_port should check this flag. Such flag allows schedulers
to refuse to work for embedded packets, for example, if they
schedule by payload.


Julian Anastasov <ja@xxxxxx>
To unsubscribe from this list: send the line "unsubscribe lvs-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at

<Prev in Thread] Current Thread [Next in Thread>