@@ -141,6 +141,7 @@
#include <linux/netfilter_ingress.h>
#include <linux/sctp.h>
#include <linux/crash_dump.h>
+#include <linux/bpf-cgroup.h>
#include "net-sysfs.h"
@@ -3329,6 +3330,10 @@ static int __dev_queue_xmit(struct sk_buff *skb, void *accel_priv)
if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_SCHED_TSTAMP))
__skb_tstamp_tx(skb, NULL, skb->sk, SCM_TSTAMP_SCHED);
+ rc = cgroup_bpf_run_filter(skb->sk, skb, BPF_CGROUP_INET_EGRESS);
+ if (rc)
+ goto free_skb_list;
+
/* Disable soft irqs for various locks below. Also
* stops preemption for RCU.
*/
@@ -3414,8 +3419,8 @@ recursion_alert:
rc = -ENETDOWN;
rcu_read_unlock_bh();
-
atomic_long_inc(&dev->tx_dropped);
+free_skb_list:
kfree_skb_list(skb);
return rc;
out:
If the cgroup associated with the receiving socket has an eBPF programs installed, run them from __dev_queue_xmit(). eBPF programs used in this context are expected to either return 1 to let the packet pass, or != 1 to drop them. The programs have access to the full skb, including the MAC headers. Note that cgroup_bpf_run_filter() is stubbed out as static inline nop for !CONFIG_CGROUP_BPF, and is otherwise guarded by a static key if the feature is unused. Signed-off-by: Daniel Mack <daniel@zonque.org> --- net/core/dev.c | 7 ++++++- 1 file changed, 6 insertions(+), 1 deletion(-)