// SPDX-License-Identifier: GPL-2.0 /* Converted from tools/testing/selftests/bpf/verifier/cgroup_skb.c */ #include #include #include "bpf_misc.h" SEC("cgroup/skb") __description("direct packet read test#1 for CGROUP_SKB") __success __failure_unpriv __msg_unpriv("invalid bpf_context access off=76 size=4") __retval(0) __naked void test_1_for_cgroup_skb(void) { asm volatile (" \ r2 = *(u32*)(r1 + %[__sk_buff_data]); \ r3 = *(u32*)(r1 + %[__sk_buff_data_end]); \ r4 = *(u32*)(r1 + %[__sk_buff_len]); \ r5 = *(u32*)(r1 + %[__sk_buff_pkt_type]); \ r6 = *(u32*)(r1 + %[__sk_buff_mark]); \ *(u32*)(r1 + %[__sk_buff_mark]) = r6; \ r7 = *(u32*)(r1 + %[__sk_buff_queue_mapping]); \ r8 = *(u32*)(r1 + %[__sk_buff_protocol]); \ r9 = *(u32*)(r1 + %[__sk_buff_vlan_present]); \ r0 = r2; \ r0 += 8; \ if r0 > r3 goto l0_%=; \ r0 = *(u8*)(r2 + 0); \ l0_%=: r0 = 0; \ exit; \ " : : __imm_const(__sk_buff_data, offsetof(struct __sk_buff, data)), __imm_const(__sk_buff_data_end, offsetof(struct __sk_buff, data_end)), __imm_const(__sk_buff_len, offsetof(struct __sk_buff, len)), __imm_const(__sk_buff_mark, offsetof(struct __sk_buff, mark)), __imm_const(__sk_buff_pkt_type, offsetof(struct __sk_buff, pkt_type)), __imm_const(__sk_buff_protocol, offsetof(struct __sk_buff, protocol)), __imm_const(__sk_buff_queue_mapping, offsetof(struct __sk_buff, queue_mapping)), __imm_const(__sk_buff_vlan_present, offsetof(struct __sk_buff, vlan_present)) : __clobber_all); } SEC("cgroup/skb") __description("direct packet read test#2 for CGROUP_SKB") __success __success_unpriv __retval(0) __naked void test_2_for_cgroup_skb(void) { asm volatile (" \ r4 = *(u32*)(r1 + %[__sk_buff_vlan_tci]); \ r5 = *(u32*)(r1 + %[__sk_buff_vlan_proto]); \ r6 = *(u32*)(r1 + %[__sk_buff_priority]); \ *(u32*)(r1 + %[__sk_buff_priority]) = r6; \ r7 = *(u32*)(r1 + %[__sk_buff_ingress_ifindex]);\ r8 = *(u32*)(r1 + %[__sk_buff_tc_index]); \ r9 = *(u32*)(r1 + %[__sk_buff_hash]); \ r0 = 0; \ exit; \ " : : __imm_const(__sk_buff_hash, offsetof(struct __sk_buff, hash)), __imm_const(__sk_buff_ingress_ifindex, offsetof(struct __sk_buff, ingress_ifindex)), __imm_const(__sk_buff_priority, offsetof(struct __sk_buff, priority)), __imm_const(__sk_buff_tc_index, offsetof(struct __sk_buff, tc_index)), __imm_const(__sk_buff_vlan_proto, offsetof(struct __sk_buff, vlan_proto)), __imm_const(__sk_buff_vlan_tci, offsetof(struct __sk_buff, vlan_tci)) : __clobber_all); } SEC("cgroup/skb") __description("direct packet read test#3 for CGROUP_SKB") __success __success_unpriv __retval(0) __naked void test_3_for_cgroup_skb(void) { asm volatile (" \ r4 = *(u32*)(r1 + %[__sk_buff_cb_0]); \ r5 = *(u32*)(r1 + %[__sk_buff_cb_1]); \ r6 = *(u32*)(r1 + %[__sk_buff_cb_2]); \ r7 = *(u32*)(r1 + %[__sk_buff_cb_3]); \ r8 = *(u32*)(r1 + %[__sk_buff_cb_4]); \ r9 = *(u32*)(r1 + %[__sk_buff_napi_id]); \ *(u32*)(r1 + %[__sk_buff_cb_0]) = r4; \ *(u32*)(r1 + %[__sk_buff_cb_1]) = r5; \ *(u32*)(r1 + %[__sk_buff_cb_2]) = r6; \ *(u32*)(r1 + %[__sk_buff_cb_3]) = r7; \ *(u32*)(r1 + %[__sk_buff_cb_4]) = r8; \ r0 = 0; \ exit; \ " : : __imm_const(__sk_buff_cb_0, offsetof(struct __sk_buff, cb[0])), __imm_const(__sk_buff_cb_1, offsetof(struct __sk_buff, cb[1])), __imm_const(__sk_buff_cb_2, offsetof(struct __sk_buff, cb[2])), __imm_const(__sk_buff_cb_3, offsetof(struct __sk_buff, cb[3])), __imm_const(__sk_buff_cb_4, offsetof(struct __sk_buff, cb[4])), __imm_const(__sk_buff_napi_id, offsetof(struct __sk_buff, napi_id)) : __clobber_all); } SEC("cgroup/skb") __description("direct packet read test#4 for CGROUP_SKB") __success __success_unpriv __retval(0) __naked void test_4_for_cgroup_skb(void) { asm volatile (" \ r2 = *(u32*)(r1 + %[__sk_buff_family]); \ r3 = *(u32*)(r1 + %[__sk_buff_remote_ip4]); \ r4 = *(u32*)(r1 + %[__sk_buff_local_ip4]); \ r5 = *(u32*)(r1 + %[__sk_buff_remote_ip6_0]); \ r5 = *(u32*)(r1 + %[__sk_buff_remote_ip6_1]); \ r5 = *(u32*)(r1 + %[__sk_buff_remote_ip6_2]); \ r5 = *(u32*)(r1 + %[__sk_buff_remote_ip6_3]); \ r6 = *(u32*)(r1 + %[__sk_buff_local_ip6_0]); \ r6 = *(u32*)(r1 + %[__sk_buff_local_ip6_1]); \ r6 = *(u32*)(r1 + %[__sk_buff_local_ip6_2]); \ r6 = *(u32*)(r1 + %[__sk_buff_local_ip6_3]); \ r7 = *(u32*)(r1 + %[__sk_buff_remote_port]); \ r8 = *(u32*)(r1 + %[__sk_buff_local_port]); \ r0 = 0; \ exit; \ " : : __imm_const(__sk_buff_family, offsetof(struct __sk_buff, family)), __imm_const(__sk_buff_local_ip4, offsetof(struct __sk_buff, local_ip4)), __imm_const(__sk_buff_local_ip6_0, offsetof(struct __sk_buff, local_ip6[0])), __imm_const(__sk_buff_local_ip6_1, offsetof(struct __sk_buff, local_ip6[1])), __imm_const(__sk_buff_local_ip6_2, offsetof(struct __sk_buff, local_ip6[2])), __imm_const(__sk_buff_local_ip6_3, offsetof(struct __sk_buff, local_ip6[3])), __imm_const(__sk_buff_local_port, offsetof(struct __sk_buff, local_port)), __imm_const(__sk_buff_remote_ip4, offsetof(struct __sk_buff, remote_ip4)), __imm_const(__sk_buff_remote_ip6_0, offsetof(struct __sk_buff, remote_ip6[0])), __imm_const(__sk_buff_remote_ip6_1, offsetof(struct __sk_buff, remote_ip6[1])), __imm_const(__sk_buff_remote_ip6_2, offsetof(struct __sk_buff, remote_ip6[2])), __imm_const(__sk_buff_remote_ip6_3, offsetof(struct __sk_buff, remote_ip6[3])), __imm_const(__sk_buff_remote_port, offsetof(struct __sk_buff, remote_port)) : __clobber_all); } SEC("cgroup/skb") __description("invalid access of tc_classid for CGROUP_SKB") __failure __msg("invalid bpf_context access") __failure_unpriv __naked void tc_classid_for_cgroup_skb(void) { asm volatile (" \ r0 = *(u32*)(r1 + %[__sk_buff_tc_classid]); \ r0 = 0; \ exit; \ " : : __imm_const(__sk_buff_tc_classid, offsetof(struct __sk_buff, tc_classid)) : __clobber_all); } SEC("cgroup/skb") __description("invalid access of data_meta for CGROUP_SKB") __failure __msg("invalid bpf_context access") __failure_unpriv __naked void data_meta_for_cgroup_skb(void) { asm volatile (" \ r0 = *(u32*)(r1 + %[__sk_buff_data_meta]); \ r0 = 0; \ exit; \ " : : __imm_const(__sk_buff_data_meta, offsetof(struct __sk_buff, data_meta)) : __clobber_all); } SEC("cgroup/skb") __description("invalid access of flow_keys for CGROUP_SKB") __failure __msg("invalid bpf_context access") __failure_unpriv __naked void flow_keys_for_cgroup_skb(void) { asm volatile (" \ r0 = *(u32*)(r1 + %[__sk_buff_flow_keys]); \ r0 = 0; \ exit; \ " : : __imm_const(__sk_buff_flow_keys, offsetof(struct __sk_buff, flow_keys)) : __clobber_all); } SEC("cgroup/skb") __description("invalid write access to napi_id for CGROUP_SKB") __failure __msg("invalid bpf_context access") __failure_unpriv __naked void napi_id_for_cgroup_skb(void) { asm volatile (" \ r9 = *(u32*)(r1 + %[__sk_buff_napi_id]); \ *(u32*)(r1 + %[__sk_buff_napi_id]) = r9; \ r0 = 0; \ exit; \ " : : __imm_const(__sk_buff_napi_id, offsetof(struct __sk_buff, napi_id)) : __clobber_all); } SEC("cgroup/skb") __description("write tstamp from CGROUP_SKB") __success __failure_unpriv __msg_unpriv("invalid bpf_context access off=152 size=8") __retval(0) __naked void write_tstamp_from_cgroup_skb(void) { asm volatile (" \ r0 = 0; \ *(u64*)(r1 + %[__sk_buff_tstamp]) = r0; \ r0 = 0; \ exit; \ " : : __imm_const(__sk_buff_tstamp, offsetof(struct __sk_buff, tstamp)) : __clobber_all); } SEC("cgroup/skb") __description("read tstamp from CGROUP_SKB") __success __success_unpriv __retval(0) __naked void read_tstamp_from_cgroup_skb(void) { asm volatile (" \ r0 = *(u64*)(r1 + %[__sk_buff_tstamp]); \ r0 = 0; \ exit; \ " : : __imm_const(__sk_buff_tstamp, offsetof(struct __sk_buff, tstamp)) : __clobber_all); } char _license[] SEC("license") = "GPL";