mirror of
https://github.com/netdata/libbpf.git
synced 2026-04-05 16:19:06 +08:00
sync: auto-generate latest BPF helpers
Latest changes to BPF helper definitions.
This commit is contained in:
committed by
Andrii Nakryiko
parent
c772c9cbde
commit
552af3d963
@@ -576,9 +576,9 @@ static __u64 (*bpf_perf_event_read)(void *map, __u64 flags) = (void *) 22;
|
|||||||
* supports redirection to the egress interface, and accepts no
|
* supports redirection to the egress interface, and accepts no
|
||||||
* flag at all.
|
* flag at all.
|
||||||
*
|
*
|
||||||
* The same effect can be attained with the more generic
|
* The same effect can also be attained with the more generic
|
||||||
* **bpf_redirect_map**\ (), which requires specific maps to be
|
* **bpf_redirect_map**\ (), which uses a BPF map to store the
|
||||||
* used but offers better performance.
|
* redirect target instead of providing it directly to the helper.
|
||||||
*
|
*
|
||||||
* Returns
|
* Returns
|
||||||
* For XDP, the helper returns **XDP_REDIRECT** on success or
|
* For XDP, the helper returns **XDP_REDIRECT** on success or
|
||||||
@@ -1242,14 +1242,12 @@ static int (*bpf_skb_adjust_room)(struct __sk_buff *skb, __s32 len_diff, __u32 m
|
|||||||
* the caller. Any higher bits in the *flags* argument must be
|
* the caller. Any higher bits in the *flags* argument must be
|
||||||
* unset.
|
* unset.
|
||||||
*
|
*
|
||||||
* When used to redirect packets to net devices, this helper
|
* See also bpf_redirect(), which only supports redirecting to an
|
||||||
* provides a high performance increase over **bpf_redirect**\ ().
|
* ifindex, but doesn't require a map to do so.
|
||||||
* This is due to various implementation details of the underlying
|
|
||||||
* mechanisms, one of which is the fact that **bpf_redirect_map**\
|
|
||||||
* () tries to send packet as a "bulk" to the device.
|
|
||||||
*
|
*
|
||||||
* Returns
|
* Returns
|
||||||
* **XDP_REDIRECT** on success, or **XDP_ABORTED** on error.
|
* **XDP_REDIRECT** on success, or the value of the two lower bits
|
||||||
|
* of the **flags* argument on error.
|
||||||
*/
|
*/
|
||||||
static int (*bpf_redirect_map)(void *map, __u32 key, __u64 flags) = (void *) 51;
|
static int (*bpf_redirect_map)(void *map, __u32 key, __u64 flags) = (void *) 51;
|
||||||
|
|
||||||
@@ -2796,4 +2794,27 @@ static int (*bpf_send_signal_thread)(__u32 sig) = (void *) 117;
|
|||||||
*/
|
*/
|
||||||
static __u64 (*bpf_jiffies64)(void) = (void *) 118;
|
static __u64 (*bpf_jiffies64)(void) = (void *) 118;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* bpf_read_branch_records
|
||||||
|
*
|
||||||
|
* For an eBPF program attached to a perf event, retrieve the
|
||||||
|
* branch records (struct perf_branch_entry) associated to *ctx*
|
||||||
|
* and store it in the buffer pointed by *buf* up to size
|
||||||
|
* *size* bytes.
|
||||||
|
*
|
||||||
|
* Returns
|
||||||
|
* On success, number of bytes written to *buf*. On error, a
|
||||||
|
* negative value.
|
||||||
|
*
|
||||||
|
* The *flags* can be set to **BPF_F_GET_BRANCH_RECORDS_SIZE** to
|
||||||
|
* instead return the number of bytes required to store all the
|
||||||
|
* branch entries. If this flag is set, *buf* may be NULL.
|
||||||
|
*
|
||||||
|
* **-EINVAL** if arguments invalid or **size** not a multiple
|
||||||
|
* of sizeof(struct perf_branch_entry).
|
||||||
|
*
|
||||||
|
* **-ENOENT** if architecture does not support branch records.
|
||||||
|
*/
|
||||||
|
static int (*bpf_read_branch_records)(struct bpf_perf_event_data *ctx, void *buf, __u32 size, __u64 flags) = (void *) 119;
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
Reference in New Issue
Block a user