Skip to content

Commit 2929ad2

Browse files
author
Alexei Starovoitov
committed
Merge branch 'improve_perf_barriers'
Daniel Borkmann says: ==================== This set first adds smp_* barrier variants to tools infrastructure and updates perf and libbpf to make use of them. For details, please see individual patches, thanks! Arnaldo, if there are no objections, could this be routed via bpf-next with Acked-by's due to later dependencies in libbpf? Alternatively, I could also get the 2nd patch out during merge window, but perhaps it's okay to do in one go as there shouldn't be much conflict in perf itself. Thanks! v1 -> v2: - add common helper and switch to acquire/release variants when possible, thanks Peter! ==================== Signed-off-by: Alexei Starovoitov <ast@kernel.org>
2 parents 78de354 + a64af0e commit 2929ad2

File tree

10 files changed

+254
-18
lines changed

10 files changed

+254
-18
lines changed

tools/arch/arm64/include/asm/barrier.h

Lines changed: 70 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -14,4 +14,74 @@
1414
#define wmb() asm volatile("dmb ishst" ::: "memory")
1515
#define rmb() asm volatile("dmb ishld" ::: "memory")
1616

17+
#define smp_store_release(p, v) \
18+
do { \
19+
union { typeof(*p) __val; char __c[1]; } __u = \
20+
{ .__val = (__force typeof(*p)) (v) }; \
21+
\
22+
switch (sizeof(*p)) { \
23+
case 1: \
24+
asm volatile ("stlrb %w1, %0" \
25+
: "=Q" (*p) \
26+
: "r" (*(__u8 *)__u.__c) \
27+
: "memory"); \
28+
break; \
29+
case 2: \
30+
asm volatile ("stlrh %w1, %0" \
31+
: "=Q" (*p) \
32+
: "r" (*(__u16 *)__u.__c) \
33+
: "memory"); \
34+
break; \
35+
case 4: \
36+
asm volatile ("stlr %w1, %0" \
37+
: "=Q" (*p) \
38+
: "r" (*(__u32 *)__u.__c) \
39+
: "memory"); \
40+
break; \
41+
case 8: \
42+
asm volatile ("stlr %1, %0" \
43+
: "=Q" (*p) \
44+
: "r" (*(__u64 *)__u.__c) \
45+
: "memory"); \
46+
break; \
47+
default: \
48+
/* Only to shut up gcc ... */ \
49+
mb(); \
50+
break; \
51+
} \
52+
} while (0)
53+
54+
#define smp_load_acquire(p) \
55+
({ \
56+
union { typeof(*p) __val; char __c[1]; } __u; \
57+
\
58+
switch (sizeof(*p)) { \
59+
case 1: \
60+
asm volatile ("ldarb %w0, %1" \
61+
: "=r" (*(__u8 *)__u.__c) \
62+
: "Q" (*p) : "memory"); \
63+
break; \
64+
case 2: \
65+
asm volatile ("ldarh %w0, %1" \
66+
: "=r" (*(__u16 *)__u.__c) \
67+
: "Q" (*p) : "memory"); \
68+
break; \
69+
case 4: \
70+
asm volatile ("ldar %w0, %1" \
71+
: "=r" (*(__u32 *)__u.__c) \
72+
: "Q" (*p) : "memory"); \
73+
break; \
74+
case 8: \
75+
asm volatile ("ldar %0, %1" \
76+
: "=r" (*(__u64 *)__u.__c) \
77+
: "Q" (*p) : "memory"); \
78+
break; \
79+
default: \
80+
/* Only to shut up gcc ... */ \
81+
mb(); \
82+
break; \
83+
} \
84+
__u.__val; \
85+
})
86+
1787
#endif /* _TOOLS_LINUX_ASM_AARCH64_BARRIER_H */

tools/arch/ia64/include/asm/barrier.h

Lines changed: 13 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -46,4 +46,17 @@
4646
#define rmb() mb()
4747
#define wmb() mb()
4848

49+
#define smp_store_release(p, v) \
50+
do { \
51+
barrier(); \
52+
WRITE_ONCE(*p, v); \
53+
} while (0)
54+
55+
#define smp_load_acquire(p) \
56+
({ \
57+
typeof(*p) ___p1 = READ_ONCE(*p); \
58+
barrier(); \
59+
___p1; \
60+
})
61+
4962
#endif /* _TOOLS_LINUX_ASM_IA64_BARRIER_H */

tools/arch/powerpc/include/asm/barrier.h

Lines changed: 16 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -27,4 +27,20 @@
2727
#define rmb() __asm__ __volatile__ ("sync" : : : "memory")
2828
#define wmb() __asm__ __volatile__ ("sync" : : : "memory")
2929

30+
#if defined(__powerpc64__)
31+
#define smp_lwsync() __asm__ __volatile__ ("lwsync" : : : "memory")
32+
33+
#define smp_store_release(p, v) \
34+
do { \
35+
smp_lwsync(); \
36+
WRITE_ONCE(*p, v); \
37+
} while (0)
38+
39+
#define smp_load_acquire(p) \
40+
({ \
41+
typeof(*p) ___p1 = READ_ONCE(*p); \
42+
smp_lwsync(); \
43+
___p1; \
44+
})
45+
#endif /* defined(__powerpc64__) */
3046
#endif /* _TOOLS_LINUX_ASM_POWERPC_BARRIER_H */

tools/arch/s390/include/asm/barrier.h

Lines changed: 13 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -28,4 +28,17 @@
2828
#define rmb() mb()
2929
#define wmb() mb()
3030

31+
#define smp_store_release(p, v) \
32+
do { \
33+
barrier(); \
34+
WRITE_ONCE(*p, v); \
35+
} while (0)
36+
37+
#define smp_load_acquire(p) \
38+
({ \
39+
typeof(*p) ___p1 = READ_ONCE(*p); \
40+
barrier(); \
41+
___p1; \
42+
})
43+
3144
#endif /* __TOOLS_LIB_ASM_BARRIER_H */

tools/arch/sparc/include/asm/barrier_64.h

Lines changed: 13 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -40,4 +40,17 @@ do { __asm__ __volatile__("ba,pt %%xcc, 1f\n\t" \
4040
#define rmb() __asm__ __volatile__("":::"memory")
4141
#define wmb() __asm__ __volatile__("":::"memory")
4242

43+
#define smp_store_release(p, v) \
44+
do { \
45+
barrier(); \
46+
WRITE_ONCE(*p, v); \
47+
} while (0)
48+
49+
#define smp_load_acquire(p) \
50+
({ \
51+
typeof(*p) ___p1 = READ_ONCE(*p); \
52+
barrier(); \
53+
___p1; \
54+
})
55+
4356
#endif /* !(__TOOLS_LINUX_SPARC64_BARRIER_H) */

tools/arch/x86/include/asm/barrier.h

Lines changed: 14 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -26,4 +26,18 @@
2626
#define wmb() asm volatile("sfence" ::: "memory")
2727
#endif
2828

29+
#if defined(__x86_64__)
30+
#define smp_store_release(p, v) \
31+
do { \
32+
barrier(); \
33+
WRITE_ONCE(*p, v); \
34+
} while (0)
35+
36+
#define smp_load_acquire(p) \
37+
({ \
38+
typeof(*p) ___p1 = READ_ONCE(*p); \
39+
barrier(); \
40+
___p1; \
41+
})
42+
#endif /* defined(__x86_64__) */
2943
#endif /* _TOOLS_LINUX_ASM_X86_BARRIER_H */

tools/include/asm/barrier.h

Lines changed: 35 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,5 @@
11
/* SPDX-License-Identifier: GPL-2.0 */
2+
#include <linux/compiler.h>
23
#if defined(__i386__) || defined(__x86_64__)
34
#include "../../arch/x86/include/asm/barrier.h"
45
#elif defined(__arm__)
@@ -26,3 +27,37 @@
2627
#else
2728
#include <asm-generic/barrier.h>
2829
#endif
30+
31+
/*
32+
* Generic fallback smp_*() definitions for archs that haven't
33+
* been updated yet.
34+
*/
35+
36+
#ifndef smp_rmb
37+
# define smp_rmb() rmb()
38+
#endif
39+
40+
#ifndef smp_wmb
41+
# define smp_wmb() wmb()
42+
#endif
43+
44+
#ifndef smp_mb
45+
# define smp_mb() mb()
46+
#endif
47+
48+
#ifndef smp_store_release
49+
# define smp_store_release(p, v) \
50+
do { \
51+
smp_mb(); \
52+
WRITE_ONCE(*p, v); \
53+
} while (0)
54+
#endif
55+
56+
#ifndef smp_load_acquire
57+
# define smp_load_acquire(p) \
58+
({ \
59+
typeof(*p) ___p1 = READ_ONCE(*p); \
60+
smp_mb(); \
61+
___p1; \
62+
})
63+
#endif

tools/include/linux/ring_buffer.h

Lines changed: 73 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,73 @@
1+
#ifndef _TOOLS_LINUX_RING_BUFFER_H_
2+
#define _TOOLS_LINUX_RING_BUFFER_H_
3+
4+
#include <asm/barrier.h>
5+
6+
/*
7+
* Contract with kernel for walking the perf ring buffer from
8+
* user space requires the following barrier pairing (quote
9+
* from kernel/events/ring_buffer.c):
10+
*
11+
* Since the mmap() consumer (userspace) can run on a
12+
* different CPU:
13+
*
14+
* kernel user
15+
*
16+
* if (LOAD ->data_tail) { LOAD ->data_head
17+
* (A) smp_rmb() (C)
18+
* STORE $data LOAD $data
19+
* smp_wmb() (B) smp_mb() (D)
20+
* STORE ->data_head STORE ->data_tail
21+
* }
22+
*
23+
* Where A pairs with D, and B pairs with C.
24+
*
25+
* In our case A is a control dependency that separates the
26+
* load of the ->data_tail and the stores of $data. In case
27+
* ->data_tail indicates there is no room in the buffer to
28+
* store $data we do not.
29+
*
30+
* D needs to be a full barrier since it separates the data
31+
* READ from the tail WRITE.
32+
*
33+
* For B a WMB is sufficient since it separates two WRITEs,
34+
* and for C an RMB is sufficient since it separates two READs.
35+
*
36+
* Note, instead of B, C, D we could also use smp_store_release()
37+
* in B and D as well as smp_load_acquire() in C.
38+
*
39+
* However, this optimization does not make sense for all kernel
40+
* supported architectures since for a fair number it would
41+
* resolve into READ_ONCE() + smp_mb() pair for smp_load_acquire(),
42+
* and smp_mb() + WRITE_ONCE() pair for smp_store_release().
43+
*
44+
* Thus for those smp_wmb() in B and smp_rmb() in C would still
45+
* be less expensive. For the case of D this has either the same
46+
* cost or is less expensive, for example, due to TSO x86 can
47+
* avoid the CPU barrier entirely.
48+
*/
49+
50+
static inline u64 ring_buffer_read_head(struct perf_event_mmap_page *base)
51+
{
52+
/*
53+
* Architectures where smp_load_acquire() does not fallback to
54+
* READ_ONCE() + smp_mb() pair.
55+
*/
56+
#if defined(__x86_64__) || defined(__aarch64__) || defined(__powerpc64__) || \
57+
defined(__ia64__) || defined(__sparc__) && defined(__arch64__)
58+
return smp_load_acquire(&base->data_head);
59+
#else
60+
u64 head = READ_ONCE(base->data_head);
61+
62+
smp_rmb();
63+
return head;
64+
#endif
65+
}
66+
67+
static inline void ring_buffer_write_tail(struct perf_event_mmap_page *base,
68+
u64 tail)
69+
{
70+
smp_store_release(&base->data_tail, tail);
71+
}
72+
73+
#endif /* _TOOLS_LINUX_RING_BUFFER_H_ */

tools/lib/bpf/libbpf.c

Lines changed: 4 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -27,6 +27,7 @@
2727
#include <linux/list.h>
2828
#include <linux/limits.h>
2929
#include <linux/perf_event.h>
30+
#include <linux/ring_buffer.h>
3031
#include <sys/stat.h>
3132
#include <sys/types.h>
3233
#include <sys/vfs.h>
@@ -2418,13 +2419,12 @@ bpf_perf_event_read_simple(void *mem, unsigned long size,
24182419
unsigned long page_size, void **buf, size_t *buf_len,
24192420
bpf_perf_event_print_t fn, void *priv)
24202421
{
2421-
volatile struct perf_event_mmap_page *header = mem;
2422+
struct perf_event_mmap_page *header = mem;
2423+
__u64 data_head = ring_buffer_read_head(header);
24222424
__u64 data_tail = header->data_tail;
2423-
__u64 data_head = header->data_head;
24242425
int ret = LIBBPF_PERF_EVENT_ERROR;
24252426
void *base, *begin, *end;
24262427

2427-
asm volatile("" ::: "memory"); /* in real code it should be smp_rmb() */
24282428
if (data_head == data_tail)
24292429
return LIBBPF_PERF_EVENT_CONT;
24302430

@@ -2467,8 +2467,6 @@ bpf_perf_event_read_simple(void *mem, unsigned long size,
24672467
data_tail += ehdr->size;
24682468
}
24692469

2470-
__sync_synchronize(); /* smp_mb() */
2471-
header->data_tail = data_tail;
2472-
2470+
ring_buffer_write_tail(header, data_tail);
24732471
return ret;
24742472
}

tools/perf/util/mmap.h

Lines changed: 3 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@
44
#include <linux/compiler.h>
55
#include <linux/refcount.h>
66
#include <linux/types.h>
7-
#include <asm/barrier.h>
7+
#include <linux/ring_buffer.h>
88
#include <stdbool.h>
99
#include "auxtrace.h"
1010
#include "event.h"
@@ -71,21 +71,12 @@ void perf_mmap__consume(struct perf_mmap *map);
7171

7272
static inline u64 perf_mmap__read_head(struct perf_mmap *mm)
7373
{
74-
struct perf_event_mmap_page *pc = mm->base;
75-
u64 head = READ_ONCE(pc->data_head);
76-
rmb();
77-
return head;
74+
return ring_buffer_read_head(mm->base);
7875
}
7976

8077
static inline void perf_mmap__write_tail(struct perf_mmap *md, u64 tail)
8178
{
82-
struct perf_event_mmap_page *pc = md->base;
83-
84-
/*
85-
* ensure all reads are done before we write the tail out.
86-
*/
87-
mb();
88-
pc->data_tail = tail;
79+
ring_buffer_write_tail(md->base, tail);
8980
}
9081

9182
union perf_event *perf_mmap__read_forward(struct perf_mmap *map);

0 commit comments

Comments
 (0)