Back to home page

OSCL-LXR

 
 

    


0001 // SPDX-License-Identifier: GPL-2.0
0002 /*
0003  * This driver enables Trace Buffer Extension (TRBE) as a per-cpu coresight
0004  * sink device could then pair with an appropriate per-cpu coresight source
0005  * device (ETE) thus generating required trace data. Trace can be enabled
0006  * via the perf framework.
0007  *
0008  * The AUX buffer handling is inspired from Arm SPE PMU driver.
0009  *
0010  * Copyright (C) 2020 ARM Ltd.
0011  *
0012  * Author: Anshuman Khandual <anshuman.khandual@arm.com>
0013  */
0014 #define DRVNAME "arm_trbe"
0015 
0016 #define pr_fmt(fmt) DRVNAME ": " fmt
0017 
0018 #include <asm/barrier.h>
0019 #include <asm/cpufeature.h>
0020 
0021 #include "coresight-self-hosted-trace.h"
0022 #include "coresight-trbe.h"
0023 
0024 #define PERF_IDX2OFF(idx, buf) ((idx) % ((buf)->nr_pages << PAGE_SHIFT))
0025 
0026 /*
0027  * A padding packet that will help the user space tools
0028  * in skipping relevant sections in the captured trace
0029  * data which could not be decoded. TRBE doesn't support
0030  * formatting the trace data, unlike the legacy CoreSight
0031  * sinks and thus we use ETE trace packets to pad the
0032  * sections of the buffer.
0033  */
0034 #define ETE_IGNORE_PACKET       0x70
0035 
0036 /*
0037  * Minimum amount of meaningful trace will contain:
0038  * A-Sync, Trace Info, Trace On, Address, Atom.
0039  * This is about 44bytes of ETE trace. To be on
0040  * the safer side, we assume 64bytes is the minimum
0041  * space required for a meaningful session, before
0042  * we hit a "WRAP" event.
0043  */
0044 #define TRBE_TRACE_MIN_BUF_SIZE     64
0045 
0046 enum trbe_fault_action {
0047     TRBE_FAULT_ACT_WRAP,
0048     TRBE_FAULT_ACT_SPURIOUS,
0049     TRBE_FAULT_ACT_FATAL,
0050 };
0051 
0052 struct trbe_buf {
0053     /*
0054      * Even though trbe_base represents vmap()
0055      * mapped allocated buffer's start address,
0056      * it's being as unsigned long for various
0057      * arithmetic and comparision operations &
0058      * also to be consistent with trbe_write &
0059      * trbe_limit sibling pointers.
0060      */
0061     unsigned long trbe_base;
0062     /* The base programmed into the TRBE */
0063     unsigned long trbe_hw_base;
0064     unsigned long trbe_limit;
0065     unsigned long trbe_write;
0066     int nr_pages;
0067     void **pages;
0068     bool snapshot;
0069     struct trbe_cpudata *cpudata;
0070 };
0071 
0072 /*
0073  * TRBE erratum list
0074  *
0075  * The errata are defined in arm64 generic cpu_errata framework.
0076  * Since the errata work arounds could be applied individually
0077  * to the affected CPUs inside the TRBE driver, we need to know if
0078  * a given CPU is affected by the erratum. Unlike the other erratum
0079  * work arounds, TRBE driver needs to check multiple times during
0080  * a trace session. Thus we need a quicker access to per-CPU
0081  * errata and not issue costly this_cpu_has_cap() everytime.
0082  * We keep a set of the affected errata in trbe_cpudata, per TRBE.
0083  *
0084  * We rely on the corresponding cpucaps to be defined for a given
0085  * TRBE erratum. We map the given cpucap into a TRBE internal number
0086  * to make the tracking of the errata lean.
0087  *
0088  * This helps in :
0089  *   - Not duplicating the detection logic
0090  *   - Streamlined detection of erratum across the system
0091  */
0092 #define TRBE_WORKAROUND_OVERWRITE_FILL_MODE 0
0093 #define TRBE_WORKAROUND_WRITE_OUT_OF_RANGE  1
0094 #define TRBE_NEEDS_DRAIN_AFTER_DISABLE      2
0095 #define TRBE_NEEDS_CTXT_SYNC_AFTER_ENABLE   3
0096 #define TRBE_IS_BROKEN              4
0097 
0098 static int trbe_errata_cpucaps[] = {
0099     [TRBE_WORKAROUND_OVERWRITE_FILL_MODE] = ARM64_WORKAROUND_TRBE_OVERWRITE_FILL_MODE,
0100     [TRBE_WORKAROUND_WRITE_OUT_OF_RANGE] = ARM64_WORKAROUND_TRBE_WRITE_OUT_OF_RANGE,
0101     [TRBE_NEEDS_DRAIN_AFTER_DISABLE] = ARM64_WORKAROUND_2064142,
0102     [TRBE_NEEDS_CTXT_SYNC_AFTER_ENABLE] = ARM64_WORKAROUND_2038923,
0103     [TRBE_IS_BROKEN] = ARM64_WORKAROUND_1902691,
0104     -1,     /* Sentinel, must be the last entry */
0105 };
0106 
0107 /* The total number of listed errata in trbe_errata_cpucaps */
0108 #define TRBE_ERRATA_MAX         (ARRAY_SIZE(trbe_errata_cpucaps) - 1)
0109 
0110 /*
0111  * Safe limit for the number of bytes that may be overwritten
0112  * when ARM64_WORKAROUND_TRBE_OVERWRITE_FILL_MODE is triggered.
0113  */
0114 #define TRBE_WORKAROUND_OVERWRITE_FILL_MODE_SKIP_BYTES  256
0115 
0116 /*
0117  * struct trbe_cpudata: TRBE instance specific data
0118  * @trbe_flag       - TRBE dirty/access flag support
0119  * @trbe_hw_align   - Actual TRBE alignment required for TRBPTR_EL1.
0120  * @trbe_align      - Software alignment used for the TRBPTR_EL1.
0121  * @cpu         - CPU this TRBE belongs to.
0122  * @mode        - Mode of current operation. (perf/disabled)
0123  * @drvdata     - TRBE specific drvdata
0124  * @errata      - Bit map for the errata on this TRBE.
0125  */
0126 struct trbe_cpudata {
0127     bool trbe_flag;
0128     u64 trbe_hw_align;
0129     u64 trbe_align;
0130     int cpu;
0131     enum cs_mode mode;
0132     struct trbe_buf *buf;
0133     struct trbe_drvdata *drvdata;
0134     DECLARE_BITMAP(errata, TRBE_ERRATA_MAX);
0135 };
0136 
0137 struct trbe_drvdata {
0138     struct trbe_cpudata __percpu *cpudata;
0139     struct perf_output_handle * __percpu *handle;
0140     struct hlist_node hotplug_node;
0141     int irq;
0142     cpumask_t supported_cpus;
0143     enum cpuhp_state trbe_online;
0144     struct platform_device *pdev;
0145 };
0146 
0147 static void trbe_check_errata(struct trbe_cpudata *cpudata)
0148 {
0149     int i;
0150 
0151     for (i = 0; i < TRBE_ERRATA_MAX; i++) {
0152         int cap = trbe_errata_cpucaps[i];
0153 
0154         if (WARN_ON_ONCE(cap < 0))
0155             return;
0156         if (this_cpu_has_cap(cap))
0157             set_bit(i, cpudata->errata);
0158     }
0159 }
0160 
0161 static inline bool trbe_has_erratum(struct trbe_cpudata *cpudata, int i)
0162 {
0163     return (i < TRBE_ERRATA_MAX) && test_bit(i, cpudata->errata);
0164 }
0165 
0166 static inline bool trbe_may_overwrite_in_fill_mode(struct trbe_cpudata *cpudata)
0167 {
0168     return trbe_has_erratum(cpudata, TRBE_WORKAROUND_OVERWRITE_FILL_MODE);
0169 }
0170 
0171 static inline bool trbe_may_write_out_of_range(struct trbe_cpudata *cpudata)
0172 {
0173     return trbe_has_erratum(cpudata, TRBE_WORKAROUND_WRITE_OUT_OF_RANGE);
0174 }
0175 
0176 static inline bool trbe_needs_drain_after_disable(struct trbe_cpudata *cpudata)
0177 {
0178     /*
0179      * Errata affected TRBE implementation will need TSB CSYNC and
0180      * DSB in order to prevent subsequent writes into certain TRBE
0181      * system registers from being ignored and not effected.
0182      */
0183     return trbe_has_erratum(cpudata, TRBE_NEEDS_DRAIN_AFTER_DISABLE);
0184 }
0185 
0186 static inline bool trbe_needs_ctxt_sync_after_enable(struct trbe_cpudata *cpudata)
0187 {
0188     /*
0189      * Errata affected TRBE implementation will need an additional
0190      * context synchronization in order to prevent an inconsistent
0191      * TRBE prohibited region view on the CPU which could possibly
0192      * corrupt the TRBE buffer or the TRBE state.
0193      */
0194     return trbe_has_erratum(cpudata, TRBE_NEEDS_CTXT_SYNC_AFTER_ENABLE);
0195 }
0196 
0197 static inline bool trbe_is_broken(struct trbe_cpudata *cpudata)
0198 {
0199     return trbe_has_erratum(cpudata, TRBE_IS_BROKEN);
0200 }
0201 
0202 static int trbe_alloc_node(struct perf_event *event)
0203 {
0204     if (event->cpu == -1)
0205         return NUMA_NO_NODE;
0206     return cpu_to_node(event->cpu);
0207 }
0208 
0209 static inline void trbe_drain_buffer(void)
0210 {
0211     tsb_csync();
0212     dsb(nsh);
0213 }
0214 
0215 static inline void set_trbe_enabled(struct trbe_cpudata *cpudata, u64 trblimitr)
0216 {
0217     /*
0218      * Enable the TRBE without clearing LIMITPTR which
0219      * might be required for fetching the buffer limits.
0220      */
0221     trblimitr |= TRBLIMITR_ENABLE;
0222     write_sysreg_s(trblimitr, SYS_TRBLIMITR_EL1);
0223 
0224     /* Synchronize the TRBE enable event */
0225     isb();
0226 
0227     if (trbe_needs_ctxt_sync_after_enable(cpudata))
0228         isb();
0229 }
0230 
0231 static inline void set_trbe_disabled(struct trbe_cpudata *cpudata)
0232 {
0233     u64 trblimitr = read_sysreg_s(SYS_TRBLIMITR_EL1);
0234 
0235     /*
0236      * Disable the TRBE without clearing LIMITPTR which
0237      * might be required for fetching the buffer limits.
0238      */
0239     trblimitr &= ~TRBLIMITR_ENABLE;
0240     write_sysreg_s(trblimitr, SYS_TRBLIMITR_EL1);
0241 
0242     if (trbe_needs_drain_after_disable(cpudata))
0243         trbe_drain_buffer();
0244     isb();
0245 }
0246 
0247 static void trbe_drain_and_disable_local(struct trbe_cpudata *cpudata)
0248 {
0249     trbe_drain_buffer();
0250     set_trbe_disabled(cpudata);
0251 }
0252 
0253 static void trbe_reset_local(struct trbe_cpudata *cpudata)
0254 {
0255     trbe_drain_and_disable_local(cpudata);
0256     write_sysreg_s(0, SYS_TRBLIMITR_EL1);
0257     write_sysreg_s(0, SYS_TRBPTR_EL1);
0258     write_sysreg_s(0, SYS_TRBBASER_EL1);
0259     write_sysreg_s(0, SYS_TRBSR_EL1);
0260 }
0261 
0262 static void trbe_report_wrap_event(struct perf_output_handle *handle)
0263 {
0264     /*
0265      * Mark the buffer to indicate that there was a WRAP event by
0266      * setting the COLLISION flag. This indicates to the user that
0267      * the TRBE trace collection was stopped without stopping the
0268      * ETE and thus there might be some amount of trace that was
0269      * lost between the time the WRAP was detected and the IRQ
0270      * was consumed by the CPU.
0271      *
0272      * Setting the TRUNCATED flag would move the event to STOPPED
0273      * state unnecessarily, even when there is space left in the
0274      * ring buffer. Using the COLLISION flag doesn't have this side
0275      * effect. We only set TRUNCATED flag when there is no space
0276      * left in the ring buffer.
0277      */
0278     perf_aux_output_flag(handle, PERF_AUX_FLAG_COLLISION);
0279 }
0280 
0281 static void trbe_stop_and_truncate_event(struct perf_output_handle *handle)
0282 {
0283     struct trbe_buf *buf = etm_perf_sink_config(handle);
0284 
0285     /*
0286      * We cannot proceed with the buffer collection and we
0287      * do not have any data for the current session. The
0288      * etm_perf driver expects to close out the aux_buffer
0289      * at event_stop(). So disable the TRBE here and leave
0290      * the update_buffer() to return a 0 size.
0291      */
0292     trbe_drain_and_disable_local(buf->cpudata);
0293     perf_aux_output_flag(handle, PERF_AUX_FLAG_TRUNCATED);
0294     perf_aux_output_end(handle, 0);
0295     *this_cpu_ptr(buf->cpudata->drvdata->handle) = NULL;
0296 }
0297 
0298 /*
0299  * TRBE Buffer Management
0300  *
0301  * The TRBE buffer spans from the base pointer till the limit pointer. When enabled,
0302  * it starts writing trace data from the write pointer onward till the limit pointer.
0303  * When the write pointer reaches the address just before the limit pointer, it gets
0304  * wrapped around again to the base pointer. This is called a TRBE wrap event, which
0305  * generates a maintenance interrupt when operated in WRAP or FILL mode. This driver
0306  * uses FILL mode, where the TRBE stops the trace collection at wrap event. The IRQ
0307  * handler updates the AUX buffer and re-enables the TRBE with updated WRITE and
0308  * LIMIT pointers.
0309  *
0310  *  Wrap around with an IRQ
0311  *  ------ < ------ < ------- < ----- < -----
0312  *  |                   |
0313  *  ------ > ------ > ------- > ----- > -----
0314  *
0315  *  +---------------+-----------------------+
0316  *  |       |           |
0317  *  +---------------+-----------------------+
0318  *  Base Pointer    Write Pointer       Limit Pointer
0319  *
0320  * The base and limit pointers always needs to be PAGE_SIZE aligned. But the write
0321  * pointer can be aligned to the implementation defined TRBE trace buffer alignment
0322  * as captured in trbe_cpudata->trbe_align.
0323  *
0324  *
0325  *      head        tail        wakeup
0326  *  +---------------------------------------+----- ~ ~ ------
0327  *  |$$$$$$$|################|$$$$$$$$$$$$$$|       |
0328  *  +---------------------------------------+----- ~ ~ ------
0329  *  Base Pointer    Write Pointer       Limit Pointer
0330  *
0331  * The perf_output_handle indices (head, tail, wakeup) are monotonically increasing
0332  * values which tracks all the driver writes and user reads from the perf auxiliary
0333  * buffer. Generally [head..tail] is the area where the driver can write into unless
0334  * the wakeup is behind the tail. Enabled TRBE buffer span needs to be adjusted and
0335  * configured depending on the perf_output_handle indices, so that the driver does
0336  * not override into areas in the perf auxiliary buffer which is being or yet to be
0337  * consumed from the user space. The enabled TRBE buffer area is a moving subset of
0338  * the allocated perf auxiliary buffer.
0339  */
0340 
0341 static void __trbe_pad_buf(struct trbe_buf *buf, u64 offset, int len)
0342 {
0343     memset((void *)buf->trbe_base + offset, ETE_IGNORE_PACKET, len);
0344 }
0345 
0346 static void trbe_pad_buf(struct perf_output_handle *handle, int len)
0347 {
0348     struct trbe_buf *buf = etm_perf_sink_config(handle);
0349     u64 head = PERF_IDX2OFF(handle->head, buf);
0350 
0351     __trbe_pad_buf(buf, head, len);
0352     if (!buf->snapshot)
0353         perf_aux_output_skip(handle, len);
0354 }
0355 
0356 static unsigned long trbe_snapshot_offset(struct perf_output_handle *handle)
0357 {
0358     struct trbe_buf *buf = etm_perf_sink_config(handle);
0359 
0360     /*
0361      * The ETE trace has alignment synchronization packets allowing
0362      * the decoder to reset in case of an overflow or corruption.
0363      * So we can use the entire buffer for the snapshot mode.
0364      */
0365     return buf->nr_pages * PAGE_SIZE;
0366 }
0367 
0368 static u64 trbe_min_trace_buf_size(struct perf_output_handle *handle)
0369 {
0370     u64 size = TRBE_TRACE_MIN_BUF_SIZE;
0371     struct trbe_buf *buf = etm_perf_sink_config(handle);
0372     struct trbe_cpudata *cpudata = buf->cpudata;
0373 
0374     /*
0375      * When the TRBE is affected by an erratum that could make it
0376      * write to the next "virtually addressed" page beyond the LIMIT.
0377      * We need to make sure there is always a PAGE after the LIMIT,
0378      * within the buffer. Thus we ensure there is at least an extra
0379      * page than normal. With this we could then adjust the LIMIT
0380      * pointer down by a PAGE later.
0381      */
0382     if (trbe_may_write_out_of_range(cpudata))
0383         size += PAGE_SIZE;
0384     return size;
0385 }
0386 
0387 /*
0388  * TRBE Limit Calculation
0389  *
0390  * The following markers are used to illustrate various TRBE buffer situations.
0391  *
0392  * $$$$ - Data area, unconsumed captured trace data, not to be overridden
0393  * #### - Free area, enabled, trace will be written
0394  * %%%% - Free area, disabled, trace will not be written
0395  * ==== - Free area, padded with ETE_IGNORE_PACKET, trace will be skipped
0396  */
0397 static unsigned long __trbe_normal_offset(struct perf_output_handle *handle)
0398 {
0399     struct trbe_buf *buf = etm_perf_sink_config(handle);
0400     struct trbe_cpudata *cpudata = buf->cpudata;
0401     const u64 bufsize = buf->nr_pages * PAGE_SIZE;
0402     u64 limit = bufsize;
0403     u64 head, tail, wakeup;
0404 
0405     head = PERF_IDX2OFF(handle->head, buf);
0406 
0407     /*
0408      *      head
0409      *  ------->|
0410      *  |
0411      *  head    TRBE align  tail
0412      * +----|-------|---------------|-------+
0413      * |$$$$|=======|###############|$$$$$$$|
0414      * +----|-------|---------------|-------+
0415      * trbe_base                trbe_base + nr_pages
0416      *
0417      * Perf aux buffer output head position can be misaligned depending on
0418      * various factors including user space reads. In case misaligned, head
0419      * needs to be aligned before TRBE can be configured. Pad the alignment
0420      * gap with ETE_IGNORE_PACKET bytes that will be ignored by user tools
0421      * and skip this section thus advancing the head.
0422      */
0423     if (!IS_ALIGNED(head, cpudata->trbe_align)) {
0424         unsigned long delta = roundup(head, cpudata->trbe_align) - head;
0425 
0426         delta = min(delta, handle->size);
0427         trbe_pad_buf(handle, delta);
0428         head = PERF_IDX2OFF(handle->head, buf);
0429     }
0430 
0431     /*
0432      *  head = tail (size = 0)
0433      * +----|-------------------------------+
0434      * |$$$$|$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$ |
0435      * +----|-------------------------------+
0436      * trbe_base                trbe_base + nr_pages
0437      *
0438      * Perf aux buffer does not have any space for the driver to write into.
0439      */
0440     if (!handle->size)
0441         return 0;
0442 
0443     /* Compute the tail and wakeup indices now that we've aligned head */
0444     tail = PERF_IDX2OFF(handle->head + handle->size, buf);
0445     wakeup = PERF_IDX2OFF(handle->wakeup, buf);
0446 
0447     /*
0448      * Lets calculate the buffer area which TRBE could write into. There
0449      * are three possible scenarios here. Limit needs to be aligned with
0450      * PAGE_SIZE per the TRBE requirement. Always avoid clobbering the
0451      * unconsumed data.
0452      *
0453      * 1) head < tail
0454      *
0455      *  head            tail
0456      * +----|-----------------------|-------+
0457      * |$$$$|#######################|$$$$$$$|
0458      * +----|-----------------------|-------+
0459      * trbe_base            limit   trbe_base + nr_pages
0460      *
0461      * TRBE could write into [head..tail] area. Unless the tail is right at
0462      * the end of the buffer, neither an wrap around nor an IRQ is expected
0463      * while being enabled.
0464      *
0465      * 2) head == tail
0466      *
0467      *  head = tail (size > 0)
0468      * +----|-------------------------------+
0469      * |%%%%|###############################|
0470      * +----|-------------------------------+
0471      * trbe_base                limit = trbe_base + nr_pages
0472      *
0473      * TRBE should just write into [head..base + nr_pages] area even though
0474      * the entire buffer is empty. Reason being, when the trace reaches the
0475      * end of the buffer, it will just wrap around with an IRQ giving an
0476      * opportunity to reconfigure the buffer.
0477      *
0478      * 3) tail < head
0479      *
0480      *  tail            head
0481      * +----|-----------------------|-------+
0482      * |%%%%|$$$$$$$$$$$$$$$$$$$$$$$|#######|
0483      * +----|-----------------------|-------+
0484      * trbe_base                limit = trbe_base + nr_pages
0485      *
0486      * TRBE should just write into [head..base + nr_pages] area even though
0487      * the [trbe_base..tail] is also empty. Reason being, when the trace
0488      * reaches the end of the buffer, it will just wrap around with an IRQ
0489      * giving an opportunity to reconfigure the buffer.
0490      */
0491     if (head < tail)
0492         limit = round_down(tail, PAGE_SIZE);
0493 
0494     /*
0495      * Wakeup may be arbitrarily far into the future. If it's not in the
0496      * current generation, either we'll wrap before hitting it, or it's
0497      * in the past and has been handled already.
0498      *
0499      * If there's a wakeup before we wrap, arrange to be woken up by the
0500      * page boundary following it. Keep the tail boundary if that's lower.
0501      *
0502      *  head        wakeup  tail
0503      * +----|---------------|-------|-------+
0504      * |$$$$|###############|%%%%%%%|$$$$$$$|
0505      * +----|---------------|-------|-------+
0506      * trbe_base        limit       trbe_base + nr_pages
0507      */
0508     if (handle->wakeup < (handle->head + handle->size) && head <= wakeup)
0509         limit = min(limit, round_up(wakeup, PAGE_SIZE));
0510 
0511     /*
0512      * There are two situation when this can happen i.e limit is before
0513      * the head and hence TRBE cannot be configured.
0514      *
0515      * 1) head < tail (aligned down with PAGE_SIZE) and also they are both
0516      * within the same PAGE size range.
0517      *
0518      *          PAGE_SIZE
0519      *      |----------------------|
0520      *
0521      *      limit   head    tail
0522      * +------------|------|--------|-------+
0523      * |$$$$$$$$$$$$$$$$$$$|========|$$$$$$$|
0524      * +------------|------|--------|-------+
0525      * trbe_base                trbe_base + nr_pages
0526      *
0527      * 2) head < wakeup (aligned up with PAGE_SIZE) < tail and also both
0528      * head and wakeup are within same PAGE size range.
0529      *
0530      *      PAGE_SIZE
0531      *  |----------------------|
0532      *
0533      *  limit   head    wakeup  tail
0534      * +----|------|-------|--------|-------+
0535      * |$$$$$$$$$$$|=======|========|$$$$$$$|
0536      * +----|------|-------|--------|-------+
0537      * trbe_base                trbe_base + nr_pages
0538      */
0539     if (limit > head)
0540         return limit;
0541 
0542     trbe_pad_buf(handle, handle->size);
0543     return 0;
0544 }
0545 
0546 static unsigned long trbe_normal_offset(struct perf_output_handle *handle)
0547 {
0548     struct trbe_buf *buf = etm_perf_sink_config(handle);
0549     u64 limit = __trbe_normal_offset(handle);
0550     u64 head = PERF_IDX2OFF(handle->head, buf);
0551 
0552     /*
0553      * If the head is too close to the limit and we don't
0554      * have space for a meaningful run, we rather pad it
0555      * and start fresh.
0556      *
0557      * We might have to do this more than once to make sure
0558      * we have enough required space.
0559      */
0560     while (limit && ((limit - head) < trbe_min_trace_buf_size(handle))) {
0561         trbe_pad_buf(handle, limit - head);
0562         limit = __trbe_normal_offset(handle);
0563         head = PERF_IDX2OFF(handle->head, buf);
0564     }
0565     return limit;
0566 }
0567 
0568 static unsigned long compute_trbe_buffer_limit(struct perf_output_handle *handle)
0569 {
0570     struct trbe_buf *buf = etm_perf_sink_config(handle);
0571     unsigned long offset;
0572 
0573     if (buf->snapshot)
0574         offset = trbe_snapshot_offset(handle);
0575     else
0576         offset = trbe_normal_offset(handle);
0577     return buf->trbe_base + offset;
0578 }
0579 
0580 static void clr_trbe_status(void)
0581 {
0582     u64 trbsr = read_sysreg_s(SYS_TRBSR_EL1);
0583 
0584     WARN_ON(is_trbe_enabled());
0585     trbsr &= ~TRBSR_IRQ;
0586     trbsr &= ~TRBSR_TRG;
0587     trbsr &= ~TRBSR_WRAP;
0588     trbsr &= ~(TRBSR_EC_MASK << TRBSR_EC_SHIFT);
0589     trbsr &= ~(TRBSR_BSC_MASK << TRBSR_BSC_SHIFT);
0590     trbsr &= ~TRBSR_STOP;
0591     write_sysreg_s(trbsr, SYS_TRBSR_EL1);
0592 }
0593 
0594 static void set_trbe_limit_pointer_enabled(struct trbe_buf *buf)
0595 {
0596     u64 trblimitr = read_sysreg_s(SYS_TRBLIMITR_EL1);
0597     unsigned long addr = buf->trbe_limit;
0598 
0599     WARN_ON(!IS_ALIGNED(addr, (1UL << TRBLIMITR_LIMIT_SHIFT)));
0600     WARN_ON(!IS_ALIGNED(addr, PAGE_SIZE));
0601 
0602     trblimitr &= ~TRBLIMITR_NVM;
0603     trblimitr &= ~(TRBLIMITR_FILL_MODE_MASK << TRBLIMITR_FILL_MODE_SHIFT);
0604     trblimitr &= ~(TRBLIMITR_TRIG_MODE_MASK << TRBLIMITR_TRIG_MODE_SHIFT);
0605     trblimitr &= ~(TRBLIMITR_LIMIT_MASK << TRBLIMITR_LIMIT_SHIFT);
0606 
0607     /*
0608      * Fill trace buffer mode is used here while configuring the
0609      * TRBE for trace capture. In this particular mode, the trace
0610      * collection is stopped and a maintenance interrupt is raised
0611      * when the current write pointer wraps. This pause in trace
0612      * collection gives the software an opportunity to capture the
0613      * trace data in the interrupt handler, before reconfiguring
0614      * the TRBE.
0615      */
0616     trblimitr |= (TRBE_FILL_MODE_FILL & TRBLIMITR_FILL_MODE_MASK) << TRBLIMITR_FILL_MODE_SHIFT;
0617 
0618     /*
0619      * Trigger mode is not used here while configuring the TRBE for
0620      * the trace capture. Hence just keep this in the ignore mode.
0621      */
0622     trblimitr |= (TRBE_TRIG_MODE_IGNORE & TRBLIMITR_TRIG_MODE_MASK) <<
0623               TRBLIMITR_TRIG_MODE_SHIFT;
0624     trblimitr |= (addr & PAGE_MASK);
0625     set_trbe_enabled(buf->cpudata, trblimitr);
0626 }
0627 
0628 static void trbe_enable_hw(struct trbe_buf *buf)
0629 {
0630     WARN_ON(buf->trbe_hw_base < buf->trbe_base);
0631     WARN_ON(buf->trbe_write < buf->trbe_hw_base);
0632     WARN_ON(buf->trbe_write >= buf->trbe_limit);
0633     set_trbe_disabled(buf->cpudata);
0634     clr_trbe_status();
0635     set_trbe_base_pointer(buf->trbe_hw_base);
0636     set_trbe_write_pointer(buf->trbe_write);
0637 
0638     /*
0639      * Synchronize all the register updates
0640      * till now before enabling the TRBE.
0641      */
0642     isb();
0643     set_trbe_limit_pointer_enabled(buf);
0644 }
0645 
0646 static enum trbe_fault_action trbe_get_fault_act(struct perf_output_handle *handle,
0647                          u64 trbsr)
0648 {
0649     int ec = get_trbe_ec(trbsr);
0650     int bsc = get_trbe_bsc(trbsr);
0651     struct trbe_buf *buf = etm_perf_sink_config(handle);
0652     struct trbe_cpudata *cpudata = buf->cpudata;
0653 
0654     WARN_ON(is_trbe_running(trbsr));
0655     if (is_trbe_trg(trbsr) || is_trbe_abort(trbsr))
0656         return TRBE_FAULT_ACT_FATAL;
0657 
0658     if ((ec == TRBE_EC_STAGE1_ABORT) || (ec == TRBE_EC_STAGE2_ABORT))
0659         return TRBE_FAULT_ACT_FATAL;
0660 
0661     /*
0662      * If the trbe is affected by TRBE_WORKAROUND_OVERWRITE_FILL_MODE,
0663      * it might write data after a WRAP event in the fill mode.
0664      * Thus the check TRBPTR == TRBBASER will not be honored.
0665      */
0666     if ((is_trbe_wrap(trbsr) && (ec == TRBE_EC_OTHERS) && (bsc == TRBE_BSC_FILLED)) &&
0667         (trbe_may_overwrite_in_fill_mode(cpudata) ||
0668          get_trbe_write_pointer() == get_trbe_base_pointer()))
0669         return TRBE_FAULT_ACT_WRAP;
0670 
0671     return TRBE_FAULT_ACT_SPURIOUS;
0672 }
0673 
0674 static unsigned long trbe_get_trace_size(struct perf_output_handle *handle,
0675                      struct trbe_buf *buf, bool wrap)
0676 {
0677     u64 write;
0678     u64 start_off, end_off;
0679     u64 size;
0680     u64 overwrite_skip = TRBE_WORKAROUND_OVERWRITE_FILL_MODE_SKIP_BYTES;
0681 
0682     /*
0683      * If the TRBE has wrapped around the write pointer has
0684      * wrapped and should be treated as limit.
0685      *
0686      * When the TRBE is affected by TRBE_WORKAROUND_WRITE_OUT_OF_RANGE,
0687      * it may write upto 64bytes beyond the "LIMIT". The driver already
0688      * keeps a valid page next to the LIMIT and we could potentially
0689      * consume the trace data that may have been collected there. But we
0690      * cannot be really sure it is available, and the TRBPTR may not
0691      * indicate the same. Also, affected cores are also affected by another
0692      * erratum which forces the PAGE_SIZE alignment on the TRBPTR, and thus
0693      * could potentially pad an entire PAGE_SIZE - 64bytes, to get those
0694      * 64bytes. Thus we ignore the potential triggering of the erratum
0695      * on WRAP and limit the data to LIMIT.
0696      */
0697     if (wrap)
0698         write = get_trbe_limit_pointer();
0699     else
0700         write = get_trbe_write_pointer();
0701 
0702     /*
0703      * TRBE may use a different base address than the base
0704      * of the ring buffer. Thus use the beginning of the ring
0705      * buffer to compute the offsets.
0706      */
0707     end_off = write - buf->trbe_base;
0708     start_off = PERF_IDX2OFF(handle->head, buf);
0709 
0710     if (WARN_ON_ONCE(end_off < start_off))
0711         return 0;
0712 
0713     size = end_off - start_off;
0714     /*
0715      * If the TRBE is affected by the following erratum, we must fill
0716      * the space we skipped with IGNORE packets. And we are always
0717      * guaranteed to have at least a PAGE_SIZE space in the buffer.
0718      */
0719     if (trbe_has_erratum(buf->cpudata, TRBE_WORKAROUND_OVERWRITE_FILL_MODE) &&
0720         !WARN_ON(size < overwrite_skip))
0721         __trbe_pad_buf(buf, start_off, overwrite_skip);
0722 
0723     return size;
0724 }
0725 
0726 static void *arm_trbe_alloc_buffer(struct coresight_device *csdev,
0727                    struct perf_event *event, void **pages,
0728                    int nr_pages, bool snapshot)
0729 {
0730     struct trbe_buf *buf;
0731     struct page **pglist;
0732     int i;
0733 
0734     /*
0735      * TRBE LIMIT and TRBE WRITE pointers must be page aligned. But with
0736      * just a single page, there would not be any room left while writing
0737      * into a partially filled TRBE buffer after the page size alignment.
0738      * Hence restrict the minimum buffer size as two pages.
0739      */
0740     if (nr_pages < 2)
0741         return NULL;
0742 
0743     buf = kzalloc_node(sizeof(*buf), GFP_KERNEL, trbe_alloc_node(event));
0744     if (!buf)
0745         return ERR_PTR(-ENOMEM);
0746 
0747     pglist = kcalloc(nr_pages, sizeof(*pglist), GFP_KERNEL);
0748     if (!pglist) {
0749         kfree(buf);
0750         return ERR_PTR(-ENOMEM);
0751     }
0752 
0753     for (i = 0; i < nr_pages; i++)
0754         pglist[i] = virt_to_page(pages[i]);
0755 
0756     buf->trbe_base = (unsigned long)vmap(pglist, nr_pages, VM_MAP, PAGE_KERNEL);
0757     if (!buf->trbe_base) {
0758         kfree(pglist);
0759         kfree(buf);
0760         return ERR_PTR(-ENOMEM);
0761     }
0762     buf->trbe_limit = buf->trbe_base + nr_pages * PAGE_SIZE;
0763     buf->trbe_write = buf->trbe_base;
0764     buf->snapshot = snapshot;
0765     buf->nr_pages = nr_pages;
0766     buf->pages = pages;
0767     kfree(pglist);
0768     return buf;
0769 }
0770 
0771 static void arm_trbe_free_buffer(void *config)
0772 {
0773     struct trbe_buf *buf = config;
0774 
0775     vunmap((void *)buf->trbe_base);
0776     kfree(buf);
0777 }
0778 
0779 static unsigned long arm_trbe_update_buffer(struct coresight_device *csdev,
0780                         struct perf_output_handle *handle,
0781                         void *config)
0782 {
0783     struct trbe_drvdata *drvdata = dev_get_drvdata(csdev->dev.parent);
0784     struct trbe_cpudata *cpudata = dev_get_drvdata(&csdev->dev);
0785     struct trbe_buf *buf = config;
0786     enum trbe_fault_action act;
0787     unsigned long size, status;
0788     unsigned long flags;
0789     bool wrap = false;
0790 
0791     WARN_ON(buf->cpudata != cpudata);
0792     WARN_ON(cpudata->cpu != smp_processor_id());
0793     WARN_ON(cpudata->drvdata != drvdata);
0794     if (cpudata->mode != CS_MODE_PERF)
0795         return 0;
0796 
0797     /*
0798      * We are about to disable the TRBE. And this could in turn
0799      * fill up the buffer triggering, an IRQ. This could be consumed
0800      * by the PE asynchronously, causing a race here against
0801      * the IRQ handler in closing out the handle. So, let us
0802      * make sure the IRQ can't trigger while we are collecting
0803      * the buffer. We also make sure that a WRAP event is handled
0804      * accordingly.
0805      */
0806     local_irq_save(flags);
0807 
0808     /*
0809      * If the TRBE was disabled due to lack of space in the AUX buffer or a
0810      * spurious fault, the driver leaves it disabled, truncating the buffer.
0811      * Since the etm_perf driver expects to close out the AUX buffer, the
0812      * driver skips it. Thus, just pass in 0 size here to indicate that the
0813      * buffer was truncated.
0814      */
0815     if (!is_trbe_enabled()) {
0816         size = 0;
0817         goto done;
0818     }
0819     /*
0820      * perf handle structure needs to be shared with the TRBE IRQ handler for
0821      * capturing trace data and restarting the handle. There is a probability
0822      * of an undefined reference based crash when etm event is being stopped
0823      * while a TRBE IRQ also getting processed. This happens due the release
0824      * of perf handle via perf_aux_output_end() in etm_event_stop(). Stopping
0825      * the TRBE here will ensure that no IRQ could be generated when the perf
0826      * handle gets freed in etm_event_stop().
0827      */
0828     trbe_drain_and_disable_local(cpudata);
0829 
0830     /* Check if there is a pending interrupt and handle it here */
0831     status = read_sysreg_s(SYS_TRBSR_EL1);
0832     if (is_trbe_irq(status)) {
0833 
0834         /*
0835          * Now that we are handling the IRQ here, clear the IRQ
0836          * from the status, to let the irq handler know that it
0837          * is taken care of.
0838          */
0839         clr_trbe_irq();
0840         isb();
0841 
0842         act = trbe_get_fault_act(handle, status);
0843         /*
0844          * If this was not due to a WRAP event, we have some
0845          * errors and as such buffer is empty.
0846          */
0847         if (act != TRBE_FAULT_ACT_WRAP) {
0848             size = 0;
0849             goto done;
0850         }
0851 
0852         trbe_report_wrap_event(handle);
0853         wrap = true;
0854     }
0855 
0856     size = trbe_get_trace_size(handle, buf, wrap);
0857 
0858 done:
0859     local_irq_restore(flags);
0860 
0861     if (buf->snapshot)
0862         handle->head += size;
0863     return size;
0864 }
0865 
0866 
0867 static int trbe_apply_work_around_before_enable(struct trbe_buf *buf)
0868 {
0869     /*
0870      * TRBE_WORKAROUND_OVERWRITE_FILL_MODE causes the TRBE to overwrite a few cache
0871      * line size from the "TRBBASER_EL1" in the event of a "FILL".
0872      * Thus, we could loose some amount of the trace at the base.
0873      *
0874      * Before Fix:
0875      *
0876      *  normal-BASE     head (normal-TRBPTR)         tail (normal-LIMIT)
0877      *  |                   \/                       /
0878      *   -------------------------------------------------------------
0879      *  |   Pg0      |   Pg1       |           |          |  PgN     |
0880      *   -------------------------------------------------------------
0881      *
0882      * In the normal course of action, we would set the TRBBASER to the
0883      * beginning of the ring-buffer (normal-BASE). But with the erratum,
0884      * the TRBE could overwrite the contents at the "normal-BASE", after
0885      * hitting the "normal-LIMIT", since it doesn't stop as expected. And
0886      * this is wrong. This could result in overwriting trace collected in
0887      * one of the previous runs, being consumed by the user. So we must
0888      * always make sure that the TRBBASER is within the region
0889      * [head, head+size]. Note that TRBBASER must be PAGE aligned,
0890      *
0891      *  After moving the BASE:
0892      *
0893      *  normal-BASE     head (normal-TRBPTR)         tail (normal-LIMIT)
0894      *  |                   \/                       /
0895      *   -------------------------------------------------------------
0896      *  |         |          |xyzdef.     |..   tuvw|                |
0897      *   -------------------------------------------------------------
0898      *                      /
0899      *              New-BASER
0900      *
0901      * Also, we would set the TRBPTR to head (after adjusting for
0902      * alignment) at normal-PTR. This would mean that the last few bytes
0903      * of the trace (say, "xyz") might overwrite the first few bytes of
0904      * trace written ("abc"). More importantly they will appear in what
0905      * userspace sees as the beginning of the trace, which is wrong. We may
0906      * not always have space to move the latest trace "xyz" to the correct
0907      * order as it must appear beyond the LIMIT. (i.e, [head..head+size]).
0908      * Thus it is easier to ignore those bytes than to complicate the
0909      * driver to move it, assuming that the erratum was triggered and
0910      * doing additional checks to see if there is indeed allowed space at
0911      * TRBLIMITR.LIMIT.
0912      *
0913      *  Thus the full workaround will move the BASE and the PTR and would
0914      *  look like (after padding at the skipped bytes at the end of
0915      *  session) :
0916      *
0917      *  normal-BASE     head (normal-TRBPTR)         tail (normal-LIMIT)
0918      *  |                   \/                       /
0919      *   -------------------------------------------------------------
0920      *  |         |          |///abc..     |..  rst|                |
0921      *   -------------------------------------------------------------
0922      *                      /    |
0923      *              New-BASER    New-TRBPTR
0924      *
0925      * To summarize, with the work around:
0926      *
0927      *  - We always align the offset for the next session to PAGE_SIZE
0928      *    (This is to ensure we can program the TRBBASER to this offset
0929      *    within the region [head...head+size]).
0930      *
0931      *  - At TRBE enable:
0932      *     - Set the TRBBASER to the page aligned offset of the current
0933      *       proposed write offset. (which is guaranteed to be aligned
0934      *       as above)
0935      *     - Move the TRBPTR to skip first 256bytes (that might be
0936      *       overwritten with the erratum). This ensures that the trace
0937      *       generated in the session is not re-written.
0938      *
0939      *  - At trace collection:
0940      *     - Pad the 256bytes skipped above again with IGNORE packets.
0941      */
0942     if (trbe_has_erratum(buf->cpudata, TRBE_WORKAROUND_OVERWRITE_FILL_MODE)) {
0943         if (WARN_ON(!IS_ALIGNED(buf->trbe_write, PAGE_SIZE)))
0944             return -EINVAL;
0945         buf->trbe_hw_base = buf->trbe_write;
0946         buf->trbe_write += TRBE_WORKAROUND_OVERWRITE_FILL_MODE_SKIP_BYTES;
0947     }
0948 
0949     /*
0950      * TRBE_WORKAROUND_WRITE_OUT_OF_RANGE could cause the TRBE to write to
0951      * the next page after the TRBLIMITR.LIMIT. For perf, the "next page"
0952      * may be:
0953      *     - The page beyond the ring buffer. This could mean, TRBE could
0954      *       corrupt another entity (kernel / user)
0955      *     - A portion of the "ring buffer" consumed by the userspace.
0956      *       i.e, a page outisde [head, head + size].
0957      *
0958      * We work around this by:
0959      *     - Making sure that we have at least an extra space of PAGE left
0960      *       in the ring buffer [head, head + size], than we normally do
0961      *       without the erratum. See trbe_min_trace_buf_size().
0962      *
0963      *     - Adjust the TRBLIMITR.LIMIT to leave the extra PAGE outside
0964      *       the TRBE's range (i.e [TRBBASER, TRBLIMITR.LIMI] ).
0965      */
0966     if (trbe_has_erratum(buf->cpudata, TRBE_WORKAROUND_WRITE_OUT_OF_RANGE)) {
0967         s64 space = buf->trbe_limit - buf->trbe_write;
0968         /*
0969          * We must have more than a PAGE_SIZE worth space in the proposed
0970          * range for the TRBE.
0971          */
0972         if (WARN_ON(space <= PAGE_SIZE ||
0973                 !IS_ALIGNED(buf->trbe_limit, PAGE_SIZE)))
0974             return -EINVAL;
0975         buf->trbe_limit -= PAGE_SIZE;
0976     }
0977 
0978     return 0;
0979 }
0980 
0981 static int __arm_trbe_enable(struct trbe_buf *buf,
0982                  struct perf_output_handle *handle)
0983 {
0984     int ret = 0;
0985 
0986     perf_aux_output_flag(handle, PERF_AUX_FLAG_CORESIGHT_FORMAT_RAW);
0987     buf->trbe_limit = compute_trbe_buffer_limit(handle);
0988     buf->trbe_write = buf->trbe_base + PERF_IDX2OFF(handle->head, buf);
0989     if (buf->trbe_limit == buf->trbe_base) {
0990         ret = -ENOSPC;
0991         goto err;
0992     }
0993     /* Set the base of the TRBE to the buffer base */
0994     buf->trbe_hw_base = buf->trbe_base;
0995 
0996     ret = trbe_apply_work_around_before_enable(buf);
0997     if (ret)
0998         goto err;
0999 
1000     *this_cpu_ptr(buf->cpudata->drvdata->handle) = handle;
1001     trbe_enable_hw(buf);
1002     return 0;
1003 err:
1004     trbe_stop_and_truncate_event(handle);
1005     return ret;
1006 }
1007 
1008 static int arm_trbe_enable(struct coresight_device *csdev, u32 mode, void *data)
1009 {
1010     struct trbe_drvdata *drvdata = dev_get_drvdata(csdev->dev.parent);
1011     struct trbe_cpudata *cpudata = dev_get_drvdata(&csdev->dev);
1012     struct perf_output_handle *handle = data;
1013     struct trbe_buf *buf = etm_perf_sink_config(handle);
1014 
1015     WARN_ON(cpudata->cpu != smp_processor_id());
1016     WARN_ON(cpudata->drvdata != drvdata);
1017     if (mode != CS_MODE_PERF)
1018         return -EINVAL;
1019 
1020     cpudata->buf = buf;
1021     cpudata->mode = mode;
1022     buf->cpudata = cpudata;
1023 
1024     return __arm_trbe_enable(buf, handle);
1025 }
1026 
1027 static int arm_trbe_disable(struct coresight_device *csdev)
1028 {
1029     struct trbe_drvdata *drvdata = dev_get_drvdata(csdev->dev.parent);
1030     struct trbe_cpudata *cpudata = dev_get_drvdata(&csdev->dev);
1031     struct trbe_buf *buf = cpudata->buf;
1032 
1033     WARN_ON(buf->cpudata != cpudata);
1034     WARN_ON(cpudata->cpu != smp_processor_id());
1035     WARN_ON(cpudata->drvdata != drvdata);
1036     if (cpudata->mode != CS_MODE_PERF)
1037         return -EINVAL;
1038 
1039     trbe_drain_and_disable_local(cpudata);
1040     buf->cpudata = NULL;
1041     cpudata->buf = NULL;
1042     cpudata->mode = CS_MODE_DISABLED;
1043     return 0;
1044 }
1045 
1046 static void trbe_handle_spurious(struct perf_output_handle *handle)
1047 {
1048     struct trbe_buf *buf = etm_perf_sink_config(handle);
1049     u64 trblimitr = read_sysreg_s(SYS_TRBLIMITR_EL1);
1050 
1051     /*
1052      * If the IRQ was spurious, simply re-enable the TRBE
1053      * back without modifying the buffer parameters to
1054      * retain the trace collected so far.
1055      */
1056     set_trbe_enabled(buf->cpudata, trblimitr);
1057 }
1058 
1059 static int trbe_handle_overflow(struct perf_output_handle *handle)
1060 {
1061     struct perf_event *event = handle->event;
1062     struct trbe_buf *buf = etm_perf_sink_config(handle);
1063     unsigned long size;
1064     struct etm_event_data *event_data;
1065 
1066     size = trbe_get_trace_size(handle, buf, true);
1067     if (buf->snapshot)
1068         handle->head += size;
1069 
1070     trbe_report_wrap_event(handle);
1071     perf_aux_output_end(handle, size);
1072     event_data = perf_aux_output_begin(handle, event);
1073     if (!event_data) {
1074         /*
1075          * We are unable to restart the trace collection,
1076          * thus leave the TRBE disabled. The etm-perf driver
1077          * is able to detect this with a disconnected handle
1078          * (handle->event = NULL).
1079          */
1080         trbe_drain_and_disable_local(buf->cpudata);
1081         *this_cpu_ptr(buf->cpudata->drvdata->handle) = NULL;
1082         return -EINVAL;
1083     }
1084 
1085     return __arm_trbe_enable(buf, handle);
1086 }
1087 
1088 static bool is_perf_trbe(struct perf_output_handle *handle)
1089 {
1090     struct trbe_buf *buf = etm_perf_sink_config(handle);
1091     struct trbe_cpudata *cpudata = buf->cpudata;
1092     struct trbe_drvdata *drvdata = cpudata->drvdata;
1093     int cpu = smp_processor_id();
1094 
1095     WARN_ON(buf->trbe_hw_base != get_trbe_base_pointer());
1096     WARN_ON(buf->trbe_limit != get_trbe_limit_pointer());
1097 
1098     if (cpudata->mode != CS_MODE_PERF)
1099         return false;
1100 
1101     if (cpudata->cpu != cpu)
1102         return false;
1103 
1104     if (!cpumask_test_cpu(cpu, &drvdata->supported_cpus))
1105         return false;
1106 
1107     return true;
1108 }
1109 
1110 static irqreturn_t arm_trbe_irq_handler(int irq, void *dev)
1111 {
1112     struct perf_output_handle **handle_ptr = dev;
1113     struct perf_output_handle *handle = *handle_ptr;
1114     struct trbe_buf *buf = etm_perf_sink_config(handle);
1115     enum trbe_fault_action act;
1116     u64 status;
1117     bool truncated = false;
1118     u64 trfcr;
1119 
1120     /* Reads to TRBSR_EL1 is fine when TRBE is active */
1121     status = read_sysreg_s(SYS_TRBSR_EL1);
1122     /*
1123      * If the pending IRQ was handled by update_buffer callback
1124      * we have nothing to do here.
1125      */
1126     if (!is_trbe_irq(status))
1127         return IRQ_NONE;
1128 
1129     /* Prohibit the CPU from tracing before we disable the TRBE */
1130     trfcr = cpu_prohibit_trace();
1131     /*
1132      * Ensure the trace is visible to the CPUs and
1133      * any external aborts have been resolved.
1134      */
1135     trbe_drain_and_disable_local(buf->cpudata);
1136     clr_trbe_irq();
1137     isb();
1138 
1139     if (WARN_ON_ONCE(!handle) || !perf_get_aux(handle))
1140         return IRQ_NONE;
1141 
1142     if (!is_perf_trbe(handle))
1143         return IRQ_NONE;
1144 
1145     act = trbe_get_fault_act(handle, status);
1146     switch (act) {
1147     case TRBE_FAULT_ACT_WRAP:
1148         truncated = !!trbe_handle_overflow(handle);
1149         break;
1150     case TRBE_FAULT_ACT_SPURIOUS:
1151         trbe_handle_spurious(handle);
1152         break;
1153     case TRBE_FAULT_ACT_FATAL:
1154         trbe_stop_and_truncate_event(handle);
1155         truncated = true;
1156         break;
1157     }
1158 
1159     /*
1160      * If the buffer was truncated, ensure perf callbacks
1161      * have completed, which will disable the event.
1162      *
1163      * Otherwise, restore the trace filter controls to
1164      * allow the tracing.
1165      */
1166     if (truncated)
1167         irq_work_run();
1168     else
1169         write_trfcr(trfcr);
1170 
1171     return IRQ_HANDLED;
1172 }
1173 
1174 static const struct coresight_ops_sink arm_trbe_sink_ops = {
1175     .enable     = arm_trbe_enable,
1176     .disable    = arm_trbe_disable,
1177     .alloc_buffer   = arm_trbe_alloc_buffer,
1178     .free_buffer    = arm_trbe_free_buffer,
1179     .update_buffer  = arm_trbe_update_buffer,
1180 };
1181 
1182 static const struct coresight_ops arm_trbe_cs_ops = {
1183     .sink_ops   = &arm_trbe_sink_ops,
1184 };
1185 
1186 static ssize_t align_show(struct device *dev, struct device_attribute *attr, char *buf)
1187 {
1188     struct trbe_cpudata *cpudata = dev_get_drvdata(dev);
1189 
1190     return sprintf(buf, "%llx\n", cpudata->trbe_hw_align);
1191 }
1192 static DEVICE_ATTR_RO(align);
1193 
1194 static ssize_t flag_show(struct device *dev, struct device_attribute *attr, char *buf)
1195 {
1196     struct trbe_cpudata *cpudata = dev_get_drvdata(dev);
1197 
1198     return sprintf(buf, "%d\n", cpudata->trbe_flag);
1199 }
1200 static DEVICE_ATTR_RO(flag);
1201 
1202 static struct attribute *arm_trbe_attrs[] = {
1203     &dev_attr_align.attr,
1204     &dev_attr_flag.attr,
1205     NULL,
1206 };
1207 
1208 static const struct attribute_group arm_trbe_group = {
1209     .attrs = arm_trbe_attrs,
1210 };
1211 
1212 static const struct attribute_group *arm_trbe_groups[] = {
1213     &arm_trbe_group,
1214     NULL,
1215 };
1216 
1217 static void arm_trbe_enable_cpu(void *info)
1218 {
1219     struct trbe_drvdata *drvdata = info;
1220     struct trbe_cpudata *cpudata = this_cpu_ptr(drvdata->cpudata);
1221 
1222     trbe_reset_local(cpudata);
1223     enable_percpu_irq(drvdata->irq, IRQ_TYPE_NONE);
1224 }
1225 
1226 static void arm_trbe_register_coresight_cpu(struct trbe_drvdata *drvdata, int cpu)
1227 {
1228     struct trbe_cpudata *cpudata = per_cpu_ptr(drvdata->cpudata, cpu);
1229     struct coresight_device *trbe_csdev = coresight_get_percpu_sink(cpu);
1230     struct coresight_desc desc = { 0 };
1231     struct device *dev;
1232 
1233     if (WARN_ON(trbe_csdev))
1234         return;
1235 
1236     /* If the TRBE was not probed on the CPU, we shouldn't be here */
1237     if (WARN_ON(!cpudata->drvdata))
1238         return;
1239 
1240     dev = &cpudata->drvdata->pdev->dev;
1241     desc.name = devm_kasprintf(dev, GFP_KERNEL, "trbe%d", cpu);
1242     if (!desc.name)
1243         goto cpu_clear;
1244 
1245     desc.type = CORESIGHT_DEV_TYPE_SINK;
1246     desc.subtype.sink_subtype = CORESIGHT_DEV_SUBTYPE_SINK_PERCPU_SYSMEM;
1247     desc.ops = &arm_trbe_cs_ops;
1248     desc.pdata = dev_get_platdata(dev);
1249     desc.groups = arm_trbe_groups;
1250     desc.dev = dev;
1251     trbe_csdev = coresight_register(&desc);
1252     if (IS_ERR(trbe_csdev))
1253         goto cpu_clear;
1254 
1255     dev_set_drvdata(&trbe_csdev->dev, cpudata);
1256     coresight_set_percpu_sink(cpu, trbe_csdev);
1257     return;
1258 cpu_clear:
1259     cpumask_clear_cpu(cpu, &drvdata->supported_cpus);
1260 }
1261 
1262 /*
1263  * Must be called with preemption disabled, for trbe_check_errata().
1264  */
1265 static void arm_trbe_probe_cpu(void *info)
1266 {
1267     struct trbe_drvdata *drvdata = info;
1268     int cpu = smp_processor_id();
1269     struct trbe_cpudata *cpudata = per_cpu_ptr(drvdata->cpudata, cpu);
1270     u64 trbidr;
1271 
1272     if (WARN_ON(!cpudata))
1273         goto cpu_clear;
1274 
1275     if (!is_trbe_available()) {
1276         pr_err("TRBE is not implemented on cpu %d\n", cpu);
1277         goto cpu_clear;
1278     }
1279 
1280     trbidr = read_sysreg_s(SYS_TRBIDR_EL1);
1281     if (!is_trbe_programmable(trbidr)) {
1282         pr_err("TRBE is owned in higher exception level on cpu %d\n", cpu);
1283         goto cpu_clear;
1284     }
1285 
1286     cpudata->trbe_hw_align = 1ULL << get_trbe_address_align(trbidr);
1287     if (cpudata->trbe_hw_align > SZ_2K) {
1288         pr_err("Unsupported alignment on cpu %d\n", cpu);
1289         goto cpu_clear;
1290     }
1291 
1292     /*
1293      * Run the TRBE erratum checks, now that we know
1294      * this instance is about to be registered.
1295      */
1296     trbe_check_errata(cpudata);
1297 
1298     if (trbe_is_broken(cpudata)) {
1299         pr_err("Disabling TRBE on cpu%d due to erratum\n", cpu);
1300         goto cpu_clear;
1301     }
1302 
1303     /*
1304      * If the TRBE is affected by erratum TRBE_WORKAROUND_OVERWRITE_FILL_MODE,
1305      * we must always program the TBRPTR_EL1, 256bytes from a page
1306      * boundary, with TRBBASER_EL1 set to the page, to prevent
1307      * TRBE over-writing 256bytes at TRBBASER_EL1 on FILL event.
1308      *
1309      * Thus make sure we always align our write pointer to a PAGE_SIZE,
1310      * which also guarantees that we have at least a PAGE_SIZE space in
1311      * the buffer (TRBLIMITR is PAGE aligned) and thus we can skip
1312      * the required bytes at the base.
1313      */
1314     if (trbe_may_overwrite_in_fill_mode(cpudata))
1315         cpudata->trbe_align = PAGE_SIZE;
1316     else
1317         cpudata->trbe_align = cpudata->trbe_hw_align;
1318 
1319     cpudata->trbe_flag = get_trbe_flag_update(trbidr);
1320     cpudata->cpu = cpu;
1321     cpudata->drvdata = drvdata;
1322     return;
1323 cpu_clear:
1324     cpumask_clear_cpu(cpu, &drvdata->supported_cpus);
1325 }
1326 
1327 static void arm_trbe_remove_coresight_cpu(void *info)
1328 {
1329     int cpu = smp_processor_id();
1330     struct trbe_drvdata *drvdata = info;
1331     struct trbe_cpudata *cpudata = per_cpu_ptr(drvdata->cpudata, cpu);
1332     struct coresight_device *trbe_csdev = coresight_get_percpu_sink(cpu);
1333 
1334     disable_percpu_irq(drvdata->irq);
1335     trbe_reset_local(cpudata);
1336     if (trbe_csdev) {
1337         coresight_unregister(trbe_csdev);
1338         cpudata->drvdata = NULL;
1339         coresight_set_percpu_sink(cpu, NULL);
1340     }
1341 }
1342 
1343 static int arm_trbe_probe_coresight(struct trbe_drvdata *drvdata)
1344 {
1345     int cpu;
1346 
1347     drvdata->cpudata = alloc_percpu(typeof(*drvdata->cpudata));
1348     if (!drvdata->cpudata)
1349         return -ENOMEM;
1350 
1351     for_each_cpu(cpu, &drvdata->supported_cpus) {
1352         /* If we fail to probe the CPU, let us defer it to hotplug callbacks */
1353         if (smp_call_function_single(cpu, arm_trbe_probe_cpu, drvdata, 1))
1354             continue;
1355         if (cpumask_test_cpu(cpu, &drvdata->supported_cpus))
1356             arm_trbe_register_coresight_cpu(drvdata, cpu);
1357         if (cpumask_test_cpu(cpu, &drvdata->supported_cpus))
1358             smp_call_function_single(cpu, arm_trbe_enable_cpu, drvdata, 1);
1359     }
1360     return 0;
1361 }
1362 
1363 static int arm_trbe_remove_coresight(struct trbe_drvdata *drvdata)
1364 {
1365     int cpu;
1366 
1367     for_each_cpu(cpu, &drvdata->supported_cpus)
1368         smp_call_function_single(cpu, arm_trbe_remove_coresight_cpu, drvdata, 1);
1369     free_percpu(drvdata->cpudata);
1370     return 0;
1371 }
1372 
1373 static void arm_trbe_probe_hotplugged_cpu(struct trbe_drvdata *drvdata)
1374 {
1375     preempt_disable();
1376     arm_trbe_probe_cpu(drvdata);
1377     preempt_enable();
1378 }
1379 
1380 static int arm_trbe_cpu_startup(unsigned int cpu, struct hlist_node *node)
1381 {
1382     struct trbe_drvdata *drvdata = hlist_entry_safe(node, struct trbe_drvdata, hotplug_node);
1383 
1384     if (cpumask_test_cpu(cpu, &drvdata->supported_cpus)) {
1385 
1386         /*
1387          * If this CPU was not probed for TRBE,
1388          * initialize it now.
1389          */
1390         if (!coresight_get_percpu_sink(cpu)) {
1391             arm_trbe_probe_hotplugged_cpu(drvdata);
1392             if (cpumask_test_cpu(cpu, &drvdata->supported_cpus))
1393                 arm_trbe_register_coresight_cpu(drvdata, cpu);
1394             if (cpumask_test_cpu(cpu, &drvdata->supported_cpus))
1395                 arm_trbe_enable_cpu(drvdata);
1396         } else {
1397             arm_trbe_enable_cpu(drvdata);
1398         }
1399     }
1400     return 0;
1401 }
1402 
1403 static int arm_trbe_cpu_teardown(unsigned int cpu, struct hlist_node *node)
1404 {
1405     struct trbe_drvdata *drvdata = hlist_entry_safe(node, struct trbe_drvdata, hotplug_node);
1406 
1407     if (cpumask_test_cpu(cpu, &drvdata->supported_cpus)) {
1408         struct trbe_cpudata *cpudata = per_cpu_ptr(drvdata->cpudata, cpu);
1409 
1410         disable_percpu_irq(drvdata->irq);
1411         trbe_reset_local(cpudata);
1412     }
1413     return 0;
1414 }
1415 
1416 static int arm_trbe_probe_cpuhp(struct trbe_drvdata *drvdata)
1417 {
1418     enum cpuhp_state trbe_online;
1419     int ret;
1420 
1421     trbe_online = cpuhp_setup_state_multi(CPUHP_AP_ONLINE_DYN, DRVNAME,
1422                           arm_trbe_cpu_startup, arm_trbe_cpu_teardown);
1423     if (trbe_online < 0)
1424         return trbe_online;
1425 
1426     ret = cpuhp_state_add_instance(trbe_online, &drvdata->hotplug_node);
1427     if (ret) {
1428         cpuhp_remove_multi_state(trbe_online);
1429         return ret;
1430     }
1431     drvdata->trbe_online = trbe_online;
1432     return 0;
1433 }
1434 
1435 static void arm_trbe_remove_cpuhp(struct trbe_drvdata *drvdata)
1436 {
1437     cpuhp_remove_multi_state(drvdata->trbe_online);
1438 }
1439 
1440 static int arm_trbe_probe_irq(struct platform_device *pdev,
1441                   struct trbe_drvdata *drvdata)
1442 {
1443     int ret;
1444 
1445     drvdata->irq = platform_get_irq(pdev, 0);
1446     if (drvdata->irq < 0) {
1447         pr_err("IRQ not found for the platform device\n");
1448         return drvdata->irq;
1449     }
1450 
1451     if (!irq_is_percpu(drvdata->irq)) {
1452         pr_err("IRQ is not a PPI\n");
1453         return -EINVAL;
1454     }
1455 
1456     if (irq_get_percpu_devid_partition(drvdata->irq, &drvdata->supported_cpus))
1457         return -EINVAL;
1458 
1459     drvdata->handle = alloc_percpu(struct perf_output_handle *);
1460     if (!drvdata->handle)
1461         return -ENOMEM;
1462 
1463     ret = request_percpu_irq(drvdata->irq, arm_trbe_irq_handler, DRVNAME, drvdata->handle);
1464     if (ret) {
1465         free_percpu(drvdata->handle);
1466         return ret;
1467     }
1468     return 0;
1469 }
1470 
1471 static void arm_trbe_remove_irq(struct trbe_drvdata *drvdata)
1472 {
1473     free_percpu_irq(drvdata->irq, drvdata->handle);
1474     free_percpu(drvdata->handle);
1475 }
1476 
1477 static int arm_trbe_device_probe(struct platform_device *pdev)
1478 {
1479     struct coresight_platform_data *pdata;
1480     struct trbe_drvdata *drvdata;
1481     struct device *dev = &pdev->dev;
1482     int ret;
1483 
1484     /* Trace capture is not possible with kernel page table isolation */
1485     if (arm64_kernel_unmapped_at_el0()) {
1486         pr_err("TRBE wouldn't work if kernel gets unmapped at EL0\n");
1487         return -EOPNOTSUPP;
1488     }
1489 
1490     drvdata = devm_kzalloc(dev, sizeof(*drvdata), GFP_KERNEL);
1491     if (!drvdata)
1492         return -ENOMEM;
1493 
1494     pdata = coresight_get_platform_data(dev);
1495     if (IS_ERR(pdata))
1496         return PTR_ERR(pdata);
1497 
1498     dev_set_drvdata(dev, drvdata);
1499     dev->platform_data = pdata;
1500     drvdata->pdev = pdev;
1501     ret = arm_trbe_probe_irq(pdev, drvdata);
1502     if (ret)
1503         return ret;
1504 
1505     ret = arm_trbe_probe_coresight(drvdata);
1506     if (ret)
1507         goto probe_failed;
1508 
1509     ret = arm_trbe_probe_cpuhp(drvdata);
1510     if (ret)
1511         goto cpuhp_failed;
1512 
1513     return 0;
1514 cpuhp_failed:
1515     arm_trbe_remove_coresight(drvdata);
1516 probe_failed:
1517     arm_trbe_remove_irq(drvdata);
1518     return ret;
1519 }
1520 
1521 static int arm_trbe_device_remove(struct platform_device *pdev)
1522 {
1523     struct trbe_drvdata *drvdata = platform_get_drvdata(pdev);
1524 
1525     arm_trbe_remove_cpuhp(drvdata);
1526     arm_trbe_remove_coresight(drvdata);
1527     arm_trbe_remove_irq(drvdata);
1528     return 0;
1529 }
1530 
1531 static const struct of_device_id arm_trbe_of_match[] = {
1532     { .compatible = "arm,trace-buffer-extension"},
1533     {},
1534 };
1535 MODULE_DEVICE_TABLE(of, arm_trbe_of_match);
1536 
1537 static struct platform_driver arm_trbe_driver = {
1538     .driver = {
1539         .name = DRVNAME,
1540         .of_match_table = of_match_ptr(arm_trbe_of_match),
1541         .suppress_bind_attrs = true,
1542     },
1543     .probe  = arm_trbe_device_probe,
1544     .remove = arm_trbe_device_remove,
1545 };
1546 
1547 static int __init arm_trbe_init(void)
1548 {
1549     int ret;
1550 
1551     ret = platform_driver_register(&arm_trbe_driver);
1552     if (!ret)
1553         return 0;
1554 
1555     pr_err("Error registering %s platform driver\n", DRVNAME);
1556     return ret;
1557 }
1558 
1559 static void __exit arm_trbe_exit(void)
1560 {
1561     platform_driver_unregister(&arm_trbe_driver);
1562 }
1563 module_init(arm_trbe_init);
1564 module_exit(arm_trbe_exit);
1565 
1566 MODULE_AUTHOR("Anshuman Khandual <anshuman.khandual@arm.com>");
1567 MODULE_DESCRIPTION("Arm Trace Buffer Extension (TRBE) driver");
1568 MODULE_LICENSE("GPL v2");