Implement {spe,sve}_supported() helpers and refactor code

Implement helpers to test if the core supports SPE/SVE.  We have a
similar helper for AMU and this patch makes all extensions consistent
in their implementation.

Change-Id: I3e6f7522535ca358259ad142550b19fcb883ca67
Signed-off-by: Dimitris Papastamos <dimitris.papastamos@arm.com>
This commit is contained in:
Dimitris Papastamos 2018-02-19 14:52:19 +00:00
parent c7aa7fdf56
commit 2ff8fbf3b0
4 changed files with 135 additions and 129 deletions

View File

@ -1,5 +1,5 @@
/* /*
* Copyright (c) 2017, ARM Limited and Contributors. All rights reserved. * Copyright (c) 2017-2018, ARM Limited and Contributors. All rights reserved.
* *
* SPDX-License-Identifier: BSD-3-Clause * SPDX-License-Identifier: BSD-3-Clause
*/ */
@ -7,6 +7,7 @@
#ifndef __SPE_H__ #ifndef __SPE_H__
#define __SPE_H__ #define __SPE_H__
int spe_supported(void);
void spe_enable(int el2_unused); void spe_enable(int el2_unused);
void spe_disable(void); void spe_disable(void);

View File

@ -1,5 +1,5 @@
/* /*
* Copyright (c) 2017, ARM Limited and Contributors. All rights reserved. * Copyright (c) 2017-2018, ARM Limited and Contributors. All rights reserved.
* *
* SPDX-License-Identifier: BSD-3-Clause * SPDX-License-Identifier: BSD-3-Clause
*/ */
@ -7,6 +7,7 @@
#ifndef __SVE_H__ #ifndef __SVE_H__
#define __SVE_H__ #define __SVE_H__
int sve_supported(void);
void sve_enable(int el2_unused); void sve_enable(int el2_unused);
#endif /* __SVE_H__ */ #endif /* __SVE_H__ */

View File

@ -1,5 +1,5 @@
/* /*
* Copyright (c) 2017, ARM Limited and Contributors. All rights reserved. * Copyright (c) 2017-2018, ARM Limited and Contributors. All rights reserved.
* *
* SPDX-License-Identifier: BSD-3-Clause * SPDX-License-Identifier: BSD-3-Clause
*/ */
@ -14,71 +14,72 @@
*/ */
#define psb_csync() asm volatile("hint #17") #define psb_csync() asm volatile("hint #17")
void spe_enable(int el2_unused) int spe_supported(void)
{ {
uint64_t features; uint64_t features;
features = read_id_aa64dfr0_el1() >> ID_AA64DFR0_PMS_SHIFT; features = read_id_aa64dfr0_el1() >> ID_AA64DFR0_PMS_SHIFT;
if ((features & ID_AA64DFR0_PMS_MASK) == 1) { return (features & ID_AA64DFR0_PMS_MASK) == 1;
uint64_t v; }
if (el2_unused) { void spe_enable(int el2_unused)
/* {
* MDCR_EL2.TPMS (ARM v8.2): Do not trap statistical uint64_t v;
* profiling controls to EL2.
*
* MDCR_EL2.E2PB (ARM v8.2): SPE enabled in Non-secure
* state. Accesses to profiling buffer controls at
* Non-secure EL1 are not trapped to EL2.
*/
v = read_mdcr_el2();
v &= ~MDCR_EL2_TPMS;
v |= MDCR_EL2_E2PB(MDCR_EL2_E2PB_EL1);
write_mdcr_el2(v);
}
if (!spe_supported())
return;
if (el2_unused) {
/* /*
* MDCR_EL2.NSPB (ARM v8.2): SPE enabled in Non-secure state * MDCR_EL2.TPMS (ARM v8.2): Do not trap statistical
* and disabled in secure state. Accesses to SPE registers at * profiling controls to EL2.
* S-EL1 generate trap exceptions to EL3. *
* MDCR_EL2.E2PB (ARM v8.2): SPE enabled in Non-secure
* state. Accesses to profiling buffer controls at
* Non-secure EL1 are not trapped to EL2.
*/ */
v = read_mdcr_el3(); v = read_mdcr_el2();
v |= MDCR_NSPB(MDCR_NSPB_EL1); v &= ~MDCR_EL2_TPMS;
write_mdcr_el3(v); v |= MDCR_EL2_E2PB(MDCR_EL2_E2PB_EL1);
write_mdcr_el2(v);
} }
/*
* MDCR_EL2.NSPB (ARM v8.2): SPE enabled in Non-secure state
* and disabled in secure state. Accesses to SPE registers at
* S-EL1 generate trap exceptions to EL3.
*/
v = read_mdcr_el3();
v |= MDCR_NSPB(MDCR_NSPB_EL1);
write_mdcr_el3(v);
} }
void spe_disable(void) void spe_disable(void)
{ {
uint64_t features; uint64_t v;
features = read_id_aa64dfr0_el1() >> ID_AA64DFR0_PMS_SHIFT; if (!spe_supported())
if ((features & ID_AA64DFR0_PMS_MASK) == 1) { return;
uint64_t v;
/* Drain buffered data */ /* Drain buffered data */
psb_csync(); psb_csync();
dsbnsh(); dsbnsh();
/* Disable profiling buffer */ /* Disable profiling buffer */
v = read_pmblimitr_el1(); v = read_pmblimitr_el1();
v &= ~(1ULL << 0); v &= ~(1ULL << 0);
write_pmblimitr_el1(v); write_pmblimitr_el1(v);
isb(); isb();
}
} }
static void *spe_drain_buffers_hook(const void *arg) static void *spe_drain_buffers_hook(const void *arg)
{ {
uint64_t features; if (!spe_supported())
return (void *)-1;
features = read_id_aa64dfr0_el1() >> ID_AA64DFR0_PMS_SHIFT;
if ((features & ID_AA64DFR0_PMS_MASK) == 1) {
/* Drain buffered data */
psb_csync();
dsbnsh();
}
/* Drain buffered data */
psb_csync();
dsbnsh();
return 0; return 0;
} }

View File

@ -1,5 +1,5 @@
/* /*
* Copyright (c) 2017, ARM Limited and Contributors. All rights reserved. * Copyright (c) 2017-2018, ARM Limited and Contributors. All rights reserved.
* *
* SPDX-License-Identifier: BSD-3-Clause * SPDX-License-Identifier: BSD-3-Clause
*/ */
@ -9,117 +9,120 @@
#include <pubsub.h> #include <pubsub.h>
#include <sve.h> #include <sve.h>
static void *disable_sve_hook(const void *arg) int sve_supported(void)
{ {
uint64_t features; uint64_t features;
features = read_id_aa64pfr0_el1() >> ID_AA64PFR0_SVE_SHIFT; features = read_id_aa64pfr0_el1() >> ID_AA64PFR0_SVE_SHIFT;
if ((features & ID_AA64PFR0_SVE_MASK) == 1) { return (features & ID_AA64PFR0_SVE_MASK) == 1;
uint64_t cptr; }
/* static void *disable_sve_hook(const void *arg)
* Disable SVE, SIMD and FP access for the Secure world. {
* As the SIMD/FP registers are part of the SVE Z-registers, any uint64_t cptr;
* use of SIMD/FP functionality will corrupt the SVE registers.
* Therefore it is necessary to prevent use of SIMD/FP support
* in the Secure world as well as SVE functionality.
*/
cptr = read_cptr_el3();
cptr = (cptr | TFP_BIT) & ~(CPTR_EZ_BIT);
write_cptr_el3(cptr);
/* if (!sve_supported())
* No explicit ISB required here as ERET to switch to Secure return (void *)-1;
* world covers it
*/ /*
} * Disable SVE, SIMD and FP access for the Secure world.
* As the SIMD/FP registers are part of the SVE Z-registers, any
* use of SIMD/FP functionality will corrupt the SVE registers.
* Therefore it is necessary to prevent use of SIMD/FP support
* in the Secure world as well as SVE functionality.
*/
cptr = read_cptr_el3();
cptr = (cptr | TFP_BIT) & ~(CPTR_EZ_BIT);
write_cptr_el3(cptr);
/*
* No explicit ISB required here as ERET to switch to Secure
* world covers it
*/
return 0; return 0;
} }
static void *enable_sve_hook(const void *arg) static void *enable_sve_hook(const void *arg)
{ {
uint64_t features; uint64_t cptr;
features = read_id_aa64pfr0_el1() >> ID_AA64PFR0_SVE_SHIFT; if (!sve_supported())
if ((features & ID_AA64PFR0_SVE_MASK) == 1) { return (void *)-1;
uint64_t cptr;
/* /*
* Enable SVE, SIMD and FP access for the Non-secure world. * Enable SVE, SIMD and FP access for the Non-secure world.
*/ */
cptr = read_cptr_el3(); cptr = read_cptr_el3();
cptr = (cptr | CPTR_EZ_BIT) & ~(TFP_BIT); cptr = (cptr | CPTR_EZ_BIT) & ~(TFP_BIT);
write_cptr_el3(cptr); write_cptr_el3(cptr);
/* /*
* No explicit ISB required here as ERET to switch to Non-secure * No explicit ISB required here as ERET to switch to Non-secure
* world covers it * world covers it
*/ */
}
return 0; return 0;
} }
void sve_enable(int el2_unused) void sve_enable(int el2_unused)
{ {
uint64_t features; uint64_t cptr;
if (!sve_supported())
return;
features = read_id_aa64pfr0_el1() >> ID_AA64PFR0_SVE_SHIFT;
if ((features & ID_AA64PFR0_SVE_MASK) == 1) {
uint64_t cptr;
#if CTX_INCLUDE_FPREGS #if CTX_INCLUDE_FPREGS
/* /*
* CTX_INCLUDE_FPREGS is not supported on SVE enabled systems. * CTX_INCLUDE_FPREGS is not supported on SVE enabled systems.
*/ */
assert(0); assert(0);
#endif #endif
/* /*
* Update CPTR_EL3 to enable access to SVE functionality for the * Update CPTR_EL3 to enable access to SVE functionality for the
* Non-secure world. * Non-secure world.
* NOTE - assumed that CPTR_EL3.TFP is set to allow access to * NOTE - assumed that CPTR_EL3.TFP is set to allow access to
* the SIMD, floating-point and SVE support. * the SIMD, floating-point and SVE support.
* *
* CPTR_EL3.EZ: Set to 1 to enable access to SVE functionality * CPTR_EL3.EZ: Set to 1 to enable access to SVE functionality
* in the Non-secure world. * in the Non-secure world.
*/ */
cptr = read_cptr_el3(); cptr = read_cptr_el3();
cptr |= CPTR_EZ_BIT; cptr |= CPTR_EZ_BIT;
write_cptr_el3(cptr); write_cptr_el3(cptr);
/*
* Need explicit ISB here to guarantee that update to ZCR_ELx
* and CPTR_EL2.TZ do not result in trap to EL3.
*/
isb();
/*
* Ensure lower ELs have access to full vector length.
*/
write_zcr_el3(ZCR_EL3_LEN_MASK);
if (el2_unused) {
/* /*
* Need explicit ISB here to guarantee that update to ZCR_ELx * Update CPTR_EL2 to enable access to SVE functionality
* and CPTR_EL2.TZ do not result in trap to EL3. * for Non-secure world, EL2 and Non-secure EL1 and EL0.
* NOTE - assumed that CPTR_EL2.TFP is set to allow
* access to the SIMD, floating-point and SVE support.
*
* CPTR_EL2.TZ: Set to 0 to enable access to SVE support
* for EL2 and Non-secure EL1 and EL0.
*/ */
isb(); cptr = read_cptr_el2();
cptr &= ~(CPTR_EL2_TZ_BIT);
write_cptr_el2(cptr);
/* /*
* Ensure lower ELs have access to full vector length. * Ensure lower ELs have access to full vector length.
*/ */
write_zcr_el3(ZCR_EL3_LEN_MASK); write_zcr_el2(ZCR_EL2_LEN_MASK);
if (el2_unused) {
/*
* Update CPTR_EL2 to enable access to SVE functionality
* for Non-secure world, EL2 and Non-secure EL1 and EL0.
* NOTE - assumed that CPTR_EL2.TFP is set to allow
* access to the SIMD, floating-point and SVE support.
*
* CPTR_EL2.TZ: Set to 0 to enable access to SVE support
* for EL2 and Non-secure EL1 and EL0.
*/
cptr = read_cptr_el2();
cptr &= ~(CPTR_EL2_TZ_BIT);
write_cptr_el2(cptr);
/*
* Ensure lower ELs have access to full vector length.
*/
write_zcr_el2(ZCR_EL2_LEN_MASK);
}
/*
* No explicit ISB required here as ERET to switch to
* Non-secure world covers it.
*/
} }
/*
* No explicit ISB required here as ERET to switch to
* Non-secure world covers it.
*/
} }
SUBSCRIBE_TO_EVENT(cm_exited_normal_world, disable_sve_hook); SUBSCRIBE_TO_EVENT(cm_exited_normal_world, disable_sve_hook);