/* * Copyright (c) 2019, ARM Limited and Contributors. All rights reserved. * * SPDX-License-Identifier: BSD-3-Clause */ #include #include #include #include #include #include #include #include #include #define MAX_CLUSTERS 5 static unsigned long cci_base_addr; static unsigned int cci_cluster_ix_to_iface[MAX_CLUSTERS]; void mcsi_init(unsigned long cci_base, unsigned int num_cci_masters) { int i; assert(cci_base); assert(num_cci_masters < MAX_CLUSTERS); cci_base_addr = cci_base; for (i = 0; i < num_cci_masters; i++) cci_cluster_ix_to_iface[i] = SLAVE_IFACE_OFFSET(i); } void mcsi_cache_flush(void) { /* timeout is 10ms */ int timeout = 10000; /* to make flush by SF safe, need to disable BIU DCM */ mmio_clrbits_32(CCI_CLK_CTRL, 1 << 8); mmio_write_32(cci_base_addr + FLUSH_SF, 0x1); for (; timeout; timeout--, udelay(1)) { if ((mmio_read_32(cci_base_addr + FLUSH_SF) & 0x1) == 0x0) break; } if (!timeout) { INFO("SF lush timeout\n"); return; } /* enable BIU DCM as it was */ mmio_setbits_32(CCI_CLK_CTRL, 1 << 8); } static inline unsigned long get_slave_iface_base(unsigned long mpidr) { /* * We assume the TF topology code allocates affinity instances * consecutively from zero. * It is a programming error if this is called without initializing * the slave interface to use for this cluster. */ unsigned int cluster_id = (mpidr >> MPIDR_AFF1_SHIFT) & MPIDR_AFFLVL_MASK; assert(cluster_id < MAX_CLUSTERS); assert(cci_cluster_ix_to_iface[cluster_id] != 0); return cci_base_addr + cci_cluster_ix_to_iface[cluster_id]; } void cci_enable_cluster_coherency(unsigned long mpidr) { unsigned long slave_base; unsigned int support_ability; unsigned int config = 0; unsigned int pending = 0; assert(cci_base_addr); slave_base = get_slave_iface_base(mpidr); support_ability = mmio_read_32(slave_base); pending = (mmio_read_32( cci_base_addr + SNP_PENDING_REG)) >> SNP_PENDING; while (pending) { pending = (mmio_read_32( cci_base_addr + SNP_PENDING_REG)) >> SNP_PENDING; } if (support_ability & SNP_SUPPORT) config |= SNOOP_EN_BIT; if (support_ability & DVM_SUPPORT) config |= DVM_EN_BIT; mmio_write_32(slave_base, support_ability | config); /* Wait for the dust to settle down */ while (mmio_read_32(cci_base_addr + SNP_PENDING_REG) >> SNP_PENDING) ; } #if ERRATA_MCSIB_SW #pragma weak mcsib_sw_workaround_main #endif void cci_disable_cluster_coherency(unsigned long mpidr) { unsigned long slave_base; unsigned int config = 0; assert(cci_base_addr); slave_base = get_slave_iface_base(mpidr); while (mmio_read_32(cci_base_addr + SNP_PENDING_REG) >> SNP_PENDING) ; config = mmio_read_32(slave_base); config &= ~(DVM_EN_BIT | SNOOP_EN_BIT); /* Disable Snoops and DVM messages */ mmio_write_32(slave_base, config); #if ERRATA_MCSIB_SW mcsib_sw_workaround_main(); #endif /* Wait for the dust to settle down */ while (mmio_read_32(cci_base_addr + SNP_PENDING_REG) >> SNP_PENDING) ; } void cci_secure_switch(unsigned int status) { unsigned int config; config = mmio_read_32(cci_base_addr + CENTRAL_CTRL_REG); if (status == NS_ACC) config |= SECURE_ACC_EN; else config &= ~SECURE_ACC_EN; mmio_write_32(cci_base_addr + CENTRAL_CTRL_REG, config); } void cci_pmu_secure_switch(unsigned int status) { unsigned int config; config = mmio_read_32(cci_base_addr + CENTRAL_CTRL_REG); if (status == NS_ACC) config |= PMU_SECURE_ACC_EN; else config &= ~PMU_SECURE_ACC_EN; mmio_write_32(cci_base_addr + CENTRAL_CTRL_REG, config); } void cci_init_sf(void) { while (mmio_read_32(cci_base_addr + SNP_PENDING_REG) >> SNP_PENDING) ; /* init sf1 */ mmio_write_32(cci_base_addr + SF_INIT_REG, TRIG_SF1_INIT); while (mmio_read_32(cci_base_addr + SF_INIT_REG) & TRIG_SF1_INIT) ; while (!(mmio_read_32(cci_base_addr + SF_INIT_REG) & SF1_INIT_DONE)) ; /* init sf2 */ mmio_write_32(cci_base_addr + SF_INIT_REG, TRIG_SF2_INIT); while (mmio_read_32(cci_base_addr + SF_INIT_REG) & TRIG_SF2_INIT) ; while (!(mmio_read_32(cci_base_addr + SF_INIT_REG) & SF2_INIT_DONE)) ; } void cci_interrupt_en(void) { mmio_setbits_32(cci_base_addr + CENTRAL_CTRL_REG, INT_EN); } unsigned long cci_reg_access(unsigned int op, unsigned long offset, unsigned long val) { unsigned long ret = 0; if ((cci_base_addr == 0) || (offset > MSCI_MEMORY_SZ)) panic(); switch (op) { case MCSI_REG_ACCESS_READ: ret = mmio_read_32(cci_base_addr + offset); break; case MCSI_REG_ACCESS_WRITE: mmio_write_32(cci_base_addr + offset, val); dsb(); break; case MCSI_REG_ACCESS_SET_BITMASK: mmio_setbits_32(cci_base_addr + offset, val); dsb(); break; case MCSI_REG_ACCESS_CLEAR_BITMASK: mmio_clrbits_32(cci_base_addr + offset, val); dsb(); break; default: break; } return ret; }