// SPDX-License-Identifier: GPL-2.0-only /* * ARM-specific support for Broadcom STB S2/S3/S5 power management * * S2: clock gate CPUs and as many peripherals as possible * S3: power off all of the chip except the Always ON (AON) island; keep DDR is * self-refresh * S5: (a.k.a. S3 cold boot) much like S3, except DDR is powered down, so we * treat this mode like a soft power-off, with wakeup allowed from AON * * Copyright © 2014-2017 Broadcom */ #define pr_fmt(fmt) "brcmstb-pm: " fmt #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include "pm.h" #include "aon_defs.h" #define SHIMPHY_DDR_PAD_CNTRL 0x8c /* Method #0 */ #define SHIMPHY_PAD_PLL_SEQUENCE BIT(8) #define SHIMPHY_PAD_GATE_PLL_S3 BIT(9) /* Method #1 */ #define PWRDWN_SEQ_NO_SEQUENCING 0 #define PWRDWN_SEQ_HOLD_CHANNEL 1 #define PWRDWN_SEQ_RESET_PLL 2 #define PWRDWN_SEQ_POWERDOWN_PLL 3 #define SHIMPHY_PAD_S3_PWRDWN_SEQ_MASK 0x00f00000 #define SHIMPHY_PAD_S3_PWRDWN_SEQ_SHIFT 20 #define DDR_FORCE_CKE_RST_N BIT(3) #define DDR_PHY_RST_N BIT(2) #define DDR_PHY_CKE BIT(1) #define DDR_PHY_NO_CHANNEL 0xffffffff #define MAX_NUM_MEMC 3 struct brcmstb_memc { void __iomem *ddr_phy_base; void __iomem *ddr_shimphy_base; void __iomem *ddr_ctrl; }; struct brcmstb_pm_control { void __iomem *aon_ctrl_base; void __iomem *aon_sram; struct brcmstb_memc memcs[MAX_NUM_MEMC]; void __iomem *boot_sram; size_t boot_sram_len; bool support_warm_boot; size_t pll_status_offset; int num_memc; struct brcmstb_s3_params *s3_params; dma_addr_t s3_params_pa; int s3entry_method; u32 warm_boot_offset; u32 phy_a_standby_ctrl_offs; u32 phy_b_standby_ctrl_offs; bool needs_ddr_pad; struct platform_device *pdev; }; enum bsp_initiate_command { BSP_CLOCK_STOP = 0x00, BSP_GEN_RANDOM_KEY = 0x4A, BSP_RESTORE_RANDOM_KEY = 0x55, BSP_GEN_FIXED_KEY = 0x63, }; #define PM_INITIATE 0x01 #define PM_INITIATE_SUCCESS 0x00 #define PM_INITIATE_FAIL 0xfe static struct brcmstb_pm_control ctrl; noinline int brcmstb_pm_s3_finish(void); static int (*brcmstb_pm_do_s2_sram)(void __iomem *aon_ctrl_base, void __iomem *ddr_phy_pll_status); static int brcmstb_init_sram(struct device_node *dn) { void __iomem *sram; struct resource res; int ret; ret = of_address_to_resource(dn, 0, &res); if (ret) return ret; /* Uncached, executable remapping of SRAM */ sram = __arm_ioremap_exec(res.start, resource_size(&res), false); if (!sram) return -ENOMEM; ctrl.boot_sram = sram; ctrl.boot_sram_len = resource_size(&res); return 0; } static const struct of_device_id sram_dt_ids[] = { { .compatible = "mmio-sram" }, { /* sentinel */ } }; static int do_bsp_initiate_command(enum bsp_initiate_command cmd) { void __iomem *base = ctrl.aon_ctrl_base; int ret; int timeo = 1000 * 1000; /* 1 second */ writel_relaxed(0, base + AON_CTRL_PM_INITIATE); (void)readl_relaxed(base + AON_CTRL_PM_INITIATE); /* Go! */ writel_relaxed((cmd << 1) | PM_INITIATE, base + AON_CTRL_PM_INITIATE); /* * If firmware doesn't support the 'ack', then just assume it's done * after 10ms. Note that this only works for command 0, BSP_CLOCK_STOP */ if (of_machine_is_compatible("brcm,bcm74371a0")) { (void)readl_relaxed(base + AON_CTRL_PM_INITIATE); mdelay(10); return 0; } for (;;) { ret = readl_relaxed(base + AON_CTRL_PM_INITIATE); if (!(ret & PM_INITIATE)) break; if (timeo <= 0) { pr_err("error: timeout waiting for BSP (%x)\n", ret); break; } timeo -= 50; udelay(50); } return (ret & 0xff) != PM_INITIATE_SUCCESS; } static int brcmstb_pm_handshake(void) { void __iomem *base = ctrl.aon_ctrl_base; u32 tmp; int ret; /* BSP power handshake, v1 */ tmp = readl_relaxed(base + AON_CTRL_HOST_MISC_CMDS); tmp &= ~1UL; writel_relaxed(tmp, base + AON_CTRL_HOST_MISC_CMDS); (void)readl_relaxed(base + AON_CTRL_HOST_MISC_CMDS); ret = do_bsp_initiate_command(BSP_CLOCK_STOP); if (ret) pr_err("BSP handshake failed\n"); /* * HACK: BSP may have internal race on the CLOCK_STOP command. * Avoid touching the BSP for a few milliseconds. */ mdelay(3); return ret; } static inline void shimphy_set(u32 value, u32 mask) { int i; if (!ctrl.needs_ddr_pad) return; for (i = 0; i < ctrl.num_memc; i++) { u32 tmp; tmp = readl_relaxed(ctrl.memcs[i].ddr_shimphy_base + SHIMPHY_DDR_PAD_CNTRL); tmp = value | (tmp & mask); writel_relaxed(tmp, ctrl.memcs[i].ddr_shimphy_base + SHIMPHY_DDR_PAD_CNTRL); } wmb(); /* Complete sequence in order. */ } static inline void ddr_ctrl_set(bool warmboot) { int i; for (i = 0; i < ctrl.num_memc; i++) { u32 tmp; tmp = readl_relaxed(ctrl.memcs[i].ddr_ctrl + ctrl.warm_boot_offset); if (warmboot) tmp |= 1; else tmp &= ~1; /* Cold boot */ writel_relaxed(tmp, ctrl.memcs[i].ddr_ctrl + ctrl.warm_boot_offset); } /* Complete sequence in order */ wmb(); } static inline void s3entry_method0(void) { shimphy_set(SHIMPHY_PAD_GATE_PLL_S3 | SHIMPHY_PAD_PLL_SEQUENCE, 0xffffffff); } static inline void s3entry_method1(void) { /* * S3 Entry Sequence * ----------------- * Step 1: SHIMPHY_ADDR_CNTL_0_DDR_PAD_CNTRL [ S3_PWRDWN_SEQ ] = 3 * Step 2: MEMC_DDR_0_WARM_BOOT [ WARM_BOOT ] = 1 */ shimphy_set((PWRDWN_SEQ_POWERDOWN_PLL << SHIMPHY_PAD_S3_PWRDWN_SEQ_SHIFT), ~SHIMPHY_PAD_S3_PWRDWN_SEQ_MASK); ddr_ctrl_set(true); } static inline void s5entry_method1(void) { int i; /* * S5 Entry Sequence * ----------------- * Step 1: SHIMPHY_ADDR_CNTL_0_DDR_PAD_CNTRL [ S3_PWRDWN_SEQ ] = 3 * Step 2: MEMC_DDR_0_WARM_BOOT [ WARM_BOOT ] = 0 * Step 3: DDR_PHY_CONTROL_REGS_[AB]_0_STANDBY_CONTROL[ CKE ] = 0 * DDR_PHY_CONTROL_REGS_[AB]_0_STANDBY_CONTROL[ RST_N ] = 0 */ shimphy_set((PWRDWN_SEQ_POWERDOWN_PLL << SHIMPHY_PAD_S3_PWRDWN_SEQ_SHIFT), ~SHIMPHY_PAD_S3_PWRDWN_SEQ_MASK); ddr_ctrl_set(false); for (i = 0; i < ctrl.num_memc; i++) { u32 tmp; /* Step 3: Channel A (RST_N = CKE = 0) */ tmp = readl_relaxed(ctrl.memcs[i].ddr_phy_base + ctrl.phy_a_standby_ctrl_offs); tmp &= ~(DDR_PHY_RST_N | DDR_PHY_RST_N); writel_relaxed(tmp, ctrl.memcs[i].ddr_phy_base + ctrl.phy_a_standby_ctrl_offs); /* Step 3: Channel B? */ if (ctrl.phy_b_standby_ctrl_offs != DDR_PHY_NO_CHANNEL) { tmp = readl_relaxed(ctrl.memcs[i].ddr_phy_base + ctrl.phy_b_standby_ctrl_offs); tmp &= ~(DDR_PHY_RST_N | DDR_PHY_RST_N); writel_relaxed(tmp, ctrl.memcs[i].ddr_phy_base + ctrl.phy_b_standby_ctrl_offs); } } /* Must complete */ wmb(); } /* * Run a Power Management State Machine (PMSM) shutdown command and put the CPU * into a low-power mode */ static void brcmstb_do_pmsm_power_down(unsigned long base_cmd, bool onewrite) { void __iomem *base = ctrl.aon_ctrl_base; if ((ctrl.s3entry_method == 1) && (base_cmd == PM_COLD_CONFIG)) s5entry_method1(); /* pm_start_pwrdn transition 0->1 */ writel_relaxed(base_cmd, base + AON_CTRL_PM_CTRL); if (!onewrite) { (void)readl_relaxed(base + AON_CTRL_PM_CTRL); writel_relaxed(base_cmd | PM_PWR_DOWN, base + AON_CTRL_PM_CTRL); (void)readl_relaxed(base + AON_CTRL_PM_CTRL); } wfi(); } /* Support S5 cold boot out of "poweroff" */ static void brcmstb_pm_poweroff(void) { brcmstb_pm_handshake(); /* Clear magic S3 warm-boot value */ writel_relaxed(0, ctrl.aon_sram + AON_REG_MAGIC_FLAGS); (void)readl_relaxed(ctrl.aon_sram + AON_REG_MAGIC_FLAGS); /* Skip wait-for-interrupt signal; just use a countdown */ writel_relaxed(0x10, ctrl.aon_ctrl_base + AON_CTRL_PM_CPU_WAIT_COUNT); (void)readl_relaxed(ctrl.aon_ctrl_base + AON_CTRL_PM_CPU_WAIT_COUNT); if (ctrl.s3entry_method == 1) { shimphy_set((PWRDWN_SEQ_POWERDOWN_PLL << SHIMPHY_PAD_S3_PWRDWN_SEQ_SHIFT), ~SHIMPHY_PAD_S3_PWRDWN_SEQ_MASK); ddr_ctrl_set(false); brcmstb_do_pmsm_power_down(M1_PM_COLD_CONFIG, true); return; /* We should never actually get here */ } brcmstb_do_pmsm_power_down(PM_COLD_CONFIG, false); } static void *brcmstb_pm_copy_to_sram(void *fn, size_t len) { unsigned int size = ALIGN(len, FNCPY_ALIGN); if (ctrl.boot_sram_len < size) { pr_err("standby code will not fit in SRAM\n"); return NULL; } return fncpy(ctrl.boot_sram, fn, size); } /* * S2 suspend/resume picks up where we left off, so we must execute carefully * from SRAM, in order to allow DDR to come back up safely before we continue. */ static int brcmstb_pm_s2(void) { /* A previous S3 can set a value hazardous to S2, so make sure. */ if (ctrl.s3entry_method == 1) { shimphy_set((PWRDWN_SEQ_NO_SEQUENCING << SHIMPHY_PAD_S3_PWRDWN_SEQ_SHIFT), ~SHIMPHY_PAD_S3_PWRDWN_SEQ_MASK); ddr_ctrl_set(false); } brcmstb_pm_do_s2_sram = brcmstb_pm_copy_to_sram(&brcmstb_pm_do_s2, brcmstb_pm_do_s2_sz); if (!brcmstb_pm_do_s2_sram) return -EINVAL; return brcmstb_pm_do_s2_sram(ctrl.aon_ctrl_base, ctrl.memcs[0].ddr_phy_base + ctrl.pll_status_offset); } /* * This function is called on a new stack, so don't allow inlining (which will * generate stack references on the old stack). It cannot be made static because * it is referenced from brcmstb_pm_s3() */ noinline int brcmstb_pm_s3_finish(void) { struct brcmstb_s3_params *params = ctrl.s3_params; dma_addr_t params_pa = ctrl.s3_params_pa; phys_addr_t reentry = virt_to_phys(&cpu_resume_arm); enum bsp_initiate_command cmd; u32 flags; /* * Clear parameter structure, but not DTU area, which has already been * filled in. We know DTU is a the end, so we can just subtract its * size. */ memset(params, 0, sizeof(*params) - sizeof(params->dtu)); flags = readl_relaxed(ctrl.aon_sram + AON_REG_MAGIC_FLAGS); flags &= S3_BOOTLOADER_RESERVED; flags |= S3_FLAG_NO_MEM_VERIFY; flags |= S3_FLAG_LOAD_RANDKEY; /* Load random / fixed key */ if (flags & S3_FLAG_LOAD_RANDKEY) cmd = BSP_GEN_RANDOM_KEY; else cmd = BSP_GEN_FIXED_KEY; if (do_bsp_initiate_command(cmd)) { pr_info("key loading failed\n"); return -EIO; } params->magic = BRCMSTB_S3_MAGIC; params->reentry = reentry; /* No more writes to DRAM */ flush_cache_all(); flags |= BRCMSTB_S3_MAGIC_SHORT; writel_relaxed(flags, ctrl.aon_sram + AON_REG_MAGIC_FLAGS); writel_relaxed(lower_32_bits(params_pa), ctrl.aon_sram + AON_REG_CONTROL_LOW); writel_relaxed(upper_32_bits(params_pa), ctrl.aon_sram + AON_REG_CONTROL_HIGH); switch (ctrl.s3entry_method) { case 0: s3entry_method0(); brcmstb_do_pmsm_power_down(PM_WARM_CONFIG, false); break; case 1: s3entry_method1(); brcmstb_do_pmsm_power_down(M1_PM_WARM_CONFIG, true); break; default: return -EINVAL; } /* Must have been interrupted from wfi()? */ return -EINTR; } static int brcmstb_pm_do_s3(unsigned long sp) { unsigned long save_sp; int ret; asm volatile ( "mov %[save], sp\n" "mov sp, %[new]\n" "bl brcmstb_pm_s3_finish\n" "mov %[ret], r0\n" "mov %[new], sp\n" "mov sp, %[save]\n" : [save] "=&r" (save_sp), [ret] "=&r" (ret) : [new] "r" (sp) ); return ret; } static int brcmstb_pm_s3(void) { void __iomem *sp = ctrl.boot_sram + ctrl.boot_sram_len; return cpu_suspend((unsigned long)sp, brcmstb_pm_do_s3); } static int brcmstb_pm_standby(bool deep_standby) { int ret; if (brcmstb_pm_handshake()) return -EIO; if (deep_standby) ret = brcmstb_pm_s3(); else ret = brcmstb_pm_s2(); if (ret) pr_err("%s: standby failed\n", __func__); return ret; } static int brcmstb_pm_enter(suspend_state_t state) { int ret = -EINVAL; switch (state) { case PM_SUSPEND_STANDBY: ret = brcmstb_pm_standby(false); break; case PM_SUSPEND_MEM: ret = brcmstb_pm_standby(true); break; } return ret; } static int brcmstb_pm_valid(suspend_state_t state) { switch (state) { case PM_SUSPEND_STANDBY: return true; case PM_SUSPEND_MEM: return ctrl.support_warm_boot; default: return false; } } static const struct platform_suspend_ops brcmstb_pm_ops = { .enter = brcmstb_pm_enter, .valid = brcmstb_pm_valid, }; static const struct of_device_id aon_ctrl_dt_ids[] = { { .compatible = "brcm,brcmstb-aon-ctrl" }, {} }; struct ddr_phy_ofdata { bool supports_warm_boot; size_t pll_status_offset; int s3entry_method; u32 warm_boot_offset; u32 phy_a_standby_ctrl_offs; u32 phy_b_standby_ctrl_offs; }; static struct ddr_phy_ofdata ddr_phy_71_1 = { .supports_warm_boot = true, .pll_status_offset = 0x0c, .s3entry_method = 1, .warm_boot_offset = 0x2c, .phy_a_standby_ctrl_offs = 0x198, .phy_b_standby_ctrl_offs = DDR_PHY_NO_CHANNEL }; static struct ddr_phy_ofdata ddr_phy_72_0 = { .supports_warm_boot = true, .pll_status_offset = 0x10, .s3entry_method = 1, .warm_boot_offset = 0x40, .phy_a_standby_ctrl_offs = 0x2a4, .phy_b_standby_ctrl_offs = 0x8a4 }; static struct ddr_phy_ofdata ddr_phy_225_1 = { .supports_warm_boot = false, .pll_status_offset = 0x4, .s3entry_method = 0 }; static struct ddr_phy_ofdata ddr_phy_240_1 = { .supports_warm_boot = true, .pll_status_offset = 0x4, .s3entry_method = 0 }; static const struct of_device_id ddr_phy_dt_ids[] = { { .compatible = "brcm,brcmstb-ddr-phy-v71.1", .data = &ddr_phy_71_1, }, { .compatible = "brcm,brcmstb-ddr-phy-v72.0", .data = &ddr_phy_72_0, }, { .compatible = "brcm,brcmstb-ddr-phy-v225.1", .data = &ddr_phy_225_1, }, { .compatible = "brcm,brcmstb-ddr-phy-v240.1", .data = &ddr_phy_240_1, }, { /* Same as v240.1, for the registers we care about */ .compatible = "brcm,brcmstb-ddr-phy-v240.2", .data = &ddr_phy_240_1, }, {} }; struct ddr_seq_ofdata { bool needs_ddr_pad; u32 warm_boot_offset; }; static const struct ddr_seq_ofdata ddr_seq_b22 = { .needs_ddr_pad = false, .warm_boot_offset = 0x2c, }; static const struct ddr_seq_ofdata ddr_seq = { .needs_ddr_pad = true, }; static const struct of_device_id ddr_shimphy_dt_ids[] = { { .compatible = "brcm,brcmstb-ddr-shimphy-v1.0" }, {} }; static const struct of_device_id brcmstb_memc_of_match[] = { { .compatible = "brcm,brcmstb-memc-ddr-rev-b.2.1", .data = &ddr_seq, }, { .compatible = "brcm,brcmstb-memc-ddr-rev-b.2.2", .data = &ddr_seq_b22, }, { .compatible = "brcm,brcmstb-memc-ddr-rev-b.2.3", .data = &ddr_seq_b22, }, { .compatible = "brcm,brcmstb-memc-ddr-rev-b.3.0", .data = &ddr_seq_b22, }, { .compatible = "brcm,brcmstb-memc-ddr-rev-b.3.1", .data = &ddr_seq_b22, }, { .compatible = "brcm,brcmstb-memc-ddr", .data = &ddr_seq, }, {}, }; static void __iomem *brcmstb_ioremap_match(const struct of_device_id *matches, int index, const void **ofdata) { struct device_node *dn; const struct of_device_id *match; dn = of_find_matching_node_and_match(NULL, matches, &match); if (!dn) return ERR_PTR(-EINVAL); if (ofdata) *ofdata = match->data; return of_io_request_and_map(dn, index, dn->full_name); } static int brcmstb_pm_panic_notify(struct notifier_block *nb, unsigned long action, void *data) { writel_relaxed(BRCMSTB_PANIC_MAGIC, ctrl.aon_sram + AON_REG_PANIC); return NOTIFY_DONE; } static struct notifier_block brcmstb_pm_panic_nb = { .notifier_call = brcmstb_pm_panic_notify, }; static int brcmstb_pm_probe(struct platform_device *pdev) { const struct ddr_phy_ofdata *ddr_phy_data; const struct ddr_seq_ofdata *ddr_seq_data; const struct of_device_id *of_id = NULL; struct device_node *dn; void __iomem *base; int ret, i; /* AON ctrl registers */ base = brcmstb_ioremap_match(aon_ctrl_dt_ids, 0, NULL); if (IS_ERR(base)) { pr_err("error mapping AON_CTRL\n"); return PTR_ERR(base); } ctrl.aon_ctrl_base = base; /* AON SRAM registers */ base = brcmstb_ioremap_match(aon_ctrl_dt_ids, 1, NULL); if (IS_ERR(base)) { /* Assume standard offset */ ctrl.aon_sram = ctrl.aon_ctrl_base + AON_CTRL_SYSTEM_DATA_RAM_OFS; } else { ctrl.aon_sram = base; } writel_relaxed(0, ctrl.aon_sram + AON_REG_PANIC); /* DDR PHY registers */ base = brcmstb_ioremap_match(ddr_phy_dt_ids, 0, (const void **)&ddr_phy_data); if (IS_ERR(base)) { pr_err("error mapping DDR PHY\n"); return PTR_ERR(base); } ctrl.support_warm_boot = ddr_phy_data->supports_warm_boot; ctrl.pll_status_offset = ddr_phy_data->pll_status_offset; /* Only need DDR PHY 0 for now? */ ctrl.memcs[0].ddr_phy_base = base; ctrl.s3entry_method = ddr_phy_data->s3entry_method; ctrl.phy_a_standby_ctrl_offs = ddr_phy_data->phy_a_standby_ctrl_offs; ctrl.phy_b_standby_ctrl_offs = ddr_phy_data->phy_b_standby_ctrl_offs; /* * Slightly gross to use the phy ver to get a memc, * offset but that is the only versioned things so far * we can test for. */ ctrl.warm_boot_offset = ddr_phy_data->warm_boot_offset; /* DDR SHIM-PHY registers */ for_each_matching_node(dn, ddr_shimphy_dt_ids) { i = ctrl.num_memc; if (i >= MAX_NUM_MEMC) { pr_warn("too many MEMCs (max %d)\n", MAX_NUM_MEMC); break; } base = of_io_request_and_map(dn, 0, dn->full_name); if (IS_ERR(base)) { if (!ctrl.support_warm_boot) break; pr_err("error mapping DDR SHIMPHY %d\n", i); return PTR_ERR(base); } ctrl.memcs[i].ddr_shimphy_base = base; ctrl.num_memc++; } /* Sequencer DRAM Param and Control Registers */ i = 0; for_each_matching_node(dn, brcmstb_memc_of_match) { base = of_iomap(dn, 0); if (!base) { pr_err("error mapping DDR Sequencer %d\n", i); return -ENOMEM; } of_id = of_match_node(brcmstb_memc_of_match, dn); if (!of_id) { iounmap(base); return -EINVAL; } ddr_seq_data = of_id->data; ctrl.needs_ddr_pad = ddr_seq_data->needs_ddr_pad; /* Adjust warm boot offset based on the DDR sequencer */ if (ddr_seq_data->warm_boot_offset) ctrl.warm_boot_offset = ddr_seq_data->warm_boot_offset; ctrl.memcs[i].ddr_ctrl = base; i++; } pr_debug("PM: supports warm boot:%d, method:%d, wboffs:%x\n", ctrl.support_warm_boot, ctrl.s3entry_method, ctrl.warm_boot_offset); dn = of_find_matching_node(NULL, sram_dt_ids); if (!dn) { pr_err("SRAM not found\n"); return -EINVAL; } ret = brcmstb_init_sram(dn); of_node_put(dn); if (ret) { pr_err("error setting up SRAM for PM\n"); return ret; } ctrl.pdev = pdev; ctrl.s3_params = kmalloc(sizeof(*ctrl.s3_params), GFP_KERNEL); if (!ctrl.s3_params) return -ENOMEM; ctrl.s3_params_pa = dma_map_single(&pdev->dev, ctrl.s3_params, sizeof(*ctrl.s3_params), DMA_TO_DEVICE); if (dma_mapping_error(&pdev->dev, ctrl.s3_params_pa)) { pr_err("error mapping DMA memory\n"); ret = -ENOMEM; goto out; } atomic_notifier_chain_register(&panic_notifier_list, &brcmstb_pm_panic_nb); pm_power_off = brcmstb_pm_poweroff; suspend_set_ops(&brcmstb_pm_ops); return 0; out: kfree(ctrl.s3_params); pr_warn("PM: initialization failed with code %d\n", ret); return ret; } static struct platform_driver brcmstb_pm_driver = { .driver = { .name = "brcmstb-pm", .of_match_table = aon_ctrl_dt_ids, }, }; static int __init brcmstb_pm_init(void) { return platform_driver_probe(&brcmstb_pm_driver, brcmstb_pm_probe); } module_init(brcmstb_pm_init);