// SPDX-License-Identifier: GPL-2.0 /* * Copyright (c) 2019 Fuzhou Rockchip Electronics Co., Ltd */ #include #include #include #include #include #include DECLARE_GLOBAL_DATA_PTR; #define SYSMEM_MAGIC 0x4D454D53 /* "SMEM" */ #define LMB_ALLOC_ANYWHERE 0 /* sync with lmb.c */ #define SYSMEM_ALLOC_NO_ALIGN 1 #define SYSMEM_ALLOC_ANYWHERE 2 #define SYSMEM_I(fmt, args...) printf("Sysmem: "fmt, ##args) #define SYSMEM_W(fmt, args...) printf("Sysmem Warn: "fmt, ##args) #define SYSMEM_E(fmt, args...) printf("Sysmem Error: "fmt, ##args) #define SYSMEM_D(fmt, args...) debug("Sysmem Debug: "fmt, ##args) struct memcheck { uint32_t magic; }; /* Global for platform, must in data section */ struct sysmem plat_sysmem __section(".data") = { .has_initf = false, .has_initr = false, }; bool sysmem_has_init(void) { return gd->flags & GD_FLG_RELOC ? plat_sysmem.has_initr : plat_sysmem.has_initf; } static inline int sysmem_is_overlap(phys_addr_t base1, phys_size_t size1, phys_addr_t base2, phys_size_t size2) { return ((base1 < (base2 + size2)) && (base2 < (base1 + size1))); } static inline int sysmem_is_sub_region(struct memblock *sub, struct memblock *main) { if (!sub || !main) return false; return ((sub->base >= main->base) && (sub->base + sub->size <= main->base + main->size)); } void sysmem_dump(void) { struct sysmem *sysmem = &plat_sysmem; struct lmb *lmb = &sysmem->lmb; struct memblock *mem; struct memcheck *check; struct list_head *node; ulong memory_size = 0; ulong reserved_size = 0; ulong allocated_size = 0; bool overflow = false; ulong i; if (!sysmem_has_init()) return; printf("\nsysmem_dump_all:\n"); /* Memory pool */ printf(" --------------------------------------------------------------------\n"); for (i = 0; i < lmb->memory.cnt; i++) { memory_size += lmb->memory.region[i].size; printf(" memory.rgn[%ld].addr = 0x%08lx - 0x%08lx (size: 0x%08lx)\n", i, (ulong)lmb->memory.region[i].base, (ulong)lmb->memory.region[i].base + (ulong)lmb->memory.region[i].size, (ulong)lmb->memory.region[i].size); } printf("\n memory.total = 0x%08lx (%ld MiB. %ld KiB)\n", (ulong)memory_size, SIZE_MB((ulong)memory_size), SIZE_KB((ulong)memory_size)); /* Allocated */ i = 0; printf(" --------------------------------------------------------------------\n"); list_for_each(node, &sysmem->allocated_head) { mem = list_entry(node, struct memblock, node); allocated_size += mem->size; if (mem->attr.flags & F_OFC) { check = (struct memcheck *) (mem->base + mem->size - sizeof(*check)); overflow = (check->magic != SYSMEM_MAGIC); } else if (mem->attr.flags & F_HOFC) { check = (struct memcheck *) (mem->base - sizeof(*check)); overflow = (check->magic != SYSMEM_MAGIC); } else { overflow = false; } printf(" allocated.rgn[%ld].name = \"%s\" %s %s\n", i, mem->attr.name, overflow ? " " : "", mem->orig_base != mem->base ? "<*>" : ""); printf(" .addr = 0x%08lx - 0x%08lx (size: 0x%08lx)\n", (ulong)mem->orig_base, (ulong)(mem->orig_base + mem->size), (ulong)mem->size); i++; } /* Kernel 'reserved-memory' */ i = 0; printf("\n"); list_for_each(node, &sysmem->kmem_resv_head) { mem = list_entry(node, struct memblock, node); allocated_size += mem->size; printf(" kmem-resv.rgn[%ld].name = \"%s\" %s\n", i, mem->attr.name, mem->orig_base != mem->base ? "<*>" : ""); printf(" .addr = 0x%08lx - 0x%08lx (size: 0x%08lx)\n", (ulong)mem->orig_base, (ulong)(mem->orig_base + mem->size), (ulong)mem->size); i++; } printf("\n framework malloc_r = %3d MiB", SIZE_MB(CONFIG_SYS_MALLOC_LEN)); printf("\n framework malloc_f = %3d KiB\n", SIZE_KB(CONFIG_SYS_MALLOC_F_LEN)); printf("\n allocated.total = 0x%08lx (%ld MiB. %ld KiB)\n", (ulong)allocated_size, SIZE_MB((ulong)allocated_size), SIZE_KB((ulong)allocated_size)); /* LMB core reserved */ printf(" --------------------------------------------------------------------\n"); reserved_size = 0; for (i = 0; i < lmb->reserved.cnt; i++) { reserved_size += lmb->reserved.region[i].size; printf(" LMB.allocated[%ld].addr = 0x%08lx - 0x%08lx (size: 0x%08lx)\n", i, (ulong)lmb->reserved.region[i].base, (ulong)lmb->reserved.region[i].base + (ulong)lmb->reserved.region[i].size, (ulong)lmb->reserved.region[i].size); } printf("\n reserved.core.total = 0x%08lx (%ld MiB. %ld KiB)\n", (ulong)reserved_size, SIZE_MB((ulong)reserved_size), SIZE_KB((ulong)reserved_size)); printf(" --------------------------------------------------------------------\n\n"); } void sysmem_overflow_check(void) { struct sysmem *sysmem = &plat_sysmem; struct list_head *node, *knode; struct memcheck *check; struct memblock *kmem; struct memblock *smem; struct memblock *rmem; int overflow = 0, overlap = 0; if (!sysmem_has_init()) return; #ifdef CONFIG_BIDRAM /* * Check kernel 'reserved-memory' overlap with invisible regions * * Here, only print warning message when overlap with invisible region */ list_for_each(knode, &sysmem->kmem_resv_head) { kmem = list_entry(knode, struct memblock, node); rmem = bidram_reserved_is_overlap(kmem->base, kmem->size); if (rmem) { const char *alias; int i, dump = 1; /* * Ignore the sub region of invisible region. * eg: ramoops of SHM. */ alias = rmem->attr.alias[0]; if (alias && sysmem_is_sub_region(kmem, rmem)) { for (i = 0; i < ALIAS_COUNT_MAX; i++, alias++) { alias = rmem->attr.alias[i]; if (!alias) continue; if (!strncasecmp(kmem->attr.name, alias, strlen(alias))) { dump = 0; break; } } } if (dump) SYSMEM_W("kernel 'reserved-memory' \"%s\"(0x%08lx - 0x%08lx) " "is overlap with [invisible] \"%s\" (0x%08lx - 0x%08lx)\n", kmem->attr.name, (ulong)kmem->base, (ulong)(kmem->base + kmem->size), rmem->attr.name, (ulong)rmem->base, (ulong)(rmem->base + rmem->size)); } } #endif list_for_each(node, &sysmem->allocated_head) { smem = list_entry(node, struct memblock, node); /* * Check kernel 'reserved-memory' overlap with sysmem allocated regions */ list_for_each(knode, &sysmem->kmem_resv_head) { kmem = list_entry(knode, struct memblock, node); if (sysmem_is_overlap(smem->base, smem->size, kmem->base, kmem->size)) { if (smem->attr.flags & F_KMEM_CAN_OVERLAP) continue; overlap = 1; SYSMEM_W("kernel 'reserved-memory' \"%s\"(0x%08lx - 0x%08lx) " "is overlap with \"%s\" (0x%08lx - 0x%08lx)\n", kmem->attr.name, (ulong)kmem->base, (ulong)(kmem->base + kmem->size), smem->attr.name, (ulong)smem->base, (ulong)(smem->base + smem->size)); } } /* * Check sysmem allocated regions overflow. */ if (smem->attr.flags & F_OFC) { check = (struct memcheck *) (smem->base + smem->size - sizeof(*check)); overflow = (check->magic != SYSMEM_MAGIC); } else if (smem->attr.flags & F_HOFC) { check = (struct memcheck *) (smem->base - sizeof(*check)); overflow = (check->magic != SYSMEM_MAGIC); } else { overflow = 0; } if (overflow) { SYSMEM_E("Found there is region overflow!\n"); break; } } if (overflow || overlap) sysmem_dump(); } static int sysmem_add(phys_addr_t base, phys_size_t size) { struct sysmem *sysmem = &plat_sysmem; int ret; if (!size) return -EINVAL; ret = lmb_add(&sysmem->lmb, base, size); if (ret < 0) SYSMEM_E("Failed to add sysmem at 0x%08lx for 0x%08lx size\n", (ulong)base, (ulong)size); return (ret >= 0) ? 0 : ret; } static const char *sysmem_alias2name(const char *name, int *id) { const char *alias; int i, j; int match = 0; for (i = 0; i < MEM_MAX; i++) { /* Pirmary name */ if (mem_attr[i].name && !strcasecmp(mem_attr[i].name, name)) { match = 1; goto finish; } /* Alias name */ alias = mem_attr[i].alias[0]; if (!alias) continue; for (j = 0; j < ALIAS_COUNT_MAX; j++) { alias = mem_attr[i].alias[j]; if (alias && !strcasecmp(alias, name)) { match = 1; goto finish; } } } finish: if (match) { *id = i; return mem_attr[i].name; } return name; } static void *sysmem_alloc_align_base(enum memblk_id id, const char *mem_name, phys_addr_t base, phys_size_t size, ulong align) { struct sysmem *sysmem = &plat_sysmem; struct memblk_attr attr; struct memblock *mem; struct memcheck *check; struct list_head *node; const char *name; phys_addr_t paddr; phys_addr_t alloc_base; phys_size_t alloc_size; phys_addr_t orig_base = base; if (!sysmem_has_init()) goto out; if (id == MEM_BY_NAME || id == MEM_KMEM_RESERVED) { if (!mem_name) { SYSMEM_E("NULL name for alloc sysmem\n"); goto out; } /* Find: name, id and attr by outer mem_name & id */ name = sysmem_alias2name(mem_name, (int *)&id); attr = mem_attr[id]; if (!attr.name) attr.name = strdup(name); /* Always make kernel 'reserved-memory' alloc successfully */ if (id == MEM_KMEM_RESERVED) { struct memblock *mem; mem = malloc(sizeof(*mem)); if (!mem) { SYSMEM_E("No memory for \"%s\" alloc sysmem\n", name); return mem; } attr.flags |= F_KMEM_RESERVED; mem->orig_base = orig_base; mem->base = base; mem->size = size; mem->attr = attr; sysmem->kmem_resv_cnt++; list_add_tail(&mem->node, &sysmem->kmem_resv_head); return (void *)base; } } else if (id > MEM_UNK && id < MEM_MAX) { attr = mem_attr[id]; name = attr.name; /* * Special handle for Android AVB alloc(on any where) * * Fixup base and place right after U-Boot stack, adding a lot * of space(4KB) maybe safer. */ if (attr.flags & F_HIGHEST_MEM) { base = gd->start_addr_sp - CONFIG_SYS_STACK_SIZE - size - 0x1000; /* * The 0x0 address is usually allocated by 32-bit uncompressed * kernel and this alloc action is just a peek. * * Due to LMB core doesn't support alloc at 0x0 address, we have * to alloc the memblk backword a few bytes. * * ARCH_DMA_MINALIGN maybe a good choice. */ } else if (!base) { base += ARCH_DMA_MINALIGN; } else if (base < gd->bd->bi_dram[0].start) { /* * On Rockchip platform: * * So far, we use F_IGNORE_INVISIBLE for uncompress * kernel alloc, and for ARMv8 enabling AArch32 mode, the * ATF is still AArch64 and ocuppies 0~1MB and shmem 1~2M. * So let's ignore the region which overlap with them. */ if (attr.flags & F_IGNORE_INVISIBLE) { base = gd->bd->bi_dram[0].start; } else { SYSMEM_E("Failed to alloc invisible sub region 0x%08lx - 0x%08lx " "of \"%s\" at 0x%08lx - 0x%08lx\n", (ulong)base, (ulong)gd->bd->bi_dram[0].start, name, (ulong)base, (ulong)(base + size)); goto out; } } } else { SYSMEM_E("Unsupport memblk id %d for alloc sysmem\n", id); goto out; } if (!size) { SYSMEM_E("\"%s\" size is 0 for alloc sysmem\n", name); goto out; } /* * Some modules use "sysmem_alloc()" to alloc region for storage * read/write buffer, it should be aligned to cacheline size. eg: AVB. * * Aligned down to cacheline size if not aligned, otherwise the tail * of region maybe overflow. */ if (attr.flags & F_CACHELINE_ALIGN && !IS_ALIGNED(base, ARCH_DMA_MINALIGN)) { base = ALIGN(base, ARCH_DMA_MINALIGN); base -= ARCH_DMA_MINALIGN; } if (base != SYSMEM_ALLOC_ANYWHERE && !IS_ALIGNED(base, 4)) { SYSMEM_E("\"%s\" base=0x%08lx is not 4-byte aligned\n", name, (ulong)base); goto out; } /* Must be sizeof(long) byte aligned */ size = ALIGN(size, sizeof(long)); SYSMEM_D("Enter alloc: \"%s\" 0x%08lx - 0x%08lx\n", name, (ulong)base, (ulong)(base + size)); /* Already allocated ? */ list_for_each(node, &sysmem->allocated_head) { mem = list_entry(node, struct memblock, node); SYSMEM_D("Has allcated: %s, 0x%08lx - 0x%08lx\n", mem->attr.name, (ulong)mem->base, (ulong)(mem->base + mem->size)); if (!strcmp(mem->attr.name, name)) { /* Allow double alloc for same but smaller region */ if (mem->base <= base && mem->size >= size) return (void *)base; SYSMEM_E("Failed to double alloc for existence \"%s\"\n", name); goto out; } else if (sysmem_is_overlap(mem->base, mem->size, base, size)) { SYSMEM_E("\"%s\" (0x%08lx - 0x%08lx) alloc is " "overlap with existence \"%s\" (0x%08lx - " "0x%08lx)\n", name, (ulong)base, (ulong)(base + size), mem->attr.name, (ulong)mem->base, (ulong)(mem->base + mem->size)); goto out; } } /* Add overflow check magic ? */ if (attr.flags & F_OFC) alloc_size = size + sizeof(*check); else alloc_size = size; /* Alloc anywhere ? */ if (base == SYSMEM_ALLOC_ANYWHERE) alloc_base = LMB_ALLOC_ANYWHERE; else alloc_base = base + alloc_size; /* LMB is align down alloc mechanism */ SYSMEM_D("DO alloc... base: 0x%08lx\n", (ulong)alloc_base); paddr = lmb_alloc_base(&sysmem->lmb, alloc_size, align, alloc_base); if (paddr) { if ((paddr == base) || (base == SYSMEM_ALLOC_ANYWHERE)) { mem = malloc(sizeof(*mem)); if (!mem) { SYSMEM_E("No memory for \"%s\" alloc sysmem\n", name); goto out; } /* Record original base for dump */ if (attr.flags & F_HIGHEST_MEM) mem->orig_base = base; else mem->orig_base = orig_base; mem->base = paddr; mem->size = alloc_size; mem->attr = attr; sysmem->allocated_cnt++; list_add_tail(&mem->node, &sysmem->allocated_head); /* Add overflow check magic */ if (mem->attr.flags & F_OFC) { check = (struct memcheck *)(paddr + size); check->magic = SYSMEM_MAGIC; } else if (mem->attr.flags & F_HOFC) { check = (struct memcheck *)(paddr - sizeof(*check)); check->magic = SYSMEM_MAGIC; } } else { SYSMEM_E("Failed to alloc \"%s\" expect at 0x%08lx - 0x%08lx " "but at 0x%08lx - 0x%08lx\n", name, (ulong)base, (ulong)(base + size), (ulong)paddr, (ulong)(paddr + size)); /* Free what we don't want allocated region */ if (lmb_free(&sysmem->lmb, paddr, alloc_size) < 0) SYSMEM_E("Failed to free \"%s\"\n", name); goto out; } } else { SYSMEM_E("Failed to alloc \"%s\" at 0x%08lx - 0x%08lx\n", name, (ulong)base, (ulong)(base + size)); goto out; } SYSMEM_D("Exit alloc: \"%s\", paddr=0x%08lx, size=0x%08lx, align=0x%x, anywhere=%d\n", name, (ulong)paddr, (ulong)size, (u32)align, !base); return (void *)paddr; out: /* * Why: base + sizeof(ulong) ? * It's not a standard way to handle the case: the input base is 0. * Because 0 equals NULL, but we don't want to return NULL when alloc * successfully, so just return a !NULL value is okay. * * When it happens ? * Maybe 32-bit platform would alloc region for uncompress kernel * at 0 address. */ if (base == 0) base = base + sizeof(ulong); return (attr.flags & (F_IGNORE_INVISIBLE | F_NO_FAIL_DUMP)) ? (void *)base : NULL; } void *sysmem_alloc(enum memblk_id id, phys_size_t size) { void *paddr; paddr = sysmem_alloc_align_base(id, NULL, SYSMEM_ALLOC_ANYWHERE, size, ARCH_DMA_MINALIGN); if (!paddr) sysmem_dump(); return paddr; } void *sysmem_alloc_by_name(const char *name, phys_size_t size) { void *paddr; paddr = sysmem_alloc_align_base(MEM_BY_NAME, name, SYSMEM_ALLOC_ANYWHERE, size, ARCH_DMA_MINALIGN); if (!paddr) sysmem_dump(); return paddr; } void *sysmem_alloc_base(enum memblk_id id, phys_addr_t base, phys_size_t size) { void *paddr; paddr = sysmem_alloc_align_base(id, NULL, base, size, SYSMEM_ALLOC_NO_ALIGN); if (!paddr) sysmem_dump(); return paddr; } void *sysmem_alloc_base_by_name(const char *name, phys_addr_t base, phys_size_t size) { void *paddr; paddr = sysmem_alloc_align_base(MEM_BY_NAME, name, base, size, SYSMEM_ALLOC_NO_ALIGN); if (!paddr) sysmem_dump(); return paddr; } void *sysmem_fdt_reserve_alloc_base(const char *name, phys_addr_t base, phys_size_t size) { void *paddr; paddr = sysmem_alloc_align_base(MEM_KMEM_RESERVED, name, base, size, SYSMEM_ALLOC_NO_ALIGN); if (!paddr) sysmem_dump(); return paddr; } ulong sysmem_alloc_temporary_mem(phys_size_t size) { struct sysmem *sysmem = &plat_sysmem; phys_addr_t alloc_base; phys_addr_t paddr; phys_addr_t base; int ret; if (!sysmem_has_init()) return false; base = (gd->start_addr_sp - CONFIG_SYS_STACK_SIZE - 0x2000) - size; /* LMB is align down alloc mechanism */ alloc_base = base + size; paddr = __lmb_alloc_base(&sysmem->lmb, size, SZ_1K, alloc_base); if (paddr) { /* If free failed, return false */ ret = lmb_free(&sysmem->lmb, paddr, size); if (ret < 0) { SYSMEM_E("Can't free at 0x%08lx - 0x%08lx, ret=%d\n", (ulong)paddr, (ulong)(paddr + size), ret); return 0; } } return paddr; } int sysmem_free(phys_addr_t base) { struct sysmem *sysmem = &plat_sysmem; struct memblock *mem; struct list_head *node; int ret, found = 0; if (!sysmem_has_init()) return -ENOSYS; /* Find existence */ list_for_each(node, &sysmem->allocated_head) { mem = list_entry(node, struct memblock, node); if (mem->base == base || mem->orig_base == base) { found = 1; break; } } if (!found) { SYSMEM_E("Failed to free no allocated sysmem at 0x%08lx\n", (ulong)base); return -EINVAL; } ret = lmb_free(&sysmem->lmb, mem->base, mem->size); if (ret >= 0) { SYSMEM_D("Free: \"%s\" 0x%08lx - 0x%08lx\n", mem->attr.name, (ulong)mem->base, (ulong)(mem->base + mem->size)); sysmem->allocated_cnt--; list_del(&mem->node); free(mem); } else { SYSMEM_E("Failed to free \"%s\" at 0x%08lx\n", mem->attr.name, (ulong)base); } return (ret >= 0) ? 0 : ret; } int sysmem_initr(void) { return sysmem_init(); } int sysmem_init(void) { struct sysmem *sysmem = &plat_sysmem; phys_addr_t mem_start; phys_size_t mem_size; int ret; lmb_init(&sysmem->lmb); INIT_LIST_HEAD(&sysmem->allocated_head); INIT_LIST_HEAD(&sysmem->kmem_resv_head); sysmem->allocated_cnt = 0; sysmem->kmem_resv_cnt = 0; if (gd->flags & GD_FLG_RELOC) { sysmem->has_initr = true; } else { SYSMEM_I("init\n"); sysmem->has_initf = true; } /* Add all available system memory */ #ifdef CONFIG_NR_DRAM_BANKS int i; for (i = 0; i < CONFIG_NR_DRAM_BANKS; i++) { if (!gd->bd->bi_dram[i].size) continue; ret = sysmem_add(gd->bd->bi_dram[i].start, gd->bd->bi_dram[i].size); if (ret) { SYSMEM_E("Failed to add sysmem from bi_dram[%d]\n", i); goto fail; } } #else mem_start = env_get_bootm_low(); mem_size = env_get_bootm_size(); ret = sysmem_add(mem_start, mem_size); if (ret) { SYSMEM_E("Failed to add sysmem from bootm_low/size\n"); goto fail; } #endif /* Reserved for board */ ret = board_sysmem_reserve(sysmem); if (ret) { SYSMEM_E("Failed to reserve sysmem for board\n"); goto fail; } /* Reserved for U-boot framework: 'reserve_xxx()' */ mem_start = gd->start_addr_sp; mem_size = gd->ram_top - mem_start; if (!sysmem_alloc_base(MEM_UBOOT, mem_start, mem_size)) { SYSMEM_E("Failed to reserve sysmem for U-Boot framework\n"); ret = -ENOMEM; goto fail; } /* Reserved for U-Boot stack */ mem_start = gd->start_addr_sp - CONFIG_SYS_STACK_SIZE; mem_size = CONFIG_SYS_STACK_SIZE; if (!sysmem_alloc_base(MEM_STACK, mem_start, mem_size)) { SYSMEM_E("Failed to reserve sysmem for stack\n"); ret = -ENOMEM; goto fail; } return 0; fail: if (ret && !(gd->flags & GD_FLG_RELOC)) { sysmem_dump(); SYSMEM_W("Maybe malloc size %d MiB is too large?\n\n", SIZE_MB(CONFIG_SYS_MALLOC_LEN)); } return ret; } __weak int board_sysmem_reserve(struct sysmem *sysmem) { /* please define platform specific board_sysmem_reserve() */ return 0; } static int do_sysmem_dump(cmd_tbl_t *cmdtp, int flag, int argc, char *const argv[]) { sysmem_dump(); return 0; } static int do_sysmem_search(cmd_tbl_t *cmdtp, int flag, int argc, char *const argv[]) { ulong addr, size; if (argc != 2) return CMD_RET_USAGE; size = simple_strtoul(argv[1], NULL, 16); if (!size) return CMD_RET_USAGE; addr = sysmem_alloc_temporary_mem(size); if (!addr) { SYSMEM_I("No available region with size 0x%08lx\n", size); } else { SYSMEM_I("Available region at address: 0x%08lx\n",addr); } env_set_hex("smem_addr", addr); return 0; } U_BOOT_CMD( sysmem_dump, 1, 1, do_sysmem_dump, "Dump sysmem layout", "" ); U_BOOT_CMD( sysmem_search, 2, 1, do_sysmem_search, "Search a available sysmem region", "" );