^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1) // SPDX-License-Identifier: GPL-2.0
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3) * linux/arch/m68k/sun3/sun3dvma.c
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4) *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 5) * Copyright (C) 2000 Sam Creasey
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 6) *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 7) * Contains common routines for sun3/sun3x DVMA management.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 8) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 9)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 10) #include <linux/memblock.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 11) #include <linux/init.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 12) #include <linux/module.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 13) #include <linux/kernel.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 14) #include <linux/gfp.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 15) #include <linux/mm.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 16) #include <linux/list.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 17)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 18) #include <asm/page.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 19) #include <asm/dvma.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 20)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 21) #undef DVMA_DEBUG
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 22)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 23) #ifdef CONFIG_SUN3X
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 24) extern void dvma_unmap_iommu(unsigned long baddr, int len);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 25) #else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 26) static inline void dvma_unmap_iommu(unsigned long a, int b)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 27) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 28) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 29) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 30)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 31) #ifdef CONFIG_SUN3
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 32) extern void sun3_dvma_init(void);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 33) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 34)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 35) static unsigned long *iommu_use;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 36)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 37) #define dvma_index(baddr) ((baddr - DVMA_START) >> DVMA_PAGE_SHIFT)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 38)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 39) #define dvma_entry_use(baddr) (iommu_use[dvma_index(baddr)])
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 40)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 41) struct hole {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 42) unsigned long start;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 43) unsigned long end;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 44) unsigned long size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 45) struct list_head list;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 46) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 47)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 48) static struct list_head hole_list;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 49) static struct list_head hole_cache;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 50) static struct hole initholes[64];
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 51)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 52) #ifdef DVMA_DEBUG
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 53)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 54) static unsigned long dvma_allocs;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 55) static unsigned long dvma_frees;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 56) static unsigned long long dvma_alloc_bytes;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 57) static unsigned long long dvma_free_bytes;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 58)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 59) static void print_use(void)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 60) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 61)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 62) int i;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 63) int j = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 64)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 65) pr_info("dvma entry usage:\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 66)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 67) for(i = 0; i < IOMMU_TOTAL_ENTRIES; i++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 68) if(!iommu_use[i])
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 69) continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 70)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 71) j++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 72)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 73) pr_info("dvma entry: %08x len %08lx\n",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 74) (i << DVMA_PAGE_SHIFT) + DVMA_START, iommu_use[i]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 75) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 76)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 77) pr_info("%d entries in use total\n", j);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 78)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 79) pr_info("allocation/free calls: %lu/%lu\n", dvma_allocs, dvma_frees);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 80) pr_info("allocation/free bytes: %Lx/%Lx\n", dvma_alloc_bytes,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 81) dvma_free_bytes);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 82) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 83)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 84) static void print_holes(struct list_head *holes)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 85) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 86)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 87) struct list_head *cur;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 88) struct hole *hole;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 89)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 90) pr_info("listing dvma holes\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 91) list_for_each(cur, holes) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 92) hole = list_entry(cur, struct hole, list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 93)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 94) if((hole->start == 0) && (hole->end == 0) && (hole->size == 0))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 95) continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 96)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 97) pr_info("hole: start %08lx end %08lx size %08lx\n",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 98) hole->start, hole->end, hole->size);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 99) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 100)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 101) pr_info("end of hole listing...\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 102) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 103) #endif /* DVMA_DEBUG */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 104)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 105) static inline int refill(void)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 106) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 107)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 108) struct hole *hole;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 109) struct hole *prev = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 110) struct list_head *cur;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 111) int ret = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 112)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 113) list_for_each(cur, &hole_list) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 114) hole = list_entry(cur, struct hole, list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 115)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 116) if(!prev) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 117) prev = hole;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 118) continue;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 119) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 120)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 121) if(hole->end == prev->start) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 122) hole->size += prev->size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 123) hole->end = prev->end;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 124) list_move(&(prev->list), &hole_cache);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 125) ret++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 126) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 127)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 128) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 129)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 130) return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 131) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 132)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 133) static inline struct hole *rmcache(void)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 134) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 135) struct hole *ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 136)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 137) if(list_empty(&hole_cache)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 138) if(!refill()) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 139) pr_crit("out of dvma hole cache!\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 140) BUG();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 141) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 142) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 143)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 144) ret = list_entry(hole_cache.next, struct hole, list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 145) list_del(&(ret->list));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 146)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 147) return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 148)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 149) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 150)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 151) static inline unsigned long get_baddr(int len, unsigned long align)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 152) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 153)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 154) struct list_head *cur;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 155) struct hole *hole;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 156)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 157) if(list_empty(&hole_list)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 158) #ifdef DVMA_DEBUG
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 159) pr_crit("out of dvma holes! (printing hole cache)\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 160) print_holes(&hole_cache);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 161) print_use();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 162) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 163) BUG();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 164) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 165)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 166) list_for_each(cur, &hole_list) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 167) unsigned long newlen;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 168)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 169) hole = list_entry(cur, struct hole, list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 170)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 171) if(align > DVMA_PAGE_SIZE)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 172) newlen = len + ((hole->end - len) & (align-1));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 173) else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 174) newlen = len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 175)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 176) if(hole->size > newlen) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 177) hole->end -= newlen;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 178) hole->size -= newlen;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 179) dvma_entry_use(hole->end) = newlen;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 180) #ifdef DVMA_DEBUG
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 181) dvma_allocs++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 182) dvma_alloc_bytes += newlen;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 183) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 184) return hole->end;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 185) } else if(hole->size == newlen) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 186) list_move(&(hole->list), &hole_cache);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 187) dvma_entry_use(hole->start) = newlen;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 188) #ifdef DVMA_DEBUG
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 189) dvma_allocs++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 190) dvma_alloc_bytes += newlen;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 191) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 192) return hole->start;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 193) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 194)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 195) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 196)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 197) pr_crit("unable to find dvma hole!\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 198) BUG();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 199) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 200) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 201)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 202) static inline int free_baddr(unsigned long baddr)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 203) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 204)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 205) unsigned long len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 206) struct hole *hole;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 207) struct list_head *cur;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 208) unsigned long orig_baddr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 209)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 210) orig_baddr = baddr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 211) len = dvma_entry_use(baddr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 212) dvma_entry_use(baddr) = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 213) baddr &= DVMA_PAGE_MASK;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 214) dvma_unmap_iommu(baddr, len);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 215)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 216) #ifdef DVMA_DEBUG
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 217) dvma_frees++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 218) dvma_free_bytes += len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 219) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 220)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 221) list_for_each(cur, &hole_list) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 222) hole = list_entry(cur, struct hole, list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 223)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 224) if(hole->end == baddr) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 225) hole->end += len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 226) hole->size += len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 227) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 228) } else if(hole->start == (baddr + len)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 229) hole->start = baddr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 230) hole->size += len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 231) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 232) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 233)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 234) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 235)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 236) hole = rmcache();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 237)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 238) hole->start = baddr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 239) hole->end = baddr + len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 240) hole->size = len;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 241)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 242) // list_add_tail(&(hole->list), cur);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 243) list_add(&(hole->list), cur);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 244)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 245) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 246)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 247) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 248)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 249) void __init dvma_init(void)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 250) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 251)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 252) struct hole *hole;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 253) int i;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 254)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 255) INIT_LIST_HEAD(&hole_list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 256) INIT_LIST_HEAD(&hole_cache);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 257)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 258) /* prepare the hole cache */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 259) for(i = 0; i < 64; i++)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 260) list_add(&(initholes[i].list), &hole_cache);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 261)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 262) hole = rmcache();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 263) hole->start = DVMA_START;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 264) hole->end = DVMA_END;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 265) hole->size = DVMA_SIZE;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 266)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 267) list_add(&(hole->list), &hole_list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 268)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 269) iommu_use = memblock_alloc(IOMMU_TOTAL_ENTRIES * sizeof(unsigned long),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 270) SMP_CACHE_BYTES);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 271) if (!iommu_use)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 272) panic("%s: Failed to allocate %zu bytes\n", __func__,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 273) IOMMU_TOTAL_ENTRIES * sizeof(unsigned long));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 274)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 275) dvma_unmap_iommu(DVMA_START, DVMA_SIZE);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 276)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 277) #ifdef CONFIG_SUN3
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 278) sun3_dvma_init();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 279) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 280)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 281) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 282)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 283) unsigned long dvma_map_align(unsigned long kaddr, int len, int align)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 284) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 285)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 286) unsigned long baddr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 287) unsigned long off;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 288)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 289) if(!len)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 290) len = 0x800;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 291)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 292) if(!kaddr || !len) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 293) // pr_err("error: kaddr %lx len %x\n", kaddr, len);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 294) // *(int *)4 = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 295) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 296) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 297)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 298) pr_debug("dvma_map request %08x bytes from %08lx\n", len, kaddr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 299) off = kaddr & ~DVMA_PAGE_MASK;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 300) kaddr &= PAGE_MASK;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 301) len += off;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 302) len = ((len + (DVMA_PAGE_SIZE-1)) & DVMA_PAGE_MASK);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 303)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 304) if(align == 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 305) align = DVMA_PAGE_SIZE;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 306) else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 307) align = ((align + (DVMA_PAGE_SIZE-1)) & DVMA_PAGE_MASK);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 308)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 309) baddr = get_baddr(len, align);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 310) // pr_info("using baddr %lx\n", baddr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 311)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 312) if(!dvma_map_iommu(kaddr, baddr, len))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 313) return (baddr + off);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 314)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 315) pr_crit("dvma_map failed kaddr %lx baddr %lx len %x\n", kaddr, baddr,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 316) len);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 317) BUG();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 318) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 319) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 320) EXPORT_SYMBOL(dvma_map_align);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 321)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 322) void dvma_unmap(void *baddr)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 323) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 324) unsigned long addr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 325)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 326) addr = (unsigned long)baddr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 327) /* check if this is a vme mapping */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 328) if(!(addr & 0x00f00000))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 329) addr |= 0xf00000;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 330)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 331) free_baddr(addr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 332)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 333) return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 334)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 335) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 336) EXPORT_SYMBOL(dvma_unmap);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 337)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 338) void *dvma_malloc_align(unsigned long len, unsigned long align)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 339) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 340) unsigned long kaddr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 341) unsigned long baddr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 342) unsigned long vaddr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 343)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 344) if(!len)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 345) return NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 346)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 347) pr_debug("dvma_malloc request %lx bytes\n", len);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 348) len = ((len + (DVMA_PAGE_SIZE-1)) & DVMA_PAGE_MASK);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 349)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 350) if((kaddr = __get_free_pages(GFP_ATOMIC, get_order(len))) == 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 351) return NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 352)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 353) if((baddr = (unsigned long)dvma_map_align(kaddr, len, align)) == 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 354) free_pages(kaddr, get_order(len));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 355) return NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 356) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 357)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 358) vaddr = dvma_btov(baddr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 359)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 360) if(dvma_map_cpu(kaddr, vaddr, len) < 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 361) dvma_unmap((void *)baddr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 362) free_pages(kaddr, get_order(len));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 363) return NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 364) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 365)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 366) pr_debug("mapped %08lx bytes %08lx kern -> %08lx bus\n", len, kaddr,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 367) baddr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 368)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 369) return (void *)vaddr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 370)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 371) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 372) EXPORT_SYMBOL(dvma_malloc_align);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 373)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 374) void dvma_free(void *vaddr)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 375) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 376)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 377) return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 378)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 379) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 380) EXPORT_SYMBOL(dvma_free);