^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1) // SPDX-License-Identifier: GPL-2.0-only
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2) /**
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3) * Copyright (C) 2008, Creative Technology Ltd. All Rights Reserved.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4) *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 5) * @File ctvmem.c
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 6) *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 7) * @Brief
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 8) * This file contains the implementation of virtual memory management object
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 9) * for card device.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 10) *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 11) * @Author Liu Chun
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 12) * @Date Apr 1 2008
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 13) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 14)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 15) #include "ctvmem.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 16) #include "ctatc.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 17) #include <linux/slab.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 18) #include <linux/mm.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 19) #include <linux/io.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 20) #include <sound/pcm.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 21)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 22) #define CT_PTES_PER_PAGE (CT_PAGE_SIZE / sizeof(void *))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 23) #define CT_ADDRS_PER_PAGE (CT_PTES_PER_PAGE * CT_PAGE_SIZE)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 24)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 25) /* *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 26) * Find or create vm block based on requested @size.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 27) * @size must be page aligned.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 28) * */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 29) static struct ct_vm_block *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 30) get_vm_block(struct ct_vm *vm, unsigned int size, struct ct_atc *atc)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 31) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 32) struct ct_vm_block *block = NULL, *entry;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 33) struct list_head *pos;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 34)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 35) size = CT_PAGE_ALIGN(size);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 36) if (size > vm->size) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 37) dev_err(atc->card->dev,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 38) "Fail! No sufficient device virtual memory space available!\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 39) return NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 40) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 41)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 42) mutex_lock(&vm->lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 43) list_for_each(pos, &vm->unused) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 44) entry = list_entry(pos, struct ct_vm_block, list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 45) if (entry->size >= size)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 46) break; /* found a block that is big enough */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 47) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 48) if (pos == &vm->unused)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 49) goto out;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 50)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 51) if (entry->size == size) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 52) /* Move the vm node from unused list to used list directly */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 53) list_move(&entry->list, &vm->used);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 54) vm->size -= size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 55) block = entry;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 56) goto out;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 57) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 58)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 59) block = kzalloc(sizeof(*block), GFP_KERNEL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 60) if (!block)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 61) goto out;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 62)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 63) block->addr = entry->addr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 64) block->size = size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 65) list_add(&block->list, &vm->used);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 66) entry->addr += size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 67) entry->size -= size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 68) vm->size -= size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 69)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 70) out:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 71) mutex_unlock(&vm->lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 72) return block;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 73) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 74)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 75) static void put_vm_block(struct ct_vm *vm, struct ct_vm_block *block)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 76) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 77) struct ct_vm_block *entry, *pre_ent;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 78) struct list_head *pos, *pre;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 79)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 80) block->size = CT_PAGE_ALIGN(block->size);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 81)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 82) mutex_lock(&vm->lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 83) list_del(&block->list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 84) vm->size += block->size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 85)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 86) list_for_each(pos, &vm->unused) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 87) entry = list_entry(pos, struct ct_vm_block, list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 88) if (entry->addr >= (block->addr + block->size))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 89) break; /* found a position */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 90) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 91) if (pos == &vm->unused) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 92) list_add_tail(&block->list, &vm->unused);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 93) entry = block;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 94) } else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 95) if ((block->addr + block->size) == entry->addr) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 96) entry->addr = block->addr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 97) entry->size += block->size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 98) kfree(block);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 99) } else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 100) __list_add(&block->list, pos->prev, pos);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 101) entry = block;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 102) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 103) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 104)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 105) pos = &entry->list;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 106) pre = pos->prev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 107) while (pre != &vm->unused) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 108) entry = list_entry(pos, struct ct_vm_block, list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 109) pre_ent = list_entry(pre, struct ct_vm_block, list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 110) if ((pre_ent->addr + pre_ent->size) > entry->addr)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 111) break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 112)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 113) pre_ent->size += entry->size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 114) list_del(pos);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 115) kfree(entry);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 116) pos = pre;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 117) pre = pos->prev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 118) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 119) mutex_unlock(&vm->lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 120) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 121)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 122) /* Map host addr (kmalloced/vmalloced) to device logical addr. */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 123) static struct ct_vm_block *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 124) ct_vm_map(struct ct_vm *vm, struct snd_pcm_substream *substream, int size)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 125) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 126) struct ct_vm_block *block;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 127) unsigned int pte_start;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 128) unsigned i, pages;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 129) unsigned long *ptp;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 130) struct ct_atc *atc = snd_pcm_substream_chip(substream);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 131)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 132) block = get_vm_block(vm, size, atc);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 133) if (block == NULL) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 134) dev_err(atc->card->dev,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 135) "No virtual memory block that is big enough to allocate!\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 136) return NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 137) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 138)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 139) ptp = (unsigned long *)vm->ptp[0].area;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 140) pte_start = (block->addr >> CT_PAGE_SHIFT);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 141) pages = block->size >> CT_PAGE_SHIFT;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 142) for (i = 0; i < pages; i++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 143) unsigned long addr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 144) addr = snd_pcm_sgbuf_get_addr(substream, i << CT_PAGE_SHIFT);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 145) ptp[pte_start + i] = addr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 146) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 147)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 148) block->size = size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 149) return block;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 150) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 151)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 152) static void ct_vm_unmap(struct ct_vm *vm, struct ct_vm_block *block)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 153) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 154) /* do unmapping */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 155) put_vm_block(vm, block);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 156) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 157)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 158) /* *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 159) * return the host physical addr of the @index-th device
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 160) * page table page on success, or ~0UL on failure.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 161) * The first returned ~0UL indicates the termination.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 162) * */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 163) static dma_addr_t
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 164) ct_get_ptp_phys(struct ct_vm *vm, int index)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 165) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 166) return (index >= CT_PTP_NUM) ? ~0UL : vm->ptp[index].addr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 167) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 168)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 169) int ct_vm_create(struct ct_vm **rvm, struct pci_dev *pci)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 170) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 171) struct ct_vm *vm;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 172) struct ct_vm_block *block;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 173) int i, err = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 174)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 175) *rvm = NULL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 176)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 177) vm = kzalloc(sizeof(*vm), GFP_KERNEL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 178) if (!vm)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 179) return -ENOMEM;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 180)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 181) mutex_init(&vm->lock);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 182)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 183) /* Allocate page table pages */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 184) for (i = 0; i < CT_PTP_NUM; i++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 185) err = snd_dma_alloc_pages(SNDRV_DMA_TYPE_DEV,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 186) &pci->dev,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 187) PAGE_SIZE, &vm->ptp[i]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 188) if (err < 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 189) break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 190) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 191) if (err < 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 192) /* no page table pages are allocated */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 193) ct_vm_destroy(vm);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 194) return -ENOMEM;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 195) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 196) vm->size = CT_ADDRS_PER_PAGE * i;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 197) vm->map = ct_vm_map;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 198) vm->unmap = ct_vm_unmap;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 199) vm->get_ptp_phys = ct_get_ptp_phys;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 200) INIT_LIST_HEAD(&vm->unused);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 201) INIT_LIST_HEAD(&vm->used);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 202) block = kzalloc(sizeof(*block), GFP_KERNEL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 203) if (NULL != block) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 204) block->addr = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 205) block->size = vm->size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 206) list_add(&block->list, &vm->unused);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 207) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 208)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 209) *rvm = vm;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 210) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 211) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 212)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 213) /* The caller must ensure no mapping pages are being used
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 214) * by hardware before calling this function */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 215) void ct_vm_destroy(struct ct_vm *vm)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 216) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 217) int i;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 218) struct list_head *pos;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 219) struct ct_vm_block *entry;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 220)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 221) /* free used and unused list nodes */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 222) while (!list_empty(&vm->used)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 223) pos = vm->used.next;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 224) list_del(pos);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 225) entry = list_entry(pos, struct ct_vm_block, list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 226) kfree(entry);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 227) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 228) while (!list_empty(&vm->unused)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 229) pos = vm->unused.next;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 230) list_del(pos);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 231) entry = list_entry(pos, struct ct_vm_block, list);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 232) kfree(entry);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 233) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 234)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 235) /* free allocated page table pages */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 236) for (i = 0; i < CT_PTP_NUM; i++)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 237) snd_dma_free_pages(&vm->ptp[i]);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 238)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 239) vm->size = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 240)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 241) kfree(vm);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 242) }