^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1) // SPDX-License-Identifier: GPL-2.0-only
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3) * Copyright 2001-2003 SuSE Labs.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4) * Distributed under the GNU public license, v2.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 5) *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 6) * This is a GART driver for the AMD Opteron/Athlon64 on-CPU northbridge.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 7) * It also includes support for the AMD 8151 AGP bridge,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 8) * although it doesn't actually do much, as all the real
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 9) * work is done in the northbridge(s).
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 10) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 11)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 12) #include <linux/module.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 13) #include <linux/pci.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 14) #include <linux/init.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 15) #include <linux/agp_backend.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 16) #include <linux/mmzone.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 17) #include <asm/page.h> /* PAGE_SIZE */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 18) #include <asm/e820/api.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 19) #include <asm/amd_nb.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 20) #include <asm/gart.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 21) #include "agp.h"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 22)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 23) /* NVIDIA K8 registers */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 24) #define NVIDIA_X86_64_0_APBASE 0x10
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 25) #define NVIDIA_X86_64_1_APBASE1 0x50
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 26) #define NVIDIA_X86_64_1_APLIMIT1 0x54
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 27) #define NVIDIA_X86_64_1_APSIZE 0xa8
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 28) #define NVIDIA_X86_64_1_APBASE2 0xd8
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 29) #define NVIDIA_X86_64_1_APLIMIT2 0xdc
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 30)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 31) /* ULi K8 registers */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 32) #define ULI_X86_64_BASE_ADDR 0x10
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 33) #define ULI_X86_64_HTT_FEA_REG 0x50
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 34) #define ULI_X86_64_ENU_SCR_REG 0x54
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 35)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 36) static struct resource *aperture_resource;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 37) static bool __initdata agp_try_unsupported = 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 38) static int agp_bridges_found;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 39)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 40) static void amd64_tlbflush(struct agp_memory *temp)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 41) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 42) amd_flush_garts();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 43) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 44)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 45) static int amd64_insert_memory(struct agp_memory *mem, off_t pg_start, int type)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 46) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 47) int i, j, num_entries;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 48) long long tmp;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 49) int mask_type;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 50) struct agp_bridge_data *bridge = mem->bridge;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 51) u32 pte;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 52)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 53) num_entries = agp_num_entries();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 54)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 55) if (type != mem->type)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 56) return -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 57) mask_type = bridge->driver->agp_type_to_mask_type(bridge, type);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 58) if (mask_type != 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 59) return -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 60)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 61)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 62) /* Make sure we can fit the range in the gatt table. */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 63) /* FIXME: could wrap */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 64) if (((unsigned long)pg_start + mem->page_count) > num_entries)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 65) return -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 66)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 67) j = pg_start;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 68)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 69) /* gatt table should be empty. */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 70) while (j < (pg_start + mem->page_count)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 71) if (!PGE_EMPTY(agp_bridge, readl(agp_bridge->gatt_table+j)))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 72) return -EBUSY;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 73) j++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 74) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 75)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 76) if (!mem->is_flushed) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 77) global_cache_flush();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 78) mem->is_flushed = true;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 79) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 80)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 81) for (i = 0, j = pg_start; i < mem->page_count; i++, j++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 82) tmp = agp_bridge->driver->mask_memory(agp_bridge,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 83) page_to_phys(mem->pages[i]),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 84) mask_type);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 85)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 86) BUG_ON(tmp & 0xffffff0000000ffcULL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 87) pte = (tmp & 0x000000ff00000000ULL) >> 28;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 88) pte |=(tmp & 0x00000000fffff000ULL);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 89) pte |= GPTE_VALID | GPTE_COHERENT;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 90)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 91) writel(pte, agp_bridge->gatt_table+j);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 92) readl(agp_bridge->gatt_table+j); /* PCI Posting. */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 93) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 94) amd64_tlbflush(mem);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 95) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 96) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 97)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 98) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 99) * This hack alters the order element according
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 100) * to the size of a long. It sucks. I totally disown this, even
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 101) * though it does appear to work for the most part.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 102) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 103) static struct aper_size_info_32 amd64_aperture_sizes[7] =
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 104) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 105) {32, 8192, 3+(sizeof(long)/8), 0 },
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 106) {64, 16384, 4+(sizeof(long)/8), 1<<1 },
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 107) {128, 32768, 5+(sizeof(long)/8), 1<<2 },
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 108) {256, 65536, 6+(sizeof(long)/8), 1<<1 | 1<<2 },
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 109) {512, 131072, 7+(sizeof(long)/8), 1<<3 },
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 110) {1024, 262144, 8+(sizeof(long)/8), 1<<1 | 1<<3},
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 111) {2048, 524288, 9+(sizeof(long)/8), 1<<2 | 1<<3}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 112) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 113)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 114)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 115) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 116) * Get the current Aperture size from the x86-64.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 117) * Note, that there may be multiple x86-64's, but we just return
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 118) * the value from the first one we find. The set_size functions
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 119) * keep the rest coherent anyway. Or at least should do.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 120) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 121) static int amd64_fetch_size(void)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 122) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 123) struct pci_dev *dev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 124) int i;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 125) u32 temp;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 126) struct aper_size_info_32 *values;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 127)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 128) dev = node_to_amd_nb(0)->misc;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 129) if (dev==NULL)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 130) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 131)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 132) pci_read_config_dword(dev, AMD64_GARTAPERTURECTL, &temp);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 133) temp = (temp & 0xe);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 134) values = A_SIZE_32(amd64_aperture_sizes);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 135)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 136) for (i = 0; i < agp_bridge->driver->num_aperture_sizes; i++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 137) if (temp == values[i].size_value) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 138) agp_bridge->previous_size =
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 139) agp_bridge->current_size = (void *) (values + i);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 140)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 141) agp_bridge->aperture_size_idx = i;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 142) return values[i].size;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 143) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 144) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 145) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 146) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 147)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 148) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 149) * In a multiprocessor x86-64 system, this function gets
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 150) * called once for each CPU.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 151) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 152) static u64 amd64_configure(struct pci_dev *hammer, u64 gatt_table)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 153) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 154) u64 aperturebase;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 155) u32 tmp;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 156) u64 aper_base;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 157)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 158) /* Address to map to */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 159) pci_read_config_dword(hammer, AMD64_GARTAPERTUREBASE, &tmp);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 160) aperturebase = (u64)tmp << 25;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 161) aper_base = (aperturebase & PCI_BASE_ADDRESS_MEM_MASK);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 162)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 163) enable_gart_translation(hammer, gatt_table);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 164)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 165) return aper_base;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 166) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 167)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 168)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 169) static const struct aper_size_info_32 amd_8151_sizes[7] =
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 170) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 171) {2048, 524288, 9, 0x00000000 }, /* 0 0 0 0 0 0 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 172) {1024, 262144, 8, 0x00000400 }, /* 1 0 0 0 0 0 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 173) {512, 131072, 7, 0x00000600 }, /* 1 1 0 0 0 0 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 174) {256, 65536, 6, 0x00000700 }, /* 1 1 1 0 0 0 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 175) {128, 32768, 5, 0x00000720 }, /* 1 1 1 1 0 0 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 176) {64, 16384, 4, 0x00000730 }, /* 1 1 1 1 1 0 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 177) {32, 8192, 3, 0x00000738 } /* 1 1 1 1 1 1 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 178) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 179)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 180) static int amd_8151_configure(void)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 181) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 182) unsigned long gatt_bus = virt_to_phys(agp_bridge->gatt_table_real);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 183) int i;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 184)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 185) if (!amd_nb_has_feature(AMD_NB_GART))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 186) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 187)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 188) /* Configure AGP regs in each x86-64 host bridge. */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 189) for (i = 0; i < amd_nb_num(); i++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 190) agp_bridge->gart_bus_addr =
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 191) amd64_configure(node_to_amd_nb(i)->misc, gatt_bus);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 192) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 193) amd_flush_garts();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 194) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 195) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 196)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 197)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 198) static void amd64_cleanup(void)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 199) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 200) u32 tmp;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 201) int i;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 202)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 203) if (!amd_nb_has_feature(AMD_NB_GART))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 204) return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 205)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 206) for (i = 0; i < amd_nb_num(); i++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 207) struct pci_dev *dev = node_to_amd_nb(i)->misc;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 208) /* disable gart translation */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 209) pci_read_config_dword(dev, AMD64_GARTAPERTURECTL, &tmp);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 210) tmp &= ~GARTEN;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 211) pci_write_config_dword(dev, AMD64_GARTAPERTURECTL, tmp);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 212) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 213) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 214)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 215)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 216) static const struct agp_bridge_driver amd_8151_driver = {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 217) .owner = THIS_MODULE,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 218) .aperture_sizes = amd_8151_sizes,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 219) .size_type = U32_APER_SIZE,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 220) .num_aperture_sizes = 7,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 221) .needs_scratch_page = true,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 222) .configure = amd_8151_configure,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 223) .fetch_size = amd64_fetch_size,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 224) .cleanup = amd64_cleanup,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 225) .tlb_flush = amd64_tlbflush,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 226) .mask_memory = agp_generic_mask_memory,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 227) .masks = NULL,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 228) .agp_enable = agp_generic_enable,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 229) .cache_flush = global_cache_flush,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 230) .create_gatt_table = agp_generic_create_gatt_table,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 231) .free_gatt_table = agp_generic_free_gatt_table,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 232) .insert_memory = amd64_insert_memory,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 233) .remove_memory = agp_generic_remove_memory,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 234) .alloc_by_type = agp_generic_alloc_by_type,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 235) .free_by_type = agp_generic_free_by_type,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 236) .agp_alloc_page = agp_generic_alloc_page,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 237) .agp_alloc_pages = agp_generic_alloc_pages,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 238) .agp_destroy_page = agp_generic_destroy_page,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 239) .agp_destroy_pages = agp_generic_destroy_pages,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 240) .agp_type_to_mask_type = agp_generic_type_to_mask_type,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 241) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 242)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 243) /* Some basic sanity checks for the aperture. */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 244) static int agp_aperture_valid(u64 aper, u32 size)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 245) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 246) if (!aperture_valid(aper, size, 32*1024*1024))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 247) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 248)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 249) /* Request the Aperture. This catches cases when someone else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 250) already put a mapping in there - happens with some very broken BIOS
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 251)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 252) Maybe better to use pci_assign_resource/pci_enable_device instead
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 253) trusting the bridges? */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 254) if (!aperture_resource &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 255) !(aperture_resource = request_mem_region(aper, size, "aperture"))) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 256) printk(KERN_ERR PFX "Aperture conflicts with PCI mapping.\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 257) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 258) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 259) return 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 260) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 261)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 262) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 263) * W*s centric BIOS sometimes only set up the aperture in the AGP
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 264) * bridge, not the northbridge. On AMD64 this is handled early
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 265) * in aperture.c, but when IOMMU is not enabled or we run
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 266) * on a 32bit kernel this needs to be redone.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 267) * Unfortunately it is impossible to fix the aperture here because it's too late
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 268) * to allocate that much memory. But at least error out cleanly instead of
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 269) * crashing.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 270) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 271) static int fix_northbridge(struct pci_dev *nb, struct pci_dev *agp, u16 cap)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 272) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 273) u64 aper, nb_aper;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 274) int order = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 275) u32 nb_order, nb_base;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 276) u16 apsize;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 277)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 278) pci_read_config_dword(nb, AMD64_GARTAPERTURECTL, &nb_order);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 279) nb_order = (nb_order >> 1) & 7;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 280) pci_read_config_dword(nb, AMD64_GARTAPERTUREBASE, &nb_base);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 281) nb_aper = (u64)nb_base << 25;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 282)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 283) /* Northbridge seems to contain crap. Try the AGP bridge. */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 284)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 285) pci_read_config_word(agp, cap+0x14, &apsize);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 286) if (apsize == 0xffff) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 287) if (agp_aperture_valid(nb_aper, (32*1024*1024)<<nb_order))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 288) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 289) return -1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 290) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 291)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 292) apsize &= 0xfff;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 293) /* Some BIOS use weird encodings not in the AGPv3 table. */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 294) if (apsize & 0xff)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 295) apsize |= 0xf00;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 296) order = 7 - hweight16(apsize);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 297)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 298) aper = pci_bus_address(agp, AGP_APERTURE_BAR);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 299)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 300) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 301) * On some sick chips APSIZE is 0. This means it wants 4G
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 302) * so let double check that order, and lets trust the AMD NB settings
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 303) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 304) if (order >=0 && aper + (32ULL<<(20 + order)) > 0x100000000ULL) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 305) dev_info(&agp->dev, "aperture size %u MB is not right, using settings from NB\n",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 306) 32 << order);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 307) order = nb_order;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 308) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 309)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 310) if (nb_order >= order) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 311) if (agp_aperture_valid(nb_aper, (32*1024*1024)<<nb_order))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 312) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 313) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 314)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 315) dev_info(&agp->dev, "aperture from AGP @ %Lx size %u MB\n",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 316) aper, 32 << order);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 317) if (order < 0 || !agp_aperture_valid(aper, (32*1024*1024)<<order))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 318) return -1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 319)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 320) gart_set_size_and_enable(nb, order);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 321) pci_write_config_dword(nb, AMD64_GARTAPERTUREBASE, aper >> 25);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 322)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 323) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 324) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 325)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 326) static int cache_nbs(struct pci_dev *pdev, u32 cap_ptr)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 327) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 328) int i;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 329)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 330) if (amd_cache_northbridges() < 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 331) return -ENODEV;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 332)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 333) if (!amd_nb_has_feature(AMD_NB_GART))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 334) return -ENODEV;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 335)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 336) i = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 337) for (i = 0; i < amd_nb_num(); i++) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 338) struct pci_dev *dev = node_to_amd_nb(i)->misc;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 339) if (fix_northbridge(dev, pdev, cap_ptr) < 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 340) dev_err(&dev->dev, "no usable aperture found\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 341) #ifdef __x86_64__
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 342) /* should port this to i386 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 343) dev_err(&dev->dev, "consider rebooting with iommu=memaper=2 to get a good aperture\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 344) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 345) return -1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 346) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 347) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 348) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 349) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 350)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 351) /* Handle AMD 8151 quirks */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 352) static void amd8151_init(struct pci_dev *pdev, struct agp_bridge_data *bridge)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 353) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 354) char *revstring;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 355)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 356) switch (pdev->revision) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 357) case 0x01: revstring="A0"; break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 358) case 0x02: revstring="A1"; break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 359) case 0x11: revstring="B0"; break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 360) case 0x12: revstring="B1"; break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 361) case 0x13: revstring="B2"; break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 362) case 0x14: revstring="B3"; break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 363) default: revstring="??"; break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 364) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 365)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 366) dev_info(&pdev->dev, "AMD 8151 AGP Bridge rev %s\n", revstring);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 367)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 368) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 369) * Work around errata.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 370) * Chips before B2 stepping incorrectly reporting v3.5
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 371) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 372) if (pdev->revision < 0x13) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 373) dev_info(&pdev->dev, "correcting AGP revision (reports 3.5, is really 3.0)\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 374) bridge->major_version = 3;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 375) bridge->minor_version = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 376) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 377) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 378)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 379)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 380) static const struct aper_size_info_32 uli_sizes[7] =
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 381) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 382) {256, 65536, 6, 10},
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 383) {128, 32768, 5, 9},
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 384) {64, 16384, 4, 8},
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 385) {32, 8192, 3, 7},
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 386) {16, 4096, 2, 6},
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 387) {8, 2048, 1, 4},
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 388) {4, 1024, 0, 3}
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 389) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 390) static int uli_agp_init(struct pci_dev *pdev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 391) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 392) u32 httfea,baseaddr,enuscr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 393) struct pci_dev *dev1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 394) int i, ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 395) unsigned size = amd64_fetch_size();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 396)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 397) dev_info(&pdev->dev, "setting up ULi AGP\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 398) dev1 = pci_get_slot (pdev->bus,PCI_DEVFN(0,0));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 399) if (dev1 == NULL) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 400) dev_info(&pdev->dev, "can't find ULi secondary device\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 401) return -ENODEV;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 402) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 403)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 404) for (i = 0; i < ARRAY_SIZE(uli_sizes); i++)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 405) if (uli_sizes[i].size == size)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 406) break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 407)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 408) if (i == ARRAY_SIZE(uli_sizes)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 409) dev_info(&pdev->dev, "no ULi size found for %d\n", size);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 410) ret = -ENODEV;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 411) goto put;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 412) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 413)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 414) /* shadow x86-64 registers into ULi registers */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 415) pci_read_config_dword (node_to_amd_nb(0)->misc, AMD64_GARTAPERTUREBASE,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 416) &httfea);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 417)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 418) /* if x86-64 aperture base is beyond 4G, exit here */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 419) if ((httfea & 0x7fff) >> (32 - 25)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 420) ret = -ENODEV;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 421) goto put;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 422) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 423)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 424) httfea = (httfea& 0x7fff) << 25;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 425)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 426) pci_read_config_dword(pdev, ULI_X86_64_BASE_ADDR, &baseaddr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 427) baseaddr&= ~PCI_BASE_ADDRESS_MEM_MASK;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 428) baseaddr|= httfea;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 429) pci_write_config_dword(pdev, ULI_X86_64_BASE_ADDR, baseaddr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 430)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 431) enuscr= httfea+ (size * 1024 * 1024) - 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 432) pci_write_config_dword(dev1, ULI_X86_64_HTT_FEA_REG, httfea);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 433) pci_write_config_dword(dev1, ULI_X86_64_ENU_SCR_REG, enuscr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 434) ret = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 435) put:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 436) pci_dev_put(dev1);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 437) return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 438) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 439)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 440)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 441) static const struct aper_size_info_32 nforce3_sizes[5] =
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 442) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 443) {512, 131072, 7, 0x00000000 },
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 444) {256, 65536, 6, 0x00000008 },
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 445) {128, 32768, 5, 0x0000000C },
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 446) {64, 16384, 4, 0x0000000E },
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 447) {32, 8192, 3, 0x0000000F }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 448) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 449)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 450) /* Handle shadow device of the Nvidia NForce3 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 451) /* CHECK-ME original 2.4 version set up some IORRs. Check if that is needed. */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 452) static int nforce3_agp_init(struct pci_dev *pdev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 453) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 454) u32 tmp, apbase, apbar, aplimit;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 455) struct pci_dev *dev1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 456) int i, ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 457) unsigned size = amd64_fetch_size();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 458)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 459) dev_info(&pdev->dev, "setting up Nforce3 AGP\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 460)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 461) dev1 = pci_get_slot(pdev->bus, PCI_DEVFN(11, 0));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 462) if (dev1 == NULL) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 463) dev_info(&pdev->dev, "can't find Nforce3 secondary device\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 464) return -ENODEV;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 465) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 466)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 467) for (i = 0; i < ARRAY_SIZE(nforce3_sizes); i++)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 468) if (nforce3_sizes[i].size == size)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 469) break;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 470)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 471) if (i == ARRAY_SIZE(nforce3_sizes)) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 472) dev_info(&pdev->dev, "no NForce3 size found for %d\n", size);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 473) ret = -ENODEV;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 474) goto put;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 475) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 476)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 477) pci_read_config_dword(dev1, NVIDIA_X86_64_1_APSIZE, &tmp);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 478) tmp &= ~(0xf);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 479) tmp |= nforce3_sizes[i].size_value;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 480) pci_write_config_dword(dev1, NVIDIA_X86_64_1_APSIZE, tmp);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 481)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 482) /* shadow x86-64 registers into NVIDIA registers */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 483) pci_read_config_dword (node_to_amd_nb(0)->misc, AMD64_GARTAPERTUREBASE,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 484) &apbase);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 485)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 486) /* if x86-64 aperture base is beyond 4G, exit here */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 487) if ( (apbase & 0x7fff) >> (32 - 25) ) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 488) dev_info(&pdev->dev, "aperture base > 4G\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 489) ret = -ENODEV;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 490) goto put;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 491) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 492)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 493) apbase = (apbase & 0x7fff) << 25;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 494)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 495) pci_read_config_dword(pdev, NVIDIA_X86_64_0_APBASE, &apbar);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 496) apbar &= ~PCI_BASE_ADDRESS_MEM_MASK;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 497) apbar |= apbase;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 498) pci_write_config_dword(pdev, NVIDIA_X86_64_0_APBASE, apbar);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 499)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 500) aplimit = apbase + (size * 1024 * 1024) - 1;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 501) pci_write_config_dword(dev1, NVIDIA_X86_64_1_APBASE1, apbase);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 502) pci_write_config_dword(dev1, NVIDIA_X86_64_1_APLIMIT1, aplimit);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 503) pci_write_config_dword(dev1, NVIDIA_X86_64_1_APBASE2, apbase);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 504) pci_write_config_dword(dev1, NVIDIA_X86_64_1_APLIMIT2, aplimit);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 505)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 506) ret = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 507) put:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 508) pci_dev_put(dev1);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 509)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 510) return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 511) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 512)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 513) static int agp_amd64_probe(struct pci_dev *pdev,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 514) const struct pci_device_id *ent)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 515) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 516) struct agp_bridge_data *bridge;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 517) u8 cap_ptr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 518) int err;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 519)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 520) /* The Highlander principle */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 521) if (agp_bridges_found)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 522) return -ENODEV;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 523)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 524) cap_ptr = pci_find_capability(pdev, PCI_CAP_ID_AGP);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 525) if (!cap_ptr)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 526) return -ENODEV;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 527)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 528) /* Could check for AGPv3 here */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 529)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 530) bridge = agp_alloc_bridge();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 531) if (!bridge)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 532) return -ENOMEM;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 533)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 534) if (pdev->vendor == PCI_VENDOR_ID_AMD &&
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 535) pdev->device == PCI_DEVICE_ID_AMD_8151_0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 536) amd8151_init(pdev, bridge);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 537) } else {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 538) dev_info(&pdev->dev, "AGP bridge [%04x/%04x]\n",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 539) pdev->vendor, pdev->device);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 540) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 541)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 542) bridge->driver = &amd_8151_driver;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 543) bridge->dev = pdev;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 544) bridge->capndx = cap_ptr;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 545)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 546) /* Fill in the mode register */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 547) pci_read_config_dword(pdev, bridge->capndx+PCI_AGP_STATUS, &bridge->mode);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 548)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 549) if (cache_nbs(pdev, cap_ptr) == -1) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 550) agp_put_bridge(bridge);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 551) return -ENODEV;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 552) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 553)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 554) if (pdev->vendor == PCI_VENDOR_ID_NVIDIA) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 555) int ret = nforce3_agp_init(pdev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 556) if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 557) agp_put_bridge(bridge);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 558) return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 559) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 560) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 561)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 562) if (pdev->vendor == PCI_VENDOR_ID_AL) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 563) int ret = uli_agp_init(pdev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 564) if (ret) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 565) agp_put_bridge(bridge);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 566) return ret;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 567) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 568) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 569)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 570) pci_set_drvdata(pdev, bridge);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 571) err = agp_add_bridge(bridge);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 572) if (err < 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 573) return err;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 574)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 575) agp_bridges_found++;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 576) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 577) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 578)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 579) static void agp_amd64_remove(struct pci_dev *pdev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 580) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 581) struct agp_bridge_data *bridge = pci_get_drvdata(pdev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 582)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 583) release_mem_region(virt_to_phys(bridge->gatt_table_real),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 584) amd64_aperture_sizes[bridge->aperture_size_idx].size);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 585) agp_remove_bridge(bridge);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 586) agp_put_bridge(bridge);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 587)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 588) agp_bridges_found--;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 589) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 590)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 591) #ifdef CONFIG_PM
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 592)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 593) static int agp_amd64_suspend(struct pci_dev *pdev, pm_message_t state)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 594) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 595) pci_save_state(pdev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 596) pci_set_power_state(pdev, pci_choose_state(pdev, state));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 597)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 598) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 599) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 600)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 601) static int agp_amd64_resume(struct pci_dev *pdev)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 602) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 603) pci_set_power_state(pdev, PCI_D0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 604) pci_restore_state(pdev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 605)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 606) if (pdev->vendor == PCI_VENDOR_ID_NVIDIA)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 607) nforce3_agp_init(pdev);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 608)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 609) return amd_8151_configure();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 610) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 611)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 612) #endif /* CONFIG_PM */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 613)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 614) static const struct pci_device_id agp_amd64_pci_table[] = {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 615) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 616) .class = (PCI_CLASS_BRIDGE_HOST << 8),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 617) .class_mask = ~0,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 618) .vendor = PCI_VENDOR_ID_AMD,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 619) .device = PCI_DEVICE_ID_AMD_8151_0,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 620) .subvendor = PCI_ANY_ID,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 621) .subdevice = PCI_ANY_ID,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 622) },
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 623) /* ULi M1689 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 624) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 625) .class = (PCI_CLASS_BRIDGE_HOST << 8),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 626) .class_mask = ~0,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 627) .vendor = PCI_VENDOR_ID_AL,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 628) .device = PCI_DEVICE_ID_AL_M1689,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 629) .subvendor = PCI_ANY_ID,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 630) .subdevice = PCI_ANY_ID,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 631) },
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 632) /* VIA K8T800Pro */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 633) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 634) .class = (PCI_CLASS_BRIDGE_HOST << 8),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 635) .class_mask = ~0,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 636) .vendor = PCI_VENDOR_ID_VIA,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 637) .device = PCI_DEVICE_ID_VIA_K8T800PRO_0,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 638) .subvendor = PCI_ANY_ID,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 639) .subdevice = PCI_ANY_ID,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 640) },
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 641) /* VIA K8T800 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 642) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 643) .class = (PCI_CLASS_BRIDGE_HOST << 8),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 644) .class_mask = ~0,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 645) .vendor = PCI_VENDOR_ID_VIA,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 646) .device = PCI_DEVICE_ID_VIA_8385_0,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 647) .subvendor = PCI_ANY_ID,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 648) .subdevice = PCI_ANY_ID,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 649) },
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 650) /* VIA K8M800 / K8N800 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 651) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 652) .class = (PCI_CLASS_BRIDGE_HOST << 8),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 653) .class_mask = ~0,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 654) .vendor = PCI_VENDOR_ID_VIA,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 655) .device = PCI_DEVICE_ID_VIA_8380_0,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 656) .subvendor = PCI_ANY_ID,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 657) .subdevice = PCI_ANY_ID,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 658) },
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 659) /* VIA K8M890 / K8N890 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 660) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 661) .class = (PCI_CLASS_BRIDGE_HOST << 8),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 662) .class_mask = ~0,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 663) .vendor = PCI_VENDOR_ID_VIA,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 664) .device = PCI_DEVICE_ID_VIA_VT3336,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 665) .subvendor = PCI_ANY_ID,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 666) .subdevice = PCI_ANY_ID,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 667) },
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 668) /* VIA K8T890 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 669) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 670) .class = (PCI_CLASS_BRIDGE_HOST << 8),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 671) .class_mask = ~0,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 672) .vendor = PCI_VENDOR_ID_VIA,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 673) .device = PCI_DEVICE_ID_VIA_3238_0,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 674) .subvendor = PCI_ANY_ID,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 675) .subdevice = PCI_ANY_ID,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 676) },
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 677) /* VIA K8T800/K8M800/K8N800 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 678) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 679) .class = (PCI_CLASS_BRIDGE_HOST << 8),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 680) .class_mask = ~0,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 681) .vendor = PCI_VENDOR_ID_VIA,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 682) .device = PCI_DEVICE_ID_VIA_838X_1,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 683) .subvendor = PCI_ANY_ID,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 684) .subdevice = PCI_ANY_ID,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 685) },
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 686) /* NForce3 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 687) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 688) .class = (PCI_CLASS_BRIDGE_HOST << 8),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 689) .class_mask = ~0,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 690) .vendor = PCI_VENDOR_ID_NVIDIA,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 691) .device = PCI_DEVICE_ID_NVIDIA_NFORCE3,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 692) .subvendor = PCI_ANY_ID,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 693) .subdevice = PCI_ANY_ID,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 694) },
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 695) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 696) .class = (PCI_CLASS_BRIDGE_HOST << 8),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 697) .class_mask = ~0,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 698) .vendor = PCI_VENDOR_ID_NVIDIA,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 699) .device = PCI_DEVICE_ID_NVIDIA_NFORCE3S,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 700) .subvendor = PCI_ANY_ID,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 701) .subdevice = PCI_ANY_ID,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 702) },
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 703) /* SIS 755 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 704) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 705) .class = (PCI_CLASS_BRIDGE_HOST << 8),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 706) .class_mask = ~0,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 707) .vendor = PCI_VENDOR_ID_SI,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 708) .device = PCI_DEVICE_ID_SI_755,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 709) .subvendor = PCI_ANY_ID,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 710) .subdevice = PCI_ANY_ID,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 711) },
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 712) /* SIS 760 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 713) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 714) .class = (PCI_CLASS_BRIDGE_HOST << 8),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 715) .class_mask = ~0,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 716) .vendor = PCI_VENDOR_ID_SI,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 717) .device = PCI_DEVICE_ID_SI_760,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 718) .subvendor = PCI_ANY_ID,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 719) .subdevice = PCI_ANY_ID,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 720) },
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 721) /* ALI/ULI M1695 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 722) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 723) .class = (PCI_CLASS_BRIDGE_HOST << 8),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 724) .class_mask = ~0,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 725) .vendor = PCI_VENDOR_ID_AL,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 726) .device = 0x1695,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 727) .subvendor = PCI_ANY_ID,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 728) .subdevice = PCI_ANY_ID,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 729) },
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 730)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 731) { }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 732) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 733)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 734) MODULE_DEVICE_TABLE(pci, agp_amd64_pci_table);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 735)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 736) static const struct pci_device_id agp_amd64_pci_promisc_table[] = {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 737) { PCI_DEVICE_CLASS(0, 0) },
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 738) { }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 739) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 740)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 741) static struct pci_driver agp_amd64_pci_driver = {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 742) .name = "agpgart-amd64",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 743) .id_table = agp_amd64_pci_table,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 744) .probe = agp_amd64_probe,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 745) .remove = agp_amd64_remove,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 746) #ifdef CONFIG_PM
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 747) .suspend = agp_amd64_suspend,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 748) .resume = agp_amd64_resume,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 749) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 750) };
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 751)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 752)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 753) /* Not static due to IOMMU code calling it early. */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 754) int __init agp_amd64_init(void)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 755) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 756) int err = 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 757)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 758) if (agp_off)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 759) return -EINVAL;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 760)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 761) err = pci_register_driver(&agp_amd64_pci_driver);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 762) if (err < 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 763) return err;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 764)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 765) if (agp_bridges_found == 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 766) if (!agp_try_unsupported && !agp_try_unsupported_boot) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 767) printk(KERN_INFO PFX "No supported AGP bridge found.\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 768) #ifdef MODULE
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 769) printk(KERN_INFO PFX "You can try agp_try_unsupported=1\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 770) #else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 771) printk(KERN_INFO PFX "You can boot with agp=try_unsupported\n");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 772) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 773) pci_unregister_driver(&agp_amd64_pci_driver);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 774) return -ENODEV;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 775) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 776)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 777) /* First check that we have at least one AMD64 NB */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 778) if (!amd_nb_num()) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 779) pci_unregister_driver(&agp_amd64_pci_driver);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 780) return -ENODEV;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 781) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 782)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 783) /* Look for any AGP bridge */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 784) agp_amd64_pci_driver.id_table = agp_amd64_pci_promisc_table;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 785) err = driver_attach(&agp_amd64_pci_driver.driver);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 786) if (err == 0 && agp_bridges_found == 0) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 787) pci_unregister_driver(&agp_amd64_pci_driver);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 788) err = -ENODEV;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 789) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 790) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 791) return err;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 792) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 793)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 794) static int __init agp_amd64_mod_init(void)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 795) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 796) #ifndef MODULE
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 797) if (gart_iommu_aperture)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 798) return agp_bridges_found ? 0 : -ENODEV;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 799) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 800) return agp_amd64_init();
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 801) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 802)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 803) static void __exit agp_amd64_cleanup(void)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 804) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 805) #ifndef MODULE
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 806) if (gart_iommu_aperture)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 807) return;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 808) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 809) if (aperture_resource)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 810) release_resource(aperture_resource);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 811) pci_unregister_driver(&agp_amd64_pci_driver);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 812) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 813)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 814) module_init(agp_amd64_mod_init);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 815) module_exit(agp_amd64_cleanup);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 816)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 817) MODULE_AUTHOR("Dave Jones, Andi Kleen");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 818) module_param(agp_try_unsupported, bool, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 819) MODULE_LICENSE("GPL");