2
* This file is subject to the terms and conditions of the GNU General Public
3
* License. See the file "COPYING" in the main directory of this archive
6
* Copyright (C) 2000, 05 by Ralf Baechle (ralf@linux-mips.org)
7
* Copyright (C) 2000 by Silicon Graphics, Inc.
8
* Copyright (C) 2004 by Christoph Hellwig
10
* On SGI IP27 the ARC memory configuration data is completly bogus but
11
* alternate easier to use mechanisms are available.
13
#include <linux/init.h>
14
#include <linux/kernel.h>
16
#include <linux/mmzone.h>
17
#include <linux/module.h>
18
#include <linux/nodemask.h>
19
#include <linux/swap.h>
20
#include <linux/bootmem.h>
21
#include <linux/pfn.h>
22
#include <linux/highmem.h>
24
#include <asm/pgalloc.h>
25
#include <asm/sections.h>
27
#include <asm/sn/arch.h>
28
#include <asm/sn/hub.h>
29
#include <asm/sn/klconfig.h>
30
#include <asm/sn/sn_private.h>
33
#define SLOT_PFNSHIFT (SLOT_SHIFT - PAGE_SHIFT)
34
#define PFN_NASIDSHFT (NASID_SHFT - PAGE_SHIFT)
36
struct node_data *__node_data[MAX_COMPACT_NODES];
38
EXPORT_SYMBOL(__node_data);
42
static int is_fine_dirmode(void)
44
return (((LOCAL_HUB_L(NI_STATUS_REV_ID) & NSRI_REGIONSIZE_MASK)
45
>> NSRI_REGIONSIZE_SHFT) & REGIONSIZE_FINE);
48
static hubreg_t get_region(cnodeid_t cnode)
51
return COMPACT_TO_NASID_NODEID(cnode) >> NASID_TO_FINEREG_SHFT;
53
return COMPACT_TO_NASID_NODEID(cnode) >> NASID_TO_COARSEREG_SHFT;
56
static hubreg_t region_mask;
58
static void gen_region_mask(hubreg_t *region_mask)
63
for_each_online_node(cnode) {
64
(*region_mask) |= 1ULL << get_region(cnode);
68
#define rou_rflag rou_flags
70
static int router_distance;
72
static void router_recurse(klrou_t *router_a, klrou_t *router_b, int depth)
78
if (router_a->rou_rflag == 1)
81
if (depth >= router_distance)
84
router_a->rou_rflag = 1;
86
for (port = 1; port <= MAX_ROUTER_PORTS; port++) {
87
if (router_a->rou_port[port].port_nasid == INVALID_NASID)
90
brd = (lboard_t *)NODE_OFFSET_TO_K0(
91
router_a->rou_port[port].port_nasid,
92
router_a->rou_port[port].port_offset);
94
if (brd->brd_type == KLTYPE_ROUTER) {
95
router = (klrou_t *)NODE_OFFSET_TO_K0(NASID_GET(brd), brd->brd_compts[0]);
96
if (router == router_b) {
97
if (depth < router_distance)
98
router_distance = depth;
101
router_recurse(router, router_b, depth + 1);
105
router_a->rou_rflag = 0;
108
unsigned char __node_distances[MAX_COMPACT_NODES][MAX_COMPACT_NODES];
110
static int __init compute_node_distance(nasid_t nasid_a, nasid_t nasid_b)
112
klrou_t *router, *router_a = NULL, *router_b = NULL;
113
lboard_t *brd, *dest_brd;
118
/* Figure out which routers nodes in question are connected to */
119
for_each_online_node(cnode) {
120
nasid = COMPACT_TO_NASID_NODEID(cnode);
122
if (nasid == -1) continue;
124
brd = find_lboard_class((lboard_t *)KL_CONFIG_INFO(nasid),
131
if (brd->brd_flags & DUPLICATE_BOARD)
134
router = (klrou_t *)NODE_OFFSET_TO_K0(NASID_GET(brd), brd->brd_compts[0]);
135
router->rou_rflag = 0;
137
for (port = 1; port <= MAX_ROUTER_PORTS; port++) {
138
if (router->rou_port[port].port_nasid == INVALID_NASID)
141
dest_brd = (lboard_t *)NODE_OFFSET_TO_K0(
142
router->rou_port[port].port_nasid,
143
router->rou_port[port].port_offset);
145
if (dest_brd->brd_type == KLTYPE_IP27) {
146
if (dest_brd->brd_nasid == nasid_a)
148
if (dest_brd->brd_nasid == nasid_b)
153
} while ((brd = find_lboard_class(KLCF_NEXT(brd), KLTYPE_ROUTER)));
156
if (router_a == NULL) {
157
printk("node_distance: router_a NULL\n");
160
if (router_b == NULL) {
161
printk("node_distance: router_b NULL\n");
165
if (nasid_a == nasid_b)
168
if (router_a == router_b)
171
router_distance = 100;
172
router_recurse(router_a, router_b, 2);
174
return router_distance;
177
static void __init init_topology_matrix(void)
179
nasid_t nasid, nasid2;
182
for (row = 0; row < MAX_COMPACT_NODES; row++)
183
for (col = 0; col < MAX_COMPACT_NODES; col++)
184
__node_distances[row][col] = -1;
186
for_each_online_node(row) {
187
nasid = COMPACT_TO_NASID_NODEID(row);
188
for_each_online_node(col) {
189
nasid2 = COMPACT_TO_NASID_NODEID(col);
190
__node_distances[row][col] =
191
compute_node_distance(nasid, nasid2);
196
static void __init dump_topology(void)
200
lboard_t *brd, *dest_brd;
206
printk("************** Topology ********************\n");
209
for_each_online_node(col)
210
printk("%02d ", col);
212
for_each_online_node(row) {
213
printk("%02d ", row);
214
for_each_online_node(col)
215
printk("%2d ", node_distance(row, col));
219
for_each_online_node(cnode) {
220
nasid = COMPACT_TO_NASID_NODEID(cnode);
222
if (nasid == -1) continue;
224
brd = find_lboard_class((lboard_t *)KL_CONFIG_INFO(nasid),
231
if (brd->brd_flags & DUPLICATE_BOARD)
233
printk("Router %d:", router_num);
236
router = (klrou_t *)NODE_OFFSET_TO_K0(NASID_GET(brd), brd->brd_compts[0]);
238
for (port = 1; port <= MAX_ROUTER_PORTS; port++) {
239
if (router->rou_port[port].port_nasid == INVALID_NASID)
242
dest_brd = (lboard_t *)NODE_OFFSET_TO_K0(
243
router->rou_port[port].port_nasid,
244
router->rou_port[port].port_offset);
246
if (dest_brd->brd_type == KLTYPE_IP27)
247
printk(" %d", dest_brd->brd_nasid);
248
if (dest_brd->brd_type == KLTYPE_ROUTER)
253
} while ( (brd = find_lboard_class(KLCF_NEXT(brd), KLTYPE_ROUTER)) );
257
static pfn_t __init slot_getbasepfn(cnodeid_t cnode, int slot)
259
nasid_t nasid = COMPACT_TO_NASID_NODEID(cnode);
261
return ((pfn_t)nasid << PFN_NASIDSHFT) | (slot << SLOT_PFNSHIFT);
264
static pfn_t __init slot_psize_compute(cnodeid_t node, int slot)
271
nasid = COMPACT_TO_NASID_NODEID(node);
272
/* Find the node board */
273
brd = find_lboard((lboard_t *)KL_CONFIG_INFO(nasid), KLTYPE_IP27);
277
/* Get the memory bank structure */
278
banks = (klmembnk_t *) find_first_component(brd, KLSTRUCT_MEMBNK);
282
/* Size in _Megabytes_ */
283
size = (unsigned long)banks->membnk_bnksz[slot/4];
285
/* hack for 128 dimm banks */
288
size <<= 20; /* size in bytes */
289
return(size >> PAGE_SHIFT);
295
return size >> PAGE_SHIFT;
299
static void __init mlreset(void)
303
master_nasid = get_nasid();
304
fine_mode = is_fine_dirmode();
307
* Probe for all CPUs - this creates the cpumask and sets up the
308
* mapping tables. We need to do this as early as possible.
314
init_topology_matrix();
317
gen_region_mask(®ion_mask);
319
setup_replication_mask();
322
* Set all nodes' calias sizes to 8k
324
for_each_online_node(i) {
327
nasid = COMPACT_TO_NASID_NODEID(i);
330
* Always have node 0 in the region mask, otherwise
331
* CALIAS accesses get exceptions since the hub
332
* thinks it is a node 0 address.
334
REMOTE_HUB_S(nasid, PI_REGION_PRESENT, (region_mask | 1));
335
#ifdef CONFIG_REPLICATE_EXHANDLERS
336
REMOTE_HUB_S(nasid, PI_CALIAS_SIZE, PI_CALIAS_SIZE_8K);
338
REMOTE_HUB_S(nasid, PI_CALIAS_SIZE, PI_CALIAS_SIZE_0);
343
* Set up all hubs to have a big window pointing at
344
* widget 0. Memory mode, widget 0, offset 0
346
REMOTE_HUB_S(nasid, IIO_ITTE(SWIN0_BIGWIN),
347
((HUB_PIO_MAP_TO_MEM << IIO_ITTE_IOSP_SHIFT) |
348
(0 << IIO_ITTE_WIDGET_SHIFT)));
353
static void __init szmem(void)
355
pfn_t slot_psize, slot0sz = 0, nodebytes; /* Hack to detect problem configs */
361
for_each_online_node(node) {
363
for (slot = 0; slot < MAX_MEM_SLOTS; slot++) {
364
slot_psize = slot_psize_compute(node, slot);
366
slot0sz = slot_psize;
368
* We need to refine the hack when we have replicated
371
nodebytes += (1LL << SLOT_SHIFT);
376
if ((nodebytes >> PAGE_SHIFT) * (sizeof(struct page)) >
377
(slot0sz << PAGE_SHIFT)) {
378
printk("Ignoring slot %d onwards on node %d\n",
380
slot = MAX_MEM_SLOTS;
383
num_physpages += slot_psize;
384
add_active_range(node, slot_getbasepfn(node, slot),
385
slot_getbasepfn(node, slot) + slot_psize);
390
static void __init node_mem_init(cnodeid_t node)
392
pfn_t slot_firstpfn = slot_getbasepfn(node, 0);
393
pfn_t slot_freepfn = node_getfirstfree(node);
394
unsigned long bootmap_size;
395
pfn_t start_pfn, end_pfn;
397
get_pfn_range_for_nid(node, &start_pfn, &end_pfn);
400
* Allocate the node data structures on the node first.
402
__node_data[node] = __va(slot_freepfn << PAGE_SHIFT);
404
NODE_DATA(node)->bdata = &bootmem_node_data[node];
405
NODE_DATA(node)->node_start_pfn = start_pfn;
406
NODE_DATA(node)->node_spanned_pages = end_pfn - start_pfn;
408
cpus_clear(hub_data(node)->h_cpus);
410
slot_freepfn += PFN_UP(sizeof(struct pglist_data) +
411
sizeof(struct hub_data));
413
bootmap_size = init_bootmem_node(NODE_DATA(node), slot_freepfn,
415
free_bootmem_with_active_regions(node, end_pfn);
416
reserve_bootmem_node(NODE_DATA(node), slot_firstpfn << PAGE_SHIFT,
417
((slot_freepfn - slot_firstpfn) << PAGE_SHIFT) + bootmap_size,
419
sparse_memory_present_with_active_regions(node);
423
* A node with nothing. We use it to avoid any special casing in
426
static struct node_data null_node = {
428
.h_cpus = CPU_MASK_NONE
433
* Currently, the intranode memory hole support assumes that each slot
434
* contains at least 32 MBytes of memory. We assume all bootmem data
435
* fits on the first slot.
437
void __init prom_meminit(void)
444
for (node = 0; node < MAX_COMPACT_NODES; node++) {
445
if (node_online(node)) {
449
__node_data[node] = &null_node;
453
void __init prom_free_prom_memory(void)
455
/* We got nothing to free here ... */
458
extern unsigned long setup_zero_pages(void);
460
void __init paging_init(void)
462
unsigned long zones_size[MAX_NR_ZONES] = {0, };
467
for_each_online_node(node) {
468
pfn_t start_pfn, end_pfn;
470
get_pfn_range_for_nid(node, &start_pfn, &end_pfn);
472
if (end_pfn > max_low_pfn)
473
max_low_pfn = end_pfn;
475
zones_size[ZONE_NORMAL] = max_low_pfn;
476
free_area_init_nodes(zones_size);
479
void __init mem_init(void)
481
unsigned long codesize, datasize, initsize, tmp;
484
high_memory = (void *) __va(num_physpages << PAGE_SHIFT);
486
for_each_online_node(node) {
488
* This will free up the bootmem, ie, slot 0 memory.
490
totalram_pages += free_all_bootmem_node(NODE_DATA(node));
493
totalram_pages -= setup_zero_pages(); /* This comes from node 0 */
495
codesize = (unsigned long) &_etext - (unsigned long) &_text;
496
datasize = (unsigned long) &_edata - (unsigned long) &_etext;
497
initsize = (unsigned long) &__init_end - (unsigned long) &__init_begin;
499
tmp = nr_free_pages();
500
printk(KERN_INFO "Memory: %luk/%luk available (%ldk kernel code, "
501
"%ldk reserved, %ldk data, %ldk init, %ldk highmem)\n",
502
tmp << (PAGE_SHIFT-10),
503
num_physpages << (PAGE_SHIFT-10),
505
(num_physpages - tmp) << (PAGE_SHIFT-10),
508
totalhigh_pages << (PAGE_SHIFT-10));