mirror of
				https://git.kernel.org/pub/scm/linux/kernel/git/chenhuacai/linux-loongson
				synced 2025-10-31 22:26:12 +00:00 
			
		
		
		
	 08677214e3
			
		
	
	
		08677214e3
		
	
	
	
	
		
			
			Finally we can use early_res to replace bootmem for x86_64 now. Still can use CONFIG_NO_BOOTMEM to enable it or not. -v2: fix 32bit compiling about MAX_DMA32_PFN -v3: folded bug fix from LKML message below Signed-off-by: Yinghai Lu <yinghai@kernel.org> LKML-Reference: <4B747239.4070907@kernel.org> Signed-off-by: H. Peter Anvin <hpa@zytor.com>
		
			
				
	
	
		
			168 lines
		
	
	
		
			5.0 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			168 lines
		
	
	
		
			5.0 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * Discontiguous memory support, Kanoj Sarcar, SGI, Nov 1999
 | |
|  */
 | |
| #ifndef _LINUX_BOOTMEM_H
 | |
| #define _LINUX_BOOTMEM_H
 | |
| 
 | |
| #include <linux/mmzone.h>
 | |
| #include <asm/dma.h>
 | |
| 
 | |
| /*
 | |
|  *  simple boot-time physical memory area allocator.
 | |
|  */
 | |
| 
 | |
| extern unsigned long max_low_pfn;
 | |
| extern unsigned long min_low_pfn;
 | |
| 
 | |
| /*
 | |
|  * highest page
 | |
|  */
 | |
| extern unsigned long max_pfn;
 | |
| 
 | |
| #ifdef CONFIG_CRASH_DUMP
 | |
| extern unsigned long saved_max_pfn;
 | |
| #endif
 | |
| 
 | |
| #ifndef CONFIG_NO_BOOTMEM
 | |
| /*
 | |
|  * node_bootmem_map is a map pointer - the bits represent all physical 
 | |
|  * memory pages (including holes) on the node.
 | |
|  */
 | |
| typedef struct bootmem_data {
 | |
| 	unsigned long node_min_pfn;
 | |
| 	unsigned long node_low_pfn;
 | |
| 	void *node_bootmem_map;
 | |
| 	unsigned long last_end_off;
 | |
| 	unsigned long hint_idx;
 | |
| 	struct list_head list;
 | |
| } bootmem_data_t;
 | |
| 
 | |
| extern bootmem_data_t bootmem_node_data[];
 | |
| #endif
 | |
| 
 | |
| extern unsigned long bootmem_bootmap_pages(unsigned long);
 | |
| 
 | |
| extern unsigned long init_bootmem_node(pg_data_t *pgdat,
 | |
| 				       unsigned long freepfn,
 | |
| 				       unsigned long startpfn,
 | |
| 				       unsigned long endpfn);
 | |
| extern unsigned long init_bootmem(unsigned long addr, unsigned long memend);
 | |
| 
 | |
| unsigned long free_all_memory_core_early(int nodeid);
 | |
| extern unsigned long free_all_bootmem_node(pg_data_t *pgdat);
 | |
| extern unsigned long free_all_bootmem(void);
 | |
| 
 | |
| extern void free_bootmem_node(pg_data_t *pgdat,
 | |
| 			      unsigned long addr,
 | |
| 			      unsigned long size);
 | |
| extern void free_bootmem(unsigned long addr, unsigned long size);
 | |
| extern void free_bootmem_late(unsigned long addr, unsigned long size);
 | |
| 
 | |
| /*
 | |
|  * Flags for reserve_bootmem (also if CONFIG_HAVE_ARCH_BOOTMEM_NODE,
 | |
|  * the architecture-specific code should honor this).
 | |
|  *
 | |
|  * If flags is 0, then the return value is always 0 (success). If
 | |
|  * flags contains BOOTMEM_EXCLUSIVE, then -EBUSY is returned if the
 | |
|  * memory already was reserved.
 | |
|  */
 | |
| #define BOOTMEM_DEFAULT		0
 | |
| #define BOOTMEM_EXCLUSIVE	(1<<0)
 | |
| 
 | |
| extern int reserve_bootmem(unsigned long addr,
 | |
| 			   unsigned long size,
 | |
| 			   int flags);
 | |
| extern int reserve_bootmem_node(pg_data_t *pgdat,
 | |
| 				unsigned long physaddr,
 | |
| 				unsigned long size,
 | |
| 				int flags);
 | |
| 
 | |
| extern void *__alloc_bootmem(unsigned long size,
 | |
| 			     unsigned long align,
 | |
| 			     unsigned long goal);
 | |
| extern void *__alloc_bootmem_nopanic(unsigned long size,
 | |
| 				     unsigned long align,
 | |
| 				     unsigned long goal);
 | |
| extern void *__alloc_bootmem_node(pg_data_t *pgdat,
 | |
| 				  unsigned long size,
 | |
| 				  unsigned long align,
 | |
| 				  unsigned long goal);
 | |
| void *__alloc_bootmem_node_high(pg_data_t *pgdat,
 | |
| 				  unsigned long size,
 | |
| 				  unsigned long align,
 | |
| 				  unsigned long goal);
 | |
| extern void *__alloc_bootmem_node_nopanic(pg_data_t *pgdat,
 | |
| 				  unsigned long size,
 | |
| 				  unsigned long align,
 | |
| 				  unsigned long goal);
 | |
| extern void *__alloc_bootmem_low(unsigned long size,
 | |
| 				 unsigned long align,
 | |
| 				 unsigned long goal);
 | |
| extern void *__alloc_bootmem_low_node(pg_data_t *pgdat,
 | |
| 				      unsigned long size,
 | |
| 				      unsigned long align,
 | |
| 				      unsigned long goal);
 | |
| 
 | |
| #define alloc_bootmem(x) \
 | |
| 	__alloc_bootmem(x, SMP_CACHE_BYTES, __pa(MAX_DMA_ADDRESS))
 | |
| #define alloc_bootmem_nopanic(x) \
 | |
| 	__alloc_bootmem_nopanic(x, SMP_CACHE_BYTES, __pa(MAX_DMA_ADDRESS))
 | |
| #define alloc_bootmem_pages(x) \
 | |
| 	__alloc_bootmem(x, PAGE_SIZE, __pa(MAX_DMA_ADDRESS))
 | |
| #define alloc_bootmem_pages_nopanic(x) \
 | |
| 	__alloc_bootmem_nopanic(x, PAGE_SIZE, __pa(MAX_DMA_ADDRESS))
 | |
| #define alloc_bootmem_node(pgdat, x) \
 | |
| 	__alloc_bootmem_node(pgdat, x, SMP_CACHE_BYTES, __pa(MAX_DMA_ADDRESS))
 | |
| #define alloc_bootmem_pages_node(pgdat, x) \
 | |
| 	__alloc_bootmem_node(pgdat, x, PAGE_SIZE, __pa(MAX_DMA_ADDRESS))
 | |
| #define alloc_bootmem_pages_node_nopanic(pgdat, x) \
 | |
| 	__alloc_bootmem_node_nopanic(pgdat, x, PAGE_SIZE, __pa(MAX_DMA_ADDRESS))
 | |
| 
 | |
| #define alloc_bootmem_low(x) \
 | |
| 	__alloc_bootmem_low(x, SMP_CACHE_BYTES, 0)
 | |
| #define alloc_bootmem_low_pages(x) \
 | |
| 	__alloc_bootmem_low(x, PAGE_SIZE, 0)
 | |
| #define alloc_bootmem_low_pages_node(pgdat, x) \
 | |
| 	__alloc_bootmem_low_node(pgdat, x, PAGE_SIZE, 0)
 | |
| 
 | |
| extern int reserve_bootmem_generic(unsigned long addr, unsigned long size,
 | |
| 				   int flags);
 | |
| 
 | |
| extern void *alloc_bootmem_section(unsigned long size,
 | |
| 				   unsigned long section_nr);
 | |
| 
 | |
| #ifdef CONFIG_HAVE_ARCH_ALLOC_REMAP
 | |
| extern void *alloc_remap(int nid, unsigned long size);
 | |
| #else
 | |
| static inline void *alloc_remap(int nid, unsigned long size)
 | |
| {
 | |
| 	return NULL;
 | |
| }
 | |
| #endif /* CONFIG_HAVE_ARCH_ALLOC_REMAP */
 | |
| 
 | |
| extern void *alloc_large_system_hash(const char *tablename,
 | |
| 				     unsigned long bucketsize,
 | |
| 				     unsigned long numentries,
 | |
| 				     int scale,
 | |
| 				     int flags,
 | |
| 				     unsigned int *_hash_shift,
 | |
| 				     unsigned int *_hash_mask,
 | |
| 				     unsigned long limit);
 | |
| 
 | |
| #define HASH_EARLY	0x00000001	/* Allocating during early boot? */
 | |
| #define HASH_SMALL	0x00000002	/* sub-page allocation allowed, min
 | |
| 					 * shift passed via *_hash_shift */
 | |
| 
 | |
| /* Only NUMA needs hash distribution. 64bit NUMA architectures have
 | |
|  * sufficient vmalloc space.
 | |
|  */
 | |
| #if defined(CONFIG_NUMA) && defined(CONFIG_64BIT)
 | |
| #define HASHDIST_DEFAULT 1
 | |
| #else
 | |
| #define HASHDIST_DEFAULT 0
 | |
| #endif
 | |
| extern int hashdist;		/* Distribute hashes across NUMA nodes? */
 | |
| 
 | |
| 
 | |
| #endif /* _LINUX_BOOTMEM_H */
 |