mirror of
				https://git.kernel.org/pub/scm/linux/kernel/git/chenhuacai/linux-loongson
				synced 2025-10-31 12:09:34 +00:00 
			
		
		
		
	 2d4dc890b5
			
		
	
	
		2d4dc890b5
		
	
	
	
	
		
			
			Mtdblock driver doesn't call flush_dcache_page for pages in request. So, this causes problems on architectures where the icache doesn't fill from the dcache or with dcache aliases. The patch fixes this. The ARCH_IMPLEMENTS_FLUSH_DCACHE_PAGE symbol was introduced to avoid pointless empty cache-thrashing loops on architectures for which flush_dcache_page() is a no-op. Every architecture was provided with this flush pages on architectires where ARCH_IMPLEMENTS_FLUSH_DCACHE_PAGE is equal 1 or do nothing otherwise. See "fix mtd_blkdevs problem with caches on some architectures" discussion on LKML for more information. Signed-off-by: Ilya Loginov <isloginov@gmail.com> Cc: Ingo Molnar <mingo@elte.hu> Cc: David Woodhouse <dwmw2@infradead.org> Cc: Peter Horton <phorton@bitbox.co.uk> Cc: "Ed L. Cashin" <ecashin@coraid.com> Signed-off-by: Jens Axboe <jens.axboe@oracle.com>
		
			
				
	
	
		
			118 lines
		
	
	
		
			4.3 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			118 lines
		
	
	
		
			4.3 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /* MN10300 Cache flushing
 | |
|  *
 | |
|  * Copyright (C) 2007 Red Hat, Inc. All Rights Reserved.
 | |
|  * Written by David Howells (dhowells@redhat.com)
 | |
|  *
 | |
|  * This program is free software; you can redistribute it and/or
 | |
|  * modify it under the terms of the GNU General Public Licence
 | |
|  * as published by the Free Software Foundation; either version
 | |
|  * 2 of the Licence, or (at your option) any later version.
 | |
|  */
 | |
| #ifndef _ASM_CACHEFLUSH_H
 | |
| #define _ASM_CACHEFLUSH_H
 | |
| 
 | |
| #ifndef __ASSEMBLY__
 | |
| 
 | |
| /* Keep includes the same across arches.  */
 | |
| #include <linux/mm.h>
 | |
| 
 | |
| /*
 | |
|  * virtually-indexed cache management (our cache is physically indexed)
 | |
|  */
 | |
| #define flush_cache_all()			do {} while (0)
 | |
| #define flush_cache_mm(mm)			do {} while (0)
 | |
| #define flush_cache_dup_mm(mm)			do {} while (0)
 | |
| #define flush_cache_range(mm, start, end)	do {} while (0)
 | |
| #define flush_cache_page(vma, vmaddr, pfn)	do {} while (0)
 | |
| #define flush_cache_vmap(start, end)		do {} while (0)
 | |
| #define flush_cache_vunmap(start, end)		do {} while (0)
 | |
| #define ARCH_IMPLEMENTS_FLUSH_DCACHE_PAGE 0
 | |
| #define flush_dcache_page(page)			do {} while (0)
 | |
| #define flush_dcache_mmap_lock(mapping)		do {} while (0)
 | |
| #define flush_dcache_mmap_unlock(mapping)	do {} while (0)
 | |
| 
 | |
| /*
 | |
|  * physically-indexed cache management
 | |
|  */
 | |
| #ifndef CONFIG_MN10300_CACHE_DISABLED
 | |
| 
 | |
| extern void flush_icache_range(unsigned long start, unsigned long end);
 | |
| extern void flush_icache_page(struct vm_area_struct *vma, struct page *pg);
 | |
| 
 | |
| #else
 | |
| 
 | |
| #define flush_icache_range(start, end)		do {} while (0)
 | |
| #define flush_icache_page(vma, pg)		do {} while (0)
 | |
| 
 | |
| #endif
 | |
| 
 | |
| #define flush_icache_user_range(vma, pg, adr, len) \
 | |
| 	flush_icache_range(adr, adr + len)
 | |
| 
 | |
| #define copy_to_user_page(vma, page, vaddr, dst, src, len) \
 | |
| 	do {					\
 | |
| 		memcpy(dst, src, len);		\
 | |
| 		flush_icache_page(vma, page);	\
 | |
| 	} while (0)
 | |
| 
 | |
| #define copy_from_user_page(vma, page, vaddr, dst, src, len) \
 | |
| 	memcpy(dst, src, len)
 | |
| 
 | |
| /*
 | |
|  * primitive routines
 | |
|  */
 | |
| #ifndef CONFIG_MN10300_CACHE_DISABLED
 | |
| extern void mn10300_icache_inv(void);
 | |
| extern void mn10300_dcache_inv(void);
 | |
| extern void mn10300_dcache_inv_page(unsigned start);
 | |
| extern void mn10300_dcache_inv_range(unsigned start, unsigned end);
 | |
| extern void mn10300_dcache_inv_range2(unsigned start, unsigned size);
 | |
| #ifdef CONFIG_MN10300_CACHE_WBACK
 | |
| extern void mn10300_dcache_flush(void);
 | |
| extern void mn10300_dcache_flush_page(unsigned start);
 | |
| extern void mn10300_dcache_flush_range(unsigned start, unsigned end);
 | |
| extern void mn10300_dcache_flush_range2(unsigned start, unsigned size);
 | |
| extern void mn10300_dcache_flush_inv(void);
 | |
| extern void mn10300_dcache_flush_inv_page(unsigned start);
 | |
| extern void mn10300_dcache_flush_inv_range(unsigned start, unsigned end);
 | |
| extern void mn10300_dcache_flush_inv_range2(unsigned start, unsigned size);
 | |
| #else
 | |
| #define mn10300_dcache_flush()				do {} while (0)
 | |
| #define mn10300_dcache_flush_page(start)		do {} while (0)
 | |
| #define mn10300_dcache_flush_range(start, end)		do {} while (0)
 | |
| #define mn10300_dcache_flush_range2(start, size)	do {} while (0)
 | |
| #define mn10300_dcache_flush_inv()			mn10300_dcache_inv()
 | |
| #define mn10300_dcache_flush_inv_page(start) \
 | |
| 	mn10300_dcache_inv_page((start))
 | |
| #define mn10300_dcache_flush_inv_range(start, end) \
 | |
| 	mn10300_dcache_inv_range((start), (end))
 | |
| #define mn10300_dcache_flush_inv_range2(start, size) \
 | |
| 	mn10300_dcache_inv_range2((start), (size))
 | |
| #endif /* CONFIG_MN10300_CACHE_WBACK */
 | |
| #else
 | |
| #define mn10300_icache_inv()				do {} while (0)
 | |
| #define mn10300_dcache_inv()				do {} while (0)
 | |
| #define mn10300_dcache_inv_page(start)			do {} while (0)
 | |
| #define mn10300_dcache_inv_range(start, end)		do {} while (0)
 | |
| #define mn10300_dcache_inv_range2(start, size)		do {} while (0)
 | |
| #define mn10300_dcache_flush()				do {} while (0)
 | |
| #define mn10300_dcache_flush_inv_page(start)		do {} while (0)
 | |
| #define mn10300_dcache_flush_inv()			do {} while (0)
 | |
| #define mn10300_dcache_flush_inv_range(start, end)	do {} while (0)
 | |
| #define mn10300_dcache_flush_inv_range2(start, size)	do {} while (0)
 | |
| #define mn10300_dcache_flush_page(start)		do {} while (0)
 | |
| #define mn10300_dcache_flush_range(start, end)		do {} while (0)
 | |
| #define mn10300_dcache_flush_range2(start, size)	do {} while (0)
 | |
| #endif /* CONFIG_MN10300_CACHE_DISABLED */
 | |
| 
 | |
| /*
 | |
|  * internal debugging function
 | |
|  */
 | |
| #ifdef CONFIG_DEBUG_PAGEALLOC
 | |
| extern void kernel_map_pages(struct page *page, int numpages, int enable);
 | |
| #endif
 | |
| 
 | |
| #endif /* __ASSEMBLY__ */
 | |
| 
 | |
| #endif /* _ASM_CACHEFLUSH_H */
 |