980 lines
		
	
	
		
			29 KiB
		
	
	
	
		
			Diff
		
	
	
	
	
	
			
		
		
	
	
			980 lines
		
	
	
		
			29 KiB
		
	
	
	
		
			Diff
		
	
	
	
	
	
| From 940b4fea5ebfde3abe03c6469a57c01ee961497a Mon Sep 17 00:00:00 2001
 | |
| From: Kurt Mahan <kmahan@freescale.com>
 | |
| Date: Wed, 18 Jun 2008 15:20:21 -0600
 | |
| Subject: [PATCH] Split 547x/548x and 5445x cache routines into separate files.
 | |
| 
 | |
| LTIBName: mcfv4e-cache-split
 | |
| Signed-off-by: Kurt Mahan <kmahan@freescale.com>
 | |
| ---
 | |
|  include/asm-m68k/cf_5445x_cacheflush.h |  447 ++++++++++++++++++++++++++++++++
 | |
|  include/asm-m68k/cf_548x_cacheflush.h  |  259 ++++++++++++++++++
 | |
|  include/asm-m68k/cf_cacheflush.h       |  244 +-----------------
 | |
|  3 files changed, 711 insertions(+), 239 deletions(-)
 | |
|  create mode 100644 include/asm-m68k/cf_5445x_cacheflush.h
 | |
|  create mode 100644 include/asm-m68k/cf_548x_cacheflush.h
 | |
| 
 | |
| --- /dev/null
 | |
| +++ b/include/asm-m68k/cf_5445x_cacheflush.h
 | |
| @@ -0,0 +1,447 @@
 | |
| +/*
 | |
| + * include/asm-m68k/cf_5445x_cacheflush.h - Coldfire 5445x Cache
 | |
| + *
 | |
| + * Based on include/asm-m68k/cacheflush.h
 | |
| + *
 | |
| + * Coldfire pieces by:
 | |
| + *   Kurt Mahan kmahan@freescale.com
 | |
| + *
 | |
| + * Copyright Freescale Semiconductor, Inc. 2007, 2008
 | |
| + *
 | |
| + * This program is free software; you can redistribute  it and/or modify it
 | |
| + * under  the terms of  the GNU General  Public License as published by the
 | |
| + * Free Software Foundation;  either version 2 of the  License, or (at your
 | |
| + * option) any later version.
 | |
| + */
 | |
| +#ifndef M68K_CF_5445x_CACHEFLUSH_H
 | |
| +#define M68K_CF_5445x_CACHEFLUSH_H
 | |
| +
 | |
| +#include <asm/cfcache.h>
 | |
| +
 | |
| +/*
 | |
| + * Coldfire Cache Model
 | |
| + *
 | |
| + * The Coldfire processors use a Harvard architecture cache configured
 | |
| + * as four-way set associative.  The cache does not implement bus snooping
 | |
| + * so cache coherency with other masters must be maintained in software.
 | |
| + *
 | |
| + * The cache is managed via the CPUSHL instruction in conjunction with
 | |
| + * bits set in the CACR (cache control register).  Currently the code
 | |
| + * uses the CPUSHL enhancement which adds the ability to
 | |
| + * invalidate/clear/push a cacheline by physical address.  This feature
 | |
| + * is designated in the Hardware Configuration Register [D1-CPES].
 | |
| + *
 | |
| + * CACR Bits:
 | |
| + *	DPI[28]		cpushl invalidate disable for d-cache
 | |
| + *	IDPI[12]	cpushl invalidate disable for i-cache
 | |
| + *	SPA[14]		cpushl search by physical address
 | |
| + *	IVO[20]		cpushl invalidate only
 | |
| + *
 | |
| + * Random Terminology:
 | |
| + *  * invalidate = reset the cache line's valid bit
 | |
| + *  * push = generate a line-sized store of the data if its contents are marked
 | |
| + *	     as modifed (the modified flag is cleared after the store)
 | |
| + *  * clear = push + invalidate
 | |
| + */
 | |
| +
 | |
| +/**
 | |
| + * flush_icache - Flush all of the instruction cache
 | |
| + */
 | |
| +static inline void flush_icache(void)
 | |
| +{
 | |
| +	asm volatile("nop\n"
 | |
| +		     "moveq%.l	#0,%%d0\n"
 | |
| +		     "moveq%.l	#0,%%d1\n"
 | |
| +		     "move%.l	%%d0,%%a0\n"
 | |
| +		     "1:\n"
 | |
| +		     "cpushl	%%ic,(%%a0)\n"
 | |
| +		     "add%.l	#0x0010,%%a0\n"
 | |
| +		     "addq%.l	#1,%%d1\n"
 | |
| +		     "cmpi%.l	%0,%%d1\n"
 | |
| +		     "bne	1b\n"
 | |
| +		     "moveq%.l	#0,%%d1\n"
 | |
| +		     "addq%.l	#1,%%d0\n"
 | |
| +		     "move%.l	%%d0,%%a0\n"
 | |
| +		     "cmpi%.l	#4,%%d0\n"
 | |
| +		     "bne	1b\n"
 | |
| +		     : : "i" (CACHE_SETS)
 | |
| +		     : "a0", "d0", "d1");
 | |
| +}
 | |
| +
 | |
| +/**
 | |
| + * flush_dcache - Flush all of the data cache
 | |
| + */
 | |
| +static inline void flush_dcache(void)
 | |
| +{
 | |
| +	asm volatile("nop\n"
 | |
| +		     "moveq%.l	#0,%%d0\n"
 | |
| +		     "moveq%.l	#0,%%d1\n"
 | |
| +		     "move%.l	%%d0,%%a0\n"
 | |
| +		     "1:\n"
 | |
| +		     "cpushl	%%dc,(%%a0)\n"
 | |
| +		     "add%.l	#0x0010,%%a0\n"
 | |
| +		     "addq%.l	#1,%%d1\n"
 | |
| +		     "cmpi%.l	%0,%%d1\n"
 | |
| +		     "bne	1b\n"
 | |
| +		     "moveq%.l	#0,%%d1\n"
 | |
| +		     "addq%.l	#1,%%d0\n"
 | |
| +		     "move%.l	%%d0,%%a0\n"
 | |
| +		     "cmpi%.l	#4,%%d0\n"
 | |
| +		     "bne	1b\n"
 | |
| +		     : : "i" (CACHE_SETS)
 | |
| +		     : "a0", "d0", "d1");
 | |
| +}
 | |
| +
 | |
| +/**
 | |
| + * flush_bcache - Flush all of both caches
 | |
| + */
 | |
| +static inline void flush_bcache(void)
 | |
| +{
 | |
| +	asm volatile("nop\n"
 | |
| +		     "moveq%.l	#0,%%d0\n"
 | |
| +		     "moveq%.l	#0,%%d1\n"
 | |
| +		     "move%.l	%%d0,%%a0\n"
 | |
| +		     "1:\n"
 | |
| +		     "cpushl	%%bc,(%%a0)\n"
 | |
| +		     "add%.l	#0x0010,%%a0\n"
 | |
| +		     "addq%.l	#1,%%d1\n"
 | |
| +		     "cmpi%.l	%0,%%d1\n"
 | |
| +		     "bne	1b\n"
 | |
| +		     "moveq%.l	#0,%%d1\n"
 | |
| +		     "addq%.l	#1,%%d0\n"
 | |
| +		     "move%.l	%%d0,%%a0\n"
 | |
| +		     "cmpi%.l	#4,%%d0\n"
 | |
| +		     "bne	1b\n"
 | |
| +		     : : "i" (CACHE_SETS)
 | |
| +		     : "a0", "d0", "d1");
 | |
| +}
 | |
| +
 | |
| +/**
 | |
| + * cf_cache_clear - invalidate cache
 | |
| + * @paddr: starting physical address
 | |
| + * @len: number of bytes
 | |
| + *
 | |
| + * Invalidate cache lines starting at paddr for len bytes.
 | |
| + * Those lines are not pushed.
 | |
| + */
 | |
| +static inline void cf_cache_clear(unsigned long paddr, int len)
 | |
| +{
 | |
| +	/* number of lines */
 | |
| +	len = (len + (CACHE_LINE_SIZE-1)) / CACHE_LINE_SIZE;
 | |
| +	if (len == 0)
 | |
| +		return;
 | |
| +
 | |
| +	/* align on set boundary */
 | |
| +	paddr &= 0xfffffff0;
 | |
| +
 | |
| +	asm volatile("nop\n"
 | |
| +		     "move%.l   %2,%%d0\n"
 | |
| +		     "or%.l	%3,%%d0\n"
 | |
| +		     "movec	%%d0,%%cacr\n"
 | |
| +		     "move%.l	%0,%%a0\n"
 | |
| +		     "move%.l	%1,%%d0\n"
 | |
| +		     "1:\n"
 | |
| +		     "cpushl	%%bc,(%%a0)\n"
 | |
| +		     "lea	0x10(%%a0),%%a0\n"
 | |
| +		     "subq%.l	#1,%%d0\n"
 | |
| +		     "bne%.b	1b\n"
 | |
| +		     "movec	%2,%%cacr\n"
 | |
| +		     : : "a" (paddr), "r" (len),
 | |
| +			 "r" (shadow_cacr),
 | |
| +			 "i" (CF_CACR_SPA+CF_CACR_IVO)
 | |
| +		     : "a0", "d0");
 | |
| +}
 | |
| +
 | |
| +/**
 | |
| + * cf_cache_push - Push dirty cache out with no invalidate
 | |
| + * @paddr: starting physical address
 | |
| + * @len: number of bytes
 | |
| + *
 | |
| + * Push the any dirty lines starting at paddr for len bytes.
 | |
| + * Those lines are not invalidated.
 | |
| + */
 | |
| +static inline void cf_cache_push(unsigned long paddr, int len)
 | |
| +{
 | |
| +	/* number of lines */
 | |
| +	len = (len + (CACHE_LINE_SIZE-1)) / CACHE_LINE_SIZE;
 | |
| +	if (len == 0)
 | |
| +		return;
 | |
| +
 | |
| +	/* align on set boundary */
 | |
| +	paddr &= 0xfffffff0;
 | |
| +
 | |
| +	asm volatile("nop\n"
 | |
| +		     "move%.l   %2,%%d0\n"
 | |
| +		     "or%.l	%3,%%d0\n"
 | |
| +		     "movec	%%d0,%%cacr\n"
 | |
| +		     "move%.l	%0,%%a0\n"
 | |
| +		     "move%.l	%1,%%d0\n"
 | |
| +		     "1:\n"
 | |
| +		     "cpushl	%%bc,(%%a0)\n"
 | |
| +		     "lea	0x10(%%a0),%%a0\n"
 | |
| +		     "subq%.l	#1,%%d0\n"
 | |
| +		     "bne.b	1b\n"
 | |
| +		     "movec	%2,%%cacr\n"
 | |
| +		     : : "a" (paddr), "r" (len),
 | |
| +			 "r" (shadow_cacr),
 | |
| +			 "i" (CF_CACR_SPA+CF_CACR_DPI+CF_CACR_IDPI)
 | |
| +		     : "a0", "d0");
 | |
| +}
 | |
| +
 | |
| +/**
 | |
| + * cf_cache_flush - Push dirty cache out and invalidate
 | |
| + * @paddr: starting physical address
 | |
| + * @len: number of bytes
 | |
| + *
 | |
| + * Push the any dirty lines starting at paddr for len bytes and
 | |
| + * invalidate those lines.
 | |
| + */
 | |
| +static inline void cf_cache_flush(unsigned long paddr, int len)
 | |
| +{
 | |
| +	/* number of lines */
 | |
| +	len = (len + (CACHE_LINE_SIZE-1)) / CACHE_LINE_SIZE;
 | |
| +	if (len == 0)
 | |
| +		return;
 | |
| +
 | |
| +	/* align on set boundary */
 | |
| +	paddr &= 0xfffffff0;
 | |
| +
 | |
| +	asm volatile("nop\n"
 | |
| +		     "move%.l   %2,%%d0\n"
 | |
| +		     "or%.l	%3,%%d0\n"
 | |
| +		     "movec	%%d0,%%cacr\n"
 | |
| +		     "move%.l	%0,%%a0\n"
 | |
| +		     "move%.l	%1,%%d0\n"
 | |
| +		     "1:\n"
 | |
| +		     "cpushl	%%bc,(%%a0)\n"
 | |
| +		     "lea	0x10(%%a0),%%a0\n"
 | |
| +		     "subq%.l	#1,%%d0\n"
 | |
| +		     "bne.b	1b\n"
 | |
| +		     "movec	%2,%%cacr\n"
 | |
| +		     : : "a" (paddr), "r" (len),
 | |
| +			 "r" (shadow_cacr),
 | |
| +			 "i" (CF_CACR_SPA)
 | |
| +		     : "a0", "d0");
 | |
| +}
 | |
| +
 | |
| +/**
 | |
| + * cf_cache_flush_range - Push dirty data/inst cache in range out and invalidate
 | |
| + * @vstart - starting virtual address
 | |
| + * @vend: ending virtual address
 | |
| + *
 | |
| + * Push the any dirty data/instr lines starting at paddr for len bytes and
 | |
| + * invalidate those lines.
 | |
| + */
 | |
| +static inline void cf_cache_flush_range(unsigned long vstart, unsigned long vend)
 | |
| +{
 | |
| +	int len;
 | |
| +
 | |
| +	/* align on set boundary */
 | |
| +	vstart &= 0xfffffff0;
 | |
| +	vend = PAGE_ALIGN((vend + (CACHE_LINE_SIZE-1))) & 0xfffffff0;
 | |
| +	len = vend - vstart;
 | |
| +	if (len == 0)
 | |
| +		return;
 | |
| +	vstart = __pa(vstart);
 | |
| +	vend = vstart + len;
 | |
| +
 | |
| +	asm volatile("nop\n"
 | |
| +		     "move%.l   %2,%%d0\n"
 | |
| +		     "or%.l	%3,%%d0\n"
 | |
| +		     "movec	%%d0,%%cacr\n"
 | |
| +		     "move%.l	%0,%%a0\n"
 | |
| +		     "move%.l	%1,%%a1\n"
 | |
| +		     "1:\n"
 | |
| +		     "cpushl	%%bc,(%%a0)\n"
 | |
| +		     "lea	0x10(%%a0),%%a0\n"
 | |
| +		     "cmpa%.l	%%a0,%%a1\n"
 | |
| +		     "bne.b	1b\n"
 | |
| +		     "movec	%2,%%cacr\n"
 | |
| +		     : /* no return */
 | |
| +		     : "a" (vstart), "a" (vend),
 | |
| +		       "r" (shadow_cacr),
 | |
| +		       "i" (CF_CACR_SPA)
 | |
| +		     : "a0", "a1", "d0");
 | |
| +}
 | |
| +
 | |
| +/**
 | |
| + * cf_dcache_flush_range - Push dirty data cache in range out and invalidate
 | |
| + * @vstart - starting virtual address
 | |
| + * @vend: ending virtual address
 | |
| + *
 | |
| + * Push the any dirty data lines starting at paddr for len bytes and
 | |
| + * invalidate those lines.
 | |
| + */
 | |
| +static inline void cf_dcache_flush_range(unsigned long vstart, unsigned long vend)
 | |
| +{
 | |
| +	/* align on set boundary */
 | |
| +	vstart &= 0xfffffff0;
 | |
| +	vend = (vend + (CACHE_LINE_SIZE-1)) & 0xfffffff0;
 | |
| +
 | |
| +	asm volatile("nop\n"
 | |
| +		     "move%.l   %2,%%d0\n"
 | |
| +		     "or%.l	%3,%%d0\n"
 | |
| +		     "movec	%%d0,%%cacr\n"
 | |
| +		     "move%.l	%0,%%a0\n"
 | |
| +		     "move%.l	%1,%%a1\n"
 | |
| +		     "1:\n"
 | |
| +		     "cpushl	%%dc,(%%a0)\n"
 | |
| +		     "lea	0x10(%%a0),%%a0\n"
 | |
| +		     "cmpa%.l	%%a0,%%a1\n"
 | |
| +		     "bne.b	1b\n"
 | |
| +		     "movec	%2,%%cacr\n"
 | |
| +		     : /* no return */
 | |
| +		     : "a" (__pa(vstart)), "a" (__pa(vend)),
 | |
| +		       "r" (shadow_cacr),
 | |
| +		       "i" (CF_CACR_SPA)
 | |
| +		     : "a0", "a1", "d0");
 | |
| +}
 | |
| +
 | |
| +/**
 | |
| + * cf_icache_flush_range - Push dirty inst cache in range out and invalidate
 | |
| + * @vstart - starting virtual address
 | |
| + * @vend: ending virtual address
 | |
| + *
 | |
| + * Push the any dirty instr lines starting at paddr for len bytes and
 | |
| + * invalidate those lines.  This should just be an invalidate since you
 | |
| + * shouldn't be able to have dirty instruction cache.
 | |
| + */
 | |
| +static inline void cf_icache_flush_range(unsigned long vstart, unsigned long vend)
 | |
| +{
 | |
| +	/* align on set boundary */
 | |
| +	vstart &= 0xfffffff0;
 | |
| +	vend = (vend + (CACHE_LINE_SIZE-1)) & 0xfffffff0;
 | |
| +
 | |
| +	asm volatile("nop\n"
 | |
| +		     "move%.l   %2,%%d0\n"
 | |
| +		     "or%.l	%3,%%d0\n"
 | |
| +		     "movec	%%d0,%%cacr\n"
 | |
| +		     "move%.l	%0,%%a0\n"
 | |
| +		     "move%.l	%1,%%a1\n"
 | |
| +		     "1:\n"
 | |
| +		     "cpushl	%%ic,(%%a0)\n"
 | |
| +		     "lea	0x10(%%a0),%%a0\n"
 | |
| +		     "cmpa%.l	%%a0,%%a1\n"
 | |
| +		     "bne.b	1b\n"
 | |
| +		     "movec	%2,%%cacr\n"
 | |
| +		     : /* no return */
 | |
| +		     : "a" (__pa(vstart)), "a" (__pa(vend)),
 | |
| +		       "r" (shadow_cacr),
 | |
| +		       "i" (CF_CACR_SPA)
 | |
| +		     : "a0", "a1", "d0");
 | |
| +}
 | |
| +
 | |
| +/**
 | |
| + * flush_cache_mm - Flush an mm_struct
 | |
| + * @mm: mm_struct to flush
 | |
| + */
 | |
| +static inline void flush_cache_mm(struct mm_struct *mm)
 | |
| +{
 | |
| +	if (mm == current->mm)
 | |
| +		flush_bcache();
 | |
| +}
 | |
| +
 | |
| +#define flush_cache_dup_mm(mm)	flush_cache_mm(mm)
 | |
| +
 | |
| +/**
 | |
| + * flush_cache_range - Flush a cache range
 | |
| + * @vma: vma struct
 | |
| + * @start: Starting address
 | |
| + * @end: Ending address
 | |
| + *
 | |
| + * flush_cache_range must be a macro to avoid a dependency on
 | |
| + * linux/mm.h which includes this file.
 | |
| + */
 | |
| +static inline void flush_cache_range(struct vm_area_struct *vma,
 | |
| +	unsigned long start, unsigned long end)
 | |
| +{
 | |
| +	if (vma->vm_mm == current->mm)
 | |
| +		cf_cache_flush_range(start, end);
 | |
| +}
 | |
| +
 | |
| +/**
 | |
| + * flush_cache_page - Flush a page of the cache
 | |
| + * @vma: vma struct
 | |
| + * @vmaddr:
 | |
| + * @pfn: page numer
 | |
| + *
 | |
| + * flush_cache_page must be a macro to avoid a dependency on
 | |
| + * linux/mm.h which includes this file.
 | |
| + */
 | |
| +static inline void flush_cache_page(struct vm_area_struct *vma,
 | |
| +	unsigned long vmaddr, unsigned long pfn)
 | |
| +{
 | |
| +	if (vma->vm_mm == current->mm)
 | |
| +		cf_cache_flush_range(vmaddr, vmaddr+PAGE_SIZE);
 | |
| +}
 | |
| +
 | |
| +/**
 | |
| + * __flush_page_to_ram - Push a page out of the cache
 | |
| + * @vaddr: Virtual address at start of page
 | |
| + *
 | |
| + * Push the page at kernel virtual address *vaddr* and clear
 | |
| + * the icache.
 | |
| + */
 | |
| +static inline void __flush_page_to_ram(void *vaddr)
 | |
| +{
 | |
| +	asm volatile("nop\n"
 | |
| +		     "move%.l   %2,%%d0\n"
 | |
| +		     "or%.l	%3,%%d0\n"
 | |
| +		     "movec	%%d0,%%cacr\n"
 | |
| +		     "move%.l	%0,%%d0\n"
 | |
| +		     "and%.l	#0xfffffff0,%%d0\n"
 | |
| +		     "move%.l	%%d0,%%a0\n"
 | |
| +		     "move%.l	%1,%%d0\n"
 | |
| +		     "1:\n"
 | |
| +		     "cpushl	%%bc,(%%a0)\n"
 | |
| +		     "lea	0x10(%%a0),%%a0\n"
 | |
| +		     "subq%.l	#1,%%d0\n"
 | |
| +		     "bne.b	1b\n"
 | |
| +		     "movec	%2,%%cacr\n"
 | |
| +		     : : "a" (__pa(vaddr)), "i" (PAGE_SIZE / CACHE_LINE_SIZE),
 | |
| +			 "r" (shadow_cacr), "i" (CF_CACR_SPA)
 | |
| +		     : "a0", "d0");
 | |
| +}
 | |
| +
 | |
| +/*
 | |
| + * Various defines for the kernel.
 | |
| + */
 | |
| +
 | |
| +extern void cache_clear(unsigned long paddr, int len);
 | |
| +extern void cache_push(unsigned long paddr, int len);
 | |
| +extern void flush_icache_range(unsigned long address, unsigned long endaddr);
 | |
| +
 | |
| +#define flush_cache_all()			flush_bcache()
 | |
| +#define flush_cache_vmap(start, end)		flush_bcache()
 | |
| +#define flush_cache_vunmap(start, end)		flush_bcache()
 | |
| +
 | |
| +#define flush_dcache_range(vstart, vend)	cf_dcache_flush_range(vstart, vend)
 | |
| +#define flush_dcache_page(page)			__flush_page_to_ram(page_address(page))
 | |
| +#define flush_dcache_mmap_lock(mapping)		do { } while (0)
 | |
| +#define flush_dcache_mmap_unlock(mapping)	do { } while (0)
 | |
| +
 | |
| +#define flush_icache_page(vma, page)		__flush_page_to_ram(page_address(page))
 | |
| +
 | |
| +/**
 | |
| + * copy_to_user_page - Copy memory to user page
 | |
| + */
 | |
| +static inline void copy_to_user_page(struct vm_area_struct *vma,
 | |
| +				     struct page *page, unsigned long vaddr,
 | |
| +				     void *dst, void *src, int len)
 | |
| +{
 | |
| +	memcpy(dst, src, len);
 | |
| +	cf_cache_flush(page_to_phys(page), PAGE_SIZE);
 | |
| +}
 | |
| +
 | |
| +/**
 | |
| + * copy_from_user_page - Copy memory from user page
 | |
| + */
 | |
| +static inline void copy_from_user_page(struct vm_area_struct *vma,
 | |
| +				       struct page *page, unsigned long vaddr,
 | |
| +				       void *dst, void *src, int len)
 | |
| +{
 | |
| +	cf_cache_flush(page_to_phys(page), PAGE_SIZE);
 | |
| +	memcpy(dst, src, len);
 | |
| +}
 | |
| +
 | |
| +#endif /* M68K_CF_5445x_CACHEFLUSH_H */
 | |
| --- /dev/null
 | |
| +++ b/include/asm-m68k/cf_548x_cacheflush.h
 | |
| @@ -0,0 +1,259 @@
 | |
| +/*
 | |
| + * include/asm-m68k/cf_548x_cacheflush.h - Coldfire 547x/548x Cache
 | |
| + *
 | |
| + * Based on include/asm-m68k/cacheflush.h
 | |
| + *
 | |
| + * Coldfire pieces by:
 | |
| + *   Kurt Mahan kmahan@freescale.com
 | |
| + *
 | |
| + * Copyright Freescale Semiconductor, Inc. 2007, 2008
 | |
| + *
 | |
| + * This program is free software; you can redistribute  it and/or modify it
 | |
| + * under  the terms of  the GNU General  Public License as published by the
 | |
| + * Free Software Foundation;  either version 2 of the  License, or (at your
 | |
| + * option) any later version.
 | |
| + */
 | |
| +#ifndef M68K_CF_548x_CACHEFLUSH_H
 | |
| +#define M68K_CF_548x_CACHEFLUSH_H
 | |
| +
 | |
| +#include <asm/cfcache.h>
 | |
| +/*
 | |
| + * Cache handling functions
 | |
| + */
 | |
| +
 | |
| +#define flush_icache()						\
 | |
| +({								\
 | |
| +  unsigned long set;						\
 | |
| +  unsigned long start_set;					\
 | |
| +  unsigned long end_set;					\
 | |
| +								\
 | |
| +  start_set = 0;						\
 | |
| +  end_set = (unsigned long)LAST_DCACHE_ADDR;			\
 | |
| +    								\
 | |
| +  for (set = start_set; set <= end_set; set += (0x10 - 3)) {	\
 | |
| +    asm volatile("cpushl %%ic,(%0)\n"				\
 | |
| +                 "\taddq%.l #1,%0\n"				\
 | |
| +                 "\tcpushl %%ic,(%0)\n"				\
 | |
| +                 "\taddq%.l #1,%0\n"				\
 | |
| +                 "\tcpushl %%ic,(%0)\n"				\
 | |
| +                 "\taddq%.l #1,%0\n"				\
 | |
| +                 "\tcpushl %%ic,(%0)" : "=a" (set) : "a" (set));		\
 | |
| +  }								\
 | |
| +})
 | |
| +
 | |
| +#define flush_dcache()						\
 | |
| +({								\
 | |
| +  unsigned long set;						\
 | |
| +  unsigned long start_set;					\
 | |
| +  unsigned long end_set;					\
 | |
| +								\
 | |
| +  start_set = 0;						\
 | |
| +  end_set = (unsigned long)LAST_DCACHE_ADDR;			\
 | |
| +    								\
 | |
| +  for (set = start_set; set <= end_set; set += (0x10 - 3)) {	\
 | |
| +    asm volatile("cpushl %%dc,(%0)\n"				\
 | |
| +                 "\taddq%.l #1,%0\n"				\
 | |
| +                 "\tcpushl %%dc,(%0)\n"				\
 | |
| +                 "\taddq%.l #1,%0\n"				\
 | |
| +                 "\tcpushl %%dc,(%0)\n"				\
 | |
| +                 "\taddq%.l #1,%0\n"				\
 | |
| +                 "\tcpushl %%dc,(%0)" : "=a" (set) : "a" (set));		\
 | |
| +  }								\
 | |
| +})
 | |
| +
 | |
| +#define flush_bcache()						\
 | |
| +({								\
 | |
| +  unsigned long set;						\
 | |
| +  unsigned long start_set;					\
 | |
| +  unsigned long end_set;					\
 | |
| +								\
 | |
| +  start_set = 0;						\
 | |
| +  end_set = (unsigned long)LAST_DCACHE_ADDR;			\
 | |
| +    								\
 | |
| +  for (set = start_set; set <= end_set; set += (0x10 - 3)) {	\
 | |
| +    asm volatile("cpushl %%bc,(%0)\n"				\
 | |
| +                 "\taddq%.l #1,%0\n"				\
 | |
| +                 "\tcpushl %%bc,(%0)\n"				\
 | |
| +                 "\taddq%.l #1,%0\n"				\
 | |
| +                 "\tcpushl %%bc,(%0)\n"				\
 | |
| +                 "\taddq%.l #1,%0\n"				\
 | |
| +                 "\tcpushl %%bc,(%0)" : "=a" (set) : "a" (set));		\
 | |
| +  }								\
 | |
| +})
 | |
| +
 | |
| +/*
 | |
| + * invalidate the cache for the specified memory range.
 | |
| + * It starts at the physical address specified for
 | |
| + * the given number of bytes.
 | |
| + */
 | |
| +extern void cache_clear(unsigned long paddr, int len);
 | |
| +/*
 | |
| + * push any dirty cache in the specified memory range.
 | |
| + * It starts at the physical address specified for
 | |
| + * the given number of bytes.
 | |
| + */
 | |
| +extern void cache_push(unsigned long paddr, int len);
 | |
| +
 | |
| +/*
 | |
| + * push and invalidate pages in the specified user virtual
 | |
| + * memory range.
 | |
| + */
 | |
| +extern void cache_push_v(unsigned long vaddr, int len);
 | |
| +
 | |
| +/* This is needed whenever the virtual mapping of the current
 | |
| +   process changes.  */
 | |
| +
 | |
| +/**
 | |
| + * flush_cache_mm - Flush an mm_struct
 | |
| + * @mm: mm_struct to flush
 | |
| + */
 | |
| +static inline void flush_cache_mm(struct mm_struct *mm)
 | |
| +{
 | |
| +	if (mm == current->mm)
 | |
| +		flush_bcache();
 | |
| +}
 | |
| +
 | |
| +#define flush_cache_dup_mm(mm)	flush_cache_mm(mm)
 | |
| +
 | |
| +#define flush_cache_all()		flush_bcache()
 | |
| +
 | |
| +/**
 | |
| + * flush_cache_range - Flush a cache range
 | |
| + * @vma: vma struct
 | |
| + * @start: Starting address
 | |
| + * @end: Ending address
 | |
| + *
 | |
| + * flush_cache_range must be a macro to avoid a dependency on
 | |
| + * linux/mm.h which includes this file.
 | |
| + */
 | |
| +static inline void flush_cache_range(struct vm_area_struct *vma,
 | |
| +	unsigned long start, unsigned long end)
 | |
| +{
 | |
| +	if (vma->vm_mm == current->mm)
 | |
| +		flush_bcache();
 | |
| +//		cf_cache_flush_range(start, end);
 | |
| +}
 | |
| +
 | |
| +/**
 | |
| + * flush_cache_page - Flush a page of the cache
 | |
| + * @vma: vma struct
 | |
| + * @vmaddr:
 | |
| + * @pfn: page numer
 | |
| + *
 | |
| + * flush_cache_page must be a macro to avoid a dependency on
 | |
| + * linux/mm.h which includes this file.
 | |
| + */
 | |
| +static inline void flush_cache_page(struct vm_area_struct *vma,
 | |
| +	unsigned long vmaddr, unsigned long pfn)
 | |
| +{
 | |
| +	if (vma->vm_mm == current->mm)
 | |
| +		flush_bcache();
 | |
| +//		cf_cache_flush_range(vmaddr, vmaddr+PAGE_SIZE);
 | |
| +}
 | |
| +
 | |
| +/* Push the page at kernel virtual address and clear the icache */
 | |
| +/* RZ: use cpush %bc instead of cpush %dc, cinv %ic */
 | |
| +#define flush_page_to_ram(page) __flush_page_to_ram((void *) page_address(page))
 | |
| +extern inline void __flush_page_to_ram(void *address)
 | |
| +{
 | |
| +  unsigned long set;
 | |
| +  unsigned long start_set;
 | |
| +  unsigned long end_set;
 | |
| +  unsigned long addr = (unsigned long) address;
 | |
| +
 | |
| +  addr &= ~(PAGE_SIZE - 1); /* round down to page start address */
 | |
| +
 | |
| +  start_set = addr & _ICACHE_SET_MASK;
 | |
| +  end_set = (addr + PAGE_SIZE-1) & _ICACHE_SET_MASK;
 | |
| +
 | |
| +  if (start_set > end_set) {
 | |
| +    /* from the begining to the lowest address */
 | |
| +    for (set = 0; set <= end_set; set += (0x10 - 3)) {
 | |
| +      asm volatile("cpushl %%bc,(%0)\n"
 | |
| +                   "\taddq%.l #1,%0\n"
 | |
| +                   "\tcpushl %%bc,(%0)\n"
 | |
| +                   "\taddq%.l #1,%0\n"
 | |
| +                   "\tcpushl %%bc,(%0)\n"
 | |
| +                   "\taddq%.l #1,%0\n"
 | |
| +                   "\tcpushl %%bc,(%0)" : "=a" (set) : "a" (set));
 | |
| +    }
 | |
| +    /* next loop will finish the cache ie pass the hole */
 | |
| +    end_set = LAST_ICACHE_ADDR;    
 | |
| +  }
 | |
| +  for (set = start_set; set <= end_set; set += (0x10 - 3)) {
 | |
| +    asm volatile("cpushl %%bc,(%0)\n"
 | |
| +                 "\taddq%.l #1,%0\n"
 | |
| +                 "\tcpushl %%bc,(%0)\n"
 | |
| +                 "\taddq%.l #1,%0\n"
 | |
| +                 "\tcpushl %%bc,(%0)\n"
 | |
| +                 "\taddq%.l #1,%0\n"
 | |
| +                 "\tcpushl %%bc,(%0)" : "=a" (set) : "a" (set));
 | |
| +  }
 | |
| +}
 | |
| +
 | |
| +/* Use __flush_page_to_ram() for flush_dcache_page all values are same - MW */
 | |
| +#define flush_dcache_page(page)			\
 | |
| +	__flush_page_to_ram((void *) page_address(page))
 | |
| +#define flush_icache_page(vma,pg)		\
 | |
| +	__flush_page_to_ram((void *) page_address(pg))
 | |
| +#define flush_icache_user_range(adr,len)	do { } while (0)
 | |
| +/* NL */
 | |
| +#define flush_icache_user_page(vma,page,addr,len)	do { } while (0)
 | |
| +
 | |
| +/* Push n pages at kernel virtual address and clear the icache */
 | |
| +/* RZ: use cpush %bc instead of cpush %dc, cinv %ic */
 | |
| +extern inline void flush_icache_range (unsigned long address,
 | |
| +				       unsigned long endaddr)
 | |
| +{
 | |
| +  unsigned long set;
 | |
| +  unsigned long start_set;
 | |
| +  unsigned long end_set;
 | |
| +
 | |
| +  start_set = address & _ICACHE_SET_MASK;
 | |
| +  end_set = endaddr & _ICACHE_SET_MASK;
 | |
| +
 | |
| +  if (start_set > end_set) {
 | |
| +    /* from the begining to the lowest address */
 | |
| +    for (set = 0; set <= end_set; set += (0x10 - 3)) {
 | |
| +      asm volatile("cpushl %%ic,(%0)\n"
 | |
| +                   "\taddq%.l #1,%0\n"
 | |
| +                   "\tcpushl %%ic,(%0)\n"
 | |
| +                   "\taddq%.l #1,%0\n"
 | |
| +                   "\tcpushl %%ic,(%0)\n"
 | |
| +                   "\taddq%.l #1,%0\n"
 | |
| +                   "\tcpushl %%ic,(%0)" : "=a" (set) : "a" (set));
 | |
| +    }
 | |
| +    /* next loop will finish the cache ie pass the hole */
 | |
| +    end_set = LAST_ICACHE_ADDR;    
 | |
| +  }
 | |
| +  for (set = start_set; set <= end_set; set += (0x10 - 3)) {
 | |
| +    asm volatile("cpushl %%ic,(%0)\n"
 | |
| +                 "\taddq%.l #1,%0\n"
 | |
| +                 "\tcpushl %%ic,(%0)\n"
 | |
| +                 "\taddq%.l #1,%0\n"
 | |
| +                 "\tcpushl %%ic,(%0)\n"
 | |
| +                 "\taddq%.l #1,%0\n"
 | |
| +                 "\tcpushl %%ic,(%0)" : "=a" (set) : "a" (set));
 | |
| +  }
 | |
| +}
 | |
| +
 | |
| +static inline void copy_to_user_page(struct vm_area_struct *vma,
 | |
| +				     struct page *page, unsigned long vaddr,
 | |
| +				     void *dst, void *src, int len)
 | |
| +{
 | |
| +	memcpy(dst, src, len);
 | |
| +	flush_icache_user_page(vma, page, vaddr, len);
 | |
| +}
 | |
| +static inline void copy_from_user_page(struct vm_area_struct *vma,
 | |
| +				       struct page *page, unsigned long vaddr,
 | |
| +				       void *dst, void *src, int len)
 | |
| +{
 | |
| +	memcpy(dst, src, len);
 | |
| +}
 | |
| +
 | |
| +#define flush_cache_vmap(start, end)		flush_cache_all()
 | |
| +#define flush_cache_vunmap(start, end)		flush_cache_all()
 | |
| +#define flush_dcache_mmap_lock(mapping)		do { } while (0)
 | |
| +#define flush_dcache_mmap_unlock(mapping)	do { } while (0)
 | |
| +
 | |
| +#endif /* M68K_CF_548x_CACHEFLUSH_H */
 | |
| --- a/include/asm-m68k/cf_cacheflush.h
 | |
| +++ b/include/asm-m68k/cf_cacheflush.h
 | |
| @@ -1,244 +1,10 @@
 | |
|  #ifndef M68K_CF_CACHEFLUSH_H
 | |
|  #define M68K_CF_CACHEFLUSH_H
 | |
|  
 | |
| -#include <asm/cfcache.h>
 | |
| -/*
 | |
| - * Cache handling functions
 | |
| - */
 | |
| -
 | |
| -#define flush_icache()						\
 | |
| -({								\
 | |
| -  unsigned long set;						\
 | |
| -  unsigned long start_set;					\
 | |
| -  unsigned long end_set;					\
 | |
| -								\
 | |
| -  start_set = 0;						\
 | |
| -  end_set = (unsigned long)LAST_DCACHE_ADDR;			\
 | |
| -    								\
 | |
| -  for (set = start_set; set <= end_set; set += (0x10 - 3)) {	\
 | |
| -    asm volatile("cpushl %%ic,(%0)\n"				\
 | |
| -                 "\taddq%.l #1,%0\n"				\
 | |
| -                 "\tcpushl %%ic,(%0)\n"				\
 | |
| -                 "\taddq%.l #1,%0\n"				\
 | |
| -                 "\tcpushl %%ic,(%0)\n"				\
 | |
| -                 "\taddq%.l #1,%0\n"				\
 | |
| -                 "\tcpushl %%ic,(%0)" : "=a" (set) : "a" (set));		\
 | |
| -  }								\
 | |
| -})
 | |
| -
 | |
| -#define flush_dcache()						\
 | |
| -({								\
 | |
| -  unsigned long set;						\
 | |
| -  unsigned long start_set;					\
 | |
| -  unsigned long end_set;					\
 | |
| -								\
 | |
| -  start_set = 0;						\
 | |
| -  end_set = (unsigned long)LAST_DCACHE_ADDR;			\
 | |
| -    								\
 | |
| -  for (set = start_set; set <= end_set; set += (0x10 - 3)) {	\
 | |
| -    asm volatile("cpushl %%dc,(%0)\n"				\
 | |
| -                 "\taddq%.l #1,%0\n"				\
 | |
| -                 "\tcpushl %%dc,(%0)\n"				\
 | |
| -                 "\taddq%.l #1,%0\n"				\
 | |
| -                 "\tcpushl %%dc,(%0)\n"				\
 | |
| -                 "\taddq%.l #1,%0\n"				\
 | |
| -                 "\tcpushl %%dc,(%0)" : "=a" (set) : "a" (set));		\
 | |
| -  }								\
 | |
| -})
 | |
| -
 | |
| -#define flush_bcache()						\
 | |
| -({								\
 | |
| -  unsigned long set;						\
 | |
| -  unsigned long start_set;					\
 | |
| -  unsigned long end_set;					\
 | |
| -								\
 | |
| -  start_set = 0;						\
 | |
| -  end_set = (unsigned long)LAST_DCACHE_ADDR;			\
 | |
| -    								\
 | |
| -  for (set = start_set; set <= end_set; set += (0x10 - 3)) {	\
 | |
| -    asm volatile("cpushl %%bc,(%0)\n"				\
 | |
| -                 "\taddq%.l #1,%0\n"				\
 | |
| -                 "\tcpushl %%bc,(%0)\n"				\
 | |
| -                 "\taddq%.l #1,%0\n"				\
 | |
| -                 "\tcpushl %%bc,(%0)\n"				\
 | |
| -                 "\taddq%.l #1,%0\n"				\
 | |
| -                 "\tcpushl %%bc,(%0)" : "=a" (set) : "a" (set));		\
 | |
| -  }								\
 | |
| -})
 | |
| -
 | |
| -/*
 | |
| - * invalidate the cache for the specified memory range.
 | |
| - * It starts at the physical address specified for
 | |
| - * the given number of bytes.
 | |
| - */
 | |
| -extern void cache_clear(unsigned long paddr, int len);
 | |
| -/*
 | |
| - * push any dirty cache in the specified memory range.
 | |
| - * It starts at the physical address specified for
 | |
| - * the given number of bytes.
 | |
| - */
 | |
| -extern void cache_push(unsigned long paddr, int len);
 | |
| -
 | |
| -/*
 | |
| - * push and invalidate pages in the specified user virtual
 | |
| - * memory range.
 | |
| - */
 | |
| -extern void cache_push_v(unsigned long vaddr, int len);
 | |
| -
 | |
| -/* This is needed whenever the virtual mapping of the current
 | |
| -   process changes.  */
 | |
| -
 | |
| -/**
 | |
| - * flush_cache_mm - Flush an mm_struct
 | |
| - * @mm: mm_struct to flush
 | |
| - */
 | |
| -static inline void flush_cache_mm(struct mm_struct *mm)
 | |
| -{
 | |
| -	if (mm == current->mm)
 | |
| -		flush_bcache();
 | |
| -}
 | |
| -
 | |
| -#define flush_cache_dup_mm(mm)	flush_cache_mm(mm)
 | |
| -
 | |
| -#define flush_cache_all()		flush_bcache()
 | |
| -
 | |
| -/**
 | |
| - * flush_cache_range - Flush a cache range
 | |
| - * @vma: vma struct
 | |
| - * @start: Starting address
 | |
| - * @end: Ending address
 | |
| - *
 | |
| - * flush_cache_range must be a macro to avoid a dependency on
 | |
| - * linux/mm.h which includes this file.
 | |
| - */
 | |
| -static inline void flush_cache_range(struct vm_area_struct *vma,
 | |
| -	unsigned long start, unsigned long end)
 | |
| -{
 | |
| -	if (vma->vm_mm == current->mm)
 | |
| -		flush_bcache();
 | |
| -//		cf_cache_flush_range(start, end);
 | |
| -}
 | |
| -
 | |
| -/**
 | |
| - * flush_cache_page - Flush a page of the cache
 | |
| - * @vma: vma struct
 | |
| - * @vmaddr:
 | |
| - * @pfn: page numer
 | |
| - *
 | |
| - * flush_cache_page must be a macro to avoid a dependency on
 | |
| - * linux/mm.h which includes this file.
 | |
| - */
 | |
| -static inline void flush_cache_page(struct vm_area_struct *vma,
 | |
| -	unsigned long vmaddr, unsigned long pfn)
 | |
| -{
 | |
| -	if (vma->vm_mm == current->mm)
 | |
| -		flush_bcache();
 | |
| -//		cf_cache_flush_range(vmaddr, vmaddr+PAGE_SIZE);
 | |
| -}
 | |
| -
 | |
| -/* Push the page at kernel virtual address and clear the icache */
 | |
| -/* RZ: use cpush %bc instead of cpush %dc, cinv %ic */
 | |
| -#define flush_page_to_ram(page) __flush_page_to_ram((void *) page_address(page))
 | |
| -extern inline void __flush_page_to_ram(void *address)
 | |
| -{
 | |
| -  unsigned long set;
 | |
| -  unsigned long start_set;
 | |
| -  unsigned long end_set;
 | |
| -  unsigned long addr = (unsigned long) address;
 | |
| -
 | |
| -  addr &= ~(PAGE_SIZE - 1); /* round down to page start address */
 | |
| -
 | |
| -  start_set = addr & _ICACHE_SET_MASK;
 | |
| -  end_set = (addr + PAGE_SIZE-1) & _ICACHE_SET_MASK;
 | |
| -
 | |
| -  if (start_set > end_set) {
 | |
| -    /* from the begining to the lowest address */
 | |
| -    for (set = 0; set <= end_set; set += (0x10 - 3)) {
 | |
| -      asm volatile("cpushl %%bc,(%0)\n"
 | |
| -                   "\taddq%.l #1,%0\n"
 | |
| -                   "\tcpushl %%bc,(%0)\n"
 | |
| -                   "\taddq%.l #1,%0\n"
 | |
| -                   "\tcpushl %%bc,(%0)\n"
 | |
| -                   "\taddq%.l #1,%0\n"
 | |
| -                   "\tcpushl %%bc,(%0)" : "=a" (set) : "a" (set));
 | |
| -    }
 | |
| -    /* next loop will finish the cache ie pass the hole */
 | |
| -    end_set = LAST_ICACHE_ADDR;    
 | |
| -  }
 | |
| -  for (set = start_set; set <= end_set; set += (0x10 - 3)) {
 | |
| -    asm volatile("cpushl %%bc,(%0)\n"
 | |
| -                 "\taddq%.l #1,%0\n"
 | |
| -                 "\tcpushl %%bc,(%0)\n"
 | |
| -                 "\taddq%.l #1,%0\n"
 | |
| -                 "\tcpushl %%bc,(%0)\n"
 | |
| -                 "\taddq%.l #1,%0\n"
 | |
| -                 "\tcpushl %%bc,(%0)" : "=a" (set) : "a" (set));
 | |
| -  }
 | |
| -}
 | |
| -
 | |
| -/* Use __flush_page_to_ram() for flush_dcache_page all values are same - MW */
 | |
| -#define flush_dcache_page(page)			\
 | |
| -	__flush_page_to_ram((void *) page_address(page))
 | |
| -#define flush_icache_page(vma,pg)		\
 | |
| -	__flush_page_to_ram((void *) page_address(pg))
 | |
| -#define flush_icache_user_range(adr,len)	do { } while (0)
 | |
| -/* NL */
 | |
| -#define flush_icache_user_page(vma,page,addr,len)	do { } while (0)
 | |
| -
 | |
| -/* Push n pages at kernel virtual address and clear the icache */
 | |
| -/* RZ: use cpush %bc instead of cpush %dc, cinv %ic */
 | |
| -extern inline void flush_icache_range (unsigned long address,
 | |
| -				       unsigned long endaddr)
 | |
| -{
 | |
| -  unsigned long set;
 | |
| -  unsigned long start_set;
 | |
| -  unsigned long end_set;
 | |
| -
 | |
| -  start_set = address & _ICACHE_SET_MASK;
 | |
| -  end_set = endaddr & _ICACHE_SET_MASK;
 | |
| -
 | |
| -  if (start_set > end_set) {
 | |
| -    /* from the begining to the lowest address */
 | |
| -    for (set = 0; set <= end_set; set += (0x10 - 3)) {
 | |
| -      asm volatile("cpushl %%ic,(%0)\n"
 | |
| -                   "\taddq%.l #1,%0\n"
 | |
| -                   "\tcpushl %%ic,(%0)\n"
 | |
| -                   "\taddq%.l #1,%0\n"
 | |
| -                   "\tcpushl %%ic,(%0)\n"
 | |
| -                   "\taddq%.l #1,%0\n"
 | |
| -                   "\tcpushl %%ic,(%0)" : "=a" (set) : "a" (set));
 | |
| -    }
 | |
| -    /* next loop will finish the cache ie pass the hole */
 | |
| -    end_set = LAST_ICACHE_ADDR;    
 | |
| -  }
 | |
| -  for (set = start_set; set <= end_set; set += (0x10 - 3)) {
 | |
| -    asm volatile("cpushl %%ic,(%0)\n"
 | |
| -                 "\taddq%.l #1,%0\n"
 | |
| -                 "\tcpushl %%ic,(%0)\n"
 | |
| -                 "\taddq%.l #1,%0\n"
 | |
| -                 "\tcpushl %%ic,(%0)\n"
 | |
| -                 "\taddq%.l #1,%0\n"
 | |
| -                 "\tcpushl %%ic,(%0)" : "=a" (set) : "a" (set));
 | |
| -  }
 | |
| -}
 | |
| -
 | |
| -static inline void copy_to_user_page(struct vm_area_struct *vma,
 | |
| -				     struct page *page, unsigned long vaddr,
 | |
| -				     void *dst, void *src, int len)
 | |
| -{
 | |
| -	memcpy(dst, src, len);
 | |
| -	flush_icache_user_page(vma, page, vaddr, len);
 | |
| -}
 | |
| -static inline void copy_from_user_page(struct vm_area_struct *vma,
 | |
| -				       struct page *page, unsigned long vaddr,
 | |
| -				       void *dst, void *src, int len)
 | |
| -{
 | |
| -	memcpy(dst, src, len);
 | |
| -}
 | |
| -
 | |
| -#define flush_cache_vmap(start, end)		flush_cache_all()
 | |
| -#define flush_cache_vunmap(start, end)		flush_cache_all()
 | |
| -#define flush_dcache_mmap_lock(mapping)		do { } while (0)
 | |
| -#define flush_dcache_mmap_unlock(mapping)	do { } while (0)
 | |
| +#ifdef CONFIG_M5445X
 | |
| +#include "cf_5445x_cacheflush.h"
 | |
| +#else
 | |
| +#include "cf_548x_cacheflush.h"
 | |
| +#endif
 | |
|  
 | |
|  #endif /* M68K_CF_CACHEFLUSH_H */
 |