84 lines
2.8 KiB
C
84 lines
2.8 KiB
C
/* SPDX-License-Identifier: GPL-2.0-only */
|
|
/*
|
|
* Cache flush operations for the Hexagon architecture
|
|
*
|
|
* Copyright (c) 2010-2011, The Linux Foundation. All rights reserved.
|
|
*/
|
|
|
|
#ifndef _ASM_CACHEFLUSH_H
|
|
#define _ASM_CACHEFLUSH_H
|
|
|
|
#include <linux/mm_types.h>
|
|
|
|
/* Cache flushing:
|
|
*
|
|
* - flush_cache_all() flushes entire cache
|
|
* - flush_cache_mm(mm) flushes the specified mm context's cache lines
|
|
* - flush_cache_page(mm, vmaddr, pfn) flushes a single page
|
|
* - flush_cache_range(vma, start, end) flushes a range of pages
|
|
* - flush_icache_range(start, end) flush a range of instructions
|
|
* - flush_dcache_page(pg) flushes(wback&invalidates) a page for dcache
|
|
* - flush_icache_pages(vma, pg, nr) flushes(invalidates) nr pages for icache
|
|
*
|
|
* Need to doublecheck which one is really needed for ptrace stuff to work.
|
|
*/
|
|
#define LINESIZE 32
|
|
#define LINEBITS 5
|
|
|
|
/*
|
|
* Flush Dcache range through current map.
|
|
*/
|
|
extern void flush_dcache_range(unsigned long start, unsigned long end);
|
|
#define flush_dcache_range flush_dcache_range
|
|
|
|
/*
|
|
* Flush Icache range through current map.
|
|
*/
|
|
extern void flush_icache_range(unsigned long start, unsigned long end);
|
|
#define flush_icache_range flush_icache_range
|
|
|
|
/*
|
|
* Memory-management related flushes are there to ensure in non-physically
|
|
* indexed cache schemes that stale lines belonging to a given ASID aren't
|
|
* in the cache to confuse things. The prototype Hexagon Virtual Machine
|
|
* only uses a single ASID for all user-mode maps, which should
|
|
* mean that they aren't necessary. A brute-force, flush-everything
|
|
* implementation, with the name xxxxx_hexagon() is present in
|
|
* arch/hexagon/mm/cache.c, but let's not wire it up until we know
|
|
* it is needed.
|
|
*/
|
|
extern void flush_cache_all_hexagon(void);
|
|
|
|
/*
|
|
* This may or may not ever have to be non-null, depending on the
|
|
* virtual machine MMU. For a native kernel, it's definitiely a no-op
|
|
*
|
|
* This is also the place where deferred cache coherency stuff seems
|
|
* to happen, classically... but instead we do it like ia64 and
|
|
* clean the cache when the PTE is set.
|
|
*
|
|
*/
|
|
static inline void update_mmu_cache_range(struct vm_fault *vmf,
|
|
struct vm_area_struct *vma, unsigned long address,
|
|
pte_t *ptep, unsigned int nr)
|
|
{
|
|
/* generic_ptrace_pokedata doesn't wind up here, does it? */
|
|
}
|
|
|
|
#define update_mmu_cache(vma, addr, ptep) \
|
|
update_mmu_cache_range(NULL, vma, addr, ptep, 1)
|
|
|
|
void copy_to_user_page(struct vm_area_struct *vma, struct page *page,
|
|
unsigned long vaddr, void *dst, void *src, int len);
|
|
#define copy_to_user_page copy_to_user_page
|
|
|
|
#define copy_from_user_page(vma, page, vaddr, dst, src, len) \
|
|
memcpy(dst, src, len)
|
|
|
|
extern void hexagon_inv_dcache_range(unsigned long start, unsigned long end);
|
|
extern void hexagon_clean_dcache_range(unsigned long start, unsigned long end);
|
|
|
|
#include <asm-generic/cacheflush.h>
|
|
|
|
#endif
|