|  | /* | 
|  | * Copyright 2010 Tilera Corporation. All Rights Reserved. | 
|  | * | 
|  | *   This program is free software; you can redistribute it and/or | 
|  | *   modify it under the terms of the GNU General Public License | 
|  | *   as published by the Free Software Foundation, version 2. | 
|  | * | 
|  | *   This program is distributed in the hope that it will be useful, but | 
|  | *   WITHOUT ANY WARRANTY; without even the implied warranty of | 
|  | *   MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE, GOOD TITLE or | 
|  | *   NON INFRINGEMENT.  See the GNU General Public License for | 
|  | *   more details. | 
|  | * | 
|  | */ | 
|  |  | 
|  | #include <linux/cpumask.h> | 
|  | #include <linux/module.h> | 
|  | #include <linux/hugetlb.h> | 
|  | #include <asm/tlbflush.h> | 
|  | #include <asm/homecache.h> | 
|  | #include <hv/hypervisor.h> | 
|  |  | 
|  | /* From tlbflush.h */ | 
|  | DEFINE_PER_CPU(int, current_asid); | 
|  | int min_asid, max_asid; | 
|  |  | 
|  | /* | 
|  | * Note that we flush the L1I (for VM_EXEC pages) as well as the TLB | 
|  | * so that when we are unmapping an executable page, we also flush it. | 
|  | * Combined with flushing the L1I at context switch time, this means | 
|  | * we don't have to do any other icache flushes. | 
|  | */ | 
|  |  | 
|  | void flush_tlb_mm(struct mm_struct *mm) | 
|  | { | 
|  | HV_Remote_ASID asids[NR_CPUS]; | 
|  | int i = 0, cpu; | 
|  | for_each_cpu(cpu, mm_cpumask(mm)) { | 
|  | HV_Remote_ASID *asid = &asids[i++]; | 
|  | asid->y = cpu / smp_topology.width; | 
|  | asid->x = cpu % smp_topology.width; | 
|  | asid->asid = per_cpu(current_asid, cpu); | 
|  | } | 
|  | flush_remote(0, HV_FLUSH_EVICT_L1I, mm_cpumask(mm), | 
|  | 0, 0, 0, NULL, asids, i); | 
|  | } | 
|  |  | 
|  | void flush_tlb_current_task(void) | 
|  | { | 
|  | flush_tlb_mm(current->mm); | 
|  | } | 
|  |  | 
|  | void flush_tlb_page_mm(struct vm_area_struct *vma, struct mm_struct *mm, | 
|  | unsigned long va) | 
|  | { | 
|  | unsigned long size = vma_kernel_pagesize(vma); | 
|  | int cache = (vma->vm_flags & VM_EXEC) ? HV_FLUSH_EVICT_L1I : 0; | 
|  | flush_remote(0, cache, mm_cpumask(mm), | 
|  | va, size, size, mm_cpumask(mm), NULL, 0); | 
|  | } | 
|  |  | 
|  | void flush_tlb_page(struct vm_area_struct *vma, unsigned long va) | 
|  | { | 
|  | flush_tlb_page_mm(vma, vma->vm_mm, va); | 
|  | } | 
|  | EXPORT_SYMBOL(flush_tlb_page); | 
|  |  | 
|  | void flush_tlb_range(struct vm_area_struct *vma, | 
|  | unsigned long start, unsigned long end) | 
|  | { | 
|  | unsigned long size = vma_kernel_pagesize(vma); | 
|  | struct mm_struct *mm = vma->vm_mm; | 
|  | int cache = (vma->vm_flags & VM_EXEC) ? HV_FLUSH_EVICT_L1I : 0; | 
|  | flush_remote(0, cache, mm_cpumask(mm), start, end - start, size, | 
|  | mm_cpumask(mm), NULL, 0); | 
|  | } | 
|  |  | 
|  | void flush_tlb_all(void) | 
|  | { | 
|  | int i; | 
|  | for (i = 0; ; ++i) { | 
|  | HV_VirtAddrRange r = hv_inquire_virtual(i); | 
|  | if (r.size == 0) | 
|  | break; | 
|  | flush_remote(0, HV_FLUSH_EVICT_L1I, cpu_online_mask, | 
|  | r.start, r.size, PAGE_SIZE, cpu_online_mask, | 
|  | NULL, 0); | 
|  | flush_remote(0, 0, NULL, | 
|  | r.start, r.size, HPAGE_SIZE, cpu_online_mask, | 
|  | NULL, 0); | 
|  | } | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Callers need to flush the L1I themselves if necessary, e.g. for | 
|  | * kernel module unload.  Otherwise we assume callers are not using | 
|  | * executable pgprot_t's.  Using EVICT_L1I means that dataplane cpus | 
|  | * will get an unnecessary interrupt otherwise. | 
|  | */ | 
|  | void flush_tlb_kernel_range(unsigned long start, unsigned long end) | 
|  | { | 
|  | flush_remote(0, 0, NULL, | 
|  | start, end - start, PAGE_SIZE, cpu_online_mask, NULL, 0); | 
|  | } |