tlb.c 2.8 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798
  1. /*
  2. * Copyright 2010 Tilera Corporation. All Rights Reserved.
  3. *
  4. * This program is free software; you can redistribute it and/or
  5. * modify it under the terms of the GNU General Public License
  6. * as published by the Free Software Foundation, version 2.
  7. *
  8. * This program is distributed in the hope that it will be useful, but
  9. * WITHOUT ANY WARRANTY; without even the implied warranty of
  10. * MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE, GOOD TITLE or
  11. * NON INFRINGEMENT. See the GNU General Public License for
  12. * more details.
  13. *
  14. */
  15. #include <linux/cpumask.h>
  16. #include <linux/module.h>
  17. #include <asm/tlbflush.h>
  18. #include <asm/homecache.h>
  19. #include <hv/hypervisor.h>
  20. /* From tlbflush.h */
  21. DEFINE_PER_CPU(int, current_asid);
  22. int min_asid, max_asid;
  23. /*
  24. * Note that we flush the L1I (for VM_EXEC pages) as well as the TLB
  25. * so that when we are unmapping an executable page, we also flush it.
  26. * Combined with flushing the L1I at context switch time, this means
  27. * we don't have to do any other icache flushes.
  28. */
  29. void flush_tlb_mm(struct mm_struct *mm)
  30. {
  31. HV_Remote_ASID asids[NR_CPUS];
  32. int i = 0, cpu;
  33. for_each_cpu(cpu, mm_cpumask(mm)) {
  34. HV_Remote_ASID *asid = &asids[i++];
  35. asid->y = cpu / smp_topology.width;
  36. asid->x = cpu % smp_topology.width;
  37. asid->asid = per_cpu(current_asid, cpu);
  38. }
  39. flush_remote(0, HV_FLUSH_EVICT_L1I, mm_cpumask(mm),
  40. 0, 0, 0, NULL, asids, i);
  41. }
  42. void flush_tlb_current_task(void)
  43. {
  44. flush_tlb_mm(current->mm);
  45. }
  46. void flush_tlb_page_mm(const struct vm_area_struct *vma, struct mm_struct *mm,
  47. unsigned long va)
  48. {
  49. unsigned long size = hv_page_size(vma);
  50. int cache = (vma->vm_flags & VM_EXEC) ? HV_FLUSH_EVICT_L1I : 0;
  51. flush_remote(0, cache, mm_cpumask(mm),
  52. va, size, size, mm_cpumask(mm), NULL, 0);
  53. }
  54. void flush_tlb_page(const struct vm_area_struct *vma, unsigned long va)
  55. {
  56. flush_tlb_page_mm(vma, vma->vm_mm, va);
  57. }
  58. EXPORT_SYMBOL(flush_tlb_page);
  59. void flush_tlb_range(const struct vm_area_struct *vma,
  60. unsigned long start, unsigned long end)
  61. {
  62. unsigned long size = hv_page_size(vma);
  63. struct mm_struct *mm = vma->vm_mm;
  64. int cache = (vma->vm_flags & VM_EXEC) ? HV_FLUSH_EVICT_L1I : 0;
  65. flush_remote(0, cache, mm_cpumask(mm), start, end - start, size,
  66. mm_cpumask(mm), NULL, 0);
  67. }
  68. void flush_tlb_all(void)
  69. {
  70. int i;
  71. for (i = 0; ; ++i) {
  72. HV_VirtAddrRange r = hv_inquire_virtual(i);
  73. if (r.size == 0)
  74. break;
  75. flush_remote(0, HV_FLUSH_EVICT_L1I, cpu_online_mask,
  76. r.start, r.size, PAGE_SIZE, cpu_online_mask,
  77. NULL, 0);
  78. flush_remote(0, 0, NULL,
  79. r.start, r.size, HPAGE_SIZE, cpu_online_mask,
  80. NULL, 0);
  81. }
  82. }
  83. void flush_tlb_kernel_range(unsigned long start, unsigned long end)
  84. {
  85. flush_remote(0, HV_FLUSH_EVICT_L1I, cpu_online_mask,
  86. start, end - start, PAGE_SIZE, cpu_online_mask, NULL, 0);
  87. }