trampoline_64.S 4.2 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172
  1. /*
  2. *
  3. * Trampoline.S Derived from Setup.S by Linus Torvalds
  4. *
  5. * 4 Jan 1997 Michael Chastain: changed to gnu as.
  6. * 15 Sept 2005 Eric Biederman: 64bit PIC support
  7. *
  8. * Entry: CS:IP point to the start of our code, we are
  9. * in real mode with no stack, but the rest of the
  10. * trampoline page to make our stack and everything else
  11. * is a mystery.
  12. *
  13. * On entry to trampoline_data, the processor is in real mode
  14. * with 16-bit addressing and 16-bit data. CS has some value
  15. * and IP is zero. Thus, data addresses need to be absolute
  16. * (no relocation) and are taken with regard to r_base.
  17. *
  18. * With the addition of trampoline_level4_pgt this code can
  19. * now enter a 64bit kernel that lives at arbitrary 64bit
  20. * physical addresses.
  21. *
  22. * If you work on this file, check the object module with objdump
  23. * --full-contents --reloc to make sure there are no relocation
  24. * entries.
  25. */
  26. #include <linux/linkage.h>
  27. #include <linux/init.h>
  28. #include <asm/pgtable_types.h>
  29. #include <asm/page_types.h>
  30. #include <asm/msr.h>
  31. #include <asm/segment.h>
  32. #include <asm/processor-flags.h>
  33. .section ".x86_trampoline","a"
  34. .balign PAGE_SIZE
  35. .code16
  36. ENTRY(trampoline_data)
  37. r_base = .
  38. cli # We should be safe anyway
  39. wbinvd
  40. mov %cs, %ax # Code and data in the same place
  41. mov %ax, %ds
  42. mov %ax, %es
  43. mov %ax, %ss
  44. movl $0xA5A5A5A5, trampoline_status - r_base
  45. # write marker for master knows we're running
  46. # Setup stack
  47. movw $(trampoline_stack_end - r_base), %sp
  48. call verify_cpu # Verify the cpu supports long mode
  49. testl %eax, %eax # Check for return code
  50. jnz no_longmode
  51. mov %cs, %ax
  52. movzx %ax, %esi # Find the 32bit trampoline location
  53. shll $4, %esi
  54. # Fixup the absolute vectors
  55. leal (startup_32 - r_base)(%esi), %eax
  56. movl %eax, startup_32_vector - r_base
  57. leal (startup_64 - r_base)(%esi), %eax
  58. movl %eax, startup_64_vector - r_base
  59. leal (tgdt - r_base)(%esi), %eax
  60. movl %eax, (tgdt + 2 - r_base)
  61. /*
  62. * GDT tables in non default location kernel can be beyond 16MB and
  63. * lgdt will not be able to load the address as in real mode default
  64. * operand size is 16bit. Use lgdtl instead to force operand size
  65. * to 32 bit.
  66. */
  67. lidtl tidt - r_base # load idt with 0, 0
  68. lgdtl tgdt - r_base # load gdt with whatever is appropriate
  69. mov $X86_CR0_PE, %ax # protected mode (PE) bit
  70. lmsw %ax # into protected mode
  71. # flush prefetch and jump to startup_32
  72. ljmpl *(startup_32_vector - r_base)
  73. .code32
  74. .balign 4
  75. startup_32:
  76. movl $__KERNEL_DS, %eax # Initialize the %ds segment register
  77. movl %eax, %ds
  78. movl $X86_CR4_PAE, %eax
  79. movl %eax, %cr4 # Enable PAE mode
  80. # Setup trampoline 4 level pagetables
  81. leal (trampoline_level4_pgt - r_base)(%esi), %eax
  82. movl %eax, %cr3
  83. movl $MSR_EFER, %ecx
  84. movl $(1 << _EFER_LME), %eax # Enable Long Mode
  85. xorl %edx, %edx
  86. wrmsr
  87. # Enable paging and in turn activate Long Mode
  88. # Enable protected mode
  89. movl $(X86_CR0_PG | X86_CR0_PE), %eax
  90. movl %eax, %cr0
  91. /*
  92. * At this point we're in long mode but in 32bit compatibility mode
  93. * with EFER.LME = 1, CS.L = 0, CS.D = 1 (and in turn
  94. * EFER.LMA = 1). Now we want to jump in 64bit mode, to do that we use
  95. * the new gdt/idt that has __KERNEL_CS with CS.L = 1.
  96. */
  97. ljmp *(startup_64_vector - r_base)(%esi)
  98. .code64
  99. .balign 4
  100. startup_64:
  101. # Now jump into the kernel using virtual addresses
  102. movq $secondary_startup_64, %rax
  103. jmp *%rax
  104. .code16
  105. no_longmode:
  106. hlt
  107. jmp no_longmode
  108. #include "verify_cpu.S"
  109. .balign 4
  110. # Careful these need to be in the same 64K segment as the above;
  111. tidt:
  112. .word 0 # idt limit = 0
  113. .word 0, 0 # idt base = 0L
  114. # Duplicate the global descriptor table
  115. # so the kernel can live anywhere
  116. .balign 4
  117. tgdt:
  118. .short tgdt_end - tgdt # gdt limit
  119. .long tgdt - r_base
  120. .short 0
  121. .quad 0x00cf9b000000ffff # __KERNEL32_CS
  122. .quad 0x00af9b000000ffff # __KERNEL_CS
  123. .quad 0x00cf93000000ffff # __KERNEL_DS
  124. tgdt_end:
  125. .balign 4
  126. startup_32_vector:
  127. .long startup_32 - r_base
  128. .word __KERNEL32_CS, 0
  129. .balign 4
  130. startup_64_vector:
  131. .long startup_64 - r_base
  132. .word __KERNEL_CS, 0
  133. .balign 4
  134. ENTRY(trampoline_status)
  135. .long 0
  136. trampoline_stack:
  137. .org 0x1000
  138. trampoline_stack_end:
  139. ENTRY(trampoline_level4_pgt)
  140. .quad level3_ident_pgt - __START_KERNEL_map + _KERNPG_TABLE
  141. .fill 510,8,0
  142. .quad level3_kernel_pgt - __START_KERNEL_map + _KERNPG_TABLE
  143. ENTRY(trampoline_end)