kmemcheck.c 14 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654
  1. /**
  2. * kmemcheck - a heavyweight memory checker for the linux kernel
  3. * Copyright (C) 2007, 2008 Vegard Nossum <vegardno@ifi.uio.no>
  4. * (With a lot of help from Ingo Molnar and Pekka Enberg.)
  5. *
  6. * This program is free software; you can redistribute it and/or modify
  7. * it under the terms of the GNU General Public License (version 2) as
  8. * published by the Free Software Foundation.
  9. */
  10. #include <linux/init.h>
  11. #include <linux/interrupt.h>
  12. #include <linux/kallsyms.h>
  13. #include <linux/kernel.h>
  14. #include <linux/kmemcheck.h>
  15. #include <linux/mm.h>
  16. #include <linux/module.h>
  17. #include <linux/page-flags.h>
  18. #include <linux/percpu.h>
  19. #include <linux/ptrace.h>
  20. #include <linux/string.h>
  21. #include <linux/types.h>
  22. #include <asm/cacheflush.h>
  23. #include <asm/kmemcheck.h>
  24. #include <asm/pgtable.h>
  25. #include <asm/tlbflush.h>
  26. #include "error.h"
  27. #include "opcode.h"
  28. #include "pte.h"
  29. #include "selftest.h"
  30. #include "shadow.h"
  31. #ifdef CONFIG_KMEMCHECK_DISABLED_BY_DEFAULT
  32. # define KMEMCHECK_ENABLED 0
  33. #endif
  34. #ifdef CONFIG_KMEMCHECK_ENABLED_BY_DEFAULT
  35. # define KMEMCHECK_ENABLED 1
  36. #endif
  37. #ifdef CONFIG_KMEMCHECK_ONESHOT_BY_DEFAULT
  38. # define KMEMCHECK_ENABLED 2
  39. #endif
  40. int kmemcheck_enabled = KMEMCHECK_ENABLED;
  41. int __init kmemcheck_init(void)
  42. {
  43. #ifdef CONFIG_SMP
  44. /*
  45. * Limit SMP to use a single CPU. We rely on the fact that this code
  46. * runs before SMP is set up.
  47. */
  48. if (setup_max_cpus > 1) {
  49. printk(KERN_INFO
  50. "kmemcheck: Limiting number of CPUs to 1.\n");
  51. setup_max_cpus = 1;
  52. }
  53. #endif
  54. if (!kmemcheck_selftest()) {
  55. printk(KERN_INFO "kmemcheck: self-tests failed; disabling\n");
  56. kmemcheck_enabled = 0;
  57. return -EINVAL;
  58. }
  59. printk(KERN_INFO "kmemcheck: Initialized\n");
  60. return 0;
  61. }
  62. early_initcall(kmemcheck_init);
  63. /*
  64. * We need to parse the kmemcheck= option before any memory is allocated.
  65. */
  66. static int __init param_kmemcheck(char *str)
  67. {
  68. if (!str)
  69. return -EINVAL;
  70. sscanf(str, "%d", &kmemcheck_enabled);
  71. return 0;
  72. }
  73. early_param("kmemcheck", param_kmemcheck);
  74. int kmemcheck_show_addr(unsigned long address)
  75. {
  76. pte_t *pte;
  77. pte = kmemcheck_pte_lookup(address);
  78. if (!pte)
  79. return 0;
  80. set_pte(pte, __pte(pte_val(*pte) | _PAGE_PRESENT));
  81. __flush_tlb_one(address);
  82. return 1;
  83. }
  84. int kmemcheck_hide_addr(unsigned long address)
  85. {
  86. pte_t *pte;
  87. pte = kmemcheck_pte_lookup(address);
  88. if (!pte)
  89. return 0;
  90. set_pte(pte, __pte(pte_val(*pte) & ~_PAGE_PRESENT));
  91. __flush_tlb_one(address);
  92. return 1;
  93. }
  94. struct kmemcheck_context {
  95. bool busy;
  96. int balance;
  97. /*
  98. * There can be at most two memory operands to an instruction, but
  99. * each address can cross a page boundary -- so we may need up to
  100. * four addresses that must be hidden/revealed for each fault.
  101. */
  102. unsigned long addr[4];
  103. unsigned long n_addrs;
  104. unsigned long flags;
  105. /* Data size of the instruction that caused a fault. */
  106. unsigned int size;
  107. };
  108. static DEFINE_PER_CPU(struct kmemcheck_context, kmemcheck_context);
  109. bool kmemcheck_active(struct pt_regs *regs)
  110. {
  111. struct kmemcheck_context *data = &__get_cpu_var(kmemcheck_context);
  112. return data->balance > 0;
  113. }
  114. /* Save an address that needs to be shown/hidden */
  115. static void kmemcheck_save_addr(unsigned long addr)
  116. {
  117. struct kmemcheck_context *data = &__get_cpu_var(kmemcheck_context);
  118. BUG_ON(data->n_addrs >= ARRAY_SIZE(data->addr));
  119. data->addr[data->n_addrs++] = addr;
  120. }
  121. static unsigned int kmemcheck_show_all(void)
  122. {
  123. struct kmemcheck_context *data = &__get_cpu_var(kmemcheck_context);
  124. unsigned int i;
  125. unsigned int n;
  126. n = 0;
  127. for (i = 0; i < data->n_addrs; ++i)
  128. n += kmemcheck_show_addr(data->addr[i]);
  129. return n;
  130. }
  131. static unsigned int kmemcheck_hide_all(void)
  132. {
  133. struct kmemcheck_context *data = &__get_cpu_var(kmemcheck_context);
  134. unsigned int i;
  135. unsigned int n;
  136. n = 0;
  137. for (i = 0; i < data->n_addrs; ++i)
  138. n += kmemcheck_hide_addr(data->addr[i]);
  139. return n;
  140. }
  141. /*
  142. * Called from the #PF handler.
  143. */
  144. void kmemcheck_show(struct pt_regs *regs)
  145. {
  146. struct kmemcheck_context *data = &__get_cpu_var(kmemcheck_context);
  147. BUG_ON(!irqs_disabled());
  148. if (unlikely(data->balance != 0)) {
  149. kmemcheck_show_all();
  150. kmemcheck_error_save_bug(regs);
  151. data->balance = 0;
  152. return;
  153. }
  154. /*
  155. * None of the addresses actually belonged to kmemcheck. Note that
  156. * this is not an error.
  157. */
  158. if (kmemcheck_show_all() == 0)
  159. return;
  160. ++data->balance;
  161. /*
  162. * The IF needs to be cleared as well, so that the faulting
  163. * instruction can run "uninterrupted". Otherwise, we might take
  164. * an interrupt and start executing that before we've had a chance
  165. * to hide the page again.
  166. *
  167. * NOTE: In the rare case of multiple faults, we must not override
  168. * the original flags:
  169. */
  170. if (!(regs->flags & X86_EFLAGS_TF))
  171. data->flags = regs->flags;
  172. regs->flags |= X86_EFLAGS_TF;
  173. regs->flags &= ~X86_EFLAGS_IF;
  174. }
  175. /*
  176. * Called from the #DB handler.
  177. */
  178. void kmemcheck_hide(struct pt_regs *regs)
  179. {
  180. struct kmemcheck_context *data = &__get_cpu_var(kmemcheck_context);
  181. int n;
  182. BUG_ON(!irqs_disabled());
  183. if (unlikely(data->balance != 1)) {
  184. kmemcheck_show_all();
  185. kmemcheck_error_save_bug(regs);
  186. data->n_addrs = 0;
  187. data->balance = 0;
  188. if (!(data->flags & X86_EFLAGS_TF))
  189. regs->flags &= ~X86_EFLAGS_TF;
  190. if (data->flags & X86_EFLAGS_IF)
  191. regs->flags |= X86_EFLAGS_IF;
  192. return;
  193. }
  194. if (kmemcheck_enabled)
  195. n = kmemcheck_hide_all();
  196. else
  197. n = kmemcheck_show_all();
  198. if (n == 0)
  199. return;
  200. --data->balance;
  201. data->n_addrs = 0;
  202. if (!(data->flags & X86_EFLAGS_TF))
  203. regs->flags &= ~X86_EFLAGS_TF;
  204. if (data->flags & X86_EFLAGS_IF)
  205. regs->flags |= X86_EFLAGS_IF;
  206. }
  207. void kmemcheck_show_pages(struct page *p, unsigned int n)
  208. {
  209. unsigned int i;
  210. for (i = 0; i < n; ++i) {
  211. unsigned long address;
  212. pte_t *pte;
  213. unsigned int level;
  214. address = (unsigned long) page_address(&p[i]);
  215. pte = lookup_address(address, &level);
  216. BUG_ON(!pte);
  217. BUG_ON(level != PG_LEVEL_4K);
  218. set_pte(pte, __pte(pte_val(*pte) | _PAGE_PRESENT));
  219. set_pte(pte, __pte(pte_val(*pte) & ~_PAGE_HIDDEN));
  220. __flush_tlb_one(address);
  221. }
  222. }
  223. bool kmemcheck_page_is_tracked(struct page *p)
  224. {
  225. /* This will also check the "hidden" flag of the PTE. */
  226. return kmemcheck_pte_lookup((unsigned long) page_address(p));
  227. }
  228. void kmemcheck_hide_pages(struct page *p, unsigned int n)
  229. {
  230. unsigned int i;
  231. for (i = 0; i < n; ++i) {
  232. unsigned long address;
  233. pte_t *pte;
  234. unsigned int level;
  235. address = (unsigned long) page_address(&p[i]);
  236. pte = lookup_address(address, &level);
  237. BUG_ON(!pte);
  238. BUG_ON(level != PG_LEVEL_4K);
  239. set_pte(pte, __pte(pte_val(*pte) & ~_PAGE_PRESENT));
  240. set_pte(pte, __pte(pte_val(*pte) | _PAGE_HIDDEN));
  241. __flush_tlb_one(address);
  242. }
  243. }
  244. /* Access may NOT cross page boundary */
  245. static void kmemcheck_read_strict(struct pt_regs *regs,
  246. unsigned long addr, unsigned int size)
  247. {
  248. void *shadow;
  249. enum kmemcheck_shadow status;
  250. shadow = kmemcheck_shadow_lookup(addr);
  251. if (!shadow)
  252. return;
  253. kmemcheck_save_addr(addr);
  254. status = kmemcheck_shadow_test(shadow, size);
  255. if (status == KMEMCHECK_SHADOW_INITIALIZED)
  256. return;
  257. if (kmemcheck_enabled)
  258. kmemcheck_error_save(status, addr, size, regs);
  259. if (kmemcheck_enabled == 2)
  260. kmemcheck_enabled = 0;
  261. /* Don't warn about it again. */
  262. kmemcheck_shadow_set(shadow, size);
  263. }
  264. bool kmemcheck_is_obj_initialized(unsigned long addr, size_t size)
  265. {
  266. enum kmemcheck_shadow status;
  267. void *shadow;
  268. shadow = kmemcheck_shadow_lookup(addr);
  269. if (!shadow)
  270. return true;
  271. status = kmemcheck_shadow_test_all(shadow, size);
  272. return status == KMEMCHECK_SHADOW_INITIALIZED;
  273. }
  274. /* Access may cross page boundary */
  275. static void kmemcheck_read(struct pt_regs *regs,
  276. unsigned long addr, unsigned int size)
  277. {
  278. unsigned long page = addr & PAGE_MASK;
  279. unsigned long next_addr = addr + size - 1;
  280. unsigned long next_page = next_addr & PAGE_MASK;
  281. if (likely(page == next_page)) {
  282. kmemcheck_read_strict(regs, addr, size);
  283. return;
  284. }
  285. /*
  286. * What we do is basically to split the access across the
  287. * two pages and handle each part separately. Yes, this means
  288. * that we may now see reads that are 3 + 5 bytes, for
  289. * example (and if both are uninitialized, there will be two
  290. * reports), but it makes the code a lot simpler.
  291. */
  292. kmemcheck_read_strict(regs, addr, next_page - addr);
  293. kmemcheck_read_strict(regs, next_page, next_addr - next_page);
  294. }
  295. static void kmemcheck_write_strict(struct pt_regs *regs,
  296. unsigned long addr, unsigned int size)
  297. {
  298. void *shadow;
  299. shadow = kmemcheck_shadow_lookup(addr);
  300. if (!shadow)
  301. return;
  302. kmemcheck_save_addr(addr);
  303. kmemcheck_shadow_set(shadow, size);
  304. }
  305. static void kmemcheck_write(struct pt_regs *regs,
  306. unsigned long addr, unsigned int size)
  307. {
  308. unsigned long page = addr & PAGE_MASK;
  309. unsigned long next_addr = addr + size - 1;
  310. unsigned long next_page = next_addr & PAGE_MASK;
  311. if (likely(page == next_page)) {
  312. kmemcheck_write_strict(regs, addr, size);
  313. return;
  314. }
  315. /* See comment in kmemcheck_read(). */
  316. kmemcheck_write_strict(regs, addr, next_page - addr);
  317. kmemcheck_write_strict(regs, next_page, next_addr - next_page);
  318. }
  319. /*
  320. * Copying is hard. We have two addresses, each of which may be split across
  321. * a page (and each page will have different shadow addresses).
  322. */
  323. static void kmemcheck_copy(struct pt_regs *regs,
  324. unsigned long src_addr, unsigned long dst_addr, unsigned int size)
  325. {
  326. uint8_t shadow[8];
  327. enum kmemcheck_shadow status;
  328. unsigned long page;
  329. unsigned long next_addr;
  330. unsigned long next_page;
  331. uint8_t *x;
  332. unsigned int i;
  333. unsigned int n;
  334. BUG_ON(size > sizeof(shadow));
  335. page = src_addr & PAGE_MASK;
  336. next_addr = src_addr + size - 1;
  337. next_page = next_addr & PAGE_MASK;
  338. if (likely(page == next_page)) {
  339. /* Same page */
  340. x = kmemcheck_shadow_lookup(src_addr);
  341. if (x) {
  342. kmemcheck_save_addr(src_addr);
  343. for (i = 0; i < size; ++i)
  344. shadow[i] = x[i];
  345. } else {
  346. for (i = 0; i < size; ++i)
  347. shadow[i] = KMEMCHECK_SHADOW_INITIALIZED;
  348. }
  349. } else {
  350. n = next_page - src_addr;
  351. BUG_ON(n > sizeof(shadow));
  352. /* First page */
  353. x = kmemcheck_shadow_lookup(src_addr);
  354. if (x) {
  355. kmemcheck_save_addr(src_addr);
  356. for (i = 0; i < n; ++i)
  357. shadow[i] = x[i];
  358. } else {
  359. /* Not tracked */
  360. for (i = 0; i < n; ++i)
  361. shadow[i] = KMEMCHECK_SHADOW_INITIALIZED;
  362. }
  363. /* Second page */
  364. x = kmemcheck_shadow_lookup(next_page);
  365. if (x) {
  366. kmemcheck_save_addr(next_page);
  367. for (i = n; i < size; ++i)
  368. shadow[i] = x[i - n];
  369. } else {
  370. /* Not tracked */
  371. for (i = n; i < size; ++i)
  372. shadow[i] = KMEMCHECK_SHADOW_INITIALIZED;
  373. }
  374. }
  375. page = dst_addr & PAGE_MASK;
  376. next_addr = dst_addr + size - 1;
  377. next_page = next_addr & PAGE_MASK;
  378. if (likely(page == next_page)) {
  379. /* Same page */
  380. x = kmemcheck_shadow_lookup(dst_addr);
  381. if (x) {
  382. kmemcheck_save_addr(dst_addr);
  383. for (i = 0; i < size; ++i) {
  384. x[i] = shadow[i];
  385. shadow[i] = KMEMCHECK_SHADOW_INITIALIZED;
  386. }
  387. }
  388. } else {
  389. n = next_page - dst_addr;
  390. BUG_ON(n > sizeof(shadow));
  391. /* First page */
  392. x = kmemcheck_shadow_lookup(dst_addr);
  393. if (x) {
  394. kmemcheck_save_addr(dst_addr);
  395. for (i = 0; i < n; ++i) {
  396. x[i] = shadow[i];
  397. shadow[i] = KMEMCHECK_SHADOW_INITIALIZED;
  398. }
  399. }
  400. /* Second page */
  401. x = kmemcheck_shadow_lookup(next_page);
  402. if (x) {
  403. kmemcheck_save_addr(next_page);
  404. for (i = n; i < size; ++i) {
  405. x[i - n] = shadow[i];
  406. shadow[i] = KMEMCHECK_SHADOW_INITIALIZED;
  407. }
  408. }
  409. }
  410. status = kmemcheck_shadow_test(shadow, size);
  411. if (status == KMEMCHECK_SHADOW_INITIALIZED)
  412. return;
  413. if (kmemcheck_enabled)
  414. kmemcheck_error_save(status, src_addr, size, regs);
  415. if (kmemcheck_enabled == 2)
  416. kmemcheck_enabled = 0;
  417. }
  418. enum kmemcheck_method {
  419. KMEMCHECK_READ,
  420. KMEMCHECK_WRITE,
  421. };
  422. static void kmemcheck_access(struct pt_regs *regs,
  423. unsigned long fallback_address, enum kmemcheck_method fallback_method)
  424. {
  425. const uint8_t *insn;
  426. const uint8_t *insn_primary;
  427. unsigned int size;
  428. struct kmemcheck_context *data = &__get_cpu_var(kmemcheck_context);
  429. /* Recursive fault -- ouch. */
  430. if (data->busy) {
  431. kmemcheck_show_addr(fallback_address);
  432. kmemcheck_error_save_bug(regs);
  433. return;
  434. }
  435. data->busy = true;
  436. insn = (const uint8_t *) regs->ip;
  437. insn_primary = kmemcheck_opcode_get_primary(insn);
  438. kmemcheck_opcode_decode(insn, &size);
  439. switch (insn_primary[0]) {
  440. #ifdef CONFIG_KMEMCHECK_BITOPS_OK
  441. /* AND, OR, XOR */
  442. /*
  443. * Unfortunately, these instructions have to be excluded from
  444. * our regular checking since they access only some (and not
  445. * all) bits. This clears out "bogus" bitfield-access warnings.
  446. */
  447. case 0x80:
  448. case 0x81:
  449. case 0x82:
  450. case 0x83:
  451. switch ((insn_primary[1] >> 3) & 7) {
  452. /* OR */
  453. case 1:
  454. /* AND */
  455. case 4:
  456. /* XOR */
  457. case 6:
  458. kmemcheck_write(regs, fallback_address, size);
  459. goto out;
  460. /* ADD */
  461. case 0:
  462. /* ADC */
  463. case 2:
  464. /* SBB */
  465. case 3:
  466. /* SUB */
  467. case 5:
  468. /* CMP */
  469. case 7:
  470. break;
  471. }
  472. break;
  473. #endif
  474. /* MOVS, MOVSB, MOVSW, MOVSD */
  475. case 0xa4:
  476. case 0xa5:
  477. /*
  478. * These instructions are special because they take two
  479. * addresses, but we only get one page fault.
  480. */
  481. kmemcheck_copy(regs, regs->si, regs->di, size);
  482. goto out;
  483. /* CMPS, CMPSB, CMPSW, CMPSD */
  484. case 0xa6:
  485. case 0xa7:
  486. kmemcheck_read(regs, regs->si, size);
  487. kmemcheck_read(regs, regs->di, size);
  488. goto out;
  489. }
  490. /*
  491. * If the opcode isn't special in any way, we use the data from the
  492. * page fault handler to determine the address and type of memory
  493. * access.
  494. */
  495. switch (fallback_method) {
  496. case KMEMCHECK_READ:
  497. kmemcheck_read(regs, fallback_address, size);
  498. goto out;
  499. case KMEMCHECK_WRITE:
  500. kmemcheck_write(regs, fallback_address, size);
  501. goto out;
  502. }
  503. out:
  504. data->busy = false;
  505. }
  506. bool kmemcheck_fault(struct pt_regs *regs, unsigned long address,
  507. unsigned long error_code)
  508. {
  509. pte_t *pte;
  510. /*
  511. * XXX: Is it safe to assume that memory accesses from virtual 86
  512. * mode or non-kernel code segments will _never_ access kernel
  513. * memory (e.g. tracked pages)? For now, we need this to avoid
  514. * invoking kmemcheck for PnP BIOS calls.
  515. */
  516. if (regs->flags & X86_VM_MASK)
  517. return false;
  518. if (regs->cs != __KERNEL_CS)
  519. return false;
  520. pte = kmemcheck_pte_lookup(address);
  521. if (!pte)
  522. return false;
  523. WARN_ON_ONCE(in_nmi());
  524. if (error_code & 2)
  525. kmemcheck_access(regs, address, KMEMCHECK_WRITE);
  526. else
  527. kmemcheck_access(regs, address, KMEMCHECK_READ);
  528. kmemcheck_show(regs);
  529. return true;
  530. }
  531. bool kmemcheck_trap(struct pt_regs *regs)
  532. {
  533. if (!kmemcheck_active(regs))
  534. return false;
  535. /* We're done. */
  536. kmemcheck_hide(regs);
  537. return true;
  538. }