• swapper_pg_dir的作用


    在内存系统初始化过程中,有如下代码:

       1: static void __init pagetable_init(void)
       2: {
       3:     pgd_t *pgd_base = swapper_pg_dir;
       4:  
       5:     permanent_kmaps_init(pgd_base);
       6: }

    这里,我们看到了神秘的swapper_pg_dir,全局搜索一下,发现了

       1: /*
       2:  * Build a proper pagetable for the kernel mappings.  Up until this
       3:  * point, we've been running on some set of pagetables constructed by
       4:  * the boot process.
       5:  *
       6:  * If we're booting on native hardware, this will be a pagetable
       7:  * constructed in arch/x86/kernel/head_32.S.  The root of the
       8:  * pagetable will be swapper_pg_dir.
       9:  *
      10:  * If we're booting paravirtualized under a hypervisor, then there are
      11:  * more options: we may already be running PAE, and the pagetable may
      12:  * or may not be based in swapper_pg_dir.  In any case,
      13:  * paravirt_pagetable_setup_start() will set up swapper_pg_dir
      14:  * appropriately for the rest of the initialization to work.
      15:  *
      16:  * In general, pagetable_init() assumes that the pagetable may already
      17:  * be partially populated, and so it avoids stomping on any existing
      18:  * mappings.
      19:  */
      20: void __init early_ioremap_page_table_range_init(void)
      21: {
      22:     pgd_t *pgd_base = swapper_pg_dir;
      23:     unsigned long vaddr, end;
      24:  
      25:     /*
      26:      * Fixed mappings, only the page table structure has to be
      27:      * created - mappings will be set by set_fixmap():
      28:      */
      29:     vaddr = __fix_to_virt(__end_of_fixed_addresses - 1) & PMD_MASK;
      30:     end = (FIXADDR_TOP + PMD_SIZE - 1) & PMD_MASK;
      31:     page_table_range_init(vaddr, end, pgd_base);
      32:     early_ioremap_reset();
      33: }


    在head_32.S中,定义了如下的BSS段,BSS段是在内核映像文件中不占空间,但是在内核被加载到内存时,会保留相应的空间。

    在BSS段,一共保留了4个页面的空间,分别用initial_page_table, initial_pg_fixmap, empty_zero_page和swapper_pg_dir来标志其地址。

       1: /*
       2:  * BSS section
       3:  */
       4: __PAGE_ALIGNED_BSS
       5:     .align PAGE_SIZE
       6: #ifdef CONFIG_X86_PAE
       7: initial_pg_pmd:
       8:     .fill 1024*KPMDS,4,0
       9: #else
      10: ENTRY(initial_page_table)
      11:     .fill 1024,4,0
      12: #endif
      13: initial_pg_fixmap:
      14:     .fill 1024,4,0
      15: ENTRY(empty_zero_page)
      16:     .fill 4096,1,0
      17: ENTRY(swapper_pg_dir)
      18:     .fill 1024,4,0

    通过如下代码,将initial_page_table设置为初始页目录

       1: /*
       2:  * Enable paging
       3:  */
       4:     movl $pa(initial_page_table), %eax
       5:     movl %eax,%cr3        /* set the page table pointer.. */
       6:     movl %cr0,%eax
       7:     orl  $X86_CR0_PG,%eax
       8:     movl %eax,%cr0        /* ..and set paging (PG) bit */
       9:     ljmp $__BOOT_CS,$1f    /* Clear prefetch and normalize %eip */

    在内核初始化阶段,setup_arch调用了如下的函数:

       1: void __init setup_arch(char **cmdline_p)
       2: {
       3:  
       4: ......
       5: /* max_pfn_mapped is updated here */
       6: max_low_pfn_mapped = init_memory_mapping(0, max_low_pfn<<PAGE_SHIFT);
       7: max_pfn_mapped = max_low_pfn_mapped;
       8: ......
       9: x86_init.paging.pagetable_setup_start(swapper_pg_dir);
      10: paging_init();
      11: x86_init.paging.pagetable_setup_done(swapper_pg_dir);
      12:  
      13: ......
      14: }

    init_memory_mapping调用了kernel_physical_mapping_init,初始化swapper_pg_dir

       1: /*
       2:  * This maps the physical memory to kernel virtual address space, a total
       3:  * of max_low_pfn pages, by creating page tables starting from address
       4:  * PAGE_OFFSET:
       5:  */
       6: unsigned long __init
       7: kernel_physical_mapping_init(unsigned long start,
       8:                  unsigned long end,
       9:                  unsigned long page_size_mask)
      10: {
      11:     int use_pse = page_size_mask == (1<<PG_LEVEL_2M);
      12:     unsigned long last_map_addr = end;
      13:     unsigned long start_pfn, end_pfn;
      14:     pgd_t *pgd_base = swapper_pg_dir;
      15:     int pgd_idx, pmd_idx, pte_ofs;
      16:     unsigned long pfn;
      17:     pgd_t *pgd;
      18:     pmd_t *pmd;
      19:     pte_t *pte;
      20:     unsigned pages_2m, pages_4k;
      21:     int mapping_iter;
      22:     
      23:     start_pfn = start >> PAGE_SHIFT;
      24:     end_pfn = end >> PAGE_SHIFT;
      25:     
      26:     /*
      27:      * First iteration will setup identity mapping using large/small pages
      28:      * based on use_pse, with other attributes same as set by
      29:      * the early code in head_32.S
      30:      *
      31:      * Second iteration will setup the appropriate attributes (NX, GLOBAL..)
      32:      * as desired for the kernel identity mapping.
      33:      *
      34:      * This two pass mechanism conforms to the TLB app note which says:
      35:      *
      36:      *     "Software should not write to a paging-structure entry in a way
      37:      *      that would change, for any linear address, both the page size
      38:      *      and either the page frame or attributes."
      39:      */
      40:     mapping_iter = 1;
      41:     
      42:     if (!cpu_has_pse)
      43:         use_pse = 0;
      44:     
      45:     at:
      46:     pages_2m = pages_4k = 0;
      47:     pfn = start_pfn;
      48:     pgd_idx = pgd_index((pfn<<PAGE_SHIFT) + PAGE_OFFSET);
      49:     pgd = pgd_base + pgd_idx;
      50:     for (; pgd_idx < PTRS_PER_PGD; pgd++, pgd_idx++) {
      51:         pmd = one_md_table_init(pgd);
      52:     
      53:         if (pfn >= end_pfn)
      54:             continue;
      55:     ef CONFIG_X86_PAE
      56:         pmd_idx = pmd_index((pfn<<PAGE_SHIFT) + PAGE_OFFSET);
      57:         pmd += pmd_idx;
      58:     e
      59:         pmd_idx = 0;
      60:     if
      61:         for (; pmd_idx < PTRS_PER_PMD && pfn < end_pfn;
      62:              pmd++, pmd_idx++) {
      63:             unsigned int addr = pfn * PAGE_SIZE + PAGE_OFFSET;
      64:     
      65:             /*
      66:              * Map with big pages if possible, otherwise
      67:              * create normal page tables:
      68:              */
      69:             if (use_pse) {
      70:                 unsigned int addr2;
      71:                 pgprot_t prot = PAGE_KERNEL_LARGE;
      72:                 /*
      73:                  * first pass will use the same initial
      74:                  * identity mapping attribute + _PAGE_PSE.
      75:                  */
      76:                 pgprot_t init_prot =
      77:                     __pgprot(PTE_IDENT_ATTR |
      78:                          _PAGE_PSE);
      79:     
      80:                 addr2 = (pfn + PTRS_PER_PTE-1) * PAGE_SIZE +
      81:                     PAGE_OFFSET + PAGE_SIZE-1;
      82:     
      83:                 if (is_kernel_text(addr) ||
      84:                     is_kernel_text(addr2))
      85:                     prot = PAGE_KERNEL_LARGE_EXEC;
      86:     
      87:                 pages_2m++;
      88:                 if (mapping_iter == 1)
      89:                     set_pmd(pmd, pfn_pmd(pfn, init_prot));
      90:                 else
      91:                     set_pmd(pmd, pfn_pmd(pfn, prot));
      92:     
      93:                 pfn += PTRS_PER_PTE;
      94:                 continue;
      95:             }
      96:             pte = one_page_table_init(pmd);
      97:     
      98:             pte_ofs = pte_index((pfn<<PAGE_SHIFT) + PAGE_OFFSET);
      99:             pte += pte_ofs;
     100:             for (; pte_ofs < PTRS_PER_PTE && pfn < end_pfn;
     101:                  pte++, pfn++, pte_ofs++, addr += PAGE_SIZE) {
     102:                 pgprot_t prot = PAGE_KERNEL;
     103:                 /*
     104:                  * first pass will use the same initial
     105:                  * identity mapping attribute.
     106:                  */
     107:                 pgprot_t init_prot = __pgprot(PTE_IDENT_ATTR);
     108:     
     109:                 if (is_kernel_text(addr))
     110:                     prot = PAGE_KERNEL_EXEC;
     111:     
     112:                 pages_4k++;
     113:                 if (mapping_iter == 1) {
     114:                     set_pte(pte, pfn_pte(pfn, init_prot));
     115:                     last_map_addr = (pfn << PAGE_SHIFT) + PAGE_SIZE;
     116:                 } else
     117:                     set_pte(pte, pfn_pte(pfn, prot));
     118:             }
     119:         }
     120:     }
     121:     if (mapping_iter == 1) {
     122:         /*
     123:          * update direct mapping page count only in the first
     124:          * iteration.
     125:          */
     126:         update_page_count(PG_LEVEL_2M, pages_2m);
     127:         update_page_count(PG_LEVEL_4K, pages_4k);
     128:     
     129:         /*
     130:          * local global flush tlb, which will flush the previous
     131:          * mappings present in both small and large page TLB's.
     132:          */
     133:         __flush_tlb_all();
     134:     
     135:         /*
     136:          * Second iteration will set the actual desired PTE attributes.
     137:          */
     138:         mapping_iter = 2;
     139:         goto repeat;
     140:     }
     141:     return last_map_addr;
     142: }

    pgd_t *pgd_base = swapper_pg_dir;

    将swapper_pg_dir作为页目录地址,赋给pgd_base

        start_pfn = start >> PAGE_SHIFT;
        end_pfn = end >> PAGE_SHIFT;

    start和end代表着,内核直接映射的虚拟地址区域的开始物理地址和结束物理地址,通过右移PAGE_SHIFT位,再加上PAGE_OFFSET,得到其对应的页表项索引。

        pgd_idx = pgd_index((pfn<<PAGE_SHIFT) + PAGE_OFFSET);
        pgd = pgd_base + pgd_idx;

    pgd_idx,pgd代表着在页目录中的索引,以及相应的页目录项

       1: pgprot_t prot = PAGE_KERNEL;
       2: /*
       3:  * first pass will use the same initial
       4:  * identity mapping attribute.
       5:  */
       6: pgprot_t init_prot = __pgprot(PTE_IDENT_ATTR);
       7:  
       8: if (is_kernel_text(addr))
       9:     prot = PAGE_KERNEL_EXEC;
      10:  
      11: pages_4k++;
      12: if (mapping_iter == 1) {
      13:     set_pte(pte, pfn_pte(pfn, init_prot));
      14:     last_map_addr = (pfn << PAGE_SHIFT) + PAGE_SIZE;
      15: } else
      16:     set_pte(pte, pfn_pte(pfn, prot));

    最后,通过两个回合的遍历,将属性设置到对应的页表项上去。

  • 相关阅读:
    使用ACEGI搭建权限系统:第三部分
    分支在版本树中的应用(使用subversion)
    acegi安全框架使用:第二部分
    错误数据导致java.lang.IllegalArgumentException:Unsupported configuration attributes
    移动中间件和wap网关的比较
    3年后,又回到了.net阵营
    android中listView的几点总结
    oracle相关分布式数据解决方案
    ajax实现用户名存在校验
    使用template method模式简化android列表页面
  • 原文地址:https://www.cnblogs.com/long123king/p/3510297.html
Copyright © 2020-2023  润新知