Merge branch 'for-rmk' of git://git.kernel.org/pub/scm/linux/kernel/git/cmarinas...
authorRussell King <rmk+kernel@arm.linux.org.uk>
Thu, 8 Dec 2011 18:02:04 +0000 (18:02 +0000)
committerRussell King <rmk+kernel@arm.linux.org.uk>
Thu, 8 Dec 2011 18:02:04 +0000 (18:02 +0000)
Conflicts:
arch/arm/mm/ioremap.c

1  2 
arch/arm/Kconfig
arch/arm/include/asm/pgtable.h
arch/arm/kernel/head.S
arch/arm/mm/ioremap.c
arch/arm/mm/mmu.c

Simple merge
index a784859cc7a93d671c9c8578ec922908750fa2e0,3ddcf66f5e04ee4089f73504790d08217c6dfa1b..3f2f0eb7621112daf7f7bf93a2f12d22582f9de3
  
  #else
  
+ #include <asm-generic/pgtable-nopud.h>
  #include <asm/memory.h>
 -#include <mach/vmalloc.h>
  #include <asm/pgtable-hwdef.h>
  
+ #ifdef CONFIG_ARM_LPAE
+ #include <asm/pgtable-3level.h>
+ #else
  #include <asm/pgtable-2level.h>
+ #endif
  
  /*
   * Just any arbitrary offset to the start of the vmalloc VM area: the
Simple merge
index 12c7ad215ce732a743475a75a8c31f6dd3e7f5a9,d1f78bacb015a9756214d704d62ec694ba687f76..80632e8d7538f33d5a6539df5a72a90fff52e7ec
@@@ -194,7 -208,14 +202,8 @@@ void __iomem * __arm_ioremap_pfn_caller
         */
        if (pfn >= 0x100000 && (__pfn_to_phys(pfn) & ~SUPERSECTION_MASK))
                return NULL;
+ #endif
  
 -      /*
 -       * Don't allow RAM to be mapped - this causes problems with ARMv6+
 -       */
 -      if (WARN_ON(pfn_valid(pfn)))
 -              return NULL;
 -
        type = get_mem_type(mtype);
        if (!type)
                return NULL;
@@@ -329,34 -322,28 +338,34 @@@ __arm_ioremap_exec(unsigned long phys_a
  void __iounmap(volatile void __iomem *io_addr)
  {
        void *addr = (void *)(PAGE_MASK & (unsigned long)io_addr);
 -#if !defined(CONFIG_SMP) && !defined(CONFIG_ARM_LPAE)
 -      struct vm_struct **p, *tmp;
 +      struct vm_struct *vm;
  
 -      /*
 -       * If this is a section based mapping we need to handle it
 -       * specially as the VM subsystem does not know how to handle
 -       * such a beast. We need the lock here b/c we need to clear
 -       * all the mappings before the area can be reclaimed
 -       * by someone else.
 -       */
 -      write_lock(&vmlist_lock);
 -      for (p = &vmlist ; (tmp = *p) ; p = &tmp->next) {
 -              if ((tmp->flags & VM_IOREMAP) && (tmp->addr == addr)) {
 -                      if (tmp->flags & VM_ARM_SECTION_MAPPING) {
 -                              unmap_area_sections((unsigned long)tmp->addr,
 -                                                  tmp->size);
 -                      }
 +      read_lock(&vmlist_lock);
 +      for (vm = vmlist; vm; vm = vm->next) {
 +              if (vm->addr > addr)
 +                      break;
 +              if (!(vm->flags & VM_IOREMAP))
 +                      continue;
 +              /* If this is a static mapping we must leave it alone */
 +              if ((vm->flags & VM_ARM_STATIC_MAPPING) &&
 +                  (vm->addr <= addr) && (vm->addr + vm->size > addr)) {
 +                      read_unlock(&vmlist_lock);
 +                      return;
 +              }
- #ifndef CONFIG_SMP
++#if !defined(CONFIG_SMP) && !defined(CONFIG_ARM_LPAE)
 +              /*
 +               * If this is a section based mapping we need to handle it
 +               * specially as the VM subsystem does not know how to handle
 +               * such a beast.
 +               */
 +              if ((vm->addr == addr) &&
 +                  (vm->flags & VM_ARM_SECTION_MAPPING)) {
 +                      unmap_area_sections((unsigned long)vm->addr, vm->size);
                        break;
                }
 -      }
 -      write_unlock(&vmlist_lock);
  #endif
 +      }
 +      read_unlock(&vmlist_lock);
  
        vunmap(addr);
  }
Simple merge