summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorAndrew Morton <akpm@digeo.com>2003-02-03 16:59:30 -0800
committerLinus Torvalds <torvalds@home.transmeta.com>2003-02-03 16:59:30 -0800
commitcecee7397a165fb2c952b0738ec008768a4d6081 (patch)
treef8c0bbda831a80fe0c1e45d662152610f7ff2ef5
parent6b2ca90b290e62797e91a51df04d26f524545aab (diff)
[PATCH] mm/mmap.c whitespace cleanups
- Don't require a 160-col xterm - Coding style consistency
-rw-r--r--include/linux/mm.h7
-rw-r--r--mm/mmap.c95
2 files changed, 66 insertions, 36 deletions
diff --git a/include/linux/mm.h b/include/linux/mm.h
index 50eb18eb0488..d2b99c852301 100644
--- a/include/linux/mm.h
+++ b/include/linux/mm.h
@@ -491,7 +491,9 @@ extern int do_munmap(struct mm_struct *, unsigned long, size_t);
extern unsigned long do_brk(unsigned long, unsigned long);
-static inline void __vma_unlink(struct mm_struct * mm, struct vm_area_struct * vma, struct vm_area_struct * prev)
+static inline void
+__vma_unlink(struct mm_struct *mm, struct vm_area_struct *vma,
+ struct vm_area_struct *prev)
{
prev->vm_next = vma->vm_next;
rb_erase(&vma->vm_rb, &mm->mm_rb);
@@ -499,7 +501,8 @@ static inline void __vma_unlink(struct mm_struct * mm, struct vm_area_struct * v
mm->mmap_cache = prev;
}
-static inline int can_vma_merge(struct vm_area_struct * vma, unsigned long vm_flags)
+static inline int
+can_vma_merge(struct vm_area_struct *vma, unsigned long vm_flags)
{
if (!vma->vm_file && vma->vm_flags == vm_flags)
return 1;
diff --git a/mm/mmap.c b/mm/mmap.c
index b8f425646511..61d0dc32646a 100644
--- a/mm/mmap.c
+++ b/mm/mmap.c
@@ -205,7 +205,8 @@ out:
* internally. Essentially, translate the "PROT_xxx" and "MAP_xxx" bits
* into "VM_xxx".
*/
-static inline unsigned long calc_vm_flags(unsigned long prot, unsigned long flags)
+static inline unsigned long
+calc_vm_flags(unsigned long prot, unsigned long flags)
{
#define _trans(x,bit1,bit2) \
((bit1==bit2)?(x&bit1):(x&bit1)?bit2:0)
@@ -254,10 +255,10 @@ static void validate_mm(struct mm_struct * mm) {
#define validate_mm(mm) do { } while (0)
#endif
-static struct vm_area_struct * find_vma_prepare(struct mm_struct * mm, unsigned long addr,
- struct vm_area_struct ** pprev,
- struct rb_node *** rb_link,
- struct rb_node ** rb_parent)
+static struct vm_area_struct *
+find_vma_prepare(struct mm_struct *mm, unsigned long addr,
+ struct vm_area_struct **pprev, struct rb_node ***rb_link,
+ struct rb_node ** rb_parent)
{
struct vm_area_struct * vma;
struct rb_node ** __rb_link, * __rb_parent, * rb_prev;
@@ -291,8 +292,9 @@ static struct vm_area_struct * find_vma_prepare(struct mm_struct * mm, unsigned
return vma;
}
-static inline void __vma_link_list(struct mm_struct * mm, struct vm_area_struct * vma,
- struct vm_area_struct * prev, struct rb_node * rb_parent)
+static inline void
+__vma_link_list(struct mm_struct *mm, struct vm_area_struct *vma,
+ struct vm_area_struct *prev, struct rb_node *rb_parent)
{
if (prev) {
vma->vm_next = prev->vm_next;
@@ -300,20 +302,21 @@ static inline void __vma_link_list(struct mm_struct * mm, struct vm_area_struct
} else {
mm->mmap = vma;
if (rb_parent)
- vma->vm_next = rb_entry(rb_parent, struct vm_area_struct, vm_rb);
+ vma->vm_next = rb_entry(rb_parent,
+ struct vm_area_struct, vm_rb);
else
vma->vm_next = NULL;
}
}
-static void __vma_link_rb(struct mm_struct * mm, struct vm_area_struct * vma,
- struct rb_node ** rb_link, struct rb_node * rb_parent)
+static void __vma_link_rb(struct mm_struct *mm, struct vm_area_struct *vma,
+ struct rb_node **rb_link, struct rb_node *rb_parent)
{
rb_link_node(&vma->vm_rb, rb_parent, rb_link);
rb_insert_color(&vma->vm_rb, &mm->mm_rb);
}
-static inline void __vma_link_file(struct vm_area_struct * vma)
+static inline void __vma_link_file(struct vm_area_struct *vma)
{
struct file * file;
@@ -332,8 +335,10 @@ static inline void __vma_link_file(struct vm_area_struct * vma)
}
}
-static void __vma_link(struct mm_struct * mm, struct vm_area_struct * vma, struct vm_area_struct * prev,
- struct rb_node ** rb_link, struct rb_node * rb_parent)
+static void
+__vma_link(struct mm_struct *mm, struct vm_area_struct *vma,
+ struct vm_area_struct *prev, struct rb_node **rb_link,
+ struct rb_node *rb_parent)
{
__vma_link_list(mm, vma, prev, rb_parent);
__vma_link_rb(mm, vma, rb_link, rb_parent);
@@ -529,7 +534,8 @@ unsigned long do_mmap_pgoff(struct file * file, unsigned long addr,
* to. we assume access permissions have been handled by the open
* of the memory object, so we don't do any here.
*/
- vm_flags = calc_vm_flags(prot,flags) | mm->def_flags | VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC;
+ vm_flags = calc_vm_flags(prot,flags) | mm->def_flags |
+ VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC;
if (flags & MAP_LOCKED) {
if (!capable(CAP_IPC_LOCK))
@@ -549,14 +555,19 @@ unsigned long do_mmap_pgoff(struct file * file, unsigned long addr,
if (file) {
switch (flags & MAP_TYPE) {
case MAP_SHARED:
- if ((prot & PROT_WRITE) && !(file->f_mode & FMODE_WRITE))
+ if ((prot&PROT_WRITE) && !(file->f_mode&FMODE_WRITE))
return -EACCES;
- /* Make sure we don't allow writing to an append-only file.. */
+ /*
+ * Make sure we don't allow writing to an append-only
+ * file..
+ */
if (IS_APPEND(inode) && (file->f_mode & FMODE_WRITE))
return -EACCES;
- /* make sure there are no mandatory locks on the file. */
+ /*
+ * Make sure there are no mandatory locks on the file.
+ */
if (locks_verify_locked(inode))
return -EAGAIN;
@@ -610,7 +621,9 @@ munmap_back:
/* Check memory availability in shmem_file_setup? */
vm_flags |= VM_ACCOUNT;
} else if (vm_flags & VM_WRITE) {
- /* Private writable mapping: check memory availability */
+ /*
+ * Private writable mapping: check memory availability
+ */
charged = len >> PAGE_SHIFT;
if (!vm_enough_memory(charged))
return -ENOMEM;
@@ -620,10 +633,12 @@ munmap_back:
/* Can we just expand an old anonymous mapping? */
if (!file && !(vm_flags & VM_SHARED) && rb_parent)
- if (vma_merge(mm, prev, rb_parent, addr, addr + len, vm_flags, NULL, 0))
+ if (vma_merge(mm, prev, rb_parent, addr, addr + len,
+ vm_flags, NULL, 0))
goto out;
- /* Determine the object being mapped and call the appropriate
+ /*
+ * Determine the object being mapped and call the appropriate
* specific mapper. the address has already been validated, but
* not unmapped, but the maps are removed from the list.
*/
@@ -734,7 +749,9 @@ unacct_error:
* This function "knows" that -ENOMEM has the bits set.
*/
#ifndef HAVE_ARCH_UNMAPPED_AREA
-static inline unsigned long arch_get_unmapped_area(struct file *filp, unsigned long addr, unsigned long len, unsigned long pgoff, unsigned long flags)
+static inline unsigned long
+arch_get_unmapped_area(struct file *filp, unsigned long addr,
+ unsigned long len, unsigned long pgoff, unsigned long flags)
{
struct mm_struct *mm = current->mm;
struct vm_area_struct *vma;
@@ -769,10 +786,14 @@ static inline unsigned long arch_get_unmapped_area(struct file *filp, unsigned l
}
}
#else
-extern unsigned long arch_get_unmapped_area(struct file *, unsigned long, unsigned long, unsigned long, unsigned long);
+extern unsigned long
+arch_get_unmapped_area(struct file *, unsigned long, unsigned long,
+ unsigned long, unsigned long);
#endif
-unsigned long get_unmapped_area(struct file *file, unsigned long addr, unsigned long len, unsigned long pgoff, unsigned long flags)
+unsigned long
+get_unmapped_area(struct file *file, unsigned long addr, unsigned long len,
+ unsigned long pgoff, unsigned long flags)
{
if (flags & MAP_FIXED) {
if (addr > TASK_SIZE - len)
@@ -783,7 +804,8 @@ unsigned long get_unmapped_area(struct file *file, unsigned long addr, unsigned
}
if (file && file->f_op && file->f_op->get_unmapped_area)
- return file->f_op->get_unmapped_area(file, addr, len, pgoff, flags);
+ return file->f_op->get_unmapped_area(file, addr, len,
+ pgoff, flags);
return arch_get_unmapped_area(file, addr, len, pgoff, flags);
}
@@ -806,7 +828,8 @@ struct vm_area_struct * find_vma(struct mm_struct * mm, unsigned long addr)
while (rb_node) {
struct vm_area_struct * vma_tmp;
- vma_tmp = rb_entry(rb_node, struct vm_area_struct, vm_rb);
+ vma_tmp = rb_entry(rb_node,
+ struct vm_area_struct, vm_rb);
if (vma_tmp->vm_end > addr) {
vma = vma_tmp;
@@ -824,8 +847,9 @@ struct vm_area_struct * find_vma(struct mm_struct * mm, unsigned long addr)
}
/* Same as find_vma, but also return a pointer to the previous VMA in *pprev. */
-struct vm_area_struct * find_vma_prev(struct mm_struct * mm, unsigned long addr,
- struct vm_area_struct **pprev)
+struct vm_area_struct *
+find_vma_prev(struct mm_struct *mm, unsigned long addr,
+ struct vm_area_struct **pprev)
{
struct vm_area_struct *vma = NULL, *prev = NULL;
struct rb_node * rb_node;
@@ -852,7 +876,7 @@ struct vm_area_struct * find_vma_prev(struct mm_struct * mm, unsigned long addr,
}
}
- out:
+out:
*pprev = prev;
return prev ? prev->vm_next : vma;
}
@@ -899,7 +923,8 @@ int expand_stack(struct vm_area_struct * vma, unsigned long address)
return 0;
}
-struct vm_area_struct * find_extend_vma(struct mm_struct * mm, unsigned long addr)
+struct vm_area_struct *
+find_extend_vma(struct mm_struct *mm, unsigned long addr)
{
struct vm_area_struct *vma, *prev;
@@ -918,7 +943,7 @@ struct vm_area_struct * find_extend_vma(struct mm_struct * mm, unsigned long add
/*
* vma is the first one with address < vma->vm_start. Have to extend vma.
*/
-int expand_stack(struct vm_area_struct * vma, unsigned long address)
+int expand_stack(struct vm_area_struct *vma, unsigned long address)
{
unsigned long grow;
@@ -953,7 +978,8 @@ int expand_stack(struct vm_area_struct * vma, unsigned long address)
return 0;
}
-struct vm_area_struct * find_extend_vma(struct mm_struct * mm, unsigned long addr)
+struct vm_area_struct *
+find_extend_vma(struct mm_struct * mm, unsigned long addr)
{
struct vm_area_struct * vma;
unsigned long start;
@@ -1023,7 +1049,7 @@ static void free_pgtables(struct mmu_gather *tlb, struct vm_area_struct *prev,
break;
}
no_mmaps:
- if (last < first) /* needed for arches with discontiguous pgd indices */
+ if (last < first) /* for arches with discontiguous pgd indices */
return;
/*
* If the PGD bits are not consecutive in the virtual address, the
@@ -1298,7 +1324,8 @@ unsigned long do_brk(unsigned long addr, unsigned long len)
flags = VM_DATA_DEFAULT_FLAGS | VM_ACCOUNT | mm->def_flags;
/* Can we just expand an old anonymous mapping? */
- if (rb_parent && vma_merge(mm, prev, rb_parent, addr, addr + len, flags, NULL, 0))
+ if (rb_parent && vma_merge(mm, prev, rb_parent, addr, addr + len,
+ flags, NULL, 0))
goto out;
/*
@@ -1408,7 +1435,7 @@ void insert_vm_struct(struct mm_struct * mm, struct vm_area_struct * vma)
struct vm_area_struct * __vma, * prev;
struct rb_node ** rb_link, * rb_parent;
- __vma = find_vma_prepare(mm, vma->vm_start, &prev, &rb_link, &rb_parent);
+ __vma = find_vma_prepare(mm,vma->vm_start,&prev,&rb_link,&rb_parent);
if (__vma && __vma->vm_start < vma->vm_end)
BUG();
vma_link(mm, vma, prev, rb_link, rb_parent);