fixed all whitespaces in .c files in the fs
folder (not in the sub-folders)

Signed-off-by: amar-laksh <amarlaksh...@gmail.com>
---
 fs/aio.c           |  8 ++++----
 fs/bad_inode.c     | 16 ++++++++--------
 fs/binfmt_aout.c   |  6 +++---
 fs/binfmt_elf.c    | 26 +++++++++++++-------------
 fs/binfmt_em86.c   |  4 ++--
 fs/binfmt_flat.c   | 26 +++++++++++++-------------
 fs/binfmt_script.c |  8 ++++----
 fs/block_dev.c     |  8 ++++----
 fs/buffer.c        | 22 +++++++++++-----------
 fs/compat.c        | 10 +++++-----
 fs/compat_ioctl.c  |  2 +-
 fs/dcache.c        | 32 ++++++++++++++++----------------
 fs/dcookies.c      |  4 ++--
 fs/direct-io.c     | 10 +++++-----
 fs/exec.c          |  2 +-
 fs/fcntl.c         | 18 +++++++++---------
 fs/file_table.c    |  4 ++--
 fs/filesystems.c   | 10 +++++-----
 fs/locks.c         | 14 +++++++-------
 fs/mbcache.c       |  2 +-
 fs/mpage.c         | 12 ++++++------
 fs/namei.c         |  4 ++--
 fs/namespace.c     |  4 ++--
 fs/open.c          |  4 ++--
 fs/pipe.c          | 10 +++++-----
 fs/pnode.c         |  2 +-
 fs/posix_acl.c     |  4 ++--
 fs/read_write.c    |  4 ++--
 fs/readdir.c       |  2 +-
 fs/select.c        |  4 ++--
 fs/signalfd.c      |  2 +-
 fs/stat.c          |  4 ++--
 fs/super.c         |  6 +++---
 fs/timerfd.c       |  2 +-
 34 files changed, 148 insertions(+), 148 deletions(-)

diff --git a/fs/aio.c b/fs/aio.c
index 155f842..0062687 100644
--- a/fs/aio.c
+++ b/fs/aio.c
@@ -1304,10 +1304,10 @@ static long read_events(struct kioctx *ctx, long 
min_nr, long nr,
  *     Create an aio_context capable of receiving at least nr_events.
  *     ctxp must not point to an aio_context that already exists, and
  *     must be initialized to 0 prior to the call.  On successful
- *     creation of the aio_context, *ctxp is filled in with the resulting 
+ *     creation of the aio_context, *ctxp is filled in with the resulting
  *     handle.  May fail with -EINVAL if *ctxp is not initialized,
- *     if the specified nr_events exceeds internal limits.  May fail 
- *     with -EAGAIN if the specified nr_events exceeds the user's limit 
+ *     if the specified nr_events exceeds internal limits.  May fail
+ *     with -EAGAIN if the specified nr_events exceeds the user's limit
  *     of available events.  May fail with -ENOMEM if insufficient kernel
  *     resources are available.  May fail with -EFAULT if an invalid
  *     pointer is passed for ctxp.  Will fail with -ENOSYS if not
@@ -1344,7 +1344,7 @@ out:
 }
 
 /* sys_io_destroy:
- *     Destroy the aio_context specified.  May cancel any outstanding 
+ *     Destroy the aio_context specified.  May cancel any outstanding
  *     AIOs and block on completion.  Will fail with -ENOSYS if not
  *     implemented.  May fail with -EINVAL if the context pointed to
  *     is invalid.
diff --git a/fs/bad_inode.c b/fs/bad_inode.c
index 103f5d7..a7959f0 100644
--- a/fs/bad_inode.c
+++ b/fs/bad_inode.c
@@ -152,12 +152,12 @@ static const struct inode_operations bad_inode_ops =
 /*
  * When a filesystem is unable to read an inode due to an I/O error in
  * its read_inode() function, it can call make_bad_inode() to return a
- * set of stubs which will return EIO errors as required. 
+ * set of stubs which will return EIO errors as required.
  *
  * We only need to do limited initialisation: all other fields are
  * preinitialised to zero automatically.
  */
- 
+
 /**
  *     make_bad_inode - mark an inode bad due to an I/O error
  *     @inode: Inode to mark bad
@@ -166,7 +166,7 @@ static const struct inode_operations bad_inode_ops =
  *     failure this function makes the inode "bad" and causes I/O operations
  *     on it to fail from this point on.
  */
- 
+
 void make_bad_inode(struct inode *inode)
 {
        remove_inode_hash(inode);
@@ -174,8 +174,8 @@ void make_bad_inode(struct inode *inode)
        inode->i_mode = S_IFREG;
        inode->i_atime = inode->i_mtime = inode->i_ctime =
                current_fs_time(inode->i_sb);
-       inode->i_op = &bad_inode_ops;   
-       inode->i_fop = &bad_file_ops;   
+       inode->i_op = &bad_inode_ops;
+       inode->i_fop = &bad_file_ops;
 }
 EXPORT_SYMBOL(make_bad_inode);
 
@@ -184,17 +184,17 @@ EXPORT_SYMBOL(make_bad_inode);
  * &bad_inode_ops to cover the case of invalidated inodes as well as
  * those created by make_bad_inode() above.
  */
- 
+
 /**
  *     is_bad_inode - is an inode errored
  *     @inode: inode to test
  *
  *     Returns true if the inode in question has been marked as bad.
  */
- 
+
 bool is_bad_inode(struct inode *inode)
 {
-       return (inode->i_op == &bad_inode_ops); 
+       return (inode->i_op == &bad_inode_ops);
 }
 
 EXPORT_SYMBOL(is_bad_inode);
diff --git a/fs/binfmt_aout.c b/fs/binfmt_aout.c
index 4c55668..1dcba71 100644
--- a/fs/binfmt_aout.c
+++ b/fs/binfmt_aout.c
@@ -291,7 +291,7 @@ static int load_aout_binary(struct linux_binprm * bprm)
 
                if ((fd_offset & ~PAGE_MASK) != 0 && printk_ratelimit())
                {
-                       printk(KERN_WARNING 
+                       printk(KERN_WARNING
                               "fd_offset is not page aligned. Please convert 
program: %pD\n",
                               bprm->file);
                }
@@ -374,12 +374,12 @@ static int load_aout_library(struct file *file)
        if ((N_TXTOFF(ex) & ~PAGE_MASK) != 0) {
                if (printk_ratelimit())
                {
-                       printk(KERN_WARNING 
+                       printk(KERN_WARNING
                               "N_TXTOFF is not page aligned. Please convert 
library: %pD\n",
                               file);
                }
                vm_brk(start_addr, ex.a_text + ex.a_data + ex.a_bss);
-               
+
                read_code(file, start_addr, N_TXTOFF(ex),
                          ex.a_text + ex.a_data);
                retval = 0;
diff --git a/fs/binfmt_elf.c b/fs/binfmt_elf.c
index 051ea48..0027233 100644
--- a/fs/binfmt_elf.c
+++ b/fs/binfmt_elf.c
@@ -224,7 +224,7 @@ create_elf_tables(struct linux_binprm *bprm, struct elfhdr 
*exec,
        } while (0)
 
 #ifdef ARCH_DLINFO
-       /* 
+       /*
         * ARCH_DLINFO must come first so PPC can do its special alignment of
         * AUXV.
         * update AT_VECTOR_SIZE_ARCH if the number of NEW_AUX_ENT() in
@@ -691,7 +691,7 @@ static int load_elf_binary(struct linux_binprm *bprm)
                retval = -ENOMEM;
                goto out_ret;
        }
-       
+
        /* Get the exec-header */
        loc->elf_ex = *((struct elfhdr *)bprm->buf);
 
@@ -727,7 +727,7 @@ static int load_elf_binary(struct linux_binprm *bprm)
                         * is an a.out format binary
                         */
                        retval = -ENOEXEC;
-                       if (elf_ppnt->p_filesz > PATH_MAX || 
+                       if (elf_ppnt->p_filesz > PATH_MAX ||
                            elf_ppnt->p_filesz < 2)
                                goto out_free_ph;
 
@@ -859,7 +859,7 @@ static int load_elf_binary(struct linux_binprm *bprm)
                                 executable_stack);
        if (retval < 0)
                goto out_free_dentry;
-       
+
        current->mm->start_stack = bprm->p;
 
        /* Now we do a little grungy work by mmapping the ELF image into
@@ -875,7 +875,7 @@ static int load_elf_binary(struct linux_binprm *bprm)
 
                if (unlikely (elf_brk > elf_bss)) {
                        unsigned long nbyte;
-                   
+
                        /* There was a PT_LOAD segment with p_memsz > p_filesz
                           before this one. Map anonymous pages, if needed,
                           and clear the area.  */
@@ -1390,7 +1390,7 @@ static void fill_elf_note_phdr(struct elf_phdr *phdr, int 
sz, loff_t offset)
        return;
 }
 
-static void fill_note(struct memelfnote *note, const char *name, int type, 
+static void fill_note(struct memelfnote *note, const char *name, int type,
                unsigned int sz, void *data)
 {
        note->name = name;
@@ -1442,7 +1442,7 @@ static int fill_psinfo(struct elf_prpsinfo *psinfo, 
struct task_struct *p,
 {
        const struct cred *cred;
        unsigned int i, len;
-       
+
        /* first copy the parameters from user space */
        memset(psinfo, 0, sizeof(struct elf_prpsinfo));
 
@@ -1476,7 +1476,7 @@ static int fill_psinfo(struct elf_prpsinfo *psinfo, 
struct task_struct *p,
        SET_GID(psinfo->pr_gid, from_kgid_munged(cred->user_ns, cred->gid));
        rcu_read_unlock();
        strncpy(psinfo->pr_fname, p->comm, sizeof(psinfo->pr_fname));
-       
+
        return 0;
 }
 
@@ -1878,8 +1878,8 @@ static int elf_dump_thread_status(long signr, struct 
elf_thread_status *t)
        t->num_notes = 0;
 
        fill_prstatus(&t->prstatus, p, signr);
-       elf_core_copy_task_regs(p, &t->prstatus.pr_reg);        
-       
+       elf_core_copy_task_regs(p, &t->prstatus.pr_reg);
+
        fill_note(&t->notes[0], "CORE", NT_PRSTATUS, sizeof(t->prstatus),
                  &(t->prstatus));
        t->num_notes++;
@@ -1900,7 +1900,7 @@ static int elf_dump_thread_status(long signr, struct 
elf_thread_status *t)
                t->num_notes++;
                sz += notesize(&t->notes[2]);
        }
-#endif 
+#endif
        return sz;
 }
 
@@ -2143,7 +2143,7 @@ static int elf_core_dump(struct coredump_params *cprm)
 
        /*
         * We no longer stop all VM operations.
-        * 
+        *
         * This is because those proceses that could possibly change map_count
         * or the mmap / vma pages are now blocked in do_exit on current
         * finishing this core dump.
@@ -2152,7 +2152,7 @@ static int elf_core_dump(struct coredump_params *cprm)
         * the map_count or the pages allocated. So no possibility of crashing
         * exists while dumping the mm->vm_next areas to the core file.
         */
-  
+
        /* alloc memory for large data structures: too large to be on stack */
        elf = kmalloc(sizeof(*elf), GFP_KERNEL);
        if (!elf)
diff --git a/fs/binfmt_em86.c b/fs/binfmt_em86.c
index 4905385..f4deca8 100644
--- a/fs/binfmt_em86.c
+++ b/fs/binfmt_em86.c
@@ -67,11 +67,11 @@ static int load_em86(struct linux_binprm *bprm)
         */
        remove_arg_zero(bprm);
        retval = copy_strings_kernel(1, &bprm->filename, bprm);
-       if (retval < 0) return retval; 
+       if (retval < 0) return retval;
        bprm->argc++;
        if (i_arg) {
                retval = copy_strings_kernel(1, &i_arg, bprm);
-               if (retval < 0) return retval; 
+               if (retval < 0) return retval;
                bprm->argc++;
        }
        retval = copy_strings_kernel(1, &i_name, bprm);
diff --git a/fs/binfmt_flat.c b/fs/binfmt_flat.c
index f723cd3..eab7b5e 100644
--- a/fs/binfmt_flat.c
+++ b/fs/binfmt_flat.c
@@ -387,7 +387,7 @@ static void old_reloc(unsigned long rl)
 #endif
        flat_v2_reloc_t r;
        unsigned long *ptr;
-       
+
        r.value = rl;
 #if defined(CONFIG_COLDFIRE)
        ptr = (unsigned long *) (current->mm->start_code + r.reloc.offset);
@@ -400,7 +400,7 @@ static void old_reloc(unsigned long rl)
                "(address %p, currently %x) into segment %s\n",
                r.reloc.offset, ptr, (int)*ptr, segment[r.reloc.type]);
 #endif
-       
+
        switch (r.reloc.type) {
        case OLD_FLAT_RELOC_TYPE_TEXT:
                *ptr += current->mm->start_code;
@@ -419,7 +419,7 @@ static void old_reloc(unsigned long rl)
 #ifdef DEBUG
        printk("Relocation became %x\n", (int)*ptr);
 #endif
-}              
+}
 
 /****************************************************************************/
 
@@ -478,7 +478,7 @@ static int load_flat_file(struct linux_binprm * bprm,
                ret = -ENOEXEC;
                goto err;
        }
-       
+
        /* Don't allow old format executables to use shared libraries */
        if (rev == OLD_FLAT_VERSION && id != 0) {
                printk("BINFMT_FLAT: shared libraries are not available before 
rev 0x%x\n",
@@ -580,7 +580,7 @@ static int load_flat_file(struct linux_binprm * bprm,
                fpos = ntohl(hdr->data_start);
 #ifdef CONFIG_BINFMT_ZFLAT
                if (flags & FLAT_FLAG_GZDATA) {
-                       result = decompress_exec(bprm, fpos, (char *) datapos, 
+                       result = decompress_exec(bprm, fpos, (char *) datapos,
                                                 full_data, 0);
                } else
 #endif
@@ -703,7 +703,7 @@ static int load_flat_file(struct linux_binprm * bprm,
        libinfo->lib_list[id].loaded = 1;
        libinfo->lib_list[id].entry = (0x00ffffff & ntohl(hdr->entry)) + 
textpos;
        libinfo->lib_list[id].build_date = ntohl(hdr->build_date);
-       
+
        /*
         * We just load the allocations into some temporary memory to
         * help simplify all this mumbo jumbo
@@ -783,11 +783,11 @@ static int load_flat_file(struct linux_binprm * bprm,
                for (i=0; i < relocs; i++)
                        old_reloc(ntohl(reloc[i]));
        }
-       
+
        flush_icache_range(start_code, end_code);
 
        /* zero the BSS,  BRK and stack areas */
-       memset((void*)(datapos + data_len), 0, bss_len + 
+       memset((void*)(datapos + data_len), 0, bss_len +
                        (memp + memp_size - stack_len -         /* end brk */
                        libinfo->lib_list[id].start_brk) +      /* start brk */
                        stack_len);
@@ -881,11 +881,11 @@ static int load_flat_binary(struct linux_binprm * bprm)
        stack_len += (bprm->argc + 1) * sizeof(char *); /* the argv array */
        stack_len += (bprm->envc + 1) * sizeof(char *); /* the envp array */
        stack_len += FLAT_STACK_ALIGN - 1;  /* reserve for upcoming alignment */
-       
+
        res = load_flat_file(bprm, &libinfo, 0, &stack_len);
        if (IS_ERR_VALUE(res))
                return res;
-       
+
        /* Update data segment pointers for all libraries */
        for (i=0; i<MAX_SHARED_LIBS; i++)
                if (libinfo.lib_list[i].loaded)
@@ -907,7 +907,7 @@ static int load_flat_binary(struct linux_binprm * bprm)
                        ((char *) page_address(bprm->page[i/PAGE_SIZE]))[i % 
PAGE_SIZE];
 
        sp = (unsigned long *) create_flat_tables(p, bprm);
-       
+
        /* Fake some return addresses to ensure the call chain will
         * initialise library in order for us.  We are required to call
         * lib 1 first, then 2, ... and finally the main program (id 0).
@@ -923,7 +923,7 @@ static int load_flat_binary(struct linux_binprm * bprm)
                }
        }
 #endif
-       
+
        /* Stash our initial stack pointer into the mm structure */
        current->mm->start_stack = (unsigned long )sp;
 
@@ -932,7 +932,7 @@ static int load_flat_binary(struct linux_binprm * bprm)
 #endif
        DBG_FLT("start_thread(regs=0x%x, entry=0x%x, start_stack=0x%x)\n",
                (int)regs, (int)start_addr, (int)current->mm->start_stack);
-       
+
        start_thread(regs, start_addr, current->mm->start_stack);
 
        return 0;
diff --git a/fs/binfmt_script.c b/fs/binfmt_script.c
index afdf4e3..4b551a5 100644
--- a/fs/binfmt_script.c
+++ b/fs/binfmt_script.c
@@ -55,7 +55,7 @@ static int load_script(struct linux_binprm *bprm)
                        break;
        }
        for (cp = bprm->buf+2; (*cp == ' ') || (*cp == '\t'); cp++);
-       if (*cp == '\0') 
+       if (*cp == '\0')
                return -ENOEXEC; /* No interpreter name found */
        i_name = cp;
        i_arg = NULL;
@@ -80,15 +80,15 @@ static int load_script(struct linux_binprm *bprm)
        if (retval)
                return retval;
        retval = copy_strings_kernel(1, &bprm->interp, bprm);
-       if (retval < 0) return retval; 
+       if (retval < 0) return retval;
        bprm->argc++;
        if (i_arg) {
                retval = copy_strings_kernel(1, &i_arg, bprm);
-               if (retval < 0) return retval; 
+               if (retval < 0) return retval;
                bprm->argc++;
        }
        retval = copy_strings_kernel(1, &i_name, bprm);
-       if (retval) return retval; 
+       if (retval) return retval;
        bprm->argc++;
        retval = bprm_change_interp(interp, bprm);
        if (retval < 0)
diff --git a/fs/block_dev.c b/fs/block_dev.c
index 39b3a17..64f94df 100644
--- a/fs/block_dev.c
+++ b/fs/block_dev.c
@@ -80,7 +80,7 @@ void kill_bdev(struct block_device *bdev)
 
        invalidate_bh_lrus();
        truncate_inode_pages(mapping, 0);
-}      
+}
 EXPORT_SYMBOL(kill_bdev);
 
 /* Invalidate clean unused buffers and pagecache. */
@@ -351,13 +351,13 @@ static loff_t block_llseek(struct file *file, loff_t 
offset, int whence)
        inode_unlock(bd_inode);
        return retval;
 }
-       
+
 int blkdev_fsync(struct file *filp, loff_t start, loff_t end, int datasync)
 {
        struct inode *bd_inode = bdev_file_inode(filp);
        struct block_device *bdev = I_BDEV(bd_inode);
        int error;
-       
+
        error = filemap_write_and_wait_range(filp->f_mapping, start, end);
        if (error)
                return error;
@@ -692,7 +692,7 @@ void bdput(struct block_device *bdev)
 }
 
 EXPORT_SYMBOL(bdput);
- 
+
 static struct block_device *bd_acquire(struct inode *inode)
 {
        struct block_device *bdev;
diff --git a/fs/buffer.c b/fs/buffer.c
index e1632ab..a75ca74 100644
--- a/fs/buffer.c
+++ b/fs/buffer.c
@@ -439,7 +439,7 @@ EXPORT_SYMBOL(mark_buffer_async_write);
  * try_to_free_buffers() will be operating against the *blockdev* mapping
  * at the time, not against the S_ISREG file which depends on those buffers.
  * So the locking for private_list is via the private_lock in the address_space
- * which backs the buffers.  Which is different from the address_space 
+ * which backs the buffers.  Which is different from the address_space
  * against which the buffers are listed.  So for a particular address_space,
  * mapping->private_lock does *not* protect mapping->private_list!  In fact,
  * mapping->private_list will always be protected by the backing blockdev's
@@ -713,7 +713,7 @@ EXPORT_SYMBOL(__set_page_dirty_buffers);
  * Do this in two main stages: first we copy dirty buffers to a
  * temporary inode list, queueing the writes as we go.  Then we clean
  * up, waiting for those writes to complete.
- * 
+ *
  * During this second stage, any subsequent updates to the file may end
  * up refiling the buffer on the original inode's dirty list again, so
  * there is a chance we will end up with a buffer queued for write but
@@ -791,7 +791,7 @@ static int fsync_buffers_list(spinlock_t *lock, struct 
list_head *list)
                brelse(bh);
                spin_lock(lock);
        }
-       
+
        spin_unlock(lock);
        err2 = osync_buffers_list(lock, list);
        if (err)
@@ -901,7 +901,7 @@ no_grow:
        /*
         * Return failure for non-async IO requests.  Async IO requests
         * are not allowed to fail, so we have to wait until buffer heads
-        * become available.  But we don't want tasks sleeping with 
+        * become available.  But we don't want tasks sleeping with
         * partially complete buffers, so all were released above.
         */
        if (!retry)
@@ -910,7 +910,7 @@ no_grow:
        /* We're _really_ low on memory. Now we just
         * wait for old buffer heads to become free due to
         * finishing IO.  Since this is an async request and
-        * the reserve list is empty, we're sure there are 
+        * the reserve list is empty, we're sure there are
         * async buffer heads in use.
         */
        free_more_memory();
@@ -946,7 +946,7 @@ static sector_t blkdev_max_block(struct block_device *bdev, 
unsigned int size)
 
 /*
  * Initialise the state of a blockdev page's buffers.
- */ 
+ */
 static sector_t
 init_page_buffers(struct page *page, struct block_device *bdev,
                        sector_t block, int size)
@@ -1448,7 +1448,7 @@ static bool has_bh_in_lru(int cpu, void *dummy)
 {
        struct bh_lru *b = per_cpu_ptr(&bh_lrus, cpu);
        int i;
-       
+
        for (i = 0; i < BH_LRU_SIZE; i++) {
                if (b->bhs[i])
                        return 1;
@@ -1952,7 +1952,7 @@ int __block_write_begin(struct page *page, loff_t pos, 
unsigned len,
                if (PageUptodate(page)) {
                        if (!buffer_uptodate(bh))
                                set_buffer_uptodate(bh);
-                       continue; 
+                       continue;
                }
                if (!buffer_uptodate(bh) && !buffer_delay(bh) &&
                    !buffer_unwritten(bh) &&
@@ -2258,7 +2258,7 @@ EXPORT_SYMBOL(block_read_full_page);
 
 /* utility function for filesystems that need to do work on expanding
  * truncates.  Uses filesystem pagecache writes to allow the filesystem to
- * deal with the hole.  
+ * deal with the hole.
  */
 int generic_cont_expand_simple(struct inode *inode, loff_t size)
 {
@@ -2819,7 +2819,7 @@ int block_truncate_page(struct address_space *mapping,
 
        length = blocksize - length;
        iblock = (sector_t)index << (PAGE_CACHE_SHIFT - inode->i_blkbits);
-       
+
        page = grab_cache_page(mapping, index);
        err = -ENOMEM;
        if (!page)
@@ -3069,7 +3069,7 @@ EXPORT_SYMBOL(submit_bh);
  *
  * ll_rw_block sets b_end_io to simple completion handler that marks
  * the buffer up-to-date (if appropriate), unlocks the buffer and wakes
- * any waiters. 
+ * any waiters.
  *
  * All of the buffers must be for the same device, and must also be a
  * multiple of the current approved size for the device.
diff --git a/fs/compat.c b/fs/compat.c
index a71936a..6ce6bdd 100644
--- a/fs/compat.c
+++ b/fs/compat.c
@@ -7,7 +7,7 @@
  *  Copyright (C) 2002       Stephen Rothwell, IBM Corporation
  *  Copyright (C) 1997-2000  Jakub Jelinek  (ja...@redhat.com)
  *  Copyright (C) 1998       Eddie C. Dost  (e...@skynet.be)
- *  Copyright (C) 2001,2002  Andi Kleen, SuSE Labs 
+ *  Copyright (C) 2001,2002  Andi Kleen, SuSE Labs
  *  Copyright (C) 2003       Pavel Machek (pa...@ucw.cz)
  *
  *  This program is free software; you can redistribute it and/or modify
@@ -211,7 +211,7 @@ COMPAT_SYSCALL_DEFINE2(newfstat, unsigned int, fd,
 
 static int put_compat_statfs(struct compat_statfs __user *ubuf, struct kstatfs 
*kbuf)
 {
-       
+
        if (sizeof ubuf->f_blocks == 4) {
                if ((kbuf->f_blocks | kbuf->f_bfree | kbuf->f_bavail |
                     kbuf->f_bsize | kbuf->f_frsize) & 0xffffffff00000000ULL)
@@ -536,7 +536,7 @@ COMPAT_SYSCALL_DEFINE5(io_getevents, compat_aio_context_t, 
ctx_id,
                ut = compat_alloc_user_space(sizeof(*ut));
                if (copy_to_user(ut, &t, sizeof(t)) )
                        return -EFAULT;
-       } 
+       }
        return sys_io_getevents(ctx_id, min_nr, nr, events, ut);
 }
 
@@ -636,7 +636,7 @@ copy_iocb(long nr, u32 __user *ptr32, struct iocb __user * 
__user *ptr64)
 COMPAT_SYSCALL_DEFINE3(io_submit, compat_aio_context_t, ctx_id,
                       int, nr, u32 __user *, iocb)
 {
-       struct iocb __user * __user *iocb64; 
+       struct iocb __user * __user *iocb64;
        long ret;
 
        if (unlikely(nr < 0))
@@ -644,7 +644,7 @@ COMPAT_SYSCALL_DEFINE3(io_submit, compat_aio_context_t, 
ctx_id,
 
        if (nr > MAX_AIO_SUBMITS)
                nr = MAX_AIO_SUBMITS;
-       
+
        iocb64 = compat_alloc_user_space(nr * sizeof(*iocb64));
        ret = copy_iocb(nr, iocb, iocb64);
        if (!ret)
diff --git a/fs/compat_ioctl.c b/fs/compat_ioctl.c
index 6402eaf..c6519f5 100644
--- a/fs/compat_ioctl.c
+++ b/fs/compat_ioctl.c
@@ -3,7 +3,7 @@
  *
  * Copyright (C) 1997-2000  Jakub Jelinek  (ja...@redhat.com)
  * Copyright (C) 1998  Eddie C. Dost  (e...@skynet.be)
- * Copyright (C) 2001,2002  Andi Kleen, SuSE Labs 
+ * Copyright (C) 2001,2002  Andi Kleen, SuSE Labs
  * Copyright (C) 2003       Pavel Machek (pa...@ucw.cz)
  *
  * These routines maintain argument size conversion between 32bit and 64bit
diff --git a/fs/dcache.c b/fs/dcache.c
index 92d5140..ec44f22 100644
--- a/fs/dcache.c
+++ b/fs/dcache.c
@@ -254,14 +254,14 @@ static void __d_free(struct rcu_head *head)
 {
        struct dentry *dentry = container_of(head, struct dentry, d_u.d_rcu);
 
-       kmem_cache_free(dentry_cache, dentry); 
+       kmem_cache_free(dentry_cache, dentry);
 }
 
 static void __d_free_external(struct rcu_head *head)
 {
        struct dentry *dentry = container_of(head, struct dentry, d_u.d_rcu);
        kfree(external_name(dentry));
-       kmem_cache_free(dentry_cache, dentry); 
+       kmem_cache_free(dentry_cache, dentry);
 }
 
 static inline int dname_external(const struct dentry *dentry)
@@ -731,7 +731,7 @@ static inline bool fast_dput(struct dentry *dentry)
 }
 
 
-/* 
+/*
  * This is dput
  *
  * This is complicated by the fact that we do not want to put
@@ -750,7 +750,7 @@ static inline bool fast_dput(struct dentry *dentry)
 
 /*
  * dput - release a dentry
- * @dentry: dentry to release 
+ * @dentry: dentry to release
  *
  * Release a dentry. This will drop the usage count and if appropriate
  * call the dentry unlink method as well as removing it from the queues and
@@ -1552,7 +1552,7 @@ EXPORT_SYMBOL(d_invalidate);
  * available. On a success the dentry is returned. The name passed in is
  * copied and the copy passed in may be reused after this call.
  */
- 
+
 struct dentry *__d_alloc(struct super_block *sb, const struct qstr *name)
 {
        struct dentry *dentry;
@@ -1574,7 +1574,7 @@ struct dentry *__d_alloc(struct super_block *sb, const 
struct qstr *name)
                struct external_name *p = kmalloc(size + name->len,
                                                  GFP_KERNEL_ACCOUNT);
                if (!p) {
-                       kmem_cache_free(dentry_cache, dentry); 
+                       kmem_cache_free(dentry_cache, dentry);
                        return NULL;
                }
                atomic_set(&p->u.count, 1);
@@ -1584,7 +1584,7 @@ struct dentry *__d_alloc(struct super_block *sb, const 
struct qstr *name)
                                round_up(name->len + 1, sizeof(unsigned long)));
        } else  {
                dname = dentry->d_iname;
-       }       
+       }
 
        dentry->d_name.len = name->len;
        dentry->d_name.hash = name->hash;
@@ -1778,7 +1778,7 @@ static void __d_instantiate(struct dentry *dentry, struct 
inode *inode)
  * (or otherwise set) by the caller to indicate that it is now
  * in use by the dcache.
  */
- 
+
 void d_instantiate(struct dentry *entry, struct inode * inode)
 {
        BUG_ON(!hlist_unhashed(&entry->d_u.d_alias));
@@ -2300,7 +2300,7 @@ struct dentry *__d_lookup(const struct dentry *parent, 
const struct qstr *name)
         * See Documentation/filesystems/path-lookup.txt for more details.
         */
        rcu_read_lock();
-       
+
        hlist_bl_for_each_entry_rcu(dentry, node, b, d_hash) {
 
                if (dentry->d_name.hash != hash)
@@ -2376,7 +2376,7 @@ EXPORT_SYMBOL(d_hash_and_lookup);
  * it from the hash queues and waiting for
  * it to be deleted later when it has no users
  */
- 
+
 /**
  * d_delete - delete a dentry
  * @dentry: The dentry to delete
@@ -2384,7 +2384,7 @@ EXPORT_SYMBOL(d_hash_and_lookup);
  * Turn the dentry into a negative dentry if possible, otherwise
  * remove it from the hash queues so it can be deleted later
  */
- 
+
 void d_delete(struct dentry * dentry)
 {
        struct inode *inode;
@@ -2437,7 +2437,7 @@ static void _d_rehash(struct dentry * entry)
  *
  * Adds a dentry to the hash according to its name.
  */
- 
+
 void d_rehash(struct dentry * entry)
 {
        spin_lock(&entry->d_lock);
@@ -3136,7 +3136,7 @@ char *simple_dname(struct dentry *dentry, char *buffer, 
int buflen)
        /* these dentries are never renamed, so d_lock is not needed */
        if (prepend(&end, &buflen, " (deleted)", 11) ||
            prepend(&end, &buflen, dentry->d_name.name, dentry->d_name.len) ||
-           prepend(&end, &buflen, "/", 1))  
+           prepend(&end, &buflen, "/", 1))
                end = ERR_PTR(-ENAMETOOLONG);
        return end;
 }
@@ -3308,7 +3308,7 @@ out:
  * Returns false otherwise.
  * Caller must ensure that "new_dentry" is pinned before calling is_subdir()
  */
-  
+
 bool is_subdir(struct dentry *new_dentry, struct dentry *old_dentry)
 {
        bool result;
@@ -3410,10 +3410,10 @@ static void __init dcache_init(void)
 {
        unsigned int loop;
 
-       /* 
+       /*
         * A constructor could be added for stable state like the lists,
         * but it is probably not worth it because of the cache nature
-        * of the dcache. 
+        * of the dcache.
         */
        dentry_cache = KMEM_CACHE(dentry,
                SLAB_RECLAIM_ACCOUNT|SLAB_PANIC|SLAB_MEM_SPREAD|SLAB_ACCOUNT);
diff --git a/fs/dcookies.c b/fs/dcookies.c
index ac44a69..ca7f7a6 100644
--- a/fs/dcookies.c
+++ b/fs/dcookies.c
@@ -187,7 +187,7 @@ SYSCALL_DEFINE3(lookup_dcookie, u64, cookie64, char __user 
*, buf, size_t, len)
        }
 
        err = -ERANGE;
- 
+
        pathlen = kbuf + PAGE_SIZE - path;
        if (pathlen <= len) {
                err = pathlen;
@@ -306,7 +306,7 @@ static void dcookie_exit(void)
 struct dcookie_user {
        struct list_head next;
 };
- 
+
 struct dcookie_user * dcookie_register(void)
 {
        struct dcookie_user * user;
diff --git a/fs/direct-io.c b/fs/direct-io.c
index d6a9012..0432f4f 100644
--- a/fs/direct-io.c
+++ b/fs/direct-io.c
@@ -189,7 +189,7 @@ static inline int dio_refill_pages(struct dio *dio, struct 
dio_submit *sdio)
                sdio->to = ((ret - 1) & (PAGE_SIZE - 1)) + 1;
                return 0;
        }
-       return ret;     
+       return ret;
 }
 
 /*
@@ -286,7 +286,7 @@ static void dio_aio_complete_work(struct work_struct *work)
 static int dio_bio_complete(struct dio *dio, struct bio *bio);
 
 /*
- * Asynchronous IO callback. 
+ * Asynchronous IO callback.
  */
 static void dio_bio_end_aio(struct bio *bio)
 {
@@ -699,7 +699,7 @@ static inline int dio_bio_add_page(struct dio_submit *sdio)
        }
        return ret;
 }
-               
+
 /*
  * Put cur_page under IO.  The section of cur_page which is described by
  * cur_page_offset,cur_page_len is put into a BIO.  The section of cur_page
@@ -761,7 +761,7 @@ out:
  * An autonomous function to put a chunk of a page under deferred IO.
  *
  * The caller doesn't actually know (or care) whether this piece of page is in
- * a BIO, or is under IO or whatever.  We just take care of all possible 
+ * a BIO, or is under IO or whatever.  We just take care of all possible
  * situations here.  The separation between the logic of do_direct_IO() and
  * that of submit_page_section() is important for clarity.  Please don't break.
  *
@@ -879,7 +879,7 @@ static inline void dio_zero_block(struct dio *dio, struct 
dio_submit *sdio,
         * We need to zero out part of an fs block.  It is either at the
         * beginning or the end of the fs block.
         */
-       if (end) 
+       if (end)
                this_chunk_blocks = dio_blocks_per_fs_block - this_chunk_blocks;
 
        this_chunk_bytes = this_chunk_blocks << sdio->blkbits;
diff --git a/fs/exec.c b/fs/exec.c
index dcd4ac7..4cf2252 100644
--- a/fs/exec.c
+++ b/fs/exec.c
@@ -19,7 +19,7 @@
  * current->executable is only used by the procfs.  This allows a dispatch
  * table to check for several different types  of binary formats.  We keep
  * trying until we recognize the file or we run out of supported binary
- * formats. 
+ * formats.
  */
 
 #include <linux/slab.h>
diff --git a/fs/fcntl.c b/fs/fcntl.c
index 350a2c8..ae351e5 100644
--- a/fs/fcntl.c
+++ b/fs/fcntl.c
@@ -355,7 +355,7 @@ static int check_fcntl_cmd(unsigned cmd)
 }
 
 SYSCALL_DEFINE3(fcntl, unsigned int, fd, unsigned int, cmd, unsigned long, arg)
-{      
+{
        struct fd f = fdget_raw(fd);
        long err = -EBADF;
 
@@ -380,7 +380,7 @@ out:
 #if BITS_PER_LONG == 32
 SYSCALL_DEFINE3(fcntl64, unsigned int, fd, unsigned int, cmd,
                unsigned long, arg)
-{      
+{
        struct fd f = fdget_raw(fd);
        long err = -EBADF;
 
@@ -395,7 +395,7 @@ SYSCALL_DEFINE3(fcntl64, unsigned int, fd, unsigned int, 
cmd,
        err = security_file_fcntl(f.file, cmd, arg);
        if (err)
                goto out1;
-       
+
        switch (cmd) {
        case F_GETLK64:
        case F_OFD_GETLK:
@@ -463,8 +463,8 @@ static void send_sigio_to_task(struct task_struct *p,
                siginfo_t si;
                default:
                        /* Queue a rt signal with the appropriate fd as its
-                          value.  We use SI_SIGIO as the source, not 
-                          SI_KERNEL, since kernel signals always get 
+                          value.  We use SI_SIGIO as the source, not
+                          SI_KERNEL, since kernel signals always get
                           delivered even if we can't queue.  Failure to
                           queue in this case _should_ be reported; we fall
                           back to SIGIO in that case. --sct */
@@ -494,7 +494,7 @@ void send_sigio(struct fown_struct *fown, int fd, int band)
        enum pid_type type;
        struct pid *pid;
        int group = 1;
-       
+
        read_lock(&fown->lock);
 
        type = fown->pid_type;
@@ -506,7 +506,7 @@ void send_sigio(struct fown_struct *fown, int fd, int band)
        pid = fown->pid;
        if (!pid)
                goto out_unlock_fown;
-       
+
        read_lock(&tasklist_lock);
        do_each_pid_task(pid, type, p) {
                send_sigio_to_task(p, fown, fd, band, group);
@@ -530,7 +530,7 @@ int send_sigurg(struct fown_struct *fown)
        struct pid *pid;
        int group = 1;
        int ret = 0;
-       
+
        read_lock(&fown->lock);
 
        type = fown->pid_type;
@@ -544,7 +544,7 @@ int send_sigurg(struct fown_struct *fown)
                goto out_unlock_fown;
 
        ret = 1;
-       
+
        read_lock(&tasklist_lock);
        do_each_pid_task(pid, type, p) {
                send_sigurg_to_task(p, fown, group);
diff --git a/fs/file_table.c b/fs/file_table.c
index ad17e05..77c98ea 100644
--- a/fs/file_table.c
+++ b/fs/file_table.c
@@ -310,7 +310,7 @@ void put_filp(struct file *file)
 }
 
 void __init files_init(void)
-{ 
+{
        filp_cachep = kmem_cache_create("filp", sizeof(struct file), 0,
                        SLAB_HWCACHE_ALIGN | SLAB_PANIC, NULL);
        percpu_counter_init(&nr_files, 0, GFP_KERNEL);
@@ -329,4 +329,4 @@ void __init files_maxfiles_init(void)
        n = ((totalram_pages - memreserve) * (PAGE_SIZE / 1024)) / 10;
 
        files_stat.max_files = max_t(unsigned long, n, NR_FILE);
-} 
+}
diff --git a/fs/filesystems.c b/fs/filesystems.c
index c5618db..5e0a168 100644
--- a/fs/filesystems.c
+++ b/fs/filesystems.c
@@ -61,11 +61,11 @@ static struct file_system_type **find_filesystem(const char 
*name, unsigned len)
  *     is aware of for mount and other syscalls. Returns 0 on success,
  *     or a negative errno code on an error.
  *
- *     The &struct file_system_type that is passed is linked into the kernel 
+ *     The &struct file_system_type that is passed is linked into the kernel
  *     structures and must not be freed until the file system has been
  *     unregistered.
  */
- 
+
 int register_filesystem(struct file_system_type * fs)
 {
        int res = 0;
@@ -93,11 +93,11 @@ EXPORT_SYMBOL(register_filesystem);
  *     Remove a file system that was previously successfully registered
  *     with the kernel. An error is returned if the file system is not found.
  *     Zero is returned on a success.
- *     
+ *
  *     Once this function has returned the &struct file_system_type structure
  *     may be freed or reused.
  */
- 
+
 int unregister_filesystem(struct file_system_type * fs)
 {
        struct file_system_type ** tmp;
@@ -179,7 +179,7 @@ static int fs_maxindex(void)
 }
 
 /*
- * Whee.. Weird sysv syscall. 
+ * Whee.. Weird sysv syscall.
  */
 SYSCALL_DEFINE3(sysfs, int, option, unsigned long, arg1, unsigned long, arg2)
 {
diff --git a/fs/locks.c b/fs/locks.c
index 7c5f91b..5daec54 100644
--- a/fs/locks.c
+++ b/fs/locks.c
@@ -11,11 +11,11 @@
  *
  *  Miscellaneous edits, and a total rewrite of posix_lock_file() code.
  *  Kai Petzke (w...@marie.physik.tu-berlin.de), 1994
- *  
+ *
  *  Converted file_lock_table to a linked list from an array, which eliminates
  *  the limits on how many active file locks are open.
  *  Chad Page (page...@netcom.com), November 27, 1994
- * 
+ *
  *  Removed dependency on file descriptors. dup()'ed file descriptors now
  *  get the same locks as the original file descriptors, and a close() on
  *  any file descriptor removes ALL the locks on the file for the current
@@ -41,7 +41,7 @@
  *  with a file pointer (filp). As a result they can be shared by a parent
  *  process and its children after a fork(). They are removed when the last
  *  file descriptor referring to the file pointer is closed (unless explicitly
- *  unlocked). 
+ *  unlocked).
  *
  *  FL_FLOCK locks never deadlock, an existing lock is always removed before
  *  upgrading from shared to exclusive (or vice versa). When this happens
@@ -50,7 +50,7 @@
  *  Andy Walker (a...@lysaker.kvaerner.no), June 09, 1995
  *
  *  Removed some race conditions in flock_lock_file(), marked other possible
- *  races. Just grep for FIXME to see them. 
+ *  races. Just grep for FIXME to see them.
  *  Dmitry Gorodchanin (pgm...@ibi.com), February 09, 1996.
  *
  *  Addressed Dmitry's concerns. Deadlock checking no longer recursive.
@@ -405,7 +405,7 @@ flock_make_lock(struct file *filp, unsigned int cmd)
 
        if (type < 0)
                return ERR_PTR(type);
-       
+
        fl = locks_alloc_lock();
        if (fl == NULL)
                return ERR_PTR(-ENOMEM);
@@ -416,7 +416,7 @@ flock_make_lock(struct file *filp, unsigned int cmd)
        fl->fl_flags = FL_FLOCK;
        fl->fl_type = type;
        fl->fl_end = OFFSET_MAX;
-       
+
        return fl;
 }
 
@@ -2078,7 +2078,7 @@ int fcntl_getlk(struct file *filp, unsigned int cmd, 
struct flock __user *l)
        error = vfs_test_lock(filp, &file_lock);
        if (error)
                goto out;
- 
+
        flock.l_type = file_lock.fl_type;
        if (file_lock.fl_type != F_UNLCK) {
                error = posix_lock_to_flock(&flock, &file_lock);
diff --git a/fs/mbcache.c b/fs/mbcache.c
index 187477d..1c519d0 100644
--- a/fs/mbcache.c
+++ b/fs/mbcache.c
@@ -587,7 +587,7 @@ mb_cache_entry_insert(struct mb_cache_entry *ce, struct 
block_device *bdev,
        struct mb_cache_entry *lce;
 
        mb_assert(ce);
-       bucket = hash_long((unsigned long)bdev + (block & 0xffffffff), 
+       bucket = hash_long((unsigned long)bdev + (block & 0xffffffff),
                           cache->c_bucket_bits);
        block_hash_p = &cache->c_block_hash[bucket];
        hlist_bl_lock(block_hash_p);
diff --git a/fs/mpage.c b/fs/mpage.c
index 1480d3a..0a99fda 100644
--- a/fs/mpage.c
+++ b/fs/mpage.c
@@ -94,8 +94,8 @@ mpage_alloc(struct block_device *bdev,
  * them.  So when the buffer is up to date and the page size == block size,
  * this marks the page up to date instead of adding new buffers.
  */
-static void 
-map_buffer_to_page(struct page *page, struct buffer_head *bh, int page_block) 
+static void
+map_buffer_to_page(struct page *page, struct buffer_head *bh, int page_block)
 {
        struct inode *inode = page->mapping->host;
        struct buffer_head *page_bh, *head;
@@ -106,9 +106,9 @@ map_buffer_to_page(struct page *page, struct buffer_head 
*bh, int page_block)
                 * don't make any buffers if there is only one buffer on
                 * the page and the page just needs to be set up to date
                 */
-               if (inode->i_blkbits == PAGE_CACHE_SHIFT && 
+               if (inode->i_blkbits == PAGE_CACHE_SHIFT &&
                    buffer_uptodate(bh)) {
-                       SetPageUptodate(page);    
+                       SetPageUptodate(page);
                        return;
                }
                create_empty_buffers(page, 1 << inode->i_blkbits, 0);
@@ -226,7 +226,7 @@ do_mpage_readpage(struct bio *bio, struct page *page, 
unsigned nr_pages,
                        map_buffer_to_page(page, map_bh, page_block);
                        goto confused;
                }
-       
+
                if (first_hole != blocks_per_page)
                        goto confused;          /* hole -> non-hole */
 
@@ -420,7 +420,7 @@ EXPORT_SYMBOL(mpage_readpage);
  *
  * If all blocks are found to be contiguous then the page can go into the
  * BIO.  Otherwise fall back to the mapping's writepage().
- * 
+ *
  * FIXME: This code wants an estimate of how many pages are still to be
  * written, so it can intelligently allocate a suitably-sized BIO.  For now,
  * just allocate full-size (16-page) BIOs.
diff --git a/fs/namei.c b/fs/namei.c
index f624d13..f9cbe3d 100644
--- a/fs/namei.c
+++ b/fs/namei.c
@@ -49,8 +49,8 @@
  * The new code replaces the old recursive symlink resolution with
  * an iterative one (in case of non-nested symlink chains).  It does
  * this with calls to <fs>_follow_link().
- * As a side effect, dir_namei(), _namei() and follow_link() are now 
- * replaced with a single function lookup_dentry() that can handle all 
+ * As a side effect, dir_namei(), _namei() and follow_link() are now
+ * replaced with a single function lookup_dentry() that can handle all
  * the special cases of the former code.
  *
  * With the new dcache, the pathname is stored at each inode, at least as
diff --git a/fs/namespace.c b/fs/namespace.c
index 4fb1691..806ef24 100644
--- a/fs/namespace.c
+++ b/fs/namespace.c
@@ -1576,7 +1576,7 @@ out_unlock:
        namespace_unlock();
 }
 
-/* 
+/*
  * Is the caller allowed to modify his namespace?
  */
 static inline bool may_mount(void)
@@ -2090,7 +2090,7 @@ static int do_loopback(struct path *path, const char 
*old_name,
 
        err = -EINVAL;
        if (mnt_ns_loop(old_path.dentry))
-               goto out; 
+               goto out;
 
        mp = lock_mount(path);
        err = PTR_ERR(mp);
diff --git a/fs/open.c b/fs/open.c
index 55bdc75..586db20 100644
--- a/fs/open.c
+++ b/fs/open.c
@@ -875,7 +875,7 @@ struct file *dentry_open(const struct path *path, int flags,
                                fput(f);
                                f = ERR_PTR(error);
                        }
-               } else { 
+               } else {
                        put_filp(f);
                        f = ERR_PTR(error);
                }
@@ -982,7 +982,7 @@ struct file *filp_open(const char *filename, int flags, 
umode_t mode)
 {
        struct filename *name = getname_kernel(filename);
        struct file *file = ERR_CAST(name);
-       
+
        if (!IS_ERR(name)) {
                file = file_open_name(name, flags, mode);
                putname(name);
diff --git a/fs/pipe.c b/fs/pipe.c
index ab8dad3..b72fe79 100644
--- a/fs/pipe.c
+++ b/fs/pipe.c
@@ -45,10 +45,10 @@ unsigned long pipe_user_pages_hard;
 unsigned long pipe_user_pages_soft = PIPE_DEF_BUFFERS * INR_OPEN_CUR;
 
 /*
- * We use a start+len construction, which provides full use of the 
+ * We use a start+len construction, which provides full use of the
  * allocated memory.
  * -- Florian Coosmann (FGC)
- * 
+ *
  * Reads with count = 0 should always return 0.
  * -- Julian Bradfield 1999-06-07.
  *
@@ -842,7 +842,7 @@ SYSCALL_DEFINE1(pipe, int __user *, fildes)
 
 static int wait_for_partner(struct pipe_inode_info *pipe, unsigned int *cnt)
 {
-       int cur = *cnt; 
+       int cur = *cnt;
 
        while (cur == *cnt) {
                pipe_wait(pipe);
@@ -917,7 +917,7 @@ static int fifo_open(struct inode *inode, struct file *filp)
                        }
                }
                break;
-       
+
        case FMODE_WRITE:
        /*
         *  O_WRONLY
@@ -937,7 +937,7 @@ static int fifo_open(struct inode *inode, struct file *filp)
                                goto err_wr;
                }
                break;
-       
+
        case FMODE_READ | FMODE_WRITE:
        /*
         *  O_RDWR
diff --git a/fs/pnode.c b/fs/pnode.c
index c524fdd..ced8017 100644
--- a/fs/pnode.c
+++ b/fs/pnode.c
@@ -240,7 +240,7 @@ static int propagate_one(struct mount *m)
                if (IS_MNT_SHARED(m))
                        type |= CL_MAKE_SHARED;
        }
-               
+
        /* Notice when we are propagating across user namespaces */
        if (m->mnt_ns->user_ns != user_ns)
                type |= CL_UNPRIVILEGED;
diff --git a/fs/posix_acl.c b/fs/posix_acl.c
index 711dd51..67e5bc5 100644
--- a/fs/posix_acl.c
+++ b/fs/posix_acl.c
@@ -675,12 +675,12 @@ posix_acl_from_xattr(struct user_namespace *user_ns,
                return ERR_PTR(-EINVAL);
        if (count == 0)
                return NULL;
-       
+
        acl = posix_acl_alloc(count, GFP_NOFS);
        if (!acl)
                return ERR_PTR(-ENOMEM);
        acl_e = acl->a_entries;
-       
+
        for (end = entry + count; entry != end; acl_e++, entry++) {
                acl_e->e_tag  = le16_to_cpu(entry->e_tag);
                acl_e->e_perm = le16_to_cpu(entry->e_perm);
diff --git a/fs/read_write.c b/fs/read_write.c
index 0c8782a..904b39d 100644
--- a/fs/read_write.c
+++ b/fs/read_write.c
@@ -4,7 +4,7 @@
  *  Copyright (C) 1991, 1992  Linus Torvalds
  */
 
-#include <linux/slab.h> 
+#include <linux/slab.h>
 #include <linux/stat.h>
 #include <linux/fcntl.h>
 #include <linux/file.h>
@@ -663,7 +663,7 @@ SYSCALL_DEFINE4(pwrite64, unsigned int, fd, const char 
__user *, buf,
        f = fdget(fd);
        if (f.file) {
                ret = -ESPIPE;
-               if (f.file->f_mode & FMODE_PWRITE)  
+               if (f.file->f_mode & FMODE_PWRITE)
                        ret = vfs_write(f.file, buf, count, &pos);
                fdput(f);
        }
diff --git a/fs/readdir.c b/fs/readdir.c
index e69ef3b..15e7e14 100644
--- a/fs/readdir.c
+++ b/fs/readdir.c
@@ -132,7 +132,7 @@ SYSCALL_DEFINE3(old_readdir, unsigned int, fd,
 
 /*
  * New, all-improved, singing, dancing, iBCS2-compliant getdents()
- * interface. 
+ * interface.
  */
 struct linux_dirent {
        unsigned long   d_ino;
diff --git a/fs/select.c b/fs/select.c
index 79d0d49..c0dd808 100644
--- a/fs/select.c
+++ b/fs/select.c
@@ -10,7 +10,7 @@
  *     parameter to reflect time remaining.
  *
  *  24 January 2000
- *     Changed sys_poll()/do_poll() to use PAGE_SIZE chunk-based allocation 
+ *     Changed sys_poll()/do_poll() to use PAGE_SIZE chunk-based allocation
  *     of fds to overcome nfds < 16390 descriptors limit (Tigran Aivazian).
  */
 
@@ -570,7 +570,7 @@ int core_sys_select(int n, fd_set __user *inp, fd_set 
__user *outp,
        /*
         * We need 6 bitmaps (in/out/ex for both incoming and outgoing),
         * since we used fdset we need to allocate memory in units of
-        * long-words. 
+        * long-words.
         */
        size = FDS_BYTES(n);
        bits = stack_fds;
diff --git a/fs/signalfd.c b/fs/signalfd.c
index 270221f..3e01ca6 100644
--- a/fs/signalfd.c
+++ b/fs/signalfd.c
@@ -117,7 +117,7 @@ static int signalfd_copyinfo(struct signalfd_siginfo __user 
*uinfo,
                err |= __put_user(kinfo->si_trapno, &uinfo->ssi_trapno);
 #endif
 #ifdef BUS_MCEERR_AO
-               /* 
+               /*
                 * Other callers might not initialize the si_lsb field,
                 * so check explicitly for the right codes here.
                 */
diff --git a/fs/stat.c b/fs/stat.c
index bc045c7..feb37c2 100644
--- a/fs/stat.c
+++ b/fs/stat.c
@@ -141,7 +141,7 @@ static int cp_old_stat(struct kstat *stat, struct 
__old_kernel_stat __user * sta
 {
        static int warncount = 5;
        struct __old_kernel_stat tmp;
-       
+
        if (warncount > 0) {
                warncount--;
                printk(KERN_WARNING "VFS: Warning: %s using old stat() call. 
Recompile your binary.\n",
@@ -166,7 +166,7 @@ static int cp_old_stat(struct kstat *stat, struct 
__old_kernel_stat __user * sta
 #if BITS_PER_LONG == 32
        if (stat->size > MAX_NON_LFS)
                return -EOVERFLOW;
-#endif 
+#endif
        tmp.st_size = stat->size;
        tmp.st_atime = stat->atime.tv_sec;
        tmp.st_mtime = stat->mtime.tv_sec;
diff --git a/fs/super.c b/fs/super.c
index 1182af8..d5d28b7 100644
--- a/fs/super.c
+++ b/fs/super.c
@@ -482,7 +482,7 @@ retry:
                        return ERR_PTR(-ENOMEM);
                goto retry;
        }
-               
+
        err = set(s, data);
        if (err) {
                spin_unlock(&sb_lock);
@@ -583,7 +583,7 @@ EXPORT_SYMBOL(iterate_supers_type);
 /**
  *     get_super - get the superblock of a device
  *     @bdev: device to get the superblock for
- *     
+ *
  *     Scans the superblock list and finds the superblock of the file system
  *     mounted on the device given. %NULL is returned if no match is found.
  */
@@ -673,7 +673,7 @@ restart:
        spin_unlock(&sb_lock);
        return NULL;
 }
- 
+
 struct super_block *user_get_super(dev_t dev)
 {
        struct super_block *sb;
diff --git a/fs/timerfd.c b/fs/timerfd.c
index 053818d..d0ccf0c 100644
--- a/fs/timerfd.c
+++ b/fs/timerfd.c
@@ -415,7 +415,7 @@ SYSCALL_DEFINE2(timerfd_create, int, clockid, int, flags)
        return ufd;
 }
 
-static int do_timerfd_settime(int ufd, int flags, 
+static int do_timerfd_settime(int ufd, int flags,
                const struct itimerspec *new,
                struct itimerspec *old)
 {
-- 
1.9.1

Reply via email to