@@ -1304,10 +1304,10 @@ static long read_events(struct kioctx *ctx, long min_nr, long nr,
* Create an aio_context capable of receiving at least nr_events.
* ctxp must not point to an aio_context that already exists, and
* must be initialized to 0 prior to the call. On successful
- * creation of the aio_context, *ctxp is filled in with the resulting
+ * creation of the aio_context, *ctxp is filled in with the resulting
* handle. May fail with -EINVAL if *ctxp is not initialized,
- * if the specified nr_events exceeds internal limits. May fail
- * with -EAGAIN if the specified nr_events exceeds the user's limit
+ * if the specified nr_events exceeds internal limits. May fail
+ * with -EAGAIN if the specified nr_events exceeds the user's limit
* of available events. May fail with -ENOMEM if insufficient kernel
* resources are available. May fail with -EFAULT if an invalid
* pointer is passed for ctxp. Will fail with -ENOSYS if not
@@ -1344,7 +1344,7 @@ out:
}
/* sys_io_destroy:
- * Destroy the aio_context specified. May cancel any outstanding
+ * Destroy the aio_context specified. May cancel any outstanding
* AIOs and block on completion. Will fail with -ENOSYS if not
* implemented. May fail with -EINVAL if the context pointed to
* is invalid.
@@ -152,12 +152,12 @@ static const struct inode_operations bad_inode_ops =
/*
* When a filesystem is unable to read an inode due to an I/O error in
* its read_inode() function, it can call make_bad_inode() to return a
- * set of stubs which will return EIO errors as required.
+ * set of stubs which will return EIO errors as required.
*
* We only need to do limited initialisation: all other fields are
* preinitialised to zero automatically.
*/
-
+
/**
* make_bad_inode - mark an inode bad due to an I/O error
* @inode: Inode to mark bad
@@ -166,7 +166,7 @@ static const struct inode_operations bad_inode_ops =
* failure this function makes the inode "bad" and causes I/O operations
* on it to fail from this point on.
*/
-
+
void make_bad_inode(struct inode *inode)
{
remove_inode_hash(inode);
@@ -174,8 +174,8 @@ void make_bad_inode(struct inode *inode)
inode->i_mode = S_IFREG;
inode->i_atime = inode->i_mtime = inode->i_ctime =
current_fs_time(inode->i_sb);
- inode->i_op = &bad_inode_ops;
- inode->i_fop = &bad_file_ops;
+ inode->i_op = &bad_inode_ops;
+ inode->i_fop = &bad_file_ops;
}
EXPORT_SYMBOL(make_bad_inode);
@@ -184,17 +184,17 @@ EXPORT_SYMBOL(make_bad_inode);
* &bad_inode_ops to cover the case of invalidated inodes as well as
* those created by make_bad_inode() above.
*/
-
+
/**
* is_bad_inode - is an inode errored
* @inode: inode to test
*
* Returns true if the inode in question has been marked as bad.
*/
-
+
bool is_bad_inode(struct inode *inode)
{
- return (inode->i_op == &bad_inode_ops);
+ return (inode->i_op == &bad_inode_ops);
}
EXPORT_SYMBOL(is_bad_inode);
@@ -291,7 +291,7 @@ static int load_aout_binary(struct linux_binprm * bprm)
if ((fd_offset & ~PAGE_MASK) != 0 && printk_ratelimit())
{
- printk(KERN_WARNING
+ printk(KERN_WARNING
"fd_offset is not page aligned. Please convert program: %pD\n",
bprm->file);
}
@@ -374,12 +374,12 @@ static int load_aout_library(struct file *file)
if ((N_TXTOFF(ex) & ~PAGE_MASK) != 0) {
if (printk_ratelimit())
{
- printk(KERN_WARNING
+ printk(KERN_WARNING
"N_TXTOFF is not page aligned. Please convert library: %pD\n",
file);
}
vm_brk(start_addr, ex.a_text + ex.a_data + ex.a_bss);
-
+
read_code(file, start_addr, N_TXTOFF(ex),
ex.a_text + ex.a_data);
retval = 0;
@@ -224,7 +224,7 @@ create_elf_tables(struct linux_binprm *bprm, struct elfhdr *exec,
} while (0)
#ifdef ARCH_DLINFO
- /*
+ /*
* ARCH_DLINFO must come first so PPC can do its special alignment of
* AUXV.
* update AT_VECTOR_SIZE_ARCH if the number of NEW_AUX_ENT() in
@@ -691,7 +691,7 @@ static int load_elf_binary(struct linux_binprm *bprm)
retval = -ENOMEM;
goto out_ret;
}
-
+
/* Get the exec-header */
loc->elf_ex = *((struct elfhdr *)bprm->buf);
@@ -727,7 +727,7 @@ static int load_elf_binary(struct linux_binprm *bprm)
* is an a.out format binary
*/
retval = -ENOEXEC;
- if (elf_ppnt->p_filesz > PATH_MAX ||
+ if (elf_ppnt->p_filesz > PATH_MAX ||
elf_ppnt->p_filesz < 2)
goto out_free_ph;
@@ -859,7 +859,7 @@ static int load_elf_binary(struct linux_binprm *bprm)
executable_stack);
if (retval < 0)
goto out_free_dentry;
-
+
current->mm->start_stack = bprm->p;
/* Now we do a little grungy work by mmapping the ELF image into
@@ -875,7 +875,7 @@ static int load_elf_binary(struct linux_binprm *bprm)
if (unlikely (elf_brk > elf_bss)) {
unsigned long nbyte;
-
+
/* There was a PT_LOAD segment with p_memsz > p_filesz
before this one. Map anonymous pages, if needed,
and clear the area. */
@@ -1390,7 +1390,7 @@ static void fill_elf_note_phdr(struct elf_phdr *phdr, int sz, loff_t offset)
return;
}
-static void fill_note(struct memelfnote *note, const char *name, int type,
+static void fill_note(struct memelfnote *note, const char *name, int type,
unsigned int sz, void *data)
{
note->name = name;
@@ -1442,7 +1442,7 @@ static int fill_psinfo(struct elf_prpsinfo *psinfo, struct task_struct *p,
{
const struct cred *cred;
unsigned int i, len;
-
+
/* first copy the parameters from user space */
memset(psinfo, 0, sizeof(struct elf_prpsinfo));
@@ -1476,7 +1476,7 @@ static int fill_psinfo(struct elf_prpsinfo *psinfo, struct task_struct *p,
SET_GID(psinfo->pr_gid, from_kgid_munged(cred->user_ns, cred->gid));
rcu_read_unlock();
strncpy(psinfo->pr_fname, p->comm, sizeof(psinfo->pr_fname));
-
+
return 0;
}
@@ -1878,8 +1878,8 @@ static int elf_dump_thread_status(long signr, struct elf_thread_status *t)
t->num_notes = 0;
fill_prstatus(&t->prstatus, p, signr);
- elf_core_copy_task_regs(p, &t->prstatus.pr_reg);
-
+ elf_core_copy_task_regs(p, &t->prstatus.pr_reg);
+
fill_note(&t->notes[0], "CORE", NT_PRSTATUS, sizeof(t->prstatus),
&(t->prstatus));
t->num_notes++;
@@ -1900,7 +1900,7 @@ static int elf_dump_thread_status(long signr, struct elf_thread_status *t)
t->num_notes++;
sz += notesize(&t->notes[2]);
}
-#endif
+#endif
return sz;
}
@@ -2143,7 +2143,7 @@ static int elf_core_dump(struct coredump_params *cprm)
/*
* We no longer stop all VM operations.
- *
+ *
* This is because those proceses that could possibly change map_count
* or the mmap / vma pages are now blocked in do_exit on current
* finishing this core dump.
@@ -2152,7 +2152,7 @@ static int elf_core_dump(struct coredump_params *cprm)
* the map_count or the pages allocated. So no possibility of crashing
* exists while dumping the mm->vm_next areas to the core file.
*/
-
+
/* alloc memory for large data structures: too large to be on stack */
elf = kmalloc(sizeof(*elf), GFP_KERNEL);
if (!elf)
@@ -67,11 +67,11 @@ static int load_em86(struct linux_binprm *bprm)
*/
remove_arg_zero(bprm);
retval = copy_strings_kernel(1, &bprm->filename, bprm);
- if (retval < 0) return retval;
+ if (retval < 0) return retval;
bprm->argc++;
if (i_arg) {
retval = copy_strings_kernel(1, &i_arg, bprm);
- if (retval < 0) return retval;
+ if (retval < 0) return retval;
bprm->argc++;
}
retval = copy_strings_kernel(1, &i_name, bprm);
@@ -387,7 +387,7 @@ static void old_reloc(unsigned long rl)
#endif
flat_v2_reloc_t r;
unsigned long *ptr;
-
+
r.value = rl;
#if defined(CONFIG_COLDFIRE)
ptr = (unsigned long *) (current->mm->start_code + r.reloc.offset);
@@ -400,7 +400,7 @@ static void old_reloc(unsigned long rl)
"(address %p, currently %x) into segment %s\n",
r.reloc.offset, ptr, (int)*ptr, segment[r.reloc.type]);
#endif
-
+
switch (r.reloc.type) {
case OLD_FLAT_RELOC_TYPE_TEXT:
*ptr += current->mm->start_code;
@@ -419,7 +419,7 @@ static void old_reloc(unsigned long rl)
#ifdef DEBUG
printk("Relocation became %x\n", (int)*ptr);
#endif
-}
+}
/****************************************************************************/
@@ -478,7 +478,7 @@ static int load_flat_file(struct linux_binprm * bprm,
ret = -ENOEXEC;
goto err;
}
-
+
/* Don't allow old format executables to use shared libraries */
if (rev == OLD_FLAT_VERSION && id != 0) {
printk("BINFMT_FLAT: shared libraries are not available before rev 0x%x\n",
@@ -580,7 +580,7 @@ static int load_flat_file(struct linux_binprm * bprm,
fpos = ntohl(hdr->data_start);
#ifdef CONFIG_BINFMT_ZFLAT
if (flags & FLAT_FLAG_GZDATA) {
- result = decompress_exec(bprm, fpos, (char *) datapos,
+ result = decompress_exec(bprm, fpos, (char *) datapos,
full_data, 0);
} else
#endif
@@ -703,7 +703,7 @@ static int load_flat_file(struct linux_binprm * bprm,
libinfo->lib_list[id].loaded = 1;
libinfo->lib_list[id].entry = (0x00ffffff & ntohl(hdr->entry)) + textpos;
libinfo->lib_list[id].build_date = ntohl(hdr->build_date);
-
+
/*
* We just load the allocations into some temporary memory to
* help simplify all this mumbo jumbo
@@ -783,11 +783,11 @@ static int load_flat_file(struct linux_binprm * bprm,
for (i=0; i < relocs; i++)
old_reloc(ntohl(reloc[i]));
}
-
+
flush_icache_range(start_code, end_code);
/* zero the BSS, BRK and stack areas */
- memset((void*)(datapos + data_len), 0, bss_len +
+ memset((void*)(datapos + data_len), 0, bss_len +
(memp + memp_size - stack_len - /* end brk */
libinfo->lib_list[id].start_brk) + /* start brk */
stack_len);
@@ -881,11 +881,11 @@ static int load_flat_binary(struct linux_binprm * bprm)
stack_len += (bprm->argc + 1) * sizeof(char *); /* the argv array */
stack_len += (bprm->envc + 1) * sizeof(char *); /* the envp array */
stack_len += FLAT_STACK_ALIGN - 1; /* reserve for upcoming alignment */
-
+
res = load_flat_file(bprm, &libinfo, 0, &stack_len);
if (IS_ERR_VALUE(res))
return res;
-
+
/* Update data segment pointers for all libraries */
for (i=0; i<MAX_SHARED_LIBS; i++)
if (libinfo.lib_list[i].loaded)
@@ -907,7 +907,7 @@ static int load_flat_binary(struct linux_binprm * bprm)
((char *) page_address(bprm->page[i/PAGE_SIZE]))[i % PAGE_SIZE];
sp = (unsigned long *) create_flat_tables(p, bprm);
-
+
/* Fake some return addresses to ensure the call chain will
* initialise library in order for us. We are required to call
* lib 1 first, then 2, ... and finally the main program (id 0).
@@ -923,7 +923,7 @@ static int load_flat_binary(struct linux_binprm * bprm)
}
}
#endif
-
+
/* Stash our initial stack pointer into the mm structure */
current->mm->start_stack = (unsigned long )sp;
@@ -932,7 +932,7 @@ static int load_flat_binary(struct linux_binprm * bprm)
#endif
DBG_FLT("start_thread(regs=0x%x, entry=0x%x, start_stack=0x%x)\n",
(int)regs, (int)start_addr, (int)current->mm->start_stack);
-
+
start_thread(regs, start_addr, current->mm->start_stack);
return 0;
@@ -55,7 +55,7 @@ static int load_script(struct linux_binprm *bprm)
break;
}
for (cp = bprm->buf+2; (*cp == ' ') || (*cp == '\t'); cp++);
- if (*cp == '\0')
+ if (*cp == '\0')
return -ENOEXEC; /* No interpreter name found */
i_name = cp;
i_arg = NULL;
@@ -80,15 +80,15 @@ static int load_script(struct linux_binprm *bprm)
if (retval)
return retval;
retval = copy_strings_kernel(1, &bprm->interp, bprm);
- if (retval < 0) return retval;
+ if (retval < 0) return retval;
bprm->argc++;
if (i_arg) {
retval = copy_strings_kernel(1, &i_arg, bprm);
- if (retval < 0) return retval;
+ if (retval < 0) return retval;
bprm->argc++;
}
retval = copy_strings_kernel(1, &i_name, bprm);
- if (retval) return retval;
+ if (retval) return retval;
bprm->argc++;
retval = bprm_change_interp(interp, bprm);
if (retval < 0)
@@ -80,7 +80,7 @@ void kill_bdev(struct block_device *bdev)
invalidate_bh_lrus();
truncate_inode_pages(mapping, 0);
-}
+}
EXPORT_SYMBOL(kill_bdev);
/* Invalidate clean unused buffers and pagecache. */
@@ -351,13 +351,13 @@ static loff_t block_llseek(struct file *file, loff_t offset, int whence)
inode_unlock(bd_inode);
return retval;
}
-
+
int blkdev_fsync(struct file *filp, loff_t start, loff_t end, int datasync)
{
struct inode *bd_inode = bdev_file_inode(filp);
struct block_device *bdev = I_BDEV(bd_inode);
int error;
-
+
error = filemap_write_and_wait_range(filp->f_mapping, start, end);
if (error)
return error;
@@ -692,7 +692,7 @@ void bdput(struct block_device *bdev)
}
EXPORT_SYMBOL(bdput);
-
+
static struct block_device *bd_acquire(struct inode *inode)
{
struct block_device *bdev;
@@ -439,7 +439,7 @@ EXPORT_SYMBOL(mark_buffer_async_write);
* try_to_free_buffers() will be operating against the *blockdev* mapping
* at the time, not against the S_ISREG file which depends on those buffers.
* So the locking for private_list is via the private_lock in the address_space
- * which backs the buffers. Which is different from the address_space
+ * which backs the buffers. Which is different from the address_space
* against which the buffers are listed. So for a particular address_space,
* mapping->private_lock does *not* protect mapping->private_list! In fact,
* mapping->private_list will always be protected by the backing blockdev's
@@ -713,7 +713,7 @@ EXPORT_SYMBOL(__set_page_dirty_buffers);
* Do this in two main stages: first we copy dirty buffers to a
* temporary inode list, queueing the writes as we go. Then we clean
* up, waiting for those writes to complete.
- *
+ *
* During this second stage, any subsequent updates to the file may end
* up refiling the buffer on the original inode's dirty list again, so
* there is a chance we will end up with a buffer queued for write but
@@ -791,7 +791,7 @@ static int fsync_buffers_list(spinlock_t *lock, struct list_head *list)
brelse(bh);
spin_lock(lock);
}
-
+
spin_unlock(lock);
err2 = osync_buffers_list(lock, list);
if (err)
@@ -901,7 +901,7 @@ no_grow:
/*
* Return failure for non-async IO requests. Async IO requests
* are not allowed to fail, so we have to wait until buffer heads
- * become available. But we don't want tasks sleeping with
+ * become available. But we don't want tasks sleeping with
* partially complete buffers, so all were released above.
*/
if (!retry)
@@ -910,7 +910,7 @@ no_grow:
/* We're _really_ low on memory. Now we just
* wait for old buffer heads to become free due to
* finishing IO. Since this is an async request and
- * the reserve list is empty, we're sure there are
+ * the reserve list is empty, we're sure there are
* async buffer heads in use.
*/
free_more_memory();
@@ -946,7 +946,7 @@ static sector_t blkdev_max_block(struct block_device *bdev, unsigned int size)
/*
* Initialise the state of a blockdev page's buffers.
- */
+ */
static sector_t
init_page_buffers(struct page *page, struct block_device *bdev,
sector_t block, int size)
@@ -1448,7 +1448,7 @@ static bool has_bh_in_lru(int cpu, void *dummy)
{
struct bh_lru *b = per_cpu_ptr(&bh_lrus, cpu);
int i;
-
+
for (i = 0; i < BH_LRU_SIZE; i++) {
if (b->bhs[i])
return 1;
@@ -1952,7 +1952,7 @@ int __block_write_begin(struct page *page, loff_t pos, unsigned len,
if (PageUptodate(page)) {
if (!buffer_uptodate(bh))
set_buffer_uptodate(bh);
- continue;
+ continue;
}
if (!buffer_uptodate(bh) && !buffer_delay(bh) &&
!buffer_unwritten(bh) &&
@@ -2258,7 +2258,7 @@ EXPORT_SYMBOL(block_read_full_page);
/* utility function for filesystems that need to do work on expanding
* truncates. Uses filesystem pagecache writes to allow the filesystem to
- * deal with the hole.
+ * deal with the hole.
*/
int generic_cont_expand_simple(struct inode *inode, loff_t size)
{
@@ -2819,7 +2819,7 @@ int block_truncate_page(struct address_space *mapping,
length = blocksize - length;
iblock = (sector_t)index << (PAGE_CACHE_SHIFT - inode->i_blkbits);
-
+
page = grab_cache_page(mapping, index);
err = -ENOMEM;
if (!page)
@@ -3069,7 +3069,7 @@ EXPORT_SYMBOL(submit_bh);
*
* ll_rw_block sets b_end_io to simple completion handler that marks
* the buffer up-to-date (if appropriate), unlocks the buffer and wakes
- * any waiters.
+ * any waiters.
*
* All of the buffers must be for the same device, and must also be a
* multiple of the current approved size for the device.
@@ -7,7 +7,7 @@
* Copyright (C) 2002 Stephen Rothwell, IBM Corporation
* Copyright (C) 1997-2000 Jakub Jelinek (jakub@redhat.com)
* Copyright (C) 1998 Eddie C. Dost (ecd@skynet.be)
- * Copyright (C) 2001,2002 Andi Kleen, SuSE Labs
+ * Copyright (C) 2001,2002 Andi Kleen, SuSE Labs
* Copyright (C) 2003 Pavel Machek (pavel@ucw.cz)
*
* This program is free software; you can redistribute it and/or modify
@@ -211,7 +211,7 @@ COMPAT_SYSCALL_DEFINE2(newfstat, unsigned int, fd,
static int put_compat_statfs(struct compat_statfs __user *ubuf, struct kstatfs *kbuf)
{
-
+
if (sizeof ubuf->f_blocks == 4) {
if ((kbuf->f_blocks | kbuf->f_bfree | kbuf->f_bavail |
kbuf->f_bsize | kbuf->f_frsize) & 0xffffffff00000000ULL)
@@ -536,7 +536,7 @@ COMPAT_SYSCALL_DEFINE5(io_getevents, compat_aio_context_t, ctx_id,
ut = compat_alloc_user_space(sizeof(*ut));
if (copy_to_user(ut, &t, sizeof(t)) )
return -EFAULT;
- }
+ }
return sys_io_getevents(ctx_id, min_nr, nr, events, ut);
}
@@ -636,7 +636,7 @@ copy_iocb(long nr, u32 __user *ptr32, struct iocb __user * __user *ptr64)
COMPAT_SYSCALL_DEFINE3(io_submit, compat_aio_context_t, ctx_id,
int, nr, u32 __user *, iocb)
{
- struct iocb __user * __user *iocb64;
+ struct iocb __user * __user *iocb64;
long ret;
if (unlikely(nr < 0))
@@ -644,7 +644,7 @@ COMPAT_SYSCALL_DEFINE3(io_submit, compat_aio_context_t, ctx_id,
if (nr > MAX_AIO_SUBMITS)
nr = MAX_AIO_SUBMITS;
-
+
iocb64 = compat_alloc_user_space(nr * sizeof(*iocb64));
ret = copy_iocb(nr, iocb, iocb64);
if (!ret)
@@ -3,7 +3,7 @@
*
* Copyright (C) 1997-2000 Jakub Jelinek (jakub@redhat.com)
* Copyright (C) 1998 Eddie C. Dost (ecd@skynet.be)
- * Copyright (C) 2001,2002 Andi Kleen, SuSE Labs
+ * Copyright (C) 2001,2002 Andi Kleen, SuSE Labs
* Copyright (C) 2003 Pavel Machek (pavel@ucw.cz)
*
* These routines maintain argument size conversion between 32bit and 64bit
@@ -254,14 +254,14 @@ static void __d_free(struct rcu_head *head)
{
struct dentry *dentry = container_of(head, struct dentry, d_u.d_rcu);
- kmem_cache_free(dentry_cache, dentry);
+ kmem_cache_free(dentry_cache, dentry);
}
static void __d_free_external(struct rcu_head *head)
{
struct dentry *dentry = container_of(head, struct dentry, d_u.d_rcu);
kfree(external_name(dentry));
- kmem_cache_free(dentry_cache, dentry);
+ kmem_cache_free(dentry_cache, dentry);
}
static inline int dname_external(const struct dentry *dentry)
@@ -731,7 +731,7 @@ static inline bool fast_dput(struct dentry *dentry)
}
-/*
+/*
* This is dput
*
* This is complicated by the fact that we do not want to put
@@ -750,7 +750,7 @@ static inline bool fast_dput(struct dentry *dentry)
/*
* dput - release a dentry
- * @dentry: dentry to release
+ * @dentry: dentry to release
*
* Release a dentry. This will drop the usage count and if appropriate
* call the dentry unlink method as well as removing it from the queues and
@@ -1552,7 +1552,7 @@ EXPORT_SYMBOL(d_invalidate);
* available. On a success the dentry is returned. The name passed in is
* copied and the copy passed in may be reused after this call.
*/
-
+
struct dentry *__d_alloc(struct super_block *sb, const struct qstr *name)
{
struct dentry *dentry;
@@ -1574,7 +1574,7 @@ struct dentry *__d_alloc(struct super_block *sb, const struct qstr *name)
struct external_name *p = kmalloc(size + name->len,
GFP_KERNEL_ACCOUNT);
if (!p) {
- kmem_cache_free(dentry_cache, dentry);
+ kmem_cache_free(dentry_cache, dentry);
return NULL;
}
atomic_set(&p->u.count, 1);
@@ -1584,7 +1584,7 @@ struct dentry *__d_alloc(struct super_block *sb, const struct qstr *name)
round_up(name->len + 1, sizeof(unsigned long)));
} else {
dname = dentry->d_iname;
- }
+ }
dentry->d_name.len = name->len;
dentry->d_name.hash = name->hash;
@@ -1778,7 +1778,7 @@ static void __d_instantiate(struct dentry *dentry, struct inode *inode)
* (or otherwise set) by the caller to indicate that it is now
* in use by the dcache.
*/
-
+
void d_instantiate(struct dentry *entry, struct inode * inode)
{
BUG_ON(!hlist_unhashed(&entry->d_u.d_alias));
@@ -2300,7 +2300,7 @@ struct dentry *__d_lookup(const struct dentry *parent, const struct qstr *name)
* See Documentation/filesystems/path-lookup.txt for more details.
*/
rcu_read_lock();
-
+
hlist_bl_for_each_entry_rcu(dentry, node, b, d_hash) {
if (dentry->d_name.hash != hash)
@@ -2376,7 +2376,7 @@ EXPORT_SYMBOL(d_hash_and_lookup);
* it from the hash queues and waiting for
* it to be deleted later when it has no users
*/
-
+
/**
* d_delete - delete a dentry
* @dentry: The dentry to delete
@@ -2384,7 +2384,7 @@ EXPORT_SYMBOL(d_hash_and_lookup);
* Turn the dentry into a negative dentry if possible, otherwise
* remove it from the hash queues so it can be deleted later
*/
-
+
void d_delete(struct dentry * dentry)
{
struct inode *inode;
@@ -2437,7 +2437,7 @@ static void _d_rehash(struct dentry * entry)
*
* Adds a dentry to the hash according to its name.
*/
-
+
void d_rehash(struct dentry * entry)
{
spin_lock(&entry->d_lock);
@@ -3136,7 +3136,7 @@ char *simple_dname(struct dentry *dentry, char *buffer, int buflen)
/* these dentries are never renamed, so d_lock is not needed */
if (prepend(&end, &buflen, " (deleted)", 11) ||
prepend(&end, &buflen, dentry->d_name.name, dentry->d_name.len) ||
- prepend(&end, &buflen, "/", 1))
+ prepend(&end, &buflen, "/", 1))
end = ERR_PTR(-ENAMETOOLONG);
return end;
}
@@ -3308,7 +3308,7 @@ out:
* Returns false otherwise.
* Caller must ensure that "new_dentry" is pinned before calling is_subdir()
*/
-
+
bool is_subdir(struct dentry *new_dentry, struct dentry *old_dentry)
{
bool result;
@@ -3410,10 +3410,10 @@ static void __init dcache_init(void)
{
unsigned int loop;
- /*
+ /*
* A constructor could be added for stable state like the lists,
* but it is probably not worth it because of the cache nature
- * of the dcache.
+ * of the dcache.
*/
dentry_cache = KMEM_CACHE(dentry,
SLAB_RECLAIM_ACCOUNT|SLAB_PANIC|SLAB_MEM_SPREAD|SLAB_ACCOUNT);
@@ -187,7 +187,7 @@ SYSCALL_DEFINE3(lookup_dcookie, u64, cookie64, char __user *, buf, size_t, len)
}
err = -ERANGE;
-
+
pathlen = kbuf + PAGE_SIZE - path;
if (pathlen <= len) {
err = pathlen;
@@ -306,7 +306,7 @@ static void dcookie_exit(void)
struct dcookie_user {
struct list_head next;
};
-
+
struct dcookie_user * dcookie_register(void)
{
struct dcookie_user * user;
@@ -189,7 +189,7 @@ static inline int dio_refill_pages(struct dio *dio, struct dio_submit *sdio)
sdio->to = ((ret - 1) & (PAGE_SIZE - 1)) + 1;
return 0;
}
- return ret;
+ return ret;
}
/*
@@ -286,7 +286,7 @@ static void dio_aio_complete_work(struct work_struct *work)
static int dio_bio_complete(struct dio *dio, struct bio *bio);
/*
- * Asynchronous IO callback.
+ * Asynchronous IO callback.
*/
static void dio_bio_end_aio(struct bio *bio)
{
@@ -699,7 +699,7 @@ static inline int dio_bio_add_page(struct dio_submit *sdio)
}
return ret;
}
-
+
/*
* Put cur_page under IO. The section of cur_page which is described by
* cur_page_offset,cur_page_len is put into a BIO. The section of cur_page
@@ -761,7 +761,7 @@ out:
* An autonomous function to put a chunk of a page under deferred IO.
*
* The caller doesn't actually know (or care) whether this piece of page is in
- * a BIO, or is under IO or whatever. We just take care of all possible
+ * a BIO, or is under IO or whatever. We just take care of all possible
* situations here. The separation between the logic of do_direct_IO() and
* that of submit_page_section() is important for clarity. Please don't break.
*
@@ -879,7 +879,7 @@ static inline void dio_zero_block(struct dio *dio, struct dio_submit *sdio,
* We need to zero out part of an fs block. It is either at the
* beginning or the end of the fs block.
*/
- if (end)
+ if (end)
this_chunk_blocks = dio_blocks_per_fs_block - this_chunk_blocks;
this_chunk_bytes = this_chunk_blocks << sdio->blkbits;
@@ -19,7 +19,7 @@
* current->executable is only used by the procfs. This allows a dispatch
* table to check for several different types of binary formats. We keep
* trying until we recognize the file or we run out of supported binary
- * formats.
+ * formats.
*/
#include <linux/slab.h>
@@ -355,7 +355,7 @@ static int check_fcntl_cmd(unsigned cmd)
}
SYSCALL_DEFINE3(fcntl, unsigned int, fd, unsigned int, cmd, unsigned long, arg)
-{
+{
struct fd f = fdget_raw(fd);
long err = -EBADF;
@@ -380,7 +380,7 @@ out:
#if BITS_PER_LONG == 32
SYSCALL_DEFINE3(fcntl64, unsigned int, fd, unsigned int, cmd,
unsigned long, arg)
-{
+{
struct fd f = fdget_raw(fd);
long err = -EBADF;
@@ -395,7 +395,7 @@ SYSCALL_DEFINE3(fcntl64, unsigned int, fd, unsigned int, cmd,
err = security_file_fcntl(f.file, cmd, arg);
if (err)
goto out1;
-
+
switch (cmd) {
case F_GETLK64:
case F_OFD_GETLK:
@@ -463,8 +463,8 @@ static void send_sigio_to_task(struct task_struct *p,
siginfo_t si;
default:
/* Queue a rt signal with the appropriate fd as its
- value. We use SI_SIGIO as the source, not
- SI_KERNEL, since kernel signals always get
+ value. We use SI_SIGIO as the source, not
+ SI_KERNEL, since kernel signals always get
delivered even if we can't queue. Failure to
queue in this case _should_ be reported; we fall
back to SIGIO in that case. --sct */
@@ -494,7 +494,7 @@ void send_sigio(struct fown_struct *fown, int fd, int band)
enum pid_type type;
struct pid *pid;
int group = 1;
-
+
read_lock(&fown->lock);
type = fown->pid_type;
@@ -506,7 +506,7 @@ void send_sigio(struct fown_struct *fown, int fd, int band)
pid = fown->pid;
if (!pid)
goto out_unlock_fown;
-
+
read_lock(&tasklist_lock);
do_each_pid_task(pid, type, p) {
send_sigio_to_task(p, fown, fd, band, group);
@@ -530,7 +530,7 @@ int send_sigurg(struct fown_struct *fown)
struct pid *pid;
int group = 1;
int ret = 0;
-
+
read_lock(&fown->lock);
type = fown->pid_type;
@@ -544,7 +544,7 @@ int send_sigurg(struct fown_struct *fown)
goto out_unlock_fown;
ret = 1;
-
+
read_lock(&tasklist_lock);
do_each_pid_task(pid, type, p) {
send_sigurg_to_task(p, fown, group);
@@ -310,7 +310,7 @@ void put_filp(struct file *file)
}
void __init files_init(void)
-{
+{
filp_cachep = kmem_cache_create("filp", sizeof(struct file), 0,
SLAB_HWCACHE_ALIGN | SLAB_PANIC, NULL);
percpu_counter_init(&nr_files, 0, GFP_KERNEL);
@@ -329,4 +329,4 @@ void __init files_maxfiles_init(void)
n = ((totalram_pages - memreserve) * (PAGE_SIZE / 1024)) / 10;
files_stat.max_files = max_t(unsigned long, n, NR_FILE);
-}
+}
@@ -61,11 +61,11 @@ static struct file_system_type **find_filesystem(const char *name, unsigned len)
* is aware of for mount and other syscalls. Returns 0 on success,
* or a negative errno code on an error.
*
- * The &struct file_system_type that is passed is linked into the kernel
+ * The &struct file_system_type that is passed is linked into the kernel
* structures and must not be freed until the file system has been
* unregistered.
*/
-
+
int register_filesystem(struct file_system_type * fs)
{
int res = 0;
@@ -93,11 +93,11 @@ EXPORT_SYMBOL(register_filesystem);
* Remove a file system that was previously successfully registered
* with the kernel. An error is returned if the file system is not found.
* Zero is returned on a success.
- *
+ *
* Once this function has returned the &struct file_system_type structure
* may be freed or reused.
*/
-
+
int unregister_filesystem(struct file_system_type * fs)
{
struct file_system_type ** tmp;
@@ -179,7 +179,7 @@ static int fs_maxindex(void)
}
/*
- * Whee.. Weird sysv syscall.
+ * Whee.. Weird sysv syscall.
*/
SYSCALL_DEFINE3(sysfs, int, option, unsigned long, arg1, unsigned long, arg2)
{
@@ -11,11 +11,11 @@
*
* Miscellaneous edits, and a total rewrite of posix_lock_file() code.
* Kai Petzke (wpp@marie.physik.tu-berlin.de), 1994
- *
+ *
* Converted file_lock_table to a linked list from an array, which eliminates
* the limits on how many active file locks are open.
* Chad Page (pageone@netcom.com), November 27, 1994
- *
+ *
* Removed dependency on file descriptors. dup()'ed file descriptors now
* get the same locks as the original file descriptors, and a close() on
* any file descriptor removes ALL the locks on the file for the current
@@ -41,7 +41,7 @@
* with a file pointer (filp). As a result they can be shared by a parent
* process and its children after a fork(). They are removed when the last
* file descriptor referring to the file pointer is closed (unless explicitly
- * unlocked).
+ * unlocked).
*
* FL_FLOCK locks never deadlock, an existing lock is always removed before
* upgrading from shared to exclusive (or vice versa). When this happens
@@ -50,7 +50,7 @@
* Andy Walker (andy@lysaker.kvaerner.no), June 09, 1995
*
* Removed some race conditions in flock_lock_file(), marked other possible
- * races. Just grep for FIXME to see them.
+ * races. Just grep for FIXME to see them.
* Dmitry Gorodchanin (pgmdsg@ibi.com), February 09, 1996.
*
* Addressed Dmitry's concerns. Deadlock checking no longer recursive.
@@ -405,7 +405,7 @@ flock_make_lock(struct file *filp, unsigned int cmd)
if (type < 0)
return ERR_PTR(type);
-
+
fl = locks_alloc_lock();
if (fl == NULL)
return ERR_PTR(-ENOMEM);
@@ -416,7 +416,7 @@ flock_make_lock(struct file *filp, unsigned int cmd)
fl->fl_flags = FL_FLOCK;
fl->fl_type = type;
fl->fl_end = OFFSET_MAX;
-
+
return fl;
}
@@ -2078,7 +2078,7 @@ int fcntl_getlk(struct file *filp, unsigned int cmd, struct flock __user *l)
error = vfs_test_lock(filp, &file_lock);
if (error)
goto out;
-
+
flock.l_type = file_lock.fl_type;
if (file_lock.fl_type != F_UNLCK) {
error = posix_lock_to_flock(&flock, &file_lock);
@@ -587,7 +587,7 @@ mb_cache_entry_insert(struct mb_cache_entry *ce, struct block_device *bdev,
struct mb_cache_entry *lce;
mb_assert(ce);
- bucket = hash_long((unsigned long)bdev + (block & 0xffffffff),
+ bucket = hash_long((unsigned long)bdev + (block & 0xffffffff),
cache->c_bucket_bits);
block_hash_p = &cache->c_block_hash[bucket];
hlist_bl_lock(block_hash_p);
@@ -94,8 +94,8 @@ mpage_alloc(struct block_device *bdev,
* them. So when the buffer is up to date and the page size == block size,
* this marks the page up to date instead of adding new buffers.
*/
-static void
-map_buffer_to_page(struct page *page, struct buffer_head *bh, int page_block)
+static void
+map_buffer_to_page(struct page *page, struct buffer_head *bh, int page_block)
{
struct inode *inode = page->mapping->host;
struct buffer_head *page_bh, *head;
@@ -106,9 +106,9 @@ map_buffer_to_page(struct page *page, struct buffer_head *bh, int page_block)
* don't make any buffers if there is only one buffer on
* the page and the page just needs to be set up to date
*/
- if (inode->i_blkbits == PAGE_CACHE_SHIFT &&
+ if (inode->i_blkbits == PAGE_CACHE_SHIFT &&
buffer_uptodate(bh)) {
- SetPageUptodate(page);
+ SetPageUptodate(page);
return;
}
create_empty_buffers(page, 1 << inode->i_blkbits, 0);
@@ -226,7 +226,7 @@ do_mpage_readpage(struct bio *bio, struct page *page, unsigned nr_pages,
map_buffer_to_page(page, map_bh, page_block);
goto confused;
}
-
+
if (first_hole != blocks_per_page)
goto confused; /* hole -> non-hole */
@@ -420,7 +420,7 @@ EXPORT_SYMBOL(mpage_readpage);
*
* If all blocks are found to be contiguous then the page can go into the
* BIO. Otherwise fall back to the mapping's writepage().
- *
+ *
* FIXME: This code wants an estimate of how many pages are still to be
* written, so it can intelligently allocate a suitably-sized BIO. For now,
* just allocate full-size (16-page) BIOs.
@@ -49,8 +49,8 @@
* The new code replaces the old recursive symlink resolution with
* an iterative one (in case of non-nested symlink chains). It does
* this with calls to <fs>_follow_link().
- * As a side effect, dir_namei(), _namei() and follow_link() are now
- * replaced with a single function lookup_dentry() that can handle all
+ * As a side effect, dir_namei(), _namei() and follow_link() are now
+ * replaced with a single function lookup_dentry() that can handle all
* the special cases of the former code.
*
* With the new dcache, the pathname is stored at each inode, at least as
@@ -1576,7 +1576,7 @@ out_unlock:
namespace_unlock();
}
-/*
+/*
* Is the caller allowed to modify his namespace?
*/
static inline bool may_mount(void)
@@ -2090,7 +2090,7 @@ static int do_loopback(struct path *path, const char *old_name,
err = -EINVAL;
if (mnt_ns_loop(old_path.dentry))
- goto out;
+ goto out;
mp = lock_mount(path);
err = PTR_ERR(mp);
@@ -875,7 +875,7 @@ struct file *dentry_open(const struct path *path, int flags,
fput(f);
f = ERR_PTR(error);
}
- } else {
+ } else {
put_filp(f);
f = ERR_PTR(error);
}
@@ -982,7 +982,7 @@ struct file *filp_open(const char *filename, int flags, umode_t mode)
{
struct filename *name = getname_kernel(filename);
struct file *file = ERR_CAST(name);
-
+
if (!IS_ERR(name)) {
file = file_open_name(name, flags, mode);
putname(name);
@@ -45,10 +45,10 @@ unsigned long pipe_user_pages_hard;
unsigned long pipe_user_pages_soft = PIPE_DEF_BUFFERS * INR_OPEN_CUR;
/*
- * We use a start+len construction, which provides full use of the
+ * We use a start+len construction, which provides full use of the
* allocated memory.
* -- Florian Coosmann (FGC)
- *
+ *
* Reads with count = 0 should always return 0.
* -- Julian Bradfield 1999-06-07.
*
@@ -842,7 +842,7 @@ SYSCALL_DEFINE1(pipe, int __user *, fildes)
static int wait_for_partner(struct pipe_inode_info *pipe, unsigned int *cnt)
{
- int cur = *cnt;
+ int cur = *cnt;
while (cur == *cnt) {
pipe_wait(pipe);
@@ -917,7 +917,7 @@ static int fifo_open(struct inode *inode, struct file *filp)
}
}
break;
-
+
case FMODE_WRITE:
/*
* O_WRONLY
@@ -937,7 +937,7 @@ static int fifo_open(struct inode *inode, struct file *filp)
goto err_wr;
}
break;
-
+
case FMODE_READ | FMODE_WRITE:
/*
* O_RDWR
@@ -240,7 +240,7 @@ static int propagate_one(struct mount *m)
if (IS_MNT_SHARED(m))
type |= CL_MAKE_SHARED;
}
-
+
/* Notice when we are propagating across user namespaces */
if (m->mnt_ns->user_ns != user_ns)
type |= CL_UNPRIVILEGED;
@@ -675,12 +675,12 @@ posix_acl_from_xattr(struct user_namespace *user_ns,
return ERR_PTR(-EINVAL);
if (count == 0)
return NULL;
-
+
acl = posix_acl_alloc(count, GFP_NOFS);
if (!acl)
return ERR_PTR(-ENOMEM);
acl_e = acl->a_entries;
-
+
for (end = entry + count; entry != end; acl_e++, entry++) {
acl_e->e_tag = le16_to_cpu(entry->e_tag);
acl_e->e_perm = le16_to_cpu(entry->e_perm);
@@ -4,7 +4,7 @@
* Copyright (C) 1991, 1992 Linus Torvalds
*/
-#include <linux/slab.h>
+#include <linux/slab.h>
#include <linux/stat.h>
#include <linux/fcntl.h>
#include <linux/file.h>
@@ -663,7 +663,7 @@ SYSCALL_DEFINE4(pwrite64, unsigned int, fd, const char __user *, buf,
f = fdget(fd);
if (f.file) {
ret = -ESPIPE;
- if (f.file->f_mode & FMODE_PWRITE)
+ if (f.file->f_mode & FMODE_PWRITE)
ret = vfs_write(f.file, buf, count, &pos);
fdput(f);
}
@@ -132,7 +132,7 @@ SYSCALL_DEFINE3(old_readdir, unsigned int, fd,
/*
* New, all-improved, singing, dancing, iBCS2-compliant getdents()
- * interface.
+ * interface.
*/
struct linux_dirent {
unsigned long d_ino;
@@ -10,7 +10,7 @@
* parameter to reflect time remaining.
*
* 24 January 2000
- * Changed sys_poll()/do_poll() to use PAGE_SIZE chunk-based allocation
+ * Changed sys_poll()/do_poll() to use PAGE_SIZE chunk-based allocation
* of fds to overcome nfds < 16390 descriptors limit (Tigran Aivazian).
*/
@@ -570,7 +570,7 @@ int core_sys_select(int n, fd_set __user *inp, fd_set __user *outp,
/*
* We need 6 bitmaps (in/out/ex for both incoming and outgoing),
* since we used fdset we need to allocate memory in units of
- * long-words.
+ * long-words.
*/
size = FDS_BYTES(n);
bits = stack_fds;
@@ -117,7 +117,7 @@ static int signalfd_copyinfo(struct signalfd_siginfo __user *uinfo,
err |= __put_user(kinfo->si_trapno, &uinfo->ssi_trapno);
#endif
#ifdef BUS_MCEERR_AO
- /*
+ /*
* Other callers might not initialize the si_lsb field,
* so check explicitly for the right codes here.
*/
@@ -141,7 +141,7 @@ static int cp_old_stat(struct kstat *stat, struct __old_kernel_stat __user * sta
{
static int warncount = 5;
struct __old_kernel_stat tmp;
-
+
if (warncount > 0) {
warncount--;
printk(KERN_WARNING "VFS: Warning: %s using old stat() call. Recompile your binary.\n",
@@ -166,7 +166,7 @@ static int cp_old_stat(struct kstat *stat, struct __old_kernel_stat __user * sta
#if BITS_PER_LONG == 32
if (stat->size > MAX_NON_LFS)
return -EOVERFLOW;
-#endif
+#endif
tmp.st_size = stat->size;
tmp.st_atime = stat->atime.tv_sec;
tmp.st_mtime = stat->mtime.tv_sec;
@@ -482,7 +482,7 @@ retry:
return ERR_PTR(-ENOMEM);
goto retry;
}
-
+
err = set(s, data);
if (err) {
spin_unlock(&sb_lock);
@@ -583,7 +583,7 @@ EXPORT_SYMBOL(iterate_supers_type);
/**
* get_super - get the superblock of a device
* @bdev: device to get the superblock for
- *
+ *
* Scans the superblock list and finds the superblock of the file system
* mounted on the device given. %NULL is returned if no match is found.
*/
@@ -673,7 +673,7 @@ restart:
spin_unlock(&sb_lock);
return NULL;
}
-
+
struct super_block *user_get_super(dev_t dev)
{
struct super_block *sb;
@@ -415,7 +415,7 @@ SYSCALL_DEFINE2(timerfd_create, int, clockid, int, flags)
return ufd;
}
-static int do_timerfd_settime(int ufd, int flags,
+static int do_timerfd_settime(int ufd, int flags,
const struct itimerspec *new,
struct itimerspec *old)
{
fixed all whitespaces in .c files in the fs folder (not in the sub-folders) Signed-off-by: amar-laksh <amarlakshya1@gmail.com> --- fs/aio.c | 8 ++++---- fs/bad_inode.c | 16 ++++++++-------- fs/binfmt_aout.c | 6 +++--- fs/binfmt_elf.c | 26 +++++++++++++------------- fs/binfmt_em86.c | 4 ++-- fs/binfmt_flat.c | 26 +++++++++++++------------- fs/binfmt_script.c | 8 ++++---- fs/block_dev.c | 8 ++++---- fs/buffer.c | 22 +++++++++++----------- fs/compat.c | 10 +++++----- fs/compat_ioctl.c | 2 +- fs/dcache.c | 32 ++++++++++++++++---------------- fs/dcookies.c | 4 ++-- fs/direct-io.c | 10 +++++----- fs/exec.c | 2 +- fs/fcntl.c | 18 +++++++++--------- fs/file_table.c | 4 ++-- fs/filesystems.c | 10 +++++----- fs/locks.c | 14 +++++++------- fs/mbcache.c | 2 +- fs/mpage.c | 12 ++++++------ fs/namei.c | 4 ++-- fs/namespace.c | 4 ++-- fs/open.c | 4 ++-- fs/pipe.c | 10 +++++----- fs/pnode.c | 2 +- fs/posix_acl.c | 4 ++-- fs/read_write.c | 4 ++-- fs/readdir.c | 2 +- fs/select.c | 4 ++-- fs/signalfd.c | 2 +- fs/stat.c | 4 ++-- fs/super.c | 6 +++--- fs/timerfd.c | 2 +- 34 files changed, 148 insertions(+), 148 deletions(-)