summaryrefslogtreecommitdiff
path: root/include
diff options
context:
space:
mode:
Diffstat (limited to 'include')
-rw-r--r--include/asm-ppc64/io.h26
-rw-r--r--include/asm-ppc64/page.h2
-rw-r--r--include/asm-ppc64/pci.h10
-rw-r--r--include/asm-ppc64/pgtable.h34
-rw-r--r--include/asm-ppc64/processor.h4
-rw-r--r--include/asm-ppc64/semaphore.h8
-rw-r--r--include/asm-ppc64/siginfo.h2
-rw-r--r--include/asm-ppc64/time.h3
-rw-r--r--include/asm-ppc64/tlbflush.h2
-rw-r--r--include/asm-ppc64/uaccess.h12
-rw-r--r--include/asm-ppc64/vga.h4
11 files changed, 54 insertions, 53 deletions
diff --git a/include/asm-ppc64/io.h b/include/asm-ppc64/io.h
index 031543d25dc0..8ad63a3e7d8f 100644
--- a/include/asm-ppc64/io.h
+++ b/include/asm-ppc64/io.h
@@ -134,7 +134,7 @@ extern void iounmap(void *addr);
* Change virtual addresses to physical addresses and vv, for
* addresses in the area where the kernel has the RAM mapped.
*/
-extern inline unsigned long virt_to_phys(volatile void * address)
+static inline unsigned long virt_to_phys(volatile void * address)
{
#ifdef __IO_DEBUG
printk("virt_to_phys: 0x%08lx -> 0x%08lx\n",
@@ -144,7 +144,7 @@ extern inline unsigned long virt_to_phys(volatile void * address)
return __pa((unsigned long)address);
}
-extern inline void * phys_to_virt(unsigned long address)
+static inline void * phys_to_virt(unsigned long address)
{
#ifdef __IO_DEBUG
printk("phys_to_virt: 0x%08lx -> 0x%08lx\n", address, __va(address));
@@ -163,7 +163,7 @@ extern inline void * phys_to_virt(unsigned long address)
#endif /* __KERNEL__ */
-extern inline void iosync(void)
+static inline void iosync(void)
{
__asm__ __volatile__ ("sync" : : : "memory");
}
@@ -178,7 +178,7 @@ extern inline void iosync(void)
/*
* 8, 16 and 32 bit, big and little endian I/O operations, with barrier.
*/
-extern inline int in_8(volatile unsigned char *addr)
+static inline int in_8(volatile unsigned char *addr)
{
int ret;
@@ -186,12 +186,12 @@ extern inline int in_8(volatile unsigned char *addr)
return ret;
}
-extern inline void out_8(volatile unsigned char *addr, int val)
+static inline void out_8(volatile unsigned char *addr, int val)
{
__asm__ __volatile__("stb%U0%X0 %1,%0" : "=m" (*addr) : "r" (val));
}
-extern inline int in_le16(volatile unsigned short *addr)
+static inline int in_le16(volatile unsigned short *addr)
{
int ret;
@@ -200,7 +200,7 @@ extern inline int in_le16(volatile unsigned short *addr)
return ret;
}
-extern inline int in_be16(volatile unsigned short *addr)
+static inline int in_be16(volatile unsigned short *addr)
{
int ret;
@@ -208,18 +208,18 @@ extern inline int in_be16(volatile unsigned short *addr)
return ret;
}
-extern inline void out_le16(volatile unsigned short *addr, int val)
+static inline void out_le16(volatile unsigned short *addr, int val)
{
__asm__ __volatile__("sthbrx %1,0,%2" : "=m" (*addr) :
"r" (val), "r" (addr));
}
-extern inline void out_be16(volatile unsigned short *addr, int val)
+static inline void out_be16(volatile unsigned short *addr, int val)
{
__asm__ __volatile__("sth%U0%X0 %1,%0" : "=m" (*addr) : "r" (val));
}
-extern inline unsigned in_le32(volatile unsigned *addr)
+static inline unsigned in_le32(volatile unsigned *addr)
{
unsigned ret;
@@ -228,7 +228,7 @@ extern inline unsigned in_le32(volatile unsigned *addr)
return ret;
}
-extern inline unsigned in_be32(volatile unsigned *addr)
+static inline unsigned in_be32(volatile unsigned *addr)
{
unsigned ret;
@@ -236,13 +236,13 @@ extern inline unsigned in_be32(volatile unsigned *addr)
return ret;
}
-extern inline void out_le32(volatile unsigned *addr, int val)
+static inline void out_le32(volatile unsigned *addr, int val)
{
__asm__ __volatile__("stwbrx %1,0,%2" : "=m" (*addr) :
"r" (val), "r" (addr));
}
-extern inline void out_be32(volatile unsigned *addr, int val)
+static inline void out_be32(volatile unsigned *addr, int val)
{
__asm__ __volatile__("stw%U0%X0 %1,%0" : "=m" (*addr) : "r" (val));
}
diff --git a/include/asm-ppc64/page.h b/include/asm-ppc64/page.h
index 8f650e4077bb..3a69566241e4 100644
--- a/include/asm-ppc64/page.h
+++ b/include/asm-ppc64/page.h
@@ -131,7 +131,7 @@ extern void xmon(struct pt_regs *excp);
#define PAGE_BUG(page) do { BUG(); } while (0)
/* Pure 2^n version of get_order */
-extern __inline__ int get_order(unsigned long size)
+static inline int get_order(unsigned long size)
{
int order;
diff --git a/include/asm-ppc64/pci.h b/include/asm-ppc64/pci.h
index 16a51c64d78c..6b06be2d3ff7 100644
--- a/include/asm-ppc64/pci.h
+++ b/include/asm-ppc64/pci.h
@@ -25,12 +25,12 @@ extern int pcibios_assign_all_busses(void);
#define PCIBIOS_MIN_IO 0x1000
#define PCIBIOS_MIN_MEM 0x10000000
-extern inline void pcibios_set_master(struct pci_dev *dev)
+static inline void pcibios_set_master(struct pci_dev *dev)
{
/* No special bus mastering setup handling */
}
-extern inline void pcibios_penalize_isa_irq(int irq)
+static inline void pcibios_penalize_isa_irq(int irq)
{
/* We don't do dynamic PCI IRQ allocation */
}
@@ -78,7 +78,7 @@ extern void pci_unmap_sg(struct pci_dev *hwdev, struct scatterlist *sg,
extern void pSeries_pcibios_init_early(void);
-extern inline void pci_dma_sync_single(struct pci_dev *hwdev,
+static inline void pci_dma_sync_single(struct pci_dev *hwdev,
dma_addr_t dma_handle,
size_t size, int direction)
{
@@ -87,7 +87,7 @@ extern inline void pci_dma_sync_single(struct pci_dev *hwdev,
/* nothing to do */
}
-extern inline void pci_dma_sync_sg(struct pci_dev *hwdev,
+static inline void pci_dma_sync_sg(struct pci_dev *hwdev,
struct scatterlist *sg,
int nelems, int direction)
{
@@ -101,7 +101,7 @@ extern inline void pci_dma_sync_sg(struct pci_dev *hwdev,
* only drive the low 24-bits during PCI bus mastering, then
* you would pass 0x00ffffff as the mask to this function.
*/
-extern inline int pci_dma_supported(struct pci_dev *hwdev, u64 mask)
+static inline int pci_dma_supported(struct pci_dev *hwdev, u64 mask)
{
return 1;
}
diff --git a/include/asm-ppc64/pgtable.h b/include/asm-ppc64/pgtable.h
index 628cb54c529c..e7dcec1e8b04 100644
--- a/include/asm-ppc64/pgtable.h
+++ b/include/asm-ppc64/pgtable.h
@@ -246,35 +246,35 @@ extern unsigned long empty_zero_page[PAGE_SIZE/sizeof(unsigned long)];
* The following only work if pte_present() is true.
* Undefined behaviour if not..
*/
-extern inline int pte_read(pte_t pte) { return pte_val(pte) & _PAGE_USER;}
-extern inline int pte_write(pte_t pte) { return pte_val(pte) & _PAGE_RW;}
-extern inline int pte_exec(pte_t pte) { return pte_val(pte) & _PAGE_EXEC;}
-extern inline int pte_dirty(pte_t pte) { return pte_val(pte) & _PAGE_DIRTY;}
-extern inline int pte_young(pte_t pte) { return pte_val(pte) & _PAGE_ACCESSED;}
+static inline int pte_read(pte_t pte) { return pte_val(pte) & _PAGE_USER;}
+static inline int pte_write(pte_t pte) { return pte_val(pte) & _PAGE_RW;}
+static inline int pte_exec(pte_t pte) { return pte_val(pte) & _PAGE_EXEC;}
+static inline int pte_dirty(pte_t pte) { return pte_val(pte) & _PAGE_DIRTY;}
+static inline int pte_young(pte_t pte) { return pte_val(pte) & _PAGE_ACCESSED;}
-extern inline void pte_uncache(pte_t pte) { pte_val(pte) |= _PAGE_NO_CACHE; }
-extern inline void pte_cache(pte_t pte) { pte_val(pte) &= ~_PAGE_NO_CACHE; }
+static inline void pte_uncache(pte_t pte) { pte_val(pte) |= _PAGE_NO_CACHE; }
+static inline void pte_cache(pte_t pte) { pte_val(pte) &= ~_PAGE_NO_CACHE; }
-extern inline pte_t pte_rdprotect(pte_t pte) {
+static inline pte_t pte_rdprotect(pte_t pte) {
pte_val(pte) &= ~_PAGE_USER; return pte; }
-extern inline pte_t pte_exprotect(pte_t pte) {
+static inline pte_t pte_exprotect(pte_t pte) {
pte_val(pte) &= ~_PAGE_EXEC; return pte; }
-extern inline pte_t pte_wrprotect(pte_t pte) {
+static inline pte_t pte_wrprotect(pte_t pte) {
pte_val(pte) &= ~(_PAGE_RW); return pte; }
-extern inline pte_t pte_mkclean(pte_t pte) {
+static inline pte_t pte_mkclean(pte_t pte) {
pte_val(pte) &= ~(_PAGE_DIRTY); return pte; }
-extern inline pte_t pte_mkold(pte_t pte) {
+static inline pte_t pte_mkold(pte_t pte) {
pte_val(pte) &= ~_PAGE_ACCESSED; return pte; }
-extern inline pte_t pte_mkread(pte_t pte) {
+static inline pte_t pte_mkread(pte_t pte) {
pte_val(pte) |= _PAGE_USER; return pte; }
-extern inline pte_t pte_mkexec(pte_t pte) {
+static inline pte_t pte_mkexec(pte_t pte) {
pte_val(pte) |= _PAGE_USER | _PAGE_EXEC; return pte; }
-extern inline pte_t pte_mkwrite(pte_t pte) {
+static inline pte_t pte_mkwrite(pte_t pte) {
pte_val(pte) |= _PAGE_RW; return pte; }
-extern inline pte_t pte_mkdirty(pte_t pte) {
+static inline pte_t pte_mkdirty(pte_t pte) {
pte_val(pte) |= _PAGE_DIRTY; return pte; }
-extern inline pte_t pte_mkyoung(pte_t pte) {
+static inline pte_t pte_mkyoung(pte_t pte) {
pte_val(pte) |= _PAGE_ACCESSED; return pte; }
/* Atomic PTE updates */
diff --git a/include/asm-ppc64/processor.h b/include/asm-ppc64/processor.h
index e05ad20fae15..cdcee424cc79 100644
--- a/include/asm-ppc64/processor.h
+++ b/include/asm-ppc64/processor.h
@@ -706,12 +706,12 @@ unsigned long get_wchan(struct task_struct *p);
#define ARCH_HAS_PREFETCHW
#define ARCH_HAS_SPINLOCK_PREFETCH
-extern inline void prefetch(const void *x)
+static inline void prefetch(const void *x)
{
__asm__ __volatile__ ("dcbt 0,%0" : : "r" (x));
}
-extern inline void prefetchw(const void *x)
+static inline void prefetchw(const void *x)
{
__asm__ __volatile__ ("dcbtst 0,%0" : : "r" (x));
}
diff --git a/include/asm-ppc64/semaphore.h b/include/asm-ppc64/semaphore.h
index 0aa25cf401f9..66ffdc30231d 100644
--- a/include/asm-ppc64/semaphore.h
+++ b/include/asm-ppc64/semaphore.h
@@ -72,7 +72,7 @@ extern void __down(struct semaphore * sem);
extern int __down_interruptible(struct semaphore * sem);
extern void __up(struct semaphore * sem);
-extern inline void down(struct semaphore * sem)
+static inline void down(struct semaphore * sem)
{
#if WAITQUEUE_DEBUG
CHECK_MAGIC(sem->__magic);
@@ -86,7 +86,7 @@ extern inline void down(struct semaphore * sem)
smp_wmb();
}
-extern inline int down_interruptible(struct semaphore * sem)
+static inline int down_interruptible(struct semaphore * sem)
{
int ret = 0;
@@ -100,7 +100,7 @@ extern inline int down_interruptible(struct semaphore * sem)
return ret;
}
-extern inline int down_trylock(struct semaphore * sem)
+static inline int down_trylock(struct semaphore * sem)
{
int ret;
@@ -113,7 +113,7 @@ extern inline int down_trylock(struct semaphore * sem)
return ret;
}
-extern inline void up(struct semaphore * sem)
+static inline void up(struct semaphore * sem)
{
#if WAITQUEUE_DEBUG
CHECK_MAGIC(sem->__magic);
diff --git a/include/asm-ppc64/siginfo.h b/include/asm-ppc64/siginfo.h
index a5e7e87a4449..1070000e3045 100644
--- a/include/asm-ppc64/siginfo.h
+++ b/include/asm-ppc64/siginfo.h
@@ -223,7 +223,7 @@ typedef struct sigevent {
#ifdef __KERNEL__
#include <linux/string.h>
-extern inline void copy_siginfo(siginfo_t *to, siginfo_t *from)
+static inline void copy_siginfo(siginfo_t *to, siginfo_t *from)
{
if (from->si_code < 0)
memcpy(to, from, sizeof(siginfo_t));
diff --git a/include/asm-ppc64/time.h b/include/asm-ppc64/time.h
index bc399ddf2686..64e0b8c90e86 100644
--- a/include/asm-ppc64/time.h
+++ b/include/asm-ppc64/time.h
@@ -87,7 +87,8 @@ static __inline__ void set_dec(int val)
mtspr(SPRN_DEC, val);
}
-extern __inline__ unsigned long tb_ticks_since(unsigned long tstamp) {
+static inline unsigned long tb_ticks_since(unsigned long tstamp)
+{
return get_tb() - tstamp;
}
diff --git a/include/asm-ppc64/tlbflush.h b/include/asm-ppc64/tlbflush.h
index dd902ba04566..b9e8a863d76b 100644
--- a/include/asm-ppc64/tlbflush.h
+++ b/include/asm-ppc64/tlbflush.h
@@ -25,7 +25,7 @@ extern void __flush_tlb_range(struct mm_struct *mm,
#define flush_tlb_kernel_range(start, end) \
__flush_tlb_range(&init_mm, (start), (end))
-extern inline void flush_tlb_pgtables(struct mm_struct *mm,
+static inline void flush_tlb_pgtables(struct mm_struct *mm,
unsigned long start, unsigned long end)
{
/* PPC has hw page tables. */
diff --git a/include/asm-ppc64/uaccess.h b/include/asm-ppc64/uaccess.h
index b6bdf5e283b3..d91ba318b751 100644
--- a/include/asm-ppc64/uaccess.h
+++ b/include/asm-ppc64/uaccess.h
@@ -38,7 +38,7 @@
#define __access_ok(addr,size) (__kernel_ok || __user_ok((addr),(size)))
#define access_ok(type,addr,size) __access_ok((unsigned long)(addr),(size))
-extern inline int verify_area(int type, const void * addr, unsigned long size)
+static inline int verify_area(int type, const void * addr, unsigned long size)
{
return access_ok(type,addr,size) ? 0 : -EFAULT;
}
@@ -200,7 +200,7 @@ do { \
extern unsigned long __copy_tofrom_user(void *to, const void *from, unsigned long size);
-extern inline unsigned long
+static inline unsigned long
copy_from_user(void *to, const void *from, unsigned long n)
{
unsigned long over;
@@ -214,7 +214,7 @@ copy_from_user(void *to, const void *from, unsigned long n)
return n;
}
-extern inline unsigned long
+static inline unsigned long
copy_to_user(void *to, const void *from, unsigned long n)
{
unsigned long over;
@@ -235,7 +235,7 @@ copy_to_user(void *to, const void *from, unsigned long n)
extern unsigned long __clear_user(void *addr, unsigned long size);
-extern inline unsigned long
+static inline unsigned long
clear_user(void *addr, unsigned long size)
{
if (access_ok(VERIFY_WRITE, addr, size))
@@ -245,7 +245,7 @@ clear_user(void *addr, unsigned long size)
extern int __strncpy_from_user(char *dst, const char *src, long count);
-extern inline long
+static inline long
strncpy_from_user(char *dst, const char *src, long count)
{
if (access_ok(VERIFY_READ, src, 1))
@@ -269,7 +269,7 @@ extern int __strnlen_user(const char *str, long len, unsigned long top);
* The `top' parameter to __strnlen_user is to make sure that
* we can never overflow from the user area into kernel space.
*/
-extern __inline__ int strnlen_user(const char *str, long len)
+static inline int strnlen_user(const char *str, long len)
{
unsigned long top = __kernel_ok? ~0UL: TASK_SIZE - 1;
diff --git a/include/asm-ppc64/vga.h b/include/asm-ppc64/vga.h
index 50fb17a02a13..c09849743f45 100644
--- a/include/asm-ppc64/vga.h
+++ b/include/asm-ppc64/vga.h
@@ -26,12 +26,12 @@
* <linux/vt_buffer.h> has already done the right job for us.
*/
-extern inline void scr_writew(u16 val, volatile u16 *addr)
+static inline void scr_writew(u16 val, volatile u16 *addr)
{
st_le16(addr, val);
}
-extern inline u16 scr_readw(volatile const u16 *addr)
+static inline u16 scr_readw(volatile const u16 *addr)
{
return ld_le16(addr);
}