mirror of
https://github.com/team-infusion-developers/android_kernel_samsung_msm8976.git
synced 2024-11-07 04:09:21 +00:00
[PARISC] parisc: "extern inline" -> "static inline"
"extern inline" will have different semantics with gcc 4.3, and "static inline" is correct here. Signed-off-by: Adrian Bunk <bunk@stusta.de> Cc: Matthew Wilcox <willy@debian.org> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Kyle McMartin <kyle@mcmartin.ca>
This commit is contained in:
parent
0ed5462927
commit
f13cec8447
7 changed files with 27 additions and 27 deletions
|
@ -139,12 +139,12 @@ DECLARE_PER_CPU(struct exception_data, exception_data);
|
|||
#define stw(_s,_t,_o,_a,_e) def_store_insn(stw,"r",_s,_t,_o,_a,_e)
|
||||
|
||||
#ifdef CONFIG_PREFETCH
|
||||
extern inline void prefetch_src(const void *addr)
|
||||
static inline void prefetch_src(const void *addr)
|
||||
{
|
||||
__asm__("ldw 0(" s_space ",%0), %%r0" : : "r" (addr));
|
||||
}
|
||||
|
||||
extern inline void prefetch_dst(const void *addr)
|
||||
static inline void prefetch_dst(const void *addr)
|
||||
{
|
||||
__asm__("ldd 0(" d_space ",%0), %%r0" : : "r" (addr));
|
||||
}
|
||||
|
|
|
@ -138,7 +138,7 @@ extern void __iomem * __ioremap(unsigned long offset, unsigned long size, unsign
|
|||
/* Most machines react poorly to I/O-space being cacheable... Instead let's
|
||||
* define ioremap() in terms of ioremap_nocache().
|
||||
*/
|
||||
extern inline void __iomem * ioremap(unsigned long offset, unsigned long size)
|
||||
static inline void __iomem * ioremap(unsigned long offset, unsigned long size)
|
||||
{
|
||||
return __ioremap(offset, size, _PAGE_NO_CACHE);
|
||||
}
|
||||
|
|
|
@ -207,7 +207,7 @@ extern struct pci_bios_ops *pci_bios;
|
|||
extern void pcibios_register_hba(struct pci_hba_data *);
|
||||
extern void pcibios_set_master(struct pci_dev *);
|
||||
#else
|
||||
extern inline void pcibios_register_hba(struct pci_hba_data *x)
|
||||
static inline void pcibios_register_hba(struct pci_hba_data *x)
|
||||
{
|
||||
}
|
||||
#endif
|
||||
|
|
|
@ -325,27 +325,27 @@ static inline void pgd_clear(pgd_t *pgd) {
|
|||
* setup: the pgd is never bad, and a pmd always exists (as it's folded
|
||||
* into the pgd entry)
|
||||
*/
|
||||
extern inline int pgd_none(pgd_t pgd) { return 0; }
|
||||
extern inline int pgd_bad(pgd_t pgd) { return 0; }
|
||||
extern inline int pgd_present(pgd_t pgd) { return 1; }
|
||||
extern inline void pgd_clear(pgd_t * pgdp) { }
|
||||
static inline int pgd_none(pgd_t pgd) { return 0; }
|
||||
static inline int pgd_bad(pgd_t pgd) { return 0; }
|
||||
static inline int pgd_present(pgd_t pgd) { return 1; }
|
||||
static inline void pgd_clear(pgd_t * pgdp) { }
|
||||
#endif
|
||||
|
||||
/*
|
||||
* The following only work if pte_present() is true.
|
||||
* Undefined behaviour if not..
|
||||
*/
|
||||
extern inline int pte_dirty(pte_t pte) { return pte_val(pte) & _PAGE_DIRTY; }
|
||||
extern inline int pte_young(pte_t pte) { return pte_val(pte) & _PAGE_ACCESSED; }
|
||||
extern inline int pte_write(pte_t pte) { return pte_val(pte) & _PAGE_WRITE; }
|
||||
extern inline int pte_file(pte_t pte) { return pte_val(pte) & _PAGE_FILE; }
|
||||
static inline int pte_dirty(pte_t pte) { return pte_val(pte) & _PAGE_DIRTY; }
|
||||
static inline int pte_young(pte_t pte) { return pte_val(pte) & _PAGE_ACCESSED; }
|
||||
static inline int pte_write(pte_t pte) { return pte_val(pte) & _PAGE_WRITE; }
|
||||
static inline int pte_file(pte_t pte) { return pte_val(pte) & _PAGE_FILE; }
|
||||
|
||||
extern inline pte_t pte_mkclean(pte_t pte) { pte_val(pte) &= ~_PAGE_DIRTY; return pte; }
|
||||
extern inline pte_t pte_mkold(pte_t pte) { pte_val(pte) &= ~_PAGE_ACCESSED; return pte; }
|
||||
extern inline pte_t pte_wrprotect(pte_t pte) { pte_val(pte) &= ~_PAGE_WRITE; return pte; }
|
||||
extern inline pte_t pte_mkdirty(pte_t pte) { pte_val(pte) |= _PAGE_DIRTY; return pte; }
|
||||
extern inline pte_t pte_mkyoung(pte_t pte) { pte_val(pte) |= _PAGE_ACCESSED; return pte; }
|
||||
extern inline pte_t pte_mkwrite(pte_t pte) { pte_val(pte) |= _PAGE_WRITE; return pte; }
|
||||
static inline pte_t pte_mkclean(pte_t pte) { pte_val(pte) &= ~_PAGE_DIRTY; return pte; }
|
||||
static inline pte_t pte_mkold(pte_t pte) { pte_val(pte) &= ~_PAGE_ACCESSED; return pte; }
|
||||
static inline pte_t pte_wrprotect(pte_t pte) { pte_val(pte) &= ~_PAGE_WRITE; return pte; }
|
||||
static inline pte_t pte_mkdirty(pte_t pte) { pte_val(pte) |= _PAGE_DIRTY; return pte; }
|
||||
static inline pte_t pte_mkyoung(pte_t pte) { pte_val(pte) |= _PAGE_ACCESSED; return pte; }
|
||||
static inline pte_t pte_mkwrite(pte_t pte) { pte_val(pte) |= _PAGE_WRITE; return pte; }
|
||||
|
||||
/*
|
||||
* Conversion functions: convert a page and protection to a page entry,
|
||||
|
@ -369,7 +369,7 @@ static inline pte_t pfn_pte(unsigned long pfn, pgprot_t pgprot)
|
|||
return pte;
|
||||
}
|
||||
|
||||
extern inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
|
||||
static inline pte_t pte_modify(pte_t pte, pgprot_t newprot)
|
||||
{ pte_val(pte) = (pte_val(pte) & _PAGE_CHG_MASK) | pgprot_val(newprot); return pte; }
|
||||
|
||||
/* Permanent address of a page. On parisc we don't have highmem. */
|
||||
|
|
|
@ -19,7 +19,7 @@
|
|||
#ifdef CONFIG_PREFETCH
|
||||
|
||||
#define ARCH_HAS_PREFETCH
|
||||
extern inline void prefetch(const void *addr)
|
||||
static inline void prefetch(const void *addr)
|
||||
{
|
||||
__asm__("ldw 0(%0), %%r0" : : "r" (addr));
|
||||
}
|
||||
|
@ -27,7 +27,7 @@ extern inline void prefetch(const void *addr)
|
|||
/* LDD is a PA2.0 addition. */
|
||||
#ifdef CONFIG_PA20
|
||||
#define ARCH_HAS_PREFETCHW
|
||||
extern inline void prefetchw(const void *addr)
|
||||
static inline void prefetchw(const void *addr)
|
||||
{
|
||||
__asm__("ldd 0(%0), %%r0" : : "r" (addr));
|
||||
}
|
||||
|
|
|
@ -54,7 +54,7 @@ struct semaphore {
|
|||
|
||||
#define DECLARE_MUTEX(name) __DECLARE_SEMAPHORE_GENERIC(name,1)
|
||||
|
||||
extern inline void sema_init (struct semaphore *sem, int val)
|
||||
static inline void sema_init (struct semaphore *sem, int val)
|
||||
{
|
||||
*sem = (struct semaphore)__SEMAPHORE_INITIALIZER((*sem),val);
|
||||
}
|
||||
|
@ -82,7 +82,7 @@ asmlinkage void __up(struct semaphore * sem);
|
|||
* interrupts while we're messing with the semaphore. Sorry.
|
||||
*/
|
||||
|
||||
extern __inline__ void down(struct semaphore * sem)
|
||||
static inline void down(struct semaphore * sem)
|
||||
{
|
||||
might_sleep();
|
||||
spin_lock_irq(&sem->sentry);
|
||||
|
@ -94,7 +94,7 @@ extern __inline__ void down(struct semaphore * sem)
|
|||
spin_unlock_irq(&sem->sentry);
|
||||
}
|
||||
|
||||
extern __inline__ int down_interruptible(struct semaphore * sem)
|
||||
static inline int down_interruptible(struct semaphore * sem)
|
||||
{
|
||||
int ret = 0;
|
||||
might_sleep();
|
||||
|
@ -112,7 +112,7 @@ extern __inline__ int down_interruptible(struct semaphore * sem)
|
|||
* down_trylock returns 0 on success, 1 if we failed to get the lock.
|
||||
* May not sleep, but must preserve irq state
|
||||
*/
|
||||
extern __inline__ int down_trylock(struct semaphore * sem)
|
||||
static inline int down_trylock(struct semaphore * sem)
|
||||
{
|
||||
unsigned long flags;
|
||||
int count;
|
||||
|
@ -129,7 +129,7 @@ extern __inline__ int down_trylock(struct semaphore * sem)
|
|||
* Note! This is subtle. We jump to wake people up only if
|
||||
* the semaphore was negative (== somebody was waiting on it).
|
||||
*/
|
||||
extern __inline__ void up(struct semaphore * sem)
|
||||
static inline void up(struct semaphore * sem)
|
||||
{
|
||||
unsigned long flags;
|
||||
|
||||
|
|
|
@ -57,7 +57,7 @@ static inline void flush_tlb_mm(struct mm_struct *mm)
|
|||
#endif
|
||||
}
|
||||
|
||||
extern __inline__ void flush_tlb_pgtables(struct mm_struct *mm, unsigned long start, unsigned long end)
|
||||
static inline void flush_tlb_pgtables(struct mm_struct *mm, unsigned long start, unsigned long end)
|
||||
{
|
||||
}
|
||||
|
||||
|
|
Loading…
Reference in a new issue