aboutsummaryrefslogtreecommitdiffstats
path: root/include/asm-sh/uaccess_64.h
diff options
context:
space:
mode:
Diffstat (limited to 'include/asm-sh/uaccess_64.h')
-rw-r--r--include/asm-sh/uaccess_64.h12
1 files changed, 0 insertions, 12 deletions
diff --git a/include/asm-sh/uaccess_64.h b/include/asm-sh/uaccess_64.h
index 644c67b65f94..24800a8045ce 100644
--- a/include/asm-sh/uaccess_64.h
+++ b/include/asm-sh/uaccess_64.h
@@ -297,18 +297,6 @@ struct exception_table_entry
297 297
298#define ARCH_HAS_SEARCH_EXTABLE 298#define ARCH_HAS_SEARCH_EXTABLE
299 299
300/* If gcc inlines memset, it will use st.q instructions. Therefore, we need
301 kmalloc allocations to be 8-byte aligned. Without this, the alignment
302 becomes BYTE_PER_WORD i.e. only 4 (since sizeof(long)==sizeof(void*)==4 on
303 sh64 at the moment). */
304#define ARCH_KMALLOC_MINALIGN 8
305
306/*
307 * We want 8-byte alignment for the slab caches as well, otherwise we have
308 * the same BYTES_PER_WORD (sizeof(void *)) min align in kmem_cache_create().
309 */
310#define ARCH_SLAB_MINALIGN 8
311
312/* Returns 0 if exception not found and fixup.unit otherwise. */ 300/* Returns 0 if exception not found and fixup.unit otherwise. */
313extern unsigned long search_exception_table(unsigned long addr); 301extern unsigned long search_exception_table(unsigned long addr);
314extern const struct exception_table_entry *search_exception_tables (unsigned long addr); 302extern const struct exception_table_entry *search_exception_tables (unsigned long addr);