On 01/18/2018 05:48 AM, Alexey Brodkin wrote:
Signed-off-by: Alexey Brodkin <abrod...@synopsys.com>
---
  arch/arc/Kconfig                 |  5 +++++
  arch/arc/include/asm/processor.h | 11 +++++++++++
  arch/arc/lib/memcpy-archs.S      | 16 ++++++++++++++++
  arch/arc/lib/memset-archs.S      |  6 ++++++
  4 files changed, 38 insertions(+)

diff --git a/arch/arc/Kconfig b/arch/arc/Kconfig
index 9d5fd00d9e91..aa5262e8e43f 100644
--- a/arch/arc/Kconfig
+++ b/arch/arc/Kconfig
@@ -532,6 +532,11 @@ config ARC_DBG_TLB_PARANOIA
        bool "Paranoia Checks in Low Level TLB Handlers"
        default n
+config ARC_DISABLE_PREFETCH
+       bool "Disable use of prefetch instructions"
+       default n
+       help
+         Suppresses usage of prefetchw and prealloc instructions.
  endif
config ARC_UBOOT_SUPPORT
diff --git a/arch/arc/include/asm/processor.h b/arch/arc/include/asm/processor.h
index 8ee41e988169..01011c0515d4 100644
--- a/arch/arc/include/asm/processor.h
+++ b/arch/arc/include/asm/processor.h
@@ -106,6 +106,17 @@ extern unsigned int get_wchan(struct task_struct *p);
   */
  #define current_text_addr() ({ __label__ _l; _l: &&_l; })
+#ifdef CONFIG_ARC_DISABLE_PREFETCH
+#define ARCH_HAS_PREFETCH
+#define prefetch(x)
+
+#define ARCH_HAS_PREFETCHW
+#define prefetchw(x)
+
+#define ARCH_HAS_SPINLOCK_PREFETCH
+#define spin_lock_prefetch(x)
+#endif /* CONFIG_ARC_DISABLE_PREFETCH */
+
  #endif /* !__ASSEMBLY__ */
/*
diff --git a/arch/arc/lib/memcpy-archs.S b/arch/arc/lib/memcpy-archs.S
index d61044dd8b58..0a29d024be01 100644
--- a/arch/arc/lib/memcpy-archs.S
+++ b/arch/arc/lib/memcpy-archs.S
@@ -41,8 +41,10 @@
  #endif
ENTRY_CFI(memcpy)
+#ifndef CONFIG_ARC_DISABLE_PREFETCH
        prefetch [r1]           ; Prefetch the read location
        prefetchw [r0]          ; Prefetch the write location
+#endif

This code is already unreadable and this change is not helping. Can we make some macros "C" / gas style and have the switch inside the macro rather than littering this all over the place please !

        mov.f   0, r2
  ;;; if size is zero
        jz.d    [blink]
@@ -72,8 +74,10 @@ ENTRY_CFI(memcpy)
        lpnz    @.Lcopy32_64bytes
        ;; LOOP START
        LOADX (r6, r1)
+#ifndef CONFIG_ARC_DISABLE_PREFETCH
        PREFETCH_READ (r1)
        PREFETCH_WRITE (r3)
+#endif
        LOADX (r8, r1)
        LOADX (r10, r1)
        LOADX (r4, r1)
@@ -117,9 +121,13 @@ ENTRY_CFI(memcpy)
        lpnz    @.Lcopy8bytes_1
        ;; LOOP START
        ld.ab   r6, [r1, 4]
+#ifndef CONFIG_ARC_DISABLE_PREFETCH
        prefetch [r1, 28]       ;Prefetch the next read location
+#endif
        ld.ab   r8, [r1,4]
+#ifndef CONFIG_ARC_DISABLE_PREFETCH
        prefetchw [r3, 32]      ;Prefetch the next write location
+#endif
SHIFT_1 (r7, r6, 24)
        or      r7, r7, r5
@@ -162,9 +170,13 @@ ENTRY_CFI(memcpy)
        lpnz    @.Lcopy8bytes_2
        ;; LOOP START
        ld.ab   r6, [r1, 4]
+#ifndef CONFIG_ARC_DISABLE_PREFETCH
        prefetch [r1, 28]       ;Prefetch the next read location
+#endif
        ld.ab   r8, [r1,4]
+#ifndef CONFIG_ARC_DISABLE_PREFETCH
        prefetchw [r3, 32]      ;Prefetch the next write location
+#endif
SHIFT_1 (r7, r6, 16)
        or      r7, r7, r5
@@ -204,9 +216,13 @@ ENTRY_CFI(memcpy)
        lpnz    @.Lcopy8bytes_3
        ;; LOOP START
        ld.ab   r6, [r1, 4]
+#ifndef CONFIG_ARC_DISABLE_PREFETCH
        prefetch [r1, 28]       ;Prefetch the next read location
+#endif
        ld.ab   r8, [r1,4]
+#ifndef CONFIG_ARC_DISABLE_PREFETCH
        prefetchw [r3, 32]      ;Prefetch the next write location
+#endif
SHIFT_1 (r7, r6, 8)
        or      r7, r7, r5
diff --git a/arch/arc/lib/memset-archs.S b/arch/arc/lib/memset-archs.S
index 62ad4bcb841a..343f292d92a0 100644
--- a/arch/arc/lib/memset-archs.S
+++ b/arch/arc/lib/memset-archs.S
@@ -11,7 +11,9 @@
  #undef PREALLOC_NOT_AVAIL
ENTRY_CFI(memset)
+#ifndef CONFIG_ARC_DISABLE_PREFETCH
        prefetchw [r0]          ; Prefetch the write location
+#endif
        mov.f   0, r2
  ;;; if size is zero
        jz.d    [blink]
@@ -48,11 +50,13 @@ ENTRY_CFI(memset)
lpnz @.Lset64bytes
        ;; LOOP START
+#ifndef CONFIG_ARC_DISABLE_PREFETCH
  #ifdef PREALLOC_NOT_AVAIL
        prefetchw [r3, 64]      ;Prefetch the next write location
  #else
        prealloc  [r3, 64]
  #endif
+#endif /* CONFIG_ARC_DISABLE_PREFETCH */
  #ifdef CONFIG_ARC_HAS_LL64
        std.ab  r4, [r3, 8]
        std.ab  r4, [r3, 8]
@@ -85,7 +89,9 @@ ENTRY_CFI(memset)
        lsr.f   lp_count, r2, 5 ;Last remaining  max 124 bytes
        lpnz    .Lset32bytes
        ;; LOOP START
+#ifndef CONFIG_ARC_DISABLE_PREFETCH
        prefetchw   [r3, 32]    ;Prefetch the next write location
+#endif
  #ifdef CONFIG_ARC_HAS_LL64
        std.ab  r4, [r3, 8]
        std.ab  r4, [r3, 8]

Reply via email to