@@ -84,7 +84,7 @@ static inline u32 arch_timer_get_cntfrq(void)
return val;
}
-static inline u64 __arch_counter_get_cntpct(void)
+static inline u64 arch_counter_get_cntpct(void)
{
u64 cval;
@@ -93,12 +93,7 @@ static inline u64 __arch_counter_get_cntpct(void)
return cval;
}
-static inline u64 __arch_counter_get_cntpct_stable(void)
-{
- return __arch_counter_get_cntpct();
-}
-
-static inline u64 __arch_counter_get_cntvct(void)
+static inline u64 arch_counter_get_cntvct(void)
{
u64 cval;
@@ -107,11 +102,6 @@ static inline u64 __arch_counter_get_cntvct(void)
return cval;
}
-static inline u64 __arch_counter_get_cntvct_stable(void)
-{
- return __arch_counter_get_cntvct();
-}
-
static inline u32 arch_timer_get_cntkctl(void)
{
u32 cntkctl;
@@ -184,7 +184,7 @@ static inline void arch_timer_set_cntkctl(u32 cntkctl)
: "=r" (tmp) : "r" (_val)); \
} while (0)
-static __always_inline u64 __arch_counter_get_cntpct_stable(void)
+static __always_inline u64 arch_counter_get_cntpct(void)
{
u64 cnt;
@@ -194,17 +194,7 @@ static __always_inline u64 __arch_counter_get_cntpct_stable(void)
return cnt;
}
-static __always_inline u64 __arch_counter_get_cntpct(void)
-{
- u64 cnt;
-
- isb();
- cnt = read_sysreg(cntpct_el0);
- arch_counter_enforce_ordering(cnt);
- return cnt;
-}
-
-static __always_inline u64 __arch_counter_get_cntvct_stable(void)
+static __always_inline u64 arch_counter_get_cntvct(void)
{
u64 cnt;
@@ -214,16 +204,6 @@ static __always_inline u64 __arch_counter_get_cntvct_stable(void)
return cnt;
}
-static __always_inline u64 __arch_counter_get_cntvct(void)
-{
- u64 cnt;
-
- isb();
- cnt = read_sysreg(cntvct_el0);
- arch_counter_enforce_ordering(cnt);
- return cnt;
-}
-
#undef arch_counter_enforce_ordering
static inline int arch_timer_arch_init(void)
@@ -150,26 +150,6 @@ u32 arch_timer_reg_read(int access, enum arch_timer_reg reg,
return val;
}
-static notrace u64 arch_counter_get_cntpct_stable(void)
-{
- return __arch_counter_get_cntpct_stable();
-}
-
-static notrace u64 arch_counter_get_cntpct(void)
-{
- return __arch_counter_get_cntpct();
-}
-
-static notrace u64 arch_counter_get_cntvct_stable(void)
-{
- return __arch_counter_get_cntvct_stable();
-}
-
-static notrace u64 arch_counter_get_cntvct(void)
-{
- return __arch_counter_get_cntvct();
-}
-
/*
* Default to cp15 based access because arm64 uses this function for
* sched_clock() before DT is probed and the cp15 method is guaranteed
@@ -383,8 +363,6 @@ static u32 notrace sun50i_a64_read_cntv_tval_el0(void)
DEFINE_PER_CPU(const struct arch_timer_erratum_workaround *, timer_unstable_counter_workaround);
EXPORT_SYMBOL_GPL(timer_unstable_counter_workaround);
-static atomic_t timer_unstable_counter_workaround_in_use = ATOMIC_INIT(0);
-
static void erratum_set_next_event_tval_generic(const int access, unsigned long evt,
struct clock_event_device *clk)
{
@@ -562,9 +540,6 @@ void arch_timer_enable_workaround(const struct arch_timer_erratum_workaround *wa
per_cpu(timer_unstable_counter_workaround, i) = wa;
}
- if (wa->read_cntvct_el0 || wa->read_cntpct_el0)
- atomic_set(&timer_unstable_counter_workaround_in_use, 1);
-
/*
* Don't use the vdso fastpath if errata require using the
* out-of-line counter accessor. We may change our mind pretty
@@ -625,14 +600,9 @@ static bool arch_timer_this_cpu_has_cntvct_wa(void)
return has_erratum_handler(read_cntvct_el0);
}
-static bool arch_timer_counter_has_wa(void)
-{
- return atomic_read(&timer_unstable_counter_workaround_in_use);
-}
#else
#define arch_timer_check_ool_workaround(t,a) do { } while(0)
#define arch_timer_this_cpu_has_cntvct_wa() ({false;})
-#define arch_timer_counter_has_wa() ({false;})
#endif /* CONFIG_ARM_ARCH_TIMER_OOL_WORKAROUND */
static __always_inline irqreturn_t timer_handler(const int access,
@@ -989,22 +959,13 @@ static void __init arch_counter_register(unsigned type)
/* Register the CP15 based counter if we have one */
if (type & ARCH_TIMER_TYPE_CP15) {
- u64 (*rd)(void);
-
if ((IS_ENABLED(CONFIG_ARM64) && !is_hyp_mode_available()) ||
arch_timer_uses_ppi == ARCH_TIMER_VIRT_PPI) {
- if (arch_timer_counter_has_wa())
- rd = arch_counter_get_cntvct_stable;
- else
- rd = arch_counter_get_cntvct;
+ arch_timer_read_counter = arch_counter_get_cntvct;
} else {
- if (arch_timer_counter_has_wa())
- rd = arch_counter_get_cntpct_stable;
- else
- rd = arch_counter_get_cntpct;
+ arch_timer_read_counter = arch_counter_get_cntpct;
}
- arch_timer_read_counter = rd;
clocksource_counter.vdso_clock_mode = vdso_default;
} else {
arch_timer_read_counter = arch_counter_get_cntvct_mem;
In commit 0ea415390cd3 (clocksource/arm_arch_timer: Use arch_timer_read_counter to access stable counters), we separate stable and normal count reader. Actually the stable reader can correctly lead us to normal reader if we has no workaround. Besides, in erratum_set_next_event_tval_generic(), we use normal reader, it is obviously wrong, so just revert this commit to solve this problem by the way. Signed-off-by: Keqian Zhu <zhukeqian1@huawei.com> --- arch/arm/include/asm/arch_timer.h | 14 ++---------- arch/arm64/include/asm/arch_timer.h | 24 ++------------------ drivers/clocksource/arm_arch_timer.c | 43 ++---------------------------------- 3 files changed, 6 insertions(+), 75 deletions(-)