clocksource: Use a plain u64 instead of cycle_t
There is no point in having an extra type for extra confusion. u64 is unambiguous. Conversion was done with the following coccinelle script: @rem@ @@ -typedef u64 cycle_t; @fix@ typedef cycle_t; @@ -cycle_t +u64 Signed-off-by: Thomas Gleixner <tglx@linutronix.de> Cc: Peter Zijlstra <peterz@infradead.org> Cc: John Stultz <john.stultz@linaro.org>
This commit is contained in:
@@ -133,7 +133,7 @@ init_rtc_clockevent(void)
|
||||
* The QEMU clock as a clocksource primitive.
|
||||
*/
|
||||
|
||||
static cycle_t
|
||||
static u64
|
||||
qemu_cs_read(struct clocksource *cs)
|
||||
{
|
||||
return qemu_get_vmtime();
|
||||
@@ -260,7 +260,7 @@ common_init_rtc(void)
|
||||
* use this method when WTINT is in use.
|
||||
*/
|
||||
|
||||
static cycle_t read_rpcc(struct clocksource *cs)
|
||||
static u64 read_rpcc(struct clocksource *cs)
|
||||
{
|
||||
return rpcc();
|
||||
}
|
||||
|
@@ -268,7 +268,7 @@ static void __init timer_init(void)
|
||||
/*
|
||||
* clocksource
|
||||
*/
|
||||
static cycle_t read_cycles(struct clocksource *cs)
|
||||
static u64 read_cycles(struct clocksource *cs)
|
||||
{
|
||||
struct timer_s *t = &timers[TID_CLOCKSOURCE];
|
||||
|
||||
|
@@ -59,13 +59,13 @@ static u64 notrace ep93xx_read_sched_clock(void)
|
||||
return ret;
|
||||
}
|
||||
|
||||
cycle_t ep93xx_clocksource_read(struct clocksource *c)
|
||||
u64 ep93xx_clocksource_read(struct clocksource *c)
|
||||
{
|
||||
u64 ret;
|
||||
|
||||
ret = readl(EP93XX_TIMER4_VALUE_LOW);
|
||||
ret |= ((u64) (readl(EP93XX_TIMER4_VALUE_HIGH) & 0xff) << 32);
|
||||
return (cycle_t) ret;
|
||||
return (u64) ret;
|
||||
}
|
||||
|
||||
static int ep93xx_clkevt_set_next_event(unsigned long next,
|
||||
|
@@ -19,7 +19,7 @@
|
||||
|
||||
#include "common.h"
|
||||
|
||||
static cycle_t cksrc_dc21285_read(struct clocksource *cs)
|
||||
static u64 cksrc_dc21285_read(struct clocksource *cs)
|
||||
{
|
||||
return cs->mask - *CSR_TIMER2_VALUE;
|
||||
}
|
||||
|
@@ -493,7 +493,7 @@ static u64 notrace ixp4xx_read_sched_clock(void)
|
||||
* clocksource
|
||||
*/
|
||||
|
||||
static cycle_t ixp4xx_clocksource_read(struct clocksource *c)
|
||||
static u64 ixp4xx_clocksource_read(struct clocksource *c)
|
||||
{
|
||||
return *IXP4XX_OSTS;
|
||||
}
|
||||
|
@@ -144,7 +144,7 @@ static struct clock_event_device ckevt = {
|
||||
.set_state_oneshot = timer_set_shutdown,
|
||||
};
|
||||
|
||||
static cycle_t clksrc_read(struct clocksource *cs)
|
||||
static u64 clksrc_read(struct clocksource *cs)
|
||||
{
|
||||
return timer_read();
|
||||
}
|
||||
|
@@ -369,9 +369,9 @@ static bool use_gptimer_clksrc __initdata;
|
||||
/*
|
||||
* clocksource
|
||||
*/
|
||||
static cycle_t clocksource_read_cycles(struct clocksource *cs)
|
||||
static u64 clocksource_read_cycles(struct clocksource *cs)
|
||||
{
|
||||
return (cycle_t)__omap_dm_timer_read_counter(&clksrc,
|
||||
return (u64)__omap_dm_timer_read_counter(&clksrc,
|
||||
OMAP_TIMER_NONPOSTED);
|
||||
}
|
||||
|
||||
|
@@ -38,7 +38,7 @@
|
||||
/*
|
||||
* IOP clocksource (free-running timer 1).
|
||||
*/
|
||||
static cycle_t notrace iop_clocksource_read(struct clocksource *unused)
|
||||
static u64 notrace iop_clocksource_read(struct clocksource *unused)
|
||||
{
|
||||
return 0xffffffffu - read_tcr1();
|
||||
}
|
||||
|
@@ -20,9 +20,9 @@
|
||||
|
||||
static bool disable_cpu_idle_poll;
|
||||
|
||||
static cycle_t read_cycle_count(struct clocksource *cs)
|
||||
static u64 read_cycle_count(struct clocksource *cs)
|
||||
{
|
||||
return (cycle_t)sysreg_read(COUNT);
|
||||
return (u64)sysreg_read(COUNT);
|
||||
}
|
||||
|
||||
/*
|
||||
|
@@ -26,7 +26,7 @@
|
||||
|
||||
#if defined(CONFIG_CYCLES_CLOCKSOURCE)
|
||||
|
||||
static notrace cycle_t bfin_read_cycles(struct clocksource *cs)
|
||||
static notrace u64 bfin_read_cycles(struct clocksource *cs)
|
||||
{
|
||||
#ifdef CONFIG_CPU_FREQ
|
||||
return __bfin_cycles_off + (get_cycles() << __bfin_cycles_mod);
|
||||
@@ -80,7 +80,7 @@ void __init setup_gptimer0(void)
|
||||
enable_gptimers(TIMER0bit);
|
||||
}
|
||||
|
||||
static cycle_t bfin_read_gptimer0(struct clocksource *cs)
|
||||
static u64 bfin_read_gptimer0(struct clocksource *cs)
|
||||
{
|
||||
return bfin_read_TIMER0_COUNTER();
|
||||
}
|
||||
|
@@ -26,7 +26,7 @@
|
||||
static u32 sched_clock_multiplier;
|
||||
#define SCHED_CLOCK_SHIFT 16
|
||||
|
||||
static cycle_t tsc_read(struct clocksource *cs)
|
||||
static u64 tsc_read(struct clocksource *cs)
|
||||
{
|
||||
return get_cycles();
|
||||
}
|
||||
|
@@ -72,9 +72,9 @@ struct adsp_hw_timer_struct {
|
||||
/* Look for "TCX0" for related constants. */
|
||||
static __iomem struct adsp_hw_timer_struct *rtos_timer;
|
||||
|
||||
static cycle_t timer_get_cycles(struct clocksource *cs)
|
||||
static u64 timer_get_cycles(struct clocksource *cs)
|
||||
{
|
||||
return (cycle_t) __vmgettime();
|
||||
return (u64) __vmgettime();
|
||||
}
|
||||
|
||||
static struct clocksource hexagon_clocksource = {
|
||||
|
@@ -21,9 +21,9 @@ void __init cyclone_setup(void)
|
||||
|
||||
static void __iomem *cyclone_mc;
|
||||
|
||||
static cycle_t read_cyclone(struct clocksource *cs)
|
||||
static u64 read_cyclone(struct clocksource *cs)
|
||||
{
|
||||
return (cycle_t)readq((void __iomem *)cyclone_mc);
|
||||
return (u64)readq((void __iomem *)cyclone_mc);
|
||||
}
|
||||
|
||||
static struct clocksource clocksource_cyclone = {
|
||||
|
@@ -9,15 +9,15 @@ struct fsyscall_gtod_data_t {
|
||||
seqcount_t seq;
|
||||
struct timespec wall_time;
|
||||
struct timespec monotonic_time;
|
||||
cycle_t clk_mask;
|
||||
u64 clk_mask;
|
||||
u32 clk_mult;
|
||||
u32 clk_shift;
|
||||
void *clk_fsys_mmio;
|
||||
cycle_t clk_cycle_last;
|
||||
u64 clk_cycle_last;
|
||||
} ____cacheline_aligned;
|
||||
|
||||
struct itc_jitter_data_t {
|
||||
int itc_jitter;
|
||||
cycle_t itc_lastcycle;
|
||||
u64 itc_lastcycle;
|
||||
} ____cacheline_aligned;
|
||||
|
||||
|
@@ -31,7 +31,7 @@
|
||||
|
||||
#include "fsyscall_gtod_data.h"
|
||||
|
||||
static cycle_t itc_get_cycles(struct clocksource *cs);
|
||||
static u64 itc_get_cycles(struct clocksource *cs);
|
||||
|
||||
struct fsyscall_gtod_data_t fsyscall_gtod_data;
|
||||
|
||||
@@ -323,7 +323,7 @@ void ia64_init_itm(void)
|
||||
}
|
||||
}
|
||||
|
||||
static cycle_t itc_get_cycles(struct clocksource *cs)
|
||||
static u64 itc_get_cycles(struct clocksource *cs)
|
||||
{
|
||||
unsigned long lcycle, now, ret;
|
||||
|
||||
@@ -397,7 +397,7 @@ void update_vsyscall_tz(void)
|
||||
}
|
||||
|
||||
void update_vsyscall_old(struct timespec *wall, struct timespec *wtm,
|
||||
struct clocksource *c, u32 mult, cycle_t cycle_last)
|
||||
struct clocksource *c, u32 mult, u64 cycle_last)
|
||||
{
|
||||
write_seqcount_begin(&fsyscall_gtod_data.seq);
|
||||
|
||||
|
@@ -22,9 +22,9 @@
|
||||
|
||||
extern unsigned long sn_rtc_cycles_per_second;
|
||||
|
||||
static cycle_t read_sn2(struct clocksource *cs)
|
||||
static u64 read_sn2(struct clocksource *cs)
|
||||
{
|
||||
return (cycle_t)readq(RTC_COUNTER_ADDR);
|
||||
return (u64)readq(RTC_COUNTER_ADDR);
|
||||
}
|
||||
|
||||
static struct clocksource clocksource_sn2 = {
|
||||
|
@@ -76,7 +76,7 @@ static struct irqaction m68328_timer_irq = {
|
||||
|
||||
/***************************************************************************/
|
||||
|
||||
static cycle_t m68328_read_clk(struct clocksource *cs)
|
||||
static u64 m68328_read_clk(struct clocksource *cs)
|
||||
{
|
||||
unsigned long flags;
|
||||
u32 cycles;
|
||||
|
@@ -34,7 +34,7 @@
|
||||
#define DMA_DTMR_CLK_DIV_16 (2 << 1)
|
||||
#define DMA_DTMR_ENABLE (1 << 0)
|
||||
|
||||
static cycle_t cf_dt_get_cycles(struct clocksource *cs)
|
||||
static u64 cf_dt_get_cycles(struct clocksource *cs)
|
||||
{
|
||||
return __raw_readl(DTCN0);
|
||||
}
|
||||
|
@@ -118,7 +118,7 @@ static struct irqaction pit_irq = {
|
||||
|
||||
/***************************************************************************/
|
||||
|
||||
static cycle_t pit_read_clk(struct clocksource *cs)
|
||||
static u64 pit_read_clk(struct clocksource *cs)
|
||||
{
|
||||
unsigned long flags;
|
||||
u32 cycles;
|
||||
|
@@ -97,7 +97,7 @@ static struct irqaction mcfslt_timer_irq = {
|
||||
.handler = mcfslt_tick,
|
||||
};
|
||||
|
||||
static cycle_t mcfslt_read_clk(struct clocksource *cs)
|
||||
static u64 mcfslt_read_clk(struct clocksource *cs)
|
||||
{
|
||||
unsigned long flags;
|
||||
u32 cycles, scnt;
|
||||
|
@@ -89,7 +89,7 @@ static struct irqaction mcftmr_timer_irq = {
|
||||
|
||||
/***************************************************************************/
|
||||
|
||||
static cycle_t mcftmr_read_clk(struct clocksource *cs)
|
||||
static u64 mcftmr_read_clk(struct clocksource *cs)
|
||||
{
|
||||
unsigned long flags;
|
||||
u32 cycles;
|
||||
|
@@ -190,17 +190,17 @@ static u64 xilinx_clock_read(void)
|
||||
return read_fn(timer_baseaddr + TCR1);
|
||||
}
|
||||
|
||||
static cycle_t xilinx_read(struct clocksource *cs)
|
||||
static u64 xilinx_read(struct clocksource *cs)
|
||||
{
|
||||
/* reading actual value of timer 1 */
|
||||
return (cycle_t)xilinx_clock_read();
|
||||
return (u64)xilinx_clock_read();
|
||||
}
|
||||
|
||||
static struct timecounter xilinx_tc = {
|
||||
.cc = NULL,
|
||||
};
|
||||
|
||||
static cycle_t xilinx_cc_read(const struct cyclecounter *cc)
|
||||
static u64 xilinx_cc_read(const struct cyclecounter *cc)
|
||||
{
|
||||
return xilinx_read(NULL);
|
||||
}
|
||||
|
@@ -44,7 +44,7 @@
|
||||
/* 32kHz clock enabled and detected */
|
||||
#define CNTR_OK (SYS_CNTRL_E0 | SYS_CNTRL_32S)
|
||||
|
||||
static cycle_t au1x_counter1_read(struct clocksource *cs)
|
||||
static u64 au1x_counter1_read(struct clocksource *cs)
|
||||
{
|
||||
return alchemy_rdsys(AU1000_SYS_RTCREAD);
|
||||
}
|
||||
|
@@ -98,7 +98,7 @@ void octeon_init_cvmcount(void)
|
||||
local_irq_restore(flags);
|
||||
}
|
||||
|
||||
static cycle_t octeon_cvmcount_read(struct clocksource *cs)
|
||||
static u64 octeon_cvmcount_read(struct clocksource *cs)
|
||||
{
|
||||
return read_c0_cvmcount();
|
||||
}
|
||||
|
@@ -34,7 +34,7 @@
|
||||
|
||||
static uint16_t jz4740_jiffies_per_tick;
|
||||
|
||||
static cycle_t jz4740_clocksource_read(struct clocksource *cs)
|
||||
static u64 jz4740_clocksource_read(struct clocksource *cs)
|
||||
{
|
||||
return jz4740_timer_get_count(TIMER_CLOCKSOURCE);
|
||||
}
|
||||
|
@@ -27,7 +27,7 @@ struct txx9_clocksource {
|
||||
struct txx9_tmr_reg __iomem *tmrptr;
|
||||
};
|
||||
|
||||
static cycle_t txx9_cs_read(struct clocksource *cs)
|
||||
static u64 txx9_cs_read(struct clocksource *cs)
|
||||
{
|
||||
struct txx9_clocksource *txx9_cs =
|
||||
container_of(cs, struct txx9_clocksource, cs);
|
||||
|
@@ -25,9 +25,9 @@
|
||||
|
||||
#include <asm/sibyte/sb1250.h>
|
||||
|
||||
static cycle_t bcm1480_hpt_read(struct clocksource *cs)
|
||||
static u64 bcm1480_hpt_read(struct clocksource *cs)
|
||||
{
|
||||
return (cycle_t) __raw_readq(IOADDR(A_SCD_ZBBUS_CYCLE_COUNT));
|
||||
return (u64) __raw_readq(IOADDR(A_SCD_ZBBUS_CYCLE_COUNT));
|
||||
}
|
||||
|
||||
struct clocksource bcm1480_clocksource = {
|
||||
|
@@ -22,7 +22,7 @@
|
||||
#include <asm/dec/ioasic.h>
|
||||
#include <asm/dec/ioasic_addrs.h>
|
||||
|
||||
static cycle_t dec_ioasic_hpt_read(struct clocksource *cs)
|
||||
static u64 dec_ioasic_hpt_read(struct clocksource *cs)
|
||||
{
|
||||
return ioasic_read(IO_REG_FCTR);
|
||||
}
|
||||
|
@@ -11,7 +11,7 @@
|
||||
|
||||
#include <asm/time.h>
|
||||
|
||||
static cycle_t c0_hpt_read(struct clocksource *cs)
|
||||
static u64 c0_hpt_read(struct clocksource *cs)
|
||||
{
|
||||
return read_c0_count();
|
||||
}
|
||||
|
@@ -30,7 +30,7 @@
|
||||
* The HPT is free running from SB1250_HPT_VALUE down to 0 then starts over
|
||||
* again.
|
||||
*/
|
||||
static inline cycle_t sb1250_hpt_get_cycles(void)
|
||||
static inline u64 sb1250_hpt_get_cycles(void)
|
||||
{
|
||||
unsigned int count;
|
||||
void __iomem *addr;
|
||||
@@ -41,7 +41,7 @@ static inline cycle_t sb1250_hpt_get_cycles(void)
|
||||
return SB1250_HPT_VALUE - count;
|
||||
}
|
||||
|
||||
static cycle_t sb1250_hpt_read(struct clocksource *cs)
|
||||
static u64 sb1250_hpt_read(struct clocksource *cs)
|
||||
{
|
||||
return sb1250_hpt_get_cycles();
|
||||
}
|
||||
|
@@ -63,7 +63,7 @@ void __init ls1x_pwmtimer_init(void)
|
||||
ls1x_pwmtimer_restart();
|
||||
}
|
||||
|
||||
static cycle_t ls1x_clocksource_read(struct clocksource *cs)
|
||||
static u64 ls1x_clocksource_read(struct clocksource *cs)
|
||||
{
|
||||
unsigned long flags;
|
||||
int count;
|
||||
@@ -107,7 +107,7 @@ static cycle_t ls1x_clocksource_read(struct clocksource *cs)
|
||||
|
||||
raw_spin_unlock_irqrestore(&ls1x_timer_lock, flags);
|
||||
|
||||
return (cycle_t) (jifs * ls1x_jiffies_per_tick) + count;
|
||||
return (u64) (jifs * ls1x_jiffies_per_tick) + count;
|
||||
}
|
||||
|
||||
static struct clocksource ls1x_clocksource = {
|
||||
|
@@ -144,7 +144,7 @@ void __init setup_mfgpt0_timer(void)
|
||||
* to just read by itself. So use jiffies to emulate a free
|
||||
* running counter:
|
||||
*/
|
||||
static cycle_t mfgpt_read(struct clocksource *cs)
|
||||
static u64 mfgpt_read(struct clocksource *cs)
|
||||
{
|
||||
unsigned long flags;
|
||||
int count;
|
||||
@@ -188,7 +188,7 @@ static cycle_t mfgpt_read(struct clocksource *cs)
|
||||
|
||||
raw_spin_unlock_irqrestore(&mfgpt_lock, flags);
|
||||
|
||||
return (cycle_t) (jifs * COMPARE) + count;
|
||||
return (u64) (jifs * COMPARE) + count;
|
||||
}
|
||||
|
||||
static struct clocksource clocksource_mfgpt = {
|
||||
|
@@ -248,9 +248,9 @@ void __init setup_hpet_timer(void)
|
||||
pr_info("hpet clock event device register\n");
|
||||
}
|
||||
|
||||
static cycle_t hpet_read_counter(struct clocksource *cs)
|
||||
static u64 hpet_read_counter(struct clocksource *cs)
|
||||
{
|
||||
return (cycle_t)hpet_read(HPET_COUNTER);
|
||||
return (u64)hpet_read(HPET_COUNTER);
|
||||
}
|
||||
|
||||
static void hpet_suspend(struct clocksource *cs)
|
||||
|
@@ -75,7 +75,7 @@ static void __init estimate_frequencies(void)
|
||||
unsigned int count, start;
|
||||
unsigned char secs1, secs2, ctrl;
|
||||
int secs;
|
||||
cycle_t giccount = 0, gicstart = 0;
|
||||
u64 giccount = 0, gicstart = 0;
|
||||
|
||||
#if defined(CONFIG_KVM_GUEST) && CONFIG_KVM_GUEST_TIMER_FREQ
|
||||
mips_hpt_frequency = CONFIG_KVM_GUEST_TIMER_FREQ * 1000000;
|
||||
|
@@ -59,14 +59,14 @@ unsigned int get_c0_compare_int(void)
|
||||
return IRQ_TIMER;
|
||||
}
|
||||
|
||||
static cycle_t nlm_get_pic_timer(struct clocksource *cs)
|
||||
static u64 nlm_get_pic_timer(struct clocksource *cs)
|
||||
{
|
||||
uint64_t picbase = nlm_get_node(0)->picbase;
|
||||
|
||||
return ~nlm_pic_read_timer(picbase, PIC_CLOCK_TIMER);
|
||||
}
|
||||
|
||||
static cycle_t nlm_get_pic_timer32(struct clocksource *cs)
|
||||
static u64 nlm_get_pic_timer32(struct clocksource *cs)
|
||||
{
|
||||
uint64_t picbase = nlm_get_node(0)->picbase;
|
||||
|
||||
|
@@ -140,7 +140,7 @@ static void __init hub_rt_clock_event_global_init(void)
|
||||
setup_irq(irq, &hub_rt_irqaction);
|
||||
}
|
||||
|
||||
static cycle_t hub_rt_read(struct clocksource *cs)
|
||||
static u64 hub_rt_read(struct clocksource *cs)
|
||||
{
|
||||
return REMOTE_HUB_L(cputonasid(0), PI_RT_COUNT);
|
||||
}
|
||||
|
@@ -13,7 +13,7 @@
|
||||
#include <asm/timex.h>
|
||||
#include "internal.h"
|
||||
|
||||
static cycle_t mn10300_read(struct clocksource *cs)
|
||||
static u64 mn10300_read(struct clocksource *cs)
|
||||
{
|
||||
return read_timestamp_counter();
|
||||
}
|
||||
|
@@ -81,7 +81,7 @@ static inline unsigned long read_timersnapshot(struct nios2_timer *timer)
|
||||
return count;
|
||||
}
|
||||
|
||||
static cycle_t nios2_timer_read(struct clocksource *cs)
|
||||
static u64 nios2_timer_read(struct clocksource *cs)
|
||||
{
|
||||
struct nios2_clocksource *nios2_cs = to_nios2_clksource(cs);
|
||||
unsigned long flags;
|
||||
|
@@ -117,9 +117,9 @@ static __init void openrisc_clockevent_init(void)
|
||||
* is 32 bits wide and runs at the CPU clock frequency.
|
||||
*/
|
||||
|
||||
static cycle_t openrisc_timer_read(struct clocksource *cs)
|
||||
static u64 openrisc_timer_read(struct clocksource *cs)
|
||||
{
|
||||
return (cycle_t) mfspr(SPR_TTCR);
|
||||
return (u64) mfspr(SPR_TTCR);
|
||||
}
|
||||
|
||||
static struct clocksource openrisc_timer = {
|
||||
|
@@ -137,7 +137,7 @@ EXPORT_SYMBOL(profile_pc);
|
||||
|
||||
/* clock source code */
|
||||
|
||||
static cycle_t notrace read_cr16(struct clocksource *cs)
|
||||
static u64 notrace read_cr16(struct clocksource *cs)
|
||||
{
|
||||
return get_cycles();
|
||||
}
|
||||
|
@@ -80,7 +80,7 @@
|
||||
#include <linux/clockchips.h>
|
||||
#include <linux/timekeeper_internal.h>
|
||||
|
||||
static cycle_t rtc_read(struct clocksource *);
|
||||
static u64 rtc_read(struct clocksource *);
|
||||
static struct clocksource clocksource_rtc = {
|
||||
.name = "rtc",
|
||||
.rating = 400,
|
||||
@@ -89,7 +89,7 @@ static struct clocksource clocksource_rtc = {
|
||||
.read = rtc_read,
|
||||
};
|
||||
|
||||
static cycle_t timebase_read(struct clocksource *);
|
||||
static u64 timebase_read(struct clocksource *);
|
||||
static struct clocksource clocksource_timebase = {
|
||||
.name = "timebase",
|
||||
.rating = 400,
|
||||
@@ -802,18 +802,18 @@ void read_persistent_clock(struct timespec *ts)
|
||||
}
|
||||
|
||||
/* clocksource code */
|
||||
static cycle_t rtc_read(struct clocksource *cs)
|
||||
static u64 rtc_read(struct clocksource *cs)
|
||||
{
|
||||
return (cycle_t)get_rtc();
|
||||
return (u64)get_rtc();
|
||||
}
|
||||
|
||||
static cycle_t timebase_read(struct clocksource *cs)
|
||||
static u64 timebase_read(struct clocksource *cs)
|
||||
{
|
||||
return (cycle_t)get_tb();
|
||||
return (u64)get_tb();
|
||||
}
|
||||
|
||||
void update_vsyscall_old(struct timespec *wall_time, struct timespec *wtm,
|
||||
struct clocksource *clock, u32 mult, cycle_t cycle_last)
|
||||
struct clocksource *clock, u32 mult, u64 cycle_last)
|
||||
{
|
||||
u64 new_tb_to_xs, new_stamp_xsec;
|
||||
u32 frac_sec;
|
||||
|
@@ -209,7 +209,7 @@ void read_boot_clock64(struct timespec64 *ts)
|
||||
tod_to_timeval(clock - TOD_UNIX_EPOCH, ts);
|
||||
}
|
||||
|
||||
static cycle_t read_tod_clock(struct clocksource *cs)
|
||||
static u64 read_tod_clock(struct clocksource *cs)
|
||||
{
|
||||
unsigned long long now, adj;
|
||||
|
||||
|
@@ -148,7 +148,7 @@ static unsigned int sbus_cycles_offset(void)
|
||||
return offset;
|
||||
}
|
||||
|
||||
static cycle_t timer_cs_read(struct clocksource *cs)
|
||||
static u64 timer_cs_read(struct clocksource *cs)
|
||||
{
|
||||
unsigned int seq, offset;
|
||||
u64 cycles;
|
||||
|
@@ -770,7 +770,7 @@ void udelay(unsigned long usecs)
|
||||
}
|
||||
EXPORT_SYMBOL(udelay);
|
||||
|
||||
static cycle_t clocksource_tick_read(struct clocksource *cs)
|
||||
static u64 clocksource_tick_read(struct clocksource *cs)
|
||||
{
|
||||
return tick_ops->get_tick();
|
||||
}
|
||||
|
@@ -83,7 +83,7 @@ static irqreturn_t um_timer(int irq, void *dev)
|
||||
return IRQ_HANDLED;
|
||||
}
|
||||
|
||||
static cycle_t timer_read(struct clocksource *cs)
|
||||
static u64 timer_read(struct clocksource *cs)
|
||||
{
|
||||
return os_nsecs() / TIMER_MULTIPLIER;
|
||||
}
|
||||
|
@@ -62,7 +62,7 @@ static struct clock_event_device ckevt_puv3_osmr0 = {
|
||||
.set_state_oneshot = puv3_osmr0_shutdown,
|
||||
};
|
||||
|
||||
static cycle_t puv3_read_oscr(struct clocksource *cs)
|
||||
static u64 puv3_read_oscr(struct clocksource *cs)
|
||||
{
|
||||
return readl(OST_OSCR);
|
||||
}
|
||||
|
@@ -92,10 +92,10 @@ static notrace const struct pvclock_vsyscall_time_info *get_pvti0(void)
|
||||
return (const struct pvclock_vsyscall_time_info *)&pvclock_page;
|
||||
}
|
||||
|
||||
static notrace cycle_t vread_pvclock(int *mode)
|
||||
static notrace u64 vread_pvclock(int *mode)
|
||||
{
|
||||
const struct pvclock_vcpu_time_info *pvti = &get_pvti0()->pvti;
|
||||
cycle_t ret;
|
||||
u64 ret;
|
||||
u64 last;
|
||||
u32 version;
|
||||
|
||||
@@ -142,9 +142,9 @@ static notrace cycle_t vread_pvclock(int *mode)
|
||||
}
|
||||
#endif
|
||||
|
||||
notrace static cycle_t vread_tsc(void)
|
||||
notrace static u64 vread_tsc(void)
|
||||
{
|
||||
cycle_t ret = (cycle_t)rdtsc_ordered();
|
||||
u64 ret = (u64)rdtsc_ordered();
|
||||
u64 last = gtod->cycle_last;
|
||||
|
||||
if (likely(ret >= last))
|
||||
|
@@ -768,7 +768,7 @@ struct kvm_arch {
|
||||
spinlock_t pvclock_gtod_sync_lock;
|
||||
bool use_master_clock;
|
||||
u64 master_kernel_ns;
|
||||
cycle_t master_cycle_now;
|
||||
u64 master_cycle_now;
|
||||
struct delayed_work kvmclock_update_work;
|
||||
struct delayed_work kvmclock_sync_work;
|
||||
|
||||
|
@@ -14,7 +14,7 @@ static inline struct pvclock_vsyscall_time_info *pvclock_pvti_cpu0_va(void)
|
||||
#endif
|
||||
|
||||
/* some helper functions for xen and kvm pv clock sources */
|
||||
cycle_t pvclock_clocksource_read(struct pvclock_vcpu_time_info *src);
|
||||
u64 pvclock_clocksource_read(struct pvclock_vcpu_time_info *src);
|
||||
u8 pvclock_read_flags(struct pvclock_vcpu_time_info *src);
|
||||
void pvclock_set_flags(u8 flags);
|
||||
unsigned long pvclock_tsc_khz(struct pvclock_vcpu_time_info *src);
|
||||
@@ -87,11 +87,10 @@ static inline u64 pvclock_scale_delta(u64 delta, u32 mul_frac, int shift)
|
||||
}
|
||||
|
||||
static __always_inline
|
||||
cycle_t __pvclock_read_cycles(const struct pvclock_vcpu_time_info *src,
|
||||
u64 tsc)
|
||||
u64 __pvclock_read_cycles(const struct pvclock_vcpu_time_info *src, u64 tsc)
|
||||
{
|
||||
u64 delta = tsc - src->tsc_timestamp;
|
||||
cycle_t offset = pvclock_scale_delta(delta, src->tsc_to_system_mul,
|
||||
u64 offset = pvclock_scale_delta(delta, src->tsc_to_system_mul,
|
||||
src->tsc_shift);
|
||||
return src->system_time + offset;
|
||||
}
|
||||
|
@@ -29,7 +29,7 @@ static inline cycles_t get_cycles(void)
|
||||
return rdtsc();
|
||||
}
|
||||
|
||||
extern struct system_counterval_t convert_art_to_tsc(cycle_t art);
|
||||
extern struct system_counterval_t convert_art_to_tsc(u64 art);
|
||||
|
||||
extern void tsc_init(void);
|
||||
extern void mark_tsc_unstable(char *reason);
|
||||
|
@@ -17,8 +17,8 @@ struct vsyscall_gtod_data {
|
||||
unsigned seq;
|
||||
|
||||
int vclock_mode;
|
||||
cycle_t cycle_last;
|
||||
cycle_t mask;
|
||||
u64 cycle_last;
|
||||
u64 mask;
|
||||
u32 mult;
|
||||
u32 shift;
|
||||
|
||||
|
@@ -247,7 +247,7 @@ void apbt_setup_secondary_clock(void) {}
|
||||
static int apbt_clocksource_register(void)
|
||||
{
|
||||
u64 start, now;
|
||||
cycle_t t1;
|
||||
u64 t1;
|
||||
|
||||
/* Start the counter, use timer 2 as source, timer 0/1 for event */
|
||||
dw_apb_clocksource_start(clocksource_apbt);
|
||||
@@ -355,7 +355,7 @@ unsigned long apbt_quick_calibrate(void)
|
||||
{
|
||||
int i, scale;
|
||||
u64 old, new;
|
||||
cycle_t t1, t2;
|
||||
u64 t1, t2;
|
||||
unsigned long khz = 0;
|
||||
u32 loop, shift;
|
||||
|
||||
|
@@ -133,9 +133,9 @@ static uint32_t __init ms_hyperv_platform(void)
|
||||
return 0;
|
||||
}
|
||||
|
||||
static cycle_t read_hv_clock(struct clocksource *arg)
|
||||
static u64 read_hv_clock(struct clocksource *arg)
|
||||
{
|
||||
cycle_t current_tick;
|
||||
u64 current_tick;
|
||||
/*
|
||||
* Read the partition counter to get the current tick count. This count
|
||||
* is set to 0 when the partition is created and is incremented in
|
||||
|
@@ -791,7 +791,7 @@ static union hpet_lock hpet __cacheline_aligned = {
|
||||
{ .lock = __ARCH_SPIN_LOCK_UNLOCKED, },
|
||||
};
|
||||
|
||||
static cycle_t read_hpet(struct clocksource *cs)
|
||||
static u64 read_hpet(struct clocksource *cs)
|
||||
{
|
||||
unsigned long flags;
|
||||
union hpet_lock old, new;
|
||||
@@ -802,7 +802,7 @@ static cycle_t read_hpet(struct clocksource *cs)
|
||||
* Read HPET directly if in NMI.
|
||||
*/
|
||||
if (in_nmi())
|
||||
return (cycle_t)hpet_readl(HPET_COUNTER);
|
||||
return (u64)hpet_readl(HPET_COUNTER);
|
||||
|
||||
/*
|
||||
* Read the current state of the lock and HPET value atomically.
|
||||
@@ -821,7 +821,7 @@ static cycle_t read_hpet(struct clocksource *cs)
|
||||
WRITE_ONCE(hpet.value, new.value);
|
||||
arch_spin_unlock(&hpet.lock);
|
||||
local_irq_restore(flags);
|
||||
return (cycle_t)new.value;
|
||||
return (u64)new.value;
|
||||
}
|
||||
local_irq_restore(flags);
|
||||
|
||||
@@ -843,15 +843,15 @@ contended:
|
||||
new.lockval = READ_ONCE(hpet.lockval);
|
||||
} while ((new.value == old.value) && arch_spin_is_locked(&new.lock));
|
||||
|
||||
return (cycle_t)new.value;
|
||||
return (u64)new.value;
|
||||
}
|
||||
#else
|
||||
/*
|
||||
* For UP or 32-bit.
|
||||
*/
|
||||
static cycle_t read_hpet(struct clocksource *cs)
|
||||
static u64 read_hpet(struct clocksource *cs)
|
||||
{
|
||||
return (cycle_t)hpet_readl(HPET_COUNTER);
|
||||
return (u64)hpet_readl(HPET_COUNTER);
|
||||
}
|
||||
#endif
|
||||
|
||||
@@ -867,7 +867,7 @@ static struct clocksource clocksource_hpet = {
|
||||
static int hpet_clocksource_register(void)
|
||||
{
|
||||
u64 start, now;
|
||||
cycle_t t1;
|
||||
u64 t1;
|
||||
|
||||
/* Start the counter */
|
||||
hpet_restart_counter();
|
||||
|
@@ -32,7 +32,7 @@
|
||||
static int kvmclock __ro_after_init = 1;
|
||||
static int msr_kvm_system_time = MSR_KVM_SYSTEM_TIME;
|
||||
static int msr_kvm_wall_clock = MSR_KVM_WALL_CLOCK;
|
||||
static cycle_t kvm_sched_clock_offset;
|
||||
static u64 kvm_sched_clock_offset;
|
||||
|
||||
static int parse_no_kvmclock(char *arg)
|
||||
{
|
||||
@@ -79,10 +79,10 @@ static int kvm_set_wallclock(const struct timespec *now)
|
||||
return -1;
|
||||
}
|
||||
|
||||
static cycle_t kvm_clock_read(void)
|
||||
static u64 kvm_clock_read(void)
|
||||
{
|
||||
struct pvclock_vcpu_time_info *src;
|
||||
cycle_t ret;
|
||||
u64 ret;
|
||||
int cpu;
|
||||
|
||||
preempt_disable_notrace();
|
||||
@@ -93,12 +93,12 @@ static cycle_t kvm_clock_read(void)
|
||||
return ret;
|
||||
}
|
||||
|
||||
static cycle_t kvm_clock_get_cycles(struct clocksource *cs)
|
||||
static u64 kvm_clock_get_cycles(struct clocksource *cs)
|
||||
{
|
||||
return kvm_clock_read();
|
||||
}
|
||||
|
||||
static cycle_t kvm_sched_clock_read(void)
|
||||
static u64 kvm_sched_clock_read(void)
|
||||
{
|
||||
return kvm_clock_read() - kvm_sched_clock_offset;
|
||||
}
|
||||
|
@@ -71,10 +71,10 @@ u8 pvclock_read_flags(struct pvclock_vcpu_time_info *src)
|
||||
return flags & valid_flags;
|
||||
}
|
||||
|
||||
cycle_t pvclock_clocksource_read(struct pvclock_vcpu_time_info *src)
|
||||
u64 pvclock_clocksource_read(struct pvclock_vcpu_time_info *src)
|
||||
{
|
||||
unsigned version;
|
||||
cycle_t ret;
|
||||
u64 ret;
|
||||
u64 last;
|
||||
u8 flags;
|
||||
|
||||
|
@@ -1101,9 +1101,9 @@ static void tsc_resume(struct clocksource *cs)
|
||||
* checking the result of read_tsc() - cycle_last for being negative.
|
||||
* That works because CLOCKSOURCE_MASK(64) does not mask out any bit.
|
||||
*/
|
||||
static cycle_t read_tsc(struct clocksource *cs)
|
||||
static u64 read_tsc(struct clocksource *cs)
|
||||
{
|
||||
return (cycle_t)rdtsc_ordered();
|
||||
return (u64)rdtsc_ordered();
|
||||
}
|
||||
|
||||
/*
|
||||
@@ -1192,7 +1192,7 @@ int unsynchronized_tsc(void)
|
||||
/*
|
||||
* Convert ART to TSC given numerator/denominator found in detect_art()
|
||||
*/
|
||||
struct system_counterval_t convert_art_to_tsc(cycle_t art)
|
||||
struct system_counterval_t convert_art_to_tsc(u64 art)
|
||||
{
|
||||
u64 tmp, res, rem;
|
||||
|
||||
|
@@ -1131,8 +1131,8 @@ struct pvclock_gtod_data {
|
||||
|
||||
struct { /* extract of a clocksource struct */
|
||||
int vclock_mode;
|
||||
cycle_t cycle_last;
|
||||
cycle_t mask;
|
||||
u64 cycle_last;
|
||||
u64 mask;
|
||||
u32 mult;
|
||||
u32 shift;
|
||||
} clock;
|
||||
@@ -1572,9 +1572,9 @@ static inline void adjust_tsc_offset_host(struct kvm_vcpu *vcpu, s64 adjustment)
|
||||
|
||||
#ifdef CONFIG_X86_64
|
||||
|
||||
static cycle_t read_tsc(void)
|
||||
static u64 read_tsc(void)
|
||||
{
|
||||
cycle_t ret = (cycle_t)rdtsc_ordered();
|
||||
u64 ret = (u64)rdtsc_ordered();
|
||||
u64 last = pvclock_gtod_data.clock.cycle_last;
|
||||
|
||||
if (likely(ret >= last))
|
||||
@@ -1592,7 +1592,7 @@ static cycle_t read_tsc(void)
|
||||
return last;
|
||||
}
|
||||
|
||||
static inline u64 vgettsc(cycle_t *cycle_now)
|
||||
static inline u64 vgettsc(u64 *cycle_now)
|
||||
{
|
||||
long v;
|
||||
struct pvclock_gtod_data *gtod = &pvclock_gtod_data;
|
||||
@@ -1603,7 +1603,7 @@ static inline u64 vgettsc(cycle_t *cycle_now)
|
||||
return v * gtod->clock.mult;
|
||||
}
|
||||
|
||||
static int do_monotonic_boot(s64 *t, cycle_t *cycle_now)
|
||||
static int do_monotonic_boot(s64 *t, u64 *cycle_now)
|
||||
{
|
||||
struct pvclock_gtod_data *gtod = &pvclock_gtod_data;
|
||||
unsigned long seq;
|
||||
@@ -1624,7 +1624,7 @@ static int do_monotonic_boot(s64 *t, cycle_t *cycle_now)
|
||||
}
|
||||
|
||||
/* returns true if host is using tsc clocksource */
|
||||
static bool kvm_get_time_and_clockread(s64 *kernel_ns, cycle_t *cycle_now)
|
||||
static bool kvm_get_time_and_clockread(s64 *kernel_ns, u64 *cycle_now)
|
||||
{
|
||||
/* checked again under seqlock below */
|
||||
if (pvclock_gtod_data.clock.vclock_mode != VCLOCK_TSC)
|
||||
|
@@ -916,7 +916,7 @@ static unsigned long lguest_tsc_khz(void)
|
||||
* If we can't use the TSC, the kernel falls back to our lower-priority
|
||||
* "lguest_clock", where we read the time value given to us by the Host.
|
||||
*/
|
||||
static cycle_t lguest_clock_read(struct clocksource *cs)
|
||||
static u64 lguest_clock_read(struct clocksource *cs)
|
||||
{
|
||||
unsigned long sec, nsec;
|
||||
|
||||
|
@@ -30,7 +30,7 @@
|
||||
|
||||
#define RTC_NAME "sgi_rtc"
|
||||
|
||||
static cycle_t uv_read_rtc(struct clocksource *cs);
|
||||
static u64 uv_read_rtc(struct clocksource *cs);
|
||||
static int uv_rtc_next_event(unsigned long, struct clock_event_device *);
|
||||
static int uv_rtc_shutdown(struct clock_event_device *evt);
|
||||
|
||||
@@ -38,7 +38,7 @@ static struct clocksource clocksource_uv = {
|
||||
.name = RTC_NAME,
|
||||
.rating = 299,
|
||||
.read = uv_read_rtc,
|
||||
.mask = (cycle_t)UVH_RTC_REAL_TIME_CLOCK_MASK,
|
||||
.mask = (u64)UVH_RTC_REAL_TIME_CLOCK_MASK,
|
||||
.flags = CLOCK_SOURCE_IS_CONTINUOUS,
|
||||
};
|
||||
|
||||
@@ -296,7 +296,7 @@ static int uv_rtc_unset_timer(int cpu, int force)
|
||||
* cachelines of it's own page. This allows faster simultaneous reads
|
||||
* from a given socket.
|
||||
*/
|
||||
static cycle_t uv_read_rtc(struct clocksource *cs)
|
||||
static u64 uv_read_rtc(struct clocksource *cs)
|
||||
{
|
||||
unsigned long offset;
|
||||
|
||||
@@ -305,7 +305,7 @@ static cycle_t uv_read_rtc(struct clocksource *cs)
|
||||
else
|
||||
offset = (uv_blade_processor_id() * L1_CACHE_BYTES) % PAGE_SIZE;
|
||||
|
||||
return (cycle_t)uv_read_local_mmr(UVH_RTC | offset);
|
||||
return (u64)uv_read_local_mmr(UVH_RTC | offset);
|
||||
}
|
||||
|
||||
/*
|
||||
|
@@ -39,10 +39,10 @@ static unsigned long xen_tsc_khz(void)
|
||||
return pvclock_tsc_khz(info);
|
||||
}
|
||||
|
||||
cycle_t xen_clocksource_read(void)
|
||||
u64 xen_clocksource_read(void)
|
||||
{
|
||||
struct pvclock_vcpu_time_info *src;
|
||||
cycle_t ret;
|
||||
u64 ret;
|
||||
|
||||
preempt_disable_notrace();
|
||||
src = &__this_cpu_read(xen_vcpu)->time;
|
||||
@@ -51,7 +51,7 @@ cycle_t xen_clocksource_read(void)
|
||||
return ret;
|
||||
}
|
||||
|
||||
static cycle_t xen_clocksource_get_cycles(struct clocksource *cs)
|
||||
static u64 xen_clocksource_get_cycles(struct clocksource *cs)
|
||||
{
|
||||
return xen_clocksource_read();
|
||||
}
|
||||
|
@@ -67,7 +67,7 @@ void xen_init_irq_ops(void);
|
||||
void xen_setup_timer(int cpu);
|
||||
void xen_setup_runstate_info(int cpu);
|
||||
void xen_teardown_timer(int cpu);
|
||||
cycle_t xen_clocksource_read(void);
|
||||
u64 xen_clocksource_read(void);
|
||||
void xen_setup_cpu_clockevents(void);
|
||||
void __init xen_init_time_ops(void);
|
||||
void __init xen_hvm_init_time_ops(void);
|
||||
|
@@ -34,9 +34,9 @@
|
||||
unsigned long ccount_freq; /* ccount Hz */
|
||||
EXPORT_SYMBOL(ccount_freq);
|
||||
|
||||
static cycle_t ccount_read(struct clocksource *cs)
|
||||
static u64 ccount_read(struct clocksource *cs)
|
||||
{
|
||||
return (cycle_t)get_ccount();
|
||||
return (u64)get_ccount();
|
||||
}
|
||||
|
||||
static u64 notrace ccount_sched_clock_read(void)
|
||||
|
مرجع در شماره جدید
Block a user