@@ -72,6 +72,7 @@
#include <xen/bug.h>
#include <xen/types.h>
#include <xen/kernel.h>
+#include <xen/linux-compat.h>
#include <asm/sysregs.h>
#include <asm/cpufeature.h>
#include <asm/arm64/cpufeature.h>
@@ -119,7 +119,7 @@ static u64 do_reloc(enum aarch64_reloc_o
static int reloc_data(enum aarch64_reloc_op op, void *place, u64 val, int len)
{
- s64 sval = do_reloc(op, place, val);
+ int64_t sval = do_reloc(op, place, val);
switch ( len )
{
@@ -136,7 +136,7 @@ static int reloc_data(enum aarch64_reloc
break;
case 64:
- *(s64 *)place = sval;
+ *(int64_t *)place = sval;
break;
default:
@@ -155,9 +155,9 @@ enum aarch64_insn_movw_imm_type {
static int reloc_insn_movw(enum aarch64_reloc_op op, void *dest, u64 val,
int lsb, enum aarch64_insn_movw_imm_type imm_type)
{
- u64 imm;
- s64 sval;
- u32 insn = *(u32 *)dest;
+ uint64_t imm;
+ int64_t sval;
+ uint32_t insn = *(uint32_t *)dest;
sval = do_reloc(op, dest, val);
imm = sval >> lsb;
@@ -200,9 +200,9 @@ static int reloc_insn_movw(enum aarch64_
static int reloc_insn_imm(enum aarch64_reloc_op op, void *dest, u64 val,
int lsb, int len, enum aarch64_insn_imm_type imm_type)
{
- u64 imm, imm_mask;
- s64 sval;
- u32 insn = *(u32 *)dest;
+ uint64_t imm, imm_mask;
+ int64_t sval;
+ uint32_t insn = *(uint32_t *)dest;
/* Calculate the relocation value. */
sval = do_reloc(op, dest, val);
@@ -220,7 +220,7 @@ static int reloc_insn_imm(enum aarch64_r
* Extract the upper value bits (including the sign bit) and
* shift them to bit 0.
*/
- sval = (s64)(sval & ~(imm_mask >> 1)) >> (len - 1);
+ sval = (int64_t)(sval & ~(imm_mask >> 1)) >> (len - 1);
/*
* Overflow has occurred if the upper bits are not all equal to
@@ -39,15 +39,15 @@ struct arm64_ftr_bits {
bool visible;
bool strict; /* CPU Sanity check: strict matching required ? */
enum ftr_type type;
- u8 shift;
- u8 width;
- s64 safe_val; /* safe value for FTR_EXACT features */
+ uint8_t shift;
+ uint8_t width;
+ int64_t safe_val; /* safe value for FTR_EXACT features */
};
static inline int attr_const
cpuid_feature_extract_signed_field_width(u64 features, int field, int width)
{
- return (s64)(features << (64 - width - field)) >> (64 - width);
+ return (int64_t)(features << (64 - width - field)) >> (64 - width);
}
static inline int attr_const
@@ -87,9 +87,12 @@ cpuid_feature_extract_field(u64 features
return cpuid_feature_extract_field_width(features, field, 4, sign);
}
-static inline s64 arm64_ftr_value(const struct arm64_ftr_bits *ftrp, u64 val)
+static inline int64_t arm64_ftr_value(const struct arm64_ftr_bits *ftrp,
+ uint64_t val)
{
- return (s64)cpuid_feature_extract_field_width(val, ftrp->shift, ftrp->width, ftrp->sign);
+ return (int64_t)cpuid_feature_extract_field_width(val, ftrp->shift,
+ ftrp->width,
+ ftrp->sign);
}
#endif /* _ASM_ARM_ARM64_CPUFEATURES_H */
@@ -1223,7 +1223,7 @@ void vlapic_tdt_msr_set(struct vlapic *v
if ( value > guest_tsc )
{
uint64_t delta = gtsc_to_gtime(v->domain, value - guest_tsc);
- delta = max_t(s64, delta, 0);
+ delta = max_t(int64_t, delta, 0);
HVM_DBG_LOG(DBG_LEVEL_VLAPIC_TIMER, "delta[0x%016"PRIx64"]", delta);
@@ -134,8 +134,8 @@ struct hvm_vcpu {
unsigned long hw_cr[5];
struct vlapic vlapic;
- s64 cache_tsc_offset;
- u64 guest_time;
+ int64_t cache_tsc_offset;
+ uint64_t guest_time;
/* Lock and list for virtual platform timers. */
spinlock_t tm_lock;
@@ -66,10 +66,10 @@ struct cpu_time {
struct platform_timesource {
const char *id;
const char *name;
- u64 frequency;
+ uint64_t frequency;
/* Post-init this hook may only be invoked via the read_counter() wrapper! */
- u64 (*read_counter)(void);
- s64 (*init)(struct platform_timesource *);
+ uint64_t (*read_counter)(void);
+ int64_t (*init)(struct platform_timesource *);
void (*resume)(struct platform_timesource *);
int counter_bits;
};
@@ -368,7 +368,7 @@ static u64 cf_check read_pit_count(void)
return count32;
}
-static s64 __init cf_check init_pit(struct platform_timesource *pts)
+static int64_t __init cf_check init_pit(struct platform_timesource *pts)
{
u8 portb = inb(0x61);
u64 start, end;
@@ -610,7 +610,7 @@ static u64 cf_check read_pmtimer_count(v
return inl(pmtmr_ioport);
}
-static s64 __init cf_check init_pmtimer(struct platform_timesource *pts)
+static int64_t __init cf_check init_pmtimer(struct platform_timesource *pts)
{
if ( !pmtmr_ioport || (pmtmr_width != 24 && pmtmr_width != 32) )
return 0;
@@ -655,7 +655,7 @@ static unsigned int __initdata tsc_flags
* Called in verify_tsc_reliability() under reliable TSC conditions
* thus reusing all the checks already performed there.
*/
-static s64 __init cf_check init_tsc(struct platform_timesource *pts)
+static int64_t __init cf_check init_tsc(struct platform_timesource *pts)
{
u64 ret = pts->frequency;
@@ -1034,9 +1034,9 @@ static void __init reset_platform_timer(
spin_unlock_irq(&platform_timer_lock);
}
-static s64 __init try_platform_timer(struct platform_timesource *pts)
+static int64_t __init try_platform_timer(struct platform_timesource *pts)
{
- s64 rc = pts->init(pts);
+ int64_t rc = pts->init(pts);
if ( rc <= 0 )
return rc;
@@ -1070,7 +1070,7 @@ static u64 __init init_platform_timer(vo
struct platform_timesource *pts = NULL;
unsigned int i;
- s64 rc = -1;
+ int64_t rc = -1;
/* clocksource=tsc is initialized via __initcalls (when CPUs are up). */
if ( (opt_clocksource[0] != '\0') && strcmp(opt_clocksource, "tsc") )
@@ -1923,7 +1923,7 @@ static void cf_check local_time_calibrat
* Weirdness can happen if we lose sync with the platform timer.
* We could be smarter here: resync platform timer with local timer?
*/
- if ( ((s64)stime_elapsed64 < (EPOCH / 2)) )
+ if ( ((int64_t)stime_elapsed64 < (EPOCH / 2)) )
goto out;
/*
@@ -2398,7 +2398,7 @@ static void __init tsc_check_writability
write_tsc(tsc | (1ULL << 32));
tmp = rdtsc();
- if ( ABS((s64)tmp - (s64)tmp2) < (1LL << 31) )
+ if ( ABS((int64_t)tmp - (int64_t)tmp2) < (1LL << 31) )
what = "only partially";
}
else
@@ -105,7 +105,7 @@ static inline int erst_errno(int command
static int erst_timedout(u64 *t, u64 spin_unit)
{
- if ((s64)*t < spin_unit) {
+ if ((int64_t)*t < spin_unit) {
printk(XENLOG_WARNING "Firmware does not respond in time\n");
return 1;
}
@@ -147,8 +147,8 @@ typedef int INT32;
/*! [End] no source code translation !*/
-typedef u64 acpi_native_uint;
-typedef s64 acpi_native_int;
+typedef uint64_t acpi_native_uint;
+typedef int64_t acpi_native_int;
typedef u64 acpi_io_address;
typedef u64 acpi_physical_address;
@@ -15,7 +15,7 @@ typedef int8_t s8, __s8;
typedef uint8_t __u8;
typedef int16_t s16, __s16;
typedef int32_t s32, __s32;
-typedef int64_t __s64;
+typedef int64_t s64, __s64;
typedef paddr_t phys_addr_t;
@@ -29,7 +29,7 @@ struct vcpu;
* of real time into system time
*/
-typedef s64 s_time_t;
+typedef int64_t s_time_t;
#define PRI_stime PRId64
s_time_t get_s_time_fixed(u64 at_tick);
@@ -8,7 +8,6 @@
typedef uint8_t u8;
typedef uint16_t u16, __u16;
typedef uint32_t u32, __u32;
-typedef int64_t s64;
typedef uint64_t u64, __u64;
#include <asm/types.h>
@@ -46,8 +46,8 @@
* one or more of the following formats.
*/
union uu {
- s64 q; /* as a (signed) quad */
- s64 uq; /* as an unsigned quad */
+ int64_t q; /* as a (signed) quad */
+ uint64_t uq; /* as an unsigned quad */
long sl[2]; /* as two signed longs */
unsigned long ul[2]; /* as two unsigned longs */
};
@@ -72,7 +72,7 @@ union uu {
* and assembly.
*/
#define CHAR_BIT 8 /* number of bits in a char */
-#define QUAD_BITS (sizeof(s64) * CHAR_BIT)
+#define QUAD_BITS (sizeof(int64_t) * CHAR_BIT)
#define LONG_BITS (sizeof(long) * CHAR_BIT)
#define HALF_BITS (sizeof(long) * CHAR_BIT / 2)
@@ -324,7 +324,7 @@ u64 __qdivrem(u64 uq, u64 vq, u64 *arq)
* Divide two signed quads.
* Truncates towards zero, as required by C99.
*/
-s64 __divdi3(s64 a, s64 b)
+int64_t __divdi3(int64_t a, int64_t b)
{
u64 ua, ub, uq;
int neg = (a < 0) ^ (b < 0);
@@ -361,7 +361,7 @@ u64 __umoddi3(u64 a, u64 b)
* 11 % -5 = 1
* -11 % -5 = -1
*/
-s64 __moddi3(s64 a, s64 b)
+int64_t __moddi3(int64_t a, int64_t b)
{
u64 ua, ub, urem;
int neg = (a < 0);
@@ -374,7 +374,7 @@ s64 __moddi3(s64 a, s64 b)
/*
* Quotient and remainder of unsigned long long division
*/
-s64 __ldivmod_helper(s64 a, s64 b, s64 *r)
+int64_t __ldivmod_helper(int64_t a, int64_t b, int64_t *r)
{
u64 ua, ub, rem, quot;