| /* | 
 |  * RISC-V linux replacement vdso. | 
 |  * | 
 |  * Copyright 2021 Linaro, Ltd. | 
 |  * | 
 |  * SPDX-License-Identifier: GPL-2.0-or-later | 
 |  */ | 
 |  | 
 | #include <asm/unistd.h> | 
 | #include <asm/errno.h> | 
 |  | 
 | #if __riscv_xlen == 32 | 
 | # define TARGET_ABI32 | 
 | #endif | 
 | #include "vdso-asmoffset.h" | 
 |  | 
 | 	.text | 
 |  | 
 | .macro endf name | 
 | 	.globl	\name | 
 | 	.type	\name, @function | 
 | 	.size	\name, . - \name | 
 | .endm | 
 |  | 
 | .macro raw_syscall nr | 
 | 	li	a7, \nr | 
 | 	ecall | 
 | .endm | 
 |  | 
 | .macro vdso_syscall name, nr | 
 | \name: | 
 | 	raw_syscall \nr | 
 | 	ret | 
 | endf	\name | 
 | .endm | 
 |  | 
 | __vdso_gettimeofday: | 
 | 	.cfi_startproc | 
 | #ifdef __NR_gettimeofday | 
 | 	raw_syscall __NR_gettimeofday | 
 | 	ret | 
 | #else | 
 | 	/* No gettimeofday, fall back to clock_gettime64. */ | 
 | 	beq	a1, zero, 1f | 
 | 	sw	zero, 0(a1)	/* tz->tz_minuteswest = 0 */ | 
 | 	sw	zero, 4(a1)	/* tz->tz_dsttime = 0 */ | 
 | 1:	addi	sp, sp, -32 | 
 | 	.cfi_adjust_cfa_offset 32 | 
 | 	sw	a0, 16(sp)	/* save tv */ | 
 | 	mv	a0, sp | 
 | 	raw_syscall __NR_clock_gettime64 | 
 | 	lw	t0, 0(sp)	/* timespec.tv_sec.low */ | 
 | 	lw	t1, 4(sp)	/* timespec.tv_sec.high */ | 
 | 	lw	t2, 8(sp)	/* timespec.tv_nsec.low */ | 
 | 	lw	a1, 16(sp)	/* restore tv */ | 
 | 	addi	sp, sp, 32 | 
 | 	.cfi_adjust_cfa_offset -32 | 
 | 	bne	a0, zero, 9f	/* syscall error? */ | 
 | 	li	a0, -EOVERFLOW | 
 | 	bne	t1, zero, 9f	/* y2038? */ | 
 | 	li	a0, 0 | 
 | 	li	t3, 1000 | 
 | 	divu	t2, t2, t3	/* nsec -> usec */ | 
 | 	sw	t0, 0(a1)	/* tz->tv_sec */ | 
 | 	sw	t2, 4(a1)	/* tz->tv_usec */ | 
 | 9:	ret | 
 | #endif | 
 | 	.cfi_endproc | 
 | endf __vdso_gettimeofday | 
 |  | 
 | 	.cfi_startproc | 
 |  | 
 | #ifdef __NR_clock_gettime | 
 | vdso_syscall __vdso_clock_gettime, __NR_clock_gettime | 
 | #else | 
 | vdso_syscall __vdso_clock_gettime, __NR_clock_gettime64 | 
 | #endif | 
 |  | 
 | #ifdef __NR_clock_getres | 
 | vdso_syscall __vdso_clock_getres, __NR_clock_getres | 
 | #else | 
 | vdso_syscall __vdso_clock_getres, __NR_clock_getres_time64 | 
 | #endif | 
 |  | 
 | vdso_syscall __vdso_getcpu, __NR_getcpu | 
 |  | 
 | __vdso_flush_icache: | 
 | 	/* qemu does not need to flush the icache */ | 
 | 	li	a0, 0 | 
 | 	ret | 
 | endf __vdso_flush_icache | 
 |  | 
 | 	.cfi_endproc | 
 |  | 
 | /* | 
 |  * Start the unwind info at least one instruction before the signal | 
 |  * trampoline, because the unwinder will assume we are returning | 
 |  * after a call site. | 
 |  */ | 
 |  | 
 | 	.cfi_startproc simple | 
 | 	.cfi_signal_frame | 
 |  | 
 | #define sizeof_reg	(__riscv_xlen / 4) | 
 | #define sizeof_freg	8 | 
 | #define B_GR	(offsetof_uc_mcontext - sizeof_rt_sigframe) | 
 | #define B_FR	(offsetof_uc_mcontext - sizeof_rt_sigframe + offsetof_freg0) | 
 |  | 
 | 	.cfi_def_cfa	2, sizeof_rt_sigframe | 
 |  | 
 | 	/* Return address */ | 
 | 	.cfi_return_column 64 | 
 | 	.cfi_offset	64, B_GR + 0			/* pc */ | 
 |  | 
 | 	/* Integer registers */ | 
 | 	.cfi_offset	1, B_GR + 1 * sizeof_reg	/* r1 (ra) */ | 
 | 	.cfi_offset	2, B_GR + 2 * sizeof_reg	/* r2 (sp) */ | 
 | 	.cfi_offset	3, B_GR + 3 * sizeof_reg | 
 | 	.cfi_offset	4, B_GR + 4 * sizeof_reg | 
 | 	.cfi_offset	5, B_GR + 5 * sizeof_reg | 
 | 	.cfi_offset	6, B_GR + 6 * sizeof_reg | 
 | 	.cfi_offset	7, B_GR + 7 * sizeof_reg | 
 | 	.cfi_offset	8, B_GR + 8 * sizeof_reg | 
 | 	.cfi_offset	9, B_GR + 9 * sizeof_reg | 
 | 	.cfi_offset	10, B_GR + 10 * sizeof_reg | 
 | 	.cfi_offset	11, B_GR + 11 * sizeof_reg | 
 | 	.cfi_offset	12, B_GR + 12 * sizeof_reg | 
 | 	.cfi_offset	13, B_GR + 13 * sizeof_reg | 
 | 	.cfi_offset	14, B_GR + 14 * sizeof_reg | 
 | 	.cfi_offset	15, B_GR + 15 * sizeof_reg | 
 | 	.cfi_offset	16, B_GR + 16 * sizeof_reg | 
 | 	.cfi_offset	17, B_GR + 17 * sizeof_reg | 
 | 	.cfi_offset	18, B_GR + 18 * sizeof_reg | 
 | 	.cfi_offset	19, B_GR + 19 * sizeof_reg | 
 | 	.cfi_offset	20, B_GR + 20 * sizeof_reg | 
 | 	.cfi_offset	21, B_GR + 21 * sizeof_reg | 
 | 	.cfi_offset	22, B_GR + 22 * sizeof_reg | 
 | 	.cfi_offset	23, B_GR + 23 * sizeof_reg | 
 | 	.cfi_offset	24, B_GR + 24 * sizeof_reg | 
 | 	.cfi_offset	25, B_GR + 25 * sizeof_reg | 
 | 	.cfi_offset	26, B_GR + 26 * sizeof_reg | 
 | 	.cfi_offset	27, B_GR + 27 * sizeof_reg | 
 | 	.cfi_offset	28, B_GR + 28 * sizeof_reg | 
 | 	.cfi_offset	29, B_GR + 29 * sizeof_reg | 
 | 	.cfi_offset	30, B_GR + 30 * sizeof_reg | 
 | 	.cfi_offset	31, B_GR + 31 * sizeof_reg	/* r31 */ | 
 |  | 
 | 	.cfi_offset	32, B_FR + 0			/* f0 */ | 
 | 	.cfi_offset	33, B_FR + 1 * sizeof_freg	/* f1 */ | 
 | 	.cfi_offset	34, B_FR + 2 * sizeof_freg | 
 | 	.cfi_offset	35, B_FR + 3 * sizeof_freg | 
 | 	.cfi_offset	36, B_FR + 4 * sizeof_freg | 
 | 	.cfi_offset	37, B_FR + 5 * sizeof_freg | 
 | 	.cfi_offset	38, B_FR + 6 * sizeof_freg | 
 | 	.cfi_offset	39, B_FR + 7 * sizeof_freg | 
 | 	.cfi_offset	40, B_FR + 8 * sizeof_freg | 
 | 	.cfi_offset	41, B_FR + 9 * sizeof_freg | 
 | 	.cfi_offset	42, B_FR + 10 * sizeof_freg | 
 | 	.cfi_offset	43, B_FR + 11 * sizeof_freg | 
 | 	.cfi_offset	44, B_FR + 12 * sizeof_freg | 
 | 	.cfi_offset	45, B_FR + 13 * sizeof_freg | 
 | 	.cfi_offset	46, B_FR + 14 * sizeof_freg | 
 | 	.cfi_offset	47, B_FR + 15 * sizeof_freg | 
 | 	.cfi_offset	48, B_FR + 16 * sizeof_freg | 
 | 	.cfi_offset	49, B_FR + 17 * sizeof_freg | 
 | 	.cfi_offset	50, B_FR + 18 * sizeof_freg | 
 | 	.cfi_offset	51, B_FR + 19 * sizeof_freg | 
 | 	.cfi_offset	52, B_FR + 20 * sizeof_freg | 
 | 	.cfi_offset	53, B_FR + 21 * sizeof_freg | 
 | 	.cfi_offset	54, B_FR + 22 * sizeof_freg | 
 | 	.cfi_offset	55, B_FR + 23 * sizeof_freg | 
 | 	.cfi_offset	56, B_FR + 24 * sizeof_freg | 
 | 	.cfi_offset	57, B_FR + 25 * sizeof_freg | 
 | 	.cfi_offset	58, B_FR + 26 * sizeof_freg | 
 | 	.cfi_offset	59, B_FR + 27 * sizeof_freg | 
 | 	.cfi_offset	60, B_FR + 28 * sizeof_freg | 
 | 	.cfi_offset	61, B_FR + 29 * sizeof_freg | 
 | 	.cfi_offset	62, B_FR + 30 * sizeof_freg | 
 | 	.cfi_offset	63, B_FR + 31 * sizeof_freg	/* f31 */ | 
 |  | 
 | 	nop | 
 |  | 
 | __vdso_rt_sigreturn: | 
 | 	raw_syscall __NR_rt_sigreturn | 
 | endf __vdso_rt_sigreturn | 
 |  | 
 | 	.cfi_endproc |