mirror of
https://github.com/ClickHouse/ClickHouse.git
synced 2024-10-04 23:50:50 +00:00
2881 lines
121 KiB
C++
2881 lines
121 KiB
C++
// -*- Mode: C++; c-basic-offset: 2; indent-tabs-mode: nil -*-
|
|
/* Copyright (c) 2005-2008, Google Inc.
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions are
|
|
* met:
|
|
*
|
|
* * Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* * Redistributions in binary form must reproduce the above
|
|
* copyright notice, this list of conditions and the following disclaimer
|
|
* in the documentation and/or other materials provided with the
|
|
* distribution.
|
|
* * Neither the name of Google Inc. nor the names of its
|
|
* contributors may be used to endorse or promote products derived from
|
|
* this software without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
|
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
|
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
|
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
|
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
|
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
*
|
|
* ---
|
|
* Author: Markus Gutschke
|
|
*/
|
|
|
|
/* This file includes Linux-specific support functions common to the
|
|
* coredumper and the thread lister; primarily, this is a collection
|
|
* of direct system calls, and a couple of symbols missing from
|
|
* standard header files.
|
|
* There are a few options that the including file can set to control
|
|
* the behavior of this file:
|
|
*
|
|
* SYS_CPLUSPLUS:
|
|
* The entire header file will normally be wrapped in 'extern "C" { }",
|
|
* making it suitable for compilation as both C and C++ source. If you
|
|
* do not want to do this, you can set the SYS_CPLUSPLUS macro to inhibit
|
|
* the wrapping. N.B. doing so will suppress inclusion of all prerequisite
|
|
* system header files, too. It is the caller's responsibility to provide
|
|
* the necessary definitions.
|
|
*
|
|
* SYS_ERRNO:
|
|
* All system calls will update "errno" unless overriden by setting the
|
|
* SYS_ERRNO macro prior to including this file. SYS_ERRNO should be
|
|
* an l-value.
|
|
*
|
|
* SYS_INLINE:
|
|
* New symbols will be defined "static inline", unless overridden by
|
|
* the SYS_INLINE macro.
|
|
*
|
|
* SYS_LINUX_SYSCALL_SUPPORT_H
|
|
* This macro is used to avoid multiple inclusions of this header file.
|
|
* If you need to include this file more than once, make sure to
|
|
* unset SYS_LINUX_SYSCALL_SUPPORT_H before each inclusion.
|
|
*
|
|
* SYS_PREFIX:
|
|
* New system calls will have a prefix of "sys_" unless overridden by
|
|
* the SYS_PREFIX macro. Valid values for this macro are [0..9] which
|
|
* results in prefixes "sys[0..9]_". It is also possible to set this
|
|
* macro to -1, which avoids all prefixes.
|
|
*
|
|
* This file defines a few internal symbols that all start with "LSS_".
|
|
* Do not access these symbols from outside this file. They are not part
|
|
* of the supported API.
|
|
*
|
|
* NOTE: This is a stripped down version of the official opensource
|
|
* version of linux_syscall_support.h, which lives at
|
|
* http://code.google.com/p/linux-syscall-support/
|
|
* It includes only the syscalls that are used in perftools, plus a
|
|
* few extra. Here's the breakdown:
|
|
* 1) Perftools uses these: grep -rho 'sys_[a-z0-9_A-Z]* *(' src | sort -u
|
|
* sys__exit(
|
|
* sys_clone(
|
|
* sys_close(
|
|
* sys_fcntl(
|
|
* sys_fstat(
|
|
* sys_futex(
|
|
* sys_getcpu(
|
|
* sys_getdents64(
|
|
* sys_getppid(
|
|
* sys_gettid(
|
|
* sys_lseek(
|
|
* sys_mmap(
|
|
* sys_mremap(
|
|
* sys_munmap(
|
|
* sys_open(
|
|
* sys_pipe(
|
|
* sys_prctl(
|
|
* sys_ptrace(
|
|
* sys_ptrace_detach(
|
|
* sys_read(
|
|
* sys_sched_yield(
|
|
* sys_sigaction(
|
|
* sys_sigaltstack(
|
|
* sys_sigdelset(
|
|
* sys_sigfillset(
|
|
* sys_sigprocmask(
|
|
* sys_socket(
|
|
* sys_stat(
|
|
* sys_waitpid(
|
|
* 2) These are used as subroutines of the above:
|
|
* sys_getpid -- gettid
|
|
* sys_kill -- ptrace_detach
|
|
* sys_restore -- sigaction
|
|
* sys_restore_rt -- sigaction
|
|
* sys_socketcall -- socket
|
|
* sys_wait4 -- waitpid
|
|
* 3) I left these in even though they're not used. They either
|
|
* complement the above (write vs read) or are variants (rt_sigaction):
|
|
* sys_fstat64
|
|
* sys_llseek
|
|
* sys_mmap2
|
|
* sys_openat
|
|
* sys_getdents
|
|
* sys_rt_sigaction
|
|
* sys_rt_sigprocmask
|
|
* sys_sigaddset
|
|
* sys_sigemptyset
|
|
* sys_stat64
|
|
* sys_write
|
|
*/
|
|
#ifndef SYS_LINUX_SYSCALL_SUPPORT_H
|
|
#define SYS_LINUX_SYSCALL_SUPPORT_H
|
|
|
|
/* We currently only support x86-32, x86-64, ARM, MIPS, PPC/PPC64, Aarch64, s390 and s390x
|
|
* on Linux.
|
|
* Porting to other related platforms should not be difficult.
|
|
*/
|
|
#if (defined(__i386__) || defined(__x86_64__) || defined(__arm__) || \
|
|
defined(__mips__) || defined(__PPC__) || \
|
|
defined(__aarch64__) || defined(__s390__)) \
|
|
&& (defined(__linux))
|
|
|
|
#ifndef SYS_CPLUSPLUS
|
|
#ifdef __cplusplus
|
|
/* Some system header files in older versions of gcc neglect to properly
|
|
* handle being included from C++. As it appears to be harmless to have
|
|
* multiple nested 'extern "C"' blocks, just add another one here.
|
|
*/
|
|
extern "C" {
|
|
#endif
|
|
|
|
#include <errno.h>
|
|
#include <signal.h>
|
|
#include <stdarg.h>
|
|
#include <stddef.h>
|
|
#include <stdint.h>
|
|
#include <string.h>
|
|
#include <sys/ptrace.h>
|
|
#include <sys/resource.h>
|
|
#include <sys/time.h>
|
|
#include <sys/types.h>
|
|
#include <syscall.h>
|
|
#include <unistd.h>
|
|
#include <linux/unistd.h>
|
|
#include <endian.h>
|
|
|
|
#ifdef __mips__
|
|
/* Include definitions of the ABI currently in use. */
|
|
#include <sgidefs.h>
|
|
#endif
|
|
|
|
#endif
|
|
|
|
/* As glibc often provides subtly incompatible data structures (and implicit
|
|
* wrapper functions that convert them), we provide our own kernel data
|
|
* structures for use by the system calls.
|
|
* These structures have been developed by using Linux 2.6.23 headers for
|
|
* reference. Note though, we do not care about exact API compatibility
|
|
* with the kernel, and in fact the kernel often does not have a single
|
|
* API that works across architectures. Instead, we try to mimic the glibc
|
|
* API where reasonable, and only guarantee ABI compatibility with the
|
|
* kernel headers.
|
|
* Most notably, here are a few changes that were made to the structures
|
|
* defined by kernel headers:
|
|
*
|
|
* - we only define structures, but not symbolic names for kernel data
|
|
* types. For the latter, we directly use the native C datatype
|
|
* (i.e. "unsigned" instead of "mode_t").
|
|
* - in a few cases, it is possible to define identical structures for
|
|
* both 32bit (e.g. i386) and 64bit (e.g. x86-64) platforms by
|
|
* standardizing on the 64bit version of the data types. In particular,
|
|
* this means that we use "unsigned" where the 32bit headers say
|
|
* "unsigned long".
|
|
* - overall, we try to minimize the number of cases where we need to
|
|
* conditionally define different structures.
|
|
* - the "struct kernel_sigaction" class of structures have been
|
|
* modified to more closely mimic glibc's API by introducing an
|
|
* anonymous union for the function pointer.
|
|
* - a small number of field names had to have an underscore appended to
|
|
* them, because glibc defines a global macro by the same name.
|
|
*/
|
|
|
|
/* include/linux/dirent.h */
|
|
struct kernel_dirent64 {
|
|
unsigned long long d_ino;
|
|
long long d_off;
|
|
unsigned short d_reclen;
|
|
unsigned char d_type;
|
|
char d_name[256];
|
|
};
|
|
|
|
/* include/linux/dirent.h */
|
|
struct kernel_dirent {
|
|
long d_ino;
|
|
long d_off;
|
|
unsigned short d_reclen;
|
|
char d_name[256];
|
|
};
|
|
|
|
/* include/linux/time.h */
|
|
struct kernel_timespec {
|
|
long tv_sec;
|
|
long tv_nsec;
|
|
};
|
|
|
|
/* include/linux/time.h */
|
|
struct kernel_timeval {
|
|
long tv_sec;
|
|
long tv_usec;
|
|
};
|
|
|
|
/* include/linux/resource.h */
|
|
struct kernel_rusage {
|
|
struct kernel_timeval ru_utime;
|
|
struct kernel_timeval ru_stime;
|
|
long ru_maxrss;
|
|
long ru_ixrss;
|
|
long ru_idrss;
|
|
long ru_isrss;
|
|
long ru_minflt;
|
|
long ru_majflt;
|
|
long ru_nswap;
|
|
long ru_inblock;
|
|
long ru_oublock;
|
|
long ru_msgsnd;
|
|
long ru_msgrcv;
|
|
long ru_nsignals;
|
|
long ru_nvcsw;
|
|
long ru_nivcsw;
|
|
};
|
|
|
|
#if defined(__i386__) || defined(__arm__) \
|
|
|| defined(__PPC__) || (defined(__s390__) && !defined(__s390x__))
|
|
|
|
/* include/asm-{arm,i386,mips,ppc}/signal.h */
|
|
struct kernel_old_sigaction {
|
|
union {
|
|
void (*sa_handler_)(int);
|
|
void (*sa_sigaction_)(int, siginfo_t *, void *);
|
|
};
|
|
unsigned long sa_mask;
|
|
unsigned long sa_flags;
|
|
void (*sa_restorer)(void);
|
|
} __attribute__((packed,aligned(4)));
|
|
#elif (defined(__mips__) && _MIPS_SIM == _MIPS_SIM_ABI32)
|
|
#define kernel_old_sigaction kernel_sigaction
|
|
#elif defined(__aarch64__)
|
|
// No kernel_old_sigaction defined for arm64.
|
|
#endif
|
|
|
|
/* Some kernel functions (e.g. sigaction() in 2.6.23) require that the
|
|
* exactly match the size of the signal set, even though the API was
|
|
* intended to be extensible. We define our own KERNEL_NSIG to deal with
|
|
* this.
|
|
* Please note that glibc provides signals [1.._NSIG-1], whereas the
|
|
* kernel (and this header) provides the range [1..KERNEL_NSIG]. The
|
|
* actual number of signals is obviously the same, but the constants
|
|
* differ by one.
|
|
*/
|
|
#ifdef __mips__
|
|
#define KERNEL_NSIG 128
|
|
#else
|
|
#define KERNEL_NSIG 64
|
|
#endif
|
|
|
|
/* include/asm-{arm,i386,mips,x86_64}/signal.h */
|
|
struct kernel_sigset_t {
|
|
unsigned long sig[(KERNEL_NSIG + 8*sizeof(unsigned long) - 1)/
|
|
(8*sizeof(unsigned long))];
|
|
};
|
|
|
|
/* include/asm-{arm,generic,i386,mips,x86_64,ppc}/signal.h */
|
|
struct kernel_sigaction {
|
|
#ifdef __mips__
|
|
unsigned long sa_flags;
|
|
union {
|
|
void (*sa_handler_)(int);
|
|
void (*sa_sigaction_)(int, siginfo_t *, void *);
|
|
};
|
|
struct kernel_sigset_t sa_mask;
|
|
#else
|
|
union {
|
|
void (*sa_handler_)(int);
|
|
void (*sa_sigaction_)(int, siginfo_t *, void *);
|
|
};
|
|
unsigned long sa_flags;
|
|
void (*sa_restorer)(void);
|
|
struct kernel_sigset_t sa_mask;
|
|
#endif
|
|
};
|
|
|
|
/* include/asm-{arm,i386,mips,ppc,s390}/stat.h */
|
|
#ifdef __mips__
|
|
#if _MIPS_SIM == _MIPS_SIM_ABI64
|
|
struct kernel_stat {
|
|
#else
|
|
struct kernel_stat64 {
|
|
#endif
|
|
unsigned st_dev;
|
|
unsigned __pad0[3];
|
|
unsigned long long st_ino;
|
|
unsigned st_mode;
|
|
unsigned st_nlink;
|
|
unsigned st_uid;
|
|
unsigned st_gid;
|
|
unsigned st_rdev;
|
|
unsigned __pad1[3];
|
|
long long st_size;
|
|
unsigned st_atime_;
|
|
unsigned st_atime_nsec_;
|
|
unsigned st_mtime_;
|
|
unsigned st_mtime_nsec_;
|
|
unsigned st_ctime_;
|
|
unsigned st_ctime_nsec_;
|
|
unsigned st_blksize;
|
|
unsigned __pad2;
|
|
unsigned long long st_blocks;
|
|
};
|
|
#elif defined __PPC__
|
|
struct kernel_stat64 {
|
|
unsigned long long st_dev;
|
|
unsigned long long st_ino;
|
|
unsigned st_nlink;
|
|
unsigned st_mode;
|
|
unsigned st_uid;
|
|
unsigned st_gid;
|
|
int __pad2;
|
|
unsigned long long st_rdev;
|
|
long long st_size;
|
|
long long st_blksize;
|
|
long long st_blocks;
|
|
kernel_timespec st_atim;
|
|
kernel_timespec st_mtim;
|
|
kernel_timespec st_ctim;
|
|
unsigned long __unused4;
|
|
unsigned long __unused5;
|
|
unsigned long __unused6;
|
|
};
|
|
#else
|
|
struct kernel_stat64 {
|
|
unsigned long long st_dev;
|
|
unsigned char __pad0[4];
|
|
unsigned __st_ino;
|
|
unsigned st_mode;
|
|
unsigned st_nlink;
|
|
unsigned st_uid;
|
|
unsigned st_gid;
|
|
unsigned long long st_rdev;
|
|
unsigned char __pad3[4];
|
|
long long st_size;
|
|
unsigned st_blksize;
|
|
unsigned long long st_blocks;
|
|
unsigned st_atime_;
|
|
unsigned st_atime_nsec_;
|
|
unsigned st_mtime_;
|
|
unsigned st_mtime_nsec_;
|
|
unsigned st_ctime_;
|
|
unsigned st_ctime_nsec_;
|
|
unsigned long long st_ino;
|
|
};
|
|
#endif
|
|
|
|
/* include/asm-{arm,generic,i386,mips,x86_64,ppc,s390}/stat.h */
|
|
#if defined(__i386__) || defined(__arm__)
|
|
struct kernel_stat {
|
|
/* The kernel headers suggest that st_dev and st_rdev should be 32bit
|
|
* quantities encoding 12bit major and 20bit minor numbers in an interleaved
|
|
* format. In reality, we do not see useful data in the top bits. So,
|
|
* we'll leave the padding in here, until we find a better solution.
|
|
*/
|
|
unsigned short st_dev;
|
|
short pad1;
|
|
unsigned st_ino;
|
|
unsigned short st_mode;
|
|
unsigned short st_nlink;
|
|
unsigned short st_uid;
|
|
unsigned short st_gid;
|
|
unsigned short st_rdev;
|
|
short pad2;
|
|
unsigned st_size;
|
|
unsigned st_blksize;
|
|
unsigned st_blocks;
|
|
unsigned st_atime_;
|
|
unsigned st_atime_nsec_;
|
|
unsigned st_mtime_;
|
|
unsigned st_mtime_nsec_;
|
|
unsigned st_ctime_;
|
|
unsigned st_ctime_nsec_;
|
|
unsigned __unused4;
|
|
unsigned __unused5;
|
|
};
|
|
#elif defined(__x86_64__)
|
|
struct kernel_stat {
|
|
uint64_t st_dev;
|
|
uint64_t st_ino;
|
|
uint64_t st_nlink;
|
|
unsigned st_mode;
|
|
unsigned st_uid;
|
|
unsigned st_gid;
|
|
unsigned __pad0;
|
|
uint64_t st_rdev;
|
|
int64_t st_size;
|
|
int64_t st_blksize;
|
|
int64_t st_blocks;
|
|
uint64_t st_atime_;
|
|
uint64_t st_atime_nsec_;
|
|
uint64_t st_mtime_;
|
|
uint64_t st_mtime_nsec_;
|
|
uint64_t st_ctime_;
|
|
uint64_t st_ctime_nsec_;
|
|
int64_t __unused[3];
|
|
};
|
|
#elif defined(__PPC__)
|
|
struct kernel_stat {
|
|
unsigned long long st_dev;
|
|
unsigned long st_ino;
|
|
unsigned long st_nlink;
|
|
unsigned long st_mode;
|
|
unsigned st_uid;
|
|
unsigned st_gid;
|
|
int __pad2;
|
|
unsigned long long st_rdev;
|
|
long st_size;
|
|
unsigned long st_blksize;
|
|
unsigned long st_blocks;
|
|
kernel_timespec st_atim;
|
|
kernel_timespec st_mtim;
|
|
kernel_timespec st_ctim;
|
|
unsigned long __unused4;
|
|
unsigned long __unused5;
|
|
unsigned long __unused6;
|
|
};
|
|
#elif (defined(__mips__) && _MIPS_SIM != _MIPS_SIM_ABI64)
|
|
struct kernel_stat {
|
|
unsigned st_dev;
|
|
int st_pad1[3];
|
|
unsigned st_ino;
|
|
unsigned st_mode;
|
|
unsigned st_nlink;
|
|
unsigned st_uid;
|
|
unsigned st_gid;
|
|
unsigned st_rdev;
|
|
int st_pad2[2];
|
|
long st_size;
|
|
int st_pad3;
|
|
long st_atime_;
|
|
long st_atime_nsec_;
|
|
long st_mtime_;
|
|
long st_mtime_nsec_;
|
|
long st_ctime_;
|
|
long st_ctime_nsec_;
|
|
int st_blksize;
|
|
int st_blocks;
|
|
int st_pad4[14];
|
|
};
|
|
#elif defined(__aarch64__)
|
|
struct kernel_stat {
|
|
unsigned long st_dev;
|
|
unsigned long st_ino;
|
|
unsigned int st_mode;
|
|
unsigned int st_nlink;
|
|
unsigned int st_uid;
|
|
unsigned int st_gid;
|
|
unsigned long st_rdev;
|
|
unsigned long __pad1;
|
|
long st_size;
|
|
int st_blksize;
|
|
int __pad2;
|
|
long st_blocks;
|
|
long st_atime_;
|
|
unsigned long st_atime_nsec_;
|
|
long st_mtime_;
|
|
unsigned long st_mtime_nsec_;
|
|
long st_ctime_;
|
|
unsigned long st_ctime_nsec_;
|
|
unsigned int __unused4;
|
|
unsigned int __unused5;
|
|
};
|
|
#elif defined(__s390x__)
|
|
struct kernel_stat {
|
|
unsigned long st_dev;
|
|
unsigned long st_ino;
|
|
unsigned long st_nlink;
|
|
unsigned int st_mode;
|
|
unsigned int st_uid;
|
|
unsigned int st_gid;
|
|
unsigned int __pad1;
|
|
unsigned long st_rdev;
|
|
unsigned long st_size;
|
|
unsigned long st_atime_;
|
|
unsigned long st_atime_nsec_;
|
|
unsigned long st_mtime_;
|
|
unsigned long st_mtime_nsec_;
|
|
unsigned long st_ctime_;
|
|
unsigned long st_ctime_nsec_;
|
|
unsigned long st_blksize;
|
|
long st_blocks;
|
|
unsigned long __unused[3];
|
|
};
|
|
#elif defined(__s390__)
|
|
struct kernel_stat {
|
|
unsigned short st_dev;
|
|
unsigned short __pad1;
|
|
unsigned long st_ino;
|
|
unsigned short st_mode;
|
|
unsigned short st_nlink;
|
|
unsigned short st_uid;
|
|
unsigned short st_gid;
|
|
unsigned short st_rdev;
|
|
unsigned short __pad2;
|
|
unsigned long st_size;
|
|
unsigned long st_blksize;
|
|
unsigned long st_blocks;
|
|
unsigned long st_atime_;
|
|
unsigned long st_atime_nsec_;
|
|
unsigned long st_mtime_;
|
|
unsigned long st_mtime_nsec_;
|
|
unsigned long st_ctime_;
|
|
unsigned long st_ctime_nsec_;
|
|
unsigned long __unused4;
|
|
unsigned long __unused5;
|
|
};
|
|
#endif
|
|
|
|
|
|
/* Definitions missing from the standard header files */
|
|
#ifndef O_DIRECTORY
|
|
#if defined(__arm__)
|
|
#define O_DIRECTORY 0040000
|
|
#else
|
|
#define O_DIRECTORY 0200000
|
|
#endif
|
|
#endif
|
|
#ifndef PR_GET_DUMPABLE
|
|
#define PR_GET_DUMPABLE 3
|
|
#endif
|
|
#ifndef PR_SET_DUMPABLE
|
|
#define PR_SET_DUMPABLE 4
|
|
#endif
|
|
#ifndef AT_FDCWD
|
|
#define AT_FDCWD (-100)
|
|
#endif
|
|
#ifndef AT_SYMLINK_NOFOLLOW
|
|
#define AT_SYMLINK_NOFOLLOW 0x100
|
|
#endif
|
|
#ifndef AT_REMOVEDIR
|
|
#define AT_REMOVEDIR 0x200
|
|
#endif
|
|
#ifndef MREMAP_FIXED
|
|
#define MREMAP_FIXED 2
|
|
#endif
|
|
#ifndef SA_RESTORER
|
|
#define SA_RESTORER 0x04000000
|
|
#endif
|
|
|
|
#if defined(__i386__)
|
|
#ifndef __NR_rt_sigaction
|
|
#define __NR_rt_sigaction 174
|
|
#define __NR_rt_sigprocmask 175
|
|
#endif
|
|
#ifndef __NR_stat64
|
|
#define __NR_stat64 195
|
|
#endif
|
|
#ifndef __NR_fstat64
|
|
#define __NR_fstat64 197
|
|
#endif
|
|
#ifndef __NR_getdents64
|
|
#define __NR_getdents64 220
|
|
#endif
|
|
#ifndef __NR_gettid
|
|
#define __NR_gettid 224
|
|
#endif
|
|
#ifndef __NR_futex
|
|
#define __NR_futex 240
|
|
#endif
|
|
#ifndef __NR_openat
|
|
#define __NR_openat 295
|
|
#endif
|
|
#ifndef __NR_getcpu
|
|
#define __NR_getcpu 318
|
|
#endif
|
|
/* End of i386 definitions */
|
|
#elif defined(__arm__)
|
|
#ifndef __syscall
|
|
#if defined(__thumb__) || defined(__ARM_EABI__)
|
|
#define __SYS_REG(name) register long __sysreg __asm__("r6") = __NR_##name;
|
|
#define __SYS_REG_LIST(regs...) [sysreg] "r" (__sysreg) , ##regs
|
|
#define __syscall(name) "swi\t0"
|
|
#define __syscall_safe(name) \
|
|
"push {r7}\n" \
|
|
"mov r7,%[sysreg]\n" \
|
|
__syscall(name)"\n" \
|
|
"pop {r7}"
|
|
#else
|
|
#define __SYS_REG(name)
|
|
#define __SYS_REG_LIST(regs...) regs
|
|
#define __syscall(name) "swi\t" __sys1(__NR_##name) ""
|
|
#define __syscall_safe(name) __syscall(name)
|
|
#endif
|
|
#endif
|
|
#ifndef __NR_rt_sigaction
|
|
#define __NR_rt_sigaction (__NR_SYSCALL_BASE + 174)
|
|
#define __NR_rt_sigprocmask (__NR_SYSCALL_BASE + 175)
|
|
#endif
|
|
#ifndef __NR_stat64
|
|
#define __NR_stat64 (__NR_SYSCALL_BASE + 195)
|
|
#endif
|
|
#ifndef __NR_fstat64
|
|
#define __NR_fstat64 (__NR_SYSCALL_BASE + 197)
|
|
#endif
|
|
#ifndef __NR_getdents64
|
|
#define __NR_getdents64 (__NR_SYSCALL_BASE + 217)
|
|
#endif
|
|
#ifndef __NR_gettid
|
|
#define __NR_gettid (__NR_SYSCALL_BASE + 224)
|
|
#endif
|
|
#ifndef __NR_futex
|
|
#define __NR_futex (__NR_SYSCALL_BASE + 240)
|
|
#endif
|
|
/* End of ARM definitions */
|
|
#elif defined(__x86_64__)
|
|
#ifndef __NR_gettid
|
|
#define __NR_gettid 186
|
|
#endif
|
|
#ifndef __NR_futex
|
|
#define __NR_futex 202
|
|
#endif
|
|
#ifndef __NR_getdents64
|
|
#define __NR_getdents64 217
|
|
#endif
|
|
#ifndef __NR_openat
|
|
#define __NR_openat 257
|
|
#endif
|
|
/* End of x86-64 definitions */
|
|
#elif defined(__mips__)
|
|
#if _MIPS_SIM == _MIPS_SIM_ABI32
|
|
#ifndef __NR_rt_sigaction
|
|
#define __NR_rt_sigaction (__NR_Linux + 194)
|
|
#define __NR_rt_sigprocmask (__NR_Linux + 195)
|
|
#endif
|
|
#ifndef __NR_stat64
|
|
#define __NR_stat64 (__NR_Linux + 213)
|
|
#endif
|
|
#ifndef __NR_fstat64
|
|
#define __NR_fstat64 (__NR_Linux + 215)
|
|
#endif
|
|
#ifndef __NR_getdents64
|
|
#define __NR_getdents64 (__NR_Linux + 219)
|
|
#endif
|
|
#ifndef __NR_gettid
|
|
#define __NR_gettid (__NR_Linux + 222)
|
|
#endif
|
|
#ifndef __NR_futex
|
|
#define __NR_futex (__NR_Linux + 238)
|
|
#endif
|
|
#ifndef __NR_openat
|
|
#define __NR_openat (__NR_Linux + 288)
|
|
#endif
|
|
#ifndef __NR_fstatat
|
|
#define __NR_fstatat (__NR_Linux + 293)
|
|
#endif
|
|
#ifndef __NR_getcpu
|
|
#define __NR_getcpu (__NR_Linux + 312)
|
|
#endif
|
|
/* End of MIPS (old 32bit API) definitions */
|
|
#elif _MIPS_SIM == _MIPS_SIM_ABI64
|
|
#ifndef __NR_gettid
|
|
#define __NR_gettid (__NR_Linux + 178)
|
|
#endif
|
|
#ifndef __NR_futex
|
|
#define __NR_futex (__NR_Linux + 194)
|
|
#endif
|
|
#ifndef __NR_openat
|
|
#define __NR_openat (__NR_Linux + 247)
|
|
#endif
|
|
#ifndef __NR_fstatat
|
|
#define __NR_fstatat (__NR_Linux + 252)
|
|
#endif
|
|
#ifndef __NR_getcpu
|
|
#define __NR_getcpu (__NR_Linux + 271)
|
|
#endif
|
|
/* End of MIPS (64bit API) definitions */
|
|
#else
|
|
#ifndef __NR_gettid
|
|
#define __NR_gettid (__NR_Linux + 178)
|
|
#endif
|
|
#ifndef __NR_futex
|
|
#define __NR_futex (__NR_Linux + 194)
|
|
#endif
|
|
#ifndef __NR_openat
|
|
#define __NR_openat (__NR_Linux + 251)
|
|
#endif
|
|
#ifndef __NR_fstatat
|
|
#define __NR_fstatat (__NR_Linux + 256)
|
|
#endif
|
|
#ifndef __NR_getcpu
|
|
#define __NR_getcpu (__NR_Linux + 275)
|
|
#endif
|
|
/* End of MIPS (new 32bit API) definitions */
|
|
#endif
|
|
/* End of MIPS definitions */
|
|
#elif defined(__PPC__)
|
|
#ifndef __NR_rt_sigaction
|
|
#define __NR_rt_sigaction 173
|
|
#define __NR_rt_sigprocmask 174
|
|
#endif
|
|
#ifndef __NR_stat64
|
|
#define __NR_stat64 195
|
|
#endif
|
|
#ifndef __NR_fstat64
|
|
#define __NR_fstat64 197
|
|
#endif
|
|
#ifndef __NR_socket
|
|
#define __NR_socket 198
|
|
#endif
|
|
#ifndef __NR_getdents64
|
|
#define __NR_getdents64 202
|
|
#endif
|
|
#ifndef __NR_gettid
|
|
#define __NR_gettid 207
|
|
#endif
|
|
#ifndef __NR_futex
|
|
#define __NR_futex 221
|
|
#endif
|
|
#ifndef __NR_openat
|
|
#define __NR_openat 286
|
|
#endif
|
|
#ifndef __NR_getcpu
|
|
#define __NR_getcpu 302
|
|
#endif
|
|
/* End of powerpc defininitions */
|
|
#elif defined(__aarch64__)
|
|
#ifndef __NR_fstatat
|
|
#define __NR_fstatat 79
|
|
#endif
|
|
/* End of aarch64 defininitions */
|
|
#elif defined(__s390__)
|
|
#ifndef __NR_quotactl
|
|
#define __NR_quotactl 131
|
|
#endif
|
|
#ifndef __NR_rt_sigreturn
|
|
#define __NR_rt_sigreturn 173
|
|
#endif
|
|
#ifndef __NR_rt_sigaction
|
|
#define __NR_rt_sigaction 174
|
|
#endif
|
|
#ifndef __NR_rt_sigprocmask
|
|
#define __NR_rt_sigprocmask 175
|
|
#endif
|
|
#ifndef __NR_rt_sigpending
|
|
#define __NR_rt_sigpending 176
|
|
#endif
|
|
#ifndef __NR_rt_sigsuspend
|
|
#define __NR_rt_sigsuspend 179
|
|
#endif
|
|
#ifndef __NR_pread64
|
|
#define __NR_pread64 180
|
|
#endif
|
|
#ifndef __NR_pwrite64
|
|
#define __NR_pwrite64 181
|
|
#endif
|
|
#ifndef __NR_getdents64
|
|
#define __NR_getdents64 220
|
|
#endif
|
|
#ifndef __NR_readahead
|
|
#define __NR_readahead 222
|
|
#endif
|
|
#ifndef __NR_setxattr
|
|
#define __NR_setxattr 224
|
|
#endif
|
|
#ifndef __NR_lsetxattr
|
|
#define __NR_lsetxattr 225
|
|
#endif
|
|
#ifndef __NR_getxattr
|
|
#define __NR_getxattr 227
|
|
#endif
|
|
#ifndef __NR_lgetxattr
|
|
#define __NR_lgetxattr 228
|
|
#endif
|
|
#ifndef __NR_listxattr
|
|
#define __NR_listxattr 230
|
|
#endif
|
|
#ifndef __NR_llistxattr
|
|
#define __NR_llistxattr 231
|
|
#endif
|
|
#ifndef __NR_gettid
|
|
#define __NR_gettid 236
|
|
#endif
|
|
#ifndef __NR_tkill
|
|
#define __NR_tkill 237
|
|
#endif
|
|
#ifndef __NR_futex
|
|
#define __NR_futex 238
|
|
#endif
|
|
#ifndef __NR_sched_setaffinity
|
|
#define __NR_sched_setaffinity 239
|
|
#endif
|
|
#ifndef __NR_sched_getaffinity
|
|
#define __NR_sched_getaffinity 240
|
|
#endif
|
|
#ifndef __NR_set_tid_address
|
|
#define __NR_set_tid_address 252
|
|
#endif
|
|
#ifndef __NR_clock_gettime
|
|
#define __NR_clock_gettime 260
|
|
#endif
|
|
#ifndef __NR_clock_getres
|
|
#define __NR_clock_getres 261
|
|
#endif
|
|
#ifndef __NR_statfs64
|
|
#define __NR_statfs64 265
|
|
#endif
|
|
#ifndef __NR_fstatfs64
|
|
#define __NR_fstatfs64 266
|
|
#endif
|
|
#ifndef __NR_ioprio_set
|
|
#define __NR_ioprio_set 282
|
|
#endif
|
|
#ifndef __NR_ioprio_get
|
|
#define __NR_ioprio_get 283
|
|
#endif
|
|
#ifndef __NR_openat
|
|
#define __NR_openat 288
|
|
#endif
|
|
#ifndef __NR_unlinkat
|
|
#define __NR_unlinkat 294
|
|
#endif
|
|
#ifndef __NR_move_pages
|
|
#define __NR_move_pages 310
|
|
#endif
|
|
#ifndef __NR_getcpu
|
|
#define __NR_getcpu 311
|
|
#endif
|
|
#ifndef __NR_fallocate
|
|
#define __NR_fallocate 314
|
|
#endif
|
|
/* Some syscalls are named/numbered differently between s390 and s390x. */
|
|
#ifdef __s390x__
|
|
# ifndef __NR_getrlimit
|
|
# define __NR_getrlimit 191
|
|
# endif
|
|
# ifndef __NR_setresuid
|
|
# define __NR_setresuid 208
|
|
# endif
|
|
# ifndef __NR_getresuid
|
|
# define __NR_getresuid 209
|
|
# endif
|
|
# ifndef __NR_setresgid
|
|
# define __NR_setresgid 210
|
|
# endif
|
|
# ifndef __NR_getresgid
|
|
# define __NR_getresgid 211
|
|
# endif
|
|
# ifndef __NR_setfsuid
|
|
# define __NR_setfsuid 215
|
|
# endif
|
|
# ifndef __NR_setfsgid
|
|
# define __NR_setfsgid 216
|
|
# endif
|
|
# ifndef __NR_fadvise64
|
|
# define __NR_fadvise64 253
|
|
# endif
|
|
# ifndef __NR_newfstatat
|
|
# define __NR_newfstatat 293
|
|
# endif
|
|
#else /* __s390x__ */
|
|
# ifndef __NR_getrlimit
|
|
# define __NR_getrlimit 76
|
|
# endif
|
|
# ifndef __NR_setfsuid
|
|
# define __NR_setfsuid 138
|
|
# endif
|
|
# ifndef __NR_setfsgid
|
|
# define __NR_setfsgid 139
|
|
# endif
|
|
# ifndef __NR_setresuid
|
|
# define __NR_setresuid 164
|
|
# endif
|
|
# ifndef __NR_getresuid
|
|
# define __NR_getresuid 165
|
|
# endif
|
|
# ifndef __NR_setresgid
|
|
# define __NR_setresgid 170
|
|
# endif
|
|
# ifndef __NR_getresgid
|
|
# define __NR_getresgid 171
|
|
# endif
|
|
# ifndef __NR_ugetrlimit
|
|
# define __NR_ugetrlimit 191
|
|
# endif
|
|
# ifndef __NR_mmap2
|
|
# define __NR_mmap2 192
|
|
# endif
|
|
# ifndef __NR_setresuid32
|
|
# define __NR_setresuid32 208
|
|
# endif
|
|
# ifndef __NR_getresuid32
|
|
# define __NR_getresuid32 209
|
|
# endif
|
|
# ifndef __NR_setresgid32
|
|
# define __NR_setresgid32 210
|
|
# endif
|
|
# ifndef __NR_getresgid32
|
|
# define __NR_getresgid32 211
|
|
# endif
|
|
# ifndef __NR_setfsuid32
|
|
# define __NR_setfsuid32 215
|
|
# endif
|
|
# ifndef __NR_setfsgid32
|
|
# define __NR_setfsgid32 216
|
|
# endif
|
|
# ifndef __NR_fadvise64_64
|
|
# define __NR_fadvise64_64 264
|
|
# endif
|
|
# ifndef __NR_fstatat64
|
|
# define __NR_fstatat64 293
|
|
# endif
|
|
#endif /* __s390__ */
|
|
/* End of s390/s390x definitions */
|
|
#endif
|
|
|
|
|
|
/* After forking, we must make sure to only call system calls. */
|
|
#if __BOUNDED_POINTERS__
|
|
#error "Need to port invocations of syscalls for bounded ptrs"
|
|
#else
|
|
/* The core dumper and the thread lister get executed after threads
|
|
* have been suspended. As a consequence, we cannot call any functions
|
|
* that acquire locks. Unfortunately, libc wraps most system calls
|
|
* (e.g. in order to implement pthread_atfork, and to make calls
|
|
* cancellable), which means we cannot call these functions. Instead,
|
|
* we have to call syscall() directly.
|
|
*/
|
|
#undef LSS_ERRNO
|
|
#ifdef SYS_ERRNO
|
|
/* Allow the including file to override the location of errno. This can
|
|
* be useful when using clone() with the CLONE_VM option.
|
|
*/
|
|
#define LSS_ERRNO SYS_ERRNO
|
|
#else
|
|
#define LSS_ERRNO errno
|
|
#endif
|
|
|
|
#undef LSS_INLINE
|
|
#ifdef SYS_INLINE
|
|
#define LSS_INLINE SYS_INLINE
|
|
#else
|
|
#define LSS_INLINE static inline
|
|
#endif
|
|
|
|
/* Allow the including file to override the prefix used for all new
|
|
* system calls. By default, it will be set to "sys_".
|
|
*/
|
|
#undef LSS_NAME
|
|
#ifndef SYS_PREFIX
|
|
#define LSS_NAME(name) sys_##name
|
|
#elif SYS_PREFIX < 0
|
|
#define LSS_NAME(name) name
|
|
#elif SYS_PREFIX == 0
|
|
#define LSS_NAME(name) sys0_##name
|
|
#elif SYS_PREFIX == 1
|
|
#define LSS_NAME(name) sys1_##name
|
|
#elif SYS_PREFIX == 2
|
|
#define LSS_NAME(name) sys2_##name
|
|
#elif SYS_PREFIX == 3
|
|
#define LSS_NAME(name) sys3_##name
|
|
#elif SYS_PREFIX == 4
|
|
#define LSS_NAME(name) sys4_##name
|
|
#elif SYS_PREFIX == 5
|
|
#define LSS_NAME(name) sys5_##name
|
|
#elif SYS_PREFIX == 6
|
|
#define LSS_NAME(name) sys6_##name
|
|
#elif SYS_PREFIX == 7
|
|
#define LSS_NAME(name) sys7_##name
|
|
#elif SYS_PREFIX == 8
|
|
#define LSS_NAME(name) sys8_##name
|
|
#elif SYS_PREFIX == 9
|
|
#define LSS_NAME(name) sys9_##name
|
|
#endif
|
|
|
|
#undef LSS_RETURN
|
|
#if (defined(__i386__) || defined(__x86_64__) || defined(__arm__) || \
|
|
defined(__aarch64__) || defined(__s390__))
|
|
/* Failing system calls return a negative result in the range of
|
|
* -1..-4095. These are "errno" values with the sign inverted.
|
|
*/
|
|
#define LSS_RETURN(type, res) \
|
|
do { \
|
|
if ((unsigned long)(res) >= (unsigned long)(-4095)) { \
|
|
LSS_ERRNO = -(res); \
|
|
res = -1; \
|
|
} \
|
|
return (type) (res); \
|
|
} while (0)
|
|
#elif defined(__mips__)
|
|
/* On MIPS, failing system calls return -1, and set errno in a
|
|
* separate CPU register.
|
|
*/
|
|
#define LSS_RETURN(type, res, err) \
|
|
do { \
|
|
if (err) { \
|
|
LSS_ERRNO = (res); \
|
|
res = -1; \
|
|
} \
|
|
return (type) (res); \
|
|
} while (0)
|
|
#elif defined(__PPC__)
|
|
/* On PPC, failing system calls return -1, and set errno in a
|
|
* separate CPU register. See linux/unistd.h.
|
|
*/
|
|
#define LSS_RETURN(type, res, err) \
|
|
do { \
|
|
if (err & 0x10000000 ) { \
|
|
LSS_ERRNO = (res); \
|
|
res = -1; \
|
|
} \
|
|
return (type) (res); \
|
|
} while (0)
|
|
#endif
|
|
#if defined(__i386__)
|
|
#if defined(NO_FRAME_POINTER) && (100 * __GNUC__ + __GNUC_MINOR__ >= 404)
|
|
/* This only works for GCC-4.4 and above -- the first version to use
|
|
.cfi directives for dwarf unwind info. */
|
|
#define CFI_ADJUST_CFA_OFFSET(adjust) \
|
|
".cfi_adjust_cfa_offset " #adjust "\n"
|
|
#else
|
|
#define CFI_ADJUST_CFA_OFFSET(adjust) /**/
|
|
#endif
|
|
|
|
/* In PIC mode (e.g. when building shared libraries), gcc for i386
|
|
* reserves ebx. Unfortunately, most distribution ship with implementations
|
|
* of _syscallX() which clobber ebx.
|
|
* Also, most definitions of _syscallX() neglect to mark "memory" as being
|
|
* clobbered. This causes problems with compilers, that do a better job
|
|
* at optimizing across __asm__ calls.
|
|
* So, we just have to redefine all of the _syscallX() macros.
|
|
*/
|
|
#undef LSS_BODY
|
|
#define LSS_BODY(type,args...) \
|
|
long __res; \
|
|
__asm__ __volatile__("push %%ebx\n" \
|
|
CFI_ADJUST_CFA_OFFSET(4) \
|
|
"movl %2,%%ebx\n" \
|
|
"int $0x80\n" \
|
|
"pop %%ebx\n" \
|
|
CFI_ADJUST_CFA_OFFSET(-4) \
|
|
args \
|
|
: "esp", "memory"); \
|
|
LSS_RETURN(type,__res)
|
|
#undef _syscall0
|
|
#define _syscall0(type,name) \
|
|
type LSS_NAME(name)(void) { \
|
|
long __res; \
|
|
__asm__ volatile("int $0x80" \
|
|
: "=a" (__res) \
|
|
: "0" (__NR_##name) \
|
|
: "memory"); \
|
|
LSS_RETURN(type,__res); \
|
|
}
|
|
#undef _syscall1
|
|
#define _syscall1(type,name,type1,arg1) \
|
|
type LSS_NAME(name)(type1 arg1) { \
|
|
LSS_BODY(type, \
|
|
: "=a" (__res) \
|
|
: "0" (__NR_##name), "ri" ((long)(arg1))); \
|
|
}
|
|
#undef _syscall2
|
|
#define _syscall2(type,name,type1,arg1,type2,arg2) \
|
|
type LSS_NAME(name)(type1 arg1,type2 arg2) { \
|
|
LSS_BODY(type, \
|
|
: "=a" (__res) \
|
|
: "0" (__NR_##name),"ri" ((long)(arg1)), "c" ((long)(arg2))); \
|
|
}
|
|
#undef _syscall3
|
|
#define _syscall3(type,name,type1,arg1,type2,arg2,type3,arg3) \
|
|
type LSS_NAME(name)(type1 arg1,type2 arg2,type3 arg3) { \
|
|
LSS_BODY(type, \
|
|
: "=a" (__res) \
|
|
: "0" (__NR_##name), "ri" ((long)(arg1)), "c" ((long)(arg2)), \
|
|
"d" ((long)(arg3))); \
|
|
}
|
|
#undef _syscall4
|
|
#define _syscall4(type,name,type1,arg1,type2,arg2,type3,arg3,type4,arg4) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2, type3 arg3, type4 arg4) { \
|
|
LSS_BODY(type, \
|
|
: "=a" (__res) \
|
|
: "0" (__NR_##name), "ri" ((long)(arg1)), "c" ((long)(arg2)), \
|
|
"d" ((long)(arg3)),"S" ((long)(arg4))); \
|
|
}
|
|
#undef _syscall5
|
|
#define _syscall5(type,name,type1,arg1,type2,arg2,type3,arg3,type4,arg4, \
|
|
type5,arg5) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2, type3 arg3, type4 arg4, \
|
|
type5 arg5) { \
|
|
long __res; \
|
|
__asm__ __volatile__("push %%ebx\n" \
|
|
"movl %2,%%ebx\n" \
|
|
"movl %1,%%eax\n" \
|
|
"int $0x80\n" \
|
|
"pop %%ebx" \
|
|
: "=a" (__res) \
|
|
: "i" (__NR_##name), "ri" ((long)(arg1)), \
|
|
"c" ((long)(arg2)), "d" ((long)(arg3)), \
|
|
"S" ((long)(arg4)), "D" ((long)(arg5)) \
|
|
: "esp", "memory"); \
|
|
LSS_RETURN(type,__res); \
|
|
}
|
|
#undef _syscall6
|
|
#define _syscall6(type,name,type1,arg1,type2,arg2,type3,arg3,type4,arg4, \
|
|
type5,arg5,type6,arg6) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2, type3 arg3, type4 arg4, \
|
|
type5 arg5, type6 arg6) { \
|
|
long __res; \
|
|
struct { long __a1; long __a6; } __s = { (long)arg1, (long) arg6 }; \
|
|
__asm__ __volatile__("push %%ebp\n" \
|
|
"push %%ebx\n" \
|
|
"movl 4(%2),%%ebp\n" \
|
|
"movl 0(%2), %%ebx\n" \
|
|
"movl %1,%%eax\n" \
|
|
"int $0x80\n" \
|
|
"pop %%ebx\n" \
|
|
"pop %%ebp" \
|
|
: "=a" (__res) \
|
|
: "i" (__NR_##name), "0" ((long)(&__s)), \
|
|
"c" ((long)(arg2)), "d" ((long)(arg3)), \
|
|
"S" ((long)(arg4)), "D" ((long)(arg5)) \
|
|
: "esp", "memory"); \
|
|
LSS_RETURN(type,__res); \
|
|
}
|
|
LSS_INLINE int LSS_NAME(clone)(int (*fn)(void *), void *child_stack,
|
|
int flags, void *arg, int *parent_tidptr,
|
|
void *newtls, int *child_tidptr) {
|
|
long __res;
|
|
__asm__ __volatile__(/* if (fn == NULL)
|
|
* return -EINVAL;
|
|
*/
|
|
"movl %3,%%ecx\n"
|
|
"jecxz 1f\n"
|
|
|
|
/* if (child_stack == NULL)
|
|
* return -EINVAL;
|
|
*/
|
|
"movl %4,%%ecx\n"
|
|
"jecxz 1f\n"
|
|
|
|
/* Set up alignment of the child stack:
|
|
* child_stack = (child_stack & ~0xF) - 20;
|
|
*/
|
|
"andl $-16,%%ecx\n"
|
|
"subl $20,%%ecx\n"
|
|
|
|
/* Push "arg" and "fn" onto the stack that will be
|
|
* used by the child.
|
|
*/
|
|
"movl %6,%%eax\n"
|
|
"movl %%eax,4(%%ecx)\n"
|
|
"movl %3,%%eax\n"
|
|
"movl %%eax,(%%ecx)\n"
|
|
|
|
/* %eax = syscall(%eax = __NR_clone,
|
|
* %ebx = flags,
|
|
* %ecx = child_stack,
|
|
* %edx = parent_tidptr,
|
|
* %esi = newtls,
|
|
* %edi = child_tidptr)
|
|
* Also, make sure that %ebx gets preserved as it is
|
|
* used in PIC mode.
|
|
*/
|
|
"movl %8,%%esi\n"
|
|
"movl %7,%%edx\n"
|
|
"movl %5,%%eax\n"
|
|
"movl %9,%%edi\n"
|
|
"pushl %%ebx\n"
|
|
"movl %%eax,%%ebx\n"
|
|
"movl %2,%%eax\n"
|
|
"int $0x80\n"
|
|
|
|
/* In the parent: restore %ebx
|
|
* In the child: move "fn" into %ebx
|
|
*/
|
|
"popl %%ebx\n"
|
|
|
|
/* if (%eax != 0)
|
|
* return %eax;
|
|
*/
|
|
"test %%eax,%%eax\n"
|
|
"jnz 1f\n"
|
|
|
|
/* In the child, now. Terminate frame pointer chain.
|
|
*/
|
|
"movl $0,%%ebp\n"
|
|
|
|
/* Call "fn". "arg" is already on the stack.
|
|
*/
|
|
"call *%%ebx\n"
|
|
|
|
/* Call _exit(%ebx). Unfortunately older versions
|
|
* of gcc restrict the number of arguments that can
|
|
* be passed to asm(). So, we need to hard-code the
|
|
* system call number.
|
|
*/
|
|
"movl %%eax,%%ebx\n"
|
|
"movl $1,%%eax\n"
|
|
"int $0x80\n"
|
|
|
|
/* Return to parent.
|
|
*/
|
|
"1:\n"
|
|
: "=a" (__res)
|
|
: "0"(-EINVAL), "i"(__NR_clone),
|
|
"m"(fn), "m"(child_stack), "m"(flags), "m"(arg),
|
|
"m"(parent_tidptr), "m"(newtls), "m"(child_tidptr)
|
|
: "esp", "memory", "ecx", "edx", "esi", "edi");
|
|
LSS_RETURN(int, __res);
|
|
}
|
|
|
|
LSS_INLINE void (*LSS_NAME(restore_rt)(void))(void) {
|
|
/* On i386, the kernel does not know how to return from a signal
|
|
* handler. Instead, it relies on user space to provide a
|
|
* restorer function that calls the {rt_,}sigreturn() system call.
|
|
* Unfortunately, we cannot just reference the glibc version of this
|
|
* function, as glibc goes out of its way to make it inaccessible.
|
|
*/
|
|
void (*res)(void);
|
|
__asm__ __volatile__("call 2f\n"
|
|
"0:.align 16\n"
|
|
"1:movl %1,%%eax\n"
|
|
"int $0x80\n"
|
|
"2:popl %0\n"
|
|
"addl $(1b-0b),%0\n"
|
|
: "=a" (res)
|
|
: "i" (__NR_rt_sigreturn));
|
|
return res;
|
|
}
|
|
LSS_INLINE void (*LSS_NAME(restore)(void))(void) {
|
|
/* On i386, the kernel does not know how to return from a signal
|
|
* handler. Instead, it relies on user space to provide a
|
|
* restorer function that calls the {rt_,}sigreturn() system call.
|
|
* Unfortunately, we cannot just reference the glibc version of this
|
|
* function, as glibc goes out of its way to make it inaccessible.
|
|
*/
|
|
void (*res)(void);
|
|
__asm__ __volatile__("call 2f\n"
|
|
"0:.align 16\n"
|
|
"1:pop %%eax\n"
|
|
"movl %1,%%eax\n"
|
|
"int $0x80\n"
|
|
"2:popl %0\n"
|
|
"addl $(1b-0b),%0\n"
|
|
: "=a" (res)
|
|
: "i" (__NR_sigreturn));
|
|
return res;
|
|
}
|
|
#elif defined(__x86_64__)
|
|
/* There are no known problems with any of the _syscallX() macros
|
|
* currently shipping for x86_64, but we still need to be able to define
|
|
* our own version so that we can override the location of the errno
|
|
* location (e.g. when using the clone() system call with the CLONE_VM
|
|
* option).
|
|
*/
|
|
#undef LSS_ENTRYPOINT
|
|
#define LSS_ENTRYPOINT "syscall\n"
|
|
|
|
/* The x32 ABI has 32 bit longs, but the syscall interface is 64 bit.
|
|
* We need to explicitly cast to an unsigned 64 bit type to avoid implicit
|
|
* sign extension. We can't cast pointers directly because those are
|
|
* 32 bits, and gcc will dump ugly warnings about casting from a pointer
|
|
* to an integer of a different size.
|
|
*/
|
|
#undef LSS_SYSCALL_ARG
|
|
#define LSS_SYSCALL_ARG(a) ((uint64_t)(uintptr_t)(a))
|
|
#undef _LSS_RETURN
|
|
#define _LSS_RETURN(type, res, cast) \
|
|
do { \
|
|
if ((uint64_t)(res) >= (uint64_t)(-4095)) { \
|
|
LSS_ERRNO = -(res); \
|
|
res = -1; \
|
|
} \
|
|
return (type)(cast)(res); \
|
|
} while (0)
|
|
#undef LSS_RETURN
|
|
#define LSS_RETURN(type, res) _LSS_RETURN(type, res, uintptr_t)
|
|
|
|
#undef _LSS_BODY
|
|
#define _LSS_BODY(nr, type, name, cast, ...) \
|
|
long long __res; \
|
|
__asm__ __volatile__(LSS_BODY_ASM##nr LSS_ENTRYPOINT \
|
|
: "=a" (__res) \
|
|
: "0" (__NR_##name) LSS_BODY_ARG##nr(__VA_ARGS__) \
|
|
: LSS_BODY_CLOBBER##nr "r11", "rcx", "memory"); \
|
|
_LSS_RETURN(type, __res, cast)
|
|
#undef LSS_BODY
|
|
#define LSS_BODY(nr, type, name, args...) \
|
|
_LSS_BODY(nr, type, name, uintptr_t, ## args)
|
|
|
|
#undef LSS_BODY_ASM0
|
|
#undef LSS_BODY_ASM1
|
|
#undef LSS_BODY_ASM2
|
|
#undef LSS_BODY_ASM3
|
|
#undef LSS_BODY_ASM4
|
|
#undef LSS_BODY_ASM5
|
|
#undef LSS_BODY_ASM6
|
|
#define LSS_BODY_ASM0
|
|
#define LSS_BODY_ASM1 LSS_BODY_ASM0
|
|
#define LSS_BODY_ASM2 LSS_BODY_ASM1
|
|
#define LSS_BODY_ASM3 LSS_BODY_ASM2
|
|
#define LSS_BODY_ASM4 LSS_BODY_ASM3 "movq %5,%%r10;"
|
|
#define LSS_BODY_ASM5 LSS_BODY_ASM4 "movq %6,%%r8;"
|
|
#define LSS_BODY_ASM6 LSS_BODY_ASM5 "movq %7,%%r9;"
|
|
|
|
#undef LSS_BODY_CLOBBER0
|
|
#undef LSS_BODY_CLOBBER1
|
|
#undef LSS_BODY_CLOBBER2
|
|
#undef LSS_BODY_CLOBBER3
|
|
#undef LSS_BODY_CLOBBER4
|
|
#undef LSS_BODY_CLOBBER5
|
|
#undef LSS_BODY_CLOBBER6
|
|
#define LSS_BODY_CLOBBER0
|
|
#define LSS_BODY_CLOBBER1 LSS_BODY_CLOBBER0
|
|
#define LSS_BODY_CLOBBER2 LSS_BODY_CLOBBER1
|
|
#define LSS_BODY_CLOBBER3 LSS_BODY_CLOBBER2
|
|
#define LSS_BODY_CLOBBER4 LSS_BODY_CLOBBER3 "r10",
|
|
#define LSS_BODY_CLOBBER5 LSS_BODY_CLOBBER4 "r8",
|
|
#define LSS_BODY_CLOBBER6 LSS_BODY_CLOBBER5 "r9",
|
|
|
|
#undef LSS_BODY_ARG0
|
|
#undef LSS_BODY_ARG1
|
|
#undef LSS_BODY_ARG2
|
|
#undef LSS_BODY_ARG3
|
|
#undef LSS_BODY_ARG4
|
|
#undef LSS_BODY_ARG5
|
|
#undef LSS_BODY_ARG6
|
|
#define LSS_BODY_ARG0()
|
|
#define LSS_BODY_ARG1(arg1) \
|
|
LSS_BODY_ARG0(), "D" (arg1)
|
|
#define LSS_BODY_ARG2(arg1, arg2) \
|
|
LSS_BODY_ARG1(arg1), "S" (arg2)
|
|
#define LSS_BODY_ARG3(arg1, arg2, arg3) \
|
|
LSS_BODY_ARG2(arg1, arg2), "d" (arg3)
|
|
#define LSS_BODY_ARG4(arg1, arg2, arg3, arg4) \
|
|
LSS_BODY_ARG3(arg1, arg2, arg3), "r" (arg4)
|
|
#define LSS_BODY_ARG5(arg1, arg2, arg3, arg4, arg5) \
|
|
LSS_BODY_ARG4(arg1, arg2, arg3, arg4), "r" (arg5)
|
|
#define LSS_BODY_ARG6(arg1, arg2, arg3, arg4, arg5, arg6) \
|
|
LSS_BODY_ARG5(arg1, arg2, arg3, arg4, arg5), "r" (arg6)
|
|
|
|
#undef _syscall0
|
|
#define _syscall0(type,name) \
|
|
type LSS_NAME(name)() { \
|
|
LSS_BODY(0, type, name); \
|
|
}
|
|
#undef _syscall1
|
|
#define _syscall1(type,name,type1,arg1) \
|
|
type LSS_NAME(name)(type1 arg1) { \
|
|
LSS_BODY(1, type, name, LSS_SYSCALL_ARG(arg1)); \
|
|
}
|
|
#undef _syscall2
|
|
#define _syscall2(type,name,type1,arg1,type2,arg2) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2) { \
|
|
LSS_BODY(2, type, name, LSS_SYSCALL_ARG(arg1), LSS_SYSCALL_ARG(arg2));\
|
|
}
|
|
#undef _syscall3
|
|
#define _syscall3(type,name,type1,arg1,type2,arg2,type3,arg3) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2, type3 arg3) { \
|
|
LSS_BODY(3, type, name, LSS_SYSCALL_ARG(arg1), LSS_SYSCALL_ARG(arg2), \
|
|
LSS_SYSCALL_ARG(arg3)); \
|
|
}
|
|
#undef _syscall4
|
|
#define _syscall4(type,name,type1,arg1,type2,arg2,type3,arg3,type4,arg4) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2, type3 arg3, type4 arg4) { \
|
|
LSS_BODY(4, type, name, LSS_SYSCALL_ARG(arg1), LSS_SYSCALL_ARG(arg2), \
|
|
LSS_SYSCALL_ARG(arg3), LSS_SYSCALL_ARG(arg4));\
|
|
}
|
|
#undef _syscall5
|
|
#define _syscall5(type,name,type1,arg1,type2,arg2,type3,arg3,type4,arg4, \
|
|
type5,arg5) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2, type3 arg3, type4 arg4, \
|
|
type5 arg5) { \
|
|
LSS_BODY(5, type, name, LSS_SYSCALL_ARG(arg1), LSS_SYSCALL_ARG(arg2), \
|
|
LSS_SYSCALL_ARG(arg3), LSS_SYSCALL_ARG(arg4), \
|
|
LSS_SYSCALL_ARG(arg5)); \
|
|
}
|
|
#undef _syscall6
|
|
#define _syscall6(type,name,type1,arg1,type2,arg2,type3,arg3,type4,arg4, \
|
|
type5,arg5,type6,arg6) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2, type3 arg3, type4 arg4, \
|
|
type5 arg5, type6 arg6) { \
|
|
LSS_BODY(6, type, name, LSS_SYSCALL_ARG(arg1), LSS_SYSCALL_ARG(arg2), \
|
|
LSS_SYSCALL_ARG(arg3), LSS_SYSCALL_ARG(arg4), \
|
|
LSS_SYSCALL_ARG(arg5), LSS_SYSCALL_ARG(arg6));\
|
|
}
|
|
LSS_INLINE int LSS_NAME(clone)(int (*fn)(void *), void *child_stack,
|
|
int flags, void *arg, int *parent_tidptr,
|
|
void *newtls, int *child_tidptr) {
|
|
long long __res;
|
|
{
|
|
__asm__ __volatile__(/* if (fn == NULL)
|
|
* return -EINVAL;
|
|
*/
|
|
"testq %4,%4\n"
|
|
"jz 1f\n"
|
|
|
|
/* if (child_stack == NULL)
|
|
* return -EINVAL;
|
|
*/
|
|
"testq %5,%5\n"
|
|
"jz 1f\n"
|
|
|
|
/* Set up alignment of the child stack:
|
|
* child_stack = (child_stack & ~0xF) - 16;
|
|
*/
|
|
"andq $-16,%5\n"
|
|
"subq $16,%5\n"
|
|
|
|
/* Push "arg" and "fn" onto the stack that will be
|
|
* used by the child.
|
|
*/
|
|
"movq %7,8(%5)\n"
|
|
"movq %4,0(%5)\n"
|
|
|
|
/* %rax = syscall(%rax = __NR_clone,
|
|
* %rdi = flags,
|
|
* %rsi = child_stack,
|
|
* %rdx = parent_tidptr,
|
|
* %r8 = new_tls,
|
|
* %r10 = child_tidptr)
|
|
*/
|
|
"movq %2,%%rax\n"
|
|
"movq %9,%%r8\n"
|
|
"movq %10,%%r10\n"
|
|
"syscall\n"
|
|
|
|
/* if (%rax != 0)
|
|
* return;
|
|
*/
|
|
"testq %%rax,%%rax\n"
|
|
"jnz 1f\n"
|
|
|
|
/* In the child. Terminate frame pointer chain.
|
|
*/
|
|
"xorq %%rbp,%%rbp\n"
|
|
|
|
/* Call "fn(arg)".
|
|
*/
|
|
"popq %%rax\n"
|
|
"popq %%rdi\n"
|
|
"call *%%rax\n"
|
|
|
|
/* Call _exit(%ebx).
|
|
*/
|
|
"movq %%rax,%%rdi\n"
|
|
"movq %3,%%rax\n"
|
|
"syscall\n"
|
|
|
|
/* Return to parent.
|
|
*/
|
|
"1:\n"
|
|
: "=a" (__res)
|
|
: "0"(-EINVAL), "i"(__NR_clone), "i"(__NR_exit),
|
|
"r"(LSS_SYSCALL_ARG(fn)),
|
|
"S"(LSS_SYSCALL_ARG(child_stack)),
|
|
"D"(LSS_SYSCALL_ARG(flags)),
|
|
"r"(LSS_SYSCALL_ARG(arg)),
|
|
"d"(LSS_SYSCALL_ARG(parent_tidptr)),
|
|
"r"(LSS_SYSCALL_ARG(newtls)),
|
|
"r"(LSS_SYSCALL_ARG(child_tidptr))
|
|
: "rsp", "memory", "r8", "r10", "r11", "rcx");
|
|
}
|
|
LSS_RETURN(int, __res);
|
|
}
|
|
|
|
LSS_INLINE void (*LSS_NAME(restore_rt)(void))(void) {
|
|
/* On x86-64, the kernel does not know how to return from
|
|
* a signal handler. Instead, it relies on user space to provide a
|
|
* restorer function that calls the rt_sigreturn() system call.
|
|
* Unfortunately, we cannot just reference the glibc version of this
|
|
* function, as glibc goes out of its way to make it inaccessible.
|
|
*/
|
|
long long res;
|
|
__asm__ __volatile__("call 2f\n"
|
|
"0:.align 16\n"
|
|
"1:movq %1,%%rax\n"
|
|
"syscall\n"
|
|
"2:popq %0\n"
|
|
"addq $(1b-0b),%0\n"
|
|
: "=a" (res)
|
|
: "i" (__NR_rt_sigreturn));
|
|
return (void (*)(void))(uintptr_t)res;
|
|
}
|
|
#elif defined(__arm__)
|
|
/* Most definitions of _syscallX() neglect to mark "memory" as being
|
|
* clobbered. This causes problems with compilers, that do a better job
|
|
* at optimizing across __asm__ calls.
|
|
* So, we just have to redefine all fo the _syscallX() macros.
|
|
*/
|
|
#undef LSS_REG
|
|
#define LSS_REG(r,a) register long __r##r __asm__("r"#r) = (long)a
|
|
|
|
/* r0..r3 are scratch registers and not preserved across function
|
|
* calls. We need to first evaluate the first 4 syscall arguments
|
|
* and store them on stack. They must be loaded into r0..r3 after
|
|
* all function calls to avoid r0..r3 being clobbered.
|
|
*/
|
|
#undef LSS_SAVE_ARG
|
|
#define LSS_SAVE_ARG(r,a) long __tmp##r = (long)a
|
|
#undef LSS_LOAD_ARG
|
|
#define LSS_LOAD_ARG(r) register long __r##r __asm__("r"#r) = __tmp##r
|
|
|
|
#undef LSS_BODY
|
|
#define LSS_BODY(type, name, args...) \
|
|
register long __res_r0 __asm__("r0"); \
|
|
long __res; \
|
|
__SYS_REG(name) \
|
|
__asm__ __volatile__ (__syscall_safe(name) \
|
|
: "=r"(__res_r0) \
|
|
: __SYS_REG_LIST(args) \
|
|
: "lr", "memory"); \
|
|
__res = __res_r0; \
|
|
LSS_RETURN(type, __res)
|
|
#undef _syscall0
|
|
#define _syscall0(type, name) \
|
|
type LSS_NAME(name)() { \
|
|
LSS_BODY(type, name); \
|
|
}
|
|
#undef _syscall1
|
|
#define _syscall1(type, name, type1, arg1) \
|
|
type LSS_NAME(name)(type1 arg1) { \
|
|
/* There is no need for using a volatile temp. */ \
|
|
LSS_REG(0, arg1); \
|
|
LSS_BODY(type, name, "r"(__r0)); \
|
|
}
|
|
#undef _syscall2
|
|
#define _syscall2(type, name, type1, arg1, type2, arg2) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2) { \
|
|
LSS_SAVE_ARG(0, arg1); \
|
|
LSS_SAVE_ARG(1, arg2); \
|
|
LSS_LOAD_ARG(0); \
|
|
LSS_LOAD_ARG(1); \
|
|
LSS_BODY(type, name, "r"(__r0), "r"(__r1)); \
|
|
}
|
|
#undef _syscall3
|
|
#define _syscall3(type, name, type1, arg1, type2, arg2, type3, arg3) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2, type3 arg3) { \
|
|
LSS_SAVE_ARG(0, arg1); \
|
|
LSS_SAVE_ARG(1, arg2); \
|
|
LSS_SAVE_ARG(2, arg3); \
|
|
LSS_LOAD_ARG(0); \
|
|
LSS_LOAD_ARG(1); \
|
|
LSS_LOAD_ARG(2); \
|
|
LSS_BODY(type, name, "r"(__r0), "r"(__r1), "r"(__r2)); \
|
|
}
|
|
#undef _syscall4
|
|
#define _syscall4(type, name, type1, arg1, type2, arg2, type3, arg3, \
|
|
type4, arg4) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2, type3 arg3, type4 arg4) { \
|
|
LSS_SAVE_ARG(0, arg1); \
|
|
LSS_SAVE_ARG(1, arg2); \
|
|
LSS_SAVE_ARG(2, arg3); \
|
|
LSS_SAVE_ARG(3, arg4); \
|
|
LSS_LOAD_ARG(0); \
|
|
LSS_LOAD_ARG(1); \
|
|
LSS_LOAD_ARG(2); \
|
|
LSS_LOAD_ARG(3); \
|
|
LSS_BODY(type, name, "r"(__r0), "r"(__r1), "r"(__r2), "r"(__r3)); \
|
|
}
|
|
#undef _syscall5
|
|
#define _syscall5(type, name, type1, arg1, type2, arg2, type3, arg3, \
|
|
type4, arg4, type5, arg5) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2, type3 arg3, type4 arg4, \
|
|
type5 arg5) { \
|
|
LSS_SAVE_ARG(0, arg1); \
|
|
LSS_SAVE_ARG(1, arg2); \
|
|
LSS_SAVE_ARG(2, arg3); \
|
|
LSS_SAVE_ARG(3, arg4); \
|
|
LSS_REG(4, arg5); \
|
|
LSS_LOAD_ARG(0); \
|
|
LSS_LOAD_ARG(1); \
|
|
LSS_LOAD_ARG(2); \
|
|
LSS_LOAD_ARG(3); \
|
|
LSS_BODY(type, name, "r"(__r0), "r"(__r1), "r"(__r2), "r"(__r3), \
|
|
"r"(__r4)); \
|
|
}
|
|
#undef _syscall6
|
|
#define _syscall6(type, name, type1, arg1, type2, arg2, type3, arg3, \
|
|
type4, arg4, type5, arg5, type6, arg6) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2, type3 arg3, type4 arg4, \
|
|
type5 arg5, type6 arg6) { \
|
|
LSS_SAVE_ARG(0, arg1); \
|
|
LSS_SAVE_ARG(1, arg2); \
|
|
LSS_SAVE_ARG(2, arg3); \
|
|
LSS_SAVE_ARG(3, arg4); \
|
|
LSS_REG(4, arg5); \
|
|
LSS_REG(5, arg6); \
|
|
LSS_LOAD_ARG(0); \
|
|
LSS_LOAD_ARG(1); \
|
|
LSS_LOAD_ARG(2); \
|
|
LSS_LOAD_ARG(3); \
|
|
LSS_BODY(type, name, "r"(__r0), "r"(__r1), "r"(__r2), "r"(__r3), \
|
|
"r"(__r4), "r"(__r5)); \
|
|
}
|
|
LSS_INLINE int LSS_NAME(clone)(int (*fn)(void *), void *child_stack,
|
|
int flags, void *arg, int *parent_tidptr,
|
|
void *newtls, int *child_tidptr) {
|
|
register long __res __asm__("r5");
|
|
{
|
|
if (fn == NULL || child_stack == NULL) {
|
|
__res = -EINVAL;
|
|
goto clone_exit;
|
|
}
|
|
|
|
/* stash first 4 arguments on stack first because we can only load
|
|
* them after all function calls.
|
|
*/
|
|
int tmp_flags = flags;
|
|
int * tmp_stack = (int*) child_stack;
|
|
void * tmp_ptid = parent_tidptr;
|
|
void * tmp_tls = newtls;
|
|
|
|
register int *__ctid __asm__("r4") = child_tidptr;
|
|
|
|
/* Push "arg" and "fn" onto the stack that will be
|
|
* used by the child.
|
|
*/
|
|
*(--tmp_stack) = (int) arg;
|
|
*(--tmp_stack) = (int) fn;
|
|
|
|
/* We must load r0..r3 last after all possible function calls. */
|
|
register int __flags __asm__("r0") = tmp_flags;
|
|
register void *__stack __asm__("r1") = tmp_stack;
|
|
register void *__ptid __asm__("r2") = tmp_ptid;
|
|
register void *__tls __asm__("r3") = tmp_tls;
|
|
|
|
/* %r0 = syscall(%r0 = flags,
|
|
* %r1 = child_stack,
|
|
* %r2 = parent_tidptr,
|
|
* %r3 = newtls,
|
|
* %r4 = child_tidptr)
|
|
*/
|
|
__SYS_REG(clone)
|
|
__asm__ __volatile__(/* %r0 = syscall(%r0 = flags,
|
|
* %r1 = child_stack,
|
|
* %r2 = parent_tidptr,
|
|
* %r3 = newtls,
|
|
* %r4 = child_tidptr)
|
|
*/
|
|
"push {r7}\n"
|
|
"mov r7,%1\n"
|
|
__syscall(clone)"\n"
|
|
|
|
/* if (%r0 != 0)
|
|
* return %r0;
|
|
*/
|
|
"movs %0,r0\n"
|
|
"bne 1f\n"
|
|
|
|
/* In the child, now. Call "fn(arg)".
|
|
*/
|
|
"ldr r0,[sp, #4]\n"
|
|
"mov lr,pc\n"
|
|
"ldr pc,[sp]\n"
|
|
|
|
/* Call _exit(%r0), which never returns. We only
|
|
* need to set r7 for EABI syscall ABI but we do
|
|
* this always to simplify code sharing between
|
|
* old and new syscall ABIs.
|
|
*/
|
|
"mov r7,%2\n"
|
|
__syscall(exit)"\n"
|
|
|
|
/* Pop r7 from the stack only in the parent.
|
|
*/
|
|
"1: pop {r7}\n"
|
|
: "=r" (__res)
|
|
: "r"(__sysreg),
|
|
"i"(__NR_exit), "r"(__stack), "r"(__flags),
|
|
"r"(__ptid), "r"(__tls), "r"(__ctid)
|
|
: "cc", "lr", "memory");
|
|
}
|
|
clone_exit:
|
|
LSS_RETURN(int, __res);
|
|
}
|
|
#elif defined(__mips__)
|
|
#undef LSS_REG
|
|
#define LSS_REG(r,a) register unsigned long __r##r __asm__("$"#r) = \
|
|
(unsigned long)(a)
|
|
|
|
#if _MIPS_SIM == _MIPS_SIM_ABI32
|
|
// See http://sources.redhat.com/ml/libc-alpha/2004-10/msg00050.html
|
|
// or http://www.linux-mips.org/archives/linux-mips/2004-10/msg00142.html
|
|
#define MIPS_SYSCALL_CLOBBERS "$1", "$3", "$8", "$9", "$10", "$11", "$12",\
|
|
"$13", "$14", "$15", "$24", "$25", "memory"
|
|
#else
|
|
#define MIPS_SYSCALL_CLOBBERS "$1", "$3", "$10", "$11", "$12", "$13", \
|
|
"$14", "$15", "$24", "$25", "memory"
|
|
#endif
|
|
|
|
#undef LSS_BODY
|
|
#define LSS_BODY(type,name,r7,...) \
|
|
register unsigned long __v0 __asm__("$2") = __NR_##name; \
|
|
__asm__ __volatile__ ("syscall\n" \
|
|
: "=&r"(__v0), r7 (__r7) \
|
|
: "0"(__v0), ##__VA_ARGS__ \
|
|
: MIPS_SYSCALL_CLOBBERS); \
|
|
LSS_RETURN(type, __v0, __r7)
|
|
#undef _syscall0
|
|
#define _syscall0(type, name) \
|
|
type LSS_NAME(name)() { \
|
|
register unsigned long __r7 __asm__("$7"); \
|
|
LSS_BODY(type, name, "=r"); \
|
|
}
|
|
#undef _syscall1
|
|
#define _syscall1(type, name, type1, arg1) \
|
|
type LSS_NAME(name)(type1 arg1) { \
|
|
register unsigned long __r7 __asm__("$7"); \
|
|
LSS_REG(4, arg1); LSS_BODY(type, name, "=r", "r"(__r4)); \
|
|
}
|
|
#undef _syscall2
|
|
#define _syscall2(type, name, type1, arg1, type2, arg2) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2) { \
|
|
register unsigned long __r7 __asm__("$7"); \
|
|
LSS_REG(4, arg1); LSS_REG(5, arg2); \
|
|
LSS_BODY(type, name, "=r", "r"(__r4), "r"(__r5)); \
|
|
}
|
|
#undef _syscall3
|
|
#define _syscall3(type, name, type1, arg1, type2, arg2, type3, arg3) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2, type3 arg3) { \
|
|
register unsigned long __r7 __asm__("$7"); \
|
|
LSS_REG(4, arg1); LSS_REG(5, arg2); LSS_REG(6, arg3); \
|
|
LSS_BODY(type, name, "=r", "r"(__r4), "r"(__r5), "r"(__r6)); \
|
|
}
|
|
#undef _syscall4
|
|
#define _syscall4(type,name,type1,arg1,type2,arg2,type3,arg3,type4,arg4) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2, type3 arg3, type4 arg4) { \
|
|
LSS_REG(4, arg1); LSS_REG(5, arg2); LSS_REG(6, arg3); \
|
|
LSS_REG(7, arg4); \
|
|
LSS_BODY(type, name, "+r", "r"(__r4), "r"(__r5), "r"(__r6)); \
|
|
}
|
|
#undef _syscall5
|
|
#if _MIPS_SIM == _MIPS_SIM_ABI32
|
|
/* The old 32bit MIPS system call API passes the fifth and sixth argument
|
|
* on the stack, whereas the new APIs use registers "r8" and "r9".
|
|
*/
|
|
#define _syscall5(type,name,type1,arg1,type2,arg2,type3,arg3,type4,arg4, \
|
|
type5,arg5) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2, type3 arg3, type4 arg4, \
|
|
type5 arg5) { \
|
|
LSS_REG(4, arg1); LSS_REG(5, arg2); LSS_REG(6, arg3); \
|
|
LSS_REG(7, arg4); \
|
|
register unsigned long __v0 __asm__("$2"); \
|
|
__asm__ __volatile__ (".set noreorder\n" \
|
|
"lw $2, %6\n" \
|
|
"subu $29, 32\n" \
|
|
"sw $2, 16($29)\n" \
|
|
"li $2, %2\n" \
|
|
"syscall\n" \
|
|
"addiu $29, 32\n" \
|
|
".set reorder\n" \
|
|
: "=&r"(__v0), "+r" (__r7) \
|
|
: "i" (__NR_##name), "r"(__r4), "r"(__r5), \
|
|
"r"(__r6), "m" ((unsigned long)arg5) \
|
|
: MIPS_SYSCALL_CLOBBERS); \
|
|
LSS_RETURN(type, __v0, __r7); \
|
|
}
|
|
#else
|
|
#define _syscall5(type,name,type1,arg1,type2,arg2,type3,arg3,type4,arg4, \
|
|
type5,arg5) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2, type3 arg3, type4 arg4, \
|
|
type5 arg5) { \
|
|
LSS_REG(4, arg1); LSS_REG(5, arg2); LSS_REG(6, arg3); \
|
|
LSS_REG(7, arg4); LSS_REG(8, arg5); \
|
|
LSS_BODY(type, name, "+r", "r"(__r4), "r"(__r5), "r"(__r6), \
|
|
"r"(__r8)); \
|
|
}
|
|
#endif
|
|
#undef _syscall6
|
|
#if _MIPS_SIM == _MIPS_SIM_ABI32
|
|
/* The old 32bit MIPS system call API passes the fifth and sixth argument
|
|
* on the stack, whereas the new APIs use registers "r8" and "r9".
|
|
*/
|
|
#define _syscall6(type,name,type1,arg1,type2,arg2,type3,arg3,type4,arg4, \
|
|
type5,arg5,type6,arg6) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2, type3 arg3, type4 arg4, \
|
|
type5 arg5, type6 arg6) { \
|
|
LSS_REG(4, arg1); LSS_REG(5, arg2); LSS_REG(6, arg3); \
|
|
LSS_REG(7, arg4); \
|
|
register unsigned long __v0 __asm__("$2"); \
|
|
__asm__ __volatile__ (".set noreorder\n" \
|
|
"lw $2, %6\n" \
|
|
"lw $8, %7\n" \
|
|
"subu $29, 32\n" \
|
|
"sw $2, 16($29)\n" \
|
|
"sw $8, 20($29)\n" \
|
|
"li $2, %2\n" \
|
|
"syscall\n" \
|
|
"addiu $29, 32\n" \
|
|
".set reorder\n" \
|
|
: "=&r"(__v0), "+r" (__r7) \
|
|
: "i" (__NR_##name), "r"(__r4), "r"(__r5), \
|
|
"r"(__r6), "m" ((unsigned long)arg5), \
|
|
"m" ((unsigned long)arg6) \
|
|
: MIPS_SYSCALL_CLOBBERS); \
|
|
LSS_RETURN(type, __v0, __r7); \
|
|
}
|
|
#else
|
|
#define _syscall6(type,name,type1,arg1,type2,arg2,type3,arg3,type4,arg4, \
|
|
type5,arg5,type6,arg6) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2, type3 arg3, type4 arg4, \
|
|
type5 arg5,type6 arg6) { \
|
|
LSS_REG(4, arg1); LSS_REG(5, arg2); LSS_REG(6, arg3); \
|
|
LSS_REG(7, arg4); LSS_REG(8, arg5); LSS_REG(9, arg6); \
|
|
LSS_BODY(type, name, "+r", "r"(__r4), "r"(__r5), "r"(__r6), \
|
|
"r"(__r8), "r"(__r9)); \
|
|
}
|
|
#endif
|
|
LSS_INLINE int LSS_NAME(clone)(int (*fn)(void *), void *child_stack,
|
|
int flags, void *arg, int *parent_tidptr,
|
|
void *newtls, int *child_tidptr) {
|
|
register unsigned long __v0 __asm__("$2");
|
|
register unsigned long __r7 __asm__("$7") = (unsigned long)newtls;
|
|
{
|
|
register int __flags __asm__("$4") = flags;
|
|
register void *__stack __asm__("$5") = child_stack;
|
|
register void *__ptid __asm__("$6") = parent_tidptr;
|
|
register int *__ctid __asm__("$8") = child_tidptr;
|
|
__asm__ __volatile__(
|
|
#if _MIPS_SIM == _MIPS_SIM_ABI32 && _MIPS_SZPTR == 32
|
|
"subu $29,24\n"
|
|
#elif _MIPS_SIM == _MIPS_SIM_NABI32
|
|
"sub $29,16\n"
|
|
#else
|
|
"dsubu $29,16\n"
|
|
#endif
|
|
|
|
/* if (fn == NULL || child_stack == NULL)
|
|
* return -EINVAL;
|
|
*/
|
|
"li %0,%2\n"
|
|
"beqz %5,1f\n"
|
|
"beqz %6,1f\n"
|
|
|
|
/* Push "arg" and "fn" onto the stack that will be
|
|
* used by the child.
|
|
*/
|
|
#if _MIPS_SIM == _MIPS_SIM_ABI32 && _MIPS_SZPTR == 32
|
|
"subu %6,32\n"
|
|
"sw %5,0(%6)\n"
|
|
"sw %8,4(%6)\n"
|
|
#elif _MIPS_SIM == _MIPS_SIM_NABI32
|
|
"sub %6,32\n"
|
|
"sw %5,0(%6)\n"
|
|
"sw %8,8(%6)\n"
|
|
#else
|
|
"dsubu %6,32\n"
|
|
"sd %5,0(%6)\n"
|
|
"sd %8,8(%6)\n"
|
|
#endif
|
|
|
|
/* $7 = syscall($4 = flags,
|
|
* $5 = child_stack,
|
|
* $6 = parent_tidptr,
|
|
* $7 = newtls,
|
|
* $8 = child_tidptr)
|
|
*/
|
|
"li $2,%3\n"
|
|
"syscall\n"
|
|
|
|
/* if ($7 != 0)
|
|
* return $2;
|
|
*/
|
|
"bnez $7,1f\n"
|
|
"bnez $2,1f\n"
|
|
|
|
/* In the child, now. Call "fn(arg)".
|
|
*/
|
|
#if _MIPS_SIM == _MIPS_SIM_ABI32 && _MIPS_SZPTR == 32
|
|
"lw $25,0($29)\n"
|
|
"lw $4,4($29)\n"
|
|
#elif _MIPS_SIM == _MIPS_SIM_NABI32
|
|
"lw $25,0($29)\n"
|
|
"lw $4,8($29)\n"
|
|
#else
|
|
"ld $25,0($29)\n"
|
|
"ld $4,8($29)\n"
|
|
#endif
|
|
"jalr $25\n"
|
|
|
|
/* Call _exit($2)
|
|
*/
|
|
"move $4,$2\n"
|
|
"li $2,%4\n"
|
|
"syscall\n"
|
|
|
|
"1:\n"
|
|
#if _MIPS_SIM == _MIPS_SIM_ABI32 && _MIPS_SZPTR == 32
|
|
"addu $29, 24\n"
|
|
#elif _MIPS_SIM == _MIPS_SIM_NABI32
|
|
"add $29, 16\n"
|
|
#else
|
|
"daddu $29,16\n"
|
|
#endif
|
|
: "=&r" (__v0), "=r" (__r7)
|
|
: "i"(-EINVAL), "i"(__NR_clone), "i"(__NR_exit),
|
|
"r"(fn), "r"(__stack), "r"(__flags), "r"(arg),
|
|
"r"(__ptid), "r"(__r7), "r"(__ctid)
|
|
: "$9", "$10", "$11", "$12", "$13", "$14", "$15",
|
|
"$24", "memory");
|
|
}
|
|
LSS_RETURN(int, __v0, __r7);
|
|
}
|
|
#elif defined (__PPC__)
|
|
#undef LSS_LOADARGS_0
|
|
#define LSS_LOADARGS_0(name, dummy...) \
|
|
__sc_0 = __NR_##name
|
|
#undef LSS_LOADARGS_1
|
|
#define LSS_LOADARGS_1(name, arg1) \
|
|
LSS_LOADARGS_0(name); \
|
|
__sc_3 = (unsigned long) (arg1)
|
|
#undef LSS_LOADARGS_2
|
|
#define LSS_LOADARGS_2(name, arg1, arg2) \
|
|
LSS_LOADARGS_1(name, arg1); \
|
|
__sc_4 = (unsigned long) (arg2)
|
|
#undef LSS_LOADARGS_3
|
|
#define LSS_LOADARGS_3(name, arg1, arg2, arg3) \
|
|
LSS_LOADARGS_2(name, arg1, arg2); \
|
|
__sc_5 = (unsigned long) (arg3)
|
|
#undef LSS_LOADARGS_4
|
|
#define LSS_LOADARGS_4(name, arg1, arg2, arg3, arg4) \
|
|
LSS_LOADARGS_3(name, arg1, arg2, arg3); \
|
|
__sc_6 = (unsigned long) (arg4)
|
|
#undef LSS_LOADARGS_5
|
|
#define LSS_LOADARGS_5(name, arg1, arg2, arg3, arg4, arg5) \
|
|
LSS_LOADARGS_4(name, arg1, arg2, arg3, arg4); \
|
|
__sc_7 = (unsigned long) (arg5)
|
|
#undef LSS_LOADARGS_6
|
|
#define LSS_LOADARGS_6(name, arg1, arg2, arg3, arg4, arg5, arg6) \
|
|
LSS_LOADARGS_5(name, arg1, arg2, arg3, arg4, arg5); \
|
|
__sc_8 = (unsigned long) (arg6)
|
|
#undef LSS_ASMINPUT_0
|
|
#define LSS_ASMINPUT_0 "0" (__sc_0)
|
|
#undef LSS_ASMINPUT_1
|
|
#define LSS_ASMINPUT_1 LSS_ASMINPUT_0, "1" (__sc_3)
|
|
#undef LSS_ASMINPUT_2
|
|
#define LSS_ASMINPUT_2 LSS_ASMINPUT_1, "2" (__sc_4)
|
|
#undef LSS_ASMINPUT_3
|
|
#define LSS_ASMINPUT_3 LSS_ASMINPUT_2, "3" (__sc_5)
|
|
#undef LSS_ASMINPUT_4
|
|
#define LSS_ASMINPUT_4 LSS_ASMINPUT_3, "4" (__sc_6)
|
|
#undef LSS_ASMINPUT_5
|
|
#define LSS_ASMINPUT_5 LSS_ASMINPUT_4, "5" (__sc_7)
|
|
#undef LSS_ASMINPUT_6
|
|
#define LSS_ASMINPUT_6 LSS_ASMINPUT_5, "6" (__sc_8)
|
|
#undef LSS_BODY
|
|
#define LSS_BODY(nr, type, name, args...) \
|
|
long __sc_ret, __sc_err; \
|
|
{ \
|
|
register unsigned long __sc_0 __asm__ ("r0"); \
|
|
register unsigned long __sc_3 __asm__ ("r3"); \
|
|
register unsigned long __sc_4 __asm__ ("r4"); \
|
|
register unsigned long __sc_5 __asm__ ("r5"); \
|
|
register unsigned long __sc_6 __asm__ ("r6"); \
|
|
register unsigned long __sc_7 __asm__ ("r7"); \
|
|
register unsigned long __sc_8 __asm__ ("r8"); \
|
|
\
|
|
LSS_LOADARGS_##nr(name, args); \
|
|
__asm__ __volatile__ \
|
|
("sc\n\t" \
|
|
"mfcr %0" \
|
|
: "=&r" (__sc_0), \
|
|
"=&r" (__sc_3), "=&r" (__sc_4), \
|
|
"=&r" (__sc_5), "=&r" (__sc_6), \
|
|
"=&r" (__sc_7), "=&r" (__sc_8) \
|
|
: LSS_ASMINPUT_##nr \
|
|
: "cr0", "ctr", "memory", \
|
|
"r9", "r10", "r11", "r12"); \
|
|
__sc_ret = __sc_3; \
|
|
__sc_err = __sc_0; \
|
|
} \
|
|
LSS_RETURN(type, __sc_ret, __sc_err)
|
|
#undef _syscall0
|
|
#define _syscall0(type, name) \
|
|
type LSS_NAME(name)(void) { \
|
|
LSS_BODY(0, type, name); \
|
|
}
|
|
#undef _syscall1
|
|
#define _syscall1(type, name, type1, arg1) \
|
|
type LSS_NAME(name)(type1 arg1) { \
|
|
LSS_BODY(1, type, name, arg1); \
|
|
}
|
|
#undef _syscall2
|
|
#define _syscall2(type, name, type1, arg1, type2, arg2) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2) { \
|
|
LSS_BODY(2, type, name, arg1, arg2); \
|
|
}
|
|
#undef _syscall3
|
|
#define _syscall3(type, name, type1, arg1, type2, arg2, type3, arg3) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2, type3 arg3) { \
|
|
LSS_BODY(3, type, name, arg1, arg2, arg3); \
|
|
}
|
|
#undef _syscall4
|
|
#define _syscall4(type, name, type1, arg1, type2, arg2, type3, arg3, \
|
|
type4, arg4) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2, type3 arg3, type4 arg4) { \
|
|
LSS_BODY(4, type, name, arg1, arg2, arg3, arg4); \
|
|
}
|
|
#undef _syscall5
|
|
#define _syscall5(type, name, type1, arg1, type2, arg2, type3, arg3, \
|
|
type4, arg4, type5, arg5) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2, type3 arg3, type4 arg4, \
|
|
type5 arg5) { \
|
|
LSS_BODY(5, type, name, arg1, arg2, arg3, arg4, arg5); \
|
|
}
|
|
#undef _syscall6
|
|
#define _syscall6(type, name, type1, arg1, type2, arg2, type3, arg3, \
|
|
type4, arg4, type5, arg5, type6, arg6) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2, type3 arg3, type4 arg4, \
|
|
type5 arg5, type6 arg6) { \
|
|
LSS_BODY(6, type, name, arg1, arg2, arg3, arg4, arg5, arg6); \
|
|
}
|
|
/* clone function adapted from glibc 2.18 clone.S */
|
|
LSS_INLINE int LSS_NAME(clone)(int (*fn)(void *), void *child_stack,
|
|
int flags, void *arg, int *parent_tidptr,
|
|
void *newtls, int *child_tidptr) {
|
|
long __ret, __err;
|
|
{
|
|
#if defined(__PPC64__)
|
|
|
|
/* Stack frame offsets. */
|
|
#if _CALL_ELF != 2
|
|
#define FRAME_MIN_SIZE 112
|
|
#define FRAME_TOC_SAVE 40
|
|
#else
|
|
#define FRAME_MIN_SIZE 32
|
|
#define FRAME_TOC_SAVE 24
|
|
#endif
|
|
|
|
|
|
register int (*__fn)(void *) __asm__ ("r3") = fn;
|
|
register void *__cstack __asm__ ("r4") = child_stack;
|
|
register int __flags __asm__ ("r5") = flags;
|
|
register void * __arg __asm__ ("r6") = arg;
|
|
register int * __ptidptr __asm__ ("r7") = parent_tidptr;
|
|
register void * __newtls __asm__ ("r8") = newtls;
|
|
register int * __ctidptr __asm__ ("r9") = child_tidptr;
|
|
__asm__ __volatile__(
|
|
/* check for fn == NULL
|
|
* and child_stack == NULL
|
|
*/
|
|
"cmpdi cr0, %6, 0\n\t"
|
|
"cmpdi cr1, %7, 0\n\t"
|
|
"cror cr0*4+eq, cr1*4+eq, cr0*4+eq\n\t"
|
|
"beq- cr0, 1f\n\t"
|
|
|
|
/* set up stack frame for child */
|
|
"clrrdi %7, %7, 4\n\t"
|
|
"li 0, 0\n\t"
|
|
"stdu 0, -%13(%7)\n\t"
|
|
|
|
/* fn, arg, child_stack are saved acrVoss the syscall */
|
|
"mr 28, %6\n\t"
|
|
"mr 29, %7\n\t"
|
|
"mr 27, %9\n\t"
|
|
|
|
/* syscall
|
|
r3 == flags
|
|
r4 == child_stack
|
|
r5 == parent_tidptr
|
|
r6 == newtls
|
|
r7 == child_tidptr */
|
|
"mr 3, %8\n\t"
|
|
"mr 5, %10\n\t"
|
|
"mr 6, %11\n\t"
|
|
"mr 7, %12\n\t"
|
|
"li 0, %4\n\t"
|
|
"sc\n\t"
|
|
|
|
/* Test if syscall was successful */
|
|
"cmpdi cr1, 3, 0\n\t"
|
|
"crandc cr1*4+eq, cr1*4+eq, cr0*4+so\n\t"
|
|
"bne- cr1, 1f\n\t"
|
|
|
|
/* Do the function call */
|
|
"std 2, %14(1)\n\t"
|
|
#if _CALL_ELF != 2
|
|
"ld 0, 0(28)\n\t"
|
|
"ld 2, 8(28)\n\t"
|
|
"mtctr 0\n\t"
|
|
#else
|
|
"mr 12, 28\n\t"
|
|
"mtctr 12\n\t"
|
|
#endif
|
|
"mr 3, 27\n\t"
|
|
"bctrl\n\t"
|
|
"ld 2, %14(1)\n\t"
|
|
|
|
/* Call _exit(r3) */
|
|
"li 0, %5\n\t"
|
|
"sc\n\t"
|
|
|
|
/* Return to parent */
|
|
"1:\n\t"
|
|
"mr %0, 3\n\t"
|
|
: "=r" (__ret), "=r" (__err)
|
|
: "0" (-1), "i" (EINVAL),
|
|
"i" (__NR_clone), "i" (__NR_exit),
|
|
"r" (__fn), "r" (__cstack), "r" (__flags),
|
|
"r" (__arg), "r" (__ptidptr), "r" (__newtls),
|
|
"r" (__ctidptr), "i" (FRAME_MIN_SIZE), "i" (FRAME_TOC_SAVE)
|
|
: "cr0", "cr1", "memory", "ctr",
|
|
"r0", "r29", "r27", "r28");
|
|
#else
|
|
register int (*__fn)(void *) __asm__ ("r8") = fn;
|
|
register void *__cstack __asm__ ("r4") = child_stack;
|
|
register int __flags __asm__ ("r3") = flags;
|
|
register void * __arg __asm__ ("r9") = arg;
|
|
register int * __ptidptr __asm__ ("r5") = parent_tidptr;
|
|
register void * __newtls __asm__ ("r6") = newtls;
|
|
register int * __ctidptr __asm__ ("r7") = child_tidptr;
|
|
__asm__ __volatile__(
|
|
/* check for fn == NULL
|
|
* and child_stack == NULL
|
|
*/
|
|
"cmpwi cr0, %6, 0\n\t"
|
|
"cmpwi cr1, %7, 0\n\t"
|
|
"cror cr0*4+eq, cr1*4+eq, cr0*4+eq\n\t"
|
|
"beq- cr0, 1f\n\t"
|
|
|
|
/* set up stack frame for child */
|
|
"clrrwi %7, %7, 4\n\t"
|
|
"li 0, 0\n\t"
|
|
"stwu 0, -16(%7)\n\t"
|
|
|
|
/* fn, arg, child_stack are saved across the syscall: r28-30 */
|
|
"mr 28, %6\n\t"
|
|
"mr 29, %7\n\t"
|
|
"mr 27, %9\n\t"
|
|
|
|
/* syscall */
|
|
"li 0, %4\n\t"
|
|
/* flags already in r3
|
|
* child_stack already in r4
|
|
* ptidptr already in r5
|
|
* newtls already in r6
|
|
* ctidptr already in r7
|
|
*/
|
|
"sc\n\t"
|
|
|
|
/* Test if syscall was successful */
|
|
"cmpwi cr1, 3, 0\n\t"
|
|
"crandc cr1*4+eq, cr1*4+eq, cr0*4+so\n\t"
|
|
"bne- cr1, 1f\n\t"
|
|
|
|
/* Do the function call */
|
|
"mtctr 28\n\t"
|
|
"mr 3, 27\n\t"
|
|
"bctrl\n\t"
|
|
|
|
/* Call _exit(r3) */
|
|
"li 0, %5\n\t"
|
|
"sc\n\t"
|
|
|
|
/* Return to parent */
|
|
"1:\n"
|
|
"mfcr %1\n\t"
|
|
"mr %0, 3\n\t"
|
|
: "=r" (__ret), "=r" (__err)
|
|
: "0" (-1), "1" (EINVAL),
|
|
"i" (__NR_clone), "i" (__NR_exit),
|
|
"r" (__fn), "r" (__cstack), "r" (__flags),
|
|
"r" (__arg), "r" (__ptidptr), "r" (__newtls),
|
|
"r" (__ctidptr)
|
|
: "cr0", "cr1", "memory", "ctr",
|
|
"r0", "r29", "r27", "r28");
|
|
|
|
#endif
|
|
}
|
|
LSS_RETURN(int, __ret, __err);
|
|
}
|
|
#elif defined(__aarch64__)
|
|
#undef LSS_REG
|
|
#define LSS_REG(r,a) register long __x##r __asm__("x"#r) = (long)a
|
|
#undef LSS_BODY
|
|
#define LSS_BODY(type,name,args...) \
|
|
register long __res_x0 __asm__("x0"); \
|
|
long __res; \
|
|
__asm__ __volatile__ ("mov x8, %1\n" \
|
|
"svc 0x0\n" \
|
|
: "=r"(__res_x0) \
|
|
: "i"(__NR_##name) , ## args \
|
|
: "memory"); \
|
|
__res = __res_x0; \
|
|
LSS_RETURN(type, __res)
|
|
#undef _syscall0
|
|
#define _syscall0(type, name) \
|
|
type LSS_NAME(name)(void) { \
|
|
LSS_BODY(type, name); \
|
|
}
|
|
#undef _syscall1
|
|
#define _syscall1(type, name, type1, arg1) \
|
|
type LSS_NAME(name)(type1 arg1) { \
|
|
LSS_REG(0, arg1); LSS_BODY(type, name, "r"(__x0)); \
|
|
}
|
|
#undef _syscall2
|
|
#define _syscall2(type, name, type1, arg1, type2, arg2) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2) { \
|
|
LSS_REG(0, arg1); LSS_REG(1, arg2); \
|
|
LSS_BODY(type, name, "r"(__x0), "r"(__x1)); \
|
|
}
|
|
#undef _syscall3
|
|
#define _syscall3(type, name, type1, arg1, type2, arg2, type3, arg3) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2, type3 arg3) { \
|
|
LSS_REG(0, arg1); LSS_REG(1, arg2); LSS_REG(2, arg3); \
|
|
LSS_BODY(type, name, "r"(__x0), "r"(__x1), "r"(__x2)); \
|
|
}
|
|
#undef _syscall4
|
|
#define _syscall4(type,name,type1,arg1,type2,arg2,type3,arg3,type4,arg4) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2, type3 arg3, type4 arg4) { \
|
|
LSS_REG(0, arg1); LSS_REG(1, arg2); LSS_REG(2, arg3); \
|
|
LSS_REG(3, arg4); \
|
|
LSS_BODY(type, name, "r"(__x0), "r"(__x1), "r"(__x2), "r"(__x3)); \
|
|
}
|
|
#undef _syscall5
|
|
#define _syscall5(type,name,type1,arg1,type2,arg2,type3,arg3,type4,arg4, \
|
|
type5,arg5) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2, type3 arg3, type4 arg4, \
|
|
type5 arg5) { \
|
|
LSS_REG(0, arg1); LSS_REG(1, arg2); LSS_REG(2, arg3); \
|
|
LSS_REG(3, arg4); LSS_REG(4, arg5); \
|
|
LSS_BODY(type, name, "r"(__x0), "r"(__x1), "r"(__x2), "r"(__x3), \
|
|
"r"(__x4)); \
|
|
}
|
|
#undef _syscall6
|
|
#define _syscall6(type,name,type1,arg1,type2,arg2,type3,arg3,type4,arg4, \
|
|
type5,arg5,type6,arg6) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2, type3 arg3, type4 arg4, \
|
|
type5 arg5, type6 arg6) { \
|
|
LSS_REG(0, arg1); LSS_REG(1, arg2); LSS_REG(2, arg3); \
|
|
LSS_REG(3, arg4); LSS_REG(4, arg5); LSS_REG(5, arg6); \
|
|
LSS_BODY(type, name, "r"(__x0), "r"(__x1), "x"(__x2), "r"(__x3), \
|
|
"r"(__x4), "r"(__x5)); \
|
|
}
|
|
/* clone function adapted from glibc 2.18 clone.S */
|
|
LSS_INLINE int LSS_NAME(clone)(int (*fn)(void *), void *child_stack,
|
|
int flags, void *arg, int *parent_tidptr,
|
|
void *newtls, int *child_tidptr) {
|
|
long __res;
|
|
{
|
|
register int (*__fn)(void *) __asm__("x0") = fn;
|
|
register void *__stack __asm__("x1") = child_stack;
|
|
register int __flags __asm__("x2") = flags;
|
|
register void *__arg __asm__("x3") = arg;
|
|
register int *__ptid __asm__("x4") = parent_tidptr;
|
|
register void *__tls __asm__("x5") = newtls;
|
|
register int *__ctid __asm__("x6") = child_tidptr;
|
|
__asm__ __volatile__(/* if (fn == NULL || child_stack == NULL)
|
|
* return -EINVAL;
|
|
*/
|
|
"cbz x0,1f\n"
|
|
"cbz x1,1f\n"
|
|
|
|
/* Push "arg" and "fn" onto the stack that will be
|
|
* used by the child.
|
|
*/
|
|
"stp x0,x3, [x1, #-16]!\n"
|
|
|
|
"mov x0,x2\n" /* flags */
|
|
"mov x2,x4\n" /* ptid */
|
|
"mov x3,x5\n" /* tls */
|
|
"mov x4,x6\n" /* ctid */
|
|
"mov x8,%9\n" /* clone */
|
|
|
|
"svc 0x0\n"
|
|
|
|
/* if (%r0 != 0)
|
|
* return %r0;
|
|
*/
|
|
"cmp x0, #0\n"
|
|
"bne 2f\n"
|
|
|
|
/* In the child, now. Call "fn(arg)".
|
|
*/
|
|
"ldp x1, x0, [sp], #16\n"
|
|
"blr x1\n"
|
|
|
|
/* Call _exit(%r0).
|
|
*/
|
|
"mov x8, %10\n"
|
|
"svc 0x0\n"
|
|
"1:\n"
|
|
"mov x8, %1\n"
|
|
"2:\n"
|
|
: "=r" (__res)
|
|
: "i"(-EINVAL),
|
|
"r"(__fn), "r"(__stack), "r"(__flags), "r"(__arg),
|
|
"r"(__ptid), "r"(__tls), "r"(__ctid),
|
|
"i"(__NR_clone), "i"(__NR_exit)
|
|
: "x30", "memory");
|
|
}
|
|
LSS_RETURN(int, __res);
|
|
}
|
|
#elif defined(__s390__)
|
|
#undef LSS_REG
|
|
#define LSS_REG(r, a) register unsigned long __r##r __asm__("r"#r) = (unsigned long) a
|
|
#undef LSS_BODY
|
|
#define LSS_BODY(type, name, args...) \
|
|
register unsigned long __nr __asm__("r1") \
|
|
= (unsigned long)(__NR_##name); \
|
|
register long __res_r2 __asm__("r2"); \
|
|
long __res; \
|
|
__asm__ __volatile__ \
|
|
("svc 0\n\t" \
|
|
: "=d"(__res_r2) \
|
|
: "d"(__nr), ## args \
|
|
: "memory"); \
|
|
__res = __res_r2; \
|
|
LSS_RETURN(type, __res)
|
|
#undef _syscall0
|
|
#define _syscall0(type, name) \
|
|
type LSS_NAME(name)(void) { \
|
|
LSS_BODY(type, name); \
|
|
}
|
|
#undef _syscall1
|
|
#define _syscall1(type, name, type1, arg1) \
|
|
type LSS_NAME(name)(type1 arg1) { \
|
|
LSS_REG(2, arg1); \
|
|
LSS_BODY(type, name, "0"(__r2)); \
|
|
}
|
|
#undef _syscall2
|
|
#define _syscall2(type, name, type1, arg1, type2, arg2) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2) { \
|
|
LSS_REG(2, arg1); LSS_REG(3, arg2); \
|
|
LSS_BODY(type, name, "0"(__r2), "d"(__r3)); \
|
|
}
|
|
#undef _syscall3
|
|
#define _syscall3(type, name, type1, arg1, type2, arg2, type3, arg3) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2, type3 arg3) { \
|
|
LSS_REG(2, arg1); LSS_REG(3, arg2); LSS_REG(4, arg3); \
|
|
LSS_BODY(type, name, "0"(__r2), "d"(__r3), "d"(__r4)); \
|
|
}
|
|
#undef _syscall4
|
|
#define _syscall4(type, name, type1, arg1, type2, arg2, type3, arg3, \
|
|
type4, arg4) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2, type3 arg3, \
|
|
type4 arg4) { \
|
|
LSS_REG(2, arg1); LSS_REG(3, arg2); LSS_REG(4, arg3); \
|
|
LSS_REG(5, arg4); \
|
|
LSS_BODY(type, name, "0"(__r2), "d"(__r3), "d"(__r4), \
|
|
"d"(__r5)); \
|
|
}
|
|
#undef _syscall5
|
|
#define _syscall5(type, name, type1, arg1, type2, arg2, type3, arg3, \
|
|
type4, arg4, type5, arg5) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2, type3 arg3, \
|
|
type4 arg4, type5 arg5) { \
|
|
LSS_REG(2, arg1); LSS_REG(3, arg2); LSS_REG(4, arg3); \
|
|
LSS_REG(5, arg4); LSS_REG(6, arg5); \
|
|
LSS_BODY(type, name, "0"(__r2), "d"(__r3), "d"(__r4), \
|
|
"d"(__r5), "d"(__r6)); \
|
|
}
|
|
#undef _syscall6
|
|
#define _syscall6(type, name, type1, arg1, type2, arg2, type3, arg3, \
|
|
type4, arg4, type5, arg5, type6, arg6) \
|
|
type LSS_NAME(name)(type1 arg1, type2 arg2, type3 arg3, \
|
|
type4 arg4, type5 arg5, type6 arg6) { \
|
|
LSS_REG(2, arg1); LSS_REG(3, arg2); LSS_REG(4, arg3); \
|
|
LSS_REG(5, arg4); LSS_REG(6, arg5); LSS_REG(7, arg6); \
|
|
LSS_BODY(type, name, "0"(__r2), "d"(__r3), "d"(__r4), \
|
|
"d"(__r5), "d"(__r6), "d"(__r7)); \
|
|
}
|
|
LSS_INLINE int LSS_NAME(clone)(int (*fn)(void *), void *child_stack,
|
|
int flags, void *arg, int *parent_tidptr,
|
|
void *newtls, int *child_tidptr) {
|
|
long __ret;
|
|
{
|
|
register int (*__fn)(void *) __asm__ ("r1") = fn;
|
|
register void *__cstack __asm__ ("r2") = child_stack;
|
|
register int __flags __asm__ ("r3") = flags;
|
|
register void *__arg __asm__ ("r0") = arg;
|
|
register int *__ptidptr __asm__ ("r4") = parent_tidptr;
|
|
register void *__newtls __asm__ ("r6") = newtls;
|
|
register int *__ctidptr __asm__ ("r5") = child_tidptr;
|
|
__asm__ __volatile__ (
|
|
#ifndef __s390x__
|
|
/* arg already in r0 */
|
|
"ltr %4, %4\n\t" /* check fn, which is already in r1 */
|
|
"jz 1f\n\t" /* NULL function pointer, return -EINVAL */
|
|
"ltr %5, %5\n\t" /* check child_stack, which is already in r2 */
|
|
"jz 1f\n\t" /* NULL stack pointer, return -EINVAL */
|
|
/* flags already in r3 */
|
|
/* parent_tidptr already in r4 */
|
|
/* child_tidptr already in r5 */
|
|
/* newtls already in r6 */
|
|
"svc %2\n\t" /* invoke clone syscall */
|
|
"ltr %0,%%r2\n\t" /* load return code into __ret and test */
|
|
"jnz 1f\n\t" /* return to parent if non-zero */
|
|
/* start child thread */
|
|
"lr %%r2, %7\n\t" /* set first parameter to void *arg */
|
|
"ahi %%r15, -96\n\t" /* make room on the stack for the save area */
|
|
"xc 0(4,%%r15), 0(%%r15)\n\t"
|
|
"basr %%r14, %4\n\t" /* jump to fn */
|
|
"svc %3\n" /* invoke exit syscall */
|
|
"1:\n"
|
|
#else
|
|
/* arg already in r0 */
|
|
"ltgr %4, %4\n\t" /* check fn, which is already in r1 */
|
|
"jz 1f\n\t" /* NULL function pointer, return -EINVAL */
|
|
"ltgr %5, %5\n\t" /* check child_stack, which is already in r2 */
|
|
"jz 1f\n\t" /* NULL stack pointer, return -EINVAL */
|
|
/* flags already in r3 */
|
|
/* parent_tidptr already in r4 */
|
|
/* child_tidptr already in r5 */
|
|
/* newtls already in r6 */
|
|
"svc %2\n\t" /* invoke clone syscall */
|
|
"ltgr %0, %%r2\n\t" /* load return code into __ret and test */
|
|
"jnz 1f\n\t" /* return to parent if non-zero */
|
|
/* start child thread */
|
|
"lgr %%r2, %7\n\t" /* set first parameter to void *arg */
|
|
"aghi %%r15, -160\n\t" /* make room on the stack for the save area */
|
|
"xc 0(8,%%r15), 0(%%r15)\n\t"
|
|
"basr %%r14, %4\n\t" /* jump to fn */
|
|
"svc %3\n" /* invoke exit syscall */
|
|
"1:\n"
|
|
#endif
|
|
: "=r" (__ret)
|
|
: "0" (-EINVAL), "i" (__NR_clone), "i" (__NR_exit),
|
|
"d" (__fn), "d" (__cstack), "d" (__flags), "d" (__arg),
|
|
"d" (__ptidptr), "d" (__newtls), "d" (__ctidptr)
|
|
: "cc", "r14", "memory"
|
|
);
|
|
}
|
|
LSS_RETURN(int, __ret);
|
|
}
|
|
#endif
|
|
#define __NR__exit __NR_exit
|
|
#define __NR__gettid __NR_gettid
|
|
#define __NR__mremap __NR_mremap
|
|
LSS_INLINE _syscall1(int, close, int, f)
|
|
LSS_INLINE _syscall1(int, _exit, int, e)
|
|
LSS_INLINE _syscall3(int, fcntl, int, f,
|
|
int, c, long, a)
|
|
LSS_INLINE _syscall2(int, fstat, int, f,
|
|
struct kernel_stat*, b)
|
|
LSS_INLINE _syscall6(int, futex, int*, a,
|
|
int, o, int, v,
|
|
struct kernel_timespec*, t,
|
|
int*, a2,
|
|
int, v3)
|
|
#ifdef __NR_getdents64
|
|
LSS_INLINE _syscall3(int, getdents64, int, f,
|
|
struct kernel_dirent64*, d, int, c)
|
|
#define KERNEL_DIRENT kernel_dirent64
|
|
#define GETDENTS sys_getdents64
|
|
#else
|
|
LSS_INLINE _syscall3(int, getdents, int, f,
|
|
struct kernel_dirent*, d, int, c)
|
|
#define KERNEL_DIRENT kernel_dirent
|
|
#define GETDENTS sys_getdents
|
|
#endif
|
|
LSS_INLINE _syscall0(pid_t, getpid)
|
|
LSS_INLINE _syscall0(pid_t, getppid)
|
|
LSS_INLINE _syscall0(pid_t, _gettid)
|
|
LSS_INLINE _syscall2(int, kill, pid_t, p,
|
|
int, s)
|
|
#if defined(__x86_64__)
|
|
/* Need to make sure off_t isn't truncated to 32-bits under x32. */
|
|
LSS_INLINE off_t LSS_NAME(lseek)(int f, off_t o, int w) {
|
|
_LSS_BODY(3, off_t, lseek, off_t, LSS_SYSCALL_ARG(f), (uint64_t)(o),
|
|
LSS_SYSCALL_ARG(w));
|
|
}
|
|
#else
|
|
LSS_INLINE _syscall3(off_t, lseek, int, f,
|
|
off_t, o, int, w)
|
|
#endif
|
|
LSS_INLINE _syscall2(int, munmap, void*, s,
|
|
size_t, l)
|
|
LSS_INLINE _syscall5(void*, _mremap, void*, o,
|
|
size_t, os, size_t, ns,
|
|
unsigned long, f, void *, a)
|
|
LSS_INLINE _syscall2(int, prctl, int, o,
|
|
long, a)
|
|
LSS_INLINE _syscall4(long, ptrace, int, r,
|
|
pid_t, p, void *, a, void *, d)
|
|
LSS_INLINE _syscall3(ssize_t, read, int, f,
|
|
void *, b, size_t, c)
|
|
LSS_INLINE _syscall4(int, rt_sigaction, int, s,
|
|
const struct kernel_sigaction*, a,
|
|
struct kernel_sigaction*, o, size_t, c)
|
|
LSS_INLINE _syscall4(int, rt_sigprocmask, int, h,
|
|
const struct kernel_sigset_t*, s,
|
|
struct kernel_sigset_t*, o, size_t, c);
|
|
LSS_INLINE _syscall0(int, sched_yield)
|
|
LSS_INLINE _syscall2(int, sigaltstack, const stack_t*, s,
|
|
const stack_t*, o)
|
|
#if defined(__NR_fstatat)
|
|
LSS_INLINE _syscall4(int, fstatat, int, d, const char *, p,
|
|
struct kernel_stat*, b, int, flags)
|
|
LSS_INLINE int LSS_NAME(stat)(const char* p, struct kernel_stat* b) {
|
|
return LSS_NAME(fstatat)(AT_FDCWD,p,b,0);
|
|
}
|
|
#else
|
|
LSS_INLINE _syscall2(int, stat, const char*, f,
|
|
struct kernel_stat*, b)
|
|
#endif
|
|
LSS_INLINE _syscall3(ssize_t, write, int, f,
|
|
const void *, b, size_t, c)
|
|
#if defined(__NR_getcpu)
|
|
LSS_INLINE _syscall3(long, getcpu, unsigned *, cpu,
|
|
unsigned *, node, void *, unused);
|
|
#endif
|
|
#if defined(__x86_64__) || defined(__aarch64__) || \
|
|
(defined(__mips__) && _MIPS_SIM != _MIPS_SIM_ABI32)
|
|
LSS_INLINE _syscall3(int, socket, int, d,
|
|
int, t, int, p)
|
|
#endif
|
|
#if defined(__x86_64__) || defined(__s390x__)
|
|
LSS_INLINE int LSS_NAME(sigaction)(int signum,
|
|
const struct kernel_sigaction *act,
|
|
struct kernel_sigaction *oldact) {
|
|
#if defined(__x86_64__)
|
|
/* On x86_64, the kernel requires us to always set our own
|
|
* SA_RESTORER in order to be able to return from a signal handler.
|
|
* This function must have a "magic" signature that the "gdb"
|
|
* (and maybe the kernel?) can recognize.
|
|
*/
|
|
if (act != NULL && !(act->sa_flags & SA_RESTORER)) {
|
|
struct kernel_sigaction a = *act;
|
|
a.sa_flags |= SA_RESTORER;
|
|
a.sa_restorer = LSS_NAME(restore_rt)();
|
|
return LSS_NAME(rt_sigaction)(signum, &a, oldact,
|
|
(KERNEL_NSIG+7)/8);
|
|
} else
|
|
#endif
|
|
return LSS_NAME(rt_sigaction)(signum, act, oldact,
|
|
(KERNEL_NSIG+7)/8);
|
|
}
|
|
|
|
LSS_INLINE int LSS_NAME(sigprocmask)(int how,
|
|
const struct kernel_sigset_t *set,
|
|
struct kernel_sigset_t *oldset) {
|
|
return LSS_NAME(rt_sigprocmask)(how, set, oldset, (KERNEL_NSIG+7)/8);
|
|
}
|
|
#endif
|
|
#if (defined(__aarch64__)) || \
|
|
(defined(__mips__) && (_MIPS_ISA == _MIPS_ISA_MIPS64))
|
|
LSS_INLINE int LSS_NAME(sigaction)(int signum,
|
|
const struct kernel_sigaction *act,
|
|
struct kernel_sigaction *oldact) {
|
|
return LSS_NAME(rt_sigaction)(signum, act, oldact, (KERNEL_NSIG+7)/8);
|
|
|
|
}
|
|
LSS_INLINE int LSS_NAME(sigprocmask)(int how,
|
|
const struct kernel_sigset_t *set,
|
|
struct kernel_sigset_t *oldset) {
|
|
return LSS_NAME(rt_sigprocmask)(how, set, oldset, (KERNEL_NSIG+7)/8);
|
|
}
|
|
#endif
|
|
#ifdef __NR_wait4
|
|
LSS_INLINE _syscall4(pid_t, wait4, pid_t, p,
|
|
int*, s, int, o,
|
|
struct kernel_rusage*, r)
|
|
LSS_INLINE pid_t LSS_NAME(waitpid)(pid_t pid, int *status, int options){
|
|
return LSS_NAME(wait4)(pid, status, options, 0);
|
|
}
|
|
#else
|
|
LSS_INLINE _syscall3(pid_t, waitpid, pid_t, p,
|
|
int*, s, int, o)
|
|
#endif
|
|
#ifdef __NR_openat
|
|
LSS_INLINE _syscall4(int, openat, int, d, const char *, p, int, f, int, m)
|
|
LSS_INLINE int LSS_NAME(open)(const char* p, int f, int m) {
|
|
return LSS_NAME(openat)(AT_FDCWD,p,f,m );
|
|
}
|
|
#else
|
|
LSS_INLINE _syscall3(int, open, const char*, p,
|
|
int, f, int, m)
|
|
#endif
|
|
LSS_INLINE int LSS_NAME(sigemptyset)(struct kernel_sigset_t *set) {
|
|
memset(&set->sig, 0, sizeof(set->sig));
|
|
return 0;
|
|
}
|
|
|
|
LSS_INLINE int LSS_NAME(sigfillset)(struct kernel_sigset_t *set) {
|
|
memset(&set->sig, -1, sizeof(set->sig));
|
|
return 0;
|
|
}
|
|
|
|
LSS_INLINE int LSS_NAME(sigaddset)(struct kernel_sigset_t *set,
|
|
int signum) {
|
|
if (signum < 1 || signum > (int)(8*sizeof(set->sig))) {
|
|
LSS_ERRNO = EINVAL;
|
|
return -1;
|
|
} else {
|
|
set->sig[(signum - 1)/(8*sizeof(set->sig[0]))]
|
|
|= 1UL << ((signum - 1) % (8*sizeof(set->sig[0])));
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
LSS_INLINE int LSS_NAME(sigdelset)(struct kernel_sigset_t *set,
|
|
int signum) {
|
|
if (signum < 1 || signum > (int)(8*sizeof(set->sig))) {
|
|
LSS_ERRNO = EINVAL;
|
|
return -1;
|
|
} else {
|
|
set->sig[(signum - 1)/(8*sizeof(set->sig[0]))]
|
|
&= ~(1UL << ((signum - 1) % (8*sizeof(set->sig[0]))));
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
#if defined(__i386__) || \
|
|
defined(__arm__) || \
|
|
(defined(__mips__) && _MIPS_SIM == _MIPS_SIM_ABI32) || \
|
|
defined(__PPC__) || \
|
|
(defined(__s390__) && !defined(__s390x__))
|
|
#define __NR__sigaction __NR_sigaction
|
|
#define __NR__sigprocmask __NR_sigprocmask
|
|
LSS_INLINE _syscall2(int, fstat64, int, f,
|
|
struct kernel_stat64 *, b)
|
|
LSS_INLINE _syscall5(int, _llseek, uint, fd, ulong, hi, ulong, lo,
|
|
loff_t *, res, uint, wh)
|
|
#if defined(__s390__) && !defined(__s390x__)
|
|
/* On s390, mmap2() arguments are passed in memory. */
|
|
LSS_INLINE void* LSS_NAME(_mmap2)(void *s, size_t l, int p, int f, int d,
|
|
off_t o) {
|
|
unsigned long buf[6] = { (unsigned long) s, (unsigned long) l,
|
|
(unsigned long) p, (unsigned long) f,
|
|
(unsigned long) d, (unsigned long) o };
|
|
LSS_REG(2, buf);
|
|
LSS_BODY(void*, mmap2, "0"(__r2));
|
|
}
|
|
#elif !defined(__PPC64__)
|
|
#define __NR__mmap2 __NR_mmap2
|
|
LSS_INLINE _syscall6(void*, _mmap2, void*, s,
|
|
size_t, l, int, p,
|
|
int, f, int, d,
|
|
off_t, o)
|
|
#endif
|
|
LSS_INLINE _syscall3(int, _sigaction, int, s,
|
|
const struct kernel_old_sigaction*, a,
|
|
struct kernel_old_sigaction*, o)
|
|
LSS_INLINE _syscall3(int, _sigprocmask, int, h,
|
|
const unsigned long*, s,
|
|
unsigned long*, o)
|
|
LSS_INLINE _syscall2(int, stat64, const char *, p,
|
|
struct kernel_stat64 *, b)
|
|
|
|
LSS_INLINE int LSS_NAME(sigaction)(int signum,
|
|
const struct kernel_sigaction *act,
|
|
struct kernel_sigaction *oldact) {
|
|
int old_errno = LSS_ERRNO;
|
|
int rc;
|
|
struct kernel_sigaction a;
|
|
if (act != NULL) {
|
|
a = *act;
|
|
#ifdef __i386__
|
|
/* On i386, the kernel requires us to always set our own
|
|
* SA_RESTORER when using realtime signals. Otherwise, it does not
|
|
* know how to return from a signal handler. This function must have
|
|
* a "magic" signature that the "gdb" (and maybe the kernel?) can
|
|
* recognize.
|
|
* Apparently, a SA_RESTORER is implicitly set by the kernel, when
|
|
* using non-realtime signals.
|
|
*
|
|
* TODO: Test whether ARM needs a restorer
|
|
*/
|
|
if (!(a.sa_flags & SA_RESTORER)) {
|
|
a.sa_flags |= SA_RESTORER;
|
|
a.sa_restorer = (a.sa_flags & SA_SIGINFO)
|
|
? LSS_NAME(restore_rt)() : LSS_NAME(restore)();
|
|
}
|
|
#endif
|
|
}
|
|
rc = LSS_NAME(rt_sigaction)(signum, act ? &a : act, oldact,
|
|
(KERNEL_NSIG+7)/8);
|
|
if (rc < 0 && LSS_ERRNO == ENOSYS) {
|
|
struct kernel_old_sigaction oa, ooa, *ptr_a = &oa, *ptr_oa = &ooa;
|
|
if (!act) {
|
|
ptr_a = NULL;
|
|
} else {
|
|
oa.sa_handler_ = act->sa_handler_;
|
|
memcpy(&oa.sa_mask, &act->sa_mask, sizeof(oa.sa_mask));
|
|
#ifndef __mips__
|
|
oa.sa_restorer = act->sa_restorer;
|
|
#endif
|
|
oa.sa_flags = act->sa_flags;
|
|
}
|
|
if (!oldact) {
|
|
ptr_oa = NULL;
|
|
}
|
|
LSS_ERRNO = old_errno;
|
|
rc = LSS_NAME(_sigaction)(signum, ptr_a, ptr_oa);
|
|
if (rc == 0 && oldact) {
|
|
if (act) {
|
|
memcpy(oldact, act, sizeof(*act));
|
|
} else {
|
|
memset(oldact, 0, sizeof(*oldact));
|
|
}
|
|
oldact->sa_handler_ = ptr_oa->sa_handler_;
|
|
oldact->sa_flags = ptr_oa->sa_flags;
|
|
memcpy(&oldact->sa_mask, &ptr_oa->sa_mask, sizeof(ptr_oa->sa_mask));
|
|
#ifndef __mips__
|
|
oldact->sa_restorer = ptr_oa->sa_restorer;
|
|
#endif
|
|
}
|
|
}
|
|
return rc;
|
|
}
|
|
|
|
LSS_INLINE int LSS_NAME(sigprocmask)(int how,
|
|
const struct kernel_sigset_t *set,
|
|
struct kernel_sigset_t *oldset) {
|
|
int olderrno = LSS_ERRNO;
|
|
int rc = LSS_NAME(rt_sigprocmask)(how, set, oldset, (KERNEL_NSIG+7)/8);
|
|
if (rc < 0 && LSS_ERRNO == ENOSYS) {
|
|
LSS_ERRNO = olderrno;
|
|
if (oldset) {
|
|
LSS_NAME(sigemptyset)(oldset);
|
|
}
|
|
rc = LSS_NAME(_sigprocmask)(how,
|
|
set ? &set->sig[0] : NULL,
|
|
oldset ? &oldset->sig[0] : NULL);
|
|
}
|
|
return rc;
|
|
}
|
|
#endif
|
|
#if defined(__i386__) || \
|
|
defined(__ARM_ARCH_3__) || defined(__ARM_EABI__) || \
|
|
(defined(__mips__) && _MIPS_SIM == _MIPS_SIM_ABI32) || \
|
|
(defined(__PPC__) && !defined(__PPC64__)) || \
|
|
(defined(__s390__) && !defined(__s390x__))
|
|
/* On these architectures, implement mmap() with mmap2(). */
|
|
LSS_INLINE void* LSS_NAME(mmap)(void *s, size_t l, int p, int f, int d,
|
|
int64_t o) {
|
|
if (o % 4096) {
|
|
LSS_ERRNO = EINVAL;
|
|
return (void *) -1;
|
|
}
|
|
return LSS_NAME(_mmap2)(s, l, p, f, d, (o / 4096));
|
|
}
|
|
#elif defined(__s390x__)
|
|
/* On s390x, mmap() arguments are passed in memory. */
|
|
LSS_INLINE void* LSS_NAME(mmap)(void *s, size_t l, int p, int f, int d,
|
|
int64_t o) {
|
|
unsigned long buf[6] = { (unsigned long) s, (unsigned long) l,
|
|
(unsigned long) p, (unsigned long) f,
|
|
(unsigned long) d, (unsigned long) o };
|
|
LSS_REG(2, buf);
|
|
LSS_BODY(void*, mmap, "0"(__r2));
|
|
}
|
|
#elif defined(__x86_64__)
|
|
/* Need to make sure __off64_t isn't truncated to 32-bits under x32. */
|
|
LSS_INLINE void* LSS_NAME(mmap)(void *s, size_t l, int p, int f, int d,
|
|
int64_t o) {
|
|
LSS_BODY(6, void*, mmap, LSS_SYSCALL_ARG(s), LSS_SYSCALL_ARG(l),
|
|
LSS_SYSCALL_ARG(p), LSS_SYSCALL_ARG(f),
|
|
LSS_SYSCALL_ARG(d), (uint64_t)(o));
|
|
}
|
|
#else
|
|
/* Remaining 64-bit architectures. */
|
|
LSS_INLINE _syscall6(void*, mmap, void*, addr, size_t, length, int, prot,
|
|
int, flags, int, fd, int64_t, offset)
|
|
#endif
|
|
#if defined(__i386__) || \
|
|
defined(__PPC__) || \
|
|
(defined(__arm__) && !defined(__ARM_EABI__)) || \
|
|
(defined(__mips__) && _MIPS_SIM == _MIPS_SIM_ABI32) || \
|
|
defined(__s390__)
|
|
|
|
/* See sys_socketcall in net/socket.c in kernel source.
|
|
* It de-multiplexes on its first arg and unpacks the arglist
|
|
* array in its second arg.
|
|
*/
|
|
LSS_INLINE _syscall2(int, socketcall, int, c, unsigned long*, a)
|
|
|
|
LSS_INLINE int LSS_NAME(socket)(int domain, int type, int protocol) {
|
|
unsigned long args[3] = {
|
|
(unsigned long) domain,
|
|
(unsigned long) type,
|
|
(unsigned long) protocol
|
|
};
|
|
return LSS_NAME(socketcall)(1, args);
|
|
}
|
|
#elif defined(__ARM_EABI__)
|
|
LSS_INLINE _syscall3(int, socket, int, d,
|
|
int, t, int, p)
|
|
#endif
|
|
#if defined(__mips__)
|
|
/* sys_pipe() on MIPS has non-standard calling conventions, as it returns
|
|
* both file handles through CPU registers.
|
|
*/
|
|
LSS_INLINE int LSS_NAME(pipe)(int *p) {
|
|
register unsigned long __v0 __asm__("$2") = __NR_pipe;
|
|
register unsigned long __v1 __asm__("$3");
|
|
register unsigned long __r7 __asm__("$7");
|
|
__asm__ __volatile__ ("syscall\n"
|
|
: "=&r"(__v0), "=&r"(__v1), "+r" (__r7)
|
|
: "0"(__v0)
|
|
: "$8", "$9", "$10", "$11", "$12",
|
|
"$13", "$14", "$15", "$24", "memory");
|
|
if (__r7) {
|
|
LSS_ERRNO = __v0;
|
|
return -1;
|
|
} else {
|
|
p[0] = __v0;
|
|
p[1] = __v1;
|
|
return 0;
|
|
}
|
|
}
|
|
#elif defined(__NR_pipe2)
|
|
LSS_INLINE _syscall2(int, pipe2, int *, p,
|
|
int, f )
|
|
LSS_INLINE int LSS_NAME(pipe)( int * p) {
|
|
return LSS_NAME(pipe2)(p, 0);
|
|
}
|
|
#else
|
|
LSS_INLINE _syscall1(int, pipe, int *, p)
|
|
#endif
|
|
|
|
LSS_INLINE pid_t LSS_NAME(gettid)() {
|
|
pid_t tid = LSS_NAME(_gettid)();
|
|
if (tid != -1) {
|
|
return tid;
|
|
}
|
|
return LSS_NAME(getpid)();
|
|
}
|
|
|
|
LSS_INLINE void *LSS_NAME(mremap)(void *old_address, size_t old_size,
|
|
size_t new_size, int flags, ...) {
|
|
va_list ap;
|
|
void *new_address, *rc;
|
|
va_start(ap, flags);
|
|
new_address = va_arg(ap, void *);
|
|
rc = LSS_NAME(_mremap)(old_address, old_size, new_size,
|
|
flags, new_address);
|
|
va_end(ap);
|
|
return rc;
|
|
}
|
|
|
|
LSS_INLINE int LSS_NAME(ptrace_detach)(pid_t pid) {
|
|
/* PTRACE_DETACH can sometimes forget to wake up the tracee and it
|
|
* then sends job control signals to the real parent, rather than to
|
|
* the tracer. We reduce the risk of this happening by starting a
|
|
* whole new time slice, and then quickly sending a SIGCONT signal
|
|
* right after detaching from the tracee.
|
|
*/
|
|
int rc, err;
|
|
LSS_NAME(sched_yield)();
|
|
rc = LSS_NAME(ptrace)(PTRACE_DETACH, pid, (void *)0, (void *)0);
|
|
err = LSS_ERRNO;
|
|
LSS_NAME(kill)(pid, SIGCONT);
|
|
LSS_ERRNO = err;
|
|
return rc;
|
|
}
|
|
#endif
|
|
|
|
#if defined(__cplusplus) && !defined(SYS_CPLUSPLUS)
|
|
}
|
|
#endif
|
|
|
|
#endif
|
|
#endif
|