sanitizer_stoptheworld_netbsd_libcdep.cpp revision 360784
1//===-- sanitizer_stoptheworld_netbsd_libcdep.cpp -------------------------===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8//
9// See sanitizer_stoptheworld.h for details.
10// This implementation was inspired by Markus Gutschke's linuxthreads.cc.
11//
12// This is a NetBSD variation of Linux stoptheworld implementation
13// See sanitizer_stoptheworld_linux_libcdep.cpp for code comments.
14//
15//===----------------------------------------------------------------------===//
16
17#include "sanitizer_platform.h"
18
19#if SANITIZER_NETBSD
20
21#include "sanitizer_stoptheworld.h"
22
23#include "sanitizer_atomic.h"
24#include "sanitizer_platform_limits_posix.h"
25
26#include <sys/types.h>
27
28#include <sys/ptrace.h>
29#include <sys/uio.h>
30#include <sys/wait.h>
31
32#include <machine/reg.h>
33
34#include <elf.h>
35#include <errno.h>
36#include <sched.h>
37#include <signal.h>
38#include <stddef.h>
39
40#define internal_sigaction_norestorer internal_sigaction
41
42#include "sanitizer_common.h"
43#include "sanitizer_flags.h"
44#include "sanitizer_libc.h"
45#include "sanitizer_linux.h"
46#include "sanitizer_mutex.h"
47#include "sanitizer_placement_new.h"
48
49namespace __sanitizer {
50
51class SuspendedThreadsListNetBSD : public SuspendedThreadsList {
52 public:
53  SuspendedThreadsListNetBSD() { thread_ids_.reserve(1024); }
54
55  tid_t GetThreadID(uptr index) const;
56  uptr ThreadCount() const;
57  bool ContainsTid(tid_t thread_id) const;
58  void Append(tid_t tid);
59
60  PtraceRegistersStatus GetRegistersAndSP(uptr index, uptr *buffer,
61                                          uptr *sp) const;
62  uptr RegisterCount() const;
63
64 private:
65  InternalMmapVector<tid_t> thread_ids_;
66};
67
68struct TracerThreadArgument {
69  StopTheWorldCallback callback;
70  void *callback_argument;
71  BlockingMutex mutex;
72  atomic_uintptr_t done;
73  uptr parent_pid;
74};
75
76class ThreadSuspender {
77 public:
78  explicit ThreadSuspender(pid_t pid, TracerThreadArgument *arg)
79      : arg(arg), pid_(pid) {
80    CHECK_GE(pid, 0);
81  }
82  bool SuspendAllThreads();
83  void ResumeAllThreads();
84  void KillAllThreads();
85  SuspendedThreadsListNetBSD &suspended_threads_list() {
86    return suspended_threads_list_;
87  }
88  TracerThreadArgument *arg;
89
90 private:
91  SuspendedThreadsListNetBSD suspended_threads_list_;
92  pid_t pid_;
93};
94
95void ThreadSuspender::ResumeAllThreads() {
96  int pterrno;
97  if (!internal_iserror(internal_ptrace(PT_DETACH, pid_, (void *)(uptr)1, 0),
98                        &pterrno)) {
99    VReport(2, "Detached from process %d.\n", pid_);
100  } else {
101    VReport(1, "Could not detach from process %d (errno %d).\n", pid_, pterrno);
102  }
103}
104
105void ThreadSuspender::KillAllThreads() {
106  internal_ptrace(PT_KILL, pid_, nullptr, 0);
107}
108
109bool ThreadSuspender::SuspendAllThreads() {
110  int pterrno;
111  if (internal_iserror(internal_ptrace(PT_ATTACH, pid_, nullptr, 0),
112                       &pterrno)) {
113    Printf("Could not attach to process %d (errno %d).\n", pid_, pterrno);
114    return false;
115  }
116
117  int status;
118  uptr waitpid_status;
119  HANDLE_EINTR(waitpid_status, internal_waitpid(pid_, &status, 0));
120
121  VReport(2, "Attached to process %d.\n", pid_);
122
123#ifdef PT_LWPNEXT
124  struct ptrace_lwpstatus pl;
125  int op = PT_LWPNEXT;
126#else
127  struct ptrace_lwpinfo pl;
128  int op = PT_LWPINFO;
129#endif
130
131  pl.pl_lwpid = 0;
132
133  int val;
134  while ((val = ptrace(op, pid_, (void *)&pl, sizeof(pl))) != -1 &&
135         pl.pl_lwpid != 0) {
136    suspended_threads_list_.Append(pl.pl_lwpid);
137    VReport(2, "Appended thread %d in process %d.\n", pl.pl_lwpid, pid_);
138  }
139  return true;
140}
141
142// Pointer to the ThreadSuspender instance for use in signal handler.
143static ThreadSuspender *thread_suspender_instance = nullptr;
144
145// Synchronous signals that should not be blocked.
146static const int kSyncSignals[] = {SIGABRT, SIGILL,  SIGFPE, SIGSEGV,
147                                   SIGBUS,  SIGXCPU, SIGXFSZ};
148
149static void TracerThreadDieCallback() {
150  ThreadSuspender *inst = thread_suspender_instance;
151  if (inst && stoptheworld_tracer_pid == internal_getpid()) {
152    inst->KillAllThreads();
153    thread_suspender_instance = nullptr;
154  }
155}
156
157// Signal handler to wake up suspended threads when the tracer thread dies.
158static void TracerThreadSignalHandler(int signum, __sanitizer_siginfo *siginfo,
159                                      void *uctx) {
160  SignalContext ctx(siginfo, uctx);
161  Printf("Tracer caught signal %d: addr=0x%zx pc=0x%zx sp=0x%zx\n", signum,
162         ctx.addr, ctx.pc, ctx.sp);
163  ThreadSuspender *inst = thread_suspender_instance;
164  if (inst) {
165    if (signum == SIGABRT)
166      inst->KillAllThreads();
167    else
168      inst->ResumeAllThreads();
169    RAW_CHECK(RemoveDieCallback(TracerThreadDieCallback));
170    thread_suspender_instance = nullptr;
171    atomic_store(&inst->arg->done, 1, memory_order_relaxed);
172  }
173  internal__exit((signum == SIGABRT) ? 1 : 2);
174}
175
176// Size of alternative stack for signal handlers in the tracer thread.
177static const int kHandlerStackSize = 8192;
178
179// This function will be run as a cloned task.
180static int TracerThread(void *argument) {
181  TracerThreadArgument *tracer_thread_argument =
182      (TracerThreadArgument *)argument;
183
184  // Check if parent is already dead.
185  if (internal_getppid() != tracer_thread_argument->parent_pid)
186    internal__exit(4);
187
188  // Wait for the parent thread to finish preparations.
189  tracer_thread_argument->mutex.Lock();
190  tracer_thread_argument->mutex.Unlock();
191
192  RAW_CHECK(AddDieCallback(TracerThreadDieCallback));
193
194  ThreadSuspender thread_suspender(internal_getppid(), tracer_thread_argument);
195  // Global pointer for the signal handler.
196  thread_suspender_instance = &thread_suspender;
197
198  // Alternate stack for signal handling.
199  InternalMmapVector<char> handler_stack_memory(kHandlerStackSize);
200  stack_t handler_stack;
201  internal_memset(&handler_stack, 0, sizeof(handler_stack));
202  handler_stack.ss_sp = handler_stack_memory.data();
203  handler_stack.ss_size = kHandlerStackSize;
204  internal_sigaltstack(&handler_stack, nullptr);
205
206  // Install our handler for synchronous signals. Other signals should be
207  // blocked by the mask we inherited from the parent thread.
208  for (uptr i = 0; i < ARRAY_SIZE(kSyncSignals); i++) {
209    __sanitizer_sigaction act;
210    internal_memset(&act, 0, sizeof(act));
211    act.sigaction = TracerThreadSignalHandler;
212    act.sa_flags = SA_ONSTACK | SA_SIGINFO;
213    internal_sigaction_norestorer(kSyncSignals[i], &act, 0);
214  }
215
216  int exit_code = 0;
217  if (!thread_suspender.SuspendAllThreads()) {
218    VReport(1, "Failed suspending threads.\n");
219    exit_code = 3;
220  } else {
221    tracer_thread_argument->callback(thread_suspender.suspended_threads_list(),
222                                     tracer_thread_argument->callback_argument);
223    thread_suspender.ResumeAllThreads();
224    exit_code = 0;
225  }
226  RAW_CHECK(RemoveDieCallback(TracerThreadDieCallback));
227  thread_suspender_instance = nullptr;
228  atomic_store(&tracer_thread_argument->done, 1, memory_order_relaxed);
229  return exit_code;
230}
231
232class ScopedStackSpaceWithGuard {
233 public:
234  explicit ScopedStackSpaceWithGuard(uptr stack_size) {
235    stack_size_ = stack_size;
236    guard_size_ = GetPageSizeCached();
237    // FIXME: Omitting MAP_STACK here works in current kernels but might break
238    // in the future.
239    guard_start_ =
240        (uptr)MmapOrDie(stack_size_ + guard_size_, "ScopedStackWithGuard");
241    CHECK(MprotectNoAccess((uptr)guard_start_, guard_size_));
242  }
243  ~ScopedStackSpaceWithGuard() {
244    UnmapOrDie((void *)guard_start_, stack_size_ + guard_size_);
245  }
246  void *Bottom() const {
247    return (void *)(guard_start_ + stack_size_ + guard_size_);
248  }
249
250 private:
251  uptr stack_size_;
252  uptr guard_size_;
253  uptr guard_start_;
254};
255
256static __sanitizer_sigset_t blocked_sigset;
257static __sanitizer_sigset_t old_sigset;
258
259struct ScopedSetTracerPID {
260  explicit ScopedSetTracerPID(uptr tracer_pid) {
261    stoptheworld_tracer_pid = tracer_pid;
262    stoptheworld_tracer_ppid = internal_getpid();
263  }
264  ~ScopedSetTracerPID() {
265    stoptheworld_tracer_pid = 0;
266    stoptheworld_tracer_ppid = 0;
267  }
268};
269
270void StopTheWorld(StopTheWorldCallback callback, void *argument) {
271  // Prepare the arguments for TracerThread.
272  struct TracerThreadArgument tracer_thread_argument;
273  tracer_thread_argument.callback = callback;
274  tracer_thread_argument.callback_argument = argument;
275  tracer_thread_argument.parent_pid = internal_getpid();
276  atomic_store(&tracer_thread_argument.done, 0, memory_order_relaxed);
277  const uptr kTracerStackSize = 2 * 1024 * 1024;
278  ScopedStackSpaceWithGuard tracer_stack(kTracerStackSize);
279
280  tracer_thread_argument.mutex.Lock();
281
282  internal_sigfillset(&blocked_sigset);
283  for (uptr i = 0; i < ARRAY_SIZE(kSyncSignals); i++)
284    internal_sigdelset(&blocked_sigset, kSyncSignals[i]);
285  int rv = internal_sigprocmask(SIG_BLOCK, &blocked_sigset, &old_sigset);
286  CHECK_EQ(rv, 0);
287  uptr tracer_pid = internal_clone(TracerThread, tracer_stack.Bottom(),
288                                   CLONE_VM | CLONE_FS | CLONE_FILES,
289                                   &tracer_thread_argument);
290  internal_sigprocmask(SIG_SETMASK, &old_sigset, 0);
291  int local_errno = 0;
292  if (internal_iserror(tracer_pid, &local_errno)) {
293    VReport(1, "Failed spawning a tracer thread (errno %d).\n", local_errno);
294    tracer_thread_argument.mutex.Unlock();
295  } else {
296    ScopedSetTracerPID scoped_set_tracer_pid(tracer_pid);
297
298    tracer_thread_argument.mutex.Unlock();
299
300    while (atomic_load(&tracer_thread_argument.done, memory_order_relaxed) == 0)
301      sched_yield();
302
303    for (;;) {
304      uptr waitpid_status = internal_waitpid(tracer_pid, nullptr, __WALL);
305      if (!internal_iserror(waitpid_status, &local_errno))
306        break;
307      if (local_errno == EINTR)
308        continue;
309      VReport(1, "Waiting on the tracer thread failed (errno %d).\n",
310              local_errno);
311      break;
312    }
313  }
314}
315
316tid_t SuspendedThreadsListNetBSD::GetThreadID(uptr index) const {
317  CHECK_LT(index, thread_ids_.size());
318  return thread_ids_[index];
319}
320
321uptr SuspendedThreadsListNetBSD::ThreadCount() const {
322  return thread_ids_.size();
323}
324
325bool SuspendedThreadsListNetBSD::ContainsTid(tid_t thread_id) const {
326  for (uptr i = 0; i < thread_ids_.size(); i++) {
327    if (thread_ids_[i] == thread_id)
328      return true;
329  }
330  return false;
331}
332
333void SuspendedThreadsListNetBSD::Append(tid_t tid) {
334  thread_ids_.push_back(tid);
335}
336
337PtraceRegistersStatus SuspendedThreadsListNetBSD::GetRegistersAndSP(
338    uptr index, uptr *buffer, uptr *sp) const {
339  lwpid_t tid = GetThreadID(index);
340  pid_t ppid = internal_getppid();
341  struct reg regs;
342  int pterrno;
343  bool isErr =
344      internal_iserror(internal_ptrace(PT_GETREGS, ppid, &regs, tid), &pterrno);
345  if (isErr) {
346    VReport(1,
347            "Could not get registers from process %d thread %d (errno %d).\n",
348            ppid, tid, pterrno);
349    return pterrno == ESRCH ? REGISTERS_UNAVAILABLE_FATAL
350                            : REGISTERS_UNAVAILABLE;
351  }
352
353  *sp = PTRACE_REG_SP(&regs);
354  internal_memcpy(buffer, &regs, sizeof(regs));
355
356  return REGISTERS_AVAILABLE;
357}
358
359uptr SuspendedThreadsListNetBSD::RegisterCount() const {
360  return sizeof(struct reg) / sizeof(uptr);
361}
362}  // namespace __sanitizer
363
364#endif
365