| /* -*- Mode: C++; tab-width: 2; indent-tabs-mode: nil; c-basic-offset: 2 -*- */ |
| /* This Source Code Form is subject to the terms of the Mozilla Public |
| * License, v. 2.0. If a copy of the MPL was not distributed with this file, |
| * You can obtain one at http://mozilla.org/MPL/2.0/. */ |
| |
| #include <map> |
| #include <memory> |
| |
| #include <dlfcn.h> |
| #include <errno.h> |
| #include <fcntl.h> |
| #include <setjmp.h> |
| #include <signal.h> |
| #include <poll.h> |
| #include <pthread.h> |
| #include <alloca.h> |
| #include <sys/epoll.h> |
| #include <sys/mman.h> |
| #include <sys/prctl.h> |
| #include <sys/types.h> |
| #include <sys/socket.h> |
| #include <sys/stat.h> |
| #include <sys/syscall.h> |
| #include <vector> |
| |
| #include "mozilla/Alignment.h" |
| #include "mozilla/LinkedList.h" |
| #include "mozilla/TaggedAnonymousMemory.h" |
| #include "Nuwa.h" |
| |
| |
| /* Support for telling Valgrind about the stack pointer changes that |
| Nuwa makes. Without this, Valgrind is unusable in Nuwa child |
| processes due to the large number of false positives resulting from |
| Nuwa's stack pointer changes. See bug 1125091. |
| */ |
| |
| #if defined(MOZ_VALGRIND) |
| # include <valgrind/memcheck.h> |
| #endif |
| |
| #define DEBUG_VALGRIND_ANNOTATIONS 1 |
| |
| /* Call this as soon as possible after a setjmp() that has returned |
| non-locally (that is, it is restoring some previous context). This |
| paints a small area -- half a page -- above SP as containing |
| defined data in any area which is currently marked accessible. |
| |
| Note that in fact there are a few memory references to the stack |
| after the setjmp but before the use of this macro, even when they |
| appear consecutively in the source code. But those accesses all |
| appear to be stores, and since that part of the stack -- before we |
| get to the VALGRIND_MAKE_MEM_DEFINED_IF_ADDRESSABLE client request |
| -- is marked as accessible-but-undefined, Memcheck doesn't |
| complain. Of course, once we get past the client request then even |
| reading from the stack is "safe". |
| |
| VALGRIND_MAKE_MEM_DEFINED_IF_ADDRESSABLE and VALGRIND_PRINTF each |
| require 6 words of stack space. In the worst case, in which the |
| compiler allocates two different pieces of stack, the required |
| extra stack is therefore 12 words, that is, 48 bytes on arm32. |
| */ |
| #if defined(MOZ_VALGRIND) && defined(VALGRIND_MAKE_MEM_DEFINED_IF_ADDRESSABLE) \ |
| && defined(__arm__) && !defined(__aarch64__) |
| # define POST_SETJMP_RESTORE(_who) \ |
| do { \ |
| /* setjmp returned 1 (meaning "restored"). Paint the area */ \ |
| /* immediately above SP as "defined where it is accessible". */ \ |
| register unsigned long int sp; \ |
| __asm__ __volatile__("mov %0, sp" : "=r"(sp)); \ |
| unsigned long int len = 1024*2; \ |
| VALGRIND_MAKE_MEM_DEFINED_IF_ADDRESSABLE(sp, len); \ |
| if (DEBUG_VALGRIND_ANNOTATIONS) { \ |
| VALGRIND_PRINTF("Nuwa: POST_SETJMP_RESTORE: marking [0x%lx, +%ld) as " \ |
| "Defined-if-Addressible, called by %s\n", \ |
| sp, len, (_who)); \ |
| } \ |
| } while (0) |
| #else |
| # define POST_SETJMP_RESTORE(_who) /* */ |
| #endif |
| |
| |
| using namespace mozilla; |
| |
| /** |
| * Provides the wrappers to a selected set of pthread and system-level functions |
| * as the basis for implementing Zygote-like preforking mechanism. |
| */ |
| |
| /** |
| * Real functions for the wrappers. |
| */ |
| extern "C" { |
| #pragma GCC visibility push(default) |
| int __real_pthread_create(pthread_t *thread, |
| const pthread_attr_t *attr, |
| void *(*start_routine) (void *), |
| void *arg); |
| int __real_pthread_key_create(pthread_key_t *key, void (*destructor)(void*)); |
| int __real_pthread_key_delete(pthread_key_t key); |
| pthread_t __real_pthread_self(); |
| int __real_pthread_join(pthread_t thread, void **retval); |
| int __real_epoll_wait(int epfd, |
| struct epoll_event *events, |
| int maxevents, |
| int timeout); |
| int __real_pthread_cond_wait(pthread_cond_t *cond, pthread_mutex_t *mtx); |
| int __real_pthread_cond_timedwait(pthread_cond_t *cond, |
| pthread_mutex_t *mtx, |
| const struct timespec *abstime); |
| int __real_pthread_mutex_lock(pthread_mutex_t *mtx); |
| int __real_pthread_mutex_trylock(pthread_mutex_t *mtx); |
| int __real_poll(struct pollfd *fds, nfds_t nfds, int timeout); |
| int __real_epoll_create(int size); |
| int __real_socketpair(int domain, int type, int protocol, int sv[2]); |
| int __real_pipe2(int __pipedes[2], int flags); |
| int __real_pipe(int __pipedes[2]); |
| int __real_epoll_ctl(int aEpollFd, int aOp, int aFd, struct epoll_event *aEvent); |
| int __real_close(int aFd); |
| #pragma GCC visibility pop |
| } |
| |
| #define REAL(s) __real_##s |
| |
| /** |
| * A Nuwa process is started by preparing. After preparing, it waits |
| * for all threads becoming frozen. Then, it is ready while all |
| * threads are frozen. |
| */ |
| static bool sIsNuwaProcess = false; // This process is a Nuwa process. |
| static bool sIsNuwaChildProcess = false; // This process is spawned from Nuwa. |
| static bool sIsFreezing = false; // Waiting for all threads getting frozen. |
| static bool sNuwaReady = false; // Nuwa process is ready. |
| static bool sNuwaPendingSpawn = false; // Are there any pending spawn requests? |
| static bool sNuwaForking = false; |
| |
| // Fds of transports of top level protocols. |
| static NuwaProtoFdInfo sProtoFdInfos[NUWA_TOPLEVEL_MAX]; |
| static int sProtoFdInfosSize = 0; |
| |
| typedef std::vector<std::pair<pthread_key_t, void *> > |
| TLSInfoList; |
| |
| /** |
| * Return the system's page size |
| */ |
| static size_t getPageSize(void) { |
| #ifdef HAVE_GETPAGESIZE |
| return getpagesize(); |
| #elif defined(_SC_PAGESIZE) |
| return sysconf(_SC_PAGESIZE); |
| #elif defined(PAGE_SIZE) |
| return PAGE_SIZE; |
| #else |
| #warning "Hard-coding page size to 4096 bytes" |
| return 4096 |
| #endif |
| } |
| |
| /** |
| * Use 1 MiB stack size as Android does. |
| */ |
| #ifndef NUWA_STACK_SIZE |
| #define NUWA_STACK_SIZE (1024 * 1024) |
| #endif |
| |
| #define NATIVE_THREAD_NAME_LENGTH 16 |
| |
| typedef struct nuwa_construct { |
| typedef void(*construct_t)(void*); |
| |
| construct_t construct; |
| void *arg; |
| |
| nuwa_construct(construct_t aConstruct, void *aArg) |
| : construct(aConstruct) |
| , arg(aArg) |
| { } |
| |
| nuwa_construct(const nuwa_construct&) = default; |
| nuwa_construct& operator=(const nuwa_construct&) = default; |
| |
| } nuwa_construct_t; |
| |
| struct thread_info : public mozilla::LinkedListElement<thread_info> { |
| pthread_t origThreadID; |
| pthread_t recreatedThreadID; |
| pthread_attr_t threadAttr; |
| jmp_buf jmpEnv; |
| jmp_buf retEnv; |
| |
| int flags; |
| |
| void *(*startupFunc)(void *arg); |
| void *startupArg; |
| |
| // The thread specific function to recreate the new thread. It's executed |
| // after the thread is recreated. |
| |
| std::vector<nuwa_construct_t> *recrFunctions; |
| void addThreadConstructor(const nuwa_construct_t *construct) { |
| if (!recrFunctions) { |
| recrFunctions = new std::vector<nuwa_construct_t>(); |
| } |
| |
| recrFunctions->push_back(*construct); |
| } |
| |
| TLSInfoList tlsInfo; |
| |
| /** |
| * We must ensure that the recreated thread has entered pthread_cond_wait() or |
| * similar functions before proceeding to recreate the next one. Otherwise, if |
| * the next thread depends on the same mutex, it may be used in an incorrect |
| * state. To do this, the main thread must unconditionally acquire the mutex. |
| * The mutex is unconditionally released when the recreated thread enters |
| * pthread_cond_wait(). The recreated thread may have locked the mutex itself |
| * (if the pthread_mutex_trylock succeeded) or another thread may have already |
| * held the lock. If the recreated thread did lock the mutex we must balance |
| * that with another unlock on the main thread, which is signaled by |
| * condMutexNeedsBalancing. |
| */ |
| pthread_mutex_t *condMutex; |
| bool condMutexNeedsBalancing; |
| |
| size_t stackSize; |
| size_t guardSize; |
| void *stk; |
| |
| pid_t origNativeThreadID; |
| pid_t recreatedNativeThreadID; |
| char nativeThreadName[NATIVE_THREAD_NAME_LENGTH]; |
| }; |
| |
| typedef struct thread_info thread_info_t; |
| |
| static thread_info_t *sCurrentRecreatingThread = nullptr; |
| |
| /** |
| * This function runs the custom recreation function registered when calling |
| * NuwaMarkCurrentThread() after thread stack is restored. |
| */ |
| static void |
| RunCustomRecreation() { |
| thread_info_t *tinfo = sCurrentRecreatingThread; |
| if (tinfo->recrFunctions) { |
| for (auto iter = tinfo->recrFunctions->begin(); |
| iter != tinfo->recrFunctions->end(); |
| iter++) { |
| iter->construct(iter->arg); |
| } |
| } |
| } |
| |
| /** |
| * Every thread should be marked as either TINFO_FLAG_NUWA_SUPPORT or |
| * TINFO_FLAG_NUWA_SKIP, or it means a potential error. We force |
| * Gecko code to mark every single thread to make sure there are no accidents |
| * when recreating threads with Nuwa. |
| * |
| * Threads marked as TINFO_FLAG_NUWA_SUPPORT can be checkpointed explicitly, by |
| * calling NuwaCheckpointCurrentThread(), or implicitly when they call into wrapped |
| * functions like pthread_mutex_lock(), epoll_wait(), etc. |
| * TINFO_FLAG_NUWA_EXPLICIT_CHECKPOINT denotes the explicitly checkpointed thread. |
| */ |
| #define TINFO_FLAG_NUWA_SUPPORT 0x1 |
| #define TINFO_FLAG_NUWA_SKIP 0x2 |
| #define TINFO_FLAG_NUWA_EXPLICIT_CHECKPOINT 0x4 |
| |
| static std::vector<nuwa_construct_t> sConstructors; |
| static std::vector<nuwa_construct_t> sFinalConstructors; |
| |
| class TLSKey |
| : public std::pair<pthread_key_t, void (*)(void*)> |
| , public LinkedListElement<TLSKey> |
| { |
| public: |
| TLSKey() {} |
| |
| TLSKey(pthread_key_t aKey, void (*aDestructor)(void*)) |
| : std::pair<pthread_key_t, void (*)(void*)>(aKey, aDestructor) |
| {} |
| |
| static void* operator new(size_t size) { |
| if (sUsed) |
| return ::operator new(size); |
| sUsed = true; |
| return sFirstElement.addr(); |
| } |
| |
| static void operator delete(void* ptr) { |
| if (ptr == sFirstElement.addr()) { |
| sUsed = false; |
| return; |
| } |
| ::operator delete(ptr); |
| } |
| |
| private: |
| static bool sUsed; |
| static AlignedStorage2<TLSKey> sFirstElement; |
| }; |
| |
| bool TLSKey::sUsed = false; |
| AlignedStorage2<TLSKey> TLSKey::sFirstElement; |
| |
| static AutoCleanLinkedList<TLSKey> sTLSKeys; |
| |
| /** |
| * This mutex is used to block the running threads and freeze their contexts. |
| * PrepareNuwaProcess() is the first one to acquire the lock. Further attempts |
| * to acquire this mutex (in the freeze point macros) will block and freeze the |
| * calling thread. |
| */ |
| static pthread_mutex_t sThreadFreezeLock = PTHREAD_MUTEX_INITIALIZER; |
| |
| static thread_info_t sMainThread; |
| static int sThreadCount = 0; |
| static int sThreadFreezeCount = 0; |
| |
| // Bug 1008254: LinkedList's destructor asserts that the list is empty. |
| // But here, on exit, when the global sAllThreads list |
| // is destroyed, it may or may not be empty. Bug 1008254 comment 395 has a log |
| // when there were 8 threads remaining on exit. So this assertion was |
| // intermittently (almost every second time) failing. |
| // As a work-around to avoid this intermittent failure, we clear the list on |
| // exit just before it gets destroyed. This is the only purpose of that |
| // AllThreadsListType subclass. |
| struct AllThreadsListType : public AutoCleanLinkedList<thread_info_t> |
| { |
| ~AllThreadsListType() |
| { |
| if (!isEmpty()) { |
| __android_log_print(ANDROID_LOG_WARN, "Nuwa", |
| "Threads remaining at exit:\n"); |
| int n = 0; |
| for (const thread_info_t* t = getFirst(); t; t = t->getNext()) { |
| __android_log_print(ANDROID_LOG_WARN, "Nuwa", |
| " %.*s (origNativeThreadID=%d recreatedNativeThreadID=%d)\n", |
| NATIVE_THREAD_NAME_LENGTH, |
| t->nativeThreadName, |
| t->origNativeThreadID, |
| t->recreatedNativeThreadID); |
| n++; |
| } |
| __android_log_print(ANDROID_LOG_WARN, "Nuwa", |
| "total: %d outstanding threads. " |
| "Please fix them so they're destroyed before this point!\n", n); |
| __android_log_print(ANDROID_LOG_WARN, "Nuwa", |
| "note: sThreadCount=%d, sThreadFreezeCount=%d\n", |
| sThreadCount, |
| sThreadFreezeCount); |
| } |
| } |
| }; |
| static AllThreadsListType sAllThreads; |
| static AllThreadsListType sExitingThreads; |
| |
| /** |
| * This mutex protects the access to thread info: |
| * sAllThreads, sThreadCount, sThreadFreezeCount, sRecreateVIPCount. |
| */ |
| static pthread_mutex_t sThreadCountLock = PTHREAD_MUTEX_INITIALIZER; |
| /** |
| * This condition variable lets MakeNuwaProcess() wait until all recreated |
| * threads are frozen. |
| */ |
| static pthread_cond_t sThreadChangeCond = PTHREAD_COND_INITIALIZER; |
| |
| /** |
| * This mutex and condition variable is used to serialize the fork requests |
| * from the parent process. |
| */ |
| static pthread_mutex_t sForkLock = PTHREAD_MUTEX_INITIALIZER; |
| static pthread_cond_t sForkWaitCond = PTHREAD_COND_INITIALIZER; |
| |
| /** |
| * sForkWaitCondChanged will be reset to false on the IPC thread before |
| * and will be changed to true on the main thread to indicate that the condition |
| * that the IPC thread is waiting for has already changed. |
| */ |
| static bool sForkWaitCondChanged = false; |
| |
| /** |
| * This mutex protects the access to sTLSKeys, which keeps track of existing |
| * TLS Keys. |
| */ |
| static pthread_mutex_t sTLSKeyLock = PTHREAD_ERRORCHECK_MUTEX_INITIALIZER; |
| static int sThreadSkipCount = 0; |
| |
| static thread_info_t * |
| GetThreadInfoInner(pthread_t threadID) { |
| for (thread_info_t *tinfo = sAllThreads.getFirst(); |
| tinfo; |
| tinfo = tinfo->getNext()) { |
| if (pthread_equal(tinfo->origThreadID, threadID)) { |
| return tinfo; |
| } |
| } |
| |
| for (thread_info_t *tinfo = sExitingThreads.getFirst(); |
| tinfo; |
| tinfo = tinfo->getNext()) { |
| if (pthread_equal(tinfo->origThreadID, threadID)) { |
| return tinfo; |
| } |
| } |
| |
| return nullptr; |
| } |
| |
| /** |
| * Get thread info using the specified thread ID. |
| * |
| * @return thread_info_t which has threadID == specified threadID |
| */ |
| static thread_info_t * |
| GetThreadInfo(pthread_t threadID) { |
| REAL(pthread_mutex_lock)(&sThreadCountLock); |
| |
| thread_info_t *tinfo = GetThreadInfoInner(threadID); |
| |
| pthread_mutex_unlock(&sThreadCountLock); |
| return tinfo; |
| } |
| |
| #if !defined(HAVE_THREAD_TLS_KEYWORD) |
| /** |
| * Get thread info of the current thread. |
| * |
| * @return thread_info_t for the current thread. |
| */ |
| static thread_info_t * |
| GetCurThreadInfo() { |
| pthread_t threadID = REAL(pthread_self)(); |
| pthread_t thread_info_t::*threadIDptr = |
| (sIsNuwaProcess ? |
| &thread_info_t::origThreadID : |
| &thread_info_t::recreatedThreadID); |
| |
| REAL(pthread_mutex_lock)(&sThreadCountLock); |
| thread_info_t *tinfo; |
| for (tinfo = sAllThreads.getFirst(); |
| tinfo; |
| tinfo = tinfo->getNext()) { |
| if (pthread_equal(tinfo->*threadIDptr, threadID)) { |
| break; |
| } |
| } |
| pthread_mutex_unlock(&sThreadCountLock); |
| return tinfo; |
| } |
| #define CUR_THREAD_INFO GetCurThreadInfo() |
| #define SET_THREAD_INFO(x) /* Nothing to do. */ |
| #else |
| // Is not nullptr only for threads created by pthread_create() in an Nuwa process. |
| // It is always nullptr for the main thread. |
| static __thread thread_info_t *sCurThreadInfo = nullptr; |
| #define CUR_THREAD_INFO sCurThreadInfo |
| #define SET_THREAD_INFO(x) do { sCurThreadInfo = (x); } while(0) |
| #endif // HAVE_THREAD_TLS_KEYWORD |
| |
| /* |
| * Track all epoll fds and handling events. |
| */ |
| class EpollManager { |
| public: |
| class EpollInfo { |
| public: |
| typedef struct epoll_event Events; |
| typedef std::map<int, Events> EpollEventsMap; |
| typedef EpollEventsMap::iterator iterator; |
| typedef EpollEventsMap::const_iterator const_iterator; |
| |
| EpollInfo(): mBackSize(0) {} |
| EpollInfo(int aBackSize): mBackSize(aBackSize) {} |
| EpollInfo(const EpollInfo &aOther): mEvents(aOther.mEvents) |
| , mBackSize(aOther.mBackSize) { |
| } |
| ~EpollInfo() { |
| mEvents.clear(); |
| } |
| |
| void AddEvents(int aFd, Events &aEvents) { |
| std::pair<iterator, bool> pair = |
| mEvents.insert(std::make_pair(aFd, aEvents)); |
| if (!pair.second) { |
| abort(); |
| } |
| } |
| |
| void RemoveEvents(int aFd) { |
| if (!mEvents.erase(aFd)) { |
| abort(); |
| } |
| } |
| |
| void ModifyEvents(int aFd, Events &aEvents) { |
| iterator it = mEvents.find(aFd); |
| if (it == mEvents.end()) { |
| abort(); |
| } |
| it->second = aEvents; |
| } |
| |
| const Events &FindEvents(int aFd) const { |
| const_iterator it = mEvents.find(aFd); |
| if (it == mEvents.end()) { |
| abort(); |
| } |
| return it->second; |
| } |
| |
| int Size() const { return mEvents.size(); } |
| |
| // Iterator with values of <fd, Events> pairs. |
| const_iterator begin() const { return mEvents.begin(); } |
| const_iterator end() const { return mEvents.end(); } |
| |
| int BackSize() const { return mBackSize; } |
| |
| private: |
| EpollEventsMap mEvents; |
| int mBackSize; |
| |
| friend class EpollManager; |
| }; |
| |
| typedef std::map<int, EpollInfo> EpollInfoMap; |
| typedef EpollInfoMap::iterator iterator; |
| typedef EpollInfoMap::const_iterator const_iterator; |
| |
| public: |
| void AddEpollInfo(int aEpollFd, int aBackSize) { |
| EpollInfo *oldinfo = FindEpollInfo(aEpollFd); |
| if (oldinfo != nullptr) { |
| abort(); |
| } |
| mEpollFdsInfo[aEpollFd] = EpollInfo(aBackSize); |
| } |
| |
| EpollInfo *FindEpollInfo(int aEpollFd) { |
| iterator it = mEpollFdsInfo.find(aEpollFd); |
| if (it == mEpollFdsInfo.end()) { |
| return nullptr; |
| } |
| return &it->second; |
| } |
| |
| void RemoveEpollInfo(int aEpollFd) { |
| if (!mEpollFdsInfo.erase(aEpollFd)) { |
| abort(); |
| } |
| } |
| |
| int Size() const { return mEpollFdsInfo.size(); } |
| |
| // Iterator of <epollfd, EpollInfo> pairs. |
| const_iterator begin() const { return mEpollFdsInfo.begin(); } |
| const_iterator end() const { return mEpollFdsInfo.end(); } |
| |
| static EpollManager *Singleton() { |
| if (!sInstance) { |
| sInstance = new EpollManager(); |
| } |
| return sInstance; |
| } |
| |
| static void Shutdown() { |
| if (!sInstance) { |
| abort(); |
| } |
| |
| delete sInstance; |
| sInstance = nullptr; |
| } |
| |
| private: |
| static EpollManager *sInstance; |
| ~EpollManager() { |
| mEpollFdsInfo.clear(); |
| } |
| |
| EpollInfoMap mEpollFdsInfo; |
| |
| EpollManager() {} |
| }; |
| |
| EpollManager* EpollManager::sInstance; |
| |
| static thread_info_t * |
| thread_info_new(void) { |
| /* link tinfo to sAllThreads */ |
| thread_info_t *tinfo = new thread_info_t(); |
| tinfo->flags = 0; |
| tinfo->recrFunctions = nullptr; |
| tinfo->recreatedThreadID = 0; |
| tinfo->recreatedNativeThreadID = 0; |
| tinfo->condMutex = nullptr; |
| tinfo->condMutexNeedsBalancing = false; |
| |
| // Default stack and guard size. |
| tinfo->stackSize = NUWA_STACK_SIZE; |
| tinfo->guardSize = getPageSize(); |
| |
| REAL(pthread_mutex_lock)(&sThreadCountLock); |
| // Insert to the tail. |
| sAllThreads.insertBack(tinfo); |
| |
| sThreadCount++; |
| pthread_cond_signal(&sThreadChangeCond); |
| pthread_mutex_unlock(&sThreadCountLock); |
| |
| return tinfo; |
| } |
| |
| static void |
| thread_attr_init(thread_info_t *tinfo, const pthread_attr_t *tattr) |
| { |
| pthread_attr_init(&tinfo->threadAttr); |
| |
| if (tattr) { |
| // Override default thread stack and guard size with tattr. |
| pthread_attr_getstacksize(tattr, &tinfo->stackSize); |
| pthread_attr_getguardsize(tattr, &tinfo->guardSize); |
| |
| size_t pageSize = getPageSize(); |
| |
| tinfo->stackSize = (tinfo->stackSize + pageSize - 1) % pageSize; |
| tinfo->guardSize = (tinfo->guardSize + pageSize - 1) % pageSize; |
| |
| int detachState = 0; |
| pthread_attr_getdetachstate(tattr, &detachState); |
| pthread_attr_setdetachstate(&tinfo->threadAttr, detachState); |
| } |
| |
| tinfo->stk = MozTaggedAnonymousMmap(nullptr, |
| tinfo->stackSize + tinfo->guardSize, |
| PROT_READ | PROT_WRITE, |
| MAP_PRIVATE | MAP_ANONYMOUS, |
| /* fd */ -1, |
| /* offset */ 0, |
| "nuwa-thread-stack"); |
| |
| // Add protection to stack overflow: mprotect() stack top (the page at the |
| // lowest address) so we crash instead of corrupt other content that is |
| // malloc()'d. |
| mprotect(tinfo->stk, tinfo->guardSize, PROT_NONE); |
| |
| pthread_attr_setstack(&tinfo->threadAttr, |
| (char*)tinfo->stk + tinfo->guardSize, |
| tinfo->stackSize); |
| } |
| |
| static void |
| thread_info_cleanup(void *arg) { |
| if (sNuwaForking) { |
| // We shouldn't have any thread exiting when we are forking a new process. |
| abort(); |
| } |
| |
| thread_info_t *tinfo = (thread_info_t *)arg; |
| pthread_attr_destroy(&tinfo->threadAttr); |
| |
| munmap(tinfo->stk, tinfo->stackSize + tinfo->guardSize); |
| |
| REAL(pthread_mutex_lock)(&sThreadCountLock); |
| /* unlink tinfo from sAllThreads */ |
| tinfo->remove(); |
| pthread_mutex_unlock(&sThreadCountLock); |
| |
| if (tinfo->recrFunctions) { |
| delete tinfo->recrFunctions; |
| } |
| |
| // while sThreadCountLock is held, since delete calls wrapped functions |
| // which try to lock sThreadCountLock. This results in deadlock. And we |
| // need to delete |tinfo| before decreasing sThreadCount, so Nuwa won't |
| // get ready before tinfo is cleaned. |
| delete tinfo; |
| |
| REAL(pthread_mutex_lock)(&sThreadCountLock); |
| sThreadCount--; |
| pthread_cond_signal(&sThreadChangeCond); |
| pthread_mutex_unlock(&sThreadCountLock); |
| } |
| |
| static void |
| EnsureThreadExited(thread_info_t *tinfo) { |
| pid_t thread = sIsNuwaProcess ? tinfo->origNativeThreadID |
| : tinfo->recreatedNativeThreadID; |
| // Wait until the target thread exits. Note that we use tgkill() instead of |
| // pthread_kill() because of: |
| // 1. Use after free inside pthread implementation. |
| // 2. Race due to pthread_t reuse when a thread is created. |
| while (!syscall(__NR_tgkill, getpid(), thread, 0)) { |
| sched_yield(); |
| } |
| } |
| |
| static void* |
| safe_thread_info_cleanup(void *arg) |
| { |
| thread_info_t *tinfo = (thread_info_t *)arg; |
| |
| // We need to ensure the thread is really dead before cleaning up tinfo. |
| EnsureThreadExited(tinfo); |
| thread_info_cleanup(tinfo); |
| |
| return nullptr; |
| } |
| |
| static void |
| MaybeCleanUpDetachedThread(thread_info_t *tinfo) |
| { |
| if (pthread_getattr_np(REAL(pthread_self()), &tinfo->threadAttr)) { |
| return; |
| } |
| |
| int detachState = 0; |
| if (pthread_attr_getdetachstate(&tinfo->threadAttr, &detachState) || |
| detachState == PTHREAD_CREATE_JOINABLE) { |
| // We only clean up tinfo of a detached thread. A joinable thread |
| // will be cleaned up in __wrap_pthread_join(). |
| return; |
| } |
| |
| // Create a detached thread to safely clean up the current thread. |
| pthread_t thread; |
| if (!REAL(pthread_create)(&thread, |
| nullptr, |
| safe_thread_info_cleanup, |
| tinfo)) { |
| pthread_detach(thread); |
| } |
| } |
| |
| static void |
| invalidate_thread_info(void *arg) { |
| REAL(pthread_mutex_lock)(&sThreadCountLock); |
| |
| // Unlink tinfo from sAllThreads to make it invisible from CUR_THREAD_INFO so |
| // it won't be misused by a newly created thread. |
| thread_info_t *tinfo = (thread_info_t*) arg; |
| tinfo->remove(); |
| sExitingThreads.insertBack(tinfo); |
| |
| pthread_mutex_unlock(&sThreadCountLock); |
| |
| MaybeCleanUpDetachedThread(tinfo); |
| } |
| |
| static void * |
| _thread_create_startup(void *arg) { |
| thread_info_t *tinfo = (thread_info_t *)arg; |
| void *r; |
| |
| // Save thread info; especially, stackaddr & stacksize. |
| // Reuse the stack in the new thread. |
| pthread_getattr_np(REAL(pthread_self)(), &tinfo->threadAttr); |
| |
| SET_THREAD_INFO(tinfo); |
| tinfo->origThreadID = REAL(pthread_self)(); |
| tinfo->origNativeThreadID = gettid(); |
| |
| r = tinfo->startupFunc(tinfo->startupArg); |
| |
| return r; |
| } |
| |
| // reserve STACK_RESERVED_SZ * 4 bytes for thread_recreate_startup(). |
| #define STACK_RESERVED_SZ 96 |
| #define STACK_SENTINEL(v) ((v)[0]) |
| #define STACK_SENTINEL_VALUE(v) ((uint32_t)(v) ^ 0xdeadbeef) |
| |
| static void * |
| thread_create_startup(void *arg) { |
| /* |
| * Dark Art!! Never try to do the same unless you are ABSOLUTELY sure of |
| * what you are doing! |
| * |
| * This function is here for reserving stack space before calling |
| * _thread_create_startup(). see also thread_create_startup(); |
| */ |
| void *r; |
| volatile uint32_t reserved[STACK_RESERVED_SZ]; |
| |
| // Reserve stack space. |
| STACK_SENTINEL(reserved) = STACK_SENTINEL_VALUE(reserved); |
| |
| r = _thread_create_startup(arg); |
| |
| // Check if the reservation is enough. |
| if (STACK_SENTINEL(reserved) != STACK_SENTINEL_VALUE(reserved)) { |
| abort(); // Did not reserve enough stack space. |
| } |
| |
| // Get tinfo before invalidating it. Note that we cannot use arg directly here |
| // because thread_recreate_startup() also runs on the same stack area and |
| // could corrupt the value. |
| thread_info_t *tinfo = CUR_THREAD_INFO; |
| invalidate_thread_info(tinfo); |
| |
| if (!sIsNuwaProcess) { |
| longjmp(tinfo->retEnv, 1); |
| |
| // Never go here! |
| abort(); |
| } |
| |
| return r; |
| } |
| |
| extern "C" MFBT_API int |
| __wrap_pthread_create(pthread_t *thread, |
| const pthread_attr_t *attr, |
| void *(*start_routine) (void *), |
| void *arg) { |
| if (!sIsNuwaProcess) { |
| return REAL(pthread_create)(thread, attr, start_routine, arg); |
| } |
| |
| thread_info_t *tinfo = thread_info_new(); |
| thread_attr_init(tinfo, attr); |
| tinfo->startupFunc = start_routine; |
| tinfo->startupArg = arg; |
| |
| int rv = REAL(pthread_create)(thread, |
| &tinfo->threadAttr, |
| thread_create_startup, |
| tinfo); |
| if (rv) { |
| thread_info_cleanup(tinfo); |
| } else { |
| tinfo->origThreadID = *thread; |
| } |
| |
| return rv; |
| } |
| |
| // TLS related |
| |
| /** |
| * Iterates over the existing TLS keys and store the TLS data for the current |
| * thread in tinfo. |
| */ |
| static void |
| SaveTLSInfo(thread_info_t *tinfo) { |
| MOZ_RELEASE_ASSERT(REAL(pthread_mutex_lock)(&sTLSKeyLock) == 0); |
| tinfo->tlsInfo.clear(); |
| for (TLSKey *it = sTLSKeys.getFirst(); it != nullptr; it = it->getNext()) { |
| void *value = pthread_getspecific(it->first); |
| if (value == nullptr) { |
| continue; |
| } |
| |
| pthread_key_t key = it->first; |
| tinfo->tlsInfo.push_back(TLSInfoList::value_type(key, value)); |
| } |
| MOZ_RELEASE_ASSERT(pthread_mutex_unlock(&sTLSKeyLock) == 0); |
| } |
| |
| /** |
| * Restores the TLS data for the current thread from tinfo. |
| */ |
| static void |
| RestoreTLSInfo(thread_info_t *tinfo) { |
| for (TLSInfoList::const_iterator it = tinfo->tlsInfo.begin(); |
| it != tinfo->tlsInfo.end(); |
| it++) { |
| pthread_key_t key = it->first; |
| const void *value = it->second; |
| if (pthread_setspecific(key, value)) { |
| abort(); |
| } |
| } |
| |
| SET_THREAD_INFO(tinfo); |
| tinfo->recreatedThreadID = REAL(pthread_self)(); |
| tinfo->recreatedNativeThreadID = gettid(); |
| } |
| |
| extern "C" MFBT_API int |
| __wrap_pthread_key_create(pthread_key_t *key, void (*destructor)(void*)) { |
| int rv = REAL(pthread_key_create)(key, destructor); |
| if (rv != 0) { |
| return rv; |
| } |
| MOZ_RELEASE_ASSERT(REAL(pthread_mutex_lock)(&sTLSKeyLock) == 0); |
| sTLSKeys.insertBack(new TLSKey(*key, destructor)); |
| MOZ_RELEASE_ASSERT(pthread_mutex_unlock(&sTLSKeyLock) == 0); |
| return 0; |
| } |
| |
| extern "C" MFBT_API int |
| __wrap_pthread_key_delete(pthread_key_t key) { |
| // Don't call pthread_key_delete() for Nuwa-forked processes because bionic's |
| // pthread_key_delete() implementation can touch the thread stack that was |
| // freed in thread_info_cleanup(). |
| int rv = sIsNuwaChildProcess ? |
| 0 : REAL(pthread_key_delete)(key); |
| if (rv != 0) { |
| return rv; |
| } |
| MOZ_RELEASE_ASSERT(REAL(pthread_mutex_lock)(&sTLSKeyLock) == 0); |
| for (TLSKey *it = sTLSKeys.getFirst(); it != nullptr; it = it->getNext()) { |
| if (key == it->first) { |
| delete it; |
| break; |
| } |
| } |
| MOZ_RELEASE_ASSERT(pthread_mutex_unlock(&sTLSKeyLock) == 0); |
| return 0; |
| } |
| |
| extern "C" MFBT_API pthread_t |
| __wrap_pthread_self() { |
| thread_info_t *tinfo = CUR_THREAD_INFO; |
| if (tinfo) { |
| // For recreated thread, masquerade as the original thread in the Nuwa |
| // process. |
| return tinfo->origThreadID; |
| } |
| return REAL(pthread_self)(); |
| } |
| |
| extern "C" MFBT_API int |
| __wrap_pthread_join(pthread_t thread, void **retval) { |
| thread_info_t *tinfo = GetThreadInfo(thread); |
| if (tinfo == nullptr) { |
| return REAL(pthread_join)(thread, retval); |
| } |
| |
| pthread_t thread_info_t::*threadIDptr = |
| (sIsNuwaProcess ? |
| &thread_info_t::origThreadID : |
| &thread_info_t::recreatedThreadID); |
| |
| // pthread_join() uses the origThreadID or recreatedThreadID depending on |
| // whether we are in Nuwa or forked processes. |
| int rc = REAL(pthread_join)(tinfo->*threadIDptr, retval); |
| |
| // Before Android L, bionic wakes up the caller of pthread_join() with |
| // pthread_cond_signal() so the thread can still use the stack for some while. |
| // Call safe_thread_info_cleanup() to destroy tinfo after the thread really |
| // exits. |
| safe_thread_info_cleanup(tinfo); |
| |
| return rc; |
| } |
| |
| /** |
| * The following are used to synchronize between the main thread and the |
| * thread being recreated. The main thread will wait until the thread is woken |
| * up from the freeze points or the blocking intercepted functions and then |
| * proceed to recreate the next frozen thread. |
| * |
| * In thread recreation, the main thread recreates the frozen threads one by |
| * one. The recreated threads will be "gated" until the main thread "opens the |
| * gate" to let them run freely as if they were created from scratch. The VIP |
| * threads gets the chance to run first after their thread stacks are recreated |
| * (using longjmp()) so they can adjust their contexts to a valid, consistent |
| * state. The threads frozen waiting for pthread condition variables are VIP |
| * threads. After woken up they need to run first to make the associated mutex |
| * in a valid state to maintain the semantics of the intercepted function calls |
| * (like pthread_cond_wait()). |
| */ |
| |
| // Used to synchronize the main thread and the thread being recreated so that |
| // only one thread is allowed to be recreated at a time. |
| static pthread_mutex_t sRecreateWaitLock = PTHREAD_MUTEX_INITIALIZER; |
| // Used to block recreated threads until the main thread "opens the gate". |
| static pthread_mutex_t sRecreateGateLock = PTHREAD_MUTEX_INITIALIZER; |
| // Used to block the main thread from "opening the gate" until all VIP threads |
| // have been recreated. |
| static pthread_mutex_t sRecreateVIPGateLock = PTHREAD_MUTEX_INITIALIZER; |
| static pthread_cond_t sRecreateVIPCond = PTHREAD_COND_INITIALIZER; |
| static int sRecreateVIPCount = 0; |
| static int sRecreateGatePassed = 0; |
| |
| /** |
| * Thread recreation macros. |
| * |
| * The following macros are used in the forked process to synchronize and |
| * control the progress of thread recreation. |
| * |
| * 1. RECREATE_START() is first called in the beginning of thread |
| * recreation to set sRecreateWaitLock and sRecreateGateLock in locked |
| * state. |
| * 2. For each frozen thread: |
| * 2.1. RECREATE_BEFORE() to set the thread being recreated. |
| * 2.2. thread_recreate() to recreate the frozen thread. |
| * 2.3. Main thread calls RECREATE_WAIT() to wait on sRecreateWaitLock until |
| * the thread is recreated from the freeze point and calls |
| * RECREATE_CONTINUE() to release sRecreateWaitLock. |
| * 2.3. Non-VIP threads are blocked on RECREATE_GATE(). VIP threads calls |
| * RECREATE_PASS_VIP() to mark that a VIP thread is successfully |
| * recreated and then is blocked by calling RECREATE_GATE_VIP(). |
| * 3. RECREATE_WAIT_ALL_VIP() to wait until all VIP threads passed, that is, |
| * VIP threads already has their contexts (mainly pthread mutex) in a valid |
| * state. |
| * 4. RECREATE_OPEN_GATE() to unblock threads blocked by sRecreateGateLock. |
| * 5. RECREATE_FINISH() to complete thread recreation. |
| */ |
| #define RECREATE_START() \ |
| do { \ |
| REAL(pthread_mutex_lock)(&sRecreateWaitLock); \ |
| REAL(pthread_mutex_lock)(&sRecreateGateLock); \ |
| } while(0) |
| #define RECREATE_BEFORE(info) do { sCurrentRecreatingThread = info; } while(0) |
| #define RECREATE_WAIT() REAL(pthread_mutex_lock)(&sRecreateWaitLock) |
| #define RECREATE_CONTINUE() do { \ |
| RunCustomRecreation(); \ |
| pthread_mutex_unlock(&sRecreateWaitLock); \ |
| } while(0) |
| #define RECREATE_FINISH() pthread_mutex_unlock(&sRecreateWaitLock) |
| #define RECREATE_GATE() \ |
| do { \ |
| REAL(pthread_mutex_lock)(&sRecreateGateLock); \ |
| sRecreateGatePassed++; \ |
| pthread_mutex_unlock(&sRecreateGateLock); \ |
| } while(0) |
| #define RECREATE_OPEN_GATE() pthread_mutex_unlock(&sRecreateGateLock) |
| #define RECREATE_GATE_VIP() \ |
| do { \ |
| REAL(pthread_mutex_lock)(&sRecreateGateLock); \ |
| pthread_mutex_unlock(&sRecreateGateLock); \ |
| } while(0) |
| #define RECREATE_PASS_VIP() \ |
| do { \ |
| REAL(pthread_mutex_lock)(&sRecreateVIPGateLock); \ |
| sRecreateGatePassed++; \ |
| pthread_cond_signal(&sRecreateVIPCond); \ |
| pthread_mutex_unlock(&sRecreateVIPGateLock); \ |
| } while(0) |
| #define RECREATE_WAIT_ALL_VIP() \ |
| do { \ |
| REAL(pthread_mutex_lock)(&sRecreateVIPGateLock); \ |
| while(sRecreateGatePassed < sRecreateVIPCount) { \ |
| REAL(pthread_cond_wait)(&sRecreateVIPCond, \ |
| &sRecreateVIPGateLock); \ |
| } \ |
| pthread_mutex_unlock(&sRecreateVIPGateLock); \ |
| } while(0) |
| |
| /** |
| * Thread freeze points. Note that the freeze points are implemented as macros |
| * so as not to garble the content of the stack after setjmp(). |
| * |
| * In the nuwa process, when a thread supporting nuwa calls a wrapper |
| * function, freeze point 1 setjmp()s to save the state. We only allow the |
| * thread to be frozen in the wrapper functions. If thread freezing is not |
| * enabled yet, the wrapper functions act like their wrapped counterparts, |
| * except for the extra actions in the freeze points. If thread freezing is |
| * enabled, the thread will be frozen by calling one of the wrapper functions. |
| * The threads can be frozen in any of the following points: |
| * |
| * 1) Freeze point 1: this is the point where we setjmp() in the nuwa process |
| * and longjmp() in the spawned process. If freezing is enabled, then the |
| * current thread blocks by acquiring an already locked mutex, |
| * sThreadFreezeLock. |
| * 2) The wrapped function: the function that might block waiting for some |
| * resource or condition. |
| * 3) Freeze point 2: blocks the current thread by acquiring sThreadFreezeLock. |
| * If freezing is not enabled then revert the counter change in freeze |
| * point 1. |
| * |
| * Note: the purpose of the '(void) variable;' statements is to avoid |
| * -Wunused-but-set-variable warnings. |
| */ |
| #define THREAD_FREEZE_POINT1() \ |
| bool freezeCountChg = false; \ |
| bool recreated = false; \ |
| (void) recreated; \ |
| volatile bool freezePoint2 = false; \ |
| (void) freezePoint2; \ |
| thread_info_t *tinfo; \ |
| if (sIsNuwaProcess && \ |
| (tinfo = CUR_THREAD_INFO) && \ |
| (tinfo->flags & TINFO_FLAG_NUWA_SUPPORT) && \ |
| !(tinfo->flags & TINFO_FLAG_NUWA_EXPLICIT_CHECKPOINT)) { \ |
| if (!setjmp(tinfo->jmpEnv)) { \ |
| REAL(pthread_mutex_lock)(&sThreadCountLock); \ |
| SaveTLSInfo(tinfo); \ |
| sThreadFreezeCount++; \ |
| freezeCountChg = true; \ |
| pthread_cond_signal(&sThreadChangeCond); \ |
| pthread_mutex_unlock(&sThreadCountLock); \ |
| \ |
| if (sIsFreezing) { \ |
| REAL(pthread_mutex_lock)(&sThreadFreezeLock); \ |
| /* Never return from the pthread_mutex_lock() call. */ \ |
| abort(); \ |
| } \ |
| } else { \ |
| POST_SETJMP_RESTORE("THREAD_FREEZE_POINT1"); \ |
| RECREATE_CONTINUE(); \ |
| RECREATE_GATE(); \ |
| freezeCountChg = false; \ |
| recreated = true; \ |
| } \ |
| } |
| |
| #define THREAD_FREEZE_POINT1_VIP() \ |
| bool freezeCountChg = false; \ |
| bool recreated = false; \ |
| volatile bool freezePoint1 = false; \ |
| volatile bool freezePoint2 = false; \ |
| thread_info_t *tinfo; \ |
| if (sIsNuwaProcess && \ |
| (tinfo = CUR_THREAD_INFO) && \ |
| (tinfo->flags & TINFO_FLAG_NUWA_SUPPORT) && \ |
| !(tinfo->flags & TINFO_FLAG_NUWA_EXPLICIT_CHECKPOINT)) { \ |
| if (!setjmp(tinfo->jmpEnv)) { \ |
| REAL(pthread_mutex_lock)(&sThreadCountLock); \ |
| SaveTLSInfo(tinfo); \ |
| sThreadFreezeCount++; \ |
| sRecreateVIPCount++; \ |
| freezeCountChg = true; \ |
| pthread_cond_signal(&sThreadChangeCond); \ |
| pthread_mutex_unlock(&sThreadCountLock); \ |
| \ |
| if (sIsFreezing) { \ |
| freezePoint1 = true; \ |
| REAL(pthread_mutex_lock)(&sThreadFreezeLock); \ |
| /* Never return from the pthread_mutex_lock() call. */ \ |
| abort(); \ |
| } \ |
| } else { \ |
| POST_SETJMP_RESTORE("THREAD_FREEZE_POINT1_VIP"); \ |
| freezeCountChg = false; \ |
| recreated = true; \ |
| } \ |
| } |
| |
| #define THREAD_FREEZE_POINT2() \ |
| if (freezeCountChg) { \ |
| REAL(pthread_mutex_lock)(&sThreadCountLock); \ |
| if (sNuwaReady && sIsNuwaProcess) { \ |
| pthread_mutex_unlock(&sThreadCountLock); \ |
| freezePoint2 = true; \ |
| REAL(pthread_mutex_lock)(&sThreadFreezeLock); \ |
| /* Never return from the pthread_mutex_lock() call. */ \ |
| abort(); \ |
| } \ |
| sThreadFreezeCount--; \ |
| pthread_cond_signal(&sThreadChangeCond); \ |
| pthread_mutex_unlock(&sThreadCountLock); \ |
| } |
| |
| #define THREAD_FREEZE_POINT2_VIP() \ |
| if (freezeCountChg) { \ |
| REAL(pthread_mutex_lock)(&sThreadCountLock); \ |
| if (sNuwaReady && sIsNuwaProcess) { \ |
| pthread_mutex_unlock(&sThreadCountLock); \ |
| freezePoint2 = true; \ |
| REAL(pthread_mutex_lock)(&sThreadFreezeLock); \ |
| /* Never return from the pthread_mutex_lock() call. */ \ |
| abort(); \ |
| } \ |
| sThreadFreezeCount--; \ |
| sRecreateVIPCount--; \ |
| pthread_cond_signal(&sThreadChangeCond); \ |
| pthread_mutex_unlock(&sThreadCountLock); \ |
| } |
| |
| /** |
| * Wrapping the blocking functions: epoll_wait(), poll(), pthread_mutex_lock(), |
| * pthread_cond_wait() and pthread_cond_timedwait(): |
| * |
| * These functions are wrapped by the above freeze point macros. Once a new |
| * process is forked, the recreated thread will be blocked in one of the wrapper |
| * functions. When recreating the thread, we longjmp() to |
| * THREAD_FREEZE_POINT1() to recover the thread stack. Care must be taken to |
| * maintain the semantics of the wrapped function: |
| * |
| * - epoll_wait() and poll(): just retry the function. |
| * - pthread_mutex_lock(): don't lock if frozen at freeze point 2 (lock is |
| * already acquired). |
| * - pthread_cond_wait() and pthread_cond_timedwait(): if the thread is frozen |
| * waiting the condition variable, the mutex is already released, we need to |
| * reacquire the mutex before calling the wrapped function again so the mutex |
| * will be in a valid state. |
| */ |
| |
| extern "C" MFBT_API int |
| __wrap_epoll_wait(int epfd, |
| struct epoll_event *events, |
| int maxevents, |
| int timeout) { |
| int rv; |
| |
| THREAD_FREEZE_POINT1(); |
| rv = REAL(epoll_wait)(epfd, events, maxevents, timeout); |
| THREAD_FREEZE_POINT2(); |
| |
| return rv; |
| } |
| |
| extern "C" MFBT_API int |
| __wrap_pthread_cond_wait(pthread_cond_t *cond, |
| pthread_mutex_t *mtx) { |
| int rv = 0; |
| |
| THREAD_FREEZE_POINT1_VIP(); |
| if (freezePoint2) { |
| RECREATE_CONTINUE(); |
| RECREATE_PASS_VIP(); |
| RECREATE_GATE_VIP(); |
| return rv; |
| } |
| if (recreated && mtx) { |
| if (!freezePoint1) { |
| tinfo->condMutex = mtx; |
| // The thread was frozen in pthread_cond_wait() after releasing mtx in the |
| // Nuwa process. In recreating this thread, We failed to reacquire mtx |
| // with the pthread_mutex_trylock() call, that is, mtx was acquired by |
| // another thread. Because of this, we need the main thread's help to |
| // reacquire mtx so that it will be in a valid state. |
| if (!pthread_mutex_trylock(mtx)) { |
| tinfo->condMutexNeedsBalancing = true; |
| } |
| } |
| RECREATE_CONTINUE(); |
| RECREATE_PASS_VIP(); |
| } |
| rv = REAL(pthread_cond_wait)(cond, mtx); |
| if (recreated && mtx) { |
| // We have reacquired mtx. The main thread also wants to acquire mtx to |
| // synchronize with us. If the main thread didn't get a chance to acquire |
| // mtx let it do that now. The main thread clears condMutex after acquiring |
| // it to signal us. |
| if (tinfo->condMutex) { |
| // We need mtx to end up locked, so tell the main thread not to unlock |
| // mtx after it locks it. |
| tinfo->condMutexNeedsBalancing = false; |
| pthread_mutex_unlock(mtx); |
| } |
| // We still need to be gated as not to acquire another mutex associated with |
| // another VIP thread and interfere with it. |
| RECREATE_GATE_VIP(); |
| } |
| THREAD_FREEZE_POINT2_VIP(); |
| |
| return rv; |
| } |
| |
| extern "C" MFBT_API int |
| __wrap_pthread_cond_timedwait(pthread_cond_t *cond, |
| pthread_mutex_t *mtx, |
| const struct timespec *abstime) { |
| int rv = 0; |
| |
| THREAD_FREEZE_POINT1_VIP(); |
| if (freezePoint2) { |
| RECREATE_CONTINUE(); |
| RECREATE_PASS_VIP(); |
| RECREATE_GATE_VIP(); |
| return rv; |
| } |
| if (recreated && mtx) { |
| if (!freezePoint1) { |
| tinfo->condMutex = mtx; |
| if (!pthread_mutex_trylock(mtx)) { |
| tinfo->condMutexNeedsBalancing = true; |
| } |
| } |
| RECREATE_CONTINUE(); |
| RECREATE_PASS_VIP(); |
| } |
| rv = REAL(pthread_cond_timedwait)(cond, mtx, abstime); |
| if (recreated && mtx) { |
| if (tinfo->condMutex) { |
| tinfo->condMutexNeedsBalancing = false; |
| pthread_mutex_unlock(mtx); |
| } |
| RECREATE_GATE_VIP(); |
| } |
| THREAD_FREEZE_POINT2_VIP(); |
| |
| return rv; |
| } |
| |
| |
| extern "C" MFBT_API int |
| __wrap_pthread_mutex_trylock(pthread_mutex_t *mtx) { |
| int rv = 0; |
| |
| THREAD_FREEZE_POINT1(); |
| if (freezePoint2) { |
| return rv; |
| } |
| rv = REAL(pthread_mutex_trylock)(mtx); |
| THREAD_FREEZE_POINT2(); |
| |
| return rv; |
| } |
| |
| extern "C" MFBT_API int |
| __wrap_pthread_mutex_lock(pthread_mutex_t *mtx) { |
| int rv = 0; |
| |
| THREAD_FREEZE_POINT1(); |
| if (freezePoint2) { |
| return rv; |
| } |
| rv = REAL(pthread_mutex_lock)(mtx); |
| THREAD_FREEZE_POINT2(); |
| |
| return rv; |
| } |
| |
| extern "C" MFBT_API int |
| __wrap_poll(struct pollfd *fds, nfds_t nfds, int timeout) { |
| int rv; |
| |
| THREAD_FREEZE_POINT1(); |
| rv = REAL(poll)(fds, nfds, timeout); |
| THREAD_FREEZE_POINT2(); |
| |
| return rv; |
| } |
| |
| extern "C" MFBT_API int |
| __wrap_epoll_create(int size) { |
| int epollfd = REAL(epoll_create)(size); |
| |
| if (!sIsNuwaProcess) { |
| return epollfd; |
| } |
| |
| if (epollfd >= 0) { |
| EpollManager::Singleton()->AddEpollInfo(epollfd, size); |
| } |
| |
| return epollfd; |
| } |
| |
| /** |
| * Wrapping the functions to create file descriptor pairs. In the child process |
| * FD pairs are created for intra-process signaling. The generation of FD pairs |
| * need to be tracked in the nuwa process so they can be recreated in the |
| * spawned process. |
| */ |
| struct FdPairInfo { |
| enum { |
| kPipe, |
| kSocketpair |
| } call; |
| |
| int FDs[2]; |
| int flags; |
| int domain; |
| int type; |
| int protocol; |
| }; |
| |
| /** |
| * Protects the access to sSingalFds. |
| */ |
| static pthread_mutex_t sSignalFdLock = PTHREAD_MUTEX_INITIALIZER; |
| static std::vector<FdPairInfo> sSignalFds; |
| |
| extern "C" MFBT_API int |
| __wrap_socketpair(int domain, int type, int protocol, int sv[2]) |
| { |
| int rv = REAL(socketpair)(domain, type, protocol, sv); |
| |
| if (!sIsNuwaProcess || rv < 0) { |
| return rv; |
| } |
| |
| REAL(pthread_mutex_lock)(&sSignalFdLock); |
| FdPairInfo signalFd; |
| signalFd.call = FdPairInfo::kSocketpair; |
| signalFd.FDs[0] = sv[0]; |
| signalFd.FDs[1] = sv[1]; |
| signalFd.domain = domain; |
| signalFd.type = type; |
| signalFd.protocol = protocol; |
| |
| sSignalFds.push_back(signalFd); |
| pthread_mutex_unlock(&sSignalFdLock); |
| |
| return rv; |
| } |
| |
| extern "C" MFBT_API int |
| __wrap_pipe2(int __pipedes[2], int flags) |
| { |
| int rv = REAL(pipe2)(__pipedes, flags); |
| if (!sIsNuwaProcess || rv < 0) { |
| return rv; |
| } |
| |
| REAL(pthread_mutex_lock)(&sSignalFdLock); |
| FdPairInfo signalFd; |
| signalFd.call = FdPairInfo::kPipe; |
| signalFd.FDs[0] = __pipedes[0]; |
| signalFd.FDs[1] = __pipedes[1]; |
| signalFd.flags = flags; |
| sSignalFds.push_back(signalFd); |
| pthread_mutex_unlock(&sSignalFdLock); |
| return rv; |
| } |
| |
| extern "C" MFBT_API int |
| __wrap_pipe(int __pipedes[2]) |
| { |
| return __wrap_pipe2(__pipedes, 0); |
| } |
| |
| static void |
| DupeSingleFd(int newFd, int origFd) |
| { |
| struct stat sb; |
| if (fstat(origFd, &sb)) { |
| // Maybe the original FD is closed. |
| return; |
| } |
| int fd = fcntl(origFd, F_GETFD); |
| int fl = fcntl(origFd, F_GETFL); |
| dup2(newFd, origFd); |
| fcntl(origFd, F_SETFD, fd); |
| fcntl(origFd, F_SETFL, fl); |
| REAL(close)(newFd); |
| } |
| |
| extern "C" MFBT_API void |
| ReplaceSignalFds() |
| { |
| for (std::vector<FdPairInfo>::iterator it = sSignalFds.begin(); |
| it < sSignalFds.end(); ++it) { |
| int fds[2]; |
| int rc = 0; |
| switch (it->call) { |
| case FdPairInfo::kPipe: |
| rc = REAL(pipe2)(fds, it->flags); |
| break; |
| case FdPairInfo::kSocketpair: |
| rc = REAL(socketpair)(it->domain, it->type, it->protocol, fds); |
| break; |
| default: |
| continue; |
| } |
| |
| if (rc == 0) { |
| DupeSingleFd(fds[0], it->FDs[0]); |
| DupeSingleFd(fds[1], it->FDs[1]); |
| } |
| } |
| } |
| |
| extern "C" MFBT_API int |
| __wrap_epoll_ctl(int aEpollFd, int aOp, int aFd, struct epoll_event *aEvent) { |
| int rv = REAL(epoll_ctl)(aEpollFd, aOp, aFd, aEvent); |
| |
| if (!sIsNuwaProcess || rv == -1) { |
| return rv; |
| } |
| |
| EpollManager::EpollInfo *info = |
| EpollManager::Singleton()->FindEpollInfo(aEpollFd); |
| if (info == nullptr) { |
| abort(); |
| } |
| |
| switch(aOp) { |
| case EPOLL_CTL_ADD: |
| info->AddEvents(aFd, *aEvent); |
| break; |
| |
| case EPOLL_CTL_MOD: |
| info->ModifyEvents(aFd, *aEvent); |
| break; |
| |
| case EPOLL_CTL_DEL: |
| info->RemoveEvents(aFd); |
| break; |
| |
| default: |
| abort(); |
| } |
| |
| return rv; |
| } |
| |
| // XXX: thinker: Maybe, we should also track dup, dup2, and other functions. |
| extern "C" MFBT_API int |
| __wrap_close(int aFd) { |
| int rv = REAL(close)(aFd); |
| if (!sIsNuwaProcess || rv == -1) { |
| return rv; |
| } |
| |
| EpollManager::EpollInfo *info = |
| EpollManager::Singleton()->FindEpollInfo(aFd); |
| if (info) { |
| EpollManager::Singleton()->RemoveEpollInfo(aFd); |
| } |
| |
| return rv; |
| } |
| |
| static void * |
| thread_recreate_startup(void *arg) { |
| /* |
| * Dark Art!! Never do the same unless you are ABSOLUTELY sure what you are |
| * doing! |
| * |
| * The stack space collapsed by this frame had been reserved by |
| * thread_create_startup(). And thread_create_startup() will |
| * return immediately after returning from real start routine, so |
| * all collapsed values does not affect the result. |
| * |
| * All outer frames of thread_create_startup() and |
| * thread_recreate_startup() are equivalent, so |
| * thread_create_startup() will return successfully. |
| */ |
| thread_info_t *tinfo = (thread_info_t *)arg; |
| |
| prctl(PR_SET_NAME, (unsigned long)&tinfo->nativeThreadName, 0, 0, 0); |
| RestoreTLSInfo(tinfo); |
| |
| if (setjmp(tinfo->retEnv) != 0) { |
| POST_SETJMP_RESTORE("thread_recreate_startup"); |
| return nullptr; |
| } |
| |
| // longjump() to recreate the stack on the new thread. |
| longjmp(tinfo->jmpEnv, 1); |
| |
| // Never go here! |
| abort(); |
| |
| return nullptr; |
| } |
| |
| /** |
| * Recreate the context given by tinfo at a new thread. |
| */ |
| static void |
| thread_recreate(thread_info_t *tinfo) { |
| pthread_t thread; |
| |
| // Note that the thread_recreate_startup() runs on the stack specified by |
| // tinfo. |
| pthread_create(&thread, &tinfo->threadAttr, thread_recreate_startup, tinfo); |
| } |
| |
| /** |
| * Recreate all threads in a process forked from an Nuwa process. |
| */ |
| static void |
| RecreateThreads() { |
| sIsNuwaProcess = false; |
| sIsFreezing = false; |
| |
| sMainThread.recreatedThreadID = pthread_self(); |
| sMainThread.recreatedNativeThreadID = gettid(); |
| |
| // Run registered constructors. |
| for (std::vector<nuwa_construct_t>::iterator ctr = sConstructors.begin(); |
| ctr != sConstructors.end(); |
| ctr++) { |
| (*ctr).construct((*ctr).arg); |
| } |
| sConstructors.clear(); |
| |
| REAL(pthread_mutex_lock)(&sThreadCountLock); |
| thread_info_t *tinfo = sAllThreads.getFirst(); |
| pthread_mutex_unlock(&sThreadCountLock); |
| |
| RECREATE_START(); |
| while (tinfo != nullptr) { |
| if (tinfo->flags & TINFO_FLAG_NUWA_SUPPORT) { |
| RECREATE_BEFORE(tinfo); |
| thread_recreate(tinfo); |
| RECREATE_WAIT(); |
| if (tinfo->condMutex) { |
| // Synchronize with the recreated thread in pthread_cond_wait(). |
| REAL(pthread_mutex_lock)(tinfo->condMutex); |
| // Tell the other thread that we have successfully locked the mutex. |
| // NB: condMutex can only be touched while it is held, so we must clear |
| // it here and store the mutex locally. |
| pthread_mutex_t *mtx = tinfo->condMutex; |
| tinfo->condMutex = nullptr; |
| if (tinfo->condMutexNeedsBalancing) { |
| pthread_mutex_unlock(mtx); |
| } |
| } |
| } else if(!(tinfo->flags & TINFO_FLAG_NUWA_SKIP)) { |
| // An unmarked thread is found other than the main thread. |
| |
| // All threads should be marked as one of SUPPORT or SKIP, or |
| // abort the process to make sure all threads in the Nuwa |
| // process are Nuwa-aware. |
| abort(); |
| } |
| |
| tinfo = tinfo->getNext(); |
| } |
| RECREATE_WAIT_ALL_VIP(); |
| RECREATE_OPEN_GATE(); |
| |
| RECREATE_FINISH(); |
| |
| // Run registered final constructors. |
| for (std::vector<nuwa_construct_t>::iterator ctr = sFinalConstructors.begin(); |
| ctr != sFinalConstructors.end(); |
| ctr++) { |
| (*ctr).construct((*ctr).arg); |
| } |
| sFinalConstructors.clear(); |
| } |
| |
| extern "C" { |
| |
| /** |
| * Recreate all epoll fds and restore status; include all events. |
| */ |
| static void |
| RecreateEpollFds() { |
| EpollManager *man = EpollManager::Singleton(); |
| |
| for (EpollManager::const_iterator info_it = man->begin(); |
| info_it != man->end(); |
| info_it++) { |
| int epollfd = info_it->first; |
| const EpollManager::EpollInfo *info = &info_it->second; |
| |
| int fdflags = fcntl(epollfd, F_GETFD); |
| if (fdflags == -1) { |
| abort(); |
| } |
| int fl = fcntl(epollfd, F_GETFL); |
| if (fl == -1) { |
| abort(); |
| } |
| |
| int newepollfd = REAL(epoll_create)(info->BackSize()); |
| if (newepollfd == -1) { |
| abort(); |
| } |
| int rv = REAL(close)(epollfd); |
| if (rv == -1) { |
| abort(); |
| } |
| rv = dup2(newepollfd, epollfd); |
| if (rv == -1) { |
| abort(); |
| } |
| rv = REAL(close)(newepollfd); |
| if (rv == -1) { |
| abort(); |
| } |
| |
| rv = fcntl(epollfd, F_SETFD, fdflags); |
| if (rv == -1) { |
| abort(); |
| } |
| rv = fcntl(epollfd, F_SETFL, fl); |
| if (rv == -1) { |
| abort(); |
| } |
| |
| for (EpollManager::EpollInfo::const_iterator events_it = info->begin(); |
| events_it != info->end(); |
| events_it++) { |
| int fd = events_it->first; |
| epoll_event events; |
| events = events_it->second; |
| rv = REAL(epoll_ctl)(epollfd, EPOLL_CTL_ADD, fd, &events); |
| if (rv == -1) { |
| abort(); |
| } |
| } |
| } |
| |
| // Shutdown EpollManager. It won't be needed in the spawned process. |
| EpollManager::Shutdown(); |
| } |
| |
| /** |
| * Fix IPC to make it ready. |
| * |
| * Especially, fix ContentChild. |
| */ |
| static void |
| ReplaceIPC(NuwaProtoFdInfo *aInfoList, int aInfoSize) { |
| int i; |
| int rv; |
| |
| for (i = 0; i < aInfoSize; i++) { |
| int fd = fcntl(aInfoList[i].originFd, F_GETFD); |
| if (fd == -1) { |
| abort(); |
| } |
| |
| int fl = fcntl(aInfoList[i].originFd, F_GETFL); |
| if (fl == -1) { |
| abort(); |
| } |
| |
| rv = dup2(aInfoList[i].newFds[NUWA_NEWFD_CHILD], aInfoList[i].originFd); |
| if (rv == -1) { |
| abort(); |
| } |
| |
| rv = fcntl(aInfoList[i].originFd, F_SETFD, fd); |
| if (rv == -1) { |
| abort(); |
| } |
| |
| rv = fcntl(aInfoList[i].originFd, F_SETFL, fl); |
| if (rv == -1) { |
| abort(); |
| } |
| } |
| } |
| |
| /** |
| * Add a new content process at the chrome process. |
| */ |
| static void |
| AddNewProcess(pid_t pid, NuwaProtoFdInfo *aInfoList, int aInfoSize) { |
| static bool (*AddNewIPCProcess)(pid_t, NuwaProtoFdInfo *, int) = nullptr; |
| |
| if (AddNewIPCProcess == nullptr) { |
| AddNewIPCProcess = (bool (*)(pid_t, NuwaProtoFdInfo *, int)) |
| dlsym(RTLD_DEFAULT, "AddNewIPCProcess"); |
| } |
| AddNewIPCProcess(pid, aInfoList, aInfoSize); |
| } |
| |
| static void |
| PrepareProtoSockets(NuwaProtoFdInfo *aInfoList, int aInfoSize) { |
| int i; |
| int rv; |
| |
| for (i = 0; i < aInfoSize; i++) { |
| rv = REAL(socketpair)(PF_UNIX, SOCK_STREAM, 0, aInfoList[i].newFds); |
| if (rv == -1) { |
| abort(); |
| } |
| } |
| } |
| |
| static void |
| CloseAllProtoSockets(NuwaProtoFdInfo *aInfoList, int aInfoSize) { |
| int i; |
| |
| for (i = 0; i < aInfoSize; i++) { |
| REAL(close)(aInfoList[i].newFds[0]); |
| REAL(close)(aInfoList[i].newFds[1]); |
| } |
| } |
| |
| static void |
| AfterForkHook() |
| { |
| void (*AfterNuwaFork)(); |
| |
| // This is defined in dom/ipc/ContentChild.cpp |
| AfterNuwaFork = (void (*)()) |
| dlsym(RTLD_DEFAULT, "AfterNuwaFork"); |
| AfterNuwaFork(); |
| } |
| |
| /** |
| * Fork a new process that is ready for running IPC. |
| * |
| * @return the PID of the new process. |
| */ |
| static int |
| ForkIPCProcess() { |
| int pid; |
| |
| REAL(pthread_mutex_lock)(&sForkLock); |
| |
| PrepareProtoSockets(sProtoFdInfos, sProtoFdInfosSize); |
| |
| sNuwaForking = true; |
| pid = fork(); |
| sNuwaForking = false; |
| if (pid == -1) { |
| abort(); |
| } |
| |
| if (pid > 0) { |
| // in the parent |
| AddNewProcess(pid, sProtoFdInfos, sProtoFdInfosSize); |
| CloseAllProtoSockets(sProtoFdInfos, sProtoFdInfosSize); |
| } else { |
| // in the child |
| sIsNuwaChildProcess = true; |
| if (getenv("MOZ_DEBUG_CHILD_PROCESS")) { |
| printf("\n\nNUWA CHILDCHILDCHILDCHILD\n debug me @ %d\n\n", getpid()); |
| sleep(30); |
| } |
| AfterForkHook(); |
| ReplaceSignalFds(); |
| ReplaceIPC(sProtoFdInfos, sProtoFdInfosSize); |
| RecreateEpollFds(); |
| RecreateThreads(); |
| CloseAllProtoSockets(sProtoFdInfos, sProtoFdInfosSize); |
| } |
| |
| sForkWaitCondChanged = true; |
| pthread_cond_signal(&sForkWaitCond); |
| pthread_mutex_unlock(&sForkLock); |
| |
| return pid; |
| } |
| |
| /** |
| * Prepare for spawning a new process. Called on the IPC thread. |
| */ |
| MFBT_API void |
| NuwaSpawnPrepare() { |
| REAL(pthread_mutex_lock)(&sForkLock); |
| |
| sForkWaitCondChanged = false; // Will be modified on the main thread. |
| } |
| |
| /** |
| * Let IPC thread wait until fork action on the main thread has completed. |
| */ |
| MFBT_API void |
| NuwaSpawnWait() { |
| while (!sForkWaitCondChanged) { |
| REAL(pthread_cond_wait)(&sForkWaitCond, &sForkLock); |
| } |
| pthread_mutex_unlock(&sForkLock); |
| } |
| |
| /** |
| * Spawn a new process. If not ready for spawn (still waiting for some threads |
| * to freeze), postpone the spawn request until ready. |
| * |
| * @return the pid of the new process, or 0 if not ready. |
| */ |
| MFBT_API pid_t |
| NuwaSpawn() { |
| if (gettid() != getpid()) { |
| // Not the main thread. |
| abort(); |
| } |
| |
| pid_t pid = 0; |
| |
| if (sNuwaReady) { |
| pid = ForkIPCProcess(); |
| } else { |
| sNuwaPendingSpawn = true; |
| } |
| |
| return pid; |
| } |
| |
| /** |
| * Prepare to freeze the Nuwa-supporting threads. |
| */ |
| MFBT_API void |
| PrepareNuwaProcess() { |
| sIsNuwaProcess = true; |
| // Explicitly ignore SIGCHLD so we don't have to call watpid() to reap |
| // dead child processes. |
| signal(SIGCHLD, SIG_IGN); |
| |
| // Make marked threads block in one freeze point. |
| REAL(pthread_mutex_lock)(&sThreadFreezeLock); |
| |
| // Populate sMainThread for mapping of tgkill. |
| sMainThread.origThreadID = pthread_self(); |
| sMainThread.origNativeThreadID = gettid(); |
| } |
| |
| // Make current process as a Nuwa process. |
| MFBT_API void |
| MakeNuwaProcess() { |
| void (*GetProtoFdInfos)(NuwaProtoFdInfo *, int, int *) = nullptr; |
| void (*OnNuwaProcessReady)() = nullptr; |
| sIsFreezing = true; |
| |
| REAL(pthread_mutex_lock)(&sThreadCountLock); |
| |
| // wait until all threads are frozen. |
| while ((sThreadFreezeCount + sThreadSkipCount) != sThreadCount) { |
| REAL(pthread_cond_wait)(&sThreadChangeCond, &sThreadCountLock); |
| } |
| |
| GetProtoFdInfos = (void (*)(NuwaProtoFdInfo *, int, int *)) |
| dlsym(RTLD_DEFAULT, "GetProtoFdInfos"); |
| GetProtoFdInfos(sProtoFdInfos, NUWA_TOPLEVEL_MAX, &sProtoFdInfosSize); |
| |
| sNuwaReady = true; |
| |
| pthread_mutex_unlock(&sThreadCountLock); |
| |
| OnNuwaProcessReady = (void (*)())dlsym(RTLD_DEFAULT, "OnNuwaProcessReady"); |
| OnNuwaProcessReady(); |
| |
| if (sNuwaPendingSpawn) { |
| sNuwaPendingSpawn = false; |
| NuwaSpawn(); |
| } |
| } |
| |
| /** |
| * Mark the current thread as supporting Nuwa. The thread will be recreated in |
| * the spawned process. |
| */ |
| MFBT_API void |
| NuwaMarkCurrentThread(void (*recreate)(void *), void *arg) { |
| if (!sIsNuwaProcess) { |
| return; |
| } |
| |
| thread_info_t *tinfo = CUR_THREAD_INFO; |
| if (tinfo == nullptr) { |
| abort(); |
| } |
| |
| tinfo->flags |= TINFO_FLAG_NUWA_SUPPORT; |
| if (recreate) { |
| nuwa_construct_t construct(recreate, arg); |
| tinfo->addThreadConstructor(&construct); |
| } |
| |
| // XXX Thread name might be set later than this call. If this is the case, we |
| // might need to delay getting the thread name. |
| prctl(PR_GET_NAME, (unsigned long)&tinfo->nativeThreadName, 0, 0, 0); |
| } |
| |
| /** |
| * Mark the current thread as not supporting Nuwa. Don't recreate this thread in |
| * the spawned process. |
| */ |
| MFBT_API void |
| NuwaSkipCurrentThread() { |
| if (!sIsNuwaProcess) return; |
| |
| thread_info_t *tinfo = CUR_THREAD_INFO; |
| if (tinfo == nullptr) { |
| abort(); |
| } |
| |
| if (!(tinfo->flags & TINFO_FLAG_NUWA_SKIP)) { |
| sThreadSkipCount++; |
| } |
| tinfo->flags |= TINFO_FLAG_NUWA_SKIP; |
| } |
| |
| /** |
| * Force to freeze the current thread. |
| * |
| * This method does not return in Nuwa process. It returns for the |
| * recreated thread. |
| */ |
| MFBT_API void |
| NuwaFreezeCurrentThread() { |
| thread_info_t *tinfo = CUR_THREAD_INFO; |
| if (sIsNuwaProcess && |
| (tinfo = CUR_THREAD_INFO) && |
| (tinfo->flags & TINFO_FLAG_NUWA_SUPPORT)) { |
| if (!setjmp(tinfo->jmpEnv)) { |
| REAL(pthread_mutex_lock)(&sThreadCountLock); |
| SaveTLSInfo(tinfo); |
| sThreadFreezeCount++; |
| pthread_cond_signal(&sThreadChangeCond); |
| pthread_mutex_unlock(&sThreadCountLock); |
| |
| REAL(pthread_mutex_lock)(&sThreadFreezeLock); |
| } else { |
| POST_SETJMP_RESTORE("NuwaFreezeCurrentThread"); |
| RECREATE_CONTINUE(); |
| RECREATE_GATE(); |
| } |
| } |
| } |
| |
| /** |
| * The caller of NuwaCheckpointCurrentThread() is at the line it wishes to |
| * return after the thread is recreated. |
| * |
| * The checkpointed thread will restart at the calling line of |
| * NuwaCheckpointCurrentThread(). This macro returns true in the Nuwa process |
| * and false on the recreated thread in the forked process. |
| * |
| * NuwaCheckpointCurrentThread() is implemented as a macro so we can place the |
| * setjmp() call in the calling method without changing its stack pointer. This |
| * is essential for not corrupting the stack when the calling thread continues |
| * to request the main thread for forking a new process. The caller of |
| * NuwaCheckpointCurrentThread() should not return before the process forking |
| * finishes. |
| * |
| * @return true for Nuwa process, and false in the forked process. |
| */ |
| MFBT_API jmp_buf* |
| NuwaCheckpointCurrentThread1() { |
| thread_info_t *tinfo = CUR_THREAD_INFO; |
| if (sIsNuwaProcess && |
| (tinfo = CUR_THREAD_INFO) && |
| (tinfo->flags & TINFO_FLAG_NUWA_SUPPORT)) { |
| return &tinfo->jmpEnv; |
| } |
| abort(); |
| return nullptr; |
| } |
| |
| MFBT_API bool |
| NuwaCheckpointCurrentThread2(int setjmpCond) { |
| thread_info_t *tinfo = CUR_THREAD_INFO; |
| if (setjmpCond == 0) { |
| REAL(pthread_mutex_lock)(&sThreadCountLock); |
| if (!(tinfo->flags & TINFO_FLAG_NUWA_EXPLICIT_CHECKPOINT)) { |
| tinfo->flags |= TINFO_FLAG_NUWA_EXPLICIT_CHECKPOINT; |
| SaveTLSInfo(tinfo); |
| sThreadFreezeCount++; |
| } |
| pthread_cond_signal(&sThreadChangeCond); |
| pthread_mutex_unlock(&sThreadCountLock); |
| return true; |
| } |
| POST_SETJMP_RESTORE("NuwaCheckpointCurrentThread2"); |
| RECREATE_CONTINUE(); |
| RECREATE_GATE(); |
| return false; // Recreated thread. |
| } |
| |
| /** |
| * Register methods to be invoked before recreating threads in the spawned |
| * process. |
| */ |
| MFBT_API void |
| NuwaAddConstructor(void (*construct)(void *), void *arg) { |
| nuwa_construct_t ctr(construct, arg); |
| sConstructors.push_back(ctr); |
| } |
| |
| /** |
| * Register methods to be invoked after recreating threads in the spawned |
| * process. |
| */ |
| MFBT_API void |
| NuwaAddFinalConstructor(void (*construct)(void *), void *arg) { |
| nuwa_construct_t ctr(construct, arg); |
| sFinalConstructors.push_back(ctr); |
| } |
| |
| MFBT_API void |
| NuwaAddThreadConstructor(void (*aConstruct)(void *), void *aArg) { |
| thread_info *tinfo = CUR_THREAD_INFO; |
| if (!tinfo || !aConstruct) { |
| return; |
| } |
| |
| nuwa_construct_t construct(aConstruct, aArg); |
| tinfo->addThreadConstructor(&construct); |
| } |
| |
| /** |
| * @return if the current process is the nuwa process. |
| */ |
| MFBT_API bool |
| IsNuwaProcess() { |
| return sIsNuwaProcess; |
| } |
| |
| /** |
| * @return if the nuwa process is ready for spawning new processes. |
| */ |
| MFBT_API bool |
| IsNuwaReady() { |
| return sNuwaReady; |
| } |
| |
| #if defined(DEBUG) || defined(ENABLE_TESTS) |
| MFBT_API void |
| NuwaAssertNotFrozen(unsigned int aThread, const char* aThreadName) { |
| if (!sIsNuwaProcess || !sIsFreezing) { |
| return; |
| } |
| |
| thread_info_t *tinfo = GetThreadInfo(static_cast<pthread_t>(aThread)); |
| if (!tinfo) { |
| return; |
| } |
| |
| if ((tinfo->flags & TINFO_FLAG_NUWA_SUPPORT) && |
| !(tinfo->flags & TINFO_FLAG_NUWA_EXPLICIT_CHECKPOINT)) { |
| __android_log_print(ANDROID_LOG_FATAL, "Nuwa", |
| "Fatal error: the Nuwa process is about to deadlock in " |
| "accessing a frozen thread (%s, tid=%d).", |
| aThreadName ? aThreadName : "(unnamed)", |
| tinfo->origNativeThreadID); |
| abort(); |
| } |
| } |
| #endif |
| |
| } // extern "C" |