13cab2bb3Spatrick //===-- fuchsia.cpp ---------------------------------------------*- C++ -*-===//
23cab2bb3Spatrick //
33cab2bb3Spatrick // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
43cab2bb3Spatrick // See https://llvm.org/LICENSE.txt for license information.
53cab2bb3Spatrick // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
63cab2bb3Spatrick //
73cab2bb3Spatrick //===----------------------------------------------------------------------===//
83cab2bb3Spatrick
93cab2bb3Spatrick #include "platform.h"
103cab2bb3Spatrick
113cab2bb3Spatrick #if SCUDO_FUCHSIA
123cab2bb3Spatrick
133cab2bb3Spatrick #include "common.h"
143cab2bb3Spatrick #include "mutex.h"
153cab2bb3Spatrick #include "string_utils.h"
163cab2bb3Spatrick
173cab2bb3Spatrick #include <lib/sync/mutex.h> // for sync_mutex_t
183cab2bb3Spatrick #include <stdlib.h> // for getenv()
193cab2bb3Spatrick #include <zircon/compiler.h>
20*810390e3Srobert #include <zircon/process.h>
213cab2bb3Spatrick #include <zircon/sanitizer.h>
223cab2bb3Spatrick #include <zircon/syscalls.h>
233cab2bb3Spatrick
243cab2bb3Spatrick namespace scudo {
253cab2bb3Spatrick
getPageSize()26d89ec533Spatrick uptr getPageSize() { return _zx_system_get_page_size(); }
273cab2bb3Spatrick
die()283cab2bb3Spatrick void NORETURN die() { __builtin_trap(); }
293cab2bb3Spatrick
303cab2bb3Spatrick // We zero-initialize the Extra parameter of map(), make sure this is consistent
313cab2bb3Spatrick // with ZX_HANDLE_INVALID.
323cab2bb3Spatrick static_assert(ZX_HANDLE_INVALID == 0, "");
333cab2bb3Spatrick
allocateVmar(uptr Size,MapPlatformData * Data,bool AllowNoMem)343cab2bb3Spatrick static void *allocateVmar(uptr Size, MapPlatformData *Data, bool AllowNoMem) {
353cab2bb3Spatrick // Only scenario so far.
363cab2bb3Spatrick DCHECK(Data);
373cab2bb3Spatrick DCHECK_EQ(Data->Vmar, ZX_HANDLE_INVALID);
383cab2bb3Spatrick
393cab2bb3Spatrick const zx_status_t Status = _zx_vmar_allocate(
403cab2bb3Spatrick _zx_vmar_root_self(),
413cab2bb3Spatrick ZX_VM_CAN_MAP_READ | ZX_VM_CAN_MAP_WRITE | ZX_VM_CAN_MAP_SPECIFIC, 0,
423cab2bb3Spatrick Size, &Data->Vmar, &Data->VmarBase);
433cab2bb3Spatrick if (UNLIKELY(Status != ZX_OK)) {
443cab2bb3Spatrick if (Status != ZX_ERR_NO_MEMORY || !AllowNoMem)
45d89ec533Spatrick dieOnMapUnmapError(Status == ZX_ERR_NO_MEMORY ? Size : 0);
463cab2bb3Spatrick return nullptr;
473cab2bb3Spatrick }
483cab2bb3Spatrick return reinterpret_cast<void *>(Data->VmarBase);
493cab2bb3Spatrick }
503cab2bb3Spatrick
map(void * Addr,uptr Size,const char * Name,uptr Flags,MapPlatformData * Data)513cab2bb3Spatrick void *map(void *Addr, uptr Size, const char *Name, uptr Flags,
523cab2bb3Spatrick MapPlatformData *Data) {
53d89ec533Spatrick DCHECK_EQ(Size % getPageSizeCached(), 0);
543cab2bb3Spatrick const bool AllowNoMem = !!(Flags & MAP_ALLOWNOMEM);
553cab2bb3Spatrick
563cab2bb3Spatrick // For MAP_NOACCESS, just allocate a Vmar and return.
573cab2bb3Spatrick if (Flags & MAP_NOACCESS)
583cab2bb3Spatrick return allocateVmar(Size, Data, AllowNoMem);
593cab2bb3Spatrick
60*810390e3Srobert const zx_handle_t Vmar = (Data && Data->Vmar != ZX_HANDLE_INVALID)
61*810390e3Srobert ? Data->Vmar
62*810390e3Srobert : _zx_vmar_root_self();
633cab2bb3Spatrick
643cab2bb3Spatrick zx_status_t Status;
653cab2bb3Spatrick zx_handle_t Vmo;
663cab2bb3Spatrick uint64_t VmoSize = 0;
673cab2bb3Spatrick if (Data && Data->Vmo != ZX_HANDLE_INVALID) {
683cab2bb3Spatrick // If a Vmo was specified, it's a resize operation.
693cab2bb3Spatrick CHECK(Addr);
703cab2bb3Spatrick DCHECK(Flags & MAP_RESIZABLE);
713cab2bb3Spatrick Vmo = Data->Vmo;
723cab2bb3Spatrick VmoSize = Data->VmoSize;
733cab2bb3Spatrick Status = _zx_vmo_set_size(Vmo, VmoSize + Size);
743cab2bb3Spatrick if (Status != ZX_OK) {
753cab2bb3Spatrick if (Status != ZX_ERR_NO_MEMORY || !AllowNoMem)
76d89ec533Spatrick dieOnMapUnmapError(Status == ZX_ERR_NO_MEMORY ? Size : 0);
773cab2bb3Spatrick return nullptr;
783cab2bb3Spatrick }
793cab2bb3Spatrick } else {
803cab2bb3Spatrick // Otherwise, create a Vmo and set its name.
813cab2bb3Spatrick Status = _zx_vmo_create(Size, ZX_VMO_RESIZABLE, &Vmo);
823cab2bb3Spatrick if (UNLIKELY(Status != ZX_OK)) {
833cab2bb3Spatrick if (Status != ZX_ERR_NO_MEMORY || !AllowNoMem)
84d89ec533Spatrick dieOnMapUnmapError(Status == ZX_ERR_NO_MEMORY ? Size : 0);
853cab2bb3Spatrick return nullptr;
863cab2bb3Spatrick }
873cab2bb3Spatrick _zx_object_set_property(Vmo, ZX_PROP_NAME, Name, strlen(Name));
883cab2bb3Spatrick }
893cab2bb3Spatrick
903cab2bb3Spatrick uintptr_t P;
913cab2bb3Spatrick zx_vm_option_t MapFlags =
923cab2bb3Spatrick ZX_VM_PERM_READ | ZX_VM_PERM_WRITE | ZX_VM_ALLOW_FAULTS;
93*810390e3Srobert if (Addr)
94*810390e3Srobert DCHECK(Data);
953cab2bb3Spatrick const uint64_t Offset =
963cab2bb3Spatrick Addr ? reinterpret_cast<uintptr_t>(Addr) - Data->VmarBase : 0;
973cab2bb3Spatrick if (Offset)
983cab2bb3Spatrick MapFlags |= ZX_VM_SPECIFIC;
993cab2bb3Spatrick Status = _zx_vmar_map(Vmar, MapFlags, Offset, Vmo, VmoSize, Size, &P);
100*810390e3Srobert if (UNLIKELY(Status != ZX_OK)) {
101*810390e3Srobert if (Status != ZX_ERR_NO_MEMORY || !AllowNoMem)
102*810390e3Srobert dieOnMapUnmapError(Status == ZX_ERR_NO_MEMORY ? Size : 0);
103*810390e3Srobert return nullptr;
104*810390e3Srobert }
105*810390e3Srobert
106*810390e3Srobert if (Flags & MAP_PRECOMMIT) {
107*810390e3Srobert Status = _zx_vmar_op_range(Vmar, ZX_VMAR_OP_COMMIT, P, Size,
108*810390e3Srobert /*buffer=*/nullptr, /*buffer_size=*/0);
109*810390e3Srobert }
110*810390e3Srobert
1113cab2bb3Spatrick // No need to track the Vmo if we don't intend on resizing it. Close it.
1123cab2bb3Spatrick if (Flags & MAP_RESIZABLE) {
1133cab2bb3Spatrick DCHECK(Data);
114d89ec533Spatrick if (Data->Vmo == ZX_HANDLE_INVALID)
1153cab2bb3Spatrick Data->Vmo = Vmo;
116d89ec533Spatrick else
117d89ec533Spatrick DCHECK_EQ(Data->Vmo, Vmo);
1183cab2bb3Spatrick } else {
1193cab2bb3Spatrick CHECK_EQ(_zx_handle_close(Vmo), ZX_OK);
1203cab2bb3Spatrick }
1213cab2bb3Spatrick if (UNLIKELY(Status != ZX_OK)) {
1223cab2bb3Spatrick if (Status != ZX_ERR_NO_MEMORY || !AllowNoMem)
123d89ec533Spatrick dieOnMapUnmapError(Status == ZX_ERR_NO_MEMORY ? Size : 0);
1243cab2bb3Spatrick return nullptr;
1253cab2bb3Spatrick }
126*810390e3Srobert
1273cab2bb3Spatrick if (Data)
1283cab2bb3Spatrick Data->VmoSize += Size;
1293cab2bb3Spatrick
1303cab2bb3Spatrick return reinterpret_cast<void *>(P);
1313cab2bb3Spatrick }
1323cab2bb3Spatrick
unmap(void * Addr,uptr Size,uptr Flags,MapPlatformData * Data)1333cab2bb3Spatrick void unmap(void *Addr, uptr Size, uptr Flags, MapPlatformData *Data) {
1343cab2bb3Spatrick if (Flags & UNMAP_ALL) {
1353cab2bb3Spatrick DCHECK_NE(Data, nullptr);
1363cab2bb3Spatrick const zx_handle_t Vmar = Data->Vmar;
1373cab2bb3Spatrick DCHECK_NE(Vmar, _zx_vmar_root_self());
1383cab2bb3Spatrick // Destroying the vmar effectively unmaps the whole mapping.
1393cab2bb3Spatrick CHECK_EQ(_zx_vmar_destroy(Vmar), ZX_OK);
1403cab2bb3Spatrick CHECK_EQ(_zx_handle_close(Vmar), ZX_OK);
1413cab2bb3Spatrick } else {
142*810390e3Srobert const zx_handle_t Vmar = (Data && Data->Vmar != ZX_HANDLE_INVALID)
143*810390e3Srobert ? Data->Vmar
144*810390e3Srobert : _zx_vmar_root_self();
1453cab2bb3Spatrick const zx_status_t Status =
1463cab2bb3Spatrick _zx_vmar_unmap(Vmar, reinterpret_cast<uintptr_t>(Addr), Size);
1473cab2bb3Spatrick if (UNLIKELY(Status != ZX_OK))
1483cab2bb3Spatrick dieOnMapUnmapError();
1493cab2bb3Spatrick }
1503cab2bb3Spatrick if (Data) {
1513cab2bb3Spatrick if (Data->Vmo != ZX_HANDLE_INVALID)
1523cab2bb3Spatrick CHECK_EQ(_zx_handle_close(Data->Vmo), ZX_OK);
1533cab2bb3Spatrick memset(Data, 0, sizeof(*Data));
1543cab2bb3Spatrick }
1553cab2bb3Spatrick }
1563cab2bb3Spatrick
setMemoryPermission(UNUSED uptr Addr,UNUSED uptr Size,UNUSED uptr Flags,UNUSED MapPlatformData * Data)157d89ec533Spatrick void setMemoryPermission(UNUSED uptr Addr, UNUSED uptr Size, UNUSED uptr Flags,
158d89ec533Spatrick UNUSED MapPlatformData *Data) {
159d89ec533Spatrick const zx_vm_option_t Prot =
160d89ec533Spatrick (Flags & MAP_NOACCESS) ? 0 : (ZX_VM_PERM_READ | ZX_VM_PERM_WRITE);
161d89ec533Spatrick DCHECK(Data);
162d89ec533Spatrick DCHECK_NE(Data->Vmar, ZX_HANDLE_INVALID);
163d89ec533Spatrick if (_zx_vmar_protect(Data->Vmar, Prot, Addr, Size) != ZX_OK)
164d89ec533Spatrick dieOnMapUnmapError();
165d89ec533Spatrick }
166d89ec533Spatrick
releasePagesToOS(UNUSED uptr BaseAddress,uptr Offset,uptr Size,MapPlatformData * Data)1673cab2bb3Spatrick void releasePagesToOS(UNUSED uptr BaseAddress, uptr Offset, uptr Size,
1683cab2bb3Spatrick MapPlatformData *Data) {
1693cab2bb3Spatrick DCHECK(Data);
1703cab2bb3Spatrick DCHECK_NE(Data->Vmar, ZX_HANDLE_INVALID);
1713cab2bb3Spatrick DCHECK_NE(Data->Vmo, ZX_HANDLE_INVALID);
1723cab2bb3Spatrick const zx_status_t Status =
1733cab2bb3Spatrick _zx_vmo_op_range(Data->Vmo, ZX_VMO_OP_DECOMMIT, Offset, Size, NULL, 0);
1743cab2bb3Spatrick CHECK_EQ(Status, ZX_OK);
1753cab2bb3Spatrick }
1763cab2bb3Spatrick
getEnv(const char * Name)1773cab2bb3Spatrick const char *getEnv(const char *Name) { return getenv(Name); }
1783cab2bb3Spatrick
1793cab2bb3Spatrick // Note: we need to flag these methods with __TA_NO_THREAD_SAFETY_ANALYSIS
1803cab2bb3Spatrick // because the Fuchsia implementation of sync_mutex_t has clang thread safety
1813cab2bb3Spatrick // annotations. Were we to apply proper capability annotations to the top level
1823cab2bb3Spatrick // HybridMutex class itself, they would not be needed. As it stands, the
1833cab2bb3Spatrick // thread analysis thinks that we are locking the mutex and accidentally leaving
1843cab2bb3Spatrick // it locked on the way out.
tryLock()1853cab2bb3Spatrick bool HybridMutex::tryLock() __TA_NO_THREAD_SAFETY_ANALYSIS {
1863cab2bb3Spatrick // Size and alignment must be compatible between both types.
1873cab2bb3Spatrick return sync_mutex_trylock(&M) == ZX_OK;
1883cab2bb3Spatrick }
1893cab2bb3Spatrick
lockSlow()1903cab2bb3Spatrick void HybridMutex::lockSlow() __TA_NO_THREAD_SAFETY_ANALYSIS {
1913cab2bb3Spatrick sync_mutex_lock(&M);
1923cab2bb3Spatrick }
1933cab2bb3Spatrick
unlock()1943cab2bb3Spatrick void HybridMutex::unlock() __TA_NO_THREAD_SAFETY_ANALYSIS {
1953cab2bb3Spatrick sync_mutex_unlock(&M);
1963cab2bb3Spatrick }
1973cab2bb3Spatrick
getMonotonicTime()1983cab2bb3Spatrick u64 getMonotonicTime() { return _zx_clock_get_monotonic(); }
1993cab2bb3Spatrick
getNumberOfCPUs()2003cab2bb3Spatrick u32 getNumberOfCPUs() { return _zx_system_get_num_cpus(); }
2013cab2bb3Spatrick
getThreadID()2021f9cb04fSpatrick u32 getThreadID() { return 0; }
2031f9cb04fSpatrick
getRandom(void * Buffer,uptr Length,UNUSED bool Blocking)2043cab2bb3Spatrick bool getRandom(void *Buffer, uptr Length, UNUSED bool Blocking) {
2053cab2bb3Spatrick static_assert(MaxRandomLength <= ZX_CPRNG_DRAW_MAX_LEN, "");
2063cab2bb3Spatrick if (UNLIKELY(!Buffer || !Length || Length > MaxRandomLength))
2073cab2bb3Spatrick return false;
2083cab2bb3Spatrick _zx_cprng_draw(Buffer, Length);
2093cab2bb3Spatrick return true;
2103cab2bb3Spatrick }
2113cab2bb3Spatrick
outputRaw(const char * Buffer)2123cab2bb3Spatrick void outputRaw(const char *Buffer) {
2133cab2bb3Spatrick __sanitizer_log_write(Buffer, strlen(Buffer));
2143cab2bb3Spatrick }
2153cab2bb3Spatrick
setAbortMessage(const char * Message)2163cab2bb3Spatrick void setAbortMessage(const char *Message) {}
2173cab2bb3Spatrick
2183cab2bb3Spatrick } // namespace scudo
2193cab2bb3Spatrick
2203cab2bb3Spatrick #endif // SCUDO_FUCHSIA
221