Wrap blocking syscalls used by LLVMCAS inside RetryAfterSignal so they are retried on EINTR instead of surfacing as error or silently dropping actions due to failed to acquire locks.
443 lines
16 KiB
C++
443 lines
16 KiB
C++
//===----------------------------------------------------------------------===//
|
|
//
|
|
// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
|
|
// See https://llvm.org/LICENSE.txt for license information.
|
|
// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
|
|
//
|
|
//===----------------------------------------------------------------------===//
|
|
/// \file Implements MappedFileRegionArena.
|
|
///
|
|
/// A bump pointer allocator, backed by a memory-mapped file.
|
|
///
|
|
/// The effect we want is:
|
|
///
|
|
/// Step 1. If it doesn't exist, create the file with an initial size.
|
|
/// Step 2. Reserve virtual memory large enough for the max file size.
|
|
/// Step 3. Map the file into memory in the reserved region.
|
|
/// Step 4. Increase the file size and update the mapping when necessary.
|
|
///
|
|
/// However, updating the mapping is challenging when it needs to work portably,
|
|
/// and across multiple processes without locking for every read. Our current
|
|
/// implementation handles the steps above in following ways:
|
|
///
|
|
/// Step 1. Use \ref sys::fs::resize_file_sparse to grow the file to its max
|
|
/// size (typically several GB). If the file system doesn't support
|
|
/// sparse file, this may return a fully allocated file.
|
|
/// Step 2. Call \ref sys::fs::mapped_file_region to map the entire file.
|
|
/// Step 3. [Automatic as part of step 2.]
|
|
/// Step 4. If supported, use \c fallocate or similiar APIs to ensure the file
|
|
/// system storage for the sparse file so we won't end up with partial
|
|
/// file if the disk is out of space.
|
|
///
|
|
/// Additionally, we attempt to resize the file to its actual data size when
|
|
/// closing the mapping, if this is the only concurrent instance. This is done
|
|
/// using file locks. Shrinking the file mitigates problems with having large
|
|
/// files: on filesystems without sparse files it avoids unnecessary space use;
|
|
/// it also avoids allocating the full size if another process copies the file,
|
|
/// which typically loses sparseness. These mitigations only work while the file
|
|
/// is not in use.
|
|
///
|
|
/// The capacity and the header offset is determined by the first user of the
|
|
/// MappedFileRegionArena instance and any future mismatched value from the
|
|
/// original will result in error on creation.
|
|
///
|
|
/// To support resizing, we use two separate file locks:
|
|
/// 1. We use a shared reader lock on a ".shared" file until destruction.
|
|
/// 2. We use a lock on the main file during initialization - shared to check
|
|
/// the status, upgraded to exclusive to resize/initialize the file.
|
|
///
|
|
/// Then during destruction we attempt to get exclusive access on (1), which
|
|
/// requires no concurrent readers. If so, we shrink the file. Using two
|
|
/// separate locks simplifies the implementation and enables it to work on
|
|
/// platforms (e.g. Windows) where a shared/reader lock prevents writing.
|
|
//===----------------------------------------------------------------------===//
|
|
|
|
#include "llvm/CAS/MappedFileRegionArena.h"
|
|
#include "OnDiskCommon.h"
|
|
#include "llvm/ADT/StringExtras.h"
|
|
#include "llvm/CAS/OnDiskCASLogger.h"
|
|
#include "llvm/Support/Errno.h"
|
|
|
|
#if LLVM_ON_UNIX
|
|
#include <sys/stat.h>
|
|
#if __has_include(<sys/param.h>)
|
|
#include <sys/param.h>
|
|
#endif
|
|
#ifdef DEV_BSIZE
|
|
#define MAPPED_FILE_BSIZE DEV_BSIZE
|
|
#elif __linux__
|
|
#define MAPPED_FILE_BSIZE 512
|
|
#endif
|
|
#endif
|
|
|
|
using namespace llvm;
|
|
using namespace llvm::cas;
|
|
using namespace llvm::cas::ondisk;
|
|
|
|
namespace {
|
|
struct FileWithLock {
|
|
std::string Path;
|
|
int FD = -1;
|
|
std::optional<sys::fs::LockKind> Locked;
|
|
|
|
private:
|
|
FileWithLock(std::string PathStr, Error &E) : Path(std::move(PathStr)) {
|
|
ErrorAsOutParameter EOP(&E);
|
|
if (std::error_code EC = sys::fs::openFileForReadWrite(
|
|
Path, FD, sys::fs::CD_OpenAlways, sys::fs::OF_None))
|
|
E = createFileError(Path, EC);
|
|
}
|
|
|
|
public:
|
|
FileWithLock(FileWithLock &) = delete;
|
|
FileWithLock(FileWithLock &&Other) {
|
|
Path = std::move(Other.Path);
|
|
FD = Other.FD;
|
|
Other.FD = -1;
|
|
Locked = Other.Locked;
|
|
Other.Locked = std::nullopt;
|
|
}
|
|
|
|
~FileWithLock() { consumeError(unlock()); }
|
|
|
|
static Expected<FileWithLock> open(StringRef Path) {
|
|
Error E = Error::success();
|
|
FileWithLock Result(Path.str(), E);
|
|
if (E)
|
|
return std::move(E);
|
|
return std::move(Result);
|
|
}
|
|
|
|
Error lock(sys::fs::LockKind LK) {
|
|
assert(!Locked && "already locked");
|
|
if (std::error_code EC = lockFileThreadSafe(FD, LK))
|
|
return createFileError(Path, EC);
|
|
Locked = LK;
|
|
return Error::success();
|
|
}
|
|
|
|
Error switchLock(sys::fs::LockKind LK) {
|
|
assert(Locked && "not locked");
|
|
if (auto E = unlock())
|
|
return E;
|
|
|
|
return lock(LK);
|
|
}
|
|
|
|
Error unlock() {
|
|
if (Locked) {
|
|
Locked = std::nullopt;
|
|
if (std::error_code EC = unlockFileThreadSafe(FD))
|
|
return createFileError(Path, EC);
|
|
}
|
|
return Error::success();
|
|
}
|
|
|
|
// Return true if succeed to lock the file exclusively.
|
|
bool tryLockExclusive() {
|
|
assert(!Locked && "can only try to lock if not locked");
|
|
if (tryLockFileThreadSafe(FD) == std::error_code()) {
|
|
Locked = sys::fs::LockKind::Exclusive;
|
|
return true;
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
// Release the lock so it will not be unlocked on destruction.
|
|
void release() {
|
|
Locked = std::nullopt;
|
|
FD = -1;
|
|
}
|
|
};
|
|
|
|
struct FileSizeInfo {
|
|
uint64_t Size;
|
|
uint64_t AllocatedSize;
|
|
|
|
static ErrorOr<FileSizeInfo> get(sys::fs::file_t File);
|
|
};
|
|
} // end anonymous namespace
|
|
|
|
Expected<MappedFileRegionArena> MappedFileRegionArena::create(
|
|
const Twine &Path, uint64_t Capacity, uint64_t HeaderOffset,
|
|
std::shared_ptr<ondisk::OnDiskCASLogger> Logger,
|
|
function_ref<Error(MappedFileRegionArena &)> NewFileConstructor) {
|
|
uint64_t MinCapacity = HeaderOffset + sizeof(Header);
|
|
if (Capacity < MinCapacity)
|
|
return createStringError(
|
|
std::make_error_code(std::errc::invalid_argument),
|
|
"capacity is too small to hold MappedFileRegionArena");
|
|
|
|
MappedFileRegionArena Result;
|
|
Result.Path = Path.str();
|
|
Result.Logger = std::move(Logger);
|
|
|
|
// Open the support file. See file comment for details of locking scheme.
|
|
SmallString<128> SharedFilePath(Result.Path);
|
|
SharedFilePath.append(".shared");
|
|
|
|
auto SharedFileLock = FileWithLock::open(SharedFilePath);
|
|
if (!SharedFileLock)
|
|
return SharedFileLock.takeError();
|
|
Result.SharedLockFD = SharedFileLock->FD;
|
|
|
|
// Take shared/reader lock that will be held until destroyImpl if construction
|
|
// is successful.
|
|
if (auto E = SharedFileLock->lock(sys::fs::LockKind::Shared))
|
|
return std::move(E);
|
|
|
|
// Take shared/reader lock for initialization.
|
|
auto MainFile = FileWithLock::open(Result.Path);
|
|
if (!MainFile)
|
|
return MainFile.takeError();
|
|
if (Error E = MainFile->lock(sys::fs::LockKind::Shared))
|
|
return std::move(E);
|
|
Result.FD = MainFile->FD;
|
|
|
|
sys::fs::file_t File = sys::fs::convertFDToNativeFile(MainFile->FD);
|
|
auto FileSize = FileSizeInfo::get(File);
|
|
if (!FileSize)
|
|
return createFileError(Result.Path, FileSize.getError());
|
|
|
|
// If the size is smaller than the capacity, we need to initialize the file.
|
|
// It maybe empty, or may have been shrunk during a previous close.
|
|
if (FileSize->Size < Capacity) {
|
|
// Lock the file exclusively so only one process will do the initialization.
|
|
if (Error E = MainFile->switchLock(sys::fs::LockKind::Exclusive))
|
|
return std::move(E);
|
|
// Retrieve the current size now that we have exclusive access.
|
|
FileSize = FileSizeInfo::get(File);
|
|
if (!FileSize)
|
|
return createFileError(Result.Path, FileSize.getError());
|
|
}
|
|
|
|
if (FileSize->Size >= MinCapacity) {
|
|
// File is initialized. Read out the header to check for capacity and
|
|
// offset.
|
|
SmallVector<char, sizeof(Header)> HeaderContent(sizeof(Header));
|
|
auto Size = sys::fs::readNativeFileSlice(File, HeaderContent, HeaderOffset);
|
|
if (!Size)
|
|
return Size.takeError();
|
|
|
|
Header H;
|
|
memcpy(&H, HeaderContent.data(), sizeof(H));
|
|
if (H.HeaderOffset != HeaderOffset)
|
|
return createStringError(
|
|
std::make_error_code(std::errc::invalid_argument),
|
|
"specified header offset (" + utostr(HeaderOffset) +
|
|
") does not match existing config (" + utostr(H.HeaderOffset) +
|
|
")");
|
|
|
|
if (H.Capacity < MinCapacity)
|
|
return createStringError(
|
|
std::make_error_code(std::errc::bad_file_descriptor),
|
|
"capacity inside the MappedFileRegionArena is too small");
|
|
|
|
// If the capacity doesn't match, use the existing capacity instead.
|
|
if (H.Capacity != Capacity)
|
|
Capacity = H.Capacity;
|
|
}
|
|
|
|
// If the size is smaller than capacity, we need to resize the file.
|
|
if (FileSize->Size < Capacity) {
|
|
// Acquire the exclusive lock before resizing the file. In the rare case
|
|
// when opening a large CAS using a small requested size, a shared lock
|
|
// needs to switch to an exclusive lock here.
|
|
if (MainFile->Locked != sys::fs::LockKind::Exclusive) {
|
|
if (Error E = MainFile->switchLock(sys::fs::LockKind::Exclusive))
|
|
return std::move(E);
|
|
}
|
|
if (std::error_code EC =
|
|
sys::fs::resize_file_sparse(MainFile->FD, Capacity))
|
|
return createFileError(Result.Path, EC);
|
|
if (Result.Logger)
|
|
Result.Logger->logMappedFileRegionArenaResizeFile(
|
|
Result.Path, FileSize->Size, Capacity);
|
|
}
|
|
|
|
// Create the mapped region.
|
|
{
|
|
std::error_code EC;
|
|
const char *Name = nullptr;
|
|
#ifdef _WIN32
|
|
// Give the file mapping a name to ensure the same mappings are
|
|
// shared across processes.
|
|
std::string MapName = Result.Path;
|
|
std::replace(MapName.begin(), MapName.end(), '\\', '/');
|
|
MapName = "Local\\" + MapName;
|
|
Name = MapName.c_str();
|
|
#endif
|
|
sys::fs::mapped_file_region Map(
|
|
File, sys::fs::mapped_file_region::readwrite, Capacity, 0, EC, Name);
|
|
if (EC)
|
|
return createFileError(Result.Path, EC);
|
|
Result.Region = std::move(Map);
|
|
}
|
|
|
|
// Initialize the header.
|
|
if (Error E = Result.initializeHeader(HeaderOffset))
|
|
return std::move(E);
|
|
|
|
if (FileSize->Size < MinCapacity) {
|
|
assert(MainFile->Locked == sys::fs::LockKind::Exclusive);
|
|
// If we need to fully initialize the file, call NewFileConstructor.
|
|
if (Error E = NewFileConstructor(Result))
|
|
return std::move(E);
|
|
|
|
Result.H->HeaderOffset.exchange(HeaderOffset);
|
|
Result.H->Capacity.exchange(Capacity);
|
|
}
|
|
|
|
if (MainFile->Locked == sys::fs::LockKind::Exclusive) {
|
|
// If holding an exclusive lock, we might have resized the file and
|
|
// performed some read/write to the file. Query the file size again to make
|
|
// sure everything is up-to-date. Otherwise, FileSize info is already
|
|
// up-to-date.
|
|
FileSize = FileSizeInfo::get(File);
|
|
if (!FileSize)
|
|
return createFileError(Result.Path, FileSize.getError());
|
|
Result.H->AllocatedSize.exchange(FileSize->AllocatedSize);
|
|
}
|
|
|
|
// Release the shared lock so it can be closed in destoryImpl().
|
|
SharedFileLock->release();
|
|
return std::move(Result);
|
|
}
|
|
|
|
void MappedFileRegionArena::destroyImpl() {
|
|
if (!FD)
|
|
return;
|
|
|
|
// Drop the shared lock indicating we are no longer accessing the file.
|
|
if (SharedLockFD)
|
|
(void)unlockFileThreadSafe(*SharedLockFD);
|
|
|
|
// Attempt to truncate the file if we can get exclusive access. Ignore any
|
|
// errors.
|
|
if (H) {
|
|
assert(SharedLockFD && "Must have shared lock file open");
|
|
if (tryLockFileThreadSafe(*SharedLockFD) == std::error_code()) {
|
|
size_t Size = size();
|
|
size_t Capacity = capacity();
|
|
// sync to file system to make sure all contents are up-to-date.
|
|
(void)Region.sync();
|
|
// unmap the file before resizing since that is the requirement for
|
|
// some platforms.
|
|
Region.unmap();
|
|
(void)sys::fs::resize_file(*FD, Size);
|
|
(void)unlockFileThreadSafe(*SharedLockFD);
|
|
if (Logger)
|
|
Logger->logMappedFileRegionArenaResizeFile(Path, Capacity, Size);
|
|
}
|
|
}
|
|
|
|
auto Close = [](std::optional<int> &FD) {
|
|
if (FD) {
|
|
sys::fs::file_t File = sys::fs::convertFDToNativeFile(*FD);
|
|
sys::fs::closeFile(File);
|
|
FD = std::nullopt;
|
|
}
|
|
};
|
|
|
|
// Close the file and shared lock.
|
|
Close(FD);
|
|
Close(SharedLockFD);
|
|
|
|
if (Logger)
|
|
Logger->logMappedFileRegionArenaClose(Path);
|
|
}
|
|
|
|
Error MappedFileRegionArena::initializeHeader(uint64_t HeaderOffset) {
|
|
if (capacity() >= static_cast<uint64_t>(INT64_MAX))
|
|
return createStringError(make_error_code(std::errc::protocol_error),
|
|
"arena capacity does not fit in int64_t");
|
|
uint64_t HeaderEndOffset = HeaderOffset + sizeof(decltype(*H));
|
|
if (HeaderEndOffset > capacity())
|
|
return createStringError(make_error_code(std::errc::protocol_error),
|
|
"arena header extends past capacity");
|
|
if (!isAligned(Align::Of<decltype(*H)>(), HeaderOffset))
|
|
return createStringError(make_error_code(std::errc::protocol_error),
|
|
"arena header offset is not aligned");
|
|
H = reinterpret_cast<decltype(H)>(data() + HeaderOffset);
|
|
|
|
uint64_t ExistingValue = 0;
|
|
if (!H->BumpPtr.compare_exchange_strong(ExistingValue, HeaderEndOffset))
|
|
if (ExistingValue < HeaderEndOffset)
|
|
return createStringError(
|
|
make_error_code(std::errc::protocol_error),
|
|
"arena bump pointer is corrupt: 0x" +
|
|
utohexstr(ExistingValue, /*LowerCase=*/true));
|
|
if (Logger)
|
|
Logger->logMappedFileRegionArenaCreate(Path, *FD, data(), capacity(),
|
|
size());
|
|
return Error::success();
|
|
}
|
|
|
|
static Error createAllocatorOutOfSpaceError() {
|
|
return createStringError(std::make_error_code(std::errc::not_enough_memory),
|
|
"memory mapped file allocator is out of space");
|
|
}
|
|
|
|
Expected<int64_t> MappedFileRegionArena::allocateOffset(uint64_t AllocSize) {
|
|
AllocSize = alignTo(AllocSize, getAlign());
|
|
uint64_t OldEnd = H->BumpPtr.fetch_add(AllocSize);
|
|
uint64_t NewEnd = OldEnd + AllocSize;
|
|
if (LLVM_UNLIKELY(NewEnd > capacity())) {
|
|
// Return the allocation. If the start already passed the end, that means
|
|
// some other concurrent allocations already consumed all the capacity.
|
|
// There is no need to return the original value. If the start was not
|
|
// passed the end, current allocation certainly bumped it passed the end.
|
|
// All other allocation afterwards must have failed and current allocation
|
|
// is in charge of return the allocation back to a valid value.
|
|
if (OldEnd <= capacity())
|
|
(void)H->BumpPtr.exchange(OldEnd);
|
|
|
|
if (Logger)
|
|
Logger->logMappedFileRegionArenaOom(Path, capacity(), OldEnd, AllocSize);
|
|
|
|
return createAllocatorOutOfSpaceError();
|
|
}
|
|
|
|
uint64_t DiskSize = H->AllocatedSize;
|
|
if (LLVM_UNLIKELY(NewEnd > DiskSize)) {
|
|
uint64_t NewSize;
|
|
// The minimum increment is a page, but allocate more to amortize the cost.
|
|
constexpr uint64_t Increment = 1 * 1024 * 1024; // 1 MB
|
|
if (Error E = preallocateFileTail(*FD, DiskSize, DiskSize + Increment)
|
|
.moveInto(NewSize))
|
|
return std::move(E);
|
|
assert(NewSize >= DiskSize + Increment);
|
|
// FIXME: on Darwin this can under-count the size if there is a race to
|
|
// preallocate disk, because the semantics of F_PREALLOCATE are to add bytes
|
|
// to the end of the file, not to allocate up to a fixed size.
|
|
// Any discrepancy will be resolved the next time the file is truncated and
|
|
// then reopend.
|
|
while (DiskSize < NewSize)
|
|
H->AllocatedSize.compare_exchange_strong(DiskSize, NewSize);
|
|
}
|
|
|
|
if (Logger)
|
|
Logger->logMappedFileRegionArenaAllocate(data(), OldEnd, AllocSize);
|
|
|
|
return OldEnd;
|
|
}
|
|
|
|
ErrorOr<FileSizeInfo> FileSizeInfo::get(sys::fs::file_t File) {
|
|
#if LLVM_ON_UNIX && defined(MAPPED_FILE_BSIZE)
|
|
struct stat Status;
|
|
int StatRet = sys::RetryAfterSignal(-1, ::fstat, File, &Status);
|
|
if (StatRet)
|
|
return errnoAsErrorCode();
|
|
uint64_t AllocatedSize = uint64_t(Status.st_blksize) * MAPPED_FILE_BSIZE;
|
|
return FileSizeInfo{uint64_t(Status.st_size), AllocatedSize};
|
|
#else
|
|
// Fallback: assume the file is fully allocated. Note: this may result in
|
|
// data loss on out-of-space.
|
|
sys::fs::file_status Status;
|
|
if (std::error_code EC = sys::fs::status(File, Status))
|
|
return EC;
|
|
return FileSizeInfo{Status.getSize(), Status.getSize()};
|
|
#endif
|
|
}
|