| 1 | /*
|
| 2 | * Souffle - A Datalog Compiler
|
| 3 | * Copyright (c) 2013, 2015, Oracle and/or its affiliates. All rights reserved
|
| 4 | * Licensed under the Universal Permissive License v 1.0 as shown at:
|
| 5 | * - https://opensource.org/licenses/UPL
|
| 6 | * - <souffle root>/licenses/SOUFFLE-UPL.txt
|
| 7 | */
|
| 8 |
|
| 9 | /************************************************************************
|
| 10 | *
|
| 11 | * @file ParallelUtil.h
|
| 12 | *
|
| 13 | * A set of utilities abstracting from the underlying parallel library.
|
| 14 | * Currently supported APIs: OpenMP and Cilk
|
| 15 | *
|
| 16 | ***********************************************************************/
|
| 17 |
|
| 18 | #pragma once
|
| 19 |
|
| 20 | #include <atomic>
|
| 21 | #include <cassert>
|
| 22 | #include <cstddef>
|
| 23 | #include <memory>
|
| 24 | #include <new>
|
| 25 |
|
| 26 | // https://bugs.llvm.org/show_bug.cgi?id=41423
|
| 27 | #if defined(__cpp_lib_hardware_interference_size) && (__cpp_lib_hardware_interference_size != 201703L)
|
| 28 | using std::hardware_constructive_interference_size;
|
| 29 | using std::hardware_destructive_interference_size;
|
| 30 | #else
|
| 31 | // 64 bytes on x86-64 │ L1_CACHE_BYTES │ L1_CACHE_SHIFT │ __cacheline_aligned │
|
| 32 | // ...
|
| 33 | constexpr std::size_t hardware_constructive_interference_size = 2 * sizeof(max_align_t);
|
| 34 | constexpr std::size_t hardware_destructive_interference_size = 2 * sizeof(max_align_t);
|
| 35 | #endif
|
| 36 |
|
| 37 | #ifdef _OPENMP
|
| 38 |
|
| 39 | /**
|
| 40 | * Implementation of parallel control flow constructs utilizing OpenMP
|
| 41 | */
|
| 42 |
|
| 43 | #include <omp.h>
|
| 44 |
|
| 45 | #ifdef __APPLE__
|
| 46 | #define pthread_yield pthread_yield_np
|
| 47 | #elif !defined(_MSC_VER)
|
| 48 | #include <sched.h>
|
| 49 | // pthread_yield is deprecated and should be replaced by sched_yield
|
| 50 | #define pthread_yield sched_yield
|
| 51 | #elif defined _MSC_VER
|
| 52 | #include <thread>
|
| 53 | #define NOMINMAX
|
| 54 | #include <windows.h>
|
| 55 | #define pthread_yield std::this_thread::yield
|
| 56 | #endif
|
| 57 |
|
| 58 | #ifdef _MSC_VER
|
| 59 | // support for a parallel region
|
| 60 | #define PARALLEL_START __pragma(omp parallel) {
|
| 61 | #define PARALLEL_END }
|
| 62 |
|
| 63 | // support for parallel loops
|
| 64 | #define pfor __pragma(omp for schedule(dynamic)) for
|
| 65 | #else
|
| 66 | // support for a parallel region
|
| 67 | #define PARALLEL_START _Pragma("omp parallel") {
|
| 68 | #define PARALLEL_END }
|
| 69 |
|
| 70 | // support for parallel loops
|
| 71 | #define pfor _Pragma("omp for schedule(dynamic)") for
|
| 72 | #endif
|
| 73 |
|
| 74 | // spawn and sync are processed sequentially (overhead to expensive)
|
| 75 | #define task_spawn
|
| 76 | #define task_sync
|
| 77 |
|
| 78 | // section start / end => corresponding OpenMP pragmas
|
| 79 | // NOTE: disabled since it causes performance losses
|
| 80 | //#define SECTIONS_START _Pragma("omp parallel sections") {
|
| 81 | // NOTE: we stick to flat-level parallelism since it is faster due to thread pooling
|
| 82 | #define SECTIONS_START {
|
| 83 | #define SECTIONS_END }
|
| 84 |
|
| 85 | // the markers for a single section
|
| 86 | //#define SECTION_START _Pragma("omp section") {
|
| 87 | #define SECTION_START {
|
| 88 | #define SECTION_END }
|
| 89 |
|
| 90 | // a macro to create an operation context
|
| 91 | #define CREATE_OP_CONTEXT(NAME, INIT) [[maybe_unused]] auto NAME = INIT;
|
| 92 | #define READ_OP_CONTEXT(NAME) NAME
|
| 93 |
|
| 94 | #else
|
| 95 |
|
| 96 | // support for a parallel region => sequential execution
|
| 97 | #define PARALLEL_START {
|
| 98 | #define PARALLEL_END }
|
| 99 |
|
| 100 | // support for parallel loops => simple sequential loop
|
| 101 | #define pfor for
|
| 102 |
|
| 103 | // spawn and sync not supported
|
| 104 | #define task_spawn
|
| 105 | #define task_sync
|
| 106 |
|
| 107 | // sections are processed sequentially
|
| 108 | #define SECTIONS_START {
|
| 109 | #define SECTIONS_END }
|
| 110 |
|
| 111 | // sections are inlined
|
| 112 | #define SECTION_START {
|
| 113 | #define SECTION_END }
|
| 114 |
|
| 115 | // a macro to create an operation context
|
| 116 | #define CREATE_OP_CONTEXT(NAME, INIT) [[maybe_unused]] auto NAME = INIT;
|
| 117 | #define READ_OP_CONTEXT(NAME) NAME
|
| 118 |
|
| 119 | // mark es sequential
|
| 120 | #define IS_SEQUENTIAL
|
| 121 |
|
| 122 | #endif
|
| 123 |
|
| 124 | #ifndef IS_SEQUENTIAL
|
| 125 | #define IS_PARALLEL
|
| 126 | #endif
|
| 127 |
|
| 128 | #ifdef IS_PARALLEL
|
| 129 | #include <mutex>
|
| 130 | #include <vector>
|
| 131 | #define MAX_THREADS (omp_get_max_threads())
|
| 132 | #else
|
| 133 | #define MAX_THREADS (1)
|
| 134 | #endif
|
| 135 |
|
| 136 | namespace souffle {
|
| 137 |
|
| 138 | struct SeqConcurrentLanes {
|
| 139 | struct TrivialLock {
|
| 140 | ~TrivialLock() {}
|
| 141 | };
|
| 142 |
|
| 143 | using lane_id = std::size_t;
|
| 144 | using unique_lock_type = TrivialLock;
|
| 145 |
|
| 146 | explicit SeqConcurrentLanes(std::size_t = 1) {}
|
| 147 | SeqConcurrentLanes(const SeqConcurrentLanes&) = delete;
|
| 148 | SeqConcurrentLanes(SeqConcurrentLanes&&) = delete;
|
| 149 |
|
| 150 | virtual ~SeqConcurrentLanes() {}
|
| 151 |
|
| 152 | std::size_t lanes() const {
|
| 153 | return 1;
|
| 154 | }
|
| 155 |
|
| 156 | void setNumLanes(const std::size_t) {}
|
| 157 |
|
| 158 | unique_lock_type guard(const lane_id) const {
|
| 159 | return TrivialLock();
|
| 160 | }
|
| 161 |
|
| 162 | void lock(const lane_id) const {
|
| 163 | return;
|
| 164 | }
|
| 165 |
|
| 166 | void unlock(const lane_id) const {
|
| 167 | return;
|
| 168 | }
|
| 169 |
|
| 170 | void beforeLockAllBut(const lane_id) const {
|
| 171 | return;
|
| 172 | }
|
| 173 |
|
| 174 | void beforeUnlockAllBut(const lane_id) const {
|
| 175 | return;
|
| 176 | }
|
| 177 |
|
| 178 | void lockAllBut(const lane_id) const {
|
| 179 | return;
|
| 180 | }
|
| 181 |
|
| 182 | void unlockAllBut(const lane_id) const {
|
| 183 | return;
|
| 184 | }
|
| 185 | };
|
| 186 |
|
| 187 | #ifdef IS_PARALLEL
|
| 188 |
|
| 189 | /**
|
| 190 | * A small utility class for implementing simple locks.
|
| 191 | */
|
| 192 | class Lock {
|
| 193 | // the underlying mutex
|
| 194 | std::mutex mux;
|
| 195 |
|
| 196 | public:
|
| 197 | struct Lease {
|
| 198 | Lease(std::mutex& mux) : mux(&mux) {
|
| 199 | mux.lock();
|
| 200 | }
|
| 201 | Lease(Lease&& other) : mux(other.mux) {
|
| 202 | other.mux = nullptr;
|
| 203 | }
|
| 204 | Lease(const Lease& other) = delete;
|
| 205 | ~Lease() {
|
| 206 | if (mux != nullptr) {
|
| 207 | mux->unlock();
|
| 208 | }
|
| 209 | }
|
| 210 |
|
| 211 | protected:
|
| 212 | std::mutex* mux;
|
| 213 | };
|
| 214 |
|
| 215 | // acquired the lock for the live-cycle of the returned guard
|
| 216 | Lease acquire() {
|
| 217 | return Lease(mux);
|
| 218 | }
|
| 219 |
|
| 220 | void lock() {
|
| 221 | mux.lock();
|
| 222 | }
|
| 223 |
|
| 224 | bool try_lock() {
|
| 225 | return mux.try_lock();
|
| 226 | }
|
| 227 |
|
| 228 | void unlock() {
|
| 229 | mux.unlock();
|
| 230 | }
|
| 231 | };
|
| 232 |
|
| 233 | // /* valuable source: http://locklessinc.com/articles/locks/ */
|
| 234 |
|
| 235 | namespace detail {
|
| 236 |
|
| 237 | /* Pause instruction to prevent excess processor bus usage */
|
| 238 | #if defined _MSC_VER
|
| 239 | #define cpu_relax() YieldProcessor()
|
| 240 | #else
|
| 241 | #ifdef __x86_64__
|
| 242 | #define cpu_relax() asm volatile("pause\n" : : : "memory")
|
| 243 | #else
|
| 244 | #define cpu_relax() asm volatile("" : : : "memory")
|
| 245 | #endif
|
| 246 | #endif
|
| 247 |
|
| 248 | /**
|
| 249 | * A utility class managing waiting operations for spin locks.
|
| 250 | */
|
| 251 | class Waiter {
|
| 252 | int i = 0;
|
| 253 |
|
| 254 | public:
|
| 255 | Waiter() = default;
|
| 256 |
|
| 257 | /**
|
| 258 | * Conducts a wait operation.
|
| 259 | */
|
| 260 | void operator()() {
|
| 261 | ++i;
|
| 262 | if ((i % 1000) == 0) {
|
| 263 | // there was no progress => let others work
|
| 264 | pthread_yield();
|
| 265 | } else {
|
| 266 | // relax this CPU
|
| 267 | cpu_relax();
|
| 268 | }
|
| 269 | }
|
| 270 | };
|
| 271 | } // namespace detail
|
| 272 |
|
| 273 | /* compare: http://en.cppreference.com/w/cpp/atomic/atomic_flag */
|
| 274 | class SpinLock {
|
| 275 | std::atomic<int> lck{0};
|
| 276 |
|
| 277 | public:
|
| 278 | SpinLock() = default;
|
| 279 |
|
| 280 | void lock() {
|
| 281 | detail::Waiter wait;
|
| 282 | while (!try_lock()) {
|
| 283 | wait();
|
| 284 | }
|
| 285 | }
|
| 286 |
|
| 287 | bool try_lock() {
|
| 288 | int should = 0;
|
| 289 | return lck.compare_exchange_weak(should, 1, std::memory_order_acquire);
|
| 290 | }
|
| 291 |
|
| 292 | void unlock() {
|
| 293 | lck.store(0, std::memory_order_release);
|
| 294 | }
|
| 295 | };
|
| 296 |
|
| 297 | /**
|
| 298 | * A read/write lock for increased access performance on a
|
| 299 | * read-heavy use case.
|
| 300 | */
|
| 301 | class ReadWriteLock {
|
| 302 | /**
|
| 303 | * Based on paper:
|
| 304 | * Scalable Reader-Writer Synchronization
|
| 305 | * for Shared-Memory Multiprocessors
|
| 306 | *
|
| 307 | * Layout of the lock:
|
| 308 | * 31 ... 2 1 0
|
| 309 | * +-------------------------+--------------------+--------------------+
|
| 310 | * | interested reader count | waiting writer | active writer flag |
|
| 311 | * +-------------------------+--------------------+--------------------+
|
| 312 | */
|
| 313 |
|
| 314 | std::atomic<int> lck{0};
|
| 315 |
|
| 316 | public:
|
| 317 | ReadWriteLock() = default;
|
| 318 |
|
| 319 | void start_read() {
|
| 320 | // add reader
|
| 321 | auto r = lck.fetch_add(4, std::memory_order_acquire);
|
| 322 |
|
| 323 | // wait until there is no writer any more
|
| 324 | detail::Waiter wait;
|
| 325 | while (r & 0x3) {
|
| 326 | // release reader
|
| 327 | end_read();
|
| 328 |
|
| 329 | // wait a bit
|
| 330 | wait();
|
| 331 |
|
| 332 | // apply as a reader again
|
| 333 | r = lck.fetch_add(4, std::memory_order_acquire);
|
| 334 |
|
| 335 | } // while there is a writer => spin
|
| 336 | }
|
| 337 |
|
| 338 | void end_read() {
|
| 339 | lck.fetch_sub(4, std::memory_order_release);
|
| 340 | }
|
| 341 |
|
| 342 | void start_write() {
|
| 343 | detail::Waiter wait;
|
| 344 |
|
| 345 | // set wait-for-write bit
|
| 346 | auto stat = lck.fetch_or(2, std::memory_order_acquire);
|
| 347 | while (stat & 0x2) {
|
| 348 | wait();
|
| 349 | stat = lck.fetch_or(2, std::memory_order_acquire);
|
| 350 | }
|
| 351 |
|
| 352 | // the caller may starve here ...
|
| 353 | int should = 2;
|
| 354 | while (!lck.compare_exchange_strong(
|
| 355 | should, 1, std::memory_order_acquire, std::memory_order_relaxed)) {
|
| 356 | wait();
|
| 357 | should = 2;
|
| 358 | }
|
| 359 | }
|
| 360 |
|
| 361 | bool try_write() {
|
| 362 | int should = 0;
|
| 363 | return lck.compare_exchange_strong(should, 1, std::memory_order_acquire, std::memory_order_relaxed);
|
| 364 | }
|
| 365 |
|
| 366 | void end_write() {
|
| 367 | lck.fetch_sub(1, std::memory_order_release);
|
| 368 | }
|
| 369 |
|
| 370 | bool try_upgrade_to_write() {
|
| 371 | int should = 4;
|
| 372 | return lck.compare_exchange_strong(should, 1, std::memory_order_acquire, std::memory_order_relaxed);
|
| 373 | }
|
| 374 |
|
| 375 | void downgrade_to_read() {
|
| 376 | // delete write bit + set num readers to 1
|
| 377 | lck.fetch_add(3, std::memory_order_release);
|
| 378 | }
|
| 379 | };
|
| 380 |
|
| 381 | /**
|
| 382 | * An implementation of an optimistic r/w lock.
|
| 383 | */
|
| 384 | class OptimisticReadWriteLock {
|
| 385 | /**
|
| 386 | * The version number utilized for the synchronization.
|
| 387 | *
|
| 388 | * Usage:
|
| 389 | * - even version numbers are stable versions, not being updated
|
| 390 | * - odd version numbers are temporary versions, currently being updated
|
| 391 | */
|
| 392 | std::atomic<int> version{0};
|
| 393 |
|
| 394 | public:
|
| 395 | /**
|
| 396 | * The lease utilized to link start and end of read phases.
|
| 397 | */
|
| 398 | class Lease {
|
| 399 | friend class OptimisticReadWriteLock;
|
| 400 | int version;
|
| 401 |
|
| 402 | public:
|
| 403 | Lease(int version = 0) : version(version) {}
|
| 404 | Lease(const Lease& lease) = default;
|
| 405 | Lease& operator=(const Lease& other) = default;
|
| 406 | Lease& operator=(Lease&& other) = default;
|
| 407 | };
|
| 408 |
|
| 409 | /**
|
| 410 | * A default constructor initializing the lock.
|
| 411 | */
|
| 412 | OptimisticReadWriteLock() = default;
|
| 413 |
|
| 414 | /**
|
| 415 | * Starts a read phase, making sure that there is currently no
|
| 416 | * active concurrent modification going on. The resulting lease
|
| 417 | * enables the invoking process to later-on verify that no
|
| 418 | * concurrent modifications took place.
|
| 419 | */
|
| 420 | Lease start_read() {
|
| 421 | detail::Waiter wait;
|
| 422 |
|
| 423 | // get a snapshot of the lease version
|
| 424 | auto v = version.load(std::memory_order_acquire);
|
| 425 |
|
| 426 | // spin while there is a write in progress
|
| 427 | while ((v & 0x1) == 1) {
|
| 428 | // wait for a moment
|
| 429 | wait();
|
| 430 | // get an updated version
|
| 431 | v = version.load(std::memory_order_acquire);
|
| 432 | }
|
| 433 |
|
| 434 | // done
|
| 435 | return Lease(v);
|
| 436 | }
|
| 437 |
|
| 438 | /**
|
| 439 | * Tests whether there have been concurrent modifications since
|
| 440 | * the given lease has been issued.
|
| 441 | *
|
| 442 | * @return true if no updates have been conducted, false otherwise
|
| 443 | */
|
| 444 | bool validate(const Lease& lease) {
|
| 445 | // check whether version number has changed in the mean-while
|
| 446 | std::atomic_thread_fence(std::memory_order_acquire);
|
| 447 | return lease.version == version.load(std::memory_order_relaxed);
|
| 448 | }
|
| 449 |
|
| 450 | /**
|
| 451 | * Ends a read phase by validating the given lease.
|
| 452 | *
|
| 453 | * @return true if no updates have been conducted since the
|
| 454 | * issuing of the lease, false otherwise
|
| 455 | */
|
| 456 | bool end_read(const Lease& lease) {
|
| 457 | // check lease in the end
|
| 458 | return validate(lease);
|
| 459 | }
|
| 460 |
|
| 461 | /**
|
| 462 | * Starts a write phase on this lock be ensuring exclusive access
|
| 463 | * and invalidating any existing read lease.
|
| 464 | */
|
| 465 | void start_write() {
|
| 466 | detail::Waiter wait;
|
| 467 |
|
| 468 | // set last bit => make it odd
|
| 469 | auto v = version.fetch_or(0x1, std::memory_order_acquire);
|
| 470 |
|
| 471 | // check for concurrent writes
|
| 472 | while ((v & 0x1) == 1) {
|
| 473 | // wait for a moment
|
| 474 | wait();
|
| 475 | // get an updated version
|
| 476 | v = version.fetch_or(0x1, std::memory_order_acquire);
|
| 477 | }
|
| 478 |
|
| 479 | // done
|
| 480 | }
|
| 481 |
|
| 482 | /**
|
| 483 | * Tries to start a write phase unless there is a currently ongoing
|
| 484 | * write operation. In this case no write permission will be obtained.
|
| 485 | *
|
| 486 | * @return true if write permission has been granted, false otherwise.
|
| 487 | */
|
| 488 | bool try_start_write() {
|
| 489 | auto v = version.fetch_or(0x1, std::memory_order_acquire);
|
| 490 | return !(v & 0x1);
|
| 491 | }
|
| 492 |
|
| 493 | /**
|
| 494 | * Updates a read-lease to a write permission by a) validating that the
|
| 495 | * given lease is still valid and b) making sure that there is no currently
|
| 496 | * ongoing write operation.
|
| 497 | *
|
| 498 | * @return true if the lease was still valid and write permissions could
|
| 499 | * be granted, false otherwise.
|
| 500 | */
|
| 501 | bool try_upgrade_to_write(const Lease& lease) {
|
| 502 | auto v = version.fetch_or(0x1, std::memory_order_acquire);
|
| 503 |
|
| 504 | // check whether write privileges have been gained
|
| 505 | if (v & 0x1) return false; // there is another writer already
|
| 506 |
|
| 507 | // check whether there was no write since the gain of the read lock
|
| 508 | if (lease.version == v) return true;
|
| 509 |
|
| 510 | // if there was, undo write update
|
| 511 | abort_write();
|
| 512 |
|
| 513 | // operation failed
|
| 514 | return false;
|
| 515 | }
|
| 516 |
|
| 517 | /**
|
| 518 | * Aborts a write operation by reverting to the version number before
|
| 519 | * starting the ongoing write, thereby re-validating existing leases.
|
| 520 | */
|
| 521 | void abort_write() {
|
| 522 | // reset version number
|
| 523 | version.fetch_sub(1, std::memory_order_release);
|
| 524 | }
|
| 525 |
|
| 526 | /**
|
| 527 | * Ends a write operation by giving up the associated exclusive access
|
| 528 | * to the protected data and abandoning the provided write permission.
|
| 529 | */
|
| 530 | void end_write() {
|
| 531 | // update version number another time
|
| 532 | version.fetch_add(1, std::memory_order_release);
|
| 533 | }
|
| 534 |
|
| 535 | /**
|
| 536 | * Tests whether currently write permissions have been granted to any
|
| 537 | * client by this lock.
|
| 538 | *
|
| 539 | * @return true if so, false otherwise
|
| 540 | */
|
| 541 | bool is_write_locked() const {
|
| 542 | return version & 0x1;
|
| 543 | }
|
| 544 | };
|
| 545 |
|
| 546 | /** Concurrent lanes locking mechanism. */
|
| 547 | struct MutexConcurrentLanes {
|
| 548 | using lane_id = std::size_t;
|
| 549 | using unique_lock_type = std::unique_lock<std::mutex>;
|
| 550 |
|
| 551 | explicit MutexConcurrentLanes(const std::size_t Sz) : Size(Sz), Attribution(attribution(Sz)) {
|
| 552 | Lanes = std::make_unique<Lane[]>(Sz);
|
| 553 | }
|
| 554 | MutexConcurrentLanes(const MutexConcurrentLanes&) = delete;
|
| 555 | MutexConcurrentLanes(MutexConcurrentLanes&&) = delete;
|
| 556 |
|
| 557 | virtual ~MutexConcurrentLanes() {}
|
| 558 |
|
| 559 | // Return the number of lanes.
|
| 560 | std::size_t lanes() const {
|
| 561 | return Size;
|
| 562 | }
|
| 563 |
|
| 564 | // Select a lane
|
| 565 | lane_id getLane(std::size_t I) const {
|
| 566 | if (Attribution == lane_attribution::mod_power_of_2) {
|
| 567 | return I & (Size - 1);
|
| 568 | } else {
|
| 569 | return I % Size;
|
| 570 | }
|
| 571 | }
|
| 572 |
|
| 573 | /** Change the number of lanes.
|
| 574 | * DO not use while threads are using this object.
|
| 575 | */
|
| 576 | void setNumLanes(const std::size_t NumLanes) {
|
| 577 | Size = (NumLanes == 0 ? 1 : NumLanes);
|
| 578 | Attribution = attribution(Size);
|
| 579 | Lanes = std::make_unique<Lane[]>(Size);
|
| 580 | }
|
| 581 |
|
| 582 | unique_lock_type guard(const lane_id Lane) const {
|
| 583 | return unique_lock_type(Lanes[Lane].Access);
|
| 584 | }
|
| 585 |
|
| 586 | // Lock the given lane.
|
| 587 | // Must eventually be followed by unlock(Lane).
|
| 588 | void lock(const lane_id Lane) const {
|
| 589 | Lanes[Lane].Access.lock();
|
| 590 | }
|
| 591 |
|
| 592 | // Unlock the given lane.
|
| 593 | // Must already be the owner of the lane's lock.
|
| 594 | void unlock(const lane_id Lane) const {
|
| 595 | Lanes[Lane].Access.unlock();
|
| 596 | }
|
| 597 |
|
| 598 | // Acquire the capability to lock all other lanes than the given one.
|
| 599 | //
|
| 600 | // Must eventually be followed by beforeUnlockAllBut(Lane).
|
| 601 | void beforeLockAllBut(const lane_id Lane) const {
|
| 602 | if (!BeforeLockAll.try_lock()) {
|
| 603 | // If we cannot get the lock immediately, it means it was acquired
|
| 604 | // concurrently by another lane that will also try to acquire our
|
| 605 | // lane lock.
|
| 606 | // So we release our lane lock to let the concurrent operation
|
| 607 | // progress.
|
| 608 | unlock(Lane);
|
| 609 | BeforeLockAll.lock();
|
| 610 | lock(Lane);
|
| 611 | }
|
| 612 | }
|
| 613 |
|
| 614 | // Release the capability to lock all other lanes than the given one.
|
| 615 | //
|
| 616 | // Must already be the owner of that capability.
|
| 617 | void beforeUnlockAllBut(const lane_id) const {
|
| 618 | BeforeLockAll.unlock();
|
| 619 | }
|
| 620 |
|
| 621 | // Lock all lanes but the given one.
|
| 622 | //
|
| 623 | // Must already have acquired the capability to lock all other lanes
|
| 624 | // by calling beforeLockAllBut(Lane).
|
| 625 | //
|
| 626 | // Must eventually be followed by unlockAllBut(Lane).
|
| 627 | void lockAllBut(const lane_id Lane) const {
|
| 628 | for (std::size_t I = 0; I < Size; ++I) {
|
| 629 | if (I != Lane) {
|
| 630 | Lanes[I].Access.lock();
|
| 631 | }
|
| 632 | }
|
| 633 | }
|
| 634 |
|
| 635 | // Unlock all lanes but the given one.
|
| 636 | // Must already be the owner of all the lanes' locks.
|
| 637 | void unlockAllBut(const lane_id Lane) const {
|
| 638 | for (std::size_t I = 0; I < Size; ++I) {
|
| 639 | if (I != Lane) {
|
| 640 | Lanes[I].Access.unlock();
|
| 641 | }
|
| 642 | }
|
| 643 | }
|
| 644 |
|
| 645 | private:
|
| 646 | enum lane_attribution { mod_power_of_2, mod_other };
|
| 647 |
|
| 648 | struct Lane {
|
| 649 | alignas(hardware_destructive_interference_size) std::mutex Access;
|
| 650 | };
|
| 651 |
|
| 652 | static constexpr lane_attribution attribution(const std::size_t Sz) {
|
| 653 | assert(Sz > 0);
|
| 654 | if ((Sz & (Sz - 1)) == 0) {
|
| 655 | // Sz is a power of 2
|
| 656 | return lane_attribution::mod_power_of_2;
|
| 657 | } else {
|
| 658 | return lane_attribution::mod_other;
|
| 659 | }
|
| 660 | }
|
| 661 |
|
| 662 | protected:
|
| 663 | std::size_t Size;
|
| 664 | lane_attribution Attribution;
|
| 665 |
|
| 666 | private:
|
| 667 | mutable std::unique_ptr<Lane[]> Lanes;
|
| 668 |
|
| 669 | alignas(hardware_destructive_interference_size) mutable std::mutex BeforeLockAll;
|
| 670 | };
|
| 671 |
|
| 672 | class ConcurrentLanes : public MutexConcurrentLanes {
|
| 673 | using Base = MutexConcurrentLanes;
|
| 674 |
|
| 675 | public:
|
| 676 | using lane_id = Base::lane_id;
|
| 677 | using Base::beforeLockAllBut;
|
| 678 | using Base::beforeUnlockAllBut;
|
| 679 | using Base::guard;
|
| 680 | using Base::lock;
|
| 681 | using Base::lockAllBut;
|
| 682 | using Base::unlock;
|
| 683 | using Base::unlockAllBut;
|
| 684 |
|
| 685 | explicit ConcurrentLanes(const std::size_t Sz) : MutexConcurrentLanes(Sz) {}
|
| 686 | ConcurrentLanes(const ConcurrentLanes&) = delete;
|
| 687 | ConcurrentLanes(ConcurrentLanes&&) = delete;
|
| 688 |
|
| 689 | lane_id threadLane() const {
|
| 690 | return getLane(static_cast<std::size_t>(omp_get_thread_num()));
|
| 691 | }
|
| 692 |
|
| 693 | void setNumLanes(const std::size_t NumLanes) {
|
| 694 | Base::setNumLanes(NumLanes == 0 ? omp_get_max_threads() : NumLanes);
|
| 695 | }
|
| 696 |
|
| 697 | unique_lock_type guard() const {
|
| 698 | return Base::guard(threadLane());
|
| 699 | }
|
| 700 |
|
| 701 | void lock() const {
|
| 702 | return Base::lock(threadLane());
|
| 703 | }
|
| 704 |
|
| 705 | void unlock() const {
|
| 706 | return Base::unlock(threadLane());
|
| 707 | }
|
| 708 |
|
| 709 | void beforeLockAllBut() const {
|
| 710 | return Base::beforeLockAllBut(threadLane());
|
| 711 | }
|
| 712 |
|
| 713 | void beforeUnlockAllBut() const {
|
| 714 | return Base::beforeUnlockAllBut(threadLane());
|
| 715 | }
|
| 716 |
|
| 717 | void lockAllBut() const {
|
| 718 | return Base::lockAllBut(threadLane());
|
| 719 | }
|
| 720 |
|
| 721 | void unlockAllBut() const {
|
| 722 | return Base::unlockAllBut(threadLane());
|
| 723 | }
|
| 724 | };
|
| 725 |
|
| 726 | #else
|
| 727 |
|
| 728 | /**
|
| 729 | * A small utility class for implementing simple locks.
|
| 730 | */
|
| 731 | struct Lock {
|
| 732 | class Lease {};
|
| 733 |
|
| 734 | // no locking if there is no parallel execution
|
| 735 | Lease acquire() {
|
| 736 | return Lease();
|
| 737 | }
|
| 738 |
|
| 739 | void lock() {}
|
| 740 |
|
| 741 | bool try_lock() {
|
| 742 | return true;
|
| 743 | }
|
| 744 |
|
| 745 | void unlock() {}
|
| 746 | };
|
| 747 |
|
| 748 | /**
|
| 749 | * A 'sequential' non-locking implementation for a spin lock.
|
| 750 | */
|
| 751 | class SpinLock {
|
| 752 | public:
|
| 753 | SpinLock() = default;
|
| 754 |
|
| 755 | void lock() {}
|
| 756 |
|
| 757 | bool try_lock() {
|
| 758 | return true;
|
| 759 | }
|
| 760 |
|
| 761 | void unlock() {}
|
| 762 | };
|
| 763 |
|
| 764 | class ReadWriteLock {
|
| 765 | public:
|
| 766 | ReadWriteLock() = default;
|
| 767 |
|
| 768 | void start_read() {}
|
| 769 |
|
| 770 | void end_read() {}
|
| 771 |
|
| 772 | void start_write() {}
|
| 773 |
|
| 774 | bool try_write() {
|
| 775 | return true;
|
| 776 | }
|
| 777 |
|
| 778 | void end_write() {}
|
| 779 |
|
| 780 | bool try_upgrade_to_write() {
|
| 781 | return true;
|
| 782 | }
|
| 783 |
|
| 784 | void downgrade_to_read() {}
|
| 785 | };
|
| 786 |
|
| 787 | /**
|
| 788 | * A 'sequential' non-locking implementation for an optimistic r/w lock.
|
| 789 | */
|
| 790 | class OptimisticReadWriteLock {
|
| 791 | public:
|
| 792 | class Lease {};
|
| 793 |
|
| 794 | OptimisticReadWriteLock() = default;
|
| 795 |
|
| 796 | Lease start_read() {
|
| 797 | return Lease();
|
| 798 | }
|
| 799 |
|
| 800 | bool validate(const Lease& /*lease*/) {
|
| 801 | return true;
|
| 802 | }
|
| 803 |
|
| 804 | bool end_read(const Lease& /*lease*/) {
|
| 805 | return true;
|
| 806 | }
|
| 807 |
|
| 808 | void start_write() {}
|
| 809 |
|
| 810 | bool try_start_write() {
|
| 811 | return true;
|
| 812 | }
|
| 813 |
|
| 814 | bool try_upgrade_to_write(const Lease& /*lease*/) {
|
| 815 | return true;
|
| 816 | }
|
| 817 |
|
| 818 | void abort_write() {}
|
| 819 |
|
| 820 | void end_write() {}
|
| 821 |
|
| 822 | bool is_write_locked() const {
|
| 823 | return true;
|
| 824 | }
|
| 825 | };
|
| 826 |
|
| 827 | struct ConcurrentLanes : protected SeqConcurrentLanes {
|
| 828 | using Base = SeqConcurrentLanes;
|
| 829 | using lane_id = SeqConcurrentLanes::lane_id;
|
| 830 | using unique_lock_type = SeqConcurrentLanes::unique_lock_type;
|
| 831 |
|
| 832 | using Base::lanes;
|
| 833 | using Base::setNumLanes;
|
| 834 |
|
| 835 | explicit ConcurrentLanes(std::size_t Sz = MAX_THREADS) : Base(Sz) {}
|
| 836 | ConcurrentLanes(const ConcurrentLanes&) = delete;
|
| 837 | ConcurrentLanes(ConcurrentLanes&&) = delete;
|
| 838 |
|
| 839 | virtual ~ConcurrentLanes() {}
|
| 840 |
|
| 841 | lane_id threadLane() const {
|
| 842 | return 0;
|
| 843 | }
|
| 844 |
|
| 845 | unique_lock_type guard() const {
|
| 846 | return Base::guard(threadLane());
|
| 847 | }
|
| 848 |
|
| 849 | void lock() const {
|
| 850 | return Base::lock(threadLane());
|
| 851 | }
|
| 852 |
|
| 853 | void unlock() const {
|
| 854 | return Base::unlock(threadLane());
|
| 855 | }
|
| 856 |
|
| 857 | void beforeLockAllBut() const {
|
| 858 | return Base::beforeLockAllBut(threadLane());
|
| 859 | }
|
| 860 |
|
| 861 | void beforeUnlockAllBut() const {
|
| 862 | return Base::beforeUnlockAllBut(threadLane());
|
| 863 | }
|
| 864 |
|
| 865 | void lockAllBut() const {
|
| 866 | return Base::lockAllBut(threadLane());
|
| 867 | }
|
| 868 |
|
| 869 | void unlockAllBut() const {
|
| 870 | return Base::unlockAllBut(threadLane());
|
| 871 | }
|
| 872 | };
|
| 873 |
|
| 874 | #endif
|
| 875 |
|
| 876 | /**
|
| 877 | * Obtains a reference to the lock synchronizing output operations.
|
| 878 | */
|
| 879 | inline Lock& getOutputLock() {
|
| 880 | static Lock outputLock;
|
| 881 | return outputLock;
|
| 882 | }
|
| 883 |
|
| 884 | } // namespace souffle
|