-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathsafe_ptr.h
247 lines (205 loc) · 10.7 KB
/
safe_ptr.h
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
#pragma once
#ifndef SAFE_PTR_H
#define SAFE_PTR_H
#include <iostream>
#include <string>
#include <vector>
#include <atomic>
#include <memory>
#include <mutex>
#include <thread>
#include <map>
#include <unordered_map>
#include <condition_variable>
#include <array>
#include <sstream>
#include <cassert>
#include <random>
#include <iomanip>
#include <algorithm>
//#define SHARED_MTX // C++14
#ifdef SHARED_MTX
#include <shared_mutex> // C++14
#endif
namespace sf {
// contention free shared mutex (same-lock-type is recursive for X->X, X->S or S->S locks), but (S->X - is UB)
template<unsigned contention_free_count = 36, bool shared_flag = false>
class contention_free_shared_mutex {
std::atomic<bool> want_x_lock;
//struct cont_free_flag_t { alignas(std::hardware_destructive_interference_size) std::atomic<int> value; cont_free_flag_t() { value = 0; } }; // C++17
struct cont_free_flag_t { char tmp[60]; std::atomic<int> value; cont_free_flag_t() { value = 0; } }; // tmp[] to avoid false sharing
typedef std::array<cont_free_flag_t, contention_free_count> array_slock_t;
const std::shared_ptr<array_slock_t> shared_locks_array_ptr; // 0 - unregistred, 1 registred & free, 2... - busy
char avoid_falsesharing_1[64];
array_slock_t &shared_locks_array;
char avoid_falsesharing_2[64];
int recursive_xlock_count;
enum index_op_t { unregister_thread_op, get_index_op, register_thread_op };
#if (_WIN32 && _MSC_VER < 1900) // only for MSVS 2013
typedef int64_t thread_id_t;
std::atomic<thread_id_t> owner_thread_id;
std::array<int64_t, contention_free_count> register_thread_array;
int64_t get_fast_this_thread_id() {
static __declspec(thread) int64_t fast_this_thread_id = 0; // MSVS 2013 thread_local partially supported - only POD
if (fast_this_thread_id == 0) {
std::stringstream ss;
ss << std::this_thread::get_id(); // https://connect.microsoft.com/VisualStudio/feedback/details/1558211
fast_this_thread_id = std::stoll(ss.str());
}
return fast_this_thread_id;
}
int get_or_set_index(index_op_t index_op = get_index_op, int set_index = -1) {
if (index_op == get_index_op) { // get index
auto const thread_id = get_fast_this_thread_id();
for (size_t i = 0; i < register_thread_array.size(); ++i) {
if (register_thread_array[i] == thread_id) {
set_index = i; // thread already registred
break;
}
}
}
else if (index_op == register_thread_op) { // register thread
register_thread_array[set_index] = get_fast_this_thread_id();
}
return set_index;
}
#else
typedef std::thread::id thread_id_t;
std::atomic<std::thread::id> owner_thread_id;
std::thread::id get_fast_this_thread_id() { return std::this_thread::get_id(); }
struct unregister_t {
int thread_index;
std::shared_ptr<array_slock_t> array_slock_ptr;
unregister_t(int index, std::shared_ptr<array_slock_t> const& ptr) : thread_index(index), array_slock_ptr(ptr) {}
unregister_t(unregister_t &&src) : thread_index(src.thread_index), array_slock_ptr(std::move(src.array_slock_ptr)) {}
~unregister_t() { if (array_slock_ptr.use_count() > 0) (*array_slock_ptr)[thread_index].value--; }
};
int get_or_set_index(index_op_t index_op = get_index_op, int set_index = -1) {
thread_local static std::unordered_map<void *, unregister_t> thread_local_index_hashmap;
// get thread index - in any cases
auto it = thread_local_index_hashmap.find(this);
if (it != thread_local_index_hashmap.cend())
set_index = it->second.thread_index;
if (index_op == unregister_thread_op) { // unregister thread
if (shared_locks_array[set_index].value == 1) // if isn't shared_lock now
thread_local_index_hashmap.erase(this);
else
return -1;
}
else if (index_op == register_thread_op) { // register thread
thread_local_index_hashmap.emplace(this, unregister_t(set_index, shared_locks_array_ptr));
// remove info about deleted contfree-mutexes
for (auto it = thread_local_index_hashmap.begin(), ite = thread_local_index_hashmap.end(); it != ite;) {
if (it->second.array_slock_ptr->at(it->second.thread_index).value < 0) // if contfree-mtx was deleted
it = thread_local_index_hashmap.erase(it);
else
++it;
}
}
return set_index;
}
#endif
public:
contention_free_shared_mutex() :
shared_locks_array_ptr(std::make_shared<array_slock_t>()), shared_locks_array(*shared_locks_array_ptr), want_x_lock(false), recursive_xlock_count(0),
owner_thread_id(thread_id_t()) {}
~contention_free_shared_mutex() {
for (auto &i : shared_locks_array) i.value = -1;
}
bool unregister_thread() { return get_or_set_index(unregister_thread_op) >= 0; }
int register_thread() {
int cur_index = get_or_set_index();
if (cur_index == -1) {
if (shared_locks_array_ptr.use_count() <= (int)shared_locks_array.size()) // try once to register thread
{
for (size_t i = 0; i < shared_locks_array.size(); ++i) {
int unregistred_value = 0;
if (shared_locks_array[i].value == 0)
if (shared_locks_array[i].value.compare_exchange_strong(unregistred_value, 1)) {
cur_index = i;
get_or_set_index(register_thread_op, cur_index); // thread registred success
break;
}
}
//std::cout << "\n thread_id = " << std::this_thread::get_id() << ", register_thread_index = " << cur_index <<
// ", shared_locks_array[cur_index].value = " << shared_locks_array[cur_index].value << std::endl;
}
}
return cur_index;
}
void lock_shared() {
int const register_index = register_thread();
if (register_index >= 0) {
int recursion_depth = shared_locks_array[register_index].value.load(std::memory_order_acquire);
assert(recursion_depth >= 1);
if (recursion_depth > 1)
shared_locks_array[register_index].value.store(recursion_depth + 1, std::memory_order_release); // if recursive -> release
else {
shared_locks_array[register_index].value.store(recursion_depth + 1, std::memory_order_seq_cst); // if first -> sequential
while (want_x_lock.load(std::memory_order_seq_cst)) {
shared_locks_array[register_index].value.store(recursion_depth, std::memory_order_seq_cst);
for (volatile size_t i = 0; want_x_lock.load(std::memory_order_seq_cst); ++i)
if (i % 100000 == 0) std::this_thread::yield();
shared_locks_array[register_index].value.store(recursion_depth + 1, std::memory_order_seq_cst);
}
}
// (shared_locks_array[register_index] == 2 && want_x_lock == false) || // first shared lock
// (shared_locks_array[register_index] > 2) // recursive shared lock
}
else {
if (owner_thread_id.load(std::memory_order_acquire) != get_fast_this_thread_id()) {
size_t i = 0;
for (bool flag = false; !want_x_lock.compare_exchange_weak(flag, true, std::memory_order_seq_cst); flag = false)
if (++i % 100000 == 0) std::this_thread::yield();
owner_thread_id.store(get_fast_this_thread_id(), std::memory_order_release);
}
++recursive_xlock_count;
}
}
void unlock_shared() {
int const register_index = get_or_set_index();
if (register_index >= 0) {
int const recursion_depth = shared_locks_array[register_index].value.load(std::memory_order_acquire);
assert(recursion_depth > 1);
shared_locks_array[register_index].value.store(recursion_depth - 1, std::memory_order_release);
}
else {
if (--recursive_xlock_count == 0) {
owner_thread_id.store(decltype(owner_thread_id)(), std::memory_order_release);
want_x_lock.store(false, std::memory_order_release);
}
}
}
void lock() {
// forbidden upgrade S-lock to X-lock - this is an excellent opportunity to get deadlock
int const register_index = get_or_set_index();
if (register_index >= 0)
assert(shared_locks_array[register_index].value.load(std::memory_order_acquire) == 1);
if (owner_thread_id.load(std::memory_order_acquire) != get_fast_this_thread_id()) {
size_t i = 0;
for (bool flag = false; !want_x_lock.compare_exchange_weak(flag, true, std::memory_order_seq_cst); flag = false)
if (++i % 1000000 == 0) std::this_thread::yield();
owner_thread_id.store(get_fast_this_thread_id(), std::memory_order_release);
for (auto &i : shared_locks_array)
while (i.value.load(std::memory_order_seq_cst) > 1);
}
++recursive_xlock_count;
}
void unlock() {
assert(recursive_xlock_count > 0);
if (--recursive_xlock_count == 0) {
owner_thread_id.store(decltype(owner_thread_id)(), std::memory_order_release);
want_x_lock.store(false, std::memory_order_release);
}
}
};
template<typename mutex_t>
struct shared_lock_guard {
mutex_t &ref_mtx;
shared_lock_guard(mutex_t &mtx) : ref_mtx(mtx) { ref_mtx.lock_shared(); }
~shared_lock_guard() { ref_mtx.unlock_shared(); }
};
using default_contention_free_shared_mutex = contention_free_shared_mutex<>;
// ---------------------------------------------------------------
}
#endif // #ifndef SAFE_PTR_H