d247a35da3c63030bca417b89482bb8fdf371cd0
[GitHub/mt8127/android_kernel_alcatel_ttab.git] / drivers / md / persistent-data / dm-transaction-manager.c
1 /*
2 * Copyright (C) 2011 Red Hat, Inc.
3 *
4 * This file is released under the GPL.
5 */
6 #include "dm-transaction-manager.h"
7 #include "dm-space-map.h"
8 #include "dm-space-map-disk.h"
9 #include "dm-space-map-metadata.h"
10 #include "dm-persistent-data-internal.h"
11
12 #include <linux/export.h>
13 #include <linux/slab.h>
14 #include <linux/device-mapper.h>
15
16 #define DM_MSG_PREFIX "transaction manager"
17
18 /*----------------------------------------------------------------*/
19
20 struct shadow_info {
21 struct hlist_node hlist;
22 dm_block_t where;
23 };
24
25 /*
26 * It would be nice if we scaled with the size of transaction.
27 */
28 #define HASH_SIZE 256
29 #define HASH_MASK (HASH_SIZE - 1)
30
31 struct dm_transaction_manager {
32 int is_clone;
33 struct dm_transaction_manager *real;
34
35 struct dm_block_manager *bm;
36 struct dm_space_map *sm;
37
38 spinlock_t lock;
39 struct hlist_head buckets[HASH_SIZE];
40 };
41
42 /*----------------------------------------------------------------*/
43
44 static int is_shadow(struct dm_transaction_manager *tm, dm_block_t b)
45 {
46 int r = 0;
47 unsigned bucket = dm_hash_block(b, HASH_MASK);
48 struct shadow_info *si;
49 struct hlist_node *n;
50
51 spin_lock(&tm->lock);
52 hlist_for_each_entry(si, n, tm->buckets + bucket, hlist)
53 if (si->where == b) {
54 r = 1;
55 break;
56 }
57 spin_unlock(&tm->lock);
58
59 return r;
60 }
61
62 /*
63 * This can silently fail if there's no memory. We're ok with this since
64 * creating redundant shadows causes no harm.
65 */
66 static void insert_shadow(struct dm_transaction_manager *tm, dm_block_t b)
67 {
68 unsigned bucket;
69 struct shadow_info *si;
70
71 si = kmalloc(sizeof(*si), GFP_NOIO);
72 if (si) {
73 si->where = b;
74 bucket = dm_hash_block(b, HASH_MASK);
75 spin_lock(&tm->lock);
76 hlist_add_head(&si->hlist, tm->buckets + bucket);
77 spin_unlock(&tm->lock);
78 }
79 }
80
81 static void wipe_shadow_table(struct dm_transaction_manager *tm)
82 {
83 struct shadow_info *si;
84 struct hlist_node *n, *tmp;
85 struct hlist_head *bucket;
86 int i;
87
88 spin_lock(&tm->lock);
89 for (i = 0; i < HASH_SIZE; i++) {
90 bucket = tm->buckets + i;
91 hlist_for_each_entry_safe(si, n, tmp, bucket, hlist)
92 kfree(si);
93
94 INIT_HLIST_HEAD(bucket);
95 }
96
97 spin_unlock(&tm->lock);
98 }
99
100 /*----------------------------------------------------------------*/
101
102 static struct dm_transaction_manager *dm_tm_create(struct dm_block_manager *bm,
103 struct dm_space_map *sm)
104 {
105 int i;
106 struct dm_transaction_manager *tm;
107
108 tm = kmalloc(sizeof(*tm), GFP_KERNEL);
109 if (!tm)
110 return ERR_PTR(-ENOMEM);
111
112 tm->is_clone = 0;
113 tm->real = NULL;
114 tm->bm = bm;
115 tm->sm = sm;
116
117 spin_lock_init(&tm->lock);
118 for (i = 0; i < HASH_SIZE; i++)
119 INIT_HLIST_HEAD(tm->buckets + i);
120
121 return tm;
122 }
123
124 struct dm_transaction_manager *dm_tm_create_non_blocking_clone(struct dm_transaction_manager *real)
125 {
126 struct dm_transaction_manager *tm;
127
128 tm = kmalloc(sizeof(*tm), GFP_KERNEL);
129 if (tm) {
130 tm->is_clone = 1;
131 tm->real = real;
132 }
133
134 return tm;
135 }
136 EXPORT_SYMBOL_GPL(dm_tm_create_non_blocking_clone);
137
138 void dm_tm_destroy(struct dm_transaction_manager *tm)
139 {
140 if (!tm->is_clone)
141 wipe_shadow_table(tm);
142
143 kfree(tm);
144 }
145 EXPORT_SYMBOL_GPL(dm_tm_destroy);
146
147 int dm_tm_pre_commit(struct dm_transaction_manager *tm)
148 {
149 int r;
150
151 if (tm->is_clone)
152 return -EWOULDBLOCK;
153
154 r = dm_sm_commit(tm->sm);
155 if (r < 0)
156 return r;
157
158 return 0;
159 }
160 EXPORT_SYMBOL_GPL(dm_tm_pre_commit);
161
162 int dm_tm_commit(struct dm_transaction_manager *tm, struct dm_block *root)
163 {
164 if (tm->is_clone)
165 return -EWOULDBLOCK;
166
167 wipe_shadow_table(tm);
168
169 return dm_bm_flush_and_unlock(tm->bm, root);
170 }
171 EXPORT_SYMBOL_GPL(dm_tm_commit);
172
173 int dm_tm_new_block(struct dm_transaction_manager *tm,
174 struct dm_block_validator *v,
175 struct dm_block **result)
176 {
177 int r;
178 dm_block_t new_block;
179
180 if (tm->is_clone)
181 return -EWOULDBLOCK;
182
183 r = dm_sm_new_block(tm->sm, &new_block);
184 if (r < 0)
185 return r;
186
187 r = dm_bm_write_lock_zero(tm->bm, new_block, v, result);
188 if (r < 0) {
189 dm_sm_dec_block(tm->sm, new_block);
190 return r;
191 }
192
193 /*
194 * New blocks count as shadows in that they don't need to be
195 * shadowed again.
196 */
197 insert_shadow(tm, new_block);
198
199 return 0;
200 }
201
202 static int __shadow_block(struct dm_transaction_manager *tm, dm_block_t orig,
203 struct dm_block_validator *v,
204 struct dm_block **result)
205 {
206 int r;
207 dm_block_t new;
208 struct dm_block *orig_block;
209
210 r = dm_sm_new_block(tm->sm, &new);
211 if (r < 0)
212 return r;
213
214 r = dm_sm_dec_block(tm->sm, orig);
215 if (r < 0)
216 return r;
217
218 r = dm_bm_read_lock(tm->bm, orig, v, &orig_block);
219 if (r < 0)
220 return r;
221
222 /*
223 * It would be tempting to use dm_bm_unlock_move here, but some
224 * code, such as the space maps, keeps using the old data structures
225 * secure in the knowledge they won't be changed until the next
226 * transaction. Using unlock_move would force a synchronous read
227 * since the old block would no longer be in the cache.
228 */
229 r = dm_bm_write_lock_zero(tm->bm, new, v, result);
230 if (r) {
231 dm_bm_unlock(orig_block);
232 return r;
233 }
234
235 memcpy(dm_block_data(*result), dm_block_data(orig_block),
236 dm_bm_block_size(tm->bm));
237
238 dm_bm_unlock(orig_block);
239 return r;
240 }
241
242 int dm_tm_shadow_block(struct dm_transaction_manager *tm, dm_block_t orig,
243 struct dm_block_validator *v, struct dm_block **result,
244 int *inc_children)
245 {
246 int r;
247
248 if (tm->is_clone)
249 return -EWOULDBLOCK;
250
251 r = dm_sm_count_is_more_than_one(tm->sm, orig, inc_children);
252 if (r < 0)
253 return r;
254
255 if (is_shadow(tm, orig) && !*inc_children)
256 return dm_bm_write_lock(tm->bm, orig, v, result);
257
258 r = __shadow_block(tm, orig, v, result);
259 if (r < 0)
260 return r;
261 insert_shadow(tm, dm_block_location(*result));
262
263 return r;
264 }
265 EXPORT_SYMBOL_GPL(dm_tm_shadow_block);
266
267 int dm_tm_read_lock(struct dm_transaction_manager *tm, dm_block_t b,
268 struct dm_block_validator *v,
269 struct dm_block **blk)
270 {
271 if (tm->is_clone)
272 return dm_bm_read_try_lock(tm->real->bm, b, v, blk);
273
274 return dm_bm_read_lock(tm->bm, b, v, blk);
275 }
276 EXPORT_SYMBOL_GPL(dm_tm_read_lock);
277
278 int dm_tm_unlock(struct dm_transaction_manager *tm, struct dm_block *b)
279 {
280 return dm_bm_unlock(b);
281 }
282 EXPORT_SYMBOL_GPL(dm_tm_unlock);
283
284 void dm_tm_inc(struct dm_transaction_manager *tm, dm_block_t b)
285 {
286 /*
287 * The non-blocking clone doesn't support this.
288 */
289 BUG_ON(tm->is_clone);
290
291 dm_sm_inc_block(tm->sm, b);
292 }
293 EXPORT_SYMBOL_GPL(dm_tm_inc);
294
295 void dm_tm_dec(struct dm_transaction_manager *tm, dm_block_t b)
296 {
297 /*
298 * The non-blocking clone doesn't support this.
299 */
300 BUG_ON(tm->is_clone);
301
302 dm_sm_dec_block(tm->sm, b);
303 }
304 EXPORT_SYMBOL_GPL(dm_tm_dec);
305
306 int dm_tm_ref(struct dm_transaction_manager *tm, dm_block_t b,
307 uint32_t *result)
308 {
309 if (tm->is_clone)
310 return -EWOULDBLOCK;
311
312 return dm_sm_get_count(tm->sm, b, result);
313 }
314
315 struct dm_block_manager *dm_tm_get_bm(struct dm_transaction_manager *tm)
316 {
317 return tm->bm;
318 }
319
320 /*----------------------------------------------------------------*/
321
322 static int dm_tm_create_internal(struct dm_block_manager *bm,
323 dm_block_t sb_location,
324 struct dm_transaction_manager **tm,
325 struct dm_space_map **sm,
326 int create,
327 void *sm_root, size_t sm_len)
328 {
329 int r;
330
331 *sm = dm_sm_metadata_init();
332 if (IS_ERR(*sm))
333 return PTR_ERR(*sm);
334
335 *tm = dm_tm_create(bm, *sm);
336 if (IS_ERR(*tm)) {
337 dm_sm_destroy(*sm);
338 return PTR_ERR(*tm);
339 }
340
341 if (create) {
342 r = dm_sm_metadata_create(*sm, *tm, dm_bm_nr_blocks(bm),
343 sb_location);
344 if (r) {
345 DMERR("couldn't create metadata space map");
346 goto bad;
347 }
348
349 } else {
350 r = dm_sm_metadata_open(*sm, *tm, sm_root, sm_len);
351 if (r) {
352 DMERR("couldn't open metadata space map");
353 goto bad;
354 }
355 }
356
357 return 0;
358
359 bad:
360 dm_tm_destroy(*tm);
361 dm_sm_destroy(*sm);
362 return r;
363 }
364
365 int dm_tm_create_with_sm(struct dm_block_manager *bm, dm_block_t sb_location,
366 struct dm_transaction_manager **tm,
367 struct dm_space_map **sm)
368 {
369 return dm_tm_create_internal(bm, sb_location, tm, sm, 1, NULL, 0);
370 }
371 EXPORT_SYMBOL_GPL(dm_tm_create_with_sm);
372
373 int dm_tm_open_with_sm(struct dm_block_manager *bm, dm_block_t sb_location,
374 void *sm_root, size_t root_len,
375 struct dm_transaction_manager **tm,
376 struct dm_space_map **sm)
377 {
378 return dm_tm_create_internal(bm, sb_location, tm, sm, 0, sm_root, root_len);
379 }
380 EXPORT_SYMBOL_GPL(dm_tm_open_with_sm);
381
382 /*----------------------------------------------------------------*/