1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23#include <linux/blkdev.h>
24#include <linux/mempool.h>
25#include <linux/export.h>
26#include <linux/bio.h>
27#include <linux/workqueue.h>
28#include <linux/slab.h>
29#include "blk.h"
30
31#define BIP_INLINE_VECS 4
32
33static struct kmem_cache *bip_slab;
34static struct workqueue_struct *kintegrityd_wq;
35
36void blk_flush_integrity(void)
37{
38 flush_workqueue(kintegrityd_wq);
39}
40
41
42
43
44
45
46
47
48
49
50
51struct bio_integrity_payload *bio_integrity_alloc(struct bio *bio,
52 gfp_t gfp_mask,
53 unsigned int nr_vecs)
54{
55 struct bio_integrity_payload *bip;
56 struct bio_set *bs = bio->bi_pool;
57 unsigned inline_vecs;
58
59 if (!bs || !mempool_initialized(&bs->bio_integrity_pool)) {
60 bip = kmalloc(sizeof(struct bio_integrity_payload) +
61 sizeof(struct bio_vec) * nr_vecs, gfp_mask);
62 inline_vecs = nr_vecs;
63 } else {
64 bip = mempool_alloc(&bs->bio_integrity_pool, gfp_mask);
65 inline_vecs = BIP_INLINE_VECS;
66 }
67
68 if (unlikely(!bip))
69 return ERR_PTR(-ENOMEM);
70
71 memset(bip, 0, sizeof(*bip));
72
73 if (nr_vecs > inline_vecs) {
74 unsigned long idx = 0;
75
76 bip->bip_vec = bvec_alloc(gfp_mask, nr_vecs, &idx,
77 &bs->bvec_integrity_pool);
78 if (!bip->bip_vec)
79 goto err;
80 bip->bip_max_vcnt = bvec_nr_vecs(idx);
81 bip->bip_slab = idx;
82 } else {
83 bip->bip_vec = bip->bip_inline_vecs;
84 bip->bip_max_vcnt = inline_vecs;
85 }
86
87 bip->bip_bio = bio;
88 bio->bi_integrity = bip;
89 bio->bi_opf |= REQ_INTEGRITY;
90
91 return bip;
92err:
93 mempool_free(bip, &bs->bio_integrity_pool);
94 return ERR_PTR(-ENOMEM);
95}
96EXPORT_SYMBOL(bio_integrity_alloc);
97
98
99
100
101
102
103
104
105void bio_integrity_free(struct bio *bio)
106{
107 struct bio_integrity_payload *bip = bio_integrity(bio);
108 struct bio_set *bs = bio->bi_pool;
109
110 if (bip->bip_flags & BIP_BLOCK_INTEGRITY)
111 kfree(page_address(bip->bip_vec->bv_page) +
112 bip->bip_vec->bv_offset);
113
114 if (bs && mempool_initialized(&bs->bio_integrity_pool)) {
115 bvec_free(&bs->bvec_integrity_pool, bip->bip_vec, bip->bip_slab);
116
117 mempool_free(bip, &bs->bio_integrity_pool);
118 } else {
119 kfree(bip);
120 }
121
122 bio->bi_integrity = NULL;
123 bio->bi_opf &= ~REQ_INTEGRITY;
124}
125
126
127
128
129
130
131
132
133
134
135int bio_integrity_add_page(struct bio *bio, struct page *page,
136 unsigned int len, unsigned int offset)
137{
138 struct bio_integrity_payload *bip = bio_integrity(bio);
139 struct bio_vec *iv;
140
141 if (bip->bip_vcnt >= bip->bip_max_vcnt) {
142 printk(KERN_ERR "%s: bip_vec full\n", __func__);
143 return 0;
144 }
145
146 iv = bip->bip_vec + bip->bip_vcnt;
147
148 if (bip->bip_vcnt &&
149 bvec_gap_to_prev(bio->bi_disk->queue,
150 &bip->bip_vec[bip->bip_vcnt - 1], offset))
151 return 0;
152
153 iv->bv_page = page;
154 iv->bv_len = len;
155 iv->bv_offset = offset;
156 bip->bip_vcnt++;
157
158 return len;
159}
160EXPORT_SYMBOL(bio_integrity_add_page);
161
162
163
164
165
166
167
168static blk_status_t bio_integrity_process(struct bio *bio,
169 struct bvec_iter *proc_iter, integrity_processing_fn *proc_fn)
170{
171 struct blk_integrity *bi = blk_get_integrity(bio->bi_disk);
172 struct blk_integrity_iter iter;
173 struct bvec_iter bviter;
174 struct bio_vec bv;
175 struct bio_integrity_payload *bip = bio_integrity(bio);
176 blk_status_t ret = BLK_STS_OK;
177 void *prot_buf = page_address(bip->bip_vec->bv_page) +
178 bip->bip_vec->bv_offset;
179
180 iter.disk_name = bio->bi_disk->disk_name;
181 iter.interval = 1 << bi->interval_exp;
182 iter.seed = proc_iter->bi_sector;
183 iter.prot_buf = prot_buf;
184
185 __bio_for_each_segment(bv, bio, bviter, *proc_iter) {
186 void *kaddr = kmap_atomic(bv.bv_page);
187
188 iter.data_buf = kaddr + bv.bv_offset;
189 iter.data_size = bv.bv_len;
190
191 ret = proc_fn(&iter);
192 if (ret) {
193 kunmap_atomic(kaddr);
194 return ret;
195 }
196
197 kunmap_atomic(kaddr);
198 }
199 return ret;
200}
201
202
203
204
205
206
207
208
209
210
211
212
213
214bool bio_integrity_prep(struct bio *bio)
215{
216 struct bio_integrity_payload *bip;
217 struct blk_integrity *bi = blk_get_integrity(bio->bi_disk);
218 struct request_queue *q = bio->bi_disk->queue;
219 void *buf;
220 unsigned long start, end;
221 unsigned int len, nr_pages;
222 unsigned int bytes, offset, i;
223 unsigned int intervals;
224 blk_status_t status;
225
226 if (!bi)
227 return true;
228
229 if (bio_op(bio) != REQ_OP_READ && bio_op(bio) != REQ_OP_WRITE)
230 return true;
231
232 if (!bio_sectors(bio))
233 return true;
234
235
236 if (bio_integrity(bio))
237 return true;
238
239 if (bio_data_dir(bio) == READ) {
240 if (!bi->profile->verify_fn ||
241 !(bi->flags & BLK_INTEGRITY_VERIFY))
242 return true;
243 } else {
244 if (!bi->profile->generate_fn ||
245 !(bi->flags & BLK_INTEGRITY_GENERATE))
246 return true;
247 }
248 intervals = bio_integrity_intervals(bi, bio_sectors(bio));
249
250
251 len = intervals * bi->tuple_size;
252 buf = kmalloc(len, GFP_NOIO | q->bounce_gfp);
253 status = BLK_STS_RESOURCE;
254 if (unlikely(buf == NULL)) {
255 printk(KERN_ERR "could not allocate integrity buffer\n");
256 goto err_end_io;
257 }
258
259 end = (((unsigned long) buf) + len + PAGE_SIZE - 1) >> PAGE_SHIFT;
260 start = ((unsigned long) buf) >> PAGE_SHIFT;
261 nr_pages = end - start;
262
263
264 bip = bio_integrity_alloc(bio, GFP_NOIO, nr_pages);
265 if (IS_ERR(bip)) {
266 printk(KERN_ERR "could not allocate data integrity bioset\n");
267 kfree(buf);
268 status = BLK_STS_RESOURCE;
269 goto err_end_io;
270 }
271
272 bip->bip_flags |= BIP_BLOCK_INTEGRITY;
273 bip->bip_iter.bi_size = len;
274 bip_set_seed(bip, bio->bi_iter.bi_sector);
275
276 if (bi->flags & BLK_INTEGRITY_IP_CHECKSUM)
277 bip->bip_flags |= BIP_IP_CHECKSUM;
278
279
280 offset = offset_in_page(buf);
281 for (i = 0 ; i < nr_pages ; i++) {
282 int ret;
283 bytes = PAGE_SIZE - offset;
284
285 if (len <= 0)
286 break;
287
288 if (bytes > len)
289 bytes = len;
290
291 ret = bio_integrity_add_page(bio, virt_to_page(buf),
292 bytes, offset);
293
294 if (ret == 0) {
295 printk(KERN_ERR "could not attach integrity payload\n");
296 kfree(buf);
297 status = BLK_STS_RESOURCE;
298 goto err_end_io;
299 }
300
301 if (ret < bytes)
302 break;
303
304 buf += bytes;
305 len -= bytes;
306 offset = 0;
307 }
308
309
310 if (bio_data_dir(bio) == WRITE) {
311 bio_integrity_process(bio, &bio->bi_iter,
312 bi->profile->generate_fn);
313 } else {
314 bip->bio_iter = bio->bi_iter;
315 }
316 return true;
317
318err_end_io:
319 bio->bi_status = status;
320 bio_endio(bio);
321 return false;
322
323}
324EXPORT_SYMBOL(bio_integrity_prep);
325
326
327
328
329
330
331
332
333
334static void bio_integrity_verify_fn(struct work_struct *work)
335{
336 struct bio_integrity_payload *bip =
337 container_of(work, struct bio_integrity_payload, bip_work);
338 struct bio *bio = bip->bip_bio;
339 struct blk_integrity *bi = blk_get_integrity(bio->bi_disk);
340
341
342
343
344
345
346 bio->bi_status = bio_integrity_process(bio, &bip->bio_iter,
347 bi->profile->verify_fn);
348 bio_integrity_free(bio);
349 bio_endio(bio);
350}
351
352
353
354
355
356
357
358
359
360
361
362
363bool __bio_integrity_endio(struct bio *bio)
364{
365 struct blk_integrity *bi = blk_get_integrity(bio->bi_disk);
366 struct bio_integrity_payload *bip = bio_integrity(bio);
367
368 if (bio_op(bio) == REQ_OP_READ && !bio->bi_status &&
369 (bip->bip_flags & BIP_BLOCK_INTEGRITY) && bi->profile->verify_fn) {
370 INIT_WORK(&bip->bip_work, bio_integrity_verify_fn);
371 queue_work(kintegrityd_wq, &bip->bip_work);
372 return false;
373 }
374
375 bio_integrity_free(bio);
376 return true;
377}
378
379
380
381
382
383
384
385
386
387
388void bio_integrity_advance(struct bio *bio, unsigned int bytes_done)
389{
390 struct bio_integrity_payload *bip = bio_integrity(bio);
391 struct blk_integrity *bi = blk_get_integrity(bio->bi_disk);
392 unsigned bytes = bio_integrity_bytes(bi, bytes_done >> 9);
393
394 bip->bip_iter.bi_sector += bytes_done >> 9;
395 bvec_iter_advance(bip->bip_vec, &bip->bip_iter, bytes);
396}
397EXPORT_SYMBOL(bio_integrity_advance);
398
399
400
401
402
403
404
405void bio_integrity_trim(struct bio *bio)
406{
407 struct bio_integrity_payload *bip = bio_integrity(bio);
408 struct blk_integrity *bi = blk_get_integrity(bio->bi_disk);
409
410 bip->bip_iter.bi_size = bio_integrity_bytes(bi, bio_sectors(bio));
411}
412EXPORT_SYMBOL(bio_integrity_trim);
413
414
415
416
417
418
419
420
421
422int bio_integrity_clone(struct bio *bio, struct bio *bio_src,
423 gfp_t gfp_mask)
424{
425 struct bio_integrity_payload *bip_src = bio_integrity(bio_src);
426 struct bio_integrity_payload *bip;
427
428 BUG_ON(bip_src == NULL);
429
430 bip = bio_integrity_alloc(bio, gfp_mask, bip_src->bip_vcnt);
431 if (IS_ERR(bip))
432 return PTR_ERR(bip);
433
434 memcpy(bip->bip_vec, bip_src->bip_vec,
435 bip_src->bip_vcnt * sizeof(struct bio_vec));
436
437 bip->bip_vcnt = bip_src->bip_vcnt;
438 bip->bip_iter = bip_src->bip_iter;
439
440 return 0;
441}
442EXPORT_SYMBOL(bio_integrity_clone);
443
444int bioset_integrity_create(struct bio_set *bs, int pool_size)
445{
446 if (mempool_initialized(&bs->bio_integrity_pool))
447 return 0;
448
449 if (mempool_init_slab_pool(&bs->bio_integrity_pool,
450 pool_size, bip_slab))
451 return -1;
452
453 if (biovec_init_pool(&bs->bvec_integrity_pool, pool_size)) {
454 mempool_exit(&bs->bio_integrity_pool);
455 return -1;
456 }
457
458 return 0;
459}
460EXPORT_SYMBOL(bioset_integrity_create);
461
462void bioset_integrity_free(struct bio_set *bs)
463{
464 mempool_exit(&bs->bio_integrity_pool);
465 mempool_exit(&bs->bvec_integrity_pool);
466}
467EXPORT_SYMBOL(bioset_integrity_free);
468
469void __init bio_integrity_init(void)
470{
471
472
473
474
475 kintegrityd_wq = alloc_workqueue("kintegrityd", WQ_MEM_RECLAIM |
476 WQ_HIGHPRI | WQ_CPU_INTENSIVE, 1);
477 if (!kintegrityd_wq)
478 panic("Failed to create kintegrityd\n");
479
480 bip_slab = kmem_cache_create("bio_integrity_payload",
481 sizeof(struct bio_integrity_payload) +
482 sizeof(struct bio_vec) * BIP_INLINE_VECS,
483 0, SLAB_HWCACHE_ALIGN|SLAB_PANIC, NULL);
484}
485