dm crypt: limit the number of allocated pages
authorMikulas Patocka <mpatocka@redhat.com>
Mon, 14 Aug 2017 02:45:08 +0000 (22:45 -0400)
committerGreg Kroah-Hartman <gregkh@linuxfoundation.org>
Tue, 24 Apr 2018 07:36:31 +0000 (09:36 +0200)
commit 5059353df86e2573ccd9d43fd9d9396dcec47ca2 upstream.

dm-crypt consumes an excessive amount memory when the user attempts to
zero a dm-crypt device with "blkdiscard -z". The command "blkdiscard -z"
calls the BLKZEROOUT ioctl, it goes to the function __blkdev_issue_zeroout,
__blkdev_issue_zeroout sends a large amount of write bios that contain
the zero page as their payload.

For each incoming page, dm-crypt allocates another page that holds the
encrypted data, so when processing "blkdiscard -z", dm-crypt tries to
allocate the amount of memory that is equal to the size of the device.
This can trigger OOM killer or cause system crash.

Fix this by limiting the amount of memory that dm-crypt allocates to 2%
of total system memory. This limit is system-wide and is divided by the
number of active dm-crypt devices and each device receives an equal
share.

Cc: stable@vger.kernel.org
Signed-off-by: Mikulas Patocka <mpatocka@redhat.com>
Signed-off-by: Mike Snitzer <snitzer@redhat.com>
Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
drivers/md/dm-crypt.c

index 554d60394c0663980d89c3bb84a48007f365470b..f575110454b6a2dd7df499e07fe15141184a7d06 100644 (file)
@@ -148,6 +148,8 @@ struct crypt_config {
        mempool_t *tag_pool;
        unsigned tag_pool_max_sectors;
 
+       struct percpu_counter n_allocated_pages;
+
        struct bio_set *bs;
        struct mutex bio_alloc_lock;
 
@@ -219,6 +221,12 @@ struct crypt_config {
 #define MAX_TAG_SIZE   480
 #define POOL_ENTRY_SIZE        512
 
+static DEFINE_SPINLOCK(dm_crypt_clients_lock);
+static unsigned dm_crypt_clients_n = 0;
+static volatile unsigned long dm_crypt_pages_per_client;
+#define DM_CRYPT_MEMORY_PERCENT                        2
+#define DM_CRYPT_MIN_PAGES_PER_CLIENT          (BIO_MAX_PAGES * 16)
+
 static void clone_init(struct dm_crypt_io *, struct bio *);
 static void kcryptd_queue_crypt(struct dm_crypt_io *io);
 static struct scatterlist *crypt_get_sg_data(struct crypt_config *cc,
@@ -2156,6 +2164,43 @@ static int crypt_wipe_key(struct crypt_config *cc)
        return r;
 }
 
+static void crypt_calculate_pages_per_client(void)
+{
+       unsigned long pages = (totalram_pages - totalhigh_pages) * DM_CRYPT_MEMORY_PERCENT / 100;
+
+       if (!dm_crypt_clients_n)
+               return;
+
+       pages /= dm_crypt_clients_n;
+       if (pages < DM_CRYPT_MIN_PAGES_PER_CLIENT)
+               pages = DM_CRYPT_MIN_PAGES_PER_CLIENT;
+       dm_crypt_pages_per_client = pages;
+}
+
+static void *crypt_page_alloc(gfp_t gfp_mask, void *pool_data)
+{
+       struct crypt_config *cc = pool_data;
+       struct page *page;
+
+       if (unlikely(percpu_counter_compare(&cc->n_allocated_pages, dm_crypt_pages_per_client) >= 0) &&
+           likely(gfp_mask & __GFP_NORETRY))
+               return NULL;
+
+       page = alloc_page(gfp_mask);
+       if (likely(page != NULL))
+               percpu_counter_add(&cc->n_allocated_pages, 1);
+
+       return page;
+}
+
+static void crypt_page_free(void *page, void *pool_data)
+{
+       struct crypt_config *cc = pool_data;
+
+       __free_page(page);
+       percpu_counter_sub(&cc->n_allocated_pages, 1);
+}
+
 static void crypt_dtr(struct dm_target *ti)
 {
        struct crypt_config *cc = ti->private;
@@ -2182,6 +2227,10 @@ static void crypt_dtr(struct dm_target *ti)
        mempool_destroy(cc->req_pool);
        mempool_destroy(cc->tag_pool);
 
+       if (cc->page_pool)
+               WARN_ON(percpu_counter_sum(&cc->n_allocated_pages) != 0);
+       percpu_counter_destroy(&cc->n_allocated_pages);
+
        if (cc->iv_gen_ops && cc->iv_gen_ops->dtr)
                cc->iv_gen_ops->dtr(cc);
 
@@ -2196,6 +2245,12 @@ static void crypt_dtr(struct dm_target *ti)
 
        /* Must zero key material before freeing */
        kzfree(cc);
+
+       spin_lock(&dm_crypt_clients_lock);
+       WARN_ON(!dm_crypt_clients_n);
+       dm_crypt_clients_n--;
+       crypt_calculate_pages_per_client();
+       spin_unlock(&dm_crypt_clients_lock);
 }
 
 static int crypt_ctr_ivmode(struct dm_target *ti, const char *ivmode)
@@ -2643,6 +2698,15 @@ static int crypt_ctr(struct dm_target *ti, unsigned int argc, char **argv)
 
        ti->private = cc;
 
+       spin_lock(&dm_crypt_clients_lock);
+       dm_crypt_clients_n++;
+       crypt_calculate_pages_per_client();
+       spin_unlock(&dm_crypt_clients_lock);
+
+       ret = percpu_counter_init(&cc->n_allocated_pages, 0, GFP_KERNEL);
+       if (ret < 0)
+               goto bad;
+
        /* Optional parameters need to be read before cipher constructor */
        if (argc > 5) {
                ret = crypt_ctr_optional(ti, argc - 5, &argv[5]);
@@ -2697,7 +2761,7 @@ static int crypt_ctr(struct dm_target *ti, unsigned int argc, char **argv)
                ALIGN(sizeof(struct dm_crypt_io) + cc->dmreq_start + additional_req_size,
                      ARCH_KMALLOC_MINALIGN);
 
-       cc->page_pool = mempool_create_page_pool(BIO_MAX_PAGES, 0);
+       cc->page_pool = mempool_create(BIO_MAX_PAGES, crypt_page_alloc, crypt_page_free, cc);
        if (!cc->page_pool) {
                ti->error = "Cannot allocate page mempool";
                goto bad;