wok-current rev 24983

Up xcursor-comix (0.9.2)
author Pascal Bellard <pascal.bellard@slitaz.org>
date Sun May 08 16:45:21 2022 +0000 (2022-05-08)
parents 0f92b8cc8086
children 2b0142c9f248
files linux-cloop/receipt linux-cloop/stuff/cloop.u linux64-cloop/receipt linux64-cloop/stuff/cloop.u wbar/receipt xa/receipt xautomation/receipt xcursor-aero/receipt xcursor-comix/receipt
line diff
     1.1 --- a/linux-cloop/receipt	Sun May 08 13:06:36 2022 +0000
     1.2 +++ b/linux-cloop/receipt	Sun May 08 16:45:21 2022 +0000
     1.3 @@ -2,16 +2,15 @@
     1.4  
     1.5  PACKAGE="linux-cloop"
     1.6  SOURCE="cloop"
     1.7 -_VERSION="2.639-2"
     1.8 -#VERSION="$(sed '/+#define CLOOP_VERSION/!d;s|.* "\(.*\)"|\1|' stuff/cloop.u)"
     1.9 +_VERSION="3.14.1.3"
    1.10  VERSION="4.12"
    1.11  CATEGORY="base-system"
    1.12  MAINTAINER="pascal.bellard@slitaz.org"
    1.13  LICENSE="GPL2"
    1.14  SHORT_DESC="The read-only compressed loop device kernel module."
    1.15  WEB_SITE="http://knoppix.net/wiki/Cloop"
    1.16 -TARBALL="${SOURCE}_${_VERSION}.tar.gz"
    1.17 -WGET_URL="http://debian-knoppix.alioth.debian.org/packages/$SOURCE/$TARBALL"
    1.18 +TARBALL="${SOURCE}_${_VERSION}.tar.xz"
    1.19 +WGET_URL="http://deb.debian.org/debian/pool/main/c/$SOURCE/$TARBALL"
    1.20  
    1.21  DEPENDS="linux"
    1.22  BUILD_DEPENDS="linux-module-headers xz"
    1.23 @@ -23,6 +22,7 @@
    1.24  {
    1.25  	patch -p0 < $stuff/cloop.u
    1.26  	make ARCH=i386 KERNEL_DIR="/usr/src/linux" cloop.ko && xz cloop.ko
    1.27 +	make cloop_suspend
    1.28  }
    1.29  	
    1.30  # Rules to gen a SliTaz package suitable for Tazpkg.
     2.1 --- a/linux-cloop/stuff/cloop.u	Sun May 08 13:06:36 2022 +0000
     2.2 +++ b/linux-cloop/stuff/cloop.u	Sun May 08 16:45:21 2022 +0000
     2.3 @@ -1,6 +1,6 @@
     2.4  --- cloop.h
     2.5  +++ cloop.h
     2.6 -@@ -1,15 +1,50 @@
     2.7 +@@ -1,3 +1,7 @@
     2.8  +#define CLOOP_SIGNATURE "#!/bin/sh"                      /* @ offset 0  */
     2.9  +#define CLOOP_SIGNATURE_SIZE 9
    2.10  +#define CLOOP_SIGNATURE_OFFSET 0x0
    2.11 @@ -8,99 +8,46 @@
    2.12   #ifndef _COMPRESSED_LOOP_H
    2.13   #define _COMPRESSED_LOOP_H
    2.14   
    2.15 --#define CLOOP_HEADROOM 128
    2.16 -+/*************************************************************************\
    2.17 -+* Starting with Format V4.0 (cloop version 4.x), cloop can now have two   *
    2.18 -+* alternative structures:                                                 *
    2.19 -+*                                                                         *
    2.20 -+* 1. Header first: "robust" format, handles missing blocks well           *
    2.21 -+* 2. Footer (header last): "streaming" format, easier to create           *
    2.22 -+*                                                                         *
    2.23 -+* The cloop kernel module autodetects both formats, and can (currently)   *
    2.24 -+* still handle the V2.0 format as well.                                   *
    2.25 -+*                                                                         *
    2.26 -+* 1. Header first:                                                        *
    2.27 -+*   +---------------------------- FIXED SIZE ---------------------------+ *
    2.28 -+*   |Signature (128 bytes)                                              | *
    2.29 -+*   |block_size (32bit number, network order)                           | *
    2.30 -+*   |num_blocks (32bit number, network order)                           | *
    2.31 -+*   +--------------------------- VARIABLE SIZE -------------------------+ *
    2.32 -+*   |num_blocks * FlagsOffset (upper 4 bits flags, lower 64 bits offset)| *
    2.33 -+*   |compressed data blocks of variable size ...                        | *
    2.34 -+*   +-------------------------------------------------------------------+ *
    2.35 -+*                                                                         *
    2.36 -+* 2. Footer (header last):                                                *
    2.37 -+*   +--------------------------- VARIABLE SIZE -------------------------+ *
    2.38 -+*   |compressed data blocks of variable size ...                        | *
    2.39 -+*   |num_blocks * FlagsOffset (upper 4 bits flags, lower 64 bits offset)| *
    2.40 -+*   +---------------------------- FIXED SIZE ---------------------------+ *
    2.41 -+*   |Signature (128 bytes)                                              | *
    2.42 -+*   |block_size (32bit number, network order)                           | *
    2.43 -+*   |num_blocks (32bit number, network order)                           | *
    2.44 -+*   +-------------------------------------------------------------------+ *
    2.45 -+*                                                                         *
    2.46 -+* Offsets are always relative to beginning of file, in all formats.       *
    2.47 -+* The block index contains num_blocks+1 offsets, followed (1) or          *
    2.48 -+* preceded (2) by the compressed blocks.                                  *
    2.49 -+\*************************************************************************/
    2.50 +@@ -38,10 +42,6 @@
    2.51   
    2.52 --/* The cloop header usually looks like this:          */
    2.53 --/* #!/bin/sh                                          */
    2.54 --/* #V2.00 Format                                      */
    2.55 --/* ...padding up to CLOOP_HEADROOM...                 */
    2.56 --/* block_size (32bit number, network order)           */
    2.57 --/* num_blocks (32bit number, network order)           */
    2.58 -+#include <linux/types.h>   /* u_int32_t */
    2.59 -+
    2.60 -+#define CLOOP_HEADROOM 128
    2.61 + #include <linux/types.h>   /* u_int32_t */
    2.62   
    2.63 -+/* Header of fixed length, can be located at beginning or end of file   */
    2.64 - struct cloop_head
    2.65 - {
    2.66 - 	char preamble[CLOOP_HEADROOM];
    2.67 -@@ -17,9 +52,163 @@
    2.68 +-#ifndef __KERNEL__
    2.69 +-#include <stdint.h> /* regular uint64_t */
    2.70 +-#endif
    2.71 +-
    2.72 + #define CLOOP_HEADROOM 128
    2.73 + 
    2.74 + /* Header of fixed length, can be located at beginning or end of file   */
    2.75 +@@ -52,13 +52,6 @@
    2.76   	u_int32_t num_blocks;
    2.77   };
    2.78   
    2.79 -+/************************************************************************\
    2.80 -+*  CLOOP4 flags for each compressed block                                *
    2.81 -+*  Value   Meaning                                                       *
    2.82 -+*    0     GZIP/7ZIP compression (compatible with V2.0 Format)           *
    2.83 -+*    1     no compression (incompressible data)                          *
    2.84 -+*    2     xz compression (currently best space saver)                   *
    2.85 -+*    3     lz4 compression                                               *
    2.86 -+*    4     lzo compression (fastest)                                     *
    2.87 -+\************************************************************************/
    2.88 -+
    2.89 -+typedef uint64_t cloop_block_ptr;
    2.90 -+
    2.91 -+/* Get value of first 4 bits */
    2.92 -+#define CLOOP_BLOCK_FLAGS(x)  ((unsigned int)(((x) & 0xf000000000000000LLU) >> 60))
    2.93 -+/* Get value of last 60 bits */
    2.94 -+#define CLOOP_BLOCK_OFFSET(x)  ((x) & 0x0fffffffffffffffLLU)
    2.95 -+
    2.96 -+#define CLOOP_COMPRESSOR_ZLIB  0x0
    2.97 -+#define CLOOP_COMPRESSOR_NONE  0x1
    2.98 -+#define CLOOP_COMPRESSOR_XZ    0x2
    2.99 -+#define CLOOP_COMPRESSOR_LZ4   0x3
   2.100 -+#define CLOOP_COMPRESSOR_LZO1X 0x4
   2.101 -+
   2.102 -+#define CLOOP_COMPRESSOR_VALID(x) ((x) >= CLOOP_COMPRESSOR_ZLIB && (x) <= CLOOP_COMPRESSOR_LZO1X)
   2.103 -+
   2.104 +-#define CLOOP2_SIGNATURE "V2.0"                       /* @ offset 0x0b  */
   2.105 +-#define CLOOP2_SIGNATURE_SIZE 4
   2.106 +-#define CLOOP2_SIGNATURE_OFFSET 0x0b
   2.107 +-#define CLOOP4_SIGNATURE "V4.0"                       /* @ offset 0x0b  */
   2.108 +-#define CLOOP4_SIGNATURE_SIZE 4
   2.109 +-#define CLOOP4_SIGNATURE_OFFSET 0x0b
   2.110 +-
   2.111 + /************************************************************************\
   2.112 + *  CLOOP4 flags for each compressed block                                *
   2.113 + *  Value   Meaning                                                       *
   2.114 +@@ -84,6 +77,134 @@
   2.115 + 
   2.116 + #define CLOOP_COMPRESSOR_VALID(x) ((x) >= CLOOP_COMPRESSOR_ZLIB && (x) <= CLOOP_COMPRESSOR_LZO1X)
   2.117 + 
   2.118  +#define CLOOP_COMPRESSOR_LINK  0xF
   2.119  +
   2.120  +
   2.121 - /* data_index (num_blocks 64bit pointers, network order)...      */
   2.122 - /* compressed data (gzip block compressed format)...             */
   2.123 - 
   2.124 ++/* data_index (num_blocks 64bit pointers, network order)...      */
   2.125 ++/* compressed data (gzip block compressed format)...             */
   2.126 ++
   2.127  +struct cloop_tail
   2.128  +{
   2.129  +	u_int32_t table_size; 
   2.130 -+	u_int32_t index_size; /* size:4 comp:3 ctrl-c:1 lastlen:24 */
   2.131 ++	u_int32_t index_size; /* size:4 unused:3 ctrl-c:1 lastlen:24 */
   2.132  +#define CLOOP3_INDEX_SIZE(x)    ((unsigned int)((x) & 0xF))
   2.133 -+#define CLOOP3_BLOCKS_FLAGS(x)  ((unsigned int)((x) & 0x70) >> 4)
   2.134 -+#define CLOOP3_TRUNCATED(x)     ((unsigned int)((x) & 0x80) >> 7)
   2.135 -+#define CLOOP3_LASTLEN(x)       (unsigned int)((x) >> 8)
   2.136  +	u_int32_t num_blocks;
   2.137  +};
   2.138  +
   2.139 @@ -114,8 +61,10 @@
   2.140  +};
   2.141  +
   2.142  +static inline char *build_index(struct block_info *offsets, unsigned long n, 
   2.143 -+			unsigned long block_size, unsigned global_flags)
   2.144 ++			unsigned long block_size)
   2.145  +{
   2.146 ++	static char v[11];
   2.147 ++	u_int32_t flags = 0;
   2.148  +	u_int32_t *ofs32 = (u_int32_t *) offsets;
   2.149  +	loff_t    *ofs64 = (loff_t *) offsets;
   2.150  +
   2.151 @@ -140,8 +89,6 @@
   2.152  +		}
   2.153  +		else { /* V2.0/V4.0 */
   2.154  +			loff_t last = CLOOP_BLOCK_OFFSET(__be64_to_cpu(ofs64[n]));
   2.155 -+			u_int32_t flags;
   2.156 -+			static char v4[11];
   2.157  +			unsigned long i = n;
   2.158  +
   2.159  +			for (flags = 0; n-- ;) {
   2.160 @@ -159,12 +106,7 @@
   2.161  +					offsets[i] = offsets[offsets[i].offset];
   2.162  +				}
   2.163  +			}
   2.164 -+			strcpy(v4, (char *) "64BE v4.0a");
   2.165 -+			v4[10] = 'a' + ((flags-1) & 0xF);	// compressors used
   2.166 -+			if (flags > 0x10) {			// with links ?
   2.167 -+				v4[10] += 'A' - 'a';
   2.168 -+			}
   2.169 -+			return v4;
   2.170 ++			strcpy(v, (char *) "64BE v4.0a");
   2.171  +		}
   2.172  +	}
   2.173  +	else if (ofs32[1] == 0 && v3_64 == 0) { /* V1.0 */
   2.174 @@ -180,7 +122,6 @@
   2.175  +	else { /* V3.0 or V0.68 */
   2.176  +		unsigned long i;
   2.177  +		loff_t j;
   2.178 -+		static char v3[11];
   2.179  +		
   2.180  +		for (i = 0; i < n && ntohl(ofs32[i]) < ntohl(ofs32[i+1]); i++);
   2.181  +		if (i == n && ntohl(ofs32[0]) == (4*n) + 0x8C) { /* V0.68 */
   2.182 @@ -195,28 +136,33 @@
   2.183  +		}
   2.184  +		
   2.185  +		v3_64 = (ofs32[1] == 0);
   2.186 -+		for (i = n; i-- != 0; )
   2.187 ++		for (i = n; i-- != 0; ) {
   2.188  +			offsets[i].size = ntohl(ofs32[i << v3_64]); 
   2.189 ++			if (offsets[i].size == 0xFFFFFFFF) {
   2.190 ++				offsets[i].size = 0x10000000 | block_size;
   2.191 ++			}
   2.192 ++			offsets[i].flags = (offsets[i].size >> 28);
   2.193 ++			offsets[i].size &= 0x0FFFFFFF; 
   2.194 ++		}
   2.195  +		for (i = 0, j = sizeof(struct cloop_head); i < n; i++) {
   2.196  +			offsets[i].offset = j;
   2.197 -+			offsets[i].flags = global_flags;
   2.198 -+			if (offsets[i].size == 0xFFFFFFFF) {
   2.199 -+				offsets[i].flags = CLOOP_COMPRESSOR_NONE;
   2.200 -+				offsets[i].size = block_size;
   2.201 -+			}
   2.202 -+			if ((offsets[i].size & 0x80000000) == 0) {
   2.203 ++			if (offsets[i].flags < 8) {
   2.204  +				j += offsets[i].size;
   2.205  +			}
   2.206  +		}
   2.207  +		for (i = 0; i < n; i++) {
   2.208 -+			if (offsets[i].size & 0x80000000) {
   2.209 -+				offsets[i] = offsets[offsets[i].size & 0x7FFFFFFF];
   2.210 ++			flags |= 1 << offsets[i].flags;
   2.211 ++			if (offsets[i].flags >= 8) {
   2.212 ++				offsets[i] = offsets[offsets[i].size];
   2.213  +			}
   2.214  +		}
   2.215 -+		strcpy(v3, (char *) (v3_64) ? "64BE v3.0a" : "32BE v3.0a");
   2.216 -+		v3[10] += global_flags;
   2.217 -+		return v3;
   2.218 ++		strcpy(v, (char *) (v3_64) ? "64BE v3.0a" : "32BE v3.0a");
   2.219  +	}
   2.220 ++	v[10] = 'a' + ((flags-1) & 0xF);	// compressors used
   2.221 ++	if (flags > 0x10) {			// with links ?
   2.222 ++		v[10] += 'A' - 'a';
   2.223 ++	}
   2.224 ++	return v;
   2.225  +}
   2.226  +
   2.227   /* Cloop suspend IOCTL */
   2.228 @@ -224,661 +170,538 @@
   2.229   
   2.230  --- cloop.c
   2.231  +++ cloop.c
   2.232 -@@ -1,26 +1,23 @@
   2.233 --/*
   2.234 -- *  compressed_loop.c: Read-only compressed loop blockdevice
   2.235 -- *  hacked up by Rusty in 1999, extended and maintained by Klaus Knopper
   2.236 -- *
   2.237 -- *  A cloop file looks like this:
   2.238 -- *  [32-bit uncompressed block size: network order]
   2.239 -- *  [32-bit number of blocks (n_blocks): network order]
   2.240 -- *  [64-bit file offsets of start of blocks: network order]
   2.241 -- *    ...
   2.242 -- *    (n_blocks + 1).
   2.243 -- * n_blocks consisting of:
   2.244 -- *   [compressed block]
   2.245 -- *
   2.246 -- * Every version greatly inspired by code seen in loop.c
   2.247 -- * by Theodore Ts'o, 3/29/93.
   2.248 -- *
   2.249 -- * Copyright 1999-2009 by Paul `Rusty' Russell & Klaus Knopper.
   2.250 -- * Redistribution of this file is permitted under the GNU Public License.
   2.251 -- *
   2.252 -- */
   2.253 -+/************************************************************************\
   2.254 -+* cloop.c: Read-only compressed loop blockdevice                         *
   2.255 -+* hacked up by Rusty in 1999, extended and maintained by Klaus Knopper   *
   2.256 -+*                                                                        *
   2.257 -+* For all supported cloop file formats, please check the file "cloop.h"  *
   2.258 -+* New in Version 4:                                                      *
   2.259 -+* - Header can be first or last in cloop file,                           *
   2.260 -+* - Different compression algorithms supported (compression type         *
   2.261 -+*   encoded in first 4 bytes of block offset address)                    *
   2.262 -+*                                                                        *
   2.263 -+* Every version greatly inspired by code seen in loop.c                  *
   2.264 -+* by Theodore Ts'o, 3/29/93.                                             *
   2.265 -+*                                                                        *
   2.266 -+* Copyright 1999-2009 by Paul `Rusty' Russell & Klaus Knopper.           *
   2.267 -+* Redistribution of this file is permitted under the GNU Public License  *
   2.268 -+* V2.                                                                    *
   2.269 -+\************************************************************************/
   2.270 +@@ -17,7 +17,7 @@
   2.271 + \************************************************************************/
   2.272   
   2.273   #define CLOOP_NAME "cloop"
   2.274 --#define CLOOP_VERSION "2.639"
   2.275 +-#define CLOOP_VERSION "5.3"
   2.276  +#define CLOOP_VERSION "4.12"
   2.277   #define CLOOP_MAX 8
   2.278   
   2.279   #ifndef KBUILD_MODNAME
   2.280 -@@ -47,8 +44,27 @@
   2.281 - #include <asm/div64.h> /* do_div() for 64bit division */
   2.282 - #include <asm/uaccess.h>
   2.283 - #include <asm/byteorder.h>
   2.284 --/* Use zlib_inflate from lib/zlib_inflate */
   2.285 -+/* Check for ZLIB, LZO1X, LZ4 decompression algorithms in kernel. */
   2.286 -+#if (defined(CONFIG_ZLIB_INFLATE) || defined(CONFIG_ZLIB_INFLATE_MODULE))
   2.287 - #include <linux/zutil.h>
   2.288 -+#endif
   2.289 -+#if (defined(CONFIG_LZO_DECOMPRESS) || defined(CONFIG_LZO_DECOMPRESS_MODULE))
   2.290 -+#include <linux/lzo.h>
   2.291 -+#endif
   2.292 -+#if (defined(CONFIG_DECOMPRESS_LZ4) || defined(CONFIG_DECOMPRESS_LZ4_MODULE))
   2.293 -+#include <linux/lz4.h>
   2.294 -+#endif
   2.295 -+#if (defined(CONFIG_DECOMPRESS_LZMA) || defined(CONFIG_DECOMPRESS_LZMA_MODULE))
   2.296 -+#include <linux/decompress/unlzma.h>
   2.297 -+#endif
   2.298 -+#if (defined(CONFIG_DECOMPRESS_XZ) || defined(CONFIG_DECOMPRESS_XZ_MODULE))
   2.299 -+#include <linux/xz.h>
   2.300 -+#endif
   2.301 -+
   2.302 -+#if (!(defined(CONFIG_ZLIB_INFLATE) || defined(CONFIG_ZLIB_INFLATE_MODULE) || defined(CONFIG_LZO_DECOMPRESS) || defined(CONFIG_LZO_DECOMPRESS_MODULE) || defined(CONFIG_DECOMPRESS_LZ4) || defined(CONFIG_DECOMPRESS_LZ4_MODULE) || defined(CONFIG_DECOMPRESS_LZMA) || defined(CONFIG_DECOMPRESS_LZMA_MODULE) || defined(CONFIG_DECOMPRESS_XZ) || defined(CONFIG_DECOMPRESS_XZ_MODULE)))
   2.303 -+#error "No decompression library selected in kernel config!"
   2.304 -+#endif
   2.305 -+
   2.306 +@@ -68,7 +68,6 @@
   2.307   #include <linux/loop.h>
   2.308   #include <linux/kthread.h>
   2.309   #include <linux/compat.h>
   2.310 -@@ -92,47 +108,64 @@
   2.311 - #define DEBUGP(format, x...)
   2.312 - #endif
   2.313 +-#include <linux/blk-mq.h> /* new multiqueue infrastructure */
   2.314 + #include "cloop.h"
   2.315   
   2.316 -+/* Default size of buffer to keep some decompressed blocks in memory to speed up access */
   2.317 -+#define BLOCK_BUFFER_MEM (16*65536)
   2.318 -+
   2.319 - /* One file can be opened at module insertion time */
   2.320 - /* insmod cloop file=/path/to/file */
   2.321 - static char *file=NULL;
   2.322 - static unsigned int preload=0;
   2.323 - static unsigned int cloop_max=CLOOP_MAX;
   2.324 -+static unsigned int buffers=BLOCK_BUFFER_MEM;
   2.325 - module_param(file, charp, 0);
   2.326 - module_param(preload, uint, 0);
   2.327 - module_param(cloop_max, uint, 0);
   2.328 - MODULE_PARM_DESC(file, "Initial cloop image file (full path) for /dev/cloop");
   2.329 - MODULE_PARM_DESC(preload, "Preload n blocks of cloop data into memory");
   2.330 - MODULE_PARM_DESC(cloop_max, "Maximum number of cloop devices (default 8)");
   2.331 -+MODULE_PARM_DESC(buffers, "Size of buffer to keep uncompressed blocks in memory in MiB (default 1)");
   2.332 + /* New License scheme */
   2.333 +@@ -93,10 +92,7 @@
   2.334 + /* Use experimental major for now */
   2.335 + #define MAJOR_NR 240
   2.336   
   2.337 - static struct file *initial_file=NULL;
   2.338 - static int cloop_major=MAJOR_NR;
   2.339 +-#ifndef DEVICE_NAME
   2.340 +-#define DEVICE_NAME CLOOP_NAME
   2.341 +-#endif
   2.342 +-
   2.343 ++/* #define DEVICE_NAME CLOOP_NAME */
   2.344 + /* #define DEVICE_NR(device) (MINOR(device)) */
   2.345 + /* #define DEVICE_ON(device) */
   2.346 + /* #define DEVICE_OFF(device) */
   2.347 +@@ -143,7 +139,7 @@
   2.348 +  u_int32_t allflags;
   2.349   
   2.350 --/* Number of buffered decompressed blocks */
   2.351 --#define BUFFERED_BLOCKS 8
   2.352 - struct cloop_device
   2.353 - {
   2.354 -- /* Copied straight from the file */
   2.355 -+ /* Header filled from the file */
   2.356 -  struct cloop_head head;
   2.357 -+ int header_first;
   2.358 -+ int file_format;
   2.359 - 
   2.360 -- /* An array of offsets of compressed blocks within the file */
   2.361 -- loff_t *offsets;
   2.362 -+ /* An or'd sum of all flags of each compressed block (v3) */
   2.363 -+ u_int32_t allflags;
   2.364 -+
   2.365 -+ /* An array of cloop_ptr flags/offset for compressed blocks within the file */
   2.366 +  /* An array of cloop_ptr flags/offset for compressed blocks within the file */
   2.367 +- cloop_block_ptr *block_ptrs;
   2.368  + struct block_info *block_ptrs;
   2.369   
   2.370    /* We buffer some uncompressed blocks for performance */
   2.371 -- int buffered_blocknum[BUFFERED_BLOCKS];
   2.372 -- int current_bufnum;
   2.373 -- void *buffer[BUFFERED_BLOCKS];
   2.374 -- void *compressed_buffer;
   2.375 -- size_t preload_array_size; /* Size of pointer array in blocks */
   2.376 -- size_t preload_size;       /* Number of successfully allocated blocks */
   2.377 -- char **preload_cache;      /* Pointers to preloaded blocks */
   2.378 -+ size_t num_buffered_blocks;	/* how many uncompressed blocks buffered for performance */
   2.379 -+ int *buffered_blocknum;        /* list of numbers of uncompressed blocks in buffer */
   2.380 -+ int current_bufnum;            /* which block is current */
   2.381 -+ unsigned char **buffer;        /* cache space for num_buffered_blocks uncompressed blocks */
   2.382 -+ void *compressed_buffer;       /* space for the largest compressed block */
   2.383 -+ size_t preload_array_size;     /* Size of pointer array in blocks */
   2.384 -+ size_t preload_size;           /* Number of successfully allocated blocks */
   2.385 -+ char **preload_cache;          /* Pointers to preloaded blocks */
   2.386 - 
   2.387 -+#if (defined(CONFIG_ZLIB_INFLATE) || defined(CONFIG_ZLIB_INFLATE_MODULE))
   2.388 -  z_stream zstream;
   2.389 -+#endif
   2.390 -+#if (defined(CONFIG_DECOMPRESS_XZ) || defined(CONFIG_DECOMPRESS_XZ_MODULE))
   2.391 -+ struct xz_dec *xzdecoderstate;
   2.392 -+ struct xz_buf xz_buffer;
   2.393 -+#endif
   2.394 - 
   2.395 -  struct file   *backing_file;  /* associated file */
   2.396 -  struct inode  *backing_inode; /* for bmap */
   2.397 - 
   2.398 -+ unsigned char *underlying_filename;
   2.399 -  unsigned long largest_block;
   2.400 -  unsigned int underlying_blksize;
   2.401 -+ loff_t underlying_total_size;
   2.402 -  int clo_number;
   2.403 -  int refcnt;
   2.404 -  struct block_device *bdev;
   2.405 -@@ -147,7 +180,6 @@
   2.406 +  size_t num_buffered_blocks;	/* how many uncompressed blocks buffered for performance */
   2.407 +@@ -178,14 +174,16 @@
   2.408 +  spinlock_t queue_lock;
   2.409 +  /* mutex for ioctl() */
   2.410 +  struct mutex clo_ctl_mutex;
   2.411 +- /* mutex for request */
   2.412 +- struct mutex clo_rq_mutex;
   2.413 ++ struct list_head clo_list;
   2.414 ++ struct task_struct *clo_thread;
   2.415 ++ wait_queue_head_t clo_event;
   2.416    struct request_queue *clo_queue;
   2.417    struct gendisk *clo_disk;
   2.418 +- struct blk_mq_tag_set tag_set;
   2.419    int suspended;
   2.420 -- char clo_file_name[LO_NAME_SIZE];
   2.421   };
   2.422   
   2.423 - /* Changed in 2.639: cloop_dev is now a an array of cloop_dev pointers,
   2.424 -@@ -156,52 +188,113 @@
   2.425 ++/* Changed in 2.639: cloop_dev is now a an array of cloop_dev pointers,
   2.426 ++   so we can specify how many devices we need via parameters. */
   2.427 + static struct cloop_device **cloop_dev;
   2.428   static const char *cloop_name=CLOOP_NAME;
   2.429   static int cloop_count = 0;
   2.430 - 
   2.431 --#if (!(defined(CONFIG_ZLIB_INFLATE) || defined(CONFIG_ZLIB_INFLATE_MODULE))) /* Must be compiled into kernel. */
   2.432 --#error  "Invalid Kernel configuration. CONFIG_ZLIB_INFLATE support is needed for cloop."
   2.433 --#endif
   2.434 --
   2.435 --/* Use __get_free_pages instead of vmalloc, allows up to 32 pages,
   2.436 -- * 2MB in one piece */
   2.437 - static void *cloop_malloc(size_t size)
   2.438 - {
   2.439 -- int order = get_order(size);
   2.440 -- if(order <= KMALLOC_MAX_ORDER)
   2.441 --   return (void *)kmalloc(size, GFP_KERNEL);
   2.442 -- else if(order < MAX_ORDER)
   2.443 --   return (void *)__get_free_pages(GFP_KERNEL, order);
   2.444 -+ /* kmalloc will fail after the system is running for a while, */
   2.445 -+ /* when large orders can't return contiguous memory. */
   2.446 -+ /* Let's just use vmalloc for now. :-/ */
   2.447 -+ /* int order = get_order(size); */
   2.448 -+ /* if(order <= KMALLOC_MAX_ORDER) */
   2.449 -+ /*  return (void *)kmalloc(size, GFP_KERNEL); */
   2.450 -+ /* else if(order < MAX_ORDER) */
   2.451 -+ /*  return (void *)__get_free_pages(GFP_KERNEL, order); */
   2.452 -  return (void *)vmalloc(size);
   2.453 +@@ -214,24 +212,21 @@
   2.454 +  vfree(mem);
   2.455   }
   2.456   
   2.457 - static void cloop_free(void *mem, size_t size)
   2.458 - {
   2.459 -- int order = get_order(size);
   2.460 -- if(order <= KMALLOC_MAX_ORDER)
   2.461 --   kfree(mem);
   2.462 -- else if(order < MAX_ORDER)
   2.463 --   free_pages((unsigned long)mem, order);
   2.464 -- else vfree(mem);
   2.465 -+ /* int order = get_order(size); */
   2.466 -+ /* if(order <= KMALLOC_MAX_ORDER) */
   2.467 -+ /*  kfree(mem); */
   2.468 -+ /* else if(order < MAX_ORDER) */
   2.469 -+ /*  free_pages((unsigned long)mem, order); */
   2.470 -+ /* else */
   2.471 -+ vfree(mem);
   2.472 - }
   2.473 - 
   2.474 --static int uncompress(struct cloop_device *clo,
   2.475 --                      unsigned char *dest, unsigned long *destLen,
   2.476 --                      unsigned char *source, unsigned long sourceLen)
   2.477 +-/* static int uncompress(struct cloop_device *clo, unsigned char *dest, unsigned long *destLen, unsigned char *source, unsigned long sourceLen) */
   2.478 +-static int uncompress(struct cloop_device *clo, u_int32_t block_num, u_int32_t compressed_length, unsigned long *uncompressed_length)
   2.479  +static int uncompress(struct cloop_device *clo, unsigned char *dest, unsigned long *destLen, unsigned char *source, unsigned long sourceLen, int flags) 
   2.480   {
   2.481 -- /* Most of this code can be found in fs/cramfs/uncompress.c */
   2.482 -- int err;
   2.483 -- clo->zstream.next_in = source;
   2.484 -- clo->zstream.avail_in = sourceLen;
   2.485 -- clo->zstream.next_out = dest;
   2.486 -- clo->zstream.avail_out = *destLen;
   2.487 -- err = zlib_inflateReset(&clo->zstream);
   2.488 -- if (err != Z_OK)
   2.489 --  {
   2.490 --   printk(KERN_ERR "%s: zlib_inflateReset error %d\n", cloop_name, err);
   2.491 --   zlib_inflateEnd(&clo->zstream); zlib_inflateInit(&clo->zstream);
   2.492 --  }
   2.493 -- err = zlib_inflate(&clo->zstream, Z_FINISH);
   2.494 -- *destLen = clo->zstream.total_out;
   2.495 -- if (err != Z_STREAM_END) return err;
   2.496 -- return Z_OK;
   2.497 -+ int err = -1;
   2.498 -+ switch(flags)
   2.499 -+ {
   2.500 -+  case CLOOP_COMPRESSOR_NONE:
   2.501 +  int err = -1;
   2.502 +- int flags = CLOOP_BLOCK_FLAGS(clo->block_ptrs[block_num]);
   2.503 +  switch(flags)
   2.504 +  {
   2.505 +   case CLOOP_COMPRESSOR_NONE:
   2.506 +-   /* block is umcompressed, swap pointers only! */
   2.507 +-   { char *tmp = clo->compressed_buffer; clo->compressed_buffer = clo->buffer[clo->current_bufnum]; clo->buffer[clo->current_bufnum] = tmp; }
   2.508 +-   DEBUGP("cloop: block %d is uncompressed (flags=%d), just swapping %u bytes\n", block_num, flags, compressed_length);
   2.509  +   memcpy(dest, source, *destLen = sourceLen);
   2.510  +   err = Z_OK;
   2.511 -+   break;
   2.512 -+#if (defined(CONFIG_ZLIB_INFLATE) || defined(CONFIG_ZLIB_INFLATE_MODULE))
   2.513 -+  case CLOOP_COMPRESSOR_ZLIB:
   2.514 +    break;
   2.515 + #if (defined(CONFIG_ZLIB_INFLATE) || defined(CONFIG_ZLIB_INFLATE_MODULE))
   2.516 +   case CLOOP_COMPRESSOR_ZLIB:
   2.517 +-   clo->zstream.next_in = clo->compressed_buffer;
   2.518 +-   clo->zstream.avail_in = compressed_length;
   2.519 +-   clo->zstream.next_out = clo->buffer[clo->current_bufnum];
   2.520 +-   clo->zstream.avail_out = clo->head.block_size;
   2.521  +   clo->zstream.next_in = source;
   2.522  +   clo->zstream.avail_in = sourceLen;
   2.523  +   clo->zstream.next_out = dest;
   2.524  +   clo->zstream.avail_out = *destLen;
   2.525 -+   err = zlib_inflateReset(&clo->zstream);
   2.526 -+   if (err != Z_OK)
   2.527 -+   {
   2.528 -+    printk(KERN_ERR "%s: zlib_inflateReset error %d\n", cloop_name, err);
   2.529 -+    zlib_inflateEnd(&clo->zstream); zlib_inflateInit(&clo->zstream);
   2.530 -+   }
   2.531 -+   err = zlib_inflate(&clo->zstream, Z_FINISH);
   2.532 +    err = zlib_inflateReset(&clo->zstream);
   2.533 +    if (err != Z_OK)
   2.534 +    {
   2.535 +@@ -239,50 +234,50 @@
   2.536 +     zlib_inflateEnd(&clo->zstream); zlib_inflateInit(&clo->zstream);
   2.537 +    }
   2.538 +    err = zlib_inflate(&clo->zstream, Z_FINISH);
   2.539 +-   *uncompressed_length = clo->zstream.total_out;
   2.540  +   *destLen = clo->zstream.total_out;
   2.541 -+   if (err == Z_STREAM_END) err = 0;
   2.542 +    if (err == Z_STREAM_END) err = 0;
   2.543 +-   DEBUGP("cloop: zlib decompression done, ret =%d, size =%lu\n", err, *uncompressed_length);
   2.544  +   DEBUGP("cloop: zlib decompression done, ret =%d, size =%lu\n", err, *destLen);
   2.545 -+   break;
   2.546 -+#endif
   2.547 -+#if (defined(CONFIG_LZO_DECOMPRESS) || defined(CONFIG_LZO_DECOMPRESS_MODULE))
   2.548 -+  case CLOOP_COMPRESSOR_LZO1X:
   2.549 -+   {
   2.550 -+    size_t tmp = (size_t) clo->head.block_size;
   2.551 +    break;
   2.552 + #endif
   2.553 + #if (defined(CONFIG_LZO_DECOMPRESS) || defined(CONFIG_LZO_DECOMPRESS_MODULE))
   2.554 +   case CLOOP_COMPRESSOR_LZO1X:
   2.555 +    {
   2.556 +     size_t tmp = (size_t) clo->head.block_size;
   2.557 +-    err = lzo1x_decompress_safe(clo->compressed_buffer, compressed_length,
   2.558 +-             clo->buffer[clo->current_bufnum], &tmp);
   2.559 +-    if (err == LZO_E_OK) *uncompressed_length = (u_int32_t) tmp;
   2.560  +    err = lzo1x_decompress_safe(source, sourceLen,
   2.561  +             dest, &tmp);
   2.562  +    if (err == LZO_E_OK) *destLen = (u_int32_t) tmp;
   2.563 -+   }
   2.564 -+   break;
   2.565 -+#endif
   2.566 -+#if (defined(CONFIG_DECOMPRESS_LZ4) || defined(CONFIG_DECOMPRESS_LZ4_MODULE))
   2.567 -+  case CLOOP_COMPRESSOR_LZ4:
   2.568 -+   {
   2.569 +    }
   2.570 +    break;
   2.571 + #endif
   2.572 + #if (defined(CONFIG_DECOMPRESS_LZ4) || defined(CONFIG_DECOMPRESS_LZ4_MODULE))
   2.573 +   case CLOOP_COMPRESSOR_LZ4:
   2.574 +    {
   2.575 +-    size_t outputSize = clo->head.block_size;
   2.576  +    size_t outputSize = *destLen;
   2.577 -+    /* We should adjust outputSize here, in case the last block is smaller than block_size */
   2.578 -+#if LINUX_VERSION_CODE < KERNEL_VERSION(4, 11, 0) /* field removed */
   2.579 +     /* We should adjust outputSize here, in case the last block is smaller than block_size */
   2.580 + #if LINUX_VERSION_CODE < KERNEL_VERSION(4, 11, 0) /* field removed */
   2.581 +-    err = lz4_decompress(clo->compressed_buffer, (size_t *) &compressed_length,
   2.582 +-                         clo->buffer[clo->current_bufnum], outputSize);
   2.583  +    err = lz4_decompress(source, (size_t *) &sourceLen,
   2.584  +                         dest, outputSize);
   2.585 -+#else
   2.586 + #else
   2.587 +-    err = LZ4_decompress_safe(clo->compressed_buffer,
   2.588 +-                              clo->buffer[clo->current_bufnum],
   2.589 +-                              compressed_length, outputSize);
   2.590  +    err = LZ4_decompress_safe(source,
   2.591  +                              dest,
   2.592  +                              sourceLen, outputSize);
   2.593 -+#endif
   2.594 -+    if (err >= 0) 
   2.595 -+    {
   2.596 -+     err = 0;
   2.597 + #endif
   2.598 +     if (err >= 0) 
   2.599 +     {
   2.600 +      err = 0;
   2.601 +-     *uncompressed_length = outputSize;
   2.602  +     *destLen = outputSize;
   2.603 -+    }
   2.604 -+   }
   2.605 -+  break;
   2.606 -+#endif
   2.607 -+#if (defined(CONFIG_DECOMPRESS_XZ) || defined(CONFIG_DECOMPRESS_XZ_MODULE))
   2.608 -+ case CLOOP_COMPRESSOR_XZ:
   2.609 +     }
   2.610 +    }
   2.611 +   break;
   2.612 + #endif
   2.613 + #if (defined(CONFIG_DECOMPRESS_XZ) || defined(CONFIG_DECOMPRESS_XZ_MODULE))
   2.614 +  case CLOOP_COMPRESSOR_XZ:
   2.615 +-  clo->xz_buffer.in = clo->compressed_buffer;
   2.616  +  clo->xz_buffer.in = source;
   2.617 -+  clo->xz_buffer.in_pos = 0;
   2.618 +   clo->xz_buffer.in_pos = 0;
   2.619 +-  clo->xz_buffer.in_size = compressed_length;
   2.620 +-  clo->xz_buffer.out = clo->buffer[clo->current_bufnum];
   2.621  +  clo->xz_buffer.in_size = sourceLen;
   2.622  +  clo->xz_buffer.out = dest;
   2.623 -+  clo->xz_buffer.out_pos = 0;
   2.624 +   clo->xz_buffer.out_pos = 0;
   2.625 +-  clo->xz_buffer.out_size = clo->head.block_size;
   2.626  +  clo->xz_buffer.out_size = *destLen;
   2.627 -+  xz_dec_reset(clo->xzdecoderstate);
   2.628 -+  err = xz_dec_run(clo->xzdecoderstate, &clo->xz_buffer);
   2.629 -+  if (err == XZ_STREAM_END || err == XZ_OK)
   2.630 -+  {
   2.631 -+   err = 0;
   2.632 -+  }
   2.633 -+  else
   2.634 -+  {
   2.635 -+   printk(KERN_ERR "%s: xz_dec_run error %d\n", cloop_name, err);
   2.636 -+   err = 1;
   2.637 -+  }
   2.638 -+  break;
   2.639 -+#endif
   2.640 -+ default:
   2.641 -+   printk(KERN_ERR "%s: compression method is not supported!\n", cloop_name);
   2.642 -+ }
   2.643 -+ return err;
   2.644 - }
   2.645 - 
   2.646 - static ssize_t cloop_read_from_file(struct cloop_device *clo, struct file *f, char *buf,
   2.647 -@@ -220,7 +313,7 @@
   2.648 +   xz_dec_reset(clo->xzdecoderstate);
   2.649 +   err = xz_dec_run(clo->xzdecoderstate, &clo->xz_buffer);
   2.650 +   if (err == XZ_STREAM_END || err == XZ_OK)
   2.651 +@@ -309,16 +304,12 @@
   2.652 +  while (buf_done < buf_len)
   2.653 +   {
   2.654 +    size_t size = buf_len - buf_done, size_read;
   2.655 +-   mm_segment_t old_fs;
   2.656 +    /* kernel_read() only supports 32 bit offsets, so we use vfs_read() instead. */
   2.657 +    /* int size_read = kernel_read(f, pos, buf + buf_done, size); */
   2.658 +-
   2.659 +-   // mutex_lock(&clo->clo_rq_mutex);
   2.660 +-   old_fs = get_fs();
   2.661 +-   set_fs(KERNEL_DS);
   2.662 ++   mm_segment_t old_fs = get_fs();
   2.663 ++   set_fs(get_ds());
   2.664 +    size_read = vfs_read(f, (void __user *)(buf + buf_done), size, &pos);
   2.665 +    set_fs(old_fs);
   2.666 +-   // mutex_unlock(&clo->clo_rq_mutex);
   2.667   
   2.668      if(size_read <= 0)
   2.669       {
   2.670 --     printk(KERN_ERR "%s: Read error %d at pos %Lu in file %s, "
   2.671 -+     printk(KERN_ERR "%s: Read error %d at pos %llu in file %s, "
   2.672 -                      "%d bytes lost.\n", cloop_name, (int)size_read, pos,
   2.673 - 		     file, (int)size);
   2.674 -      memset(buf + buf_len - size, 0, size);
   2.675 -@@ -232,72 +325,84 @@
   2.676 - }
   2.677 +@@ -358,8 +349,8 @@
   2.678 +    return i;
   2.679 +   }
   2.680   
   2.681 - /* This looks more complicated than it is */
   2.682 --/* Returns number of block buffer to use for this request */
   2.683 -+/* Returns number of cache block buffer to use for this request */
   2.684 - static int cloop_load_buffer(struct cloop_device *clo, int blocknum)
   2.685 - {
   2.686 -- unsigned int buf_done = 0;
   2.687 -- unsigned long buflen;
   2.688 -- unsigned int buf_length;
   2.689 -+ loff_t compressed_block_offset;
   2.690 -+ long compressed_block_len;
   2.691 -+ long uncompressed_block_len=0;
   2.692 -  int ret;
   2.693 -  int i;
   2.694 -- if(blocknum > ntohl(clo->head.num_blocks) || blocknum < 0)
   2.695 --  {
   2.696 --   printk(KERN_WARNING "%s: Invalid block number %d requested.\n",
   2.697 --                       cloop_name, blocknum);
   2.698 --   return -1;
   2.699 --  }
   2.700 -+ if(blocknum > clo->head.num_blocks || blocknum < 0)
   2.701 -+ {
   2.702 -+  printk(KERN_WARNING "%s: Invalid block number %d requested.\n",
   2.703 -+         cloop_name, blocknum);
   2.704 -+  return -1;
   2.705 -+ }
   2.706 - 
   2.707 -  /* Quick return if the block we seek is already in one of the buffers. */
   2.708 -  /* Return number of buffer */
   2.709 -- for(i=0; i<BUFFERED_BLOCKS; i++)
   2.710 -+ for(i=0; i<clo->num_buffered_blocks; i++)
   2.711 -   if (blocknum == clo->buffered_blocknum[i])
   2.712 --   {
   2.713 --    DEBUGP(KERN_INFO "cloop_load_buffer: Found buffered block %d\n", i);
   2.714 --    return i;
   2.715 --   }
   2.716 --
   2.717 -- buf_length = be64_to_cpu(clo->offsets[blocknum+1]) - be64_to_cpu(clo->offsets[blocknum]);
   2.718 --
   2.719 --/* Load one compressed block from the file. */
   2.720 -- cloop_read_from_file(clo, clo->backing_file, (char *)clo->compressed_buffer,
   2.721 --                    be64_to_cpu(clo->offsets[blocknum]), buf_length);
   2.722 -+  {
   2.723 -+   DEBUGP(KERN_INFO "cloop_load_buffer: Found buffered block %d\n", i);
   2.724 -+   return i;
   2.725 -+  }
   2.726 - 
   2.727 -- buflen = ntohl(clo->head.block_size);
   2.728 +- compressed_block_offset = CLOOP_BLOCK_OFFSET(clo->block_ptrs[blocknum]);
   2.729 +- compressed_block_len = (long) (CLOOP_BLOCK_OFFSET(clo->block_ptrs[blocknum+1]) - compressed_block_offset) ;
   2.730  + compressed_block_offset = clo->block_ptrs[blocknum].offset;
   2.731  + compressed_block_len = (long) (clo->block_ptrs[blocknum].size) ;
   2.732   
   2.733 -- /* Go to next position in the block ring buffer */
   2.734 -- clo->current_bufnum++;
   2.735 -- if(clo->current_bufnum >= BUFFERED_BLOCKS) clo->current_bufnum = 0;
   2.736 -+ /* Load one compressed block from the file. */
   2.737 -+ if(compressed_block_offset > 0 && compressed_block_len >= 0) /* sanity check */
   2.738 -+ {
   2.739 -+  size_t n = cloop_read_from_file(clo, clo->backing_file, (char *)clo->compressed_buffer,
   2.740 -+                    compressed_block_offset, compressed_block_len);
   2.741 -+  if (n!= compressed_block_len)
   2.742 -+   {
   2.743 -+    printk(KERN_ERR "%s: error while reading %lu bytes @ %llu from file %s\n",
   2.744 +  /* Load one compressed block from the file. */
   2.745 +  if(compressed_block_offset > 0 && compressed_block_len >= 0) /* sanity check */
   2.746 +@@ -369,12 +360,12 @@
   2.747 +   if (n!= compressed_block_len)
   2.748 +    {
   2.749 +     printk(KERN_ERR "%s: error while reading %lu bytes @ %llu from file %s\n",
   2.750 +-     cloop_name, compressed_block_len, clo->block_ptrs[blocknum], clo->underlying_filename);
   2.751  +     cloop_name, compressed_block_len, clo->block_ptrs[blocknum].offset, clo->underlying_filename);
   2.752 -+    /* return -1; */
   2.753 -+   }
   2.754 -+ } else {
   2.755 -+  printk(KERN_ERR "%s: invalid data block len %ld bytes @ %lld from file %s\n",
   2.756 +     /* return -1; */
   2.757 +    }
   2.758 +  } else {
   2.759 +   printk(KERN_ERR "%s: invalid data block len %ld bytes @ %lld from file %s\n",
   2.760 +-  cloop_name, compressed_block_len, clo->block_ptrs[blocknum], clo->underlying_filename);
   2.761  +  cloop_name, compressed_block_len, clo->block_ptrs[blocknum].offset, clo->underlying_filename);
   2.762 -+  return -1;
   2.763 -+ }
   2.764 -+  
   2.765 -+ /* Go to next position in the cache block buffer (which is used as a cyclic buffer) */
   2.766 -+ if(++clo->current_bufnum >= clo->num_buffered_blocks) clo->current_bufnum = 0;
   2.767 +   return -1;
   2.768 +  }
   2.769 +   
   2.770 +@@ -382,14 +373,16 @@
   2.771 +  if(++clo->current_bufnum >= clo->num_buffered_blocks) clo->current_bufnum = 0;
   2.772   
   2.773    /* Do the uncompression */
   2.774 -- ret = uncompress(clo, clo->buffer[clo->current_bufnum], &buflen, clo->compressed_buffer,
   2.775 --                  buf_length);
   2.776 +- ret = uncompress(clo, blocknum, compressed_block_len, &uncompressed_block_len);
   2.777  + uncompressed_block_len = clo->head.block_size;
   2.778  + ret = uncompress(clo, clo->buffer[clo->current_bufnum], &uncompressed_block_len,
   2.779  +	 clo->compressed_buffer, compressed_block_len, clo->block_ptrs[blocknum].flags);
   2.780    /* DEBUGP("cloop: buflen after uncompress: %ld\n",buflen); */
   2.781    if (ret != 0)
   2.782 --  {
   2.783 --   printk(KERN_ERR "%s: zlib decompression error %i uncompressing block %u %u/%lu/%u/%u "
   2.784 --          "%Lu-%Lu\n", cloop_name, ret, blocknum,
   2.785 --	  ntohl(clo->head.block_size), buflen, buf_length, buf_done,
   2.786 --	  be64_to_cpu(clo->offsets[blocknum]), be64_to_cpu(clo->offsets[blocknum+1]));
   2.787 --   clo->buffered_blocknum[clo->current_bufnum] = -1;
   2.788 --   return -1;
   2.789 --  }
   2.790 -+ {
   2.791 -+  printk(KERN_ERR "%s: decompression error %i uncompressing block %u %lu bytes @ %llu, flags %u\n",
   2.792 -+         cloop_name, ret, blocknum,
   2.793 +  {
   2.794 +   printk(KERN_ERR "%s: decompression error %i uncompressing block %u %lu bytes @ %llu, flags %u\n",
   2.795 +          cloop_name, ret, blocknum,
   2.796 +-         compressed_block_len, CLOOP_BLOCK_OFFSET(clo->block_ptrs[blocknum]),
   2.797 +-         CLOOP_BLOCK_FLAGS(clo->block_ptrs[blocknum]));
   2.798  +         compressed_block_len, clo->block_ptrs[blocknum].offset,
   2.799  +         clo->block_ptrs[blocknum].flags);
   2.800 -+         clo->buffered_blocknum[clo->current_bufnum] = -1;
   2.801 -+  return -1;
   2.802 -+ }
   2.803 -  clo->buffered_blocknum[clo->current_bufnum] = blocknum;
   2.804 +          clo->buffered_blocknum[clo->current_bufnum] = -1;
   2.805 +   return -1;
   2.806 +  }
   2.807 +@@ -397,107 +390,146 @@
   2.808    return clo->current_bufnum;
   2.809   }
   2.810   
   2.811 - /* This function does all the real work. */
   2.812 --/* returns "uptodate" */
   2.813 +-static blk_status_t cloop_handle_request(struct cloop_device *clo, struct request *req)
   2.814 ++/* This function does all the real work. */
   2.815  +/* returns "uptodate"                    */
   2.816 - static int cloop_handle_request(struct cloop_device *clo, struct request *req)
   2.817 ++static int cloop_handle_request(struct cloop_device *clo, struct request *req)
   2.818   {
   2.819    int buffered_blocknum = -1;
   2.820    int preloaded = 0;
   2.821 -  loff_t offset     = (loff_t) blk_rq_pos(req)<<9; /* req->sector<<9 */
   2.822 -- struct bio_vec *bvec;
   2.823 -+ struct bio_vec bvec;
   2.824 +- loff_t offset = (loff_t) blk_rq_pos(req)<<9;
   2.825 ++ loff_t offset     = (loff_t) blk_rq_pos(req)<<9; /* req->sector<<9 */
   2.826 +  struct bio_vec bvec;
   2.827    struct req_iterator iter;
   2.828 +- blk_status_t ret = BLK_STS_OK;
   2.829 +-
   2.830 +- if (unlikely(req_op(req) != REQ_OP_READ ))
   2.831 +- {
   2.832 +-  blk_dump_rq_flags(req, DEVICE_NAME " bad request");
   2.833 +-  return BLK_STS_IOERR;
   2.834 +- }
   2.835 +-
   2.836 +- if (unlikely(!clo->backing_file && !clo->suspended))
   2.837 +- {
   2.838 +-  DEBUGP("cloop_handle_request: not connected to a file\n");
   2.839 +-  return BLK_STS_IOERR;
   2.840 +- }
   2.841 +-
   2.842    rq_for_each_segment(bvec, req, iter)
   2.843 +- {
   2.844 +-  unsigned long len = bvec.bv_len;
   2.845 +-  loff_t to_offset  = bvec.bv_offset;
   2.846 +-
   2.847 +-  while(len > 0)
   2.848     {
   2.849 --   unsigned long len = bvec->bv_len;
   2.850 --   char *to_ptr      = kmap(bvec->bv_page) + bvec->bv_offset;
   2.851 +-   u_int32_t length_in_buffer;
   2.852 +-   loff_t block_offset = offset;
   2.853 +-   u_int32_t offset_in_buffer;
   2.854 +-   char *from_ptr, *to_ptr;
   2.855 +-   /* do_div (div64.h) returns the 64bit division remainder and  */
   2.856 +-   /* puts the result in the first argument, i.e. block_offset   */
   2.857 +-   /* becomes the blocknumber to load, and offset_in_buffer the  */
   2.858 +-   /* position in the buffer */
   2.859 +-   offset_in_buffer = do_div(block_offset, clo->head.block_size);
   2.860 +-   /* Lookup preload cache */
   2.861 +-   if(block_offset < clo->preload_size && clo->preload_cache != NULL && clo->preload_cache[block_offset] != NULL)
   2.862 +-   { /* Copy from cache */
   2.863 +-    preloaded = 1;
   2.864 +-    from_ptr = clo->preload_cache[block_offset];
   2.865 +-   }
   2.866 +-   else
   2.867 +-   {
   2.868 +-    preloaded = 0;
   2.869 +-    buffered_blocknum = cloop_load_buffer(clo,block_offset);
   2.870 +-    if(buffered_blocknum == -1)
   2.871  +   unsigned long len = bvec.bv_len;
   2.872  +   char *to_ptr      = kmap(bvec.bv_page) + bvec.bv_offset;
   2.873 -    while(len > 0)
   2.874 ++   while(len > 0)
   2.875       {
   2.876 -      u_int32_t length_in_buffer;
   2.877 -@@ -308,7 +413,7 @@
   2.878 -      /* puts the result in the first argument, i.e. block_offset   */
   2.879 -      /* becomes the blocknumber to load, and offset_in_buffer the  */
   2.880 -      /* position in the buffer */
   2.881 --     offset_in_buffer = do_div(block_offset, ntohl(clo->head.block_size));
   2.882 +-     ret = BLK_STS_IOERR;
   2.883 +-     break; /* invalid data, leave inner loop */
   2.884 ++     u_int32_t length_in_buffer;
   2.885 ++     loff_t block_offset = offset;
   2.886 ++     u_int32_t offset_in_buffer;
   2.887 ++     char *from_ptr;
   2.888 ++     /* do_div (div64.h) returns the 64bit division remainder and  */
   2.889 ++     /* puts the result in the first argument, i.e. block_offset   */
   2.890 ++     /* becomes the blocknumber to load, and offset_in_buffer the  */
   2.891 ++     /* position in the buffer */
   2.892  +     offset_in_buffer = do_div(block_offset, clo->head.block_size);
   2.893 -      /* Lookup preload cache */
   2.894 -      if(block_offset < clo->preload_size && clo->preload_cache != NULL &&
   2.895 -         clo->preload_cache[block_offset] != NULL)
   2.896 -@@ -325,7 +430,7 @@
   2.897 -        from_ptr = clo->buffer[buffered_blocknum];
   2.898 -       }
   2.899 -      /* Now, at least part of what we want will be in the buffer. */
   2.900 --     length_in_buffer = ntohl(clo->head.block_size) - offset_in_buffer;
   2.901 ++     /* Lookup preload cache */
   2.902 ++     if(block_offset < clo->preload_size && clo->preload_cache != NULL &&
   2.903 ++        clo->preload_cache[block_offset] != NULL)
   2.904 ++      { /* Copy from cache */
   2.905 ++       preloaded = 1;
   2.906 ++       from_ptr = clo->preload_cache[block_offset];
   2.907 ++      }
   2.908 ++     else
   2.909 ++      {
   2.910 ++       preloaded = 0;
   2.911 ++       buffered_blocknum = cloop_load_buffer(clo,block_offset);
   2.912 ++       if(buffered_blocknum == -1) break; /* invalid data, leave inner loop */
   2.913 ++       /* Copy from buffer */
   2.914 ++       from_ptr = clo->buffer[buffered_blocknum];
   2.915 ++      }
   2.916 ++     /* Now, at least part of what we want will be in the buffer. */
   2.917  +     length_in_buffer = clo->head.block_size - offset_in_buffer;
   2.918 -      if(length_in_buffer > len)
   2.919 -       {
   2.920 - /*   DEBUGP("Warning: length_in_buffer=%u > len=%u\n",
   2.921 -@@ -337,18 +442,19 @@
   2.922 -      len         -= length_in_buffer;
   2.923 -      offset      += length_in_buffer;
   2.924 -     } /* while inner loop */
   2.925 --   kunmap(bvec->bv_page);
   2.926 ++     if(length_in_buffer > len)
   2.927 ++      {
   2.928 ++/*   DEBUGP("Warning: length_in_buffer=%u > len=%u\n",
   2.929 ++                      length_in_buffer,len); */
   2.930 ++       length_in_buffer = len;
   2.931 ++      }
   2.932 ++     memcpy(to_ptr, from_ptr + offset_in_buffer, length_in_buffer);
   2.933 ++     to_ptr      += length_in_buffer;
   2.934 ++     len         -= length_in_buffer;
   2.935 ++     offset      += length_in_buffer;
   2.936 ++    } /* while inner loop */
   2.937  +   kunmap(bvec.bv_page);
   2.938  +   cond_resched();
   2.939 -   } /* end rq_for_each_segment*/
   2.940 -  return ((buffered_blocknum!=-1) || preloaded);
   2.941 - }
   2.942 - 
   2.943 - /* Adopted from loop.c, a kernel thread to handle physical reads and
   2.944 -- * decompression. */
   2.945 ++  } /* end rq_for_each_segment*/
   2.946 ++ return ((buffered_blocknum!=-1) || preloaded);
   2.947 ++}
   2.948 ++
   2.949 ++/* Adopted from loop.c, a kernel thread to handle physical reads and
   2.950  +   decompression. */
   2.951 - static int cloop_thread(void *data)
   2.952 - {
   2.953 -  struct cloop_device *clo = data;
   2.954 -  current->flags |= PF_NOFREEZE;
   2.955 -- set_user_nice(current, -15);
   2.956 ++static int cloop_thread(void *data)
   2.957 ++{
   2.958 ++ struct cloop_device *clo = data;
   2.959 ++ current->flags |= PF_NOFREEZE;
   2.960  + set_user_nice(current, 10);
   2.961 -  while (!kthread_should_stop()||!list_empty(&clo->clo_list))
   2.962 -   {
   2.963 -    int err;
   2.964 -@@ -390,10 +496,18 @@
   2.965 -    int rw;
   2.966 -  /* quick sanity checks */
   2.967 -    /* blk_fs_request() was removed in 2.6.36 */
   2.968 --   if (unlikely(req == NULL || (req->cmd_type != REQ_TYPE_FS)))
   2.969 ++ while (!kthread_should_stop()||!list_empty(&clo->clo_list))
   2.970 ++  {
   2.971 ++   int err;
   2.972 ++   err = wait_event_interruptible(clo->clo_event, !list_empty(&clo->clo_list) || 
   2.973 ++                                  kthread_should_stop());
   2.974 ++   if(unlikely(err))
   2.975 ++    {
   2.976 ++     DEBUGP(KERN_ERR "cloop thread activated on error!? Continuing.\n");
   2.977 ++     continue;
   2.978 +     }
   2.979 +-    /* Copy from buffer */
   2.980 +-    from_ptr = clo->buffer[buffered_blocknum];
   2.981 +-   }
   2.982 +-   /* Now, at least part of what we want will be in the buffer. */
   2.983 +-   length_in_buffer = clo->head.block_size - offset_in_buffer;
   2.984 +-   if(length_in_buffer > len)
   2.985 +-   {
   2.986 +-   /* DEBUGP("Warning: length_in_buffer=%u > len=%u\n", length_in_buffer,len); */
   2.987 +-    length_in_buffer = len;
   2.988 +-   }
   2.989 +-   to_ptr      = kmap_atomic(bvec.bv_page);
   2.990 +-   memcpy(to_ptr + to_offset, from_ptr + offset_in_buffer, length_in_buffer);
   2.991 +-   kunmap_atomic(to_ptr);
   2.992 +-   to_offset   += length_in_buffer;
   2.993 +-   len         -= length_in_buffer;
   2.994 +-   offset      += length_in_buffer;
   2.995 +-  } /* while inner loop */
   2.996 +- } /* rq_for_each_segment */
   2.997 +- return ret;
   2.998 +-}
   2.999 +-
  2.1000 +-static blk_status_t cloop_queue_rq(struct blk_mq_hw_ctx *hctx, const struct blk_mq_queue_data *bd)
  2.1001 +-{
  2.1002 +-//  struct request_queue *q  = hctx->queue;
  2.1003 +-//  struct cloop_device *clo = q->queuedata;
  2.1004 +- struct request *req = bd->rq;
  2.1005 +- struct cloop_device *clo = req->rq_disk->private_data;
  2.1006 +- blk_status_t ret         = BLK_STS_OK;
  2.1007 +-
  2.1008 +-#if 1 /* Does it work when loading libraries? */
  2.1009 +- /* Since we have a buffered block list as well as data to read */
  2.1010 +- /* from disk (slow), and are (probably) never called from an   */
  2.1011 +- /* interrupt, we use a simple mutex lock right here to ensure  */
  2.1012 +- /* consistency.                                                */
  2.1013 +-  mutex_lock(&clo->clo_rq_mutex);
  2.1014 +- #else
  2.1015 +-  spin_lock_irq(&clo->queue_lock);
  2.1016 +- #endif
  2.1017 +- blk_mq_start_request(req);
  2.1018 +- do {
  2.1019 +-  ret = cloop_handle_request(clo, req);
  2.1020 +- } while(blk_update_request(req, ret, blk_rq_cur_bytes(req)));
  2.1021 +- blk_mq_end_request(req, ret);
  2.1022 +- #if 1 /* See above */
  2.1023 +-  mutex_unlock(&clo->clo_rq_mutex);
  2.1024 +- #else
  2.1025 +-  spin_unlock_irq(&clo->queue_lock);
  2.1026 +- #endif
  2.1027 +- return ret;
  2.1028 ++   if(!list_empty(&clo->clo_list))
  2.1029 ++    {
  2.1030 ++     struct request *req;
  2.1031 ++     unsigned long flags;
  2.1032 ++     int uptodate;
  2.1033 ++     spin_lock_irq(&clo->queue_lock);
  2.1034 ++     req = list_entry(clo->clo_list.next, struct request, queuelist);
  2.1035 ++     list_del_init(&req->queuelist);
  2.1036 ++     spin_unlock_irq(&clo->queue_lock);
  2.1037 ++     uptodate = cloop_handle_request(clo, req);
  2.1038 ++     spin_lock_irqsave(&clo->queue_lock, flags);
  2.1039 ++     __blk_end_request_all(req, uptodate ? 0 : -EIO);
  2.1040 ++     spin_unlock_irqrestore(&clo->queue_lock, flags);
  2.1041 ++    }
  2.1042 ++  }
  2.1043 ++ DEBUGP(KERN_ERR "cloop_thread exited.\n");
  2.1044 ++ return 0;
  2.1045 ++}
  2.1046 ++
  2.1047 ++/* This is called by the kernel block queue management every now and then,
  2.1048 ++ * with successive read requests qeued and sorted in a (hopefully)
  2.1049 ++ * "most efficient way". spin_lock_irq() is being held by the kernel. */
  2.1050 ++static void cloop_do_request(struct request_queue *q)
  2.1051 ++{
  2.1052 ++ struct request *req;
  2.1053 ++ while((req = blk_fetch_request(q)) != NULL)
  2.1054 ++  {
  2.1055 ++   struct cloop_device *clo;
  2.1056 ++   int rw;
  2.1057 ++ /* quick sanity checks */
  2.1058 ++   /* blk_fs_request() was removed in 2.6.36 */
  2.1059  +   if (unlikely(req == NULL
  2.1060  +#if LINUX_VERSION_CODE < KERNEL_VERSION(4, 11, 0) /* field removed */
  2.1061  +   || (req->cmd_type != REQ_TYPE_FS)
  2.1062  +#endif
  2.1063  +   ))
  2.1064 -     goto error_continue;
  2.1065 -    rw = rq_data_dir(req);
  2.1066 --   if (unlikely(rw != READ && rw != READA))
  2.1067 ++    goto error_continue;
  2.1068 ++   rw = rq_data_dir(req);
  2.1069  +   if (unlikely(rw != READ
  2.1070  +#if LINUX_VERSION_CODE < KERNEL_VERSION(4, 8, 0)
  2.1071  +                && rw != READA
  2.1072  +#endif
  2.1073  +    ))
  2.1074 -     {
  2.1075 -      DEBUGP("cloop_do_request: bad command\n");
  2.1076 -      goto error_continue;
  2.1077 -@@ -409,40 +523,51 @@
  2.1078 -    continue; /* next request */
  2.1079 -   error_continue:
  2.1080 -    DEBUGP(KERN_ERR "cloop_do_request: Discarding request %p.\n", req);
  2.1081 ++    {
  2.1082 ++     DEBUGP("cloop_do_request: bad command\n");
  2.1083 ++     goto error_continue;
  2.1084 ++    }
  2.1085 ++   clo = req->rq_disk->private_data;
  2.1086 ++   if (unlikely(!clo->backing_file && !clo->suspended))
  2.1087 ++    {
  2.1088 ++     DEBUGP("cloop_do_request: not connected to a file\n");
  2.1089 ++     goto error_continue;
  2.1090 ++    }
  2.1091 ++   list_add_tail(&req->queuelist, &clo->clo_list); /* Add to working list for thread */
  2.1092 ++   wake_up(&clo->clo_event);    /* Wake up cloop_thread */
  2.1093 ++   continue; /* next request */
  2.1094 ++  error_continue:
  2.1095 ++   DEBUGP(KERN_ERR "cloop_do_request: Discarding request %p.\n", req);
  2.1096  +#if LINUX_VERSION_CODE < KERNEL_VERSION(4, 12, 0)
  2.1097 -    req->errors++;
  2.1098 ++   req->errors++;
  2.1099  +#else
  2.1100  +   req->error_count++;
  2.1101  +#endif
  2.1102 -    __blk_end_request_all(req, -EIO);
  2.1103 -   }
  2.1104 ++   __blk_end_request_all(req, -EIO);
  2.1105 ++  }
  2.1106   }
  2.1107   
  2.1108 --/* Read header and offsets from already opened file */
  2.1109 --static int cloop_set_file(int cloop_num, struct file *file, char *filename)
  2.1110 -+/* Read header, flags and offsets from already opened file */
  2.1111 -+static int cloop_set_file(int cloop_num, struct file *file)
  2.1112 - {
  2.1113 -  struct cloop_device *clo = cloop_dev[cloop_num];
  2.1114 -  struct inode *inode;
  2.1115 + /* Read header, flags and offsets from already opened file */
  2.1116 +@@ -508,7 +540,7 @@
  2.1117    char *bbuf=NULL;
  2.1118 -- unsigned int i, offsets_read, total_offsets;
  2.1119 -- int isblkdev;
  2.1120 -- int error = 0;
  2.1121 -+ unsigned int bbuf_size = 0;
  2.1122 -+ const unsigned int header_size = sizeof(struct cloop_head);
  2.1123 +  unsigned int bbuf_size = 0;
  2.1124 +  const unsigned int header_size = sizeof(struct cloop_head);
  2.1125 +- unsigned int i, offsets_read=0, total_offsets=0;
  2.1126  + unsigned int i, total_offsets=0;
  2.1127 -+ loff_t fs_read_position = 0, header_pos[2];
  2.1128 -+ int flags, isblkdev, bytes_read, error = 0;
  2.1129 -+ if (clo->suspended) return error;
  2.1130 -+ #if LINUX_VERSION_CODE < KERNEL_VERSION(4, 0, 0)
  2.1131 -  inode = file->f_dentry->d_inode;
  2.1132 -+ clo->underlying_filename = kstrdup(file->f_dentry->d_name.name ? file->f_dentry->d_name.name : (const unsigned char *)"anonymous filename", GFP_KERNEL);
  2.1133 -+ #else
  2.1134 -+ inode = file->f_path.dentry->d_inode;
  2.1135 -+ clo->underlying_filename = kstrdup(file->f_path.dentry->d_name.name ? file->f_path.dentry->d_name.name : (const unsigned char *)"anonymous filename", GFP_KERNEL);
  2.1136 -+ #endif
  2.1137 -  isblkdev=S_ISBLK(inode->i_mode)?1:0;
  2.1138 -  if(!isblkdev&&!S_ISREG(inode->i_mode))
  2.1139 +  loff_t fs_read_position = 0, header_pos[2];
  2.1140 +  int isblkdev, bytes_read, error = 0;
  2.1141 +  if (clo->suspended) return error;
  2.1142 +@@ -581,29 +613,19 @@
  2.1143 +     goto error_release;
  2.1144 +    }
  2.1145 +    memcpy(&clo->head, bbuf, header_size);
  2.1146 +-   if (strncmp(bbuf+CLOOP4_SIGNATURE_OFFSET, CLOOP4_SIGNATURE, CLOOP4_SIGNATURE_SIZE)==0)
  2.1147 ++   if (strncmp(bbuf+CLOOP_SIGNATURE_OFFSET, CLOOP_SIGNATURE, CLOOP_SIGNATURE_SIZE)==0)
  2.1148 +    {
  2.1149 +-    clo->file_format=4;
  2.1150 ++    clo->file_format++;
  2.1151 +     clo->head.block_size=ntohl(clo->head.block_size);
  2.1152 +     clo->head.num_blocks=ntohl(clo->head.num_blocks);
  2.1153 +     clo->header_first =  (i==0) ? 1 : 0;
  2.1154 +-    printk(KERN_INFO "%s: file %s version %d, %d blocks of %d bytes, header %s.\n", cloop_name, clo->underlying_filename, clo->file_format, clo->head.num_blocks, clo->head.block_size, (i==0)?"first":"last");
  2.1155 +-    break;
  2.1156 +-   }
  2.1157 +-   else if (strncmp(bbuf+CLOOP2_SIGNATURE_OFFSET, CLOOP2_SIGNATURE, CLOOP2_SIGNATURE_SIZE)==0)
  2.1158 +-   {
  2.1159 +-    clo->file_format=2;
  2.1160 +-    clo->head.block_size=ntohl(clo->head.block_size);
  2.1161 +-    clo->head.num_blocks=ntohl(clo->head.num_blocks);
  2.1162 +-    clo->header_first =  (i==0) ? 1 : 0;
  2.1163 +-    printk(KERN_INFO "%s: file %s version %d, %d blocks of %d bytes, header %s.\n", cloop_name, clo->underlying_filename, clo->file_format, clo->head.num_blocks, clo->head.block_size, (i==0)?"first":"last");
  2.1164 ++    printk(KERN_INFO "%s: file %s, %d blocks of %d bytes, header %s.\n", cloop_name, clo->underlying_filename, clo->head.num_blocks, clo->head.block_size, (i==0)?"first":"last");
  2.1165 +     break;
  2.1166 +    }
  2.1167 +   }
  2.1168 +  if (clo->file_format == 0)
  2.1169     {
  2.1170 -    printk(KERN_ERR "%s: %s not a regular file or block device\n",
  2.1171 --		   cloop_name, filename);
  2.1172 -+		   cloop_name, clo->underlying_filename);
  2.1173 +-   printk(KERN_ERR "%s: Cannot read old 32-bit (version 0.68) images, "
  2.1174 +-                   "please use an older version of %s for this file.\n",
  2.1175 ++   printk(KERN_ERR "%s: Cannot detect %s format.\n",
  2.1176 +                    cloop_name, cloop_name);
  2.1177 +        error=-EBADF; goto error_release;
  2.1178 +   }
  2.1179 +@@ -613,67 +635,133 @@
  2.1180 +           cloop_name, clo->head.block_size);
  2.1181      error=-EBADF; goto error_release;
  2.1182     }
  2.1183 -  clo->backing_file = file;
  2.1184 -  clo->backing_inode= inode ;
  2.1185 -- if(!isblkdev&&inode->i_size<sizeof(struct cloop_head))
  2.1186 -+ clo->underlying_total_size = (isblkdev) ? inode->i_bdev->bd_inode->i_size : inode->i_size;
  2.1187 -+ if(clo->underlying_total_size < header_size)
  2.1188 +- total_offsets=clo->head.num_blocks+1;
  2.1189 +- if (!isblkdev && (sizeof(struct cloop_head)+sizeof(loff_t)*
  2.1190 ++ total_offsets=clo->head.num_blocks;
  2.1191 ++ if (!isblkdev && (sizeof(struct cloop_head)+sizeof(struct block_info)*
  2.1192 +                       total_offsets > inode->i_size))
  2.1193     {
  2.1194 --   printk(KERN_ERR "%s: %lu bytes (must be >= %u bytes)\n",
  2.1195 --                   cloop_name, (unsigned long)inode->i_size,
  2.1196 --		   (unsigned)sizeof(struct cloop_head));
  2.1197 -+   printk(KERN_ERR "%s: %llu bytes (must be >= %u bytes)\n",
  2.1198 -+                   cloop_name, clo->underlying_total_size,
  2.1199 -+		   (unsigned int)header_size);
  2.1200 +    printk(KERN_ERR "%s: file %s too small for %u blocks\n",
  2.1201 +           cloop_name, clo->underlying_filename, clo->head.num_blocks);
  2.1202      error=-EBADF; goto error_release;
  2.1203     }
  2.1204 -- /* In suspended mode, we have done all checks necessary - FF */
  2.1205 -- if (clo->suspended)
  2.1206 --   return error;
  2.1207 -  if(isblkdev)
  2.1208 -   {
  2.1209 -    struct request_queue *q = bdev_get_queue(inode->i_bdev);
  2.1210 -@@ -451,104 +576,225 @@
  2.1211 -    /* blk_queue_max_hw_segments(clo->clo_queue, queue_max_hw_segments(q)); */ /* Removed in 2.6.34 */
  2.1212 -    blk_queue_max_segment_size(clo->clo_queue, queue_max_segment_size(q));
  2.1213 -    blk_queue_segment_boundary(clo->clo_queue, queue_segment_boundary(q));
  2.1214 -+#if LINUX_VERSION_CODE < KERNEL_VERSION(4, 3, 0)
  2.1215 -    blk_queue_merge_bvec(clo->clo_queue, q->merge_bvec_fn);
  2.1216 -+#endif
  2.1217 -    clo->underlying_blksize = block_size(inode->i_bdev);
  2.1218 -   }
  2.1219 -  else
  2.1220 -    clo->underlying_blksize = PAGE_SIZE;
  2.1221 -- DEBUGP("Underlying blocksize is %u\n", clo->underlying_blksize);
  2.1222 -- bbuf = cloop_malloc(clo->underlying_blksize);
  2.1223 -+
  2.1224 -+ DEBUGP(KERN_INFO "Underlying blocksize of %s is %u\n", clo->underlying_filename, clo->underlying_blksize);
  2.1225 -+ DEBUGP(KERN_INFO "Underlying total size of %s is %llu\n", clo->underlying_filename, clo->underlying_total_size);
  2.1226 -+
  2.1227 -+ /* clo->underlying_blksize should be larger than header_size, even if it's only PAGE_SIZE */
  2.1228 -+ bbuf_size = clo->underlying_blksize;
  2.1229 -+ bbuf = cloop_malloc(bbuf_size);
  2.1230 -  if(!bbuf)
  2.1231 -   {
  2.1232 --   printk(KERN_ERR "%s: out of kernel mem for block buffer (%lu bytes)\n",
  2.1233 --                   cloop_name, (unsigned long)clo->underlying_blksize);
  2.1234 -+   printk(KERN_ERR "%s: out of kernel mem for buffer (%u bytes)\n",
  2.1235 -+                   cloop_name, (unsigned int) bbuf_size);
  2.1236 -+   error=-ENOMEM; goto error_release;
  2.1237 -+  }
  2.1238 -+
  2.1239 -+ header_pos[0] = 0; /* header first */
  2.1240 -+ header_pos[1] = clo->underlying_total_size - sizeof(struct cloop_head); /* header last */
  2.1241 -+ for(i=0; i<2; i++)
  2.1242 -+  {
  2.1243 -+   /* Check for header */
  2.1244 -+   size_t bytes_readable = MIN(clo->underlying_blksize, clo->underlying_total_size - header_pos[i]);
  2.1245 -+   size_t bytes_read = cloop_read_from_file(clo, file, bbuf, header_pos[i], bytes_readable);
  2.1246 -+   if(bytes_read != bytes_readable)
  2.1247 -+   {
  2.1248 -+    printk(KERN_ERR "%s: Bad file %s, read() of %s %u bytes returned %d.\n",
  2.1249 -+                    cloop_name, clo->underlying_filename, (i==0)?"first":"last",
  2.1250 -+		    (unsigned int)header_size, (int)bytes_read);
  2.1251 -+    error=-EBADF;
  2.1252 -+    goto error_release;
  2.1253 -+   }
  2.1254 -+   memcpy(&clo->head, bbuf, header_size);
  2.1255 -+   if (strncmp(bbuf+CLOOP_SIGNATURE_OFFSET, CLOOP_SIGNATURE, CLOOP_SIGNATURE_SIZE)==0)
  2.1256 -+   {
  2.1257 -+    clo->file_format++;
  2.1258 -+    clo->head.block_size=ntohl(clo->head.block_size);
  2.1259 -+    clo->head.num_blocks=ntohl(clo->head.num_blocks);
  2.1260 -+    clo->header_first =  (i==0) ? 1 : 0;
  2.1261 -+    printk(KERN_INFO "%s: file %s, %d blocks of %d bytes, header %s.\n", cloop_name, clo->underlying_filename, clo->head.num_blocks, clo->head.block_size, (i==0)?"first":"last");
  2.1262 -+    break;
  2.1263 -+   }
  2.1264 -+  }
  2.1265 -+ if (clo->file_format == 0)
  2.1266 -+  {
  2.1267 -+   printk(KERN_ERR "%s: Cannot detect %s format.\n",
  2.1268 -+                   cloop_name, cloop_name);
  2.1269 -+       error=-EBADF; goto error_release;
  2.1270 -+  }
  2.1271 -+ if (clo->head.block_size % 512 != 0)
  2.1272 -+  {
  2.1273 -+   printk(KERN_ERR "%s: blocksize %u not multiple of 512\n",
  2.1274 -+          cloop_name, clo->head.block_size);
  2.1275 -+   error=-EBADF; goto error_release;
  2.1276 -+  }
  2.1277 -+ total_offsets=clo->head.num_blocks;
  2.1278 -+ if (!isblkdev && (sizeof(struct cloop_head)+sizeof(struct block_info)*
  2.1279 -+                      total_offsets > inode->i_size))
  2.1280 -+  {
  2.1281 -+   printk(KERN_ERR "%s: file %s too small for %u blocks\n",
  2.1282 -+          cloop_name, clo->underlying_filename, clo->head.num_blocks);
  2.1283 -+   error=-EBADF; goto error_release;
  2.1284 -+  }
  2.1285 +- clo->block_ptrs = cloop_malloc(sizeof(cloop_block_ptr) * total_offsets);
  2.1286 +- if (!clo->block_ptrs)
  2.1287  + /* Allocate Memory for decompressors */
  2.1288  +#if (defined(CONFIG_ZLIB_INFLATE) || defined(CONFIG_ZLIB_INFLATE_MODULE))
  2.1289  + clo->zstream.workspace = cloop_malloc(zlib_inflate_workspacesize());
  2.1290  + if(!clo->zstream.workspace)
  2.1291 -+  {
  2.1292 +   {
  2.1293 +-   printk(KERN_ERR "%s: out of kernel mem for offsets\n", cloop_name);
  2.1294  +   printk(KERN_ERR "%s: out of mem for zlib working area %u\n",
  2.1295  +          cloop_name, zlib_inflate_workspacesize());
  2.1296      error=-ENOMEM; goto error_release;
  2.1297     }
  2.1298 -- total_offsets = 1; /* Dummy total_offsets: will be filled in first time around */
  2.1299 -- for (i = 0, offsets_read = 0; offsets_read < total_offsets; i++)
  2.1300 +- /* Read them offsets! */
  2.1301 +- if(clo->header_first)
  2.1302  + zlib_inflateInit(&clo->zstream);
  2.1303  +#endif
  2.1304  +#if (defined(CONFIG_DECOMPRESS_XZ) || defined(CONFIG_DECOMPRESS_XZ_MODULE))
  2.1305 @@ -890,33 +713,20 @@
  2.1306  +#endif
  2.1307  + if (total_offsets + 1 == 0) /* Version 3 */
  2.1308     {
  2.1309 --   unsigned int offset = 0, num_readable;
  2.1310 --   size_t bytes_read = cloop_read_from_file(clo, file, bbuf,
  2.1311 --                                          i*clo->underlying_blksize,
  2.1312 --                                          clo->underlying_blksize);
  2.1313 --   if(bytes_read != clo->underlying_blksize)
  2.1314 +-   fs_read_position = sizeof(struct cloop_head);
  2.1315  +   struct cloop_tail tail;
  2.1316  +   if (isblkdev)
  2.1317 -     {
  2.1318 --     printk(KERN_ERR "%s: Bad file, read() of first %lu bytes returned %d.\n",
  2.1319 --                   cloop_name, (unsigned long)clo->underlying_blksize, (int)bytes_read);
  2.1320 --     error=-EBADF;
  2.1321 --     goto error_release;
  2.1322 ++    {
  2.1323  +    /* No end of file: can't find index */
  2.1324  +     printk(KERN_ERR "%s: no V3 support for block device\n", 
  2.1325  +            cloop_name);
  2.1326  +     error=-EBADF; goto error_release;
  2.1327 -     }
  2.1328 --   /* Header will be in block zero */
  2.1329 --   if(i==0)
  2.1330 ++    }
  2.1331  +   bytes_read = cloop_read_from_file(clo, file, (void *) &tail,
  2.1332  +			inode->i_size - sizeof(struct cloop_tail),
  2.1333  +			sizeof(struct cloop_tail));
  2.1334  +   if (bytes_read == sizeof(struct cloop_tail))
  2.1335 -     {
  2.1336 --     memcpy(&clo->head, bbuf, sizeof(struct cloop_head));
  2.1337 --     offset = sizeof(struct cloop_head);
  2.1338 --     if (ntohl(clo->head.block_size) % 512 != 0)
  2.1339 ++    {
  2.1340  +     unsigned long len, zlen;
  2.1341  +     int ret;
  2.1342  +     void *zbuf;
  2.1343 @@ -926,79 +736,47 @@
  2.1344  +     zlen = ntohl(tail.table_size);
  2.1345  +     zbuf = cloop_malloc(zlen);
  2.1346  +     if (!clo->block_ptrs || !zbuf)
  2.1347 -       {
  2.1348 --       printk(KERN_ERR "%s: blocksize %u not multiple of 512\n",
  2.1349 --              cloop_name, ntohl(clo->head.block_size));
  2.1350 --       error=-EBADF; goto error_release;
  2.1351 --      }
  2.1352 --     if (clo->head.preamble[0x0B]!='V'||clo->head.preamble[0x0C]<'1')
  2.1353 --      {
  2.1354 --       printk(KERN_ERR "%s: Cannot read old 32-bit (version 0.68) images, "
  2.1355 --		       "please use an older version of %s for this file.\n",
  2.1356 --		       cloop_name, cloop_name);
  2.1357 --       error=-EBADF; goto error_release;
  2.1358 ++      {
  2.1359  +       printk(KERN_ERR "%s: out of kernel mem for index\n", cloop_name);
  2.1360  +       error=-ENOMEM; goto error_release;
  2.1361 -       }
  2.1362 --     if (clo->head.preamble[0x0C]<'2')
  2.1363 ++      }
  2.1364  +     bytes_read = cloop_read_from_file(clo, file, zbuf,
  2.1365  +			inode->i_size - zlen - sizeof(struct cloop_tail),
  2.1366  +			zlen);
  2.1367  +     if (bytes_read != zlen)
  2.1368 -       {
  2.1369 --       printk(KERN_ERR "%s: Cannot read old architecture-dependent "
  2.1370 --		       "(format <= 1.0) images, please use an older "
  2.1371 --		       "version of %s for this file.\n",
  2.1372 --		       cloop_name, cloop_name);
  2.1373 ++      {
  2.1374  +       printk(KERN_ERR "%s: can't read index\n", cloop_name);
  2.1375 -        error=-EBADF; goto error_release;
  2.1376 -       }
  2.1377 --     total_offsets=ntohl(clo->head.num_blocks)+1;
  2.1378 --     if (!isblkdev && (sizeof(struct cloop_head)+sizeof(loff_t)*
  2.1379 --                       total_offsets > inode->i_size))
  2.1380 ++       error=-EBADF; goto error_release;
  2.1381 ++      }
  2.1382  +     len = CLOOP3_INDEX_SIZE(ntohl(tail.index_size)) * total_offsets;
  2.1383 -+     flags = CLOOP3_BLOCKS_FLAGS(ntohl(tail.index_size));
  2.1384 -+// May  3 19:45:20 (none) user.info kernel: cloop: uncompress(clo=e0a78000, block_ptrs=e0c9c000, &len(1440)=ddc05e6c, zbuf=e0c9f000, zlen=43, flag=0)
  2.1385 -+printk(KERN_INFO "%s: uncompress(clo=%p, block_ptrs=%p, &len(%ld)=%p, zbuf=%p, zlen=%ld, flag=%d)\n", cloop_name, 
  2.1386 -+		clo, clo->block_ptrs, len, &len, zbuf, zlen, flags);
  2.1387 -+     ret = uncompress(clo, (void *) clo->block_ptrs, &len, zbuf, zlen, flags);
  2.1388 -+// May  3 19:45:20 (none) user.alert kernel: BUG: unable to handle kernel NULL pointer dereference at   (null)
  2.1389 -+printk(KERN_INFO "%s: uncompressed !\n", cloop_name);
  2.1390 ++     ret = uncompress(clo, (void *) clo->block_ptrs, &len, zbuf, zlen, CLOOP_COMPRESSOR_ZLIB);
  2.1391  +     cloop_free(zbuf, zlen);
  2.1392  +     if (ret != 0)
  2.1393 -       {
  2.1394 --       printk(KERN_ERR "%s: file too small for %u blocks\n",
  2.1395 --              cloop_name, ntohl(clo->head.num_blocks));
  2.1396 -+        printk(KERN_ERR "%s: decompression error %i uncompressing index, flags %u\n",
  2.1397 -+               cloop_name, ret, flags);
  2.1398 -        error=-EBADF; goto error_release;
  2.1399 -       }
  2.1400 --     clo->offsets = cloop_malloc(sizeof(loff_t) * total_offsets);
  2.1401 --     if (!clo->offsets)
  2.1402 --      {
  2.1403 --       printk(KERN_ERR "%s: out of kernel mem for offsets\n", cloop_name);
  2.1404 --       error=-ENOMEM; goto error_release;
  2.1405 --      }
  2.1406 -     }
  2.1407 --   num_readable = MIN(total_offsets - offsets_read,
  2.1408 --                      (clo->underlying_blksize - offset) 
  2.1409 --                      / sizeof(loff_t));
  2.1410 --   memcpy(&clo->offsets[offsets_read], bbuf+offset, num_readable * sizeof(loff_t));
  2.1411 --   offsets_read += num_readable;
  2.1412 --  }
  2.1413 --  { /* Search for largest block rather than estimate. KK. */
  2.1414 --   int i;
  2.1415 --   for(i=0;i<total_offsets-1;i++)
  2.1416 ++      {
  2.1417 ++        printk(KERN_ERR "%s: decompression error %i uncompressing index\n",
  2.1418 ++               cloop_name, ret);
  2.1419 ++       error=-EBADF; goto error_release;
  2.1420 ++      }
  2.1421 ++    }
  2.1422  +   else
  2.1423  +    {
  2.1424  +     printk(KERN_ERR "%s: can't find index\n", cloop_name);
  2.1425  +     error=-ENOMEM; goto error_release;
  2.1426  +    }
  2.1427 -+  }
  2.1428 -+ else
  2.1429 -+  {
  2.1430 +   }
  2.1431 +  else
  2.1432 +   {
  2.1433 +-   fs_read_position = clo->underlying_total_size - sizeof(struct cloop_head) - total_offsets * sizeof(loff_t);
  2.1434 +-  }
  2.1435 +- for(offsets_read=0;offsets_read<total_offsets;)
  2.1436 +-  {
  2.1437 +-   size_t bytes_readable;
  2.1438 +-   unsigned int num_readable, offset = 0;
  2.1439 +-   bytes_readable = MIN(bbuf_size, clo->underlying_total_size - fs_read_position);
  2.1440 +-   if(bytes_readable <= 0) break; /* Done */
  2.1441 +-   bytes_read = cloop_read_from_file(clo, file, bbuf, fs_read_position, bytes_readable);
  2.1442 +-   if(bytes_read != bytes_readable)
  2.1443  +   unsigned int n, total_bytes;
  2.1444 -+   flags = 0;
  2.1445  +   clo->block_ptrs = cloop_malloc(sizeof(struct block_info) * total_offsets);
  2.1446  +   if (!clo->block_ptrs)
  2.1447  +    {
  2.1448 @@ -1007,14 +785,26 @@
  2.1449  +    }
  2.1450  +   /* Read them offsets! */
  2.1451  +   if(clo->header_first)
  2.1452 -+    {
  2.1453 +     {
  2.1454 +-     printk(KERN_ERR "%s: Bad file %s, read() %lu bytes @ %llu returned %d.\n",
  2.1455 +-            cloop_name, clo->underlying_filename, (unsigned long)clo->underlying_blksize, fs_read_position, (int)bytes_read);
  2.1456 +-     error=-EBADF;
  2.1457 +-     goto error_release;
  2.1458  +     total_bytes = total_offsets * sizeof(struct block_info);
  2.1459  +     fs_read_position = sizeof(struct cloop_head);
  2.1460 -+    }
  2.1461 +     }
  2.1462 +-   /* remember where to read the next blk from file */
  2.1463 +-   fs_read_position += bytes_read;
  2.1464 +-   /* calculate how many offsets can be taken from current bbuf */
  2.1465 +-   num_readable = MIN(total_offsets - offsets_read,
  2.1466 +-                      bytes_read / sizeof(loff_t));
  2.1467 +-   DEBUGP(KERN_INFO "cloop: parsing %d offsets %d to %d\n", num_readable, offsets_read, offsets_read+num_readable-1);
  2.1468 +-   for (i=0,offset=0; i<num_readable; i++)
  2.1469  +   else
  2.1470       {
  2.1471 --     loff_t d=be64_to_cpu(clo->offsets[i+1]) - be64_to_cpu(clo->offsets[i]);
  2.1472 --     clo->largest_block=MAX(clo->largest_block,d);
  2.1473 +-     loff_t tmp = be64_to_cpu( *(loff_t*) (bbuf+offset) );
  2.1474 +-     if (i%50==0) DEBUGP(KERN_INFO "cloop: offset %03d: %llu\n", offsets_read, tmp);
  2.1475 +-     if(offsets_read > 0)
  2.1476  +     total_bytes = total_offsets * sizeof(loff_t);
  2.1477  +     fs_read_position = clo->underlying_total_size - sizeof(struct cloop_head) - total_bytes;
  2.1478  +    }
  2.1479 @@ -1025,35 +815,28 @@
  2.1480  +     if(bytes_readable <= 0) break; /* Done */
  2.1481  +     bytes_read = cloop_read_from_file(clo, file, bbuf, fs_read_position, bytes_readable);
  2.1482  +     if(bytes_read != bytes_readable)
  2.1483 -+      {
  2.1484 +       {
  2.1485 +-       loff_t d = CLOOP_BLOCK_OFFSET(tmp) - CLOOP_BLOCK_OFFSET(clo->block_ptrs[offsets_read-1]);
  2.1486 +-       if(d > clo->largest_block) clo->largest_block = d;
  2.1487  +       printk(KERN_ERR "%s: Bad file %s, read() %lu bytes @ %llu returned %d.\n",
  2.1488  +              cloop_name, clo->underlying_filename, (unsigned long)clo->underlying_blksize, fs_read_position, (int)bytes_read);
  2.1489  +       error=-EBADF;
  2.1490  +       goto error_release;
  2.1491 -+      }
  2.1492 +       }
  2.1493 +-     clo->block_ptrs[offsets_read++] = tmp;
  2.1494 +-     offset += sizeof(loff_t);
  2.1495  +     memcpy(((char *)clo->block_ptrs) + n, bbuf, bytes_read);
  2.1496  +     /* remember where to read the next blk from file */
  2.1497  +     fs_read_position += bytes_read;
  2.1498  +     n += bytes_read;
  2.1499       }
  2.1500 --   printk(KERN_INFO "%s: %s: %u blocks, %u bytes/block, largest block is %lu bytes.\n",
  2.1501 --          cloop_name, filename, ntohl(clo->head.num_blocks),
  2.1502 --          ntohl(clo->head.block_size), clo->largest_block);
  2.1503     }
  2.1504 --/* Combo kmalloc used too large chunks (>130000). */
  2.1505 +-  printk(KERN_INFO "%s: %s: %u blocks, %u bytes/block, largest block is %lu bytes.\n",
  2.1506 +-         cloop_name, clo->underlying_filename, clo->head.num_blocks,
  2.1507 +-         clo->head.block_size, clo->largest_block);
  2.1508    {
  2.1509     int i;
  2.1510 --  for(i=0;i<BUFFERED_BLOCKS;i++)
  2.1511 --   {
  2.1512 --    clo->buffer[i] = cloop_malloc(ntohl(clo->head.block_size));
  2.1513 --    if(!clo->buffer[i])
  2.1514 --     {
  2.1515 --      printk(KERN_ERR "%s: out of memory for buffer %lu\n",
  2.1516 --             cloop_name, (unsigned long) ntohl(clo->head.block_size));
  2.1517 --      error=-ENOMEM; goto error_release_free;
  2.1518 --     }
  2.1519 --   }
  2.1520 -+  char *version = build_index(clo->block_ptrs, clo->head.num_blocks, clo->head.block_size, flags);
  2.1521 ++  char *version = build_index(clo->block_ptrs, clo->head.num_blocks, clo->head.block_size);
  2.1522  +  clo->largest_block = 0;
  2.1523  +  for (i = 0; i < clo->head.num_blocks; i++)
  2.1524  +    if (clo->block_ptrs[i].size > clo->largest_block)
  2.1525 @@ -1061,39 +844,15 @@
  2.1526  +  printk(KERN_INFO "%s: %s: %s: %u blocks, %u bytes/block, largest block is %lu bytes.\n",
  2.1527  +         cloop_name, clo->underlying_filename, version, clo->head.num_blocks,
  2.1528  +         clo->head.block_size, clo->largest_block);
  2.1529 -+ }
  2.1530 -+ {
  2.1531 -+  int i;
  2.1532 -+  clo->num_buffered_blocks = (buffers > 0 && clo->head.block_size >= 512) ?
  2.1533 -+                              (buffers / clo->head.block_size) : 1;
  2.1534 -+  clo->buffered_blocknum = cloop_malloc(clo->num_buffered_blocks * sizeof (u_int32_t));
  2.1535 -+  clo->buffer = cloop_malloc(clo->num_buffered_blocks * sizeof (char*));
  2.1536 -+  if (!clo->buffered_blocknum || !clo->buffer)
  2.1537 -+  {
  2.1538 -+   printk(KERN_ERR "%s: out of memory for index of cache buffer (%lu bytes)\n",
  2.1539 -+                    cloop_name, (unsigned long)clo->num_buffered_blocks * sizeof (u_int32_t) + sizeof(char*) );
  2.1540 -+                    error=-ENOMEM; goto error_release;
  2.1541 -+  }
  2.1542 -+  memset(clo->buffer, 0, clo->num_buffered_blocks * sizeof (char*));
  2.1543 -+  for(i=0;i<clo->num_buffered_blocks;i++)
  2.1544 -+  {
  2.1545 -+   clo->buffered_blocknum[i] = -1;
  2.1546 -+   clo->buffer[i] = cloop_malloc(clo->head.block_size);
  2.1547 -+   if(!clo->buffer[i])
  2.1548 -+    {
  2.1549 -+     printk(KERN_ERR "%s: out of memory for cache buffer %lu\n",
  2.1550 -+            cloop_name, (unsigned long) clo->head.block_size);
  2.1551 -+     error=-ENOMEM; goto error_release_free;
  2.1552 -+    }
  2.1553 -+  }
  2.1554 -+  clo->current_bufnum = 0;
  2.1555 -  }
  2.1556 -  clo->compressed_buffer = cloop_malloc(clo->largest_block);
  2.1557 -  if(!clo->compressed_buffer)
  2.1558 -@@ -557,31 +803,7 @@
  2.1559 +   clo->num_buffered_blocks = (buffers > 0 && clo->head.block_size >= 512) ?
  2.1560 +                               (buffers / clo->head.block_size) : 1;
  2.1561 +   clo->buffered_blocknum = cloop_malloc(clo->num_buffered_blocks * sizeof (u_int32_t));
  2.1562 +@@ -705,36 +793,14 @@
  2.1563             cloop_name, clo->largest_block);
  2.1564      error=-ENOMEM; goto error_release_free_buffer;
  2.1565     }
  2.1566 +- /* Allocate Memory for decompressors */
  2.1567 +-#if (defined(CONFIG_ZLIB_INFLATE) || defined(CONFIG_ZLIB_INFLATE_MODULE))
  2.1568  - clo->zstream.workspace = cloop_malloc(zlib_inflate_workspacesize());
  2.1569  - if(!clo->zstream.workspace)
  2.1570  -  {
  2.1571 @@ -1102,443 +861,48 @@
  2.1572  -   error=-ENOMEM; goto error_release_free_all;
  2.1573  -  }
  2.1574  - zlib_inflateInit(&clo->zstream);
  2.1575 -- if(!isblkdev &&
  2.1576 --    be64_to_cpu(clo->offsets[ntohl(clo->head.num_blocks)]) != inode->i_size)
  2.1577 --  {
  2.1578 --   printk(KERN_ERR "%s: final offset wrong (%Lu not %Lu)\n",
  2.1579 +-#endif
  2.1580 +-#if (defined(CONFIG_DECOMPRESS_XZ) || defined(CONFIG_DECOMPRESS_XZ_MODULE))
  2.1581 +-#if XZ_INTERNAL_CRC32
  2.1582 +-  /* This must be called before any other xz_* function to initialize the CRC32 lookup table. */
  2.1583 +-  xz_crc32_init(void);
  2.1584 +-#endif
  2.1585 +-  clo->xzdecoderstate = xz_dec_init(XZ_SINGLE, 0);
  2.1586 +-#endif
  2.1587 +- if(CLOOP_BLOCK_OFFSET(clo->block_ptrs[clo->head.num_blocks]) > clo->underlying_total_size)
  2.1588 ++ set_capacity(clo->clo_disk, (sector_t)(clo->head.num_blocks*(clo->head.block_size>>9)));
  2.1589 ++ clo->clo_thread = kthread_create(cloop_thread, clo, "cloop%d", cloop_num);
  2.1590 ++ if(IS_ERR(clo->clo_thread))
  2.1591 +   {
  2.1592 +-   printk(KERN_ERR "%s: final offset wrong (%llu > %llu)\n",
  2.1593  -          cloop_name,
  2.1594 --          be64_to_cpu(clo->offsets[ntohl(clo->head.num_blocks)]),
  2.1595 --          inode->i_size);
  2.1596 +-	  CLOOP_BLOCK_OFFSET(clo->block_ptrs[clo->head.num_blocks]),
  2.1597 +-          clo->underlying_total_size);
  2.1598 +-#if (defined(CONFIG_ZLIB_INFLATE) || defined(CONFIG_ZLIB_INFLATE_MODULE))
  2.1599  -   cloop_free(clo->zstream.workspace, zlib_inflate_workspacesize()); clo->zstream.workspace=NULL;
  2.1600 --   goto error_release_free_all;
  2.1601 --  }
  2.1602 -- {
  2.1603 --  int i;
  2.1604 --  for(i=0; i<BUFFERED_BLOCKS; i++) clo->buffered_blocknum[i] = -1;
  2.1605 --  clo->current_bufnum=0;
  2.1606 -- }
  2.1607 -- set_capacity(clo->clo_disk, (sector_t)(ntohl(clo->head.num_blocks)*
  2.1608 --              (ntohl(clo->head.block_size)>>9)));
  2.1609 -+ set_capacity(clo->clo_disk, (sector_t)(clo->head.num_blocks*(clo->head.block_size>>9)));
  2.1610 -  clo->clo_thread = kthread_create(cloop_thread, clo, "cloop%d", cloop_num);
  2.1611 -  if(IS_ERR(clo->clo_thread))
  2.1612 -   {
  2.1613 -@@ -591,17 +813,17 @@
  2.1614 +-#endif
  2.1615 ++   error = PTR_ERR(clo->clo_thread);
  2.1616 ++   clo->clo_thread=NULL;
  2.1617 +    goto error_release_free_all;
  2.1618     }
  2.1619 +- set_capacity(clo->clo_disk, (sector_t)(clo->head.num_blocks*(clo->head.block_size>>9)));
  2.1620    if(preload > 0)
  2.1621     {
  2.1622 --   clo->preload_array_size = ((preload<=ntohl(clo->head.num_blocks))?preload:ntohl(clo->head.num_blocks));
  2.1623 -+   clo->preload_array_size = ((preload<=clo->head.num_blocks)?preload:clo->head.num_blocks);
  2.1624 -    clo->preload_size = 0;
  2.1625 -    if((clo->preload_cache = cloop_malloc(clo->preload_array_size * sizeof(char *))) != NULL)
  2.1626 -     {
  2.1627 -      int i;
  2.1628 -      for(i=0; i<clo->preload_array_size; i++)
  2.1629 -       {
  2.1630 --       if((clo->preload_cache[i] = cloop_malloc(ntohl(clo->head.block_size))) == NULL)
  2.1631 -+       if((clo->preload_cache[i] = cloop_malloc(clo->head.block_size)) == NULL)
  2.1632 -         { /* Out of memory */
  2.1633 -          printk(KERN_WARNING "%s: cloop_malloc(%d) failed for preload_cache[%d] (ignored).\n",
  2.1634 --                             cloop_name, ntohl(clo->head.block_size), i);
  2.1635 -+                             cloop_name, clo->head.block_size, i);
  2.1636 - 	 break;
  2.1637 - 	}
  2.1638 -       }
  2.1639 -@@ -612,13 +834,13 @@
  2.1640 -        if(buffered_blocknum >= 0)
  2.1641 -         {
  2.1642 - 	 memcpy(clo->preload_cache[i], clo->buffer[buffered_blocknum],
  2.1643 --	        ntohl(clo->head.block_size));
  2.1644 -+	        clo->head.block_size);
  2.1645 - 	}
  2.1646 -        else
  2.1647 -         {
  2.1648 -          printk(KERN_WARNING "%s: can't read block %d into preload cache, set to zero.\n",
  2.1649 - 	                     cloop_name, i);
  2.1650 --	 memset(clo->preload_cache[i], 0, ntohl(clo->head.block_size));
  2.1651 -+	 memset(clo->preload_cache[i], 0, clo->head.block_size);
  2.1652 - 	}
  2.1653 -       }
  2.1654 -      printk(KERN_INFO "%s: preloaded %d blocks into cache.\n", cloop_name,
  2.1655 -@@ -641,22 +863,19 @@
  2.1656 -  cloop_free(clo->compressed_buffer, clo->largest_block);
  2.1657 -  clo->compressed_buffer=NULL;
  2.1658 - error_release_free_buffer:
  2.1659 -+ if(clo->buffer)
  2.1660 -  {
  2.1661 -   int i;
  2.1662 --  for(i=0; i<BUFFERED_BLOCKS; i++)
  2.1663 --   { 
  2.1664 --    if(clo->buffer[i])
  2.1665 --     {
  2.1666 --      cloop_free(clo->buffer[i], ntohl(clo->head.block_size));
  2.1667 --      clo->buffer[i]=NULL;
  2.1668 --     }
  2.1669 --   }
  2.1670 -+  for(i=0; i<clo->num_buffered_blocks; i++) { if(clo->buffer[i]) { cloop_free(clo->buffer[i], clo->head.block_size); clo->buffer[i]=NULL; }}
  2.1671 -+  cloop_free(clo->buffer, clo->num_buffered_blocks*sizeof(char*)); clo->buffer=NULL;
  2.1672 +    clo->preload_array_size = ((preload<=clo->head.num_blocks)?preload:clo->head.num_blocks);
  2.1673 +@@ -780,6 +846,7 @@
  2.1674 +      clo->preload_array_size = clo->preload_size = 0;
  2.1675 +     }
  2.1676 +   }
  2.1677 ++ wake_up_process(clo->clo_thread);
  2.1678 +  /* Uncheck */
  2.1679 +  return error;
  2.1680 + error_release_free_all:
  2.1681 +@@ -794,9 +861,13 @@
  2.1682    }
  2.1683 -+ if (clo->buffered_blocknum) { cloop_free(clo->buffered_blocknum, sizeof(int)*clo->num_buffered_blocks); clo->buffered_blocknum=NULL; }
  2.1684 +  if (clo->buffered_blocknum) { cloop_free(clo->buffered_blocknum, sizeof(int)*clo->num_buffered_blocks); clo->buffered_blocknum=NULL; }
  2.1685   error_release_free:
  2.1686 -- cloop_free(clo->offsets, sizeof(loff_t) * total_offsets);
  2.1687 -- clo->offsets=NULL;
  2.1688 +- cloop_free(clo->block_ptrs, sizeof(cloop_block_ptr) * total_offsets);
  2.1689  + cloop_free(clo->block_ptrs, sizeof(struct block_info) * total_offsets);
  2.1690 -+ clo->block_ptrs=NULL;
  2.1691 - error_release:
  2.1692 -  if(bbuf) cloop_free(bbuf, clo->underlying_blksize);
  2.1693 -+ if(clo->underlying_filename) { kfree(clo->underlying_filename); clo->underlying_filename=NULL; }
  2.1694 -  clo->backing_file=NULL;
  2.1695 -  return error;
  2.1696 - }
  2.1697 -@@ -673,7 +892,7 @@
  2.1698 -  if(clo->backing_file) return -EBUSY;
  2.1699 -  file = fget(arg); /* get filp struct from ioctl arg fd */
  2.1700 -  if(!file) return -EBADF;
  2.1701 -- error=cloop_set_file(cloop_num,file,"losetup_file");
  2.1702 -+ error=cloop_set_file(cloop_num,file);
  2.1703 -  set_device_ro(bdev, 1);
  2.1704 -  if(error) fput(file);
  2.1705 -  return error;
  2.1706 -@@ -684,29 +903,48 @@
  2.1707 - {
  2.1708 -  struct cloop_device *clo = cloop_dev[cloop_num];
  2.1709 -  struct file *filp = clo->backing_file;
  2.1710 -- int i;
  2.1711 -  if(clo->refcnt > 1)	/* we needed one fd for the ioctl */
  2.1712 -    return -EBUSY;
  2.1713 -  if(filp==NULL) return -EINVAL;
  2.1714 -  if(clo->clo_thread) { kthread_stop(clo->clo_thread); clo->clo_thread=NULL; }
  2.1715 -- if(filp!=initial_file) fput(filp);
  2.1716 -- else { filp_close(initial_file,0); initial_file=NULL; }
  2.1717 -+ if(filp!=initial_file)
  2.1718 -+  fput(filp);
  2.1719 -+ else
  2.1720 -+ {
  2.1721 -+  filp_close(initial_file,0);
  2.1722 -+  initial_file=NULL;
  2.1723 -+ }
  2.1724 -  clo->backing_file  = NULL;
  2.1725 -  clo->backing_inode = NULL;
  2.1726 -- if(clo->offsets) { cloop_free(clo->offsets, clo->underlying_blksize); clo->offsets = NULL; }
  2.1727 -+ if(clo->underlying_filename) { kfree(clo->underlying_filename); clo->underlying_filename=NULL; }
  2.1728 -+ if(clo->block_ptrs) { cloop_free(clo->block_ptrs, clo->head.num_blocks); clo->block_ptrs = NULL; }
  2.1729 -  if(clo->preload_cache)
  2.1730 --  {
  2.1731 --   for(i=0; i < clo->preload_size; i++)
  2.1732 --    cloop_free(clo->preload_cache[i], ntohl(clo->head.block_size));
  2.1733 --   cloop_free(clo->preload_cache, clo->preload_array_size * sizeof(char *));
  2.1734 --   clo->preload_cache = NULL;
  2.1735 --   clo->preload_size = clo->preload_array_size = 0;
  2.1736 --  }
  2.1737 -- for(i=0; i<BUFFERED_BLOCKS; i++)
  2.1738 --      if(clo->buffer[i]) { cloop_free(clo->buffer[i], ntohl(clo->head.block_size)); clo->buffer[i]=NULL; }
  2.1739 -+ {
  2.1740 -+  int i;
  2.1741 -+  for(i=0; i < clo->preload_size; i++)
  2.1742 -+   cloop_free(clo->preload_cache[i], clo->head.block_size);
  2.1743 -+  cloop_free(clo->preload_cache, clo->preload_array_size * sizeof(char *));
  2.1744 -+  clo->preload_cache = NULL;
  2.1745 -+  clo->preload_size = clo->preload_array_size = 0;
  2.1746 -+ }
  2.1747 -+ if (clo->buffered_blocknum)
  2.1748 -+ {
  2.1749 -+  cloop_free(clo->buffered_blocknum, sizeof(int) * clo->num_buffered_blocks); clo->buffered_blocknum = NULL;
  2.1750 -+ }
  2.1751 -+ if (clo->buffer)
  2.1752 -+ {
  2.1753 -+  int i;
  2.1754 -+  for(i=0; i<clo->num_buffered_blocks; i++) { if(clo->buffer[i]) cloop_free(clo->buffer[i], clo->head.block_size); }
  2.1755 -+  cloop_free(clo->buffer, sizeof(char*) * clo->num_buffered_blocks); clo->buffer = NULL;
  2.1756 -+ }
  2.1757 -  if(clo->compressed_buffer) { cloop_free(clo->compressed_buffer, clo->largest_block); clo->compressed_buffer = NULL; }
  2.1758 -+#if (defined(CONFIG_ZLIB_INFLATE) || defined(CONFIG_ZLIB_INFLATE_MODULE))
  2.1759 -  zlib_inflateEnd(&clo->zstream);
  2.1760 -  if(clo->zstream.workspace) { cloop_free(clo->zstream.workspace, zlib_inflate_workspacesize()); clo->zstream.workspace = NULL; }
  2.1761 -+#endif
  2.1762 -+#if (defined(CONFIG_DECOMPRESS_XZ) || defined(CONFIG_DECOMPRESS_XZ_MODULE))
  2.1763 -+  xz_dec_end(clo->xzdecoderstate);
  2.1764 -+#endif
  2.1765 -  if(bdev) invalidate_bdev(bdev);
  2.1766 -  if(clo->clo_disk) set_capacity(clo->clo_disk, 0);
  2.1767 -  return 0;
  2.1768 -@@ -731,8 +969,8 @@
  2.1769 -                             const struct loop_info64 *info)
  2.1770 - {
  2.1771 -  if (!clo->backing_file) return -ENXIO;
  2.1772 -- memcpy(clo->clo_file_name, info->lo_file_name, LO_NAME_SIZE);
  2.1773 -- clo->clo_file_name[LO_NAME_SIZE-1] = 0;
  2.1774 -+ if(clo->underlying_filename) kfree(clo->underlying_filename);
  2.1775 -+ clo->underlying_filename = kstrdup(info->lo_file_name, GFP_KERNEL);
  2.1776 -  return 0;
  2.1777 - }
  2.1778 - 
  2.1779 -@@ -743,7 +981,11 @@
  2.1780 -  struct kstat stat;
  2.1781 -  int err;
  2.1782 -  if (!file) return -ENXIO;
  2.1783 -- err = vfs_getattr(file->f_path.mnt, file->f_path.dentry, &stat);
  2.1784 -+#if LINUX_VERSION_CODE < KERNEL_VERSION(4, 11, 0)
  2.1785 -+ err = vfs_getattr(&file->f_path, &stat);
  2.1786 -+#else
  2.1787 -+ err = vfs_getattr(&file->f_path, &stat, STATX_INO, AT_STATX_SYNC_AS_STAT);
  2.1788 -+#endif
  2.1789 -  if (err) return err;
  2.1790 -  memset(info, 0, sizeof(*info));
  2.1791 -  info->lo_number  = clo->clo_number;
  2.1792 -@@ -753,7 +995,8 @@
  2.1793 -  info->lo_offset  = 0;
  2.1794 -  info->lo_sizelimit = 0;
  2.1795 -  info->lo_flags   = 0;
  2.1796 -- memcpy(info->lo_file_name, clo->clo_file_name, LO_NAME_SIZE);
  2.1797 -+ strncpy(info->lo_file_name, clo->underlying_filename, LO_NAME_SIZE);
  2.1798 -+ info->lo_file_name[LO_NAME_SIZE-1]=0;
  2.1799 -  return 0;
  2.1800 - }
  2.1801 - 
  2.1802 -@@ -833,8 +1076,6 @@
  2.1803 -  if (!err && copy_to_user(arg, &info64, sizeof(info64))) err = -EFAULT;
  2.1804 -  return err;
  2.1805 - }
  2.1806 --/* EOF get/set_status */
  2.1807 --
  2.1808 - 
  2.1809 - static int cloop_ioctl(struct block_device *bdev, fmode_t mode,
  2.1810 - 	unsigned int cmd, unsigned long arg)
  2.1811 -@@ -914,21 +1155,20 @@
  2.1812 -  /* losetup uses write-open and flags=0x8002 to set a new file */
  2.1813 -  if(mode & FMODE_WRITE)
  2.1814 -   {
  2.1815 --   printk(KERN_WARNING "%s: Can't open device read-write in mode 0x%x\n", cloop_name, mode);
  2.1816 -+   printk(KERN_INFO "%s: Open in read-write mode 0x%x requested, ignored.\n", cloop_name, mode);
  2.1817 -    return -EROFS;
  2.1818 -   }
  2.1819 -  cloop_dev[cloop_num]->refcnt+=1;
  2.1820 -  return 0;
  2.1821 - }
  2.1822 - 
  2.1823 --static int cloop_close(struct gendisk *disk, fmode_t mode)
  2.1824 -+static void cloop_close(struct gendisk *disk, fmode_t mode)
  2.1825 - {
  2.1826 -- int cloop_num, err=0;
  2.1827 -- if(!disk) return 0;
  2.1828 -+ int cloop_num;
  2.1829 -+ if(!disk) return;
  2.1830 -  cloop_num=((struct cloop_device *)disk->private_data)->clo_number;
  2.1831 -- if(cloop_num < 0 || cloop_num > (cloop_count-1)) return 0;
  2.1832 -+ if(cloop_num < 0 || cloop_num > (cloop_count-1)) return;
  2.1833 -  cloop_dev[cloop_num]->refcnt-=1;
  2.1834 -- return err;
  2.1835 - }
  2.1836 - 
  2.1837 - static struct block_device_operations clo_fops =
  2.1838 -@@ -973,6 +1213,10 @@
  2.1839 -    goto error_out;
  2.1840 -   }
  2.1841 -  clo->clo_queue->queuedata = clo;
  2.1842 -+#if LINUX_VERSION_CODE < KERNEL_VERSION(4, 12, 0)
  2.1843 -+ queue_flag_set_unlocked(QUEUE_FLAG_NONROT, clo->clo_queue);
  2.1844 -+ queue_flag_set_unlocked(QUEUE_FLAG_NOMERGES, clo->clo_queue);
  2.1845 -+#endif
  2.1846 -  clo->clo_disk = alloc_disk(1);
  2.1847 -  if(!clo->clo_disk)
  2.1848 -   {
  2.1849 -@@ -1004,6 +1248,11 @@
  2.1850 -  cloop_dev[cloop_num] = NULL;
  2.1851 - }
  2.1852 - 
  2.1853 -+/* LZ4 Stuff */
  2.1854 -+#if (defined USE_LZ4_INTERNAL)
  2.1855 -+#include "lz4_kmod.c"
  2.1856 -+#endif
  2.1857 -+
  2.1858 - static int __init cloop_init(void)
  2.1859 - {
  2.1860 -  int error=0;
  2.1861 -@@ -1044,7 +1293,7 @@
  2.1862 -      initial_file=NULL; /* if IS_ERR, it's NOT open. */
  2.1863 -     }
  2.1864 -    else
  2.1865 --     error=cloop_set_file(0,initial_file,file);
  2.1866 -+     error=cloop_set_file(0,initial_file);
  2.1867 -    if(error)
  2.1868 -     {
  2.1869 -      printk(KERN_ERR
  2.1870 -@@ -1052,9 +1301,6 @@
  2.1871 -             cloop_name, file, error);
  2.1872 -      goto init_out_dealloc;
  2.1873 -     }
  2.1874 --   if(namelen >= LO_NAME_SIZE) namelen = LO_NAME_SIZE-1;
  2.1875 --   memcpy(cloop_dev[0]->clo_file_name, file, namelen);
  2.1876 --   cloop_dev[0]->clo_file_name[namelen] = 0;
  2.1877 -   }
  2.1878 -  return 0;
  2.1879 - init_out_dealloc:
  2.1880 ---- cloop.h
  2.1881 -+++ cloop.h
  2.1882 -@@ -86,11 +86,8 @@
  2.1883 - struct cloop_tail
  2.1884 - {
  2.1885 - 	u_int32_t table_size; 
  2.1886 --	u_int32_t index_size; /* size:4 comp:3 ctrl-c:1 lastlen:24 */
  2.1887 -+	u_int32_t index_size; /* size:4 unused:3 ctrl-c:1 lastlen:24 */
  2.1888 - #define CLOOP3_INDEX_SIZE(x)    ((unsigned int)((x) & 0xF))
  2.1889 --#define CLOOP3_BLOCKS_FLAGS(x)  ((unsigned int)((x) & 0x70) >> 4)
  2.1890 --#define CLOOP3_TRUNCATED(x)     ((unsigned int)((x) & 0x80) >> 7)
  2.1891 --#define CLOOP3_LASTLEN(x)       (unsigned int)((x) >> 8)
  2.1892 - 	u_int32_t num_blocks;
  2.1893 - };
  2.1894 - 
  2.1895 -@@ -104,8 +101,10 @@
  2.1896 - };
  2.1897 - 
  2.1898 - static inline char *build_index(struct block_info *offsets, unsigned long n, 
  2.1899 --			unsigned long block_size, unsigned global_flags)
  2.1900 -+			unsigned long block_size)
  2.1901 - {
  2.1902 -+	static char v[11];
  2.1903 -+	u_int32_t flags = 0;
  2.1904 - 	u_int32_t *ofs32 = (u_int32_t *) offsets;
  2.1905 - 	loff_t    *ofs64 = (loff_t *) offsets;
  2.1906 - 
  2.1907 -@@ -130,8 +129,6 @@
  2.1908 - 		}
  2.1909 - 		else { /* V2.0/V4.0 */
  2.1910 - 			loff_t last = CLOOP_BLOCK_OFFSET(__be64_to_cpu(ofs64[n]));
  2.1911 --			u_int32_t flags;
  2.1912 --			static char v4[11];
  2.1913 - 			unsigned long i = n;
  2.1914 - 
  2.1915 - 			for (flags = 0; n-- ;) {
  2.1916 -@@ -149,12 +146,7 @@
  2.1917 - 					offsets[i] = offsets[offsets[i].offset];
  2.1918 - 				}
  2.1919 - 			}
  2.1920 --			strcpy(v4, (char *) "64BE v4.0a");
  2.1921 --			v4[10] = 'a' + ((flags-1) & 0xF);	// compressors used
  2.1922 --			if (flags > 0x10) {			// with links ?
  2.1923 --				v4[10] += 'A' - 'a';
  2.1924 --			}
  2.1925 --			return v4;
  2.1926 -+			strcpy(v, (char *) "64BE v4.0a");
  2.1927 - 		}
  2.1928 - 	}
  2.1929 - 	else if (ofs32[1] == 0 && v3_64 == 0) { /* V1.0 */
  2.1930 -@@ -170,7 +162,6 @@
  2.1931 - 	else { /* V3.0 or V0.68 */
  2.1932 - 		unsigned long i;
  2.1933 - 		loff_t j;
  2.1934 --		static char v3[11];
  2.1935 - 		
  2.1936 - 		for (i = 0; i < n && ntohl(ofs32[i]) < ntohl(ofs32[i+1]); i++);
  2.1937 - 		if (i == n && ntohl(ofs32[0]) == (4*n) + 0x8C) { /* V0.68 */
  2.1938 -@@ -185,28 +176,33 @@
  2.1939 - 		}
  2.1940 - 		
  2.1941 - 		v3_64 = (ofs32[1] == 0);
  2.1942 --		for (i = n; i-- != 0; )
  2.1943 -+		for (i = n; i-- != 0; ) {
  2.1944 - 			offsets[i].size = ntohl(ofs32[i << v3_64]); 
  2.1945 --		for (i = 0, j = sizeof(struct cloop_head); i < n; i++) {
  2.1946 --			offsets[i].offset = j;
  2.1947 --			offsets[i].flags = global_flags;
  2.1948 - 			if (offsets[i].size == 0xFFFFFFFF) {
  2.1949 --				offsets[i].flags = CLOOP_COMPRESSOR_NONE;
  2.1950 --				offsets[i].size = block_size;
  2.1951 -+				offsets[i].size = 0x10000000 | block_size;
  2.1952 - 			}
  2.1953 --			if ((offsets[i].size & 0x80000000) == 0) {
  2.1954 -+			offsets[i].flags = (offsets[i].size >> 28);
  2.1955 -+			offsets[i].size &= 0x0FFFFFFF; 
  2.1956 -+		}
  2.1957 -+		for (i = 0, j = sizeof(struct cloop_head); i < n; i++) {
  2.1958 -+			offsets[i].offset = j;
  2.1959 -+			if (offsets[i].flags < 8) {
  2.1960 - 				j += offsets[i].size;
  2.1961 - 			}
  2.1962 - 		}
  2.1963 - 		for (i = 0; i < n; i++) {
  2.1964 --			if (offsets[i].size & 0x80000000) {
  2.1965 --				offsets[i] = offsets[offsets[i].size & 0x7FFFFFFF];
  2.1966 -+			flags |= 1 << offsets[i].flags;
  2.1967 -+			if (offsets[i].flags >= 8) {
  2.1968 -+				offsets[i] = offsets[offsets[i].size];
  2.1969 - 			}
  2.1970 - 		}
  2.1971 --		strcpy(v3, (char *) (v3_64) ? "64BE v3.0a" : "32BE v3.0a");
  2.1972 --		v3[10] += global_flags;
  2.1973 --		return v3;
  2.1974 -+		strcpy(v, (char *) (v3_64) ? "64BE v3.0a" : "32BE v3.0a");
  2.1975 -+	}
  2.1976 -+	v[10] = 'a' + ((flags-1) & 0xF);	// compressors used
  2.1977 -+	if (flags > 0x10) {			// with links ?
  2.1978 -+		v[10] += 'A' - 'a';
  2.1979 - 	}
  2.1980 -+	return v;
  2.1981 - }
  2.1982 - 
  2.1983 - /* Cloop suspend IOCTL */
  2.1984 ---- cloop.c
  2.1985 -+++ cloop.c
  2.1986 -@@ -542,7 +542,7 @@
  2.1987 -  const unsigned int header_size = sizeof(struct cloop_head);
  2.1988 -  unsigned int i, total_offsets=0;
  2.1989 -  loff_t fs_read_position = 0, header_pos[2];
  2.1990 -- int flags, isblkdev, bytes_read, error = 0;
  2.1991 -+ int isblkdev, bytes_read, error = 0;
  2.1992 -  if (clo->suspended) return error;
  2.1993 -  #if LINUX_VERSION_CODE < KERNEL_VERSION(4, 0, 0)
  2.1994 -  inode = file->f_dentry->d_inode;
  2.1995 -@@ -698,18 +698,12 @@
  2.1996 -        error=-EBADF; goto error_release;
  2.1997 -       }
  2.1998 -      len = CLOOP3_INDEX_SIZE(ntohl(tail.index_size)) * total_offsets;
  2.1999 --     flags = CLOOP3_BLOCKS_FLAGS(ntohl(tail.index_size));
  2.2000 --// May  3 19:45:20 (none) user.info kernel: cloop: uncompress(clo=e0a78000, block_ptrs=e0c9c000, &len(1440)=ddc05e6c, zbuf=e0c9f000, zlen=43, flag=0)
  2.2001 --printk(KERN_INFO "%s: uncompress(clo=%p, block_ptrs=%p, &len(%ld)=%p, zbuf=%p, zlen=%ld, flag=%d)\n", cloop_name, 
  2.2002 --		clo, clo->block_ptrs, len, &len, zbuf, zlen, flags);
  2.2003 --     ret = uncompress(clo, (void *) clo->block_ptrs, &len, zbuf, zlen, flags);
  2.2004 --// May  3 19:45:20 (none) user.alert kernel: BUG: unable to handle kernel NULL pointer dereference at   (null)
  2.2005 --printk(KERN_INFO "%s: uncompressed !\n", cloop_name);
  2.2006 -+     ret = uncompress(clo, (void *) clo->block_ptrs, &len, zbuf, zlen, CLOOP_COMPRESSOR_ZLIB);
  2.2007 -      cloop_free(zbuf, zlen);
  2.2008 -      if (ret != 0)
  2.2009 -       {
  2.2010 --        printk(KERN_ERR "%s: decompression error %i uncompressing index, flags %u\n",
  2.2011 --               cloop_name, ret, flags);
  2.2012 -+        printk(KERN_ERR "%s: decompression error %i uncompressing index\n",
  2.2013 -+               cloop_name, ret);
  2.2014 -        error=-EBADF; goto error_release;
  2.2015 -       }
  2.2016 -     }
  2.2017 -@@ -722,7 +716,6 @@
  2.2018 -  else
  2.2019 -   {
  2.2020 -    unsigned int n, total_bytes;
  2.2021 --   flags = 0;
  2.2022 -    clo->block_ptrs = cloop_malloc(sizeof(struct block_info) * total_offsets);
  2.2023 -    if (!clo->block_ptrs)
  2.2024 -     {
  2.2025 -@@ -761,7 +754,7 @@
  2.2026 -   }
  2.2027 -  {
  2.2028 -   int i;
  2.2029 --  char *version = build_index(clo->block_ptrs, clo->head.num_blocks, clo->head.block_size, flags);
  2.2030 -+  char *version = build_index(clo->block_ptrs, clo->head.num_blocks, clo->head.block_size);
  2.2031 -   clo->largest_block = 0;
  2.2032 -   for (i = 0; i < clo->head.num_blocks; i++)
  2.2033 -     if (clo->block_ptrs[i].size > clo->largest_block)
  2.2034 -@@ -769,9 +762,6 @@
  2.2035 -   printk(KERN_INFO "%s: %s: %s: %u blocks, %u bytes/block, largest block is %lu bytes.\n",
  2.2036 -          cloop_name, clo->underlying_filename, version, clo->head.num_blocks,
  2.2037 -          clo->head.block_size, clo->largest_block);
  2.2038 -- }
  2.2039 -- {
  2.2040 --  int i;
  2.2041 -   clo->num_buffered_blocks = (buffers > 0 && clo->head.block_size >= 512) ?
  2.2042 -                               (buffers / clo->head.block_size) : 1;
  2.2043 -   clo->buffered_blocknum = cloop_malloc(clo->num_buffered_blocks * sizeof (u_int32_t));
  2.2044 -@@ -874,6 +864,10 @@
  2.2045 -  cloop_free(clo->block_ptrs, sizeof(struct block_info) * total_offsets);
  2.2046    clo->block_ptrs=NULL;
  2.2047   error_release:
  2.2048  +#if (defined(CONFIG_ZLIB_INFLATE) || defined(CONFIG_ZLIB_INFLATE_MODULE))
  2.2049 @@ -1548,3 +912,146 @@
  2.2050    if(bbuf) cloop_free(bbuf, clo->underlying_blksize);
  2.2051    if(clo->underlying_filename) { kfree(clo->underlying_filename); clo->underlying_filename=NULL; }
  2.2052    clo->backing_file=NULL;
  2.2053 +@@ -829,6 +900,7 @@
  2.2054 +  if(clo->refcnt > 1)	/* we needed one fd for the ioctl */
  2.2055 +    return -EBUSY;
  2.2056 +  if(filp==NULL) return -EINVAL;
  2.2057 ++ if(clo->clo_thread) { kthread_stop(clo->clo_thread); clo->clo_thread=NULL; }
  2.2058 +  if(filp!=initial_file)
  2.2059 +   fput(filp);
  2.2060 +  else
  2.2061 +@@ -839,7 +911,7 @@
  2.2062 +  clo->backing_file  = NULL;
  2.2063 +  clo->backing_inode = NULL;
  2.2064 +  if(clo->underlying_filename) { kfree(clo->underlying_filename); clo->underlying_filename=NULL; }
  2.2065 +- if(clo->block_ptrs) { cloop_free(clo->block_ptrs, clo->head.num_blocks+1); clo->block_ptrs = NULL; }
  2.2066 ++ if(clo->block_ptrs) { cloop_free(clo->block_ptrs, clo->head.num_blocks); clo->block_ptrs = NULL; }
  2.2067 +  if(clo->preload_cache)
  2.2068 +  {
  2.2069 +   int i;
  2.2070 +@@ -1054,15 +1126,15 @@
  2.2071 +   case LOOP_CLR_FD:       /* Change arg */ 
  2.2072 +   case LOOP_GET_STATUS64: /* Change arg */ 
  2.2073 +   case LOOP_SET_STATUS64: /* Change arg */ 
  2.2074 +-    return cloop_ioctl(bdev, mode, cmd, (unsigned long) compat_ptr(arg));
  2.2075 ++	arg = (unsigned long) compat_ptr(arg);
  2.2076 +   case LOOP_SET_STATUS:   /* unchanged */
  2.2077 +   case LOOP_GET_STATUS:   /* unchanged */
  2.2078 +   case LOOP_SET_FD:       /* unchanged */
  2.2079 +   case LOOP_CHANGE_FD:    /* unchanged */
  2.2080 +-    return cloop_ioctl(bdev, mode, cmd, arg);
  2.2081 +-  default:
  2.2082 +-    return -ENOIOCTLCMD;
  2.2083 ++	return cloop_ioctl(bdev, mode, cmd, arg);
  2.2084 ++	break;
  2.2085 +  }
  2.2086 ++ return -ENOIOCTLCMD;
  2.2087 + }
  2.2088 + #endif
  2.2089 + 
  2.2090 +@@ -1093,7 +1165,7 @@
  2.2091 +  cloop_dev[cloop_num]->refcnt-=1;
  2.2092 + }
  2.2093 + 
  2.2094 +-static const struct block_device_operations clo_fops =
  2.2095 ++static struct block_device_operations clo_fops =
  2.2096 + {
  2.2097 +         owner:		THIS_MODULE,
  2.2098 +         open:           cloop_open,
  2.2099 +@@ -1105,12 +1177,6 @@
  2.2100 + 	/* locked_ioctl ceased to exist in 2.6.36 */
  2.2101 + };
  2.2102 + 
  2.2103 +-static const struct blk_mq_ops cloop_mq_ops = {
  2.2104 +-	.queue_rq       = cloop_queue_rq,
  2.2105 +-/*	.init_request	= cloop_init_request, */
  2.2106 +-/*	.complete	= cloop_complete_rq, */
  2.2107 +-};
  2.2108 +-
  2.2109 + static int cloop_register_blkdev(int major_nr)
  2.2110 + {
  2.2111 +  return register_blkdev(major_nr, cloop_name);
  2.2112 +@@ -1124,37 +1190,33 @@
  2.2113 + 
  2.2114 + static int cloop_alloc(int cloop_num)
  2.2115 + {
  2.2116 +- struct cloop_device *clo = (struct cloop_device *) cloop_malloc(sizeof(struct cloop_device));
  2.2117 ++ struct cloop_device *clo = (struct cloop_device *) cloop_malloc(sizeof(struct cloop_device));;
  2.2118 +  if(clo == NULL) goto error_out;
  2.2119 +  cloop_dev[cloop_num] = clo;
  2.2120 +  memset(clo, 0, sizeof(struct cloop_device));
  2.2121 +  clo->clo_number = cloop_num;
  2.2122 +- clo->tag_set.ops = &cloop_mq_ops;
  2.2123 +- clo->tag_set.nr_hw_queues = 1;
  2.2124 +- clo->tag_set.queue_depth = 128;
  2.2125 +- clo->tag_set.numa_node = NUMA_NO_NODE;
  2.2126 +- clo->tag_set.cmd_size = 0; /* No extra data needed */
  2.2127 +- /* BLK_MQ_F_BLOCKING is extremely important if we want to call blocking functions like vfs_read */
  2.2128 +- clo->tag_set.flags = BLK_MQ_F_SHOULD_MERGE | BLK_MQ_F_BLOCKING;
  2.2129 +- clo->tag_set.driver_data = clo;
  2.2130 +- if(blk_mq_alloc_tag_set(&clo->tag_set)) goto error_out_free_clo;
  2.2131 +- clo->clo_queue = blk_mq_init_queue(&clo->tag_set);
  2.2132 +- if(IS_ERR(clo->clo_queue))
  2.2133 ++ clo->clo_thread = NULL;
  2.2134 ++ init_waitqueue_head(&clo->clo_event);
  2.2135 ++ spin_lock_init(&clo->queue_lock);
  2.2136 ++ mutex_init(&clo->clo_ctl_mutex);
  2.2137 ++ INIT_LIST_HEAD(&clo->clo_list);
  2.2138 ++ clo->clo_queue = blk_init_queue(cloop_do_request, &clo->queue_lock);
  2.2139 ++ if(!clo->clo_queue)
  2.2140 +   {
  2.2141 +    printk(KERN_ERR "%s: Unable to alloc queue[%d]\n", cloop_name, cloop_num);
  2.2142 +-   goto error_out_free_tags;
  2.2143 ++   goto error_out;
  2.2144 +   }
  2.2145 +  clo->clo_queue->queuedata = clo;
  2.2146 +- blk_queue_max_hw_sectors(clo->clo_queue, BLK_DEF_MAX_SECTORS);
  2.2147 ++#if LINUX_VERSION_CODE < KERNEL_VERSION(4, 12, 0)
  2.2148 ++ queue_flag_set_unlocked(QUEUE_FLAG_NONROT, clo->clo_queue);
  2.2149 ++ queue_flag_set_unlocked(QUEUE_FLAG_NOMERGES, clo->clo_queue);
  2.2150 ++#endif
  2.2151 +  clo->clo_disk = alloc_disk(1);
  2.2152 +  if(!clo->clo_disk)
  2.2153 +   {
  2.2154 +    printk(KERN_ERR "%s: Unable to alloc disk[%d]\n", cloop_name, cloop_num);
  2.2155 +-   goto error_out_free_queue;
  2.2156 ++   goto error_disk;
  2.2157 +   }
  2.2158 +- spin_lock_init(&clo->queue_lock);
  2.2159 +- mutex_init(&clo->clo_ctl_mutex);
  2.2160 +- mutex_init(&clo->clo_rq_mutex);
  2.2161 +  clo->clo_disk->major = cloop_major;
  2.2162 +  clo->clo_disk->first_minor = cloop_num;
  2.2163 +  clo->clo_disk->fops = &clo_fops;
  2.2164 +@@ -1163,12 +1225,8 @@
  2.2165 +  sprintf(clo->clo_disk->disk_name, "%s%d", cloop_name, cloop_num);
  2.2166 +  add_disk(clo->clo_disk);
  2.2167 +  return 0;
  2.2168 +-error_out_free_queue:
  2.2169 ++error_disk:
  2.2170 +  blk_cleanup_queue(clo->clo_queue);
  2.2171 +-error_out_free_tags:
  2.2172 +- blk_mq_free_tag_set(&clo->tag_set);
  2.2173 +-error_out_free_clo:
  2.2174 +- cloop_free(clo, sizeof(struct cloop_device));
  2.2175 + error_out:
  2.2176 +  return -ENOMEM;
  2.2177 + }
  2.2178 +@@ -1179,7 +1237,6 @@
  2.2179 +  if(clo == NULL) return;
  2.2180 +  del_gendisk(clo->clo_disk);
  2.2181 +  blk_cleanup_queue(clo->clo_queue);
  2.2182 +- blk_mq_free_tag_set(&clo->tag_set);
  2.2183 +  put_disk(clo->clo_disk);
  2.2184 +  cloop_free(clo, sizeof(struct cloop_device));
  2.2185 +  cloop_dev[cloop_num] = NULL;
  2.2186 +--- cloop_suspend.c
  2.2187 ++++ cloop_suspend.c
  2.2188 +@@ -14,6 +14,7 @@
  2.2189 + #include <fcntl.h>
  2.2190 + #include <unistd.h>
  2.2191 + #include <stdio.h>
  2.2192 ++#include <stdint.h>
  2.2193 + 
  2.2194 + /* We don't use the structure, so that define does not hurt */
  2.2195 + #define dev_t int
     3.1 --- a/linux64-cloop/receipt	Sun May 08 13:06:36 2022 +0000
     3.2 +++ b/linux64-cloop/receipt	Sun May 08 16:45:21 2022 +0000
     3.3 @@ -2,16 +2,15 @@
     3.4  
     3.5  PACKAGE="linux64-cloop"
     3.6  SOURCE="cloop"
     3.7 -_VERSION="2.639-2"
     3.8 -#VERSION="$(sed '/+#define CLOOP_VERSION/!d;s|.* "\(.*\)"|\1|' stuff/cloop.u)"
     3.9 +_VERSION="3.14.1.3"
    3.10  VERSION="4.12"
    3.11  CATEGORY="base-system"
    3.12  MAINTAINER="pascal.bellard@slitaz.org"
    3.13  LICENSE="GPL2"
    3.14  SHORT_DESC="The read-only compressed loop device kernel module."
    3.15  WEB_SITE="http://knoppix.net/wiki/Cloop"
    3.16 -TARBALL="${SOURCE}_${_VERSION}.tar.gz"
    3.17 -WGET_URL="http://debian-knoppix.alioth.debian.org/packages/$SOURCE/$TARBALL"
    3.18 +TARBALL="${SOURCE}_${_VERSION}.tar.xz"
    3.19 +WGET_URL="http://deb.debian.org/debian/pool/main/c/$SOURCE/$TARBALL"
    3.20  PROVIDE="linux-cloop:linux64"
    3.21  
    3.22  DEPENDS="linux64"
    3.23 @@ -26,6 +25,7 @@
    3.24  {
    3.25  	patch -p0 < $stuff/cloop.u
    3.26  	make ARCH=x86_64 KERNEL_DIR="/usr/src/linux" cloop.ko && xz cloop.ko
    3.27 +	make cloop_suspend
    3.28  }
    3.29  	
    3.30  # Rules to gen a SliTaz package suitable for Tazpkg.
     4.1 --- a/linux64-cloop/stuff/cloop.u	Sun May 08 13:06:36 2022 +0000
     4.2 +++ b/linux64-cloop/stuff/cloop.u	Sun May 08 16:45:21 2022 +0000
     4.3 @@ -1,6 +1,6 @@
     4.4  --- cloop.h
     4.5  +++ cloop.h
     4.6 -@@ -1,15 +1,50 @@
     4.7 +@@ -1,3 +1,7 @@
     4.8  +#define CLOOP_SIGNATURE "#!/bin/sh"                      /* @ offset 0  */
     4.9  +#define CLOOP_SIGNATURE_SIZE 9
    4.10  +#define CLOOP_SIGNATURE_OFFSET 0x0
    4.11 @@ -8,99 +8,46 @@
    4.12   #ifndef _COMPRESSED_LOOP_H
    4.13   #define _COMPRESSED_LOOP_H
    4.14   
    4.15 --#define CLOOP_HEADROOM 128
    4.16 -+/*************************************************************************\
    4.17 -+* Starting with Format V4.0 (cloop version 4.x), cloop can now have two   *
    4.18 -+* alternative structures:                                                 *
    4.19 -+*                                                                         *
    4.20 -+* 1. Header first: "robust" format, handles missing blocks well           *
    4.21 -+* 2. Footer (header last): "streaming" format, easier to create           *
    4.22 -+*                                                                         *
    4.23 -+* The cloop kernel module autodetects both formats, and can (currently)   *
    4.24 -+* still handle the V2.0 format as well.                                   *
    4.25 -+*                                                                         *
    4.26 -+* 1. Header first:                                                        *
    4.27 -+*   +---------------------------- FIXED SIZE ---------------------------+ *
    4.28 -+*   |Signature (128 bytes)                                              | *
    4.29 -+*   |block_size (32bit number, network order)                           | *
    4.30 -+*   |num_blocks (32bit number, network order)                           | *
    4.31 -+*   +--------------------------- VARIABLE SIZE -------------------------+ *
    4.32 -+*   |num_blocks * FlagsOffset (upper 4 bits flags, lower 64 bits offset)| *
    4.33 -+*   |compressed data blocks of variable size ...                        | *
    4.34 -+*   +-------------------------------------------------------------------+ *
    4.35 -+*                                                                         *
    4.36 -+* 2. Footer (header last):                                                *
    4.37 -+*   +--------------------------- VARIABLE SIZE -------------------------+ *
    4.38 -+*   |compressed data blocks of variable size ...                        | *
    4.39 -+*   |num_blocks * FlagsOffset (upper 4 bits flags, lower 64 bits offset)| *
    4.40 -+*   +---------------------------- FIXED SIZE ---------------------------+ *
    4.41 -+*   |Signature (128 bytes)                                              | *
    4.42 -+*   |block_size (32bit number, network order)                           | *
    4.43 -+*   |num_blocks (32bit number, network order)                           | *
    4.44 -+*   +-------------------------------------------------------------------+ *
    4.45 -+*                                                                         *
    4.46 -+* Offsets are always relative to beginning of file, in all formats.       *
    4.47 -+* The block index contains num_blocks+1 offsets, followed (1) or          *
    4.48 -+* preceded (2) by the compressed blocks.                                  *
    4.49 -+\*************************************************************************/
    4.50 +@@ -38,10 +42,6 @@
    4.51   
    4.52 --/* The cloop header usually looks like this:          */
    4.53 --/* #!/bin/sh                                          */
    4.54 --/* #V2.00 Format                                      */
    4.55 --/* ...padding up to CLOOP_HEADROOM...                 */
    4.56 --/* block_size (32bit number, network order)           */
    4.57 --/* num_blocks (32bit number, network order)           */
    4.58 -+#include <linux/types.h>   /* u_int32_t */
    4.59 -+
    4.60 -+#define CLOOP_HEADROOM 128
    4.61 + #include <linux/types.h>   /* u_int32_t */
    4.62   
    4.63 -+/* Header of fixed length, can be located at beginning or end of file   */
    4.64 - struct cloop_head
    4.65 - {
    4.66 - 	char preamble[CLOOP_HEADROOM];
    4.67 -@@ -17,9 +52,163 @@
    4.68 +-#ifndef __KERNEL__
    4.69 +-#include <stdint.h> /* regular uint64_t */
    4.70 +-#endif
    4.71 +-
    4.72 + #define CLOOP_HEADROOM 128
    4.73 + 
    4.74 + /* Header of fixed length, can be located at beginning or end of file   */
    4.75 +@@ -52,13 +52,6 @@
    4.76   	u_int32_t num_blocks;
    4.77   };
    4.78   
    4.79 -+/************************************************************************\
    4.80 -+*  CLOOP4 flags for each compressed block                                *
    4.81 -+*  Value   Meaning                                                       *
    4.82 -+*    0     GZIP/7ZIP compression (compatible with V2.0 Format)           *
    4.83 -+*    1     no compression (incompressible data)                          *
    4.84 -+*    2     xz compression (currently best space saver)                   *
    4.85 -+*    3     lz4 compression                                               *
    4.86 -+*    4     lzo compression (fastest)                                     *
    4.87 -+\************************************************************************/
    4.88 -+
    4.89 -+typedef uint64_t cloop_block_ptr;
    4.90 -+
    4.91 -+/* Get value of first 4 bits */
    4.92 -+#define CLOOP_BLOCK_FLAGS(x)  ((unsigned int)(((x) & 0xf000000000000000LLU) >> 60))
    4.93 -+/* Get value of last 60 bits */
    4.94 -+#define CLOOP_BLOCK_OFFSET(x)  ((x) & 0x0fffffffffffffffLLU)
    4.95 -+
    4.96 -+#define CLOOP_COMPRESSOR_ZLIB  0x0
    4.97 -+#define CLOOP_COMPRESSOR_NONE  0x1
    4.98 -+#define CLOOP_COMPRESSOR_XZ    0x2
    4.99 -+#define CLOOP_COMPRESSOR_LZ4   0x3
   4.100 -+#define CLOOP_COMPRESSOR_LZO1X 0x4
   4.101 -+
   4.102 -+#define CLOOP_COMPRESSOR_VALID(x) ((x) >= CLOOP_COMPRESSOR_ZLIB && (x) <= CLOOP_COMPRESSOR_LZO1X)
   4.103 -+
   4.104 +-#define CLOOP2_SIGNATURE "V2.0"                       /* @ offset 0x0b  */
   4.105 +-#define CLOOP2_SIGNATURE_SIZE 4
   4.106 +-#define CLOOP2_SIGNATURE_OFFSET 0x0b
   4.107 +-#define CLOOP4_SIGNATURE "V4.0"                       /* @ offset 0x0b  */
   4.108 +-#define CLOOP4_SIGNATURE_SIZE 4
   4.109 +-#define CLOOP4_SIGNATURE_OFFSET 0x0b
   4.110 +-
   4.111 + /************************************************************************\
   4.112 + *  CLOOP4 flags for each compressed block                                *
   4.113 + *  Value   Meaning                                                       *
   4.114 +@@ -84,6 +77,134 @@
   4.115 + 
   4.116 + #define CLOOP_COMPRESSOR_VALID(x) ((x) >= CLOOP_COMPRESSOR_ZLIB && (x) <= CLOOP_COMPRESSOR_LZO1X)
   4.117 + 
   4.118  +#define CLOOP_COMPRESSOR_LINK  0xF
   4.119  +
   4.120  +
   4.121 - /* data_index (num_blocks 64bit pointers, network order)...      */
   4.122 - /* compressed data (gzip block compressed format)...             */
   4.123 - 
   4.124 ++/* data_index (num_blocks 64bit pointers, network order)...      */
   4.125 ++/* compressed data (gzip block compressed format)...             */
   4.126 ++
   4.127  +struct cloop_tail
   4.128  +{
   4.129  +	u_int32_t table_size; 
   4.130 -+	u_int32_t index_size; /* size:4 comp:3 ctrl-c:1 lastlen:24 */
   4.131 ++	u_int32_t index_size; /* size:4 unused:3 ctrl-c:1 lastlen:24 */
   4.132  +#define CLOOP3_INDEX_SIZE(x)    ((unsigned int)((x) & 0xF))
   4.133 -+#define CLOOP3_BLOCKS_FLAGS(x)  ((unsigned int)((x) & 0x70) >> 4)
   4.134 -+#define CLOOP3_TRUNCATED(x)     ((unsigned int)((x) & 0x80) >> 7)
   4.135 -+#define CLOOP3_LASTLEN(x)       (unsigned int)((x) >> 8)
   4.136  +	u_int32_t num_blocks;
   4.137  +};
   4.138  +
   4.139 @@ -114,8 +61,10 @@
   4.140  +};
   4.141  +
   4.142  +static inline char *build_index(struct block_info *offsets, unsigned long n, 
   4.143 -+			unsigned long block_size, unsigned global_flags)
   4.144 ++			unsigned long block_size)
   4.145  +{
   4.146 ++	static char v[11];
   4.147 ++	u_int32_t flags = 0;
   4.148  +	u_int32_t *ofs32 = (u_int32_t *) offsets;
   4.149  +	loff_t    *ofs64 = (loff_t *) offsets;
   4.150  +
   4.151 @@ -140,8 +89,6 @@
   4.152  +		}
   4.153  +		else { /* V2.0/V4.0 */
   4.154  +			loff_t last = CLOOP_BLOCK_OFFSET(__be64_to_cpu(ofs64[n]));
   4.155 -+			u_int32_t flags;
   4.156 -+			static char v4[11];
   4.157  +			unsigned long i = n;
   4.158  +
   4.159  +			for (flags = 0; n-- ;) {
   4.160 @@ -159,12 +106,7 @@
   4.161  +					offsets[i] = offsets[offsets[i].offset];
   4.162  +				}
   4.163  +			}
   4.164 -+			strcpy(v4, (char *) "64BE v4.0a");
   4.165 -+			v4[10] = 'a' + ((flags-1) & 0xF);	// compressors used
   4.166 -+			if (flags > 0x10) {			// with links ?
   4.167 -+				v4[10] += 'A' - 'a';
   4.168 -+			}
   4.169 -+			return v4;
   4.170 ++			strcpy(v, (char *) "64BE v4.0a");
   4.171  +		}
   4.172  +	}
   4.173  +	else if (ofs32[1] == 0 && v3_64 == 0) { /* V1.0 */
   4.174 @@ -180,7 +122,6 @@
   4.175  +	else { /* V3.0 or V0.68 */
   4.176  +		unsigned long i;
   4.177  +		loff_t j;
   4.178 -+		static char v3[11];
   4.179  +		
   4.180  +		for (i = 0; i < n && ntohl(ofs32[i]) < ntohl(ofs32[i+1]); i++);
   4.181  +		if (i == n && ntohl(ofs32[0]) == (4*n) + 0x8C) { /* V0.68 */
   4.182 @@ -195,28 +136,33 @@
   4.183  +		}
   4.184  +		
   4.185  +		v3_64 = (ofs32[1] == 0);
   4.186 -+		for (i = n; i-- != 0; )
   4.187 ++		for (i = n; i-- != 0; ) {
   4.188  +			offsets[i].size = ntohl(ofs32[i << v3_64]); 
   4.189 ++			if (offsets[i].size == 0xFFFFFFFF) {
   4.190 ++				offsets[i].size = 0x10000000 | block_size;
   4.191 ++			}
   4.192 ++			offsets[i].flags = (offsets[i].size >> 28);
   4.193 ++			offsets[i].size &= 0x0FFFFFFF; 
   4.194 ++		}
   4.195  +		for (i = 0, j = sizeof(struct cloop_head); i < n; i++) {
   4.196  +			offsets[i].offset = j;
   4.197 -+			offsets[i].flags = global_flags;
   4.198 -+			if (offsets[i].size == 0xFFFFFFFF) {
   4.199 -+				offsets[i].flags = CLOOP_COMPRESSOR_NONE;
   4.200 -+				offsets[i].size = block_size;
   4.201 -+			}
   4.202 -+			if ((offsets[i].size & 0x80000000) == 0) {
   4.203 ++			if (offsets[i].flags < 8) {
   4.204  +				j += offsets[i].size;
   4.205  +			}
   4.206  +		}
   4.207  +		for (i = 0; i < n; i++) {
   4.208 -+			if (offsets[i].size & 0x80000000) {
   4.209 -+				offsets[i] = offsets[offsets[i].size & 0x7FFFFFFF];
   4.210 ++			flags |= 1 << offsets[i].flags;
   4.211 ++			if (offsets[i].flags >= 8) {
   4.212 ++				offsets[i] = offsets[offsets[i].size];
   4.213  +			}
   4.214  +		}
   4.215 -+		strcpy(v3, (char *) (v3_64) ? "64BE v3.0a" : "32BE v3.0a");
   4.216 -+		v3[10] += global_flags;
   4.217 -+		return v3;
   4.218 ++		strcpy(v, (char *) (v3_64) ? "64BE v3.0a" : "32BE v3.0a");
   4.219  +	}
   4.220 ++	v[10] = 'a' + ((flags-1) & 0xF);	// compressors used
   4.221 ++	if (flags > 0x10) {			// with links ?
   4.222 ++		v[10] += 'A' - 'a';
   4.223 ++	}
   4.224 ++	return v;
   4.225  +}
   4.226  +
   4.227   /* Cloop suspend IOCTL */
   4.228 @@ -224,661 +170,538 @@
   4.229   
   4.230  --- cloop.c
   4.231  +++ cloop.c
   4.232 -@@ -1,26 +1,23 @@
   4.233 --/*
   4.234 -- *  compressed_loop.c: Read-only compressed loop blockdevice
   4.235 -- *  hacked up by Rusty in 1999, extended and maintained by Klaus Knopper
   4.236 -- *
   4.237 -- *  A cloop file looks like this:
   4.238 -- *  [32-bit uncompressed block size: network order]
   4.239 -- *  [32-bit number of blocks (n_blocks): network order]
   4.240 -- *  [64-bit file offsets of start of blocks: network order]
   4.241 -- *    ...
   4.242 -- *    (n_blocks + 1).
   4.243 -- * n_blocks consisting of:
   4.244 -- *   [compressed block]
   4.245 -- *
   4.246 -- * Every version greatly inspired by code seen in loop.c
   4.247 -- * by Theodore Ts'o, 3/29/93.
   4.248 -- *
   4.249 -- * Copyright 1999-2009 by Paul `Rusty' Russell & Klaus Knopper.
   4.250 -- * Redistribution of this file is permitted under the GNU Public License.
   4.251 -- *
   4.252 -- */
   4.253 -+/************************************************************************\
   4.254 -+* cloop.c: Read-only compressed loop blockdevice                         *
   4.255 -+* hacked up by Rusty in 1999, extended and maintained by Klaus Knopper   *
   4.256 -+*                                                                        *
   4.257 -+* For all supported cloop file formats, please check the file "cloop.h"  *
   4.258 -+* New in Version 4:                                                      *
   4.259 -+* - Header can be first or last in cloop file,                           *
   4.260 -+* - Different compression algorithms supported (compression type         *
   4.261 -+*   encoded in first 4 bytes of block offset address)                    *
   4.262 -+*                                                                        *
   4.263 -+* Every version greatly inspired by code seen in loop.c                  *
   4.264 -+* by Theodore Ts'o, 3/29/93.                                             *
   4.265 -+*                                                                        *
   4.266 -+* Copyright 1999-2009 by Paul `Rusty' Russell & Klaus Knopper.           *
   4.267 -+* Redistribution of this file is permitted under the GNU Public License  *
   4.268 -+* V2.                                                                    *
   4.269 -+\************************************************************************/
   4.270 +@@ -17,7 +17,7 @@
   4.271 + \************************************************************************/
   4.272   
   4.273   #define CLOOP_NAME "cloop"
   4.274 --#define CLOOP_VERSION "2.639"
   4.275 +-#define CLOOP_VERSION "5.3"
   4.276  +#define CLOOP_VERSION "4.12"
   4.277   #define CLOOP_MAX 8
   4.278   
   4.279   #ifndef KBUILD_MODNAME
   4.280 -@@ -47,8 +44,27 @@
   4.281 - #include <asm/div64.h> /* do_div() for 64bit division */
   4.282 - #include <asm/uaccess.h>
   4.283 - #include <asm/byteorder.h>
   4.284 --/* Use zlib_inflate from lib/zlib_inflate */
   4.285 -+/* Check for ZLIB, LZO1X, LZ4 decompression algorithms in kernel. */
   4.286 -+#if (defined(CONFIG_ZLIB_INFLATE) || defined(CONFIG_ZLIB_INFLATE_MODULE))
   4.287 - #include <linux/zutil.h>
   4.288 -+#endif
   4.289 -+#if (defined(CONFIG_LZO_DECOMPRESS) || defined(CONFIG_LZO_DECOMPRESS_MODULE))
   4.290 -+#include <linux/lzo.h>
   4.291 -+#endif
   4.292 -+#if (defined(CONFIG_DECOMPRESS_LZ4) || defined(CONFIG_DECOMPRESS_LZ4_MODULE))
   4.293 -+#include <linux/lz4.h>
   4.294 -+#endif
   4.295 -+#if (defined(CONFIG_DECOMPRESS_LZMA) || defined(CONFIG_DECOMPRESS_LZMA_MODULE))
   4.296 -+#include <linux/decompress/unlzma.h>
   4.297 -+#endif
   4.298 -+#if (defined(CONFIG_DECOMPRESS_XZ) || defined(CONFIG_DECOMPRESS_XZ_MODULE))
   4.299 -+#include <linux/xz.h>
   4.300 -+#endif
   4.301 -+
   4.302 -+#if (!(defined(CONFIG_ZLIB_INFLATE) || defined(CONFIG_ZLIB_INFLATE_MODULE) || defined(CONFIG_LZO_DECOMPRESS) || defined(CONFIG_LZO_DECOMPRESS_MODULE) || defined(CONFIG_DECOMPRESS_LZ4) || defined(CONFIG_DECOMPRESS_LZ4_MODULE) || defined(CONFIG_DECOMPRESS_LZMA) || defined(CONFIG_DECOMPRESS_LZMA_MODULE) || defined(CONFIG_DECOMPRESS_XZ) || defined(CONFIG_DECOMPRESS_XZ_MODULE)))
   4.303 -+#error "No decompression library selected in kernel config!"
   4.304 -+#endif
   4.305 -+
   4.306 +@@ -68,7 +68,6 @@
   4.307   #include <linux/loop.h>
   4.308   #include <linux/kthread.h>
   4.309   #include <linux/compat.h>
   4.310 -@@ -92,47 +108,64 @@
   4.311 - #define DEBUGP(format, x...)
   4.312 - #endif
   4.313 +-#include <linux/blk-mq.h> /* new multiqueue infrastructure */
   4.314 + #include "cloop.h"
   4.315   
   4.316 -+/* Default size of buffer to keep some decompressed blocks in memory to speed up access */
   4.317 -+#define BLOCK_BUFFER_MEM (16*65536)
   4.318 -+
   4.319 - /* One file can be opened at module insertion time */
   4.320 - /* insmod cloop file=/path/to/file */
   4.321 - static char *file=NULL;
   4.322 - static unsigned int preload=0;
   4.323 - static unsigned int cloop_max=CLOOP_MAX;
   4.324 -+static unsigned int buffers=BLOCK_BUFFER_MEM;
   4.325 - module_param(file, charp, 0);
   4.326 - module_param(preload, uint, 0);
   4.327 - module_param(cloop_max, uint, 0);
   4.328 - MODULE_PARM_DESC(file, "Initial cloop image file (full path) for /dev/cloop");
   4.329 - MODULE_PARM_DESC(preload, "Preload n blocks of cloop data into memory");
   4.330 - MODULE_PARM_DESC(cloop_max, "Maximum number of cloop devices (default 8)");
   4.331 -+MODULE_PARM_DESC(buffers, "Size of buffer to keep uncompressed blocks in memory in MiB (default 1)");
   4.332 + /* New License scheme */
   4.333 +@@ -93,10 +92,7 @@
   4.334 + /* Use experimental major for now */
   4.335 + #define MAJOR_NR 240
   4.336   
   4.337 - static struct file *initial_file=NULL;
   4.338 - static int cloop_major=MAJOR_NR;
   4.339 +-#ifndef DEVICE_NAME
   4.340 +-#define DEVICE_NAME CLOOP_NAME
   4.341 +-#endif
   4.342 +-
   4.343 ++/* #define DEVICE_NAME CLOOP_NAME */
   4.344 + /* #define DEVICE_NR(device) (MINOR(device)) */
   4.345 + /* #define DEVICE_ON(device) */
   4.346 + /* #define DEVICE_OFF(device) */
   4.347 +@@ -143,7 +139,7 @@
   4.348 +  u_int32_t allflags;
   4.349   
   4.350 --/* Number of buffered decompressed blocks */
   4.351 --#define BUFFERED_BLOCKS 8
   4.352 - struct cloop_device
   4.353 - {
   4.354 -- /* Copied straight from the file */
   4.355 -+ /* Header filled from the file */
   4.356 -  struct cloop_head head;
   4.357 -+ int header_first;
   4.358 -+ int file_format;
   4.359 - 
   4.360 -- /* An array of offsets of compressed blocks within the file */
   4.361 -- loff_t *offsets;
   4.362 -+ /* An or'd sum of all flags of each compressed block (v3) */
   4.363 -+ u_int32_t allflags;
   4.364 -+
   4.365 -+ /* An array of cloop_ptr flags/offset for compressed blocks within the file */
   4.366 +  /* An array of cloop_ptr flags/offset for compressed blocks within the file */
   4.367 +- cloop_block_ptr *block_ptrs;
   4.368  + struct block_info *block_ptrs;
   4.369   
   4.370    /* We buffer some uncompressed blocks for performance */
   4.371 -- int buffered_blocknum[BUFFERED_BLOCKS];
   4.372 -- int current_bufnum;
   4.373 -- void *buffer[BUFFERED_BLOCKS];
   4.374 -- void *compressed_buffer;
   4.375 -- size_t preload_array_size; /* Size of pointer array in blocks */
   4.376 -- size_t preload_size;       /* Number of successfully allocated blocks */
   4.377 -- char **preload_cache;      /* Pointers to preloaded blocks */
   4.378 -+ size_t num_buffered_blocks;	/* how many uncompressed blocks buffered for performance */
   4.379 -+ int *buffered_blocknum;        /* list of numbers of uncompressed blocks in buffer */
   4.380 -+ int current_bufnum;            /* which block is current */
   4.381 -+ unsigned char **buffer;        /* cache space for num_buffered_blocks uncompressed blocks */
   4.382 -+ void *compressed_buffer;       /* space for the largest compressed block */
   4.383 -+ size_t preload_array_size;     /* Size of pointer array in blocks */
   4.384 -+ size_t preload_size;           /* Number of successfully allocated blocks */
   4.385 -+ char **preload_cache;          /* Pointers to preloaded blocks */
   4.386 - 
   4.387 -+#if (defined(CONFIG_ZLIB_INFLATE) || defined(CONFIG_ZLIB_INFLATE_MODULE))
   4.388 -  z_stream zstream;
   4.389 -+#endif
   4.390 -+#if (defined(CONFIG_DECOMPRESS_XZ) || defined(CONFIG_DECOMPRESS_XZ_MODULE))
   4.391 -+ struct xz_dec *xzdecoderstate;
   4.392 -+ struct xz_buf xz_buffer;
   4.393 -+#endif
   4.394 - 
   4.395 -  struct file   *backing_file;  /* associated file */
   4.396 -  struct inode  *backing_inode; /* for bmap */
   4.397 - 
   4.398 -+ unsigned char *underlying_filename;
   4.399 -  unsigned long largest_block;
   4.400 -  unsigned int underlying_blksize;
   4.401 -+ loff_t underlying_total_size;
   4.402 -  int clo_number;
   4.403 -  int refcnt;
   4.404 -  struct block_device *bdev;
   4.405 -@@ -147,7 +180,6 @@
   4.406 +  size_t num_buffered_blocks;	/* how many uncompressed blocks buffered for performance */
   4.407 +@@ -178,14 +174,16 @@
   4.408 +  spinlock_t queue_lock;
   4.409 +  /* mutex for ioctl() */
   4.410 +  struct mutex clo_ctl_mutex;
   4.411 +- /* mutex for request */
   4.412 +- struct mutex clo_rq_mutex;
   4.413 ++ struct list_head clo_list;
   4.414 ++ struct task_struct *clo_thread;
   4.415 ++ wait_queue_head_t clo_event;
   4.416    struct request_queue *clo_queue;
   4.417    struct gendisk *clo_disk;
   4.418 +- struct blk_mq_tag_set tag_set;
   4.419    int suspended;
   4.420 -- char clo_file_name[LO_NAME_SIZE];
   4.421   };
   4.422   
   4.423 - /* Changed in 2.639: cloop_dev is now a an array of cloop_dev pointers,
   4.424 -@@ -156,52 +188,113 @@
   4.425 ++/* Changed in 2.639: cloop_dev is now a an array of cloop_dev pointers,
   4.426 ++   so we can specify how many devices we need via parameters. */
   4.427 + static struct cloop_device **cloop_dev;
   4.428   static const char *cloop_name=CLOOP_NAME;
   4.429   static int cloop_count = 0;
   4.430 - 
   4.431 --#if (!(defined(CONFIG_ZLIB_INFLATE) || defined(CONFIG_ZLIB_INFLATE_MODULE))) /* Must be compiled into kernel. */
   4.432 --#error  "Invalid Kernel configuration. CONFIG_ZLIB_INFLATE support is needed for cloop."
   4.433 --#endif
   4.434 --
   4.435 --/* Use __get_free_pages instead of vmalloc, allows up to 32 pages,
   4.436 -- * 2MB in one piece */
   4.437 - static void *cloop_malloc(size_t size)
   4.438 - {
   4.439 -- int order = get_order(size);
   4.440 -- if(order <= KMALLOC_MAX_ORDER)
   4.441 --   return (void *)kmalloc(size, GFP_KERNEL);
   4.442 -- else if(order < MAX_ORDER)
   4.443 --   return (void *)__get_free_pages(GFP_KERNEL, order);
   4.444 -+ /* kmalloc will fail after the system is running for a while, */
   4.445 -+ /* when large orders can't return contiguous memory. */
   4.446 -+ /* Let's just use vmalloc for now. :-/ */
   4.447 -+ /* int order = get_order(size); */
   4.448 -+ /* if(order <= KMALLOC_MAX_ORDER) */
   4.449 -+ /*  return (void *)kmalloc(size, GFP_KERNEL); */
   4.450 -+ /* else if(order < MAX_ORDER) */
   4.451 -+ /*  return (void *)__get_free_pages(GFP_KERNEL, order); */
   4.452 -  return (void *)vmalloc(size);
   4.453 +@@ -214,24 +212,21 @@
   4.454 +  vfree(mem);
   4.455   }
   4.456   
   4.457 - static void cloop_free(void *mem, size_t size)
   4.458 - {
   4.459 -- int order = get_order(size);
   4.460 -- if(order <= KMALLOC_MAX_ORDER)
   4.461 --   kfree(mem);
   4.462 -- else if(order < MAX_ORDER)
   4.463 --   free_pages((unsigned long)mem, order);
   4.464 -- else vfree(mem);
   4.465 -+ /* int order = get_order(size); */
   4.466 -+ /* if(order <= KMALLOC_MAX_ORDER) */
   4.467 -+ /*  kfree(mem); */
   4.468 -+ /* else if(order < MAX_ORDER) */
   4.469 -+ /*  free_pages((unsigned long)mem, order); */
   4.470 -+ /* else */
   4.471 -+ vfree(mem);
   4.472 - }
   4.473 - 
   4.474 --static int uncompress(struct cloop_device *clo,
   4.475 --                      unsigned char *dest, unsigned long *destLen,
   4.476 --                      unsigned char *source, unsigned long sourceLen)
   4.477 +-/* static int uncompress(struct cloop_device *clo, unsigned char *dest, unsigned long *destLen, unsigned char *source, unsigned long sourceLen) */
   4.478 +-static int uncompress(struct cloop_device *clo, u_int32_t block_num, u_int32_t compressed_length, unsigned long *uncompressed_length)
   4.479  +static int uncompress(struct cloop_device *clo, unsigned char *dest, unsigned long *destLen, unsigned char *source, unsigned long sourceLen, int flags) 
   4.480   {
   4.481 -- /* Most of this code can be found in fs/cramfs/uncompress.c */
   4.482 -- int err;
   4.483 -- clo->zstream.next_in = source;
   4.484 -- clo->zstream.avail_in = sourceLen;
   4.485 -- clo->zstream.next_out = dest;
   4.486 -- clo->zstream.avail_out = *destLen;
   4.487 -- err = zlib_inflateReset(&clo->zstream);
   4.488 -- if (err != Z_OK)
   4.489 --  {
   4.490 --   printk(KERN_ERR "%s: zlib_inflateReset error %d\n", cloop_name, err);
   4.491 --   zlib_inflateEnd(&clo->zstream); zlib_inflateInit(&clo->zstream);
   4.492 --  }
   4.493 -- err = zlib_inflate(&clo->zstream, Z_FINISH);
   4.494 -- *destLen = clo->zstream.total_out;
   4.495 -- if (err != Z_STREAM_END) return err;
   4.496 -- return Z_OK;
   4.497 -+ int err = -1;
   4.498 -+ switch(flags)
   4.499 -+ {
   4.500 -+  case CLOOP_COMPRESSOR_NONE:
   4.501 +  int err = -1;
   4.502 +- int flags = CLOOP_BLOCK_FLAGS(clo->block_ptrs[block_num]);
   4.503 +  switch(flags)
   4.504 +  {
   4.505 +   case CLOOP_COMPRESSOR_NONE:
   4.506 +-   /* block is umcompressed, swap pointers only! */
   4.507 +-   { char *tmp = clo->compressed_buffer; clo->compressed_buffer = clo->buffer[clo->current_bufnum]; clo->buffer[clo->current_bufnum] = tmp; }
   4.508 +-   DEBUGP("cloop: block %d is uncompressed (flags=%d), just swapping %u bytes\n", block_num, flags, compressed_length);
   4.509  +   memcpy(dest, source, *destLen = sourceLen);
   4.510  +   err = Z_OK;
   4.511 -+   break;
   4.512 -+#if (defined(CONFIG_ZLIB_INFLATE) || defined(CONFIG_ZLIB_INFLATE_MODULE))
   4.513 -+  case CLOOP_COMPRESSOR_ZLIB:
   4.514 +    break;
   4.515 + #if (defined(CONFIG_ZLIB_INFLATE) || defined(CONFIG_ZLIB_INFLATE_MODULE))
   4.516 +   case CLOOP_COMPRESSOR_ZLIB:
   4.517 +-   clo->zstream.next_in = clo->compressed_buffer;
   4.518 +-   clo->zstream.avail_in = compressed_length;
   4.519 +-   clo->zstream.next_out = clo->buffer[clo->current_bufnum];
   4.520 +-   clo->zstream.avail_out = clo->head.block_size;
   4.521  +   clo->zstream.next_in = source;
   4.522  +   clo->zstream.avail_in = sourceLen;
   4.523  +   clo->zstream.next_out = dest;
   4.524  +   clo->zstream.avail_out = *destLen;
   4.525 -+   err = zlib_inflateReset(&clo->zstream);
   4.526 -+   if (err != Z_OK)
   4.527 -+   {
   4.528 -+    printk(KERN_ERR "%s: zlib_inflateReset error %d\n", cloop_name, err);
   4.529 -+    zlib_inflateEnd(&clo->zstream); zlib_inflateInit(&clo->zstream);
   4.530 -+   }
   4.531 -+   err = zlib_inflate(&clo->zstream, Z_FINISH);
   4.532 +    err = zlib_inflateReset(&clo->zstream);
   4.533 +    if (err != Z_OK)
   4.534 +    {
   4.535 +@@ -239,50 +234,50 @@
   4.536 +     zlib_inflateEnd(&clo->zstream); zlib_inflateInit(&clo->zstream);
   4.537 +    }
   4.538 +    err = zlib_inflate(&clo->zstream, Z_FINISH);
   4.539 +-   *uncompressed_length = clo->zstream.total_out;
   4.540  +   *destLen = clo->zstream.total_out;
   4.541 -+   if (err == Z_STREAM_END) err = 0;
   4.542 +    if (err == Z_STREAM_END) err = 0;
   4.543 +-   DEBUGP("cloop: zlib decompression done, ret =%d, size =%lu\n", err, *uncompressed_length);
   4.544  +   DEBUGP("cloop: zlib decompression done, ret =%d, size =%lu\n", err, *destLen);
   4.545 -+   break;
   4.546 -+#endif
   4.547 -+#if (defined(CONFIG_LZO_DECOMPRESS) || defined(CONFIG_LZO_DECOMPRESS_MODULE))
   4.548 -+  case CLOOP_COMPRESSOR_LZO1X:
   4.549 -+   {
   4.550 -+    size_t tmp = (size_t) clo->head.block_size;
   4.551 +    break;
   4.552 + #endif
   4.553 + #if (defined(CONFIG_LZO_DECOMPRESS) || defined(CONFIG_LZO_DECOMPRESS_MODULE))
   4.554 +   case CLOOP_COMPRESSOR_LZO1X:
   4.555 +    {
   4.556 +     size_t tmp = (size_t) clo->head.block_size;
   4.557 +-    err = lzo1x_decompress_safe(clo->compressed_buffer, compressed_length,
   4.558 +-             clo->buffer[clo->current_bufnum], &tmp);
   4.559 +-    if (err == LZO_E_OK) *uncompressed_length = (u_int32_t) tmp;
   4.560  +    err = lzo1x_decompress_safe(source, sourceLen,
   4.561  +             dest, &tmp);
   4.562  +    if (err == LZO_E_OK) *destLen = (u_int32_t) tmp;
   4.563 -+   }
   4.564 -+   break;
   4.565 -+#endif
   4.566 -+#if (defined(CONFIG_DECOMPRESS_LZ4) || defined(CONFIG_DECOMPRESS_LZ4_MODULE))
   4.567 -+  case CLOOP_COMPRESSOR_LZ4:
   4.568 -+   {
   4.569 +    }
   4.570 +    break;
   4.571 + #endif
   4.572 + #if (defined(CONFIG_DECOMPRESS_LZ4) || defined(CONFIG_DECOMPRESS_LZ4_MODULE))
   4.573 +   case CLOOP_COMPRESSOR_LZ4:
   4.574 +    {
   4.575 +-    size_t outputSize = clo->head.block_size;
   4.576  +    size_t outputSize = *destLen;
   4.577 -+    /* We should adjust outputSize here, in case the last block is smaller than block_size */
   4.578 -+#if LINUX_VERSION_CODE < KERNEL_VERSION(4, 11, 0) /* field removed */
   4.579 +     /* We should adjust outputSize here, in case the last block is smaller than block_size */
   4.580 + #if LINUX_VERSION_CODE < KERNEL_VERSION(4, 11, 0) /* field removed */
   4.581 +-    err = lz4_decompress(clo->compressed_buffer, (size_t *) &compressed_length,
   4.582 +-                         clo->buffer[clo->current_bufnum], outputSize);
   4.583  +    err = lz4_decompress(source, (size_t *) &sourceLen,
   4.584  +                         dest, outputSize);
   4.585 -+#else
   4.586 + #else
   4.587 +-    err = LZ4_decompress_safe(clo->compressed_buffer,
   4.588 +-                              clo->buffer[clo->current_bufnum],
   4.589 +-                              compressed_length, outputSize);
   4.590  +    err = LZ4_decompress_safe(source,
   4.591  +                              dest,
   4.592  +                              sourceLen, outputSize);
   4.593 -+#endif
   4.594 -+    if (err >= 0) 
   4.595 -+    {
   4.596 -+     err = 0;
   4.597 + #endif
   4.598 +     if (err >= 0) 
   4.599 +     {
   4.600 +      err = 0;
   4.601 +-     *uncompressed_length = outputSize;
   4.602  +     *destLen = outputSize;
   4.603 -+    }
   4.604 -+   }
   4.605 -+  break;
   4.606 -+#endif
   4.607 -+#if (defined(CONFIG_DECOMPRESS_XZ) || defined(CONFIG_DECOMPRESS_XZ_MODULE))
   4.608 -+ case CLOOP_COMPRESSOR_XZ:
   4.609 +     }
   4.610 +    }
   4.611 +   break;
   4.612 + #endif
   4.613 + #if (defined(CONFIG_DECOMPRESS_XZ) || defined(CONFIG_DECOMPRESS_XZ_MODULE))
   4.614 +  case CLOOP_COMPRESSOR_XZ:
   4.615 +-  clo->xz_buffer.in = clo->compressed_buffer;
   4.616  +  clo->xz_buffer.in = source;
   4.617 -+  clo->xz_buffer.in_pos = 0;
   4.618 +   clo->xz_buffer.in_pos = 0;
   4.619 +-  clo->xz_buffer.in_size = compressed_length;
   4.620 +-  clo->xz_buffer.out = clo->buffer[clo->current_bufnum];
   4.621  +  clo->xz_buffer.in_size = sourceLen;
   4.622  +  clo->xz_buffer.out = dest;
   4.623 -+  clo->xz_buffer.out_pos = 0;
   4.624 +   clo->xz_buffer.out_pos = 0;
   4.625 +-  clo->xz_buffer.out_size = clo->head.block_size;
   4.626  +  clo->xz_buffer.out_size = *destLen;
   4.627 -+  xz_dec_reset(clo->xzdecoderstate);
   4.628 -+  err = xz_dec_run(clo->xzdecoderstate, &clo->xz_buffer);
   4.629 -+  if (err == XZ_STREAM_END || err == XZ_OK)
   4.630 -+  {
   4.631 -+   err = 0;
   4.632 -+  }
   4.633 -+  else
   4.634 -+  {
   4.635 -+   printk(KERN_ERR "%s: xz_dec_run error %d\n", cloop_name, err);
   4.636 -+   err = 1;
   4.637 -+  }
   4.638 -+  break;
   4.639 -+#endif
   4.640 -+ default:
   4.641 -+   printk(KERN_ERR "%s: compression method is not supported!\n", cloop_name);
   4.642 -+ }
   4.643 -+ return err;
   4.644 - }
   4.645 - 
   4.646 - static ssize_t cloop_read_from_file(struct cloop_device *clo, struct file *f, char *buf,
   4.647 -@@ -220,7 +313,7 @@
   4.648 +   xz_dec_reset(clo->xzdecoderstate);
   4.649 +   err = xz_dec_run(clo->xzdecoderstate, &clo->xz_buffer);
   4.650 +   if (err == XZ_STREAM_END || err == XZ_OK)
   4.651 +@@ -309,16 +304,12 @@
   4.652 +  while (buf_done < buf_len)
   4.653 +   {
   4.654 +    size_t size = buf_len - buf_done, size_read;
   4.655 +-   mm_segment_t old_fs;
   4.656 +    /* kernel_read() only supports 32 bit offsets, so we use vfs_read() instead. */
   4.657 +    /* int size_read = kernel_read(f, pos, buf + buf_done, size); */
   4.658 +-
   4.659 +-   // mutex_lock(&clo->clo_rq_mutex);
   4.660 +-   old_fs = get_fs();
   4.661 +-   set_fs(KERNEL_DS);
   4.662 ++   mm_segment_t old_fs = get_fs();
   4.663 ++   set_fs(get_ds());
   4.664 +    size_read = vfs_read(f, (void __user *)(buf + buf_done), size, &pos);
   4.665 +    set_fs(old_fs);
   4.666 +-   // mutex_unlock(&clo->clo_rq_mutex);
   4.667   
   4.668      if(size_read <= 0)
   4.669       {
   4.670 --     printk(KERN_ERR "%s: Read error %d at pos %Lu in file %s, "
   4.671 -+     printk(KERN_ERR "%s: Read error %d at pos %llu in file %s, "
   4.672 -                      "%d bytes lost.\n", cloop_name, (int)size_read, pos,
   4.673 - 		     file, (int)size);
   4.674 -      memset(buf + buf_len - size, 0, size);
   4.675 -@@ -232,72 +325,84 @@
   4.676 - }
   4.677 +@@ -358,8 +349,8 @@
   4.678 +    return i;
   4.679 +   }
   4.680   
   4.681 - /* This looks more complicated than it is */
   4.682 --/* Returns number of block buffer to use for this request */
   4.683 -+/* Returns number of cache block buffer to use for this request */
   4.684 - static int cloop_load_buffer(struct cloop_device *clo, int blocknum)
   4.685 - {
   4.686 -- unsigned int buf_done = 0;
   4.687 -- unsigned long buflen;
   4.688 -- unsigned int buf_length;
   4.689 -+ loff_t compressed_block_offset;
   4.690 -+ long compressed_block_len;
   4.691 -+ long uncompressed_block_len=0;
   4.692 -  int ret;
   4.693 -  int i;
   4.694 -- if(blocknum > ntohl(clo->head.num_blocks) || blocknum < 0)
   4.695 --  {
   4.696 --   printk(KERN_WARNING "%s: Invalid block number %d requested.\n",
   4.697 --                       cloop_name, blocknum);
   4.698 --   return -1;
   4.699 --  }
   4.700 -+ if(blocknum > clo->head.num_blocks || blocknum < 0)
   4.701 -+ {
   4.702 -+  printk(KERN_WARNING "%s: Invalid block number %d requested.\n",
   4.703 -+         cloop_name, blocknum);
   4.704 -+  return -1;
   4.705 -+ }
   4.706 - 
   4.707 -  /* Quick return if the block we seek is already in one of the buffers. */
   4.708 -  /* Return number of buffer */
   4.709 -- for(i=0; i<BUFFERED_BLOCKS; i++)
   4.710 -+ for(i=0; i<clo->num_buffered_blocks; i++)
   4.711 -   if (blocknum == clo->buffered_blocknum[i])
   4.712 --   {
   4.713 --    DEBUGP(KERN_INFO "cloop_load_buffer: Found buffered block %d\n", i);
   4.714 --    return i;
   4.715 --   }
   4.716 --
   4.717 -- buf_length = be64_to_cpu(clo->offsets[blocknum+1]) - be64_to_cpu(clo->offsets[blocknum]);
   4.718 --
   4.719 --/* Load one compressed block from the file. */
   4.720 -- cloop_read_from_file(clo, clo->backing_file, (char *)clo->compressed_buffer,
   4.721 --                    be64_to_cpu(clo->offsets[blocknum]), buf_length);
   4.722 -+  {
   4.723 -+   DEBUGP(KERN_INFO "cloop_load_buffer: Found buffered block %d\n", i);
   4.724 -+   return i;
   4.725 -+  }
   4.726 - 
   4.727 -- buflen = ntohl(clo->head.block_size);
   4.728 +- compressed_block_offset = CLOOP_BLOCK_OFFSET(clo->block_ptrs[blocknum]);
   4.729 +- compressed_block_len = (long) (CLOOP_BLOCK_OFFSET(clo->block_ptrs[blocknum+1]) - compressed_block_offset) ;
   4.730  + compressed_block_offset = clo->block_ptrs[blocknum].offset;
   4.731  + compressed_block_len = (long) (clo->block_ptrs[blocknum].size) ;
   4.732   
   4.733 -- /* Go to next position in the block ring buffer */
   4.734 -- clo->current_bufnum++;
   4.735 -- if(clo->current_bufnum >= BUFFERED_BLOCKS) clo->current_bufnum = 0;
   4.736 -+ /* Load one compressed block from the file. */
   4.737 -+ if(compressed_block_offset > 0 && compressed_block_len >= 0) /* sanity check */
   4.738 -+ {
   4.739 -+  size_t n = cloop_read_from_file(clo, clo->backing_file, (char *)clo->compressed_buffer,
   4.740 -+                    compressed_block_offset, compressed_block_len);
   4.741 -+  if (n!= compressed_block_len)
   4.742 -+   {
   4.743 -+    printk(KERN_ERR "%s: error while reading %lu bytes @ %llu from file %s\n",
   4.744 +  /* Load one compressed block from the file. */
   4.745 +  if(compressed_block_offset > 0 && compressed_block_len >= 0) /* sanity check */
   4.746 +@@ -369,12 +360,12 @@
   4.747 +   if (n!= compressed_block_len)
   4.748 +    {
   4.749 +     printk(KERN_ERR "%s: error while reading %lu bytes @ %llu from file %s\n",
   4.750 +-     cloop_name, compressed_block_len, clo->block_ptrs[blocknum], clo->underlying_filename);
   4.751  +     cloop_name, compressed_block_len, clo->block_ptrs[blocknum].offset, clo->underlying_filename);
   4.752 -+    /* return -1; */
   4.753 -+   }
   4.754 -+ } else {
   4.755 -+  printk(KERN_ERR "%s: invalid data block len %ld bytes @ %lld from file %s\n",
   4.756 +     /* return -1; */
   4.757 +    }
   4.758 +  } else {
   4.759 +   printk(KERN_ERR "%s: invalid data block len %ld bytes @ %lld from file %s\n",
   4.760 +-  cloop_name, compressed_block_len, clo->block_ptrs[blocknum], clo->underlying_filename);
   4.761  +  cloop_name, compressed_block_len, clo->block_ptrs[blocknum].offset, clo->underlying_filename);
   4.762 -+  return -1;
   4.763 -+ }
   4.764 -+  
   4.765 -+ /* Go to next position in the cache block buffer (which is used as a cyclic buffer) */
   4.766 -+ if(++clo->current_bufnum >= clo->num_buffered_blocks) clo->current_bufnum = 0;
   4.767 +   return -1;
   4.768 +  }
   4.769 +   
   4.770 +@@ -382,14 +373,16 @@
   4.771 +  if(++clo->current_bufnum >= clo->num_buffered_blocks) clo->current_bufnum = 0;
   4.772   
   4.773    /* Do the uncompression */
   4.774 -- ret = uncompress(clo, clo->buffer[clo->current_bufnum], &buflen, clo->compressed_buffer,
   4.775 --                  buf_length);
   4.776 +- ret = uncompress(clo, blocknum, compressed_block_len, &uncompressed_block_len);
   4.777  + uncompressed_block_len = clo->head.block_size;
   4.778  + ret = uncompress(clo, clo->buffer[clo->current_bufnum], &uncompressed_block_len,
   4.779  +	 clo->compressed_buffer, compressed_block_len, clo->block_ptrs[blocknum].flags);
   4.780    /* DEBUGP("cloop: buflen after uncompress: %ld\n",buflen); */
   4.781    if (ret != 0)
   4.782 --  {
   4.783 --   printk(KERN_ERR "%s: zlib decompression error %i uncompressing block %u %u/%lu/%u/%u "
   4.784 --          "%Lu-%Lu\n", cloop_name, ret, blocknum,
   4.785 --	  ntohl(clo->head.block_size), buflen, buf_length, buf_done,
   4.786 --	  be64_to_cpu(clo->offsets[blocknum]), be64_to_cpu(clo->offsets[blocknum+1]));
   4.787 --   clo->buffered_blocknum[clo->current_bufnum] = -1;
   4.788 --   return -1;
   4.789 --  }
   4.790 -+ {
   4.791 -+  printk(KERN_ERR "%s: decompression error %i uncompressing block %u %lu bytes @ %llu, flags %u\n",
   4.792 -+         cloop_name, ret, blocknum,
   4.793 +  {
   4.794 +   printk(KERN_ERR "%s: decompression error %i uncompressing block %u %lu bytes @ %llu, flags %u\n",
   4.795 +          cloop_name, ret, blocknum,
   4.796 +-         compressed_block_len, CLOOP_BLOCK_OFFSET(clo->block_ptrs[blocknum]),
   4.797 +-         CLOOP_BLOCK_FLAGS(clo->block_ptrs[blocknum]));
   4.798  +         compressed_block_len, clo->block_ptrs[blocknum].offset,
   4.799  +         clo->block_ptrs[blocknum].flags);
   4.800 -+         clo->buffered_blocknum[clo->current_bufnum] = -1;
   4.801 -+  return -1;
   4.802 -+ }
   4.803 -  clo->buffered_blocknum[clo->current_bufnum] = blocknum;
   4.804 +          clo->buffered_blocknum[clo->current_bufnum] = -1;
   4.805 +   return -1;
   4.806 +  }
   4.807 +@@ -397,107 +390,146 @@
   4.808    return clo->current_bufnum;
   4.809   }
   4.810   
   4.811 - /* This function does all the real work. */
   4.812 --/* returns "uptodate" */
   4.813 +-static blk_status_t cloop_handle_request(struct cloop_device *clo, struct request *req)
   4.814 ++/* This function does all the real work. */
   4.815  +/* returns "uptodate"                    */
   4.816 - static int cloop_handle_request(struct cloop_device *clo, struct request *req)
   4.817 ++static int cloop_handle_request(struct cloop_device *clo, struct request *req)
   4.818   {
   4.819    int buffered_blocknum = -1;
   4.820    int preloaded = 0;
   4.821 -  loff_t offset     = (loff_t) blk_rq_pos(req)<<9; /* req->sector<<9 */
   4.822 -- struct bio_vec *bvec;
   4.823 -+ struct bio_vec bvec;
   4.824 +- loff_t offset = (loff_t) blk_rq_pos(req)<<9;
   4.825 ++ loff_t offset     = (loff_t) blk_rq_pos(req)<<9; /* req->sector<<9 */
   4.826 +  struct bio_vec bvec;
   4.827    struct req_iterator iter;
   4.828 +- blk_status_t ret = BLK_STS_OK;
   4.829 +-
   4.830 +- if (unlikely(req_op(req) != REQ_OP_READ ))
   4.831 +- {
   4.832 +-  blk_dump_rq_flags(req, DEVICE_NAME " bad request");
   4.833 +-  return BLK_STS_IOERR;
   4.834 +- }
   4.835 +-
   4.836 +- if (unlikely(!clo->backing_file && !clo->suspended))
   4.837 +- {
   4.838 +-  DEBUGP("cloop_handle_request: not connected to a file\n");
   4.839 +-  return BLK_STS_IOERR;
   4.840 +- }
   4.841 +-
   4.842    rq_for_each_segment(bvec, req, iter)
   4.843 +- {
   4.844 +-  unsigned long len = bvec.bv_len;
   4.845 +-  loff_t to_offset  = bvec.bv_offset;
   4.846 +-
   4.847 +-  while(len > 0)
   4.848     {
   4.849 --   unsigned long len = bvec->bv_len;
   4.850 --   char *to_ptr      = kmap(bvec->bv_page) + bvec->bv_offset;
   4.851 +-   u_int32_t length_in_buffer;
   4.852 +-   loff_t block_offset = offset;
   4.853 +-   u_int32_t offset_in_buffer;
   4.854 +-   char *from_ptr, *to_ptr;
   4.855 +-   /* do_div (div64.h) returns the 64bit division remainder and  */
   4.856 +-   /* puts the result in the first argument, i.e. block_offset   */
   4.857 +-   /* becomes the blocknumber to load, and offset_in_buffer the  */
   4.858 +-   /* position in the buffer */
   4.859 +-   offset_in_buffer = do_div(block_offset, clo->head.block_size);
   4.860 +-   /* Lookup preload cache */
   4.861 +-   if(block_offset < clo->preload_size && clo->preload_cache != NULL && clo->preload_cache[block_offset] != NULL)
   4.862 +-   { /* Copy from cache */
   4.863 +-    preloaded = 1;
   4.864 +-    from_ptr = clo->preload_cache[block_offset];
   4.865 +-   }
   4.866 +-   else
   4.867 +-   {
   4.868 +-    preloaded = 0;
   4.869 +-    buffered_blocknum = cloop_load_buffer(clo,block_offset);
   4.870 +-    if(buffered_blocknum == -1)
   4.871  +   unsigned long len = bvec.bv_len;
   4.872  +   char *to_ptr      = kmap(bvec.bv_page) + bvec.bv_offset;
   4.873 -    while(len > 0)
   4.874 ++   while(len > 0)
   4.875       {
   4.876 -      u_int32_t length_in_buffer;
   4.877 -@@ -308,7 +413,7 @@
   4.878 -      /* puts the result in the first argument, i.e. block_offset   */
   4.879 -      /* becomes the blocknumber to load, and offset_in_buffer the  */
   4.880 -      /* position in the buffer */
   4.881 --     offset_in_buffer = do_div(block_offset, ntohl(clo->head.block_size));
   4.882 +-     ret = BLK_STS_IOERR;
   4.883 +-     break; /* invalid data, leave inner loop */
   4.884 ++     u_int32_t length_in_buffer;
   4.885 ++     loff_t block_offset = offset;
   4.886 ++     u_int32_t offset_in_buffer;
   4.887 ++     char *from_ptr;
   4.888 ++     /* do_div (div64.h) returns the 64bit division remainder and  */
   4.889 ++     /* puts the result in the first argument, i.e. block_offset   */
   4.890 ++     /* becomes the blocknumber to load, and offset_in_buffer the  */
   4.891 ++     /* position in the buffer */
   4.892  +     offset_in_buffer = do_div(block_offset, clo->head.block_size);
   4.893 -      /* Lookup preload cache */
   4.894 -      if(block_offset < clo->preload_size && clo->preload_cache != NULL &&
   4.895 -         clo->preload_cache[block_offset] != NULL)
   4.896 -@@ -325,7 +430,7 @@
   4.897 -        from_ptr = clo->buffer[buffered_blocknum];
   4.898 -       }
   4.899 -      /* Now, at least part of what we want will be in the buffer. */
   4.900 --     length_in_buffer = ntohl(clo->head.block_size) - offset_in_buffer;
   4.901 ++     /* Lookup preload cache */
   4.902 ++     if(block_offset < clo->preload_size && clo->preload_cache != NULL &&
   4.903 ++        clo->preload_cache[block_offset] != NULL)
   4.904 ++      { /* Copy from cache */
   4.905 ++       preloaded = 1;
   4.906 ++       from_ptr = clo->preload_cache[block_offset];
   4.907 ++      }
   4.908 ++     else
   4.909 ++      {
   4.910 ++       preloaded = 0;
   4.911 ++       buffered_blocknum = cloop_load_buffer(clo,block_offset);
   4.912 ++       if(buffered_blocknum == -1) break; /* invalid data, leave inner loop */
   4.913 ++       /* Copy from buffer */
   4.914 ++       from_ptr = clo->buffer[buffered_blocknum];
   4.915 ++      }
   4.916 ++     /* Now, at least part of what we want will be in the buffer. */
   4.917  +     length_in_buffer = clo->head.block_size - offset_in_buffer;
   4.918 -      if(length_in_buffer > len)
   4.919 -       {
   4.920 - /*   DEBUGP("Warning: length_in_buffer=%u > len=%u\n",
   4.921 -@@ -337,18 +442,19 @@
   4.922 -      len         -= length_in_buffer;
   4.923 -      offset      += length_in_buffer;
   4.924 -     } /* while inner loop */
   4.925 --   kunmap(bvec->bv_page);
   4.926 ++     if(length_in_buffer > len)
   4.927 ++      {
   4.928 ++/*   DEBUGP("Warning: length_in_buffer=%u > len=%u\n",
   4.929 ++                      length_in_buffer,len); */
   4.930 ++       length_in_buffer = len;
   4.931 ++      }
   4.932 ++     memcpy(to_ptr, from_ptr + offset_in_buffer, length_in_buffer);
   4.933 ++     to_ptr      += length_in_buffer;
   4.934 ++     len         -= length_in_buffer;
   4.935 ++     offset      += length_in_buffer;
   4.936 ++    } /* while inner loop */
   4.937  +   kunmap(bvec.bv_page);
   4.938  +   cond_resched();
   4.939 -   } /* end rq_for_each_segment*/
   4.940 -  return ((buffered_blocknum!=-1) || preloaded);
   4.941 - }
   4.942 - 
   4.943 - /* Adopted from loop.c, a kernel thread to handle physical reads and
   4.944 -- * decompression. */
   4.945 ++  } /* end rq_for_each_segment*/
   4.946 ++ return ((buffered_blocknum!=-1) || preloaded);
   4.947 ++}
   4.948 ++
   4.949 ++/* Adopted from loop.c, a kernel thread to handle physical reads and
   4.950  +   decompression. */
   4.951 - static int cloop_thread(void *data)
   4.952 - {
   4.953 -  struct cloop_device *clo = data;
   4.954 -  current->flags |= PF_NOFREEZE;
   4.955 -- set_user_nice(current, -15);
   4.956 ++static int cloop_thread(void *data)
   4.957 ++{
   4.958 ++ struct cloop_device *clo = data;
   4.959 ++ current->flags |= PF_NOFREEZE;
   4.960  + set_user_nice(current, 10);
   4.961 -  while (!kthread_should_stop()||!list_empty(&clo->clo_list))
   4.962 -   {
   4.963 -    int err;
   4.964 -@@ -390,10 +496,18 @@
   4.965 -    int rw;
   4.966 -  /* quick sanity checks */
   4.967 -    /* blk_fs_request() was removed in 2.6.36 */
   4.968 --   if (unlikely(req == NULL || (req->cmd_type != REQ_TYPE_FS)))
   4.969 ++ while (!kthread_should_stop()||!list_empty(&clo->clo_list))
   4.970 ++  {
   4.971 ++   int err;
   4.972 ++   err = wait_event_interruptible(clo->clo_event, !list_empty(&clo->clo_list) || 
   4.973 ++                                  kthread_should_stop());
   4.974 ++   if(unlikely(err))
   4.975 ++    {
   4.976 ++     DEBUGP(KERN_ERR "cloop thread activated on error!? Continuing.\n");
   4.977 ++     continue;
   4.978 +     }
   4.979 +-    /* Copy from buffer */
   4.980 +-    from_ptr = clo->buffer[buffered_blocknum];
   4.981 +-   }
   4.982 +-   /* Now, at least part of what we want will be in the buffer. */
   4.983 +-   length_in_buffer = clo->head.block_size - offset_in_buffer;
   4.984 +-   if(length_in_buffer > len)
   4.985 +-   {
   4.986 +-   /* DEBUGP("Warning: length_in_buffer=%u > len=%u\n", length_in_buffer,len); */
   4.987 +-    length_in_buffer = len;
   4.988 +-   }
   4.989 +-   to_ptr      = kmap_atomic(bvec.bv_page);
   4.990 +-   memcpy(to_ptr + to_offset, from_ptr + offset_in_buffer, length_in_buffer);
   4.991 +-   kunmap_atomic(to_ptr);
   4.992 +-   to_offset   += length_in_buffer;
   4.993 +-   len         -= length_in_buffer;
   4.994 +-   offset      += length_in_buffer;
   4.995 +-  } /* while inner loop */
   4.996 +- } /* rq_for_each_segment */
   4.997 +- return ret;
   4.998 +-}
   4.999 +-
  4.1000 +-static blk_status_t cloop_queue_rq(struct blk_mq_hw_ctx *hctx, const struct blk_mq_queue_data *bd)
  4.1001 +-{
  4.1002 +-//  struct request_queue *q  = hctx->queue;
  4.1003 +-//  struct cloop_device *clo = q->queuedata;
  4.1004 +- struct request *req = bd->rq;
  4.1005 +- struct cloop_device *clo = req->rq_disk->private_data;
  4.1006 +- blk_status_t ret         = BLK_STS_OK;
  4.1007 +-
  4.1008 +-#if 1 /* Does it work when loading libraries? */
  4.1009 +- /* Since we have a buffered block list as well as data to read */
  4.1010 +- /* from disk (slow), and are (probably) never called from an   */
  4.1011 +- /* interrupt, we use a simple mutex lock right here to ensure  */
  4.1012 +- /* consistency.                                                */
  4.1013 +-  mutex_lock(&clo->clo_rq_mutex);
  4.1014 +- #else
  4.1015 +-  spin_lock_irq(&clo->queue_lock);
  4.1016 +- #endif
  4.1017 +- blk_mq_start_request(req);
  4.1018 +- do {
  4.1019 +-  ret = cloop_handle_request(clo, req);
  4.1020 +- } while(blk_update_request(req, ret, blk_rq_cur_bytes(req)));
  4.1021 +- blk_mq_end_request(req, ret);
  4.1022 +- #if 1 /* See above */
  4.1023 +-  mutex_unlock(&clo->clo_rq_mutex);
  4.1024 +- #else
  4.1025 +-  spin_unlock_irq(&clo->queue_lock);
  4.1026 +- #endif
  4.1027 +- return ret;
  4.1028 ++   if(!list_empty(&clo->clo_list))
  4.1029 ++    {
  4.1030 ++     struct request *req;
  4.1031 ++     unsigned long flags;
  4.1032 ++     int uptodate;
  4.1033 ++     spin_lock_irq(&clo->queue_lock);
  4.1034 ++     req = list_entry(clo->clo_list.next, struct request, queuelist);
  4.1035 ++     list_del_init(&req->queuelist);
  4.1036 ++     spin_unlock_irq(&clo->queue_lock);
  4.1037 ++     uptodate = cloop_handle_request(clo, req);
  4.1038 ++     spin_lock_irqsave(&clo->queue_lock, flags);
  4.1039 ++     __blk_end_request_all(req, uptodate ? 0 : -EIO);
  4.1040 ++     spin_unlock_irqrestore(&clo->queue_lock, flags);
  4.1041 ++    }
  4.1042 ++  }
  4.1043 ++ DEBUGP(KERN_ERR "cloop_thread exited.\n");
  4.1044 ++ return 0;
  4.1045 ++}
  4.1046 ++
  4.1047 ++/* This is called by the kernel block queue management every now and then,
  4.1048 ++ * with successive read requests qeued and sorted in a (hopefully)
  4.1049 ++ * "most efficient way". spin_lock_irq() is being held by the kernel. */
  4.1050 ++static void cloop_do_request(struct request_queue *q)
  4.1051 ++{
  4.1052 ++ struct request *req;
  4.1053 ++ while((req = blk_fetch_request(q)) != NULL)
  4.1054 ++  {
  4.1055 ++   struct cloop_device *clo;
  4.1056 ++   int rw;
  4.1057 ++ /* quick sanity checks */
  4.1058 ++   /* blk_fs_request() was removed in 2.6.36 */
  4.1059  +   if (unlikely(req == NULL
  4.1060  +#if LINUX_VERSION_CODE < KERNEL_VERSION(4, 11, 0) /* field removed */
  4.1061  +   || (req->cmd_type != REQ_TYPE_FS)
  4.1062  +#endif
  4.1063  +   ))
  4.1064 -     goto error_continue;
  4.1065 -    rw = rq_data_dir(req);
  4.1066 --   if (unlikely(rw != READ && rw != READA))
  4.1067 ++    goto error_continue;
  4.1068 ++   rw = rq_data_dir(req);
  4.1069  +   if (unlikely(rw != READ
  4.1070  +#if LINUX_VERSION_CODE < KERNEL_VERSION(4, 8, 0)
  4.1071  +                && rw != READA
  4.1072  +#endif
  4.1073  +    ))
  4.1074 -     {
  4.1075 -      DEBUGP("cloop_do_request: bad command\n");
  4.1076 -      goto error_continue;
  4.1077 -@@ -409,40 +523,51 @@
  4.1078 -    continue; /* next request */
  4.1079 -   error_continue:
  4.1080 -    DEBUGP(KERN_ERR "cloop_do_request: Discarding request %p.\n", req);
  4.1081 ++    {
  4.1082 ++     DEBUGP("cloop_do_request: bad command\n");
  4.1083 ++     goto error_continue;
  4.1084 ++    }
  4.1085 ++   clo = req->rq_disk->private_data;
  4.1086 ++   if (unlikely(!clo->backing_file && !clo->suspended))
  4.1087 ++    {
  4.1088 ++     DEBUGP("cloop_do_request: not connected to a file\n");
  4.1089 ++     goto error_continue;
  4.1090 ++    }
  4.1091 ++   list_add_tail(&req->queuelist, &clo->clo_list); /* Add to working list for thread */
  4.1092 ++   wake_up(&clo->clo_event);    /* Wake up cloop_thread */
  4.1093 ++   continue; /* next request */
  4.1094 ++  error_continue:
  4.1095 ++   DEBUGP(KERN_ERR "cloop_do_request: Discarding request %p.\n", req);
  4.1096  +#if LINUX_VERSION_CODE < KERNEL_VERSION(4, 12, 0)
  4.1097 -    req->errors++;
  4.1098 ++   req->errors++;
  4.1099  +#else
  4.1100  +   req->error_count++;
  4.1101  +#endif
  4.1102 -    __blk_end_request_all(req, -EIO);
  4.1103 -   }
  4.1104 ++   __blk_end_request_all(req, -EIO);
  4.1105 ++  }
  4.1106   }
  4.1107   
  4.1108 --/* Read header and offsets from already opened file */
  4.1109 --static int cloop_set_file(int cloop_num, struct file *file, char *filename)
  4.1110 -+/* Read header, flags and offsets from already opened file */
  4.1111 -+static int cloop_set_file(int cloop_num, struct file *file)
  4.1112 - {
  4.1113 -  struct cloop_device *clo = cloop_dev[cloop_num];
  4.1114 -  struct inode *inode;
  4.1115 + /* Read header, flags and offsets from already opened file */
  4.1116 +@@ -508,7 +540,7 @@
  4.1117    char *bbuf=NULL;
  4.1118 -- unsigned int i, offsets_read, total_offsets;
  4.1119 -- int isblkdev;
  4.1120 -- int error = 0;
  4.1121 -+ unsigned int bbuf_size = 0;
  4.1122 -+ const unsigned int header_size = sizeof(struct cloop_head);
  4.1123 +  unsigned int bbuf_size = 0;
  4.1124 +  const unsigned int header_size = sizeof(struct cloop_head);
  4.1125 +- unsigned int i, offsets_read=0, total_offsets=0;
  4.1126  + unsigned int i, total_offsets=0;
  4.1127 -+ loff_t fs_read_position = 0, header_pos[2];
  4.1128 -+ int flags, isblkdev, bytes_read, error = 0;
  4.1129 -+ if (clo->suspended) return error;
  4.1130 -+ #if LINUX_VERSION_CODE < KERNEL_VERSION(4, 0, 0)
  4.1131 -  inode = file->f_dentry->d_inode;
  4.1132 -+ clo->underlying_filename = kstrdup(file->f_dentry->d_name.name ? file->f_dentry->d_name.name : (const unsigned char *)"anonymous filename", GFP_KERNEL);
  4.1133 -+ #else
  4.1134 -+ inode = file->f_path.dentry->d_inode;
  4.1135 -+ clo->underlying_filename = kstrdup(file->f_path.dentry->d_name.name ? file->f_path.dentry->d_name.name : (const unsigned char *)"anonymous filename", GFP_KERNEL);
  4.1136 -+ #endif
  4.1137 -  isblkdev=S_ISBLK(inode->i_mode)?1:0;
  4.1138 -  if(!isblkdev&&!S_ISREG(inode->i_mode))
  4.1139 +  loff_t fs_read_position = 0, header_pos[2];
  4.1140 +  int isblkdev, bytes_read, error = 0;
  4.1141 +  if (clo->suspended) return error;
  4.1142 +@@ -581,29 +613,19 @@
  4.1143 +     goto error_release;
  4.1144 +    }
  4.1145 +    memcpy(&clo->head, bbuf, header_size);
  4.1146 +-   if (strncmp(bbuf+CLOOP4_SIGNATURE_OFFSET, CLOOP4_SIGNATURE, CLOOP4_SIGNATURE_SIZE)==0)
  4.1147 ++   if (strncmp(bbuf+CLOOP_SIGNATURE_OFFSET, CLOOP_SIGNATURE, CLOOP_SIGNATURE_SIZE)==0)
  4.1148 +    {
  4.1149 +-    clo->file_format=4;
  4.1150 ++    clo->file_format++;
  4.1151 +     clo->head.block_size=ntohl(clo->head.block_size);
  4.1152 +     clo->head.num_blocks=ntohl(clo->head.num_blocks);
  4.1153 +     clo->header_first =  (i==0) ? 1 : 0;
  4.1154 +-    printk(KERN_INFO "%s: file %s version %d, %d blocks of %d bytes, header %s.\n", cloop_name, clo->underlying_filename, clo->file_format, clo->head.num_blocks, clo->head.block_size, (i==0)?"first":"last");
  4.1155 +-    break;
  4.1156 +-   }
  4.1157 +-   else if (strncmp(bbuf+CLOOP2_SIGNATURE_OFFSET, CLOOP2_SIGNATURE, CLOOP2_SIGNATURE_SIZE)==0)
  4.1158 +-   {
  4.1159 +-    clo->file_format=2;
  4.1160 +-    clo->head.block_size=ntohl(clo->head.block_size);
  4.1161 +-    clo->head.num_blocks=ntohl(clo->head.num_blocks);
  4.1162 +-    clo->header_first =  (i==0) ? 1 : 0;
  4.1163 +-    printk(KERN_INFO "%s: file %s version %d, %d blocks of %d bytes, header %s.\n", cloop_name, clo->underlying_filename, clo->file_format, clo->head.num_blocks, clo->head.block_size, (i==0)?"first":"last");
  4.1164 ++    printk(KERN_INFO "%s: file %s, %d blocks of %d bytes, header %s.\n", cloop_name, clo->underlying_filename, clo->head.num_blocks, clo->head.block_size, (i==0)?"first":"last");
  4.1165 +     break;
  4.1166 +    }
  4.1167 +   }
  4.1168 +  if (clo->file_format == 0)
  4.1169     {
  4.1170 -    printk(KERN_ERR "%s: %s not a regular file or block device\n",
  4.1171 --		   cloop_name, filename);
  4.1172 -+		   cloop_name, clo->underlying_filename);
  4.1173 +-   printk(KERN_ERR "%s: Cannot read old 32-bit (version 0.68) images, "
  4.1174 +-                   "please use an older version of %s for this file.\n",
  4.1175 ++   printk(KERN_ERR "%s: Cannot detect %s format.\n",
  4.1176 +                    cloop_name, cloop_name);
  4.1177 +        error=-EBADF; goto error_release;
  4.1178 +   }
  4.1179 +@@ -613,67 +635,133 @@
  4.1180 +           cloop_name, clo->head.block_size);
  4.1181      error=-EBADF; goto error_release;
  4.1182     }
  4.1183 -  clo->backing_file = file;
  4.1184 -  clo->backing_inode= inode ;
  4.1185 -- if(!isblkdev&&inode->i_size<sizeof(struct cloop_head))
  4.1186 -+ clo->underlying_total_size = (isblkdev) ? inode->i_bdev->bd_inode->i_size : inode->i_size;
  4.1187 -+ if(clo->underlying_total_size < header_size)
  4.1188 +- total_offsets=clo->head.num_blocks+1;
  4.1189 +- if (!isblkdev && (sizeof(struct cloop_head)+sizeof(loff_t)*
  4.1190 ++ total_offsets=clo->head.num_blocks;
  4.1191 ++ if (!isblkdev && (sizeof(struct cloop_head)+sizeof(struct block_info)*
  4.1192 +                       total_offsets > inode->i_size))
  4.1193     {
  4.1194 --   printk(KERN_ERR "%s: %lu bytes (must be >= %u bytes)\n",
  4.1195 --                   cloop_name, (unsigned long)inode->i_size,
  4.1196 --		   (unsigned)sizeof(struct cloop_head));
  4.1197 -+   printk(KERN_ERR "%s: %llu bytes (must be >= %u bytes)\n",
  4.1198 -+                   cloop_name, clo->underlying_total_size,
  4.1199 -+		   (unsigned int)header_size);
  4.1200 +    printk(KERN_ERR "%s: file %s too small for %u blocks\n",
  4.1201 +           cloop_name, clo->underlying_filename, clo->head.num_blocks);
  4.1202      error=-EBADF; goto error_release;
  4.1203     }
  4.1204 -- /* In suspended mode, we have done all checks necessary - FF */
  4.1205 -- if (clo->suspended)
  4.1206 --   return error;
  4.1207 -  if(isblkdev)
  4.1208 -   {
  4.1209 -    struct request_queue *q = bdev_get_queue(inode->i_bdev);
  4.1210 -@@ -451,104 +576,225 @@
  4.1211 -    /* blk_queue_max_hw_segments(clo->clo_queue, queue_max_hw_segments(q)); */ /* Removed in 2.6.34 */
  4.1212 -    blk_queue_max_segment_size(clo->clo_queue, queue_max_segment_size(q));
  4.1213 -    blk_queue_segment_boundary(clo->clo_queue, queue_segment_boundary(q));
  4.1214 -+#if LINUX_VERSION_CODE < KERNEL_VERSION(4, 3, 0)
  4.1215 -    blk_queue_merge_bvec(clo->clo_queue, q->merge_bvec_fn);
  4.1216 -+#endif
  4.1217 -    clo->underlying_blksize = block_size(inode->i_bdev);
  4.1218 -   }
  4.1219 -  else
  4.1220 -    clo->underlying_blksize = PAGE_SIZE;
  4.1221 -- DEBUGP("Underlying blocksize is %u\n", clo->underlying_blksize);
  4.1222 -- bbuf = cloop_malloc(clo->underlying_blksize);
  4.1223 -+
  4.1224 -+ DEBUGP(KERN_INFO "Underlying blocksize of %s is %u\n", clo->underlying_filename, clo->underlying_blksize);
  4.1225 -+ DEBUGP(KERN_INFO "Underlying total size of %s is %llu\n", clo->underlying_filename, clo->underlying_total_size);
  4.1226 -+
  4.1227 -+ /* clo->underlying_blksize should be larger than header_size, even if it's only PAGE_SIZE */
  4.1228 -+ bbuf_size = clo->underlying_blksize;
  4.1229 -+ bbuf = cloop_malloc(bbuf_size);
  4.1230 -  if(!bbuf)
  4.1231 -   {
  4.1232 --   printk(KERN_ERR "%s: out of kernel mem for block buffer (%lu bytes)\n",
  4.1233 --                   cloop_name, (unsigned long)clo->underlying_blksize);
  4.1234 -+   printk(KERN_ERR "%s: out of kernel mem for buffer (%u bytes)\n",
  4.1235 -+                   cloop_name, (unsigned int) bbuf_size);
  4.1236 -+   error=-ENOMEM; goto error_release;
  4.1237 -+  }
  4.1238 -+
  4.1239 -+ header_pos[0] = 0; /* header first */
  4.1240 -+ header_pos[1] = clo->underlying_total_size - sizeof(struct cloop_head); /* header last */
  4.1241 -+ for(i=0; i<2; i++)
  4.1242 -+  {
  4.1243 -+   /* Check for header */
  4.1244 -+   size_t bytes_readable = MIN(clo->underlying_blksize, clo->underlying_total_size - header_pos[i]);
  4.1245 -+   size_t bytes_read = cloop_read_from_file(clo, file, bbuf, header_pos[i], bytes_readable);
  4.1246 -+   if(bytes_read != bytes_readable)
  4.1247 -+   {
  4.1248 -+    printk(KERN_ERR "%s: Bad file %s, read() of %s %u bytes returned %d.\n",
  4.1249 -+                    cloop_name, clo->underlying_filename, (i==0)?"first":"last",
  4.1250 -+		    (unsigned int)header_size, (int)bytes_read);
  4.1251 -+    error=-EBADF;
  4.1252 -+    goto error_release;
  4.1253 -+   }
  4.1254 -+   memcpy(&clo->head, bbuf, header_size);
  4.1255 -+   if (strncmp(bbuf+CLOOP_SIGNATURE_OFFSET, CLOOP_SIGNATURE, CLOOP_SIGNATURE_SIZE)==0)
  4.1256 -+   {
  4.1257 -+    clo->file_format++;
  4.1258 -+    clo->head.block_size=ntohl(clo->head.block_size);
  4.1259 -+    clo->head.num_blocks=ntohl(clo->head.num_blocks);
  4.1260 -+    clo->header_first =  (i==0) ? 1 : 0;
  4.1261 -+    printk(KERN_INFO "%s: file %s, %d blocks of %d bytes, header %s.\n", cloop_name, clo->underlying_filename, clo->head.num_blocks, clo->head.block_size, (i==0)?"first":"last");
  4.1262 -+    break;
  4.1263 -+   }
  4.1264 -+  }
  4.1265 -+ if (clo->file_format == 0)
  4.1266 -+  {
  4.1267 -+   printk(KERN_ERR "%s: Cannot detect %s format.\n",
  4.1268 -+                   cloop_name, cloop_name);
  4.1269 -+       error=-EBADF; goto error_release;
  4.1270 -+  }
  4.1271 -+ if (clo->head.block_size % 512 != 0)
  4.1272 -+  {
  4.1273 -+   printk(KERN_ERR "%s: blocksize %u not multiple of 512\n",
  4.1274 -+          cloop_name, clo->head.block_size);
  4.1275 -+   error=-EBADF; goto error_release;
  4.1276 -+  }
  4.1277 -+ total_offsets=clo->head.num_blocks;
  4.1278 -+ if (!isblkdev && (sizeof(struct cloop_head)+sizeof(struct block_info)*
  4.1279 -+                      total_offsets > inode->i_size))
  4.1280 -+  {
  4.1281 -+   printk(KERN_ERR "%s: file %s too small for %u blocks\n",
  4.1282 -+          cloop_name, clo->underlying_filename, clo->head.num_blocks);
  4.1283 -+   error=-EBADF; goto error_release;
  4.1284 -+  }
  4.1285 +- clo->block_ptrs = cloop_malloc(sizeof(cloop_block_ptr) * total_offsets);
  4.1286 +- if (!clo->block_ptrs)
  4.1287  + /* Allocate Memory for decompressors */
  4.1288  +#if (defined(CONFIG_ZLIB_INFLATE) || defined(CONFIG_ZLIB_INFLATE_MODULE))
  4.1289  + clo->zstream.workspace = cloop_malloc(zlib_inflate_workspacesize());
  4.1290  + if(!clo->zstream.workspace)
  4.1291 -+  {
  4.1292 +   {
  4.1293 +-   printk(KERN_ERR "%s: out of kernel mem for offsets\n", cloop_name);
  4.1294  +   printk(KERN_ERR "%s: out of mem for zlib working area %u\n",
  4.1295  +          cloop_name, zlib_inflate_workspacesize());
  4.1296      error=-ENOMEM; goto error_release;
  4.1297     }
  4.1298 -- total_offsets = 1; /* Dummy total_offsets: will be filled in first time around */
  4.1299 -- for (i = 0, offsets_read = 0; offsets_read < total_offsets; i++)
  4.1300 +- /* Read them offsets! */
  4.1301 +- if(clo->header_first)
  4.1302  + zlib_inflateInit(&clo->zstream);
  4.1303  +#endif
  4.1304  +#if (defined(CONFIG_DECOMPRESS_XZ) || defined(CONFIG_DECOMPRESS_XZ_MODULE))
  4.1305 @@ -890,33 +713,20 @@
  4.1306  +#endif
  4.1307  + if (total_offsets + 1 == 0) /* Version 3 */
  4.1308     {
  4.1309 --   unsigned int offset = 0, num_readable;
  4.1310 --   size_t bytes_read = cloop_read_from_file(clo, file, bbuf,
  4.1311 --                                          i*clo->underlying_blksize,
  4.1312 --                                          clo->underlying_blksize);
  4.1313 --   if(bytes_read != clo->underlying_blksize)
  4.1314 +-   fs_read_position = sizeof(struct cloop_head);
  4.1315  +   struct cloop_tail tail;
  4.1316  +   if (isblkdev)
  4.1317 -     {
  4.1318 --     printk(KERN_ERR "%s: Bad file, read() of first %lu bytes returned %d.\n",
  4.1319 --                   cloop_name, (unsigned long)clo->underlying_blksize, (int)bytes_read);
  4.1320 --     error=-EBADF;
  4.1321 --     goto error_release;
  4.1322 ++    {
  4.1323  +    /* No end of file: can't find index */
  4.1324  +     printk(KERN_ERR "%s: no V3 support for block device\n", 
  4.1325  +            cloop_name);
  4.1326  +     error=-EBADF; goto error_release;
  4.1327 -     }
  4.1328 --   /* Header will be in block zero */
  4.1329 --   if(i==0)
  4.1330 ++    }
  4.1331  +   bytes_read = cloop_read_from_file(clo, file, (void *) &tail,
  4.1332  +			inode->i_size - sizeof(struct cloop_tail),
  4.1333  +			sizeof(struct cloop_tail));
  4.1334  +   if (bytes_read == sizeof(struct cloop_tail))
  4.1335 -     {
  4.1336 --     memcpy(&clo->head, bbuf, sizeof(struct cloop_head));
  4.1337 --     offset = sizeof(struct cloop_head);
  4.1338 --     if (ntohl(clo->head.block_size) % 512 != 0)
  4.1339 ++    {
  4.1340  +     unsigned long len, zlen;
  4.1341  +     int ret;
  4.1342  +     void *zbuf;
  4.1343 @@ -926,79 +736,47 @@
  4.1344  +     zlen = ntohl(tail.table_size);
  4.1345  +     zbuf = cloop_malloc(zlen);
  4.1346  +     if (!clo->block_ptrs || !zbuf)
  4.1347 -       {
  4.1348 --       printk(KERN_ERR "%s: blocksize %u not multiple of 512\n",
  4.1349 --              cloop_name, ntohl(clo->head.block_size));
  4.1350 --       error=-EBADF; goto error_release;
  4.1351 --      }
  4.1352 --     if (clo->head.preamble[0x0B]!='V'||clo->head.preamble[0x0C]<'1')
  4.1353 --      {
  4.1354 --       printk(KERN_ERR "%s: Cannot read old 32-bit (version 0.68) images, "
  4.1355 --		       "please use an older version of %s for this file.\n",
  4.1356 --		       cloop_name, cloop_name);
  4.1357 --       error=-EBADF; goto error_release;
  4.1358 ++      {
  4.1359  +       printk(KERN_ERR "%s: out of kernel mem for index\n", cloop_name);
  4.1360  +       error=-ENOMEM; goto error_release;
  4.1361 -       }
  4.1362 --     if (clo->head.preamble[0x0C]<'2')
  4.1363 ++      }
  4.1364  +     bytes_read = cloop_read_from_file(clo, file, zbuf,
  4.1365  +			inode->i_size - zlen - sizeof(struct cloop_tail),
  4.1366  +			zlen);
  4.1367  +     if (bytes_read != zlen)
  4.1368 -       {
  4.1369 --       printk(KERN_ERR "%s: Cannot read old architecture-dependent "
  4.1370 --		       "(format <= 1.0) images, please use an older "
  4.1371 --		       "version of %s for this file.\n",
  4.1372 --		       cloop_name, cloop_name);
  4.1373 ++      {
  4.1374  +       printk(KERN_ERR "%s: can't read index\n", cloop_name);
  4.1375 -        error=-EBADF; goto error_release;
  4.1376 -       }
  4.1377 --     total_offsets=ntohl(clo->head.num_blocks)+1;
  4.1378 --     if (!isblkdev && (sizeof(struct cloop_head)+sizeof(loff_t)*
  4.1379 --                       total_offsets > inode->i_size))
  4.1380 ++       error=-EBADF; goto error_release;
  4.1381 ++      }
  4.1382  +     len = CLOOP3_INDEX_SIZE(ntohl(tail.index_size)) * total_offsets;
  4.1383 -+     flags = CLOOP3_BLOCKS_FLAGS(ntohl(tail.index_size));
  4.1384 -+// May  3 19:45:20 (none) user.info kernel: cloop: uncompress(clo=e0a78000, block_ptrs=e0c9c000, &len(1440)=ddc05e6c, zbuf=e0c9f000, zlen=43, flag=0)
  4.1385 -+printk(KERN_INFO "%s: uncompress(clo=%p, block_ptrs=%p, &len(%ld)=%p, zbuf=%p, zlen=%ld, flag=%d)\n", cloop_name, 
  4.1386 -+		clo, clo->block_ptrs, len, &len, zbuf, zlen, flags);
  4.1387 -+     ret = uncompress(clo, (void *) clo->block_ptrs, &len, zbuf, zlen, flags);
  4.1388 -+// May  3 19:45:20 (none) user.alert kernel: BUG: unable to handle kernel NULL pointer dereference at   (null)
  4.1389 -+printk(KERN_INFO "%s: uncompressed !\n", cloop_name);
  4.1390 ++     ret = uncompress(clo, (void *) clo->block_ptrs, &len, zbuf, zlen, CLOOP_COMPRESSOR_ZLIB);
  4.1391  +     cloop_free(zbuf, zlen);
  4.1392  +     if (ret != 0)
  4.1393 -       {
  4.1394 --       printk(KERN_ERR "%s: file too small for %u blocks\n",
  4.1395 --              cloop_name, ntohl(clo->head.num_blocks));
  4.1396 -+        printk(KERN_ERR "%s: decompression error %i uncompressing index, flags %u\n",
  4.1397 -+               cloop_name, ret, flags);
  4.1398 -        error=-EBADF; goto error_release;
  4.1399 -       }
  4.1400 --     clo->offsets = cloop_malloc(sizeof(loff_t) * total_offsets);
  4.1401 --     if (!clo->offsets)
  4.1402 --      {
  4.1403 --       printk(KERN_ERR "%s: out of kernel mem for offsets\n", cloop_name);
  4.1404 --       error=-ENOMEM; goto error_release;
  4.1405 --      }
  4.1406 -     }
  4.1407 --   num_readable = MIN(total_offsets - offsets_read,
  4.1408 --                      (clo->underlying_blksize - offset) 
  4.1409 --                      / sizeof(loff_t));
  4.1410 --   memcpy(&clo->offsets[offsets_read], bbuf+offset, num_readable * sizeof(loff_t));
  4.1411 --   offsets_read += num_readable;
  4.1412 --  }
  4.1413 --  { /* Search for largest block rather than estimate. KK. */
  4.1414 --   int i;
  4.1415 --   for(i=0;i<total_offsets-1;i++)
  4.1416 ++      {
  4.1417 ++        printk(KERN_ERR "%s: decompression error %i uncompressing index\n",
  4.1418 ++               cloop_name, ret);
  4.1419 ++       error=-EBADF; goto error_release;
  4.1420 ++      }
  4.1421 ++    }
  4.1422  +   else
  4.1423  +    {
  4.1424  +     printk(KERN_ERR "%s: can't find index\n", cloop_name);
  4.1425  +     error=-ENOMEM; goto error_release;
  4.1426  +    }
  4.1427 -+  }
  4.1428 -+ else
  4.1429 -+  {
  4.1430 +   }
  4.1431 +  else
  4.1432 +   {
  4.1433 +-   fs_read_position = clo->underlying_total_size - sizeof(struct cloop_head) - total_offsets * sizeof(loff_t);
  4.1434 +-  }
  4.1435 +- for(offsets_read=0;offsets_read<total_offsets;)
  4.1436 +-  {
  4.1437 +-   size_t bytes_readable;
  4.1438 +-   unsigned int num_readable, offset = 0;
  4.1439 +-   bytes_readable = MIN(bbuf_size, clo->underlying_total_size - fs_read_position);
  4.1440 +-   if(bytes_readable <= 0) break; /* Done */
  4.1441 +-   bytes_read = cloop_read_from_file(clo, file, bbuf, fs_read_position, bytes_readable);
  4.1442 +-   if(bytes_read != bytes_readable)
  4.1443  +   unsigned int n, total_bytes;
  4.1444 -+   flags = 0;
  4.1445  +   clo->block_ptrs = cloop_malloc(sizeof(struct block_info) * total_offsets);
  4.1446  +   if (!clo->block_ptrs)
  4.1447  +    {
  4.1448 @@ -1007,14 +785,26 @@
  4.1449  +    }
  4.1450  +   /* Read them offsets! */
  4.1451  +   if(clo->header_first)
  4.1452 -+    {
  4.1453 +     {
  4.1454 +-     printk(KERN_ERR "%s: Bad file %s, read() %lu bytes @ %llu returned %d.\n",
  4.1455 +-            cloop_name, clo->underlying_filename, (unsigned long)clo->underlying_blksize, fs_read_position, (int)bytes_read);
  4.1456 +-     error=-EBADF;
  4.1457 +-     goto error_release;
  4.1458  +     total_bytes = total_offsets * sizeof(struct block_info);
  4.1459  +     fs_read_position = sizeof(struct cloop_head);
  4.1460 -+    }
  4.1461 +     }
  4.1462 +-   /* remember where to read the next blk from file */
  4.1463 +-   fs_read_position += bytes_read;
  4.1464 +-   /* calculate how many offsets can be taken from current bbuf */
  4.1465 +-   num_readable = MIN(total_offsets - offsets_read,
  4.1466 +-                      bytes_read / sizeof(loff_t));
  4.1467 +-   DEBUGP(KERN_INFO "cloop: parsing %d offsets %d to %d\n", num_readable, offsets_read, offsets_read+num_readable-1);
  4.1468 +-   for (i=0,offset=0; i<num_readable; i++)
  4.1469  +   else
  4.1470       {
  4.1471 --     loff_t d=be64_to_cpu(clo->offsets[i+1]) - be64_to_cpu(clo->offsets[i]);
  4.1472 --     clo->largest_block=MAX(clo->largest_block,d);
  4.1473 +-     loff_t tmp = be64_to_cpu( *(loff_t*) (bbuf+offset) );
  4.1474 +-     if (i%50==0) DEBUGP(KERN_INFO "cloop: offset %03d: %llu\n", offsets_read, tmp);
  4.1475 +-     if(offsets_read > 0)
  4.1476  +     total_bytes = total_offsets * sizeof(loff_t);
  4.1477  +     fs_read_position = clo->underlying_total_size - sizeof(struct cloop_head) - total_bytes;
  4.1478  +    }
  4.1479 @@ -1025,35 +815,28 @@
  4.1480  +     if(bytes_readable <= 0) break; /* Done */
  4.1481  +     bytes_read = cloop_read_from_file(clo, file, bbuf, fs_read_position, bytes_readable);
  4.1482  +     if(bytes_read != bytes_readable)
  4.1483 -+      {
  4.1484 +       {
  4.1485 +-       loff_t d = CLOOP_BLOCK_OFFSET(tmp) - CLOOP_BLOCK_OFFSET(clo->block_ptrs[offsets_read-1]);
  4.1486 +-       if(d > clo->largest_block) clo->largest_block = d;
  4.1487  +       printk(KERN_ERR "%s: Bad file %s, read() %lu bytes @ %llu returned %d.\n",
  4.1488  +              cloop_name, clo->underlying_filename, (unsigned long)clo->underlying_blksize, fs_read_position, (int)bytes_read);
  4.1489  +       error=-EBADF;
  4.1490  +       goto error_release;
  4.1491 -+      }
  4.1492 +       }
  4.1493 +-     clo->block_ptrs[offsets_read++] = tmp;
  4.1494 +-     offset += sizeof(loff_t);
  4.1495  +     memcpy(((char *)clo->block_ptrs) + n, bbuf, bytes_read);
  4.1496  +     /* remember where to read the next blk from file */
  4.1497  +     fs_read_position += bytes_read;
  4.1498  +     n += bytes_read;
  4.1499       }
  4.1500 --   printk(KERN_INFO "%s: %s: %u blocks, %u bytes/block, largest block is %lu bytes.\n",
  4.1501 --          cloop_name, filename, ntohl(clo->head.num_blocks),
  4.1502 --          ntohl(clo->head.block_size), clo->largest_block);
  4.1503     }
  4.1504 --/* Combo kmalloc used too large chunks (>130000). */
  4.1505 +-  printk(KERN_INFO "%s: %s: %u blocks, %u bytes/block, largest block is %lu bytes.\n",
  4.1506 +-         cloop_name, clo->underlying_filename, clo->head.num_blocks,
  4.1507 +-         clo->head.block_size, clo->largest_block);
  4.1508    {
  4.1509     int i;
  4.1510 --  for(i=0;i<BUFFERED_BLOCKS;i++)
  4.1511 --   {
  4.1512 --    clo->buffer[i] = cloop_malloc(ntohl(clo->head.block_size));
  4.1513 --    if(!clo->buffer[i])
  4.1514 --     {
  4.1515 --      printk(KERN_ERR "%s: out of memory for buffer %lu\n",
  4.1516 --             cloop_name, (unsigned long) ntohl(clo->head.block_size));
  4.1517 --      error=-ENOMEM; goto error_release_free;
  4.1518 --     }
  4.1519 --   }
  4.1520 -+  char *version = build_index(clo->block_ptrs, clo->head.num_blocks, clo->head.block_size, flags);
  4.1521 ++  char *version = build_index(clo->block_ptrs, clo->head.num_blocks, clo->head.block_size);
  4.1522  +  clo->largest_block = 0;
  4.1523  +  for (i = 0; i < clo->head.num_blocks; i++)
  4.1524  +    if (clo->block_ptrs[i].size > clo->largest_block)
  4.1525 @@ -1061,39 +844,15 @@
  4.1526  +  printk(KERN_INFO "%s: %s: %s: %u blocks, %u bytes/block, largest block is %lu bytes.\n",
  4.1527  +         cloop_name, clo->underlying_filename, version, clo->head.num_blocks,
  4.1528  +         clo->head.block_size, clo->largest_block);
  4.1529 -+ }
  4.1530 -+ {
  4.1531 -+  int i;
  4.1532 -+  clo->num_buffered_blocks = (buffers > 0 && clo->head.block_size >= 512) ?
  4.1533 -+                              (buffers / clo->head.block_size) : 1;
  4.1534 -+  clo->buffered_blocknum = cloop_malloc(clo->num_buffered_blocks * sizeof (u_int32_t));
  4.1535 -+  clo->buffer = cloop_malloc(clo->num_buffered_blocks * sizeof (char*));
  4.1536 -+  if (!clo->buffered_blocknum || !clo->buffer)
  4.1537 -+  {
  4.1538 -+   printk(KERN_ERR "%s: out of memory for index of cache buffer (%lu bytes)\n",
  4.1539 -+                    cloop_name, (unsigned long)clo->num_buffered_blocks * sizeof (u_int32_t) + sizeof(char*) );
  4.1540 -+                    error=-ENOMEM; goto error_release;
  4.1541 -+  }
  4.1542 -+  memset(clo->buffer, 0, clo->num_buffered_blocks * sizeof (char*));
  4.1543 -+  for(i=0;i<clo->num_buffered_blocks;i++)
  4.1544 -+  {
  4.1545 -+   clo->buffered_blocknum[i] = -1;
  4.1546 -+   clo->buffer[i] = cloop_malloc(clo->head.block_size);
  4.1547 -+   if(!clo->buffer[i])
  4.1548 -+    {
  4.1549 -+     printk(KERN_ERR "%s: out of memory for cache buffer %lu\n",
  4.1550 -+            cloop_name, (unsigned long) clo->head.block_size);
  4.1551 -+     error=-ENOMEM; goto error_release_free;
  4.1552 -+    }
  4.1553 -+  }
  4.1554 -+  clo->current_bufnum = 0;
  4.1555 -  }
  4.1556 -  clo->compressed_buffer = cloop_malloc(clo->largest_block);
  4.1557 -  if(!clo->compressed_buffer)
  4.1558 -@@ -557,31 +803,7 @@
  4.1559 +   clo->num_buffered_blocks = (buffers > 0 && clo->head.block_size >= 512) ?
  4.1560 +                               (buffers / clo->head.block_size) : 1;
  4.1561 +   clo->buffered_blocknum = cloop_malloc(clo->num_buffered_blocks * sizeof (u_int32_t));
  4.1562 +@@ -705,36 +793,14 @@
  4.1563             cloop_name, clo->largest_block);
  4.1564      error=-ENOMEM; goto error_release_free_buffer;
  4.1565     }
  4.1566 +- /* Allocate Memory for decompressors */
  4.1567 +-#if (defined(CONFIG_ZLIB_INFLATE) || defined(CONFIG_ZLIB_INFLATE_MODULE))
  4.1568  - clo->zstream.workspace = cloop_malloc(zlib_inflate_workspacesize());
  4.1569  - if(!clo->zstream.workspace)
  4.1570  -  {
  4.1571 @@ -1102,443 +861,48 @@
  4.1572  -   error=-ENOMEM; goto error_release_free_all;
  4.1573  -  }
  4.1574  - zlib_inflateInit(&clo->zstream);
  4.1575 -- if(!isblkdev &&
  4.1576 --    be64_to_cpu(clo->offsets[ntohl(clo->head.num_blocks)]) != inode->i_size)
  4.1577 --  {
  4.1578 --   printk(KERN_ERR "%s: final offset wrong (%Lu not %Lu)\n",
  4.1579 +-#endif
  4.1580 +-#if (defined(CONFIG_DECOMPRESS_XZ) || defined(CONFIG_DECOMPRESS_XZ_MODULE))
  4.1581 +-#if XZ_INTERNAL_CRC32
  4.1582 +-  /* This must be called before any other xz_* function to initialize the CRC32 lookup table. */
  4.1583 +-  xz_crc32_init(void);
  4.1584 +-#endif
  4.1585 +-  clo->xzdecoderstate = xz_dec_init(XZ_SINGLE, 0);
  4.1586 +-#endif
  4.1587 +- if(CLOOP_BLOCK_OFFSET(clo->block_ptrs[clo->head.num_blocks]) > clo->underlying_total_size)
  4.1588 ++ set_capacity(clo->clo_disk, (sector_t)(clo->head.num_blocks*(clo->head.block_size>>9)));
  4.1589 ++ clo->clo_thread = kthread_create(cloop_thread, clo, "cloop%d", cloop_num);
  4.1590 ++ if(IS_ERR(clo->clo_thread))
  4.1591 +   {
  4.1592 +-   printk(KERN_ERR "%s: final offset wrong (%llu > %llu)\n",
  4.1593  -          cloop_name,
  4.1594 --          be64_to_cpu(clo->offsets[ntohl(clo->head.num_blocks)]),
  4.1595 --          inode->i_size);
  4.1596 +-	  CLOOP_BLOCK_OFFSET(clo->block_ptrs[clo->head.num_blocks]),
  4.1597 +-          clo->underlying_total_size);
  4.1598 +-#if (defined(CONFIG_ZLIB_INFLATE) || defined(CONFIG_ZLIB_INFLATE_MODULE))
  4.1599  -   cloop_free(clo->zstream.workspace, zlib_inflate_workspacesize()); clo->zstream.workspace=NULL;
  4.1600 --   goto error_release_free_all;
  4.1601 --  }
  4.1602 -- {
  4.1603 --  int i;
  4.1604 --  for(i=0; i<BUFFERED_BLOCKS; i++) clo->buffered_blocknum[i] = -1;
  4.1605 --  clo->current_bufnum=0;
  4.1606 -- }
  4.1607 -- set_capacity(clo->clo_disk, (sector_t)(ntohl(clo->head.num_blocks)*
  4.1608 --              (ntohl(clo->head.block_size)>>9)));
  4.1609 -+ set_capacity(clo->clo_disk, (sector_t)(clo->head.num_blocks*(clo->head.block_size>>9)));
  4.1610 -  clo->clo_thread = kthread_create(cloop_thread, clo, "cloop%d", cloop_num);
  4.1611 -  if(IS_ERR(clo->clo_thread))
  4.1612 -   {
  4.1613 -@@ -591,17 +813,17 @@
  4.1614 +-#endif
  4.1615 ++   error = PTR_ERR(clo->clo_thread);
  4.1616 ++   clo->clo_thread=NULL;
  4.1617 +    goto error_release_free_all;
  4.1618     }
  4.1619 +- set_capacity(clo->clo_disk, (sector_t)(clo->head.num_blocks*(clo->head.block_size>>9)));
  4.1620    if(preload > 0)
  4.1621     {
  4.1622 --   clo->preload_array_size = ((preload<=ntohl(clo->head.num_blocks))?preload:ntohl(clo->head.num_blocks));
  4.1623 -+   clo->preload_array_size = ((preload<=clo->head.num_blocks)?preload:clo->head.num_blocks);
  4.1624 -    clo->preload_size = 0;
  4.1625 -    if((clo->preload_cache = cloop_malloc(clo->preload_array_size * sizeof(char *))) != NULL)
  4.1626 -     {
  4.1627 -      int i;
  4.1628 -      for(i=0; i<clo->preload_array_size; i++)
  4.1629 -       {
  4.1630 --       if((clo->preload_cache[i] = cloop_malloc(ntohl(clo->head.block_size))) == NULL)
  4.1631 -+       if((clo->preload_cache[i] = cloop_malloc(clo->head.block_size)) == NULL)
  4.1632 -         { /* Out of memory */
  4.1633 -          printk(KERN_WARNING "%s: cloop_malloc(%d) failed for preload_cache[%d] (ignored).\n",
  4.1634 --                             cloop_name, ntohl(clo->head.block_size), i);
  4.1635 -+                             cloop_name, clo->head.block_size, i);
  4.1636 - 	 break;
  4.1637 - 	}
  4.1638 -       }
  4.1639 -@@ -612,13 +834,13 @@
  4.1640 -        if(buffered_blocknum >= 0)
  4.1641 -         {
  4.1642 - 	 memcpy(clo->preload_cache[i], clo->buffer[buffered_blocknum],
  4.1643 --	        ntohl(clo->head.block_size));
  4.1644 -+	        clo->head.block_size);
  4.1645 - 	}
  4.1646 -        else
  4.1647 -         {
  4.1648 -          printk(KERN_WARNING "%s: can't read block %d into preload cache, set to zero.\n",
  4.1649 - 	                     cloop_name, i);
  4.1650 --	 memset(clo->preload_cache[i], 0, ntohl(clo->head.block_size));
  4.1651 -+	 memset(clo->preload_cache[i], 0, clo->head.block_size);
  4.1652 - 	}
  4.1653 -       }
  4.1654 -      printk(KERN_INFO "%s: preloaded %d blocks into cache.\n", cloop_name,
  4.1655 -@@ -641,22 +863,19 @@
  4.1656 -  cloop_free(clo->compressed_buffer, clo->largest_block);
  4.1657 -  clo->compressed_buffer=NULL;
  4.1658 - error_release_free_buffer:
  4.1659 -+ if(clo->buffer)
  4.1660 -  {
  4.1661 -   int i;
  4.1662 --  for(i=0; i<BUFFERED_BLOCKS; i++)
  4.1663 --   { 
  4.1664 --    if(clo->buffer[i])
  4.1665 --     {
  4.1666 --      cloop_free(clo->buffer[i], ntohl(clo->head.block_size));
  4.1667 --      clo->buffer[i]=NULL;
  4.1668 --     }
  4.1669 --   }
  4.1670 -+  for(i=0; i<clo->num_buffered_blocks; i++) { if(clo->buffer[i]) { cloop_free(clo->buffer[i], clo->head.block_size); clo->buffer[i]=NULL; }}
  4.1671 -+  cloop_free(clo->buffer, clo->num_buffered_blocks*sizeof(char*)); clo->buffer=NULL;
  4.1672 +    clo->preload_array_size = ((preload<=clo->head.num_blocks)?preload:clo->head.num_blocks);
  4.1673 +@@ -780,6 +846,7 @@
  4.1674 +      clo->preload_array_size = clo->preload_size = 0;
  4.1675 +     }
  4.1676 +   }
  4.1677 ++ wake_up_process(clo->clo_thread);
  4.1678 +  /* Uncheck */
  4.1679 +  return error;
  4.1680 + error_release_free_all:
  4.1681 +@@ -794,9 +861,13 @@
  4.1682    }
  4.1683 -+ if (clo->buffered_blocknum) { cloop_free(clo->buffered_blocknum, sizeof(int)*clo->num_buffered_blocks); clo->buffered_blocknum=NULL; }
  4.1684 +  if (clo->buffered_blocknum) { cloop_free(clo->buffered_blocknum, sizeof(int)*clo->num_buffered_blocks); clo->buffered_blocknum=NULL; }
  4.1685   error_release_free:
  4.1686 -- cloop_free(clo->offsets, sizeof(loff_t) * total_offsets);
  4.1687 -- clo->offsets=NULL;
  4.1688 +- cloop_free(clo->block_ptrs, sizeof(cloop_block_ptr) * total_offsets);
  4.1689  + cloop_free(clo->block_ptrs, sizeof(struct block_info) * total_offsets);
  4.1690 -+ clo->block_ptrs=NULL;
  4.1691 - error_release:
  4.1692 -  if(bbuf) cloop_free(bbuf, clo->underlying_blksize);
  4.1693 -+ if(clo->underlying_filename) { kfree(clo->underlying_filename); clo->underlying_filename=NULL; }
  4.1694 -  clo->backing_file=NULL;
  4.1695 -  return error;
  4.1696 - }
  4.1697 -@@ -673,7 +892,7 @@
  4.1698 -  if(clo->backing_file) return -EBUSY;
  4.1699 -  file = fget(arg); /* get filp struct from ioctl arg fd */
  4.1700 -  if(!file) return -EBADF;
  4.1701 -- error=cloop_set_file(cloop_num,file,"losetup_file");
  4.1702 -+ error=cloop_set_file(cloop_num,file);
  4.1703 -  set_device_ro(bdev, 1);
  4.1704 -  if(error) fput(file);
  4.1705 -  return error;
  4.1706 -@@ -684,29 +903,48 @@
  4.1707 - {
  4.1708 -  struct cloop_device *clo = cloop_dev[cloop_num];
  4.1709 -  struct file *filp = clo->backing_file;
  4.1710 -- int i;
  4.1711 -  if(clo->refcnt > 1)	/* we needed one fd for the ioctl */
  4.1712 -    return -EBUSY;
  4.1713 -  if(filp==NULL) return -EINVAL;
  4.1714 -  if(clo->clo_thread) { kthread_stop(clo->clo_thread); clo->clo_thread=NULL; }
  4.1715 -- if(filp!=initial_file) fput(filp);
  4.1716 -- else { filp_close(initial_file,0); initial_file=NULL; }
  4.1717 -+ if(filp!=initial_file)
  4.1718 -+  fput(filp);
  4.1719 -+ else
  4.1720 -+ {
  4.1721 -+  filp_close(initial_file,0);
  4.1722 -+  initial_file=NULL;
  4.1723 -+ }
  4.1724 -  clo->backing_file  = NULL;
  4.1725 -  clo->backing_inode = NULL;
  4.1726 -- if(clo->offsets) { cloop_free(clo->offsets, clo->underlying_blksize); clo->offsets = NULL; }
  4.1727 -+ if(clo->underlying_filename) { kfree(clo->underlying_filename); clo->underlying_filename=NULL; }
  4.1728 -+ if(clo->block_ptrs) { cloop_free(clo->block_ptrs, clo->head.num_blocks); clo->block_ptrs = NULL; }
  4.1729 -  if(clo->preload_cache)
  4.1730 --  {
  4.1731 --   for(i=0; i < clo->preload_size; i++)
  4.1732 --    cloop_free(clo->preload_cache[i], ntohl(clo->head.block_size));
  4.1733 --   cloop_free(clo->preload_cache, clo->preload_array_size * sizeof(char *));
  4.1734 --   clo->preload_cache = NULL;
  4.1735 --   clo->preload_size = clo->preload_array_size = 0;
  4.1736 --  }
  4.1737 -- for(i=0; i<BUFFERED_BLOCKS; i++)
  4.1738 --      if(clo->buffer[i]) { cloop_free(clo->buffer[i], ntohl(clo->head.block_size)); clo->buffer[i]=NULL; }
  4.1739 -+ {
  4.1740 -+  int i;
  4.1741 -+  for(i=0; i < clo->preload_size; i++)
  4.1742 -+   cloop_free(clo->preload_cache[i], clo->head.block_size);
  4.1743 -+  cloop_free(clo->preload_cache, clo->preload_array_size * sizeof(char *));
  4.1744 -+  clo->preload_cache = NULL;
  4.1745 -+  clo->preload_size = clo->preload_array_size = 0;
  4.1746 -+ }
  4.1747 -+ if (clo->buffered_blocknum)
  4.1748 -+ {
  4.1749 -+  cloop_free(clo->buffered_blocknum, sizeof(int) * clo->num_buffered_blocks); clo->buffered_blocknum = NULL;
  4.1750 -+ }
  4.1751 -+ if (clo->buffer)
  4.1752 -+ {
  4.1753 -+  int i;
  4.1754 -+  for(i=0; i<clo->num_buffered_blocks; i++) { if(clo->buffer[i]) cloop_free(clo->buffer[i], clo->head.block_size); }
  4.1755 -+  cloop_free(clo->buffer, sizeof(char*) * clo->num_buffered_blocks); clo->buffer = NULL;
  4.1756 -+ }
  4.1757 -  if(clo->compressed_buffer) { cloop_free(clo->compressed_buffer, clo->largest_block); clo->compressed_buffer = NULL; }
  4.1758 -+#if (defined(CONFIG_ZLIB_INFLATE) || defined(CONFIG_ZLIB_INFLATE_MODULE))
  4.1759 -  zlib_inflateEnd(&clo->zstream);
  4.1760 -  if(clo->zstream.workspace) { cloop_free(clo->zstream.workspace, zlib_inflate_workspacesize()); clo->zstream.workspace = NULL; }
  4.1761 -+#endif
  4.1762 -+#if (defined(CONFIG_DECOMPRESS_XZ) || defined(CONFIG_DECOMPRESS_XZ_MODULE))
  4.1763 -+  xz_dec_end(clo->xzdecoderstate);
  4.1764 -+#endif
  4.1765 -  if(bdev) invalidate_bdev(bdev);
  4.1766 -  if(clo->clo_disk) set_capacity(clo->clo_disk, 0);
  4.1767 -  return 0;
  4.1768 -@@ -731,8 +969,8 @@
  4.1769 -                             const struct loop_info64 *info)
  4.1770 - {
  4.1771 -  if (!clo->backing_file) return -ENXIO;
  4.1772 -- memcpy(clo->clo_file_name, info->lo_file_name, LO_NAME_SIZE);
  4.1773 -- clo->clo_file_name[LO_NAME_SIZE-1] = 0;
  4.1774 -+ if(clo->underlying_filename) kfree(clo->underlying_filename);
  4.1775 -+ clo->underlying_filename = kstrdup(info->lo_file_name, GFP_KERNEL);
  4.1776 -  return 0;
  4.1777 - }
  4.1778 - 
  4.1779 -@@ -743,7 +981,11 @@
  4.1780 -  struct kstat stat;
  4.1781 -  int err;
  4.1782 -  if (!file) return -ENXIO;
  4.1783 -- err = vfs_getattr(file->f_path.mnt, file->f_path.dentry, &stat);
  4.1784 -+#if LINUX_VERSION_CODE < KERNEL_VERSION(4, 11, 0)
  4.1785 -+ err = vfs_getattr(&file->f_path, &stat);
  4.1786 -+#else
  4.1787 -+ err = vfs_getattr(&file->f_path, &stat, STATX_INO, AT_STATX_SYNC_AS_STAT);
  4.1788 -+#endif
  4.1789 -  if (err) return err;
  4.1790 -  memset(info, 0, sizeof(*info));
  4.1791 -  info->lo_number  = clo->clo_number;
  4.1792 -@@ -753,7 +995,8 @@
  4.1793 -  info->lo_offset  = 0;
  4.1794 -  info->lo_sizelimit = 0;
  4.1795 -  info->lo_flags   = 0;
  4.1796 -- memcpy(info->lo_file_name, clo->clo_file_name, LO_NAME_SIZE);
  4.1797 -+ strncpy(info->lo_file_name, clo->underlying_filename, LO_NAME_SIZE);
  4.1798 -+ info->lo_file_name[LO_NAME_SIZE-1]=0;
  4.1799 -  return 0;
  4.1800 - }
  4.1801 - 
  4.1802 -@@ -833,8 +1076,6 @@
  4.1803 -  if (!err && copy_to_user(arg, &info64, sizeof(info64))) err = -EFAULT;
  4.1804 -  return err;
  4.1805 - }
  4.1806 --/* EOF get/set_status */
  4.1807 --
  4.1808 - 
  4.1809 - static int cloop_ioctl(struct block_device *bdev, fmode_t mode,
  4.1810 - 	unsigned int cmd, unsigned long arg)
  4.1811 -@@ -914,21 +1155,20 @@
  4.1812 -  /* losetup uses write-open and flags=0x8002 to set a new file */
  4.1813 -  if(mode & FMODE_WRITE)
  4.1814 -   {
  4.1815 --   printk(KERN_WARNING "%s: Can't open device read-write in mode 0x%x\n", cloop_name, mode);
  4.1816 -+   printk(KERN_INFO "%s: Open in read-write mode 0x%x requested, ignored.\n", cloop_name, mode);
  4.1817 -    return -EROFS;
  4.1818 -   }
  4.1819 -  cloop_dev[cloop_num]->refcnt+=1;
  4.1820 -  return 0;
  4.1821 - }
  4.1822 - 
  4.1823 --static int cloop_close(struct gendisk *disk, fmode_t mode)
  4.1824 -+static void cloop_close(struct gendisk *disk, fmode_t mode)
  4.1825 - {
  4.1826 -- int cloop_num, err=0;
  4.1827 -- if(!disk) return 0;
  4.1828 -+ int cloop_num;
  4.1829 -+ if(!disk) return;
  4.1830 -  cloop_num=((struct cloop_device *)disk->private_data)->clo_number;
  4.1831 -- if(cloop_num < 0 || cloop_num > (cloop_count-1)) return 0;
  4.1832 -+ if(cloop_num < 0 || cloop_num > (cloop_count-1)) return;
  4.1833 -  cloop_dev[cloop_num]->refcnt-=1;
  4.1834 -- return err;
  4.1835 - }
  4.1836 - 
  4.1837 - static struct block_device_operations clo_fops =
  4.1838 -@@ -973,6 +1213,10 @@
  4.1839 -    goto error_out;
  4.1840 -   }
  4.1841 -  clo->clo_queue->queuedata = clo;
  4.1842 -+#if LINUX_VERSION_CODE < KERNEL_VERSION(4, 12, 0)
  4.1843 -+ queue_flag_set_unlocked(QUEUE_FLAG_NONROT, clo->clo_queue);
  4.1844 -+ queue_flag_set_unlocked(QUEUE_FLAG_NOMERGES, clo->clo_queue);
  4.1845 -+#endif
  4.1846 -  clo->clo_disk = alloc_disk(1);
  4.1847 -  if(!clo->clo_disk)
  4.1848 -   {
  4.1849 -@@ -1004,6 +1248,11 @@
  4.1850 -  cloop_dev[cloop_num] = NULL;
  4.1851 - }
  4.1852 - 
  4.1853 -+/* LZ4 Stuff */
  4.1854 -+#if (defined USE_LZ4_INTERNAL)
  4.1855 -+#include "lz4_kmod.c"
  4.1856 -+#endif
  4.1857 -+
  4.1858 - static int __init cloop_init(void)
  4.1859 - {
  4.1860 -  int error=0;
  4.1861 -@@ -1044,7 +1293,7 @@
  4.1862 -      initial_file=NULL; /* if IS_ERR, it's NOT open. */
  4.1863 -     }
  4.1864 -    else
  4.1865 --     error=cloop_set_file(0,initial_file,file);
  4.1866 -+     error=cloop_set_file(0,initial_file);
  4.1867 -    if(error)
  4.1868 -     {
  4.1869 -      printk(KERN_ERR
  4.1870 -@@ -1052,9 +1301,6 @@
  4.1871 -             cloop_name, file, error);
  4.1872 -      goto init_out_dealloc;
  4.1873 -     }
  4.1874 --   if(namelen >= LO_NAME_SIZE) namelen = LO_NAME_SIZE-1;
  4.1875 --   memcpy(cloop_dev[0]->clo_file_name, file, namelen);
  4.1876 --   cloop_dev[0]->clo_file_name[namelen] = 0;
  4.1877 -   }
  4.1878 -  return 0;
  4.1879 - init_out_dealloc:
  4.1880 ---- cloop.h
  4.1881 -+++ cloop.h
  4.1882 -@@ -86,11 +86,8 @@
  4.1883 - struct cloop_tail
  4.1884 - {
  4.1885 - 	u_int32_t table_size; 
  4.1886 --	u_int32_t index_size; /* size:4 comp:3 ctrl-c:1 lastlen:24 */
  4.1887 -+	u_int32_t index_size; /* size:4 unused:3 ctrl-c:1 lastlen:24 */
  4.1888 - #define CLOOP3_INDEX_SIZE(x)    ((unsigned int)((x) & 0xF))
  4.1889 --#define CLOOP3_BLOCKS_FLAGS(x)  ((unsigned int)((x) & 0x70) >> 4)
  4.1890 --#define CLOOP3_TRUNCATED(x)     ((unsigned int)((x) & 0x80) >> 7)
  4.1891 --#define CLOOP3_LASTLEN(x)       (unsigned int)((x) >> 8)
  4.1892 - 	u_int32_t num_blocks;
  4.1893 - };
  4.1894 - 
  4.1895 -@@ -104,8 +101,10 @@
  4.1896 - };
  4.1897 - 
  4.1898 - static inline char *build_index(struct block_info *offsets, unsigned long n, 
  4.1899 --			unsigned long block_size, unsigned global_flags)
  4.1900 -+			unsigned long block_size)
  4.1901 - {
  4.1902 -+	static char v[11];
  4.1903 -+	u_int32_t flags = 0;
  4.1904 - 	u_int32_t *ofs32 = (u_int32_t *) offsets;
  4.1905 - 	loff_t    *ofs64 = (loff_t *) offsets;
  4.1906 - 
  4.1907 -@@ -130,8 +129,6 @@
  4.1908 - 		}
  4.1909 - 		else { /* V2.0/V4.0 */
  4.1910 - 			loff_t last = CLOOP_BLOCK_OFFSET(__be64_to_cpu(ofs64[n]));
  4.1911 --			u_int32_t flags;
  4.1912 --			static char v4[11];
  4.1913 - 			unsigned long i = n;
  4.1914 - 
  4.1915 - 			for (flags = 0; n-- ;) {
  4.1916 -@@ -149,12 +146,7 @@
  4.1917 - 					offsets[i] = offsets[offsets[i].offset];
  4.1918 - 				}
  4.1919 - 			}
  4.1920 --			strcpy(v4, (char *) "64BE v4.0a");
  4.1921 --			v4[10] = 'a' + ((flags-1) & 0xF);	// compressors used
  4.1922 --			if (flags > 0x10) {			// with links ?
  4.1923 --				v4[10] += 'A' - 'a';
  4.1924 --			}
  4.1925 --			return v4;
  4.1926 -+			strcpy(v, (char *) "64BE v4.0a");
  4.1927 - 		}
  4.1928 - 	}
  4.1929 - 	else if (ofs32[1] == 0 && v3_64 == 0) { /* V1.0 */
  4.1930 -@@ -170,7 +162,6 @@
  4.1931 - 	else { /* V3.0 or V0.68 */
  4.1932 - 		unsigned long i;
  4.1933 - 		loff_t j;
  4.1934 --		static char v3[11];
  4.1935 - 		
  4.1936 - 		for (i = 0; i < n && ntohl(ofs32[i]) < ntohl(ofs32[i+1]); i++);
  4.1937 - 		if (i == n && ntohl(ofs32[0]) == (4*n) + 0x8C) { /* V0.68 */
  4.1938 -@@ -185,28 +176,33 @@
  4.1939 - 		}
  4.1940 - 		
  4.1941 - 		v3_64 = (ofs32[1] == 0);
  4.1942 --		for (i = n; i-- != 0; )
  4.1943 -+		for (i = n; i-- != 0; ) {
  4.1944 - 			offsets[i].size = ntohl(ofs32[i << v3_64]); 
  4.1945 --		for (i = 0, j = sizeof(struct cloop_head); i < n; i++) {
  4.1946 --			offsets[i].offset = j;
  4.1947 --			offsets[i].flags = global_flags;
  4.1948 - 			if (offsets[i].size == 0xFFFFFFFF) {
  4.1949 --				offsets[i].flags = CLOOP_COMPRESSOR_NONE;
  4.1950 --				offsets[i].size = block_size;
  4.1951 -+				offsets[i].size = 0x10000000 | block_size;
  4.1952 - 			}
  4.1953 --			if ((offsets[i].size & 0x80000000) == 0) {
  4.1954 -+			offsets[i].flags = (offsets[i].size >> 28);
  4.1955 -+			offsets[i].size &= 0x0FFFFFFF; 
  4.1956 -+		}
  4.1957 -+		for (i = 0, j = sizeof(struct cloop_head); i < n; i++) {
  4.1958 -+			offsets[i].offset = j;
  4.1959 -+			if (offsets[i].flags < 8) {
  4.1960 - 				j += offsets[i].size;
  4.1961 - 			}
  4.1962 - 		}
  4.1963 - 		for (i = 0; i < n; i++) {
  4.1964 --			if (offsets[i].size & 0x80000000) {
  4.1965 --				offsets[i] = offsets[offsets[i].size & 0x7FFFFFFF];
  4.1966 -+			flags |= 1 << offsets[i].flags;
  4.1967 -+			if (offsets[i].flags >= 8) {
  4.1968 -+				offsets[i] = offsets[offsets[i].size];
  4.1969 - 			}
  4.1970 - 		}
  4.1971 --		strcpy(v3, (char *) (v3_64) ? "64BE v3.0a" : "32BE v3.0a");
  4.1972 --		v3[10] += global_flags;
  4.1973 --		return v3;
  4.1974 -+		strcpy(v, (char *) (v3_64) ? "64BE v3.0a" : "32BE v3.0a");
  4.1975 -+	}
  4.1976 -+	v[10] = 'a' + ((flags-1) & 0xF);	// compressors used
  4.1977 -+	if (flags > 0x10) {			// with links ?
  4.1978 -+		v[10] += 'A' - 'a';
  4.1979 - 	}
  4.1980 -+	return v;
  4.1981 - }
  4.1982 - 
  4.1983 - /* Cloop suspend IOCTL */
  4.1984 ---- cloop.c
  4.1985 -+++ cloop.c
  4.1986 -@@ -542,7 +542,7 @@
  4.1987 -  const unsigned int header_size = sizeof(struct cloop_head);
  4.1988 -  unsigned int i, total_offsets=0;
  4.1989 -  loff_t fs_read_position = 0, header_pos[2];
  4.1990 -- int flags, isblkdev, bytes_read, error = 0;
  4.1991 -+ int isblkdev, bytes_read, error = 0;
  4.1992 -  if (clo->suspended) return error;
  4.1993 -  #if LINUX_VERSION_CODE < KERNEL_VERSION(4, 0, 0)
  4.1994 -  inode = file->f_dentry->d_inode;
  4.1995 -@@ -698,18 +698,12 @@
  4.1996 -        error=-EBADF; goto error_release;
  4.1997 -       }
  4.1998 -      len = CLOOP3_INDEX_SIZE(ntohl(tail.index_size)) * total_offsets;
  4.1999 --     flags = CLOOP3_BLOCKS_FLAGS(ntohl(tail.index_size));
  4.2000 --// May  3 19:45:20 (none) user.info kernel: cloop: uncompress(clo=e0a78000, block_ptrs=e0c9c000, &len(1440)=ddc05e6c, zbuf=e0c9f000, zlen=43, flag=0)
  4.2001 --printk(KERN_INFO "%s: uncompress(clo=%p, block_ptrs=%p, &len(%ld)=%p, zbuf=%p, zlen=%ld, flag=%d)\n", cloop_name, 
  4.2002 --		clo, clo->block_ptrs, len, &len, zbuf, zlen, flags);
  4.2003 --     ret = uncompress(clo, (void *) clo->block_ptrs, &len, zbuf, zlen, flags);
  4.2004 --// May  3 19:45:20 (none) user.alert kernel: BUG: unable to handle kernel NULL pointer dereference at   (null)
  4.2005 --printk(KERN_INFO "%s: uncompressed !\n", cloop_name);
  4.2006 -+     ret = uncompress(clo, (void *) clo->block_ptrs, &len, zbuf, zlen, CLOOP_COMPRESSOR_ZLIB);
  4.2007 -      cloop_free(zbuf, zlen);
  4.2008 -      if (ret != 0)
  4.2009 -       {
  4.2010 --        printk(KERN_ERR "%s: decompression error %i uncompressing index, flags %u\n",
  4.2011 --               cloop_name, ret, flags);
  4.2012 -+        printk(KERN_ERR "%s: decompression error %i uncompressing index\n",
  4.2013 -+               cloop_name, ret);
  4.2014 -        error=-EBADF; goto error_release;
  4.2015 -       }
  4.2016 -     }
  4.2017 -@@ -722,7 +716,6 @@
  4.2018 -  else
  4.2019 -   {
  4.2020 -    unsigned int n, total_bytes;
  4.2021 --   flags = 0;
  4.2022 -    clo->block_ptrs = cloop_malloc(sizeof(struct block_info) * total_offsets);
  4.2023 -    if (!clo->block_ptrs)
  4.2024 -     {
  4.2025 -@@ -761,7 +754,7 @@
  4.2026 -   }
  4.2027 -  {
  4.2028 -   int i;
  4.2029 --  char *version = build_index(clo->block_ptrs, clo->head.num_blocks, clo->head.block_size, flags);
  4.2030 -+  char *version = build_index(clo->block_ptrs, clo->head.num_blocks, clo->head.block_size);
  4.2031 -   clo->largest_block = 0;
  4.2032 -   for (i = 0; i < clo->head.num_blocks; i++)
  4.2033 -     if (clo->block_ptrs[i].size > clo->largest_block)
  4.2034 -@@ -769,9 +762,6 @@
  4.2035 -   printk(KERN_INFO "%s: %s: %s: %u blocks, %u bytes/block, largest block is %lu bytes.\n",
  4.2036 -          cloop_name, clo->underlying_filename, version, clo->head.num_blocks,
  4.2037 -          clo->head.block_size, clo->largest_block);
  4.2038 -- }
  4.2039 -- {
  4.2040 --  int i;
  4.2041 -   clo->num_buffered_blocks = (buffers > 0 && clo->head.block_size >= 512) ?
  4.2042 -                               (buffers / clo->head.block_size) : 1;
  4.2043 -   clo->buffered_blocknum = cloop_malloc(clo->num_buffered_blocks * sizeof (u_int32_t));
  4.2044 -@@ -874,6 +864,10 @@
  4.2045 -  cloop_free(clo->block_ptrs, sizeof(struct block_info) * total_offsets);
  4.2046    clo->block_ptrs=NULL;
  4.2047   error_release:
  4.2048  +#if (defined(CONFIG_ZLIB_INFLATE) || defined(CONFIG_ZLIB_INFLATE_MODULE))
  4.2049 @@ -1548,3 +912,146 @@
  4.2050    if(bbuf) cloop_free(bbuf, clo->underlying_blksize);
  4.2051    if(clo->underlying_filename) { kfree(clo->underlying_filename); clo->underlying_filename=NULL; }
  4.2052    clo->backing_file=NULL;
  4.2053 +@@ -829,6 +900,7 @@
  4.2054 +  if(clo->refcnt > 1)	/* we needed one fd for the ioctl */
  4.2055 +    return -EBUSY;
  4.2056 +  if(filp==NULL) return -EINVAL;
  4.2057 ++ if(clo->clo_thread) { kthread_stop(clo->clo_thread); clo->clo_thread=NULL; }
  4.2058 +  if(filp!=initial_file)
  4.2059 +   fput(filp);
  4.2060 +  else
  4.2061 +@@ -839,7 +911,7 @@
  4.2062 +  clo->backing_file  = NULL;
  4.2063 +  clo->backing_inode = NULL;
  4.2064 +  if(clo->underlying_filename) { kfree(clo->underlying_filename); clo->underlying_filename=NULL; }
  4.2065 +- if(clo->block_ptrs) { cloop_free(clo->block_ptrs, clo->head.num_blocks+1); clo->block_ptrs = NULL; }
  4.2066 ++ if(clo->block_ptrs) { cloop_free(clo->block_ptrs, clo->head.num_blocks); clo->block_ptrs = NULL; }
  4.2067 +  if(clo->preload_cache)
  4.2068 +  {
  4.2069 +   int i;
  4.2070 +@@ -1054,15 +1126,15 @@
  4.2071 +   case LOOP_CLR_FD:       /* Change arg */ 
  4.2072 +   case LOOP_GET_STATUS64: /* Change arg */ 
  4.2073 +   case LOOP_SET_STATUS64: /* Change arg */ 
  4.2074 +-    return cloop_ioctl(bdev, mode, cmd, (unsigned long) compat_ptr(arg));
  4.2075 ++	arg = (unsigned long) compat_ptr(arg);
  4.2076 +   case LOOP_SET_STATUS:   /* unchanged */
  4.2077 +   case LOOP_GET_STATUS:   /* unchanged */
  4.2078 +   case LOOP_SET_FD:       /* unchanged */
  4.2079 +   case LOOP_CHANGE_FD:    /* unchanged */
  4.2080 +-    return cloop_ioctl(bdev, mode, cmd, arg);
  4.2081 +-  default:
  4.2082 +-    return -ENOIOCTLCMD;
  4.2083 ++	return cloop_ioctl(bdev, mode, cmd, arg);
  4.2084 ++	break;
  4.2085 +  }
  4.2086 ++ return -ENOIOCTLCMD;
  4.2087 + }
  4.2088 + #endif
  4.2089 + 
  4.2090 +@@ -1093,7 +1165,7 @@
  4.2091 +  cloop_dev[cloop_num]->refcnt-=1;
  4.2092 + }
  4.2093 + 
  4.2094 +-static const struct block_device_operations clo_fops =
  4.2095 ++static struct block_device_operations clo_fops =
  4.2096 + {
  4.2097 +         owner:		THIS_MODULE,
  4.2098 +         open:           cloop_open,
  4.2099 +@@ -1105,12 +1177,6 @@
  4.2100 + 	/* locked_ioctl ceased to exist in 2.6.36 */
  4.2101 + };
  4.2102 + 
  4.2103 +-static const struct blk_mq_ops cloop_mq_ops = {
  4.2104 +-	.queue_rq       = cloop_queue_rq,
  4.2105 +-/*	.init_request	= cloop_init_request, */
  4.2106 +-/*	.complete	= cloop_complete_rq, */
  4.2107 +-};
  4.2108 +-
  4.2109 + static int cloop_register_blkdev(int major_nr)
  4.2110 + {
  4.2111 +  return register_blkdev(major_nr, cloop_name);
  4.2112 +@@ -1124,37 +1190,33 @@
  4.2113 + 
  4.2114 + static int cloop_alloc(int cloop_num)
  4.2115 + {
  4.2116 +- struct cloop_device *clo = (struct cloop_device *) cloop_malloc(sizeof(struct cloop_device));
  4.2117 ++ struct cloop_device *clo = (struct cloop_device *) cloop_malloc(sizeof(struct cloop_device));;
  4.2118 +  if(clo == NULL) goto error_out;
  4.2119 +  cloop_dev[cloop_num] = clo;
  4.2120 +  memset(clo, 0, sizeof(struct cloop_device));
  4.2121 +  clo->clo_number = cloop_num;
  4.2122 +- clo->tag_set.ops = &cloop_mq_ops;
  4.2123 +- clo->tag_set.nr_hw_queues = 1;
  4.2124 +- clo->tag_set.queue_depth = 128;
  4.2125 +- clo->tag_set.numa_node = NUMA_NO_NODE;
  4.2126 +- clo->tag_set.cmd_size = 0; /* No extra data needed */
  4.2127 +- /* BLK_MQ_F_BLOCKING is extremely important if we want to call blocking functions like vfs_read */
  4.2128 +- clo->tag_set.flags = BLK_MQ_F_SHOULD_MERGE | BLK_MQ_F_BLOCKING;
  4.2129 +- clo->tag_set.driver_data = clo;
  4.2130 +- if(blk_mq_alloc_tag_set(&clo->tag_set)) goto error_out_free_clo;
  4.2131 +- clo->clo_queue = blk_mq_init_queue(&clo->tag_set);
  4.2132 +- if(IS_ERR(clo->clo_queue))
  4.2133 ++ clo->clo_thread = NULL;
  4.2134 ++ init_waitqueue_head(&clo->clo_event);
  4.2135 ++ spin_lock_init(&clo->queue_lock);
  4.2136 ++ mutex_init(&clo->clo_ctl_mutex);
  4.2137 ++ INIT_LIST_HEAD(&clo->clo_list);
  4.2138 ++ clo->clo_queue = blk_init_queue(cloop_do_request, &clo->queue_lock);
  4.2139 ++ if(!clo->clo_queue)
  4.2140 +   {
  4.2141 +    printk(KERN_ERR "%s: Unable to alloc queue[%d]\n", cloop_name, cloop_num);
  4.2142 +-   goto error_out_free_tags;
  4.2143 ++   goto error_out;
  4.2144 +   }
  4.2145 +  clo->clo_queue->queuedata = clo;
  4.2146 +- blk_queue_max_hw_sectors(clo->clo_queue, BLK_DEF_MAX_SECTORS);
  4.2147 ++#if LINUX_VERSION_CODE < KERNEL_VERSION(4, 12, 0)
  4.2148 ++ queue_flag_set_unlocked(QUEUE_FLAG_NONROT, clo->clo_queue);
  4.2149 ++ queue_flag_set_unlocked(QUEUE_FLAG_NOMERGES, clo->clo_queue);
  4.2150 ++#endif
  4.2151 +  clo->clo_disk = alloc_disk(1);
  4.2152 +  if(!clo->clo_disk)
  4.2153 +   {
  4.2154 +    printk(KERN_ERR "%s: Unable to alloc disk[%d]\n", cloop_name, cloop_num);
  4.2155 +-   goto error_out_free_queue;
  4.2156 ++   goto error_disk;
  4.2157 +   }
  4.2158 +- spin_lock_init(&clo->queue_lock);
  4.2159 +- mutex_init(&clo->clo_ctl_mutex);
  4.2160 +- mutex_init(&clo->clo_rq_mutex);
  4.2161 +  clo->clo_disk->major = cloop_major;
  4.2162 +  clo->clo_disk->first_minor = cloop_num;
  4.2163 +  clo->clo_disk->fops = &clo_fops;
  4.2164 +@@ -1163,12 +1225,8 @@
  4.2165 +  sprintf(clo->clo_disk->disk_name, "%s%d", cloop_name, cloop_num);
  4.2166 +  add_disk(clo->clo_disk);
  4.2167 +  return 0;
  4.2168 +-error_out_free_queue:
  4.2169 ++error_disk:
  4.2170 +  blk_cleanup_queue(clo->clo_queue);
  4.2171 +-error_out_free_tags:
  4.2172 +- blk_mq_free_tag_set(&clo->tag_set);
  4.2173 +-error_out_free_clo:
  4.2174 +- cloop_free(clo, sizeof(struct cloop_device));
  4.2175 + error_out:
  4.2176 +  return -ENOMEM;
  4.2177 + }
  4.2178 +@@ -1179,7 +1237,6 @@
  4.2179 +  if(clo == NULL) return;
  4.2180 +  del_gendisk(clo->clo_disk);
  4.2181 +  blk_cleanup_queue(clo->clo_queue);
  4.2182 +- blk_mq_free_tag_set(&clo->tag_set);
  4.2183 +  put_disk(clo->clo_disk);
  4.2184 +  cloop_free(clo, sizeof(struct cloop_device));
  4.2185 +  cloop_dev[cloop_num] = NULL;
  4.2186 +--- cloop_suspend.c
  4.2187 ++++ cloop_suspend.c
  4.2188 +@@ -14,6 +14,7 @@
  4.2189 + #include <fcntl.h>
  4.2190 + #include <unistd.h>
  4.2191 + #include <stdio.h>
  4.2192 ++#include <stdint.h>
  4.2193 + 
  4.2194 + /* We don't use the structure, so that define does not hurt */
  4.2195 + #define dev_t int
     5.1 --- a/wbar/receipt	Sun May 08 13:06:36 2022 +0000
     5.2 +++ b/wbar/receipt	Sun May 08 16:45:21 2022 +0000
     5.3 @@ -8,7 +8,7 @@
     5.4  LICENSE="GPL"
     5.5  TARBALL="$PACKAGE-$VERSION.tbz2"
     5.6  WEB_SITE="https://github.com/rodolf0/wbar"
     5.7 -WGET_URL="http://www.tecapli.com.ar/warlock/$TARBALL"
     5.8 +WGET_URL="https://storage.googleapis.com/google-code-archive-downloads/v2/code.google.com/$PACKAGE/$TARBALL"
     5.9  CONFIG_FILES="/etc/wbar/dot.wbar"
    5.10  TAGS="desktop launchbar"
    5.11  
     6.1 --- a/xa/receipt	Sun May 08 13:06:36 2022 +0000
     6.2 +++ b/xa/receipt	Sun May 08 16:45:21 2022 +0000
     6.3 @@ -10,7 +10,7 @@
     6.4  WEB_SITE="https://www.floodgap.com/retrotech/xa/"
     6.5  
     6.6  TARBALL="$PACKAGE-$VERSION.tar.gz"
     6.7 -WGET_URL="${WEB_SITE}dists/$TARBALL"
     6.8 +WGET_URL="${WEB_SITE}dists/unsupported/$TARBALL"
     6.9  
    6.10  # What is the latest version available today?
    6.11  current_version()
     7.1 --- a/xautomation/receipt	Sun May 08 13:06:36 2022 +0000
     7.2 +++ b/xautomation/receipt	Sun May 08 16:45:21 2022 +0000
     7.3 @@ -7,9 +7,8 @@
     7.4  MAINTAINER="pankso@slitaz.org"
     7.5  LICENSE="GPL2"
     7.6  TARBALL="$PACKAGE-$VERSION.tar.gz"
     7.7 -#WEB_SITE="http://hoopajoo.net/projects/xautomation.html"
     7.8 -WEB_SITE="https://sourceforge.net/projects/xautomation/"
     7.9 -WGET_URL="http://hoopajoo.net/static/projects/$TARBALL"
    7.10 +WEB_SITE="https://www.hoopajoo.net/projects/xautomation.html"
    7.11 +WGET_URL="https://www.hoopajoo.net/static/projects/$TARBALL"
    7.12  #HOST_ARCH="i486 arm"
    7.13  
    7.14  DEPENDS="xorg-libX11 libpng"
     8.1 --- a/xcursor-aero/receipt	Sun May 08 13:06:36 2022 +0000
     8.2 +++ b/xcursor-aero/receipt	Sun May 08 16:45:21 2022 +0000
     8.3 @@ -8,7 +8,7 @@
     8.4  LICENSE="GPL"
     8.5  WEB_SITE="https://www.gnome-look.org/p/999972/"
     8.6  TARBALL="$PACKAGE-$VERSION.tar.gz"
     8.7 -WGET_URL="http://www.infinality.net/files/aero.tar.gz"
     8.8 +WGET_URL="https://github.com/Infinality/mouse-cursors/raw/master/aero.tar.gz"
     8.9  TAGS="cursor-theme"
    8.10  
    8.11  DEPENDS="xorg-libXcursor"
     9.1 --- a/xcursor-comix/receipt	Sun May 08 13:06:36 2022 +0000
     9.2 +++ b/xcursor-comix/receipt	Sun May 08 16:45:21 2022 +0000
     9.3 @@ -1,7 +1,7 @@
     9.4  # SliTaz package receipt.
     9.5  
     9.6  PACKAGE="xcursor-comix"
     9.7 -VERSION="0.9.1"
     9.8 +VERSION="0.9.2"
     9.9  CATEGORY="customization"
    9.10  TAGS="cursor-theme"
    9.11  SHORT_DESC="Comix cursor theme."
    9.12 @@ -10,6 +10,7 @@
    9.13  WEB_SITE="https://limitland.de/comixcursors"
    9.14  
    9.15  TARBALL="ComixCursors-$VERSION.tar.bz2"
    9.16 +#WGET_URL="https://gitlab.com/limitland/comixcursors/-/archive/$VERSION/comixcursors-$VERSION.tar.bz2"
    9.17  WGET_URL="https://limitland.gitlab.io/comixcursors/$TARBALL"
    9.18  
    9.19  DEPENDS="xorg-libXcursor"