diff options
-rw-r--r-- | recipes-core/microcode/intel-microcode_20220207.bb (renamed from recipes-core/microcode/intel-microcode_20210608.bb) | 2 | ||||
-rw-r--r-- | recipes-core/zlib/files/CVE-2018-25032-fuzz-fixed.patch | 351 | ||||
-rw-r--r-- | recipes-core/zlib/files/fix-removed-last-lit.patch | 41 | ||||
-rw-r--r-- | recipes-core/zlib/zlib-intel_1.2.11.1.jtkv6.3.bb | 8 | ||||
-rw-r--r-- | recipes-kernel/linux/linux-intel-rt_5.10.bb | 6 | ||||
-rw-r--r-- | recipes-kernel/linux/linux-intel-rt_5.4.bb | 6 | ||||
-rw-r--r-- | recipes-kernel/linux/linux-intel/0001-io-mapping-Cleanup-atomic-iomap.patch | 94 | ||||
-rw-r--r-- | recipes-kernel/linux/linux-intel_5.10.bb | 7 | ||||
-rw-r--r-- | recipes-kernel/linux/linux-intel_5.4.bb | 6 |
9 files changed, 506 insertions, 15 deletions
diff --git a/recipes-core/microcode/intel-microcode_20210608.bb b/recipes-core/microcode/intel-microcode_20220207.bb index 8a53421f..cfeb4150 100644 --- a/recipes-core/microcode/intel-microcode_20210608.bb +++ b/recipes-core/microcode/intel-microcode_20220207.bb @@ -16,7 +16,7 @@ LIC_FILES_CHKSUM = "file://license;md5=d8405101ec6e90c1d84b082b0c40c721" SRC_URI = "git://github.com/intel/Intel-Linux-Processor-Microcode-Data-Files.git;protocol=https;branch=main \ " -SRCREV = "3f97690f0da8011f52209b232450a1e5c4f2e1f6" +SRCREV = "115c3e4cdad6a9d84bf06e066162c5c546a9d2c3" DEPENDS = "iucode-tool-native" S = "${WORKDIR}/git" diff --git a/recipes-core/zlib/files/CVE-2018-25032-fuzz-fixed.patch b/recipes-core/zlib/files/CVE-2018-25032-fuzz-fixed.patch new file mode 100644 index 00000000..778b3b9a --- /dev/null +++ b/recipes-core/zlib/files/CVE-2018-25032-fuzz-fixed.patch @@ -0,0 +1,351 @@ +From c3970bb352f4f8cdb27d31e78c558bd1eb332f48 Mon Sep 17 00:00:00 2001 +From: Mark Adler <madler@alumni.caltech.edu> +Date: Tue, 17 Apr 2018 22:09:22 -0700 +Subject: [PATCH] Fix a bug that can crash deflate on some input when using + Z_FIXED. + +This bug was reported by Danilo Ramos of Eideticom, Inc. It has +lain in wait 13 years before being found! The bug was introduced +in zlib 1.2.2.2, with the addition of the Z_FIXED option. That +option forces the use of fixed Huffman codes. For rare inputs with +a large number of distant matches, the pending buffer into which +the compressed data is written can overwrite the distance symbol +table which it overlays. That results in corrupted output due to +invalid distances, and can result in out-of-bound accesses, +crashing the application. + +The fix here combines the distance buffer and literal/length +buffers into a single symbol buffer. Now three bytes of pending +buffer space are opened up for each literal or length/distance +pair consumed, instead of the previous two bytes. This assures +that the pending buffer cannot overwrite the symbol table, since +the maximum fixed code compressed length/distance is 31 bits, and +since there are four bytes of pending space for every three bytes +of symbol space. + +Upstream-Status: Backport +[https://github.com/madler/zlib/commit/5c44459c3b28a9bd3283aaceab7c615f8020c531] +CVE: CVE-2018-25032 +Signed-off-by: Chee Yang Lee <chee.yang.lee@intel.com> +--- + deflate.c | 74 ++++++++++++++++++++++++++++++++++++++++--------------- + deflate.h | 25 +++++++++---------- + trees.c | 50 +++++++++++-------------------------- + 3 files changed, 79 insertions(+), 70 deletions(-) + +diff --git a/deflate.c b/deflate.c +index b5e7849..2730035 100644 +--- a/deflate.c ++++ b/deflate.c +@@ -237,11 +237,6 @@ int ZEXPORT deflateInit2_(strm, level, method, windowBits, memLevel, strategy, + int wrap = 1; + static const char my_version[] = ZLIB_VERSION; + +- ushf *overlay; +- /* We overlay pending_buf and d_buf+l_buf. This works since the average +- * output size for (length,distance) codes is <= 24 bits. +- */ +- + if (version == Z_NULL || version[0] != my_version[0] || + stream_size != sizeof(z_stream)) { + return Z_VERSION_ERROR; +@@ -324,9 +319,47 @@ int ZEXPORT deflateInit2_(strm, level, method, windowBits, memLevel, strategy, + + s->lit_bufsize = 1 << (memLevel + 6); /* 16K elements by default */ + +- overlay = (ushf *) ZALLOC(strm, s->lit_bufsize, sizeof(ush)+2); +- s->pending_buf = (uchf *) overlay; +- s->pending_buf_size = (ulg)s->lit_bufsize * (sizeof(ush)+2L); ++ /* We overlay pending_buf and sym_buf. This works since the average size ++ * for length/distance pairs over any compressed block is assured to be 31 ++ * bits or less. ++ * ++ * Analysis: The longest fixed codes are a length code of 8 bits plus 5 ++ * extra bits, for lengths 131 to 257. The longest fixed distance codes are ++ * 5 bits plus 13 extra bits, for distances 16385 to 32768. The longest ++ * possible fixed-codes length/distance pair is then 31 bits total. ++ * ++ * sym_buf starts one-fourth of the way into pending_buf. So there are ++ * three bytes in sym_buf for every four bytes in pending_buf. Each symbol ++ * in sym_buf is three bytes -- two for the distance and one for the ++ * literal/length. As each symbol is consumed, the pointer to the next ++ * sym_buf value to read moves forward three bytes. From that symbol, up to ++ * 31 bits are written to pending_buf. The closest the written pending_buf ++ * bits gets to the next sym_buf symbol to read is just before the last ++ * code is written. At that time, 31*(n-2) bits have been written, just ++ * after 24*(n-2) bits have been consumed from sym_buf. sym_buf starts at ++ * 8*n bits into pending_buf. (Note that the symbol buffer fills when n-1 ++ * symbols are written.) The closest the writing gets to what is unread is ++ * then n+14 bits. Here n is lit_bufsize, which is 16384 by default, and ++ * can range from 128 to 32768. ++ * ++ * Therefore, at a minimum, there are 142 bits of space between what is ++ * written and what is read in the overlain buffers, so the symbols cannot ++ * be overwritten by the compressed data. That space is actually 139 bits, ++ * due to the three-bit fixed-code block header. ++ * ++ * That covers the case where either Z_FIXED is specified, forcing fixed ++ * codes, or when the use of fixed codes is chosen, because that choice ++ * results in a smaller compressed block than dynamic codes. That latter ++ * condition then assures that the above analysis also covers all dynamic ++ * blocks. A dynamic-code block will only be chosen to be emitted if it has ++ * fewer bits than a fixed-code block would for the same set of symbols. ++ * Therefore its average symbol length is assured to be less than 31. So ++ * the compressed data for a dynamic block also cannot overwrite the ++ * symbols from which it is being constructed. ++ */ ++ ++ s->pending_buf = (uchf *) ZALLOC(strm, s->lit_bufsize, 4); ++ s->pending_buf_size = (ulg)s->lit_bufsize * 4; + + if (s->window == Z_NULL || s->prev == Z_NULL || s->head == Z_NULL || + s->pending_buf == Z_NULL) { +@@ -335,8 +368,12 @@ int ZEXPORT deflateInit2_(strm, level, method, windowBits, memLevel, strategy, + deflateEnd (strm); + return Z_MEM_ERROR; + } +- s->d_buf = overlay + s->lit_bufsize/sizeof(ush); +- s->l_buf = s->pending_buf + (1+sizeof(ush))*s->lit_bufsize; ++ s->sym_buf = s->pending_buf + s->lit_bufsize; ++ s->sym_end = (s->lit_bufsize - 1) * 3; ++ /* We avoid equality with lit_bufsize*3 because of wraparound at 64K ++ * on 16 bit machines and because stored blocks are restricted to ++ * 64K-1 bytes. ++ */ + + s->level = level; + s->strategy = strategy; +@@ -553,7 +590,7 @@ int ZEXPORT deflatePrime (strm, bits, value) + + if (deflateStateCheck(strm)) return Z_STREAM_ERROR; + s = strm->state; +- if ((Bytef *)(s->d_buf) < s->pending_out + ((Buf_size + 7) >> 3)) ++ if (s->sym_buf < s->pending_out + ((Buf_size + 7) >> 3)) + return Z_BUF_ERROR; + do { + put = Buf_size - s->bi_valid; +@@ -1125,7 +1162,6 @@ int ZEXPORT deflateCopy (dest, source) + #else + deflate_state *ds; + deflate_state *ss; +- ushf *overlay; + + + if (deflateStateCheck(source) || dest == Z_NULL) { +@@ -1145,8 +1181,7 @@ int ZEXPORT deflateCopy (dest, source) + ds->window = (Bytef *) ZALLOC(dest, ds->w_size, 2*sizeof(Byte)); + ds->prev = (Posf *) ZALLOC(dest, ds->w_size, sizeof(Pos)); + ds->head = (Posf *) ZALLOC(dest, ds->hash_size, sizeof(Pos)); +- overlay = (ushf *) ZALLOC(dest, ds->lit_bufsize, sizeof(ush)+2); +- ds->pending_buf = (uchf *) overlay; ++ ds->pending_buf = (uchf *) ZALLOC(dest, ds->lit_bufsize, 4); + + if (ds->window == Z_NULL || ds->prev == Z_NULL || ds->head == Z_NULL || + ds->pending_buf == Z_NULL) { +@@ -1160,8 +1195,7 @@ int ZEXPORT deflateCopy (dest, source) + zmemcpy(ds->pending_buf, ss->pending_buf, (uInt)ds->pending_buf_size); + + ds->pending_out = ds->pending_buf + (ss->pending_out - ss->pending_buf); +- ds->d_buf = overlay + ds->lit_bufsize/sizeof(ush); +- ds->l_buf = ds->pending_buf + (1+sizeof(ush))*ds->lit_bufsize; ++ ds->sym_buf = ds->pending_buf + ds->lit_bufsize; + + ds->l_desc.dyn_tree = ds->dyn_ltree; + ds->d_desc.dyn_tree = ds->dyn_dtree; +@@ -1717,7 +1751,7 @@ local block_state deflate_fast(s, flush) + FLUSH_BLOCK(s, 1); + return finish_done; + } +- if (s->last_lit) ++ if (s->sym_next) + FLUSH_BLOCK(s, 0); + return block_done; + } +@@ -1848,7 +1882,7 @@ local block_state deflate_slow(s, flush) + FLUSH_BLOCK(s, 1); + return finish_done; + } +- if (s->last_lit) ++ if (s->sym_next) + FLUSH_BLOCK(s, 0); + return block_done; + } +@@ -1923,7 +1957,7 @@ local block_state deflate_rle(s, flush) + FLUSH_BLOCK(s, 1); + return finish_done; + } +- if (s->last_lit) ++ if (s->sym_next) + FLUSH_BLOCK(s, 0); + return block_done; + } +@@ -1962,7 +1996,7 @@ local block_state deflate_huff(s, flush) + FLUSH_BLOCK(s, 1); + return finish_done; + } +- if (s->last_lit) ++ if (s->sym_next) + FLUSH_BLOCK(s, 0); + return block_done; + } +diff --git a/deflate.h b/deflate.h +index 7f31b54..8d84f8e 100644 +--- a/deflate.h ++++ b/deflate.h +@@ -230,7 +230,7 @@ typedef struct internal_state { + /* Depth of each subtree used as tie breaker for trees of equal frequency + */ + +- uchf *l_buf; /* buffer for literals or lengths */ ++ uchf *sym_buf; /* buffer for distances and literals/lengths */ + + uInt lit_bufsize; + /* Size of match buffer for literals/lengths. There are 4 reasons for +@@ -252,13 +252,8 @@ typedef struct internal_state { + * - I can't count above 4 + */ + +- uInt last_lit; /* running index in l_buf */ +- +- ushf *d_buf; +- /* Buffer for distances. To simplify the code, d_buf and l_buf have +- * the same number of elements. To use different lengths, an extra flag +- * array would be necessary. +- */ ++ uInt sym_next; /* running index in sym_buf */ ++ uInt sym_end; /* symbol table full when sym_next reaches this */ + + ulg opt_len; /* bit length of current block with optimal trees */ + ulg static_len; /* bit length of current block with static trees */ +@@ -344,20 +339,22 @@ void ZLIB_INTERNAL bi_windup OF((deflate_state *s)); + + # define _tr_tally_lit(s, c, flush) \ + { uch cc = (c); \ +- s->d_buf[s->last_lit] = 0; \ +- s->l_buf[s->last_lit++] = cc; \ ++ s->sym_buf[s->sym_next++] = 0; \ ++ s->sym_buf[s->sym_next++] = 0; \ ++ s->sym_buf[s->sym_next++] = cc; \ + s->dyn_ltree[cc].Freq++; \ +- flush = (s->last_lit == s->lit_bufsize-1); \ ++ flush = (s->sym_next == s->sym_end); \ + } + # define _tr_tally_dist(s, distance, length, flush) \ + { uch len = (uch)(length); \ + ush dist = (ush)(distance); \ +- s->d_buf[s->last_lit] = dist; \ +- s->l_buf[s->last_lit++] = len; \ ++ s->sym_buf[s->sym_next++] = dist; \ ++ s->sym_buf[s->sym_next++] = dist >> 8; \ ++ s->sym_buf[s->sym_next++] = len; \ + dist--; \ + s->dyn_ltree[_length_code[len]+LITERALS+1].Freq++; \ + s->dyn_dtree[d_code(dist)].Freq++; \ +- flush = (s->last_lit == s->lit_bufsize-1); \ ++ flush = (s->sym_next == s->sym_end); \ + } + #else + # define _tr_tally_lit(s, c, flush) flush = _tr_tally(s, 0, c) +diff --git a/trees.c b/trees.c +index 7480efc..2fc7860 100644 +--- a/trees.c ++++ b/trees.c +@@ -343,7 +343,7 @@ local void init_block(s) + + s->dyn_ltree[END_BLOCK].Freq = 1; + s->opt_len = s->static_len = 0L; +- s->last_lit = s->matches = 0; ++ s->sym_next = s->matches = 0; + } + + #define SMALLEST 1 +@@ -875,7 +875,7 @@ void ZLIB_INTERNAL _tr_flush_block(s, buf, stored_len, last) + + Tracev((stderr, "\nopt %lu(%lu) stat %lu(%lu) stored %lu lit %u ", + opt_lenb, s->opt_len, static_lenb, s->static_len, stored_len, +- s->last_lit)); ++ s->sym_next / 3)); + + if (static_lenb <= opt_lenb) opt_lenb = static_lenb; + +@@ -944,8 +944,9 @@ int ZLIB_INTERNAL _tr_tally (s, dist, lc) + unsigned dist; /* distance of matched string */ + unsigned lc; /* match length-MIN_MATCH or unmatched char (if dist==0) */ + { +- s->d_buf[s->last_lit] = (ush)dist; +- s->l_buf[s->last_lit++] = (uch)lc; ++ s->sym_buf[s->sym_next++] = dist; ++ s->sym_buf[s->sym_next++] = dist >> 8; ++ s->sym_buf[s->sym_next++] = lc; + if (dist == 0) { + /* lc is the unmatched char */ + s->dyn_ltree[lc].Freq++; +@@ -960,30 +961,7 @@ int ZLIB_INTERNAL _tr_tally (s, dist, lc) + s->dyn_ltree[_length_code[lc]+LITERALS+1].Freq++; + s->dyn_dtree[d_code(dist)].Freq++; + } +- +-#ifdef TRUNCATE_BLOCK +- /* Try to guess if it is profitable to stop the current block here */ +- if ((s->last_lit & 0x1fff) == 0 && s->level > 2) { +- /* Compute an upper bound for the compressed length */ +- ulg out_length = (ulg)s->last_lit*8L; +- ulg in_length = (ulg)((long)s->strstart - s->block_start); +- int dcode; +- for (dcode = 0; dcode < D_CODES; dcode++) { +- out_length += (ulg)s->dyn_dtree[dcode].Freq * +- (5L+extra_dbits[dcode]); +- } +- out_length >>= 3; +- Tracev((stderr,"\nlast_lit %u, in %ld, out ~%ld(%ld%%) ", +- s->last_lit, in_length, out_length, +- 100L - out_length*100L/in_length)); +- if (s->matches < s->last_lit/2 && out_length < in_length/2) return 1; +- } +-#endif +- return (s->last_lit == s->lit_bufsize-1); +- /* We avoid equality with lit_bufsize because of wraparound at 64K +- * on 16 bit machines and because stored blocks are restricted to +- * 64K-1 bytes. +- */ ++ return (s->sym_next == s->sym_end); + } + + /* =========================================================================== +@@ -996,13 +974,14 @@ local void compress_block(s, ltree, dtree) + { + unsigned dist; /* distance of matched string */ + int lc; /* match length or unmatched char (if dist == 0) */ +- unsigned lx = 0; /* running index in l_buf */ ++ unsigned sx = 0; /* running index in sym_buf */ + unsigned code; /* the code to send */ + int extra; /* number of extra bits to send */ + +- if (s->last_lit != 0) do { +- dist = s->d_buf[lx]; +- lc = s->l_buf[lx++]; ++ if (s->sym_next != 0) do { ++ dist = s->sym_buf[sx++] & 0xff; ++ dist += (unsigned)(s->sym_buf[sx++] & 0xff) << 8; ++ lc = s->sym_buf[sx++]; + if (dist == 0) { + send_code(s, lc, ltree); /* send a literal byte */ + Tracecv(isgraph(lc), (stderr," '%c' ", lc)); +@@ -1027,11 +1006,10 @@ local void compress_block(s, ltree, dtree) + } + } /* literal or match pair ? */ + +- /* Check that the overlay between pending_buf and d_buf+l_buf is ok: */ +- Assert((uInt)(s->pending) < s->lit_bufsize + 2*lx, +- "pendingBuf overflow"); ++ /* Check that the overlay between pending_buf and sym_buf is ok: */ ++ Assert(s->pending < s->lit_bufsize + sx, "pendingBuf overflow"); + +- } while (lx < s->last_lit); ++ } while (sx < s->sym_next); + + send_code(s, END_BLOCK, ltree); + } +-- +2.35.1 + diff --git a/recipes-core/zlib/files/fix-removed-last-lit.patch b/recipes-core/zlib/files/fix-removed-last-lit.patch new file mode 100644 index 00000000..238ecfbe --- /dev/null +++ b/recipes-core/zlib/files/fix-removed-last-lit.patch @@ -0,0 +1,41 @@ +From bf7f8fd281afee27c160cd1e904b2ff52eb9e493 Mon Sep 17 00:00:00 2001 +From: Chee Yang Lee <chee.yang.lee@intel.com> +Date: Thu, 31 Mar 2022 22:00:10 +0800 +Subject: [PATCH] fix removed last_lit + +last_lit removed and replaced with sym_next while fix for +CVE-2018-25032. causing build error for deflate_medium.c. + +| deflate_medium.c: In function 'deflate_medium': +| deflate_medium.c:308:12: error: 'deflate_state' {aka 'struct +internal_state'} has no member named 'last_lit'; did you mean +'last_flush'? +| 308 | if (s->last_lit) +| | ^~~~~~~~ +| | last_flush +| make: *** [Makefile:307: deflate_medium.lo] Error 1 + + +Upstream-Status: Inappropriate +[upstream have not taken the CVE fix patches] +Signed-off-by: Chee Yang Lee <chee.yang.lee@intel.com> +--- + deflate_medium.c | 2 +- + 1 file changed, 1 insertion(+), 1 deletion(-) + +diff --git a/deflate_medium.c b/deflate_medium.c +index 582e8b2..30f7633 100644 +--- a/deflate_medium.c ++++ b/deflate_medium.c +@@ -305,7 +305,7 @@ block_state deflate_medium(deflate_state *s, int flush) + FLUSH_BLOCK(s, 1); + return finish_done; + } +- if (s->last_lit) ++ if (s->sym_next) + FLUSH_BLOCK(s, 0); + return block_done; + } +-- +2.35.1 + diff --git a/recipes-core/zlib/zlib-intel_1.2.11.1.jtkv6.3.bb b/recipes-core/zlib/zlib-intel_1.2.11.1.jtkv6.3.bb index 65303bff..27bfd9b1 100644 --- a/recipes-core/zlib/zlib-intel_1.2.11.1.jtkv6.3.bb +++ b/recipes-core/zlib/zlib-intel_1.2.11.1.jtkv6.3.bb @@ -6,8 +6,12 @@ SUMMARY = "Zlib Compression Library (Intel-tuned)" PROVIDES = "zlib" -SRC_URI_remove = "${SOURCEFORGE_MIRROR}/libpng/${BPN}/${PV}/${BPN}-${PV}.tar.xz" -SRC_URI_prepend = "git://github.com/jtkukunas/zlib.git;protocol=git " +SRC_URI = "git://github.com/jtkukunas/zlib.git;protocol=https;branch=master \ + file://ldflags-tests.patch \ + file://run-ptest \ + file://CVE-2018-25032-fuzz-fixed.patch \ + file://fix-removed-last-lit.patch \ + " SRCREV = "a43a247bfa16ec5368747b5b64f11ea5ca033010" S = "${WORKDIR}/git" diff --git a/recipes-kernel/linux/linux-intel-rt_5.10.bb b/recipes-kernel/linux/linux-intel-rt_5.10.bb index b01fa0ae..f752b842 100644 --- a/recipes-kernel/linux/linux-intel-rt_5.10.bb +++ b/recipes-kernel/linux/linux-intel-rt_5.10.bb @@ -18,9 +18,9 @@ KMETA_BRANCH = "yocto-5.10" DEPENDS += "elfutils-native openssl-native util-linux-native" -LINUX_VERSION ?= "5.10.56" -SRCREV_machine ?= "8394053722549076e54a9dc82a92c513907d227f" -SRCREV_meta ?= "873ef26ef35da24d217a1aeb44717278d77e818c" +LINUX_VERSION ?= "5.10.78" +SRCREV_machine ?= "c287378636ba3f627b4da448d62a75ed41a35982" +SRCREV_meta ?= "64fb693a6c11f21bab3ff9bb8dcb65a70abe05e3" LINUX_KERNEL_TYPE = "preempt-rt" diff --git a/recipes-kernel/linux/linux-intel-rt_5.4.bb b/recipes-kernel/linux/linux-intel-rt_5.4.bb index b26fa2bc..16e3fcac 100644 --- a/recipes-kernel/linux/linux-intel-rt_5.4.bb +++ b/recipes-kernel/linux/linux-intel-rt_5.4.bb @@ -18,9 +18,9 @@ KMETA_BRANCH = "yocto-5.4" DEPENDS += "elfutils-native openssl-native util-linux-native" -LINUX_VERSION ?= "5.4.129" -SRCREV_machine ?= "9003195c5c421640e6ac2726e5f02fc90f85ffde" -SRCREV_meta ?= "231d3a07e10680c7c89ea101cd803b0684482b11" +LINUX_VERSION ?= "5.4.170" +SRCREV_machine ?= "196e38246d15096460031f4a17913922c006b12d" +SRCREV_meta ?= "98cce1c95fcc9a26965cbc5f038fd71d53c387c8" LINUX_KERNEL_TYPE = "preempt-rt" diff --git a/recipes-kernel/linux/linux-intel/0001-io-mapping-Cleanup-atomic-iomap.patch b/recipes-kernel/linux/linux-intel/0001-io-mapping-Cleanup-atomic-iomap.patch new file mode 100644 index 00000000..e3608238 --- /dev/null +++ b/recipes-kernel/linux/linux-intel/0001-io-mapping-Cleanup-atomic-iomap.patch @@ -0,0 +1,94 @@ +From cc3ff5fe73139251070775308dcaadf474566b4b Mon Sep 17 00:00:00 2001 +From: Thomas Gleixner <tglx@linutronix.de> +Date: Tue, 3 Nov 2020 10:27:32 +0100 +Subject: [PATCH] io-mapping: Cleanup atomic iomap + +Switch the atomic iomap implementation over to kmap_local and stick the +preempt/pagefault mechanics into the generic code similar to the +kmap_atomic variants. + +Rename the x86 map function in preparation for a non-atomic variant. + +Upstream-Status: Backport [https://github.com/torvalds/linux/commit/351191ad55c8a1eccaf23e4187c62056229c0779] + +Signed-off-by: Thomas Gleixner <tglx@linutronix.de> +Cc: Linus Torvalds <torvalds@linuxfoundation.org> +Cc: Christoph Hellwig <hch@lst.de> +Cc: Andrew Morton <akpm@linux-foundation.org> +Link: https://lore.kernel.org/r/20201103095858.625310005@linutronix.de +--- + arch/x86/include/asm/iomap.h | 9 +-------- + arch/x86/mm/iomap_32.c | 6 ++---- + include/linux/io-mapping.h | 8 ++++++-- + 3 files changed, 9 insertions(+), 14 deletions(-) + +diff --git a/arch/x86/include/asm/iomap.h b/arch/x86/include/asm/iomap.h +index 0be7a30fd6bc..e2de092fc38c 100644 +--- a/arch/x86/include/asm/iomap.h ++++ b/arch/x86/include/asm/iomap.h +@@ -13,14 +13,7 @@ + #include <asm/cacheflush.h> + #include <asm/tlbflush.h> + +-void __iomem *iomap_atomic_pfn_prot(unsigned long pfn, pgprot_t prot); +- +-static inline void iounmap_atomic(void __iomem *vaddr) +-{ +- kunmap_local_indexed((void __force *)vaddr); +- pagefault_enable(); +- preempt_enable(); +-} ++void __iomem *__iomap_local_pfn_prot(unsigned long pfn, pgprot_t prot); + + int iomap_create_wc(resource_size_t base, unsigned long size, pgprot_t *prot); + +diff --git a/arch/x86/mm/iomap_32.c b/arch/x86/mm/iomap_32.c +index e0a40d7cc66c..9aaa756ddf21 100644 +--- a/arch/x86/mm/iomap_32.c ++++ b/arch/x86/mm/iomap_32.c +@@ -44,7 +44,7 @@ void iomap_free(resource_size_t base, unsigned long size) + } + EXPORT_SYMBOL_GPL(iomap_free); + +-void __iomem *iomap_atomic_pfn_prot(unsigned long pfn, pgprot_t prot) ++void __iomem *__iomap_local_pfn_prot(unsigned long pfn, pgprot_t prot) + { + /* + * For non-PAT systems, translate non-WB request to UC- just in +@@ -60,8 +60,6 @@ void __iomem *iomap_atomic_pfn_prot(unsigned long pfn, pgprot_t prot) + /* Filter out unsupported __PAGE_KERNEL* bits: */ + pgprot_val(prot) &= __default_kernel_pte_mask; + +- preempt_disable(); +- pagefault_disable(); + return (void __force __iomem *)__kmap_local_pfn_prot(pfn, prot); + } +-EXPORT_SYMBOL_GPL(iomap_atomic_pfn_prot); ++EXPORT_SYMBOL_GPL(__iomap_local_pfn_prot); +diff --git a/include/linux/io-mapping.h b/include/linux/io-mapping.h +index 0a1a186a43c5..e9743cfd8585 100644 +--- a/include/linux/io-mapping.h ++++ b/include/linux/io-mapping.h +@@ -69,13 +69,17 @@ io_mapping_map_atomic_wc(struct io_mapping *mapping, + + BUG_ON(offset >= mapping->size); + phys_addr = mapping->base + offset; +- return iomap_atomic_pfn_prot(PHYS_PFN(phys_addr), mapping->prot); ++ preempt_disable(); ++ pagefault_disable(); ++ return __iomap_local_pfn_prot(PHYS_PFN(phys_addr), mapping->prot); + } + + static inline void + io_mapping_unmap_atomic(void __iomem *vaddr) + { +- iounmap_atomic(vaddr); ++ kunmap_local_indexed((void __force *)vaddr); ++ pagefault_enable(); ++ preempt_enable(); + } + + static inline void __iomem * +-- +2.17.1 + diff --git a/recipes-kernel/linux/linux-intel_5.10.bb b/recipes-kernel/linux/linux-intel_5.10.bb index 04ac17cd..839eb297 100644 --- a/recipes-kernel/linux/linux-intel_5.10.bb +++ b/recipes-kernel/linux/linux-intel_5.10.bb @@ -6,13 +6,14 @@ KMETA_BRANCH = "yocto-5.10" LIC_FILES_CHKSUM = "file://COPYING;md5=6bc538ed5bd9a7fc9398086aedcd7e46" SRC_URI_append = " file://0001-menuconfig-mconf-cfg-Allow-specification-of-ncurses-.patch \ + file://0001-io-mapping-Cleanup-atomic-iomap.patch \ " DEPENDS += "elfutils-native openssl-native util-linux-native" -LINUX_VERSION ?= "5.10.56" -SRCREV_machine ?= "cef410ba7c738624a98b855c955a0092789b224e" -SRCREV_meta ?= "873ef26ef35da24d217a1aeb44717278d77e818c" +LINUX_VERSION ?= "5.10.78" +SRCREV_machine ?= "4e8162d8163c74e46a9f0bafb860f42249702791" +SRCREV_meta ?= "64fb693a6c11f21bab3ff9bb8dcb65a70abe05e3" # For Crystalforest and Romley KERNEL_MODULE_AUTOLOAD_append_core2-32-intel-common = " uio" diff --git a/recipes-kernel/linux/linux-intel_5.4.bb b/recipes-kernel/linux/linux-intel_5.4.bb index 7ab7d8eb..0553de92 100644 --- a/recipes-kernel/linux/linux-intel_5.4.bb +++ b/recipes-kernel/linux/linux-intel_5.4.bb @@ -9,9 +9,9 @@ SRC_URI_append = " file://0001-menuconfig-mconf-cfg-Allow-specification-of-ncurs DEPENDS += "elfutils-native openssl-native util-linux-native" -LINUX_VERSION ?= "5.4.138" -SRCREV_machine ?= "cfef018a5efc728040e938c270bbc68fc8124d33" -SRCREV_meta ?= "231d3a07e10680c7c89ea101cd803b0684482b11" +LINUX_VERSION ?= "5.4.170" +SRCREV_machine ?= "36f93ff941f127f4137ab369aecbdd995fb58c66" +SRCREV_meta ?= "98cce1c95fcc9a26965cbc5f038fd71d53c387c8" # For Crystalforest and Romley KERNEL_MODULE_AUTOLOAD_append_core2-32-intel-common = " uio" |