This is the 5.4.256 stable release

-----BEGIN PGP SIGNATURE-----
 
 iQIzBAABCAAdFiEEZH8oZUiU471FcZm+ONu9yGCSaT4FAmTy4fwACgkQONu9yGCS
 aT7I2g//eixiEl0BJa3zbu0Y6EZ4LNGw7y20k66+TUZnr/2o7LR01RKTcRI4k9/n
 bjxqIfMGdVTPySFdWaieNk46aTPD9RBtgmium2Am2XoN2JbLqDKipAGZMqU2aqNU
 FLmiHLBRCbN3LL5QZnT1h3ExyfzpLszebogdkMHozp/+jHhbVxgqKn779NmbMfdh
 P4fBgc7IwfNra99y2pEmVv10S3zJG9wrx+UkM/AF9dGQIpgvzXqnOU+Zo+5Vy7fz
 DS/GRhportazBON4BFRQY1FA5rQvyFxPRlPdq762S9btPijS16n7FCuBxOucJ3JY
 mMhKepMhPTMuPGrgaamqn5QXgmdaiK/oFmaCAD/5FrSJ+e2+UmLjI7AmOqgj9QDa
 tvwa3Bgc29P2AjD9aq78RvjZ8aOXRH3H3AuDI99zvsK9equGsIMRvKl98IM+6m0x
 RgwwHk9zkzXMDHPROrNlrWqtQnaxfo3oG1xKYrk9IvRoIKPVspsNgnZREVO+uODp
 zmuPxW4Ol1QirqfyczbNAYL0YuumLTR2TfcumfZ1BuEvp9YNS5RMeNuyykCaaDyV
 yBlo6HvNeGtJBnLVHej9rICpWW1jcK7jbLQRjPf8IrnOyTdVHQqgy0XCvuzEw1yb
 dpd9Mp2Tsy/rKLGzS57ij3qy8/D5KgCiUQb11/CXbd/0xhtfZUQ=
 =R+7L
 -----END PGP SIGNATURE-----

Merge 5.4.256 into android11-5.4-lts

Changes in 5.4.256
	powerpc/pmac/smp: Avoid unused-variable warnings
	powerpc/pmac/smp: Drop unnecessary volatile qualifier
	Revert "MIPS: Alchemy: fix dbdma2"
	Linux 5.4.256

Change-Id: Id963efe437b2c99033d2ae4496b1081333f628bb
Signed-off-by: Greg Kroah-Hartman <gregkh@google.com>
This commit is contained in:
Greg Kroah-Hartman 2023-09-02 10:34:30 +00:00
commit ab08e05dca
3 changed files with 17 additions and 20 deletions

View File

@ -1,7 +1,7 @@
# SPDX-License-Identifier: GPL-2.0
VERSION = 5
PATCHLEVEL = 4
SUBLEVEL = 255
SUBLEVEL = 256
EXTRAVERSION =
NAME = Kleptomaniac Octopus

View File

@ -30,7 +30,6 @@
*
*/
#include <linux/dma-map-ops.h> /* for dma_default_coherent */
#include <linux/init.h>
#include <linux/kernel.h>
#include <linux/slab.h>
@ -624,18 +623,17 @@ u32 au1xxx_dbdma_put_source(u32 chanid, dma_addr_t buf, int nbytes, u32 flags)
dp->dscr_cmd0 &= ~DSCR_CMD0_IE;
/*
* There is an erratum on certain Au1200/Au1550 revisions that could
* result in "stale" data being DMA'ed. It has to do with the snoop
* logic on the cache eviction buffer. dma_default_coherent is set
* to false on these parts.
* There is an errata on the Au1200/Au1550 parts that could result
* in "stale" data being DMA'ed. It has to do with the snoop logic on
* the cache eviction buffer. DMA_NONCOHERENT is on by default for
* these parts. If it is fixed in the future, these dma_cache_inv will
* just be nothing more than empty macros. See io.h.
*/
if (!dma_default_coherent)
dma_cache_wback_inv(KSEG0ADDR(buf), nbytes);
dma_cache_wback_inv((unsigned long)buf, nbytes);
dp->dscr_cmd0 |= DSCR_CMD0_V; /* Let it rip */
wmb(); /* drain writebuffer */
dma_cache_wback_inv((unsigned long)dp, sizeof(*dp));
ctp->chan_ptr->ddma_dbell = 0;
wmb(); /* force doorbell write out to dma engine */
/* Get next descriptor pointer. */
ctp->put_ptr = phys_to_virt(DSCR_GET_NXTPTR(dp->dscr_nxtptr));
@ -687,18 +685,17 @@ u32 au1xxx_dbdma_put_dest(u32 chanid, dma_addr_t buf, int nbytes, u32 flags)
dp->dscr_source1, dp->dscr_dest0, dp->dscr_dest1);
#endif
/*
* There is an erratum on certain Au1200/Au1550 revisions that could
* result in "stale" data being DMA'ed. It has to do with the snoop
* logic on the cache eviction buffer. dma_default_coherent is set
* to false on these parts.
* There is an errata on the Au1200/Au1550 parts that could result in
* "stale" data being DMA'ed. It has to do with the snoop logic on the
* cache eviction buffer. DMA_NONCOHERENT is on by default for these
* parts. If it is fixed in the future, these dma_cache_inv will just
* be nothing more than empty macros. See io.h.
*/
if (!dma_default_coherent)
dma_cache_inv(KSEG0ADDR(buf), nbytes);
dma_cache_inv((unsigned long)buf, nbytes);
dp->dscr_cmd0 |= DSCR_CMD0_V; /* Let it rip */
wmb(); /* drain writebuffer */
dma_cache_wback_inv((unsigned long)dp, sizeof(*dp));
ctp->chan_ptr->ddma_dbell = 0;
wmb(); /* force doorbell write out to dma engine */
/* Get next descriptor pointer. */
ctp->put_ptr = phys_to_virt(DSCR_GET_NXTPTR(dp->dscr_nxtptr));

View File

@ -660,13 +660,13 @@ static void smp_core99_gpio_tb_freeze(int freeze)
#endif /* !CONFIG_PPC64 */
/* L2 and L3 cache settings to pass from CPU0 to CPU1 on G4 cpus */
volatile static long int core99_l2_cache;
volatile static long int core99_l3_cache;
static void core99_init_caches(int cpu)
{
#ifndef CONFIG_PPC64
/* L2 and L3 cache settings to pass from CPU0 to CPU1 on G4 cpus */
static long int core99_l2_cache;
static long int core99_l3_cache;
if (!cpu_has_feature(CPU_FTR_L2CR))
return;