mirror of
https://github.com/hardkernel/linux.git
synced 2026-03-25 03:50:24 +09:00
Merge tag 'v3.4.40' of git://git.kernel.org/pub/scm/linux/kernel/git/stable/linux-stable into odroidxu-update
This is the 3.4.40 stable release
This commit is contained in:
2
Makefile
2
Makefile
@@ -1,6 +1,6 @@
|
||||
VERSION = 3
|
||||
PATCHLEVEL = 4
|
||||
SUBLEVEL = 39
|
||||
SUBLEVEL = 40
|
||||
EXTRAVERSION =
|
||||
NAME = Saber-toothed Squirrel
|
||||
|
||||
|
||||
@@ -188,6 +188,10 @@ nautilus_machine_check(unsigned long vector, unsigned long la_ptr)
|
||||
extern void free_reserved_mem(void *, void *);
|
||||
extern void pcibios_claim_one_bus(struct pci_bus *);
|
||||
|
||||
static struct resource irongate_io = {
|
||||
.name = "Irongate PCI IO",
|
||||
.flags = IORESOURCE_IO,
|
||||
};
|
||||
static struct resource irongate_mem = {
|
||||
.name = "Irongate PCI MEM",
|
||||
.flags = IORESOURCE_MEM,
|
||||
@@ -209,6 +213,7 @@ nautilus_init_pci(void)
|
||||
|
||||
irongate = pci_get_bus_and_slot(0, 0);
|
||||
bus->self = irongate;
|
||||
bus->resource[0] = &irongate_io;
|
||||
bus->resource[1] = &irongate_mem;
|
||||
|
||||
pci_bus_size_bridges(bus);
|
||||
|
||||
@@ -187,7 +187,13 @@ static long pSeries_lpar_hpte_remove(unsigned long hpte_group)
|
||||
(0x1UL << 4), &dummy1, &dummy2);
|
||||
if (lpar_rc == H_SUCCESS)
|
||||
return i;
|
||||
BUG_ON(lpar_rc != H_NOT_FOUND);
|
||||
|
||||
/*
|
||||
* The test for adjunct partition is performed before the
|
||||
* ANDCOND test. H_RESOURCE may be returned, so we need to
|
||||
* check for that as well.
|
||||
*/
|
||||
BUG_ON(lpar_rc != H_NOT_FOUND && lpar_rc != H_RESOURCE);
|
||||
|
||||
slot_offset++;
|
||||
slot_offset &= 0x7;
|
||||
|
||||
@@ -4,7 +4,7 @@
|
||||
# create a compressed vmlinux image from the original vmlinux
|
||||
#
|
||||
|
||||
targets := vmlinux.lds vmlinux vmlinux.bin vmlinux.bin.gz vmlinux.bin.bz2 vmlinux.bin.lzma vmlinux.bin.xz vmlinux.bin.lzo head_$(BITS).o misc.o string.o cmdline.o early_serial_console.o piggy.o
|
||||
targets := vmlinux vmlinux.bin vmlinux.bin.gz vmlinux.bin.bz2 vmlinux.bin.lzma vmlinux.bin.xz vmlinux.bin.lzo
|
||||
|
||||
KBUILD_CFLAGS := -m$(BITS) -D__KERNEL__ $(LINUX_INCLUDE) -O2
|
||||
KBUILD_CFLAGS += -fno-strict-aliasing -fPIC
|
||||
@@ -29,7 +29,6 @@ VMLINUX_OBJS = $(obj)/vmlinux.lds $(obj)/head_$(BITS).o $(obj)/misc.o \
|
||||
$(obj)/piggy.o
|
||||
|
||||
$(obj)/eboot.o: KBUILD_CFLAGS += -fshort-wchar -mno-red-zone
|
||||
$(obj)/efi_stub_$(BITS).o: KBUILD_CLFAGS += -fshort-wchar -mno-red-zone
|
||||
|
||||
ifeq ($(CONFIG_EFI_STUB), y)
|
||||
VMLINUX_OBJS += $(obj)/eboot.o $(obj)/efi_stub_$(BITS).o
|
||||
@@ -43,7 +42,7 @@ OBJCOPYFLAGS_vmlinux.bin := -R .comment -S
|
||||
$(obj)/vmlinux.bin: vmlinux FORCE
|
||||
$(call if_changed,objcopy)
|
||||
|
||||
targets += vmlinux.bin.all vmlinux.relocs
|
||||
targets += $(patsubst $(obj)/%,%,$(VMLINUX_OBJS)) vmlinux.bin.all vmlinux.relocs
|
||||
|
||||
CMD_RELOCS = arch/x86/tools/relocs
|
||||
quiet_cmd_relocs = RELOCS $@
|
||||
|
||||
@@ -27,13 +27,13 @@ extern const unsigned long sys_call_table[];
|
||||
*/
|
||||
static inline int syscall_get_nr(struct task_struct *task, struct pt_regs *regs)
|
||||
{
|
||||
return regs->orig_ax & __SYSCALL_MASK;
|
||||
return regs->orig_ax;
|
||||
}
|
||||
|
||||
static inline void syscall_rollback(struct task_struct *task,
|
||||
struct pt_regs *regs)
|
||||
{
|
||||
regs->ax = regs->orig_ax & __SYSCALL_MASK;
|
||||
regs->ax = regs->orig_ax;
|
||||
}
|
||||
|
||||
static inline long syscall_get_error(struct task_struct *task,
|
||||
|
||||
@@ -200,6 +200,8 @@ queue_store_##name(struct request_queue *q, const char *page, size_t count) \
|
||||
unsigned long val; \
|
||||
ssize_t ret; \
|
||||
ret = queue_var_store(&val, page, count); \
|
||||
if (ret < 0) \
|
||||
return ret; \
|
||||
if (neg) \
|
||||
val = !val; \
|
||||
\
|
||||
|
||||
17
crypto/gcm.c
17
crypto/gcm.c
@@ -44,6 +44,7 @@ struct crypto_rfc4543_ctx {
|
||||
|
||||
struct crypto_rfc4543_req_ctx {
|
||||
u8 auth_tag[16];
|
||||
u8 assocbuf[32];
|
||||
struct scatterlist cipher[1];
|
||||
struct scatterlist payload[2];
|
||||
struct scatterlist assoc[2];
|
||||
@@ -1142,9 +1143,19 @@ static struct aead_request *crypto_rfc4543_crypt(struct aead_request *req,
|
||||
scatterwalk_crypto_chain(payload, dst, vdst == req->iv + 8, 2);
|
||||
assoclen += 8 + req->cryptlen - (enc ? 0 : authsize);
|
||||
|
||||
sg_init_table(assoc, 2);
|
||||
sg_set_page(assoc, sg_page(req->assoc), req->assoc->length,
|
||||
req->assoc->offset);
|
||||
if (req->assoc->length == req->assoclen) {
|
||||
sg_init_table(assoc, 2);
|
||||
sg_set_page(assoc, sg_page(req->assoc), req->assoc->length,
|
||||
req->assoc->offset);
|
||||
} else {
|
||||
BUG_ON(req->assoclen > sizeof(rctx->assocbuf));
|
||||
|
||||
scatterwalk_map_and_copy(rctx->assocbuf, req->assoc, 0,
|
||||
req->assoclen, 0);
|
||||
|
||||
sg_init_table(assoc, 2);
|
||||
sg_set_buf(assoc, rctx->assocbuf, req->assoclen);
|
||||
}
|
||||
scatterwalk_crypto_chain(assoc, payload, 0, 2);
|
||||
|
||||
aead_request_set_tfm(subreq, ctx->child);
|
||||
|
||||
@@ -150,6 +150,7 @@ enum piix_controller_ids {
|
||||
tolapai_sata,
|
||||
piix_pata_vmw, /* PIIX4 for VMware, spurious DMA_ERR */
|
||||
ich8_sata_snb,
|
||||
ich8_2port_sata_snb,
|
||||
};
|
||||
|
||||
struct piix_map_db {
|
||||
@@ -326,7 +327,7 @@ static const struct pci_device_id piix_pci_tbl[] = {
|
||||
/* SATA Controller IDE (Lynx Point) */
|
||||
{ 0x8086, 0x8c01, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich8_sata_snb },
|
||||
/* SATA Controller IDE (Lynx Point) */
|
||||
{ 0x8086, 0x8c08, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich8_2port_sata },
|
||||
{ 0x8086, 0x8c08, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich8_2port_sata_snb },
|
||||
/* SATA Controller IDE (Lynx Point) */
|
||||
{ 0x8086, 0x8c09, PCI_ANY_ID, PCI_ANY_ID, 0, 0, ich8_2port_sata },
|
||||
/* SATA Controller IDE (DH89xxCC) */
|
||||
@@ -511,6 +512,7 @@ static const struct piix_map_db *piix_map_db_table[] = {
|
||||
[ich8m_apple_sata] = &ich8m_apple_map_db,
|
||||
[tolapai_sata] = &tolapai_map_db,
|
||||
[ich8_sata_snb] = &ich8_map_db,
|
||||
[ich8_2port_sata_snb] = &ich8_2port_map_db,
|
||||
};
|
||||
|
||||
static struct ata_port_info piix_port_info[] = {
|
||||
@@ -652,6 +654,15 @@ static struct ata_port_info piix_port_info[] = {
|
||||
.port_ops = &piix_sata_ops,
|
||||
},
|
||||
|
||||
[ich8_2port_sata_snb] =
|
||||
{
|
||||
.flags = PIIX_SATA_FLAGS | PIIX_FLAG_SIDPR
|
||||
| PIIX_FLAG_PIO16,
|
||||
.pio_mask = ATA_PIO4,
|
||||
.mwdma_mask = ATA_MWDMA2,
|
||||
.udma_mask = ATA_UDMA6,
|
||||
.port_ops = &piix_sata_ops,
|
||||
},
|
||||
};
|
||||
|
||||
static struct pci_bits piix_enable_bits[] = {
|
||||
|
||||
@@ -2408,6 +2408,9 @@ int ata_dev_configure(struct ata_device *dev)
|
||||
dev->max_sectors = min_t(unsigned int, ATA_MAX_SECTORS_128,
|
||||
dev->max_sectors);
|
||||
|
||||
if (dev->horkage & ATA_HORKAGE_MAX_SEC_LBA48)
|
||||
dev->max_sectors = ATA_MAX_SECTORS_LBA48;
|
||||
|
||||
if (ap->ops->dev_config)
|
||||
ap->ops->dev_config(dev);
|
||||
|
||||
@@ -4064,6 +4067,7 @@ static const struct ata_blacklist_entry ata_device_blacklist [] = {
|
||||
/* Weird ATAPI devices */
|
||||
{ "TORiSAN DVD-ROM DRD-N216", NULL, ATA_HORKAGE_MAX_SEC_128 },
|
||||
{ "QUANTUM DAT DAT72-000", NULL, ATA_HORKAGE_ATAPI_MOD16_DMA },
|
||||
{ "Slimtype DVD A DS8A8SH", NULL, ATA_HORKAGE_MAX_SEC_LBA48 },
|
||||
|
||||
/* Devices we expect to fail diagnostics */
|
||||
|
||||
|
||||
@@ -392,7 +392,7 @@ static int regcache_rbtree_sync(struct regmap *map, unsigned int min,
|
||||
base = 0;
|
||||
|
||||
if (max < rbnode->base_reg + rbnode->blklen)
|
||||
end = rbnode->base_reg + rbnode->blklen - max;
|
||||
end = max - rbnode->base_reg + 1;
|
||||
else
|
||||
end = rbnode->blklen;
|
||||
|
||||
|
||||
@@ -19,10 +19,10 @@
|
||||
/* There is only *one* pci_eisa device per machine, right ? */
|
||||
static struct eisa_root_device pci_eisa_root;
|
||||
|
||||
static int __init pci_eisa_init(struct pci_dev *pdev,
|
||||
const struct pci_device_id *ent)
|
||||
static int __init pci_eisa_init(struct pci_dev *pdev)
|
||||
{
|
||||
int rc;
|
||||
int rc, i;
|
||||
struct resource *res, *bus_res = NULL;
|
||||
|
||||
if ((rc = pci_enable_device (pdev))) {
|
||||
printk (KERN_ERR "pci_eisa : Could not enable device %s\n",
|
||||
@@ -30,9 +30,30 @@ static int __init pci_eisa_init(struct pci_dev *pdev,
|
||||
return rc;
|
||||
}
|
||||
|
||||
/*
|
||||
* The Intel 82375 PCI-EISA bridge is a subtractive-decode PCI
|
||||
* device, so the resources available on EISA are the same as those
|
||||
* available on the 82375 bus. This works the same as a PCI-PCI
|
||||
* bridge in subtractive-decode mode (see pci_read_bridge_bases()).
|
||||
* We assume other PCI-EISA bridges are similar.
|
||||
*
|
||||
* eisa_root_register() can only deal with a single io port resource,
|
||||
* so we use the first valid io port resource.
|
||||
*/
|
||||
pci_bus_for_each_resource(pdev->bus, res, i)
|
||||
if (res && (res->flags & IORESOURCE_IO)) {
|
||||
bus_res = res;
|
||||
break;
|
||||
}
|
||||
|
||||
if (!bus_res) {
|
||||
dev_err(&pdev->dev, "No resources available\n");
|
||||
return -1;
|
||||
}
|
||||
|
||||
pci_eisa_root.dev = &pdev->dev;
|
||||
pci_eisa_root.res = pdev->bus->resource[0];
|
||||
pci_eisa_root.bus_base_addr = pdev->bus->resource[0]->start;
|
||||
pci_eisa_root.res = bus_res;
|
||||
pci_eisa_root.bus_base_addr = bus_res->start;
|
||||
pci_eisa_root.slots = EISA_MAX_SLOTS;
|
||||
pci_eisa_root.dma_mask = pdev->dma_mask;
|
||||
dev_set_drvdata(pci_eisa_root.dev, &pci_eisa_root);
|
||||
@@ -45,22 +66,26 @@ static int __init pci_eisa_init(struct pci_dev *pdev,
|
||||
return 0;
|
||||
}
|
||||
|
||||
static struct pci_device_id pci_eisa_pci_tbl[] = {
|
||||
{ PCI_ANY_ID, PCI_ANY_ID, PCI_ANY_ID, PCI_ANY_ID,
|
||||
PCI_CLASS_BRIDGE_EISA << 8, 0xffff00, 0 },
|
||||
{ 0, }
|
||||
};
|
||||
|
||||
static struct pci_driver __refdata pci_eisa_driver = {
|
||||
.name = "pci_eisa",
|
||||
.id_table = pci_eisa_pci_tbl,
|
||||
.probe = pci_eisa_init,
|
||||
};
|
||||
|
||||
static int __init pci_eisa_init_module (void)
|
||||
/*
|
||||
* We have to call pci_eisa_init_early() before pnpacpi_init()/isapnp_init().
|
||||
* Otherwise pnp resource will get enabled early and could prevent eisa
|
||||
* to be initialized.
|
||||
* Also need to make sure pci_eisa_init_early() is called after
|
||||
* x86/pci_subsys_init().
|
||||
* So need to use subsys_initcall_sync with it.
|
||||
*/
|
||||
static int __init pci_eisa_init_early(void)
|
||||
{
|
||||
return pci_register_driver (&pci_eisa_driver);
|
||||
}
|
||||
struct pci_dev *dev = NULL;
|
||||
int ret;
|
||||
|
||||
device_initcall(pci_eisa_init_module);
|
||||
MODULE_DEVICE_TABLE(pci, pci_eisa_pci_tbl);
|
||||
for_each_pci_dev(dev)
|
||||
if ((dev->class >> 8) == PCI_CLASS_BRIDGE_EISA) {
|
||||
ret = pci_eisa_init(dev);
|
||||
if (ret)
|
||||
return ret;
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
subsys_initcall_sync(pci_eisa_init_early);
|
||||
|
||||
@@ -416,6 +416,8 @@ static int __hwspin_lock_request(struct hwspinlock *hwlock)
|
||||
ret = pm_runtime_get_sync(dev);
|
||||
if (ret < 0) {
|
||||
dev_err(dev, "%s: can't power on device\n", __func__);
|
||||
pm_runtime_put_noidle(dev);
|
||||
module_put(dev->driver->owner);
|
||||
return ret;
|
||||
}
|
||||
|
||||
|
||||
@@ -584,14 +584,6 @@ mwifiex_shutdown_drv(struct mwifiex_adapter *adapter)
|
||||
return ret;
|
||||
}
|
||||
|
||||
/* cancel current command */
|
||||
if (adapter->curr_cmd) {
|
||||
dev_warn(adapter->dev, "curr_cmd is still in processing\n");
|
||||
del_timer(&adapter->cmd_timer);
|
||||
mwifiex_insert_cmd_to_free_q(adapter, adapter->curr_cmd);
|
||||
adapter->curr_cmd = NULL;
|
||||
}
|
||||
|
||||
/* shut down mwifiex */
|
||||
dev_dbg(adapter->dev, "info: shutdown mwifiex...\n");
|
||||
|
||||
|
||||
@@ -52,8 +52,8 @@ int rt2x00pci_regbusy_read(struct rt2x00_dev *rt2x00dev,
|
||||
udelay(REGISTER_BUSY_DELAY);
|
||||
}
|
||||
|
||||
ERROR(rt2x00dev, "Indirect register access failed: "
|
||||
"offset=0x%.08x, value=0x%.08x\n", offset, *reg);
|
||||
printk_once(KERN_ERR "%s() Indirect register access failed: "
|
||||
"offset=0x%.08x, value=0x%.08x\n", __func__, offset, *reg);
|
||||
*reg = ~0;
|
||||
|
||||
return 0;
|
||||
|
||||
@@ -164,7 +164,7 @@ static int mpc512x_psc_spi_transfer_rxtx(struct spi_device *spi,
|
||||
|
||||
for (i = count; i > 0; i--) {
|
||||
data = tx_buf ? *tx_buf++ : 0;
|
||||
if (len == EOFBYTE)
|
||||
if (len == EOFBYTE && t->cs_change)
|
||||
setbits32(&fifo->txcmd, MPC512x_PSC_FIFO_EOF);
|
||||
out_8(&fifo->txdata_8, data);
|
||||
len--;
|
||||
|
||||
@@ -963,25 +963,30 @@ static irqreturn_t s3c64xx_spi_irq(int irq, void *data)
|
||||
{
|
||||
struct s3c64xx_spi_driver_data *sdd = data;
|
||||
struct spi_master *spi = sdd->master;
|
||||
unsigned int val;
|
||||
unsigned int val, clr = 0;
|
||||
|
||||
val = readl(sdd->regs + S3C64XX_SPI_PENDING_CLR);
|
||||
val = readl(sdd->regs + S3C64XX_SPI_STATUS);
|
||||
|
||||
val &= S3C64XX_SPI_PND_RX_OVERRUN_CLR |
|
||||
S3C64XX_SPI_PND_RX_UNDERRUN_CLR |
|
||||
S3C64XX_SPI_PND_TX_OVERRUN_CLR |
|
||||
S3C64XX_SPI_PND_TX_UNDERRUN_CLR;
|
||||
|
||||
writel(val, sdd->regs + S3C64XX_SPI_PENDING_CLR);
|
||||
|
||||
if (val & S3C64XX_SPI_PND_RX_OVERRUN_CLR)
|
||||
if (val & S3C64XX_SPI_ST_RX_OVERRUN_ERR) {
|
||||
clr = S3C64XX_SPI_PND_RX_OVERRUN_CLR;
|
||||
dev_err(&spi->dev, "RX overrun\n");
|
||||
if (val & S3C64XX_SPI_PND_RX_UNDERRUN_CLR)
|
||||
}
|
||||
if (val & S3C64XX_SPI_ST_RX_UNDERRUN_ERR) {
|
||||
clr |= S3C64XX_SPI_PND_RX_UNDERRUN_CLR;
|
||||
dev_err(&spi->dev, "RX underrun\n");
|
||||
if (val & S3C64XX_SPI_PND_TX_OVERRUN_CLR)
|
||||
}
|
||||
if (val & S3C64XX_SPI_ST_TX_OVERRUN_ERR) {
|
||||
clr |= S3C64XX_SPI_PND_TX_OVERRUN_CLR;
|
||||
dev_err(&spi->dev, "TX overrun\n");
|
||||
if (val & S3C64XX_SPI_PND_TX_UNDERRUN_CLR)
|
||||
}
|
||||
if (val & S3C64XX_SPI_ST_TX_UNDERRUN_ERR) {
|
||||
clr |= S3C64XX_SPI_PND_TX_UNDERRUN_CLR;
|
||||
dev_err(&spi->dev, "TX underrun\n");
|
||||
}
|
||||
|
||||
/* Clear the pending irq by setting and then clearing it */
|
||||
writel(clr, sdd->regs + S3C64XX_SPI_PENDING_CLR);
|
||||
writel(0, sdd->regs + S3C64XX_SPI_PENDING_CLR);
|
||||
|
||||
return IRQ_HANDLED;
|
||||
}
|
||||
@@ -1005,9 +1010,13 @@ static void s3c64xx_spi_hwinit(struct s3c64xx_spi_driver_data *sdd, int channel)
|
||||
writel(0, regs + S3C64XX_SPI_MODE_CFG);
|
||||
writel(0, regs + S3C64XX_SPI_PACKET_CNT);
|
||||
|
||||
/* Clear any irq pending bits */
|
||||
writel(readl(regs + S3C64XX_SPI_PENDING_CLR),
|
||||
regs + S3C64XX_SPI_PENDING_CLR);
|
||||
/* Clear any irq pending bits, should set and clear the bits */
|
||||
val = S3C64XX_SPI_PND_RX_OVERRUN_CLR |
|
||||
S3C64XX_SPI_PND_RX_UNDERRUN_CLR |
|
||||
S3C64XX_SPI_PND_TX_OVERRUN_CLR |
|
||||
S3C64XX_SPI_PND_TX_UNDERRUN_CLR;
|
||||
writel(val, regs + S3C64XX_SPI_PENDING_CLR);
|
||||
writel(0, regs + S3C64XX_SPI_PENDING_CLR);
|
||||
|
||||
writel(0, regs + S3C64XX_SPI_SWAP_CFG);
|
||||
|
||||
|
||||
@@ -1549,6 +1549,7 @@ static int __init thermal_init(void)
|
||||
idr_destroy(&thermal_cdev_idr);
|
||||
mutex_destroy(&thermal_idr_lock);
|
||||
mutex_destroy(&thermal_list_lock);
|
||||
return result;
|
||||
}
|
||||
result = genetlink_init();
|
||||
return result;
|
||||
|
||||
@@ -187,8 +187,8 @@ fill_with_dentries(void *buf, const char *name, int namelen, loff_t offset,
|
||||
if (dbuf->count == ARRAY_SIZE(dbuf->dentries))
|
||||
return -ENOSPC;
|
||||
|
||||
if (name[0] == '.' && (name[1] == '\0' ||
|
||||
(name[1] == '.' && name[2] == '\0')))
|
||||
if (name[0] == '.' && (namelen < 2 ||
|
||||
(namelen == 2 && name[1] == '.')))
|
||||
return 0;
|
||||
|
||||
dentry = lookup_one_len(name, dbuf->xadir, namelen);
|
||||
|
||||
@@ -1582,6 +1582,12 @@ static int ubifs_remount_rw(struct ubifs_info *c)
|
||||
c->remounting_rw = 1;
|
||||
c->ro_mount = 0;
|
||||
|
||||
if (c->space_fixup) {
|
||||
err = ubifs_fixup_free_space(c);
|
||||
if (err)
|
||||
return err;
|
||||
}
|
||||
|
||||
err = check_free_space(c);
|
||||
if (err)
|
||||
goto out;
|
||||
@@ -1698,12 +1704,6 @@ static int ubifs_remount_rw(struct ubifs_info *c)
|
||||
err = dbg_check_space_info(c);
|
||||
}
|
||||
|
||||
if (c->space_fixup) {
|
||||
err = ubifs_fixup_free_space(c);
|
||||
if (err)
|
||||
goto out;
|
||||
}
|
||||
|
||||
mutex_unlock(&c->umount_mutex);
|
||||
return err;
|
||||
|
||||
|
||||
@@ -937,7 +937,7 @@ static inline int atapi_cdb_len(const u16 *dev_id)
|
||||
}
|
||||
}
|
||||
|
||||
static inline bool atapi_command_packet_set(const u16 *dev_id)
|
||||
static inline int atapi_command_packet_set(const u16 *dev_id)
|
||||
{
|
||||
return (dev_id[ATA_ID_CONFIG] >> 8) & 0x1f;
|
||||
}
|
||||
|
||||
@@ -393,6 +393,7 @@ enum {
|
||||
ATA_HORKAGE_NOSETXFER = (1 << 14), /* skip SETXFER, SATA only */
|
||||
ATA_HORKAGE_BROKEN_FPDMA_AA = (1 << 15), /* skip AA */
|
||||
ATA_HORKAGE_DUMP_ID = (1 << 16), /* dump IDENTIFY data */
|
||||
ATA_HORKAGE_MAX_SEC_LBA48 = (1 << 17), /* Set max sects to 65535 */
|
||||
|
||||
/* DMA mask for user DMA control: User visible values; DO NOT
|
||||
renumber */
|
||||
|
||||
@@ -93,14 +93,20 @@ do { \
|
||||
|
||||
#else /* !CONFIG_PREEMPT_COUNT */
|
||||
|
||||
#define preempt_disable() do { } while (0)
|
||||
#define sched_preempt_enable_no_resched() do { } while (0)
|
||||
#define preempt_enable_no_resched() do { } while (0)
|
||||
#define preempt_enable() do { } while (0)
|
||||
/*
|
||||
* Even if we don't have any preemption, we need preempt disable/enable
|
||||
* to be barriers, so that we don't have things like get_user/put_user
|
||||
* that can cause faults and scheduling migrate into our preempt-protected
|
||||
* region.
|
||||
*/
|
||||
#define preempt_disable() barrier()
|
||||
#define sched_preempt_enable_no_resched() barrier()
|
||||
#define preempt_enable_no_resched() barrier()
|
||||
#define preempt_enable() barrier()
|
||||
|
||||
#define preempt_disable_notrace() do { } while (0)
|
||||
#define preempt_enable_no_resched_notrace() do { } while (0)
|
||||
#define preempt_enable_notrace() do { } while (0)
|
||||
#define preempt_disable_notrace() barrier()
|
||||
#define preempt_enable_no_resched_notrace() barrier()
|
||||
#define preempt_enable_notrace() barrier()
|
||||
|
||||
#endif /* CONFIG_PREEMPT_COUNT */
|
||||
|
||||
|
||||
@@ -16,7 +16,10 @@
|
||||
* In the debug case, 1 means unlocked, 0 means locked. (the values
|
||||
* are inverted, to catch initialization bugs)
|
||||
*
|
||||
* No atomicity anywhere, we are on UP.
|
||||
* No atomicity anywhere, we are on UP. However, we still need
|
||||
* the compiler barriers, because we do not want the compiler to
|
||||
* move potentially faulting instructions (notably user accesses)
|
||||
* into the locked sequence, resulting in non-atomic execution.
|
||||
*/
|
||||
|
||||
#ifdef CONFIG_DEBUG_SPINLOCK
|
||||
@@ -25,6 +28,7 @@
|
||||
static inline void arch_spin_lock(arch_spinlock_t *lock)
|
||||
{
|
||||
lock->slock = 0;
|
||||
barrier();
|
||||
}
|
||||
|
||||
static inline void
|
||||
@@ -32,6 +36,7 @@ arch_spin_lock_flags(arch_spinlock_t *lock, unsigned long flags)
|
||||
{
|
||||
local_irq_save(flags);
|
||||
lock->slock = 0;
|
||||
barrier();
|
||||
}
|
||||
|
||||
static inline int arch_spin_trylock(arch_spinlock_t *lock)
|
||||
@@ -39,32 +44,34 @@ static inline int arch_spin_trylock(arch_spinlock_t *lock)
|
||||
char oldval = lock->slock;
|
||||
|
||||
lock->slock = 0;
|
||||
barrier();
|
||||
|
||||
return oldval > 0;
|
||||
}
|
||||
|
||||
static inline void arch_spin_unlock(arch_spinlock_t *lock)
|
||||
{
|
||||
barrier();
|
||||
lock->slock = 1;
|
||||
}
|
||||
|
||||
/*
|
||||
* Read-write spinlocks. No debug version.
|
||||
*/
|
||||
#define arch_read_lock(lock) do { (void)(lock); } while (0)
|
||||
#define arch_write_lock(lock) do { (void)(lock); } while (0)
|
||||
#define arch_read_trylock(lock) ({ (void)(lock); 1; })
|
||||
#define arch_write_trylock(lock) ({ (void)(lock); 1; })
|
||||
#define arch_read_unlock(lock) do { (void)(lock); } while (0)
|
||||
#define arch_write_unlock(lock) do { (void)(lock); } while (0)
|
||||
#define arch_read_lock(lock) do { barrier(); (void)(lock); } while (0)
|
||||
#define arch_write_lock(lock) do { barrier(); (void)(lock); } while (0)
|
||||
#define arch_read_trylock(lock) ({ barrier(); (void)(lock); 1; })
|
||||
#define arch_write_trylock(lock) ({ barrier(); (void)(lock); 1; })
|
||||
#define arch_read_unlock(lock) do { barrier(); (void)(lock); } while (0)
|
||||
#define arch_write_unlock(lock) do { barrier(); (void)(lock); } while (0)
|
||||
|
||||
#else /* DEBUG_SPINLOCK */
|
||||
#define arch_spin_is_locked(lock) ((void)(lock), 0)
|
||||
/* for sched.c and kernel_lock.c: */
|
||||
# define arch_spin_lock(lock) do { (void)(lock); } while (0)
|
||||
# define arch_spin_lock_flags(lock, flags) do { (void)(lock); } while (0)
|
||||
# define arch_spin_unlock(lock) do { (void)(lock); } while (0)
|
||||
# define arch_spin_trylock(lock) ({ (void)(lock); 1; })
|
||||
# define arch_spin_lock(lock) do { barrier(); (void)(lock); } while (0)
|
||||
# define arch_spin_lock_flags(lock, flags) do { barrier(); (void)(lock); } while (0)
|
||||
# define arch_spin_unlock(lock) do { barrier(); (void)(lock); } while (0)
|
||||
# define arch_spin_trylock(lock) ({ barrier(); (void)(lock); 1; })
|
||||
#endif /* DEBUG_SPINLOCK */
|
||||
|
||||
#define arch_spin_is_contended(lock) (((void)(lock), 0))
|
||||
|
||||
@@ -80,6 +80,14 @@ void panic(const char *fmt, ...)
|
||||
long i, i_next = 0;
|
||||
int state = 0;
|
||||
|
||||
/*
|
||||
* Disable local interrupts. This will prevent panic_smp_self_stop
|
||||
* from deadlocking the first cpu that invokes the panic, since
|
||||
* there is nothing to prevent an interrupt handler (that runs
|
||||
* after the panic_lock is acquired) from invoking panic again.
|
||||
*/
|
||||
local_irq_disable();
|
||||
|
||||
/*
|
||||
* It's possible to come here directly from a panic-assertion and
|
||||
* not have preempt disabled. Some functions called from here want
|
||||
|
||||
@@ -4381,12 +4381,8 @@ ftrace_enable_sysctl(struct ctl_table *table, int write,
|
||||
ftrace_startup_sysctl();
|
||||
|
||||
/* we are starting ftrace again */
|
||||
if (ftrace_ops_list != &ftrace_list_end) {
|
||||
if (ftrace_ops_list->next == &ftrace_list_end)
|
||||
ftrace_trace_function = ftrace_ops_list->func;
|
||||
else
|
||||
ftrace_trace_function = ftrace_ops_list_func;
|
||||
}
|
||||
if (ftrace_ops_list != &ftrace_list_end)
|
||||
update_ftrace_function();
|
||||
|
||||
} else {
|
||||
/* stopping ftrace calls (just send to ftrace_stub) */
|
||||
|
||||
@@ -1619,7 +1619,7 @@ struct vm_area_struct *find_vma(struct mm_struct *mm, unsigned long addr)
|
||||
if (mm) {
|
||||
/* Check the cache first. */
|
||||
/* (Cache hit rate is typically around 35%.) */
|
||||
vma = mm->mmap_cache;
|
||||
vma = ACCESS_ONCE(mm->mmap_cache);
|
||||
if (!(vma && vma->vm_end > addr && vma->vm_start <= addr)) {
|
||||
struct rb_node * rb_node;
|
||||
|
||||
|
||||
@@ -807,7 +807,7 @@ struct vm_area_struct *find_vma(struct mm_struct *mm, unsigned long addr)
|
||||
struct vm_area_struct *vma;
|
||||
|
||||
/* check the cache first */
|
||||
vma = mm->mmap_cache;
|
||||
vma = ACCESS_ONCE(mm->mmap_cache);
|
||||
if (vma && vma->vm_start <= addr && vma->vm_end > addr)
|
||||
return vma;
|
||||
|
||||
|
||||
@@ -436,7 +436,7 @@ static int cgw_notifier(struct notifier_block *nb,
|
||||
if (gwj->src.dev == dev || gwj->dst.dev == dev) {
|
||||
hlist_del(&gwj->list);
|
||||
cgw_unregister_filter(gwj);
|
||||
kfree(gwj);
|
||||
kmem_cache_free(cgw_cache, gwj);
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -850,7 +850,7 @@ static void cgw_remove_all_jobs(void)
|
||||
hlist_for_each_entry_safe(gwj, n, nx, &cgw_list, list) {
|
||||
hlist_del(&gwj->list);
|
||||
cgw_unregister_filter(gwj);
|
||||
kfree(gwj);
|
||||
kmem_cache_free(cgw_cache, gwj);
|
||||
}
|
||||
}
|
||||
|
||||
@@ -903,7 +903,7 @@ static int cgw_remove_job(struct sk_buff *skb, struct nlmsghdr *nlh, void *arg)
|
||||
|
||||
hlist_del(&gwj->list);
|
||||
cgw_unregister_filter(gwj);
|
||||
kfree(gwj);
|
||||
kmem_cache_free(cgw_cache, gwj);
|
||||
err = 0;
|
||||
break;
|
||||
}
|
||||
|
||||
@@ -165,7 +165,7 @@ const char *snd_hda_get_jack_type(u32 cfg)
|
||||
"Line Out", "Speaker", "HP Out", "CD",
|
||||
"SPDIF Out", "Digital Out", "Modem Line", "Modem Hand",
|
||||
"Line In", "Aux", "Mic", "Telephony",
|
||||
"SPDIF In", "Digitial In", "Reserved", "Other"
|
||||
"SPDIF In", "Digital In", "Reserved", "Other"
|
||||
};
|
||||
|
||||
return jack_types[(cfg & AC_DEFCFG_DEVICE)
|
||||
|
||||
@@ -322,7 +322,7 @@ int snd_hdmi_get_eld(struct hdmi_eld *eld,
|
||||
struct hda_codec *codec, hda_nid_t nid)
|
||||
{
|
||||
int i;
|
||||
int ret;
|
||||
int ret = 0;
|
||||
int size;
|
||||
unsigned char *buf;
|
||||
|
||||
|
||||
@@ -6629,7 +6629,8 @@ static int alc662_parse_auto_config(struct hda_codec *codec)
|
||||
const hda_nid_t *ssids;
|
||||
|
||||
if (codec->vendor_id == 0x10ec0272 || codec->vendor_id == 0x10ec0663 ||
|
||||
codec->vendor_id == 0x10ec0665 || codec->vendor_id == 0x10ec0670)
|
||||
codec->vendor_id == 0x10ec0665 || codec->vendor_id == 0x10ec0670 ||
|
||||
codec->vendor_id == 0x10ec0671)
|
||||
ssids = alc663_ssids;
|
||||
else
|
||||
ssids = alc662_ssids;
|
||||
@@ -7056,6 +7057,7 @@ static const struct hda_codec_preset snd_hda_preset_realtek[] = {
|
||||
{ .id = 0x10ec0665, .name = "ALC665", .patch = patch_alc662 },
|
||||
{ .id = 0x10ec0668, .name = "ALC668", .patch = patch_alc662 },
|
||||
{ .id = 0x10ec0670, .name = "ALC670", .patch = patch_alc662 },
|
||||
{ .id = 0x10ec0671, .name = "ALC671", .patch = patch_alc662 },
|
||||
{ .id = 0x10ec0680, .name = "ALC680", .patch = patch_alc680 },
|
||||
{ .id = 0x10ec0880, .name = "ALC880", .patch = patch_alc880 },
|
||||
{ .id = 0x10ec0882, .name = "ALC882", .patch = patch_alc882 },
|
||||
|
||||
@@ -342,8 +342,8 @@ static int camelot_pcm_new(struct snd_soc_pcm_runtime *rtd)
|
||||
return 0;
|
||||
}
|
||||
|
||||
static struct snd_soc_platform sh7760_soc_platform = {
|
||||
.pcm_ops = &camelot_pcm_ops,
|
||||
static struct snd_soc_platform_driver sh7760_soc_platform = {
|
||||
.ops = &camelot_pcm_ops,
|
||||
.pcm_new = camelot_pcm_new,
|
||||
.pcm_free = camelot_pcm_free,
|
||||
};
|
||||
|
||||
Reference in New Issue
Block a user