mirror of
https://github.com/hardkernel/linux.git
synced 2026-03-24 19:40:21 +09:00
Merge tag 'v4.9.277' of git://git.kernel.org/pub/scm/linux/kernel/git/stable/linux-stable into odroidg12-4.9.y
This is the 4.9.277 stable release Change-Id: If6b64078940d789b507791c7bbe010b5de2765cb
This commit is contained in:
2
Makefile
2
Makefile
@@ -1,6 +1,6 @@
|
||||
VERSION = 4
|
||||
PATCHLEVEL = 9
|
||||
SUBLEVEL = 276
|
||||
SUBLEVEL = 277
|
||||
EXTRAVERSION =
|
||||
NAME = Roaring Lionus
|
||||
|
||||
|
||||
@@ -174,7 +174,7 @@
|
||||
status = "disabled";
|
||||
};
|
||||
|
||||
nand: nand@2000 {
|
||||
nand_controller: nand-controller@2000 {
|
||||
#address-cells = <1>;
|
||||
#size-cells = <0>;
|
||||
compatible = "brcm,nand-bcm63138", "brcm,brcmnand-v7.0", "brcm,brcmnand";
|
||||
|
||||
@@ -13,10 +13,10 @@
|
||||
};
|
||||
};
|
||||
|
||||
&nand {
|
||||
&nand_controller {
|
||||
status = "okay";
|
||||
|
||||
nandcs@1 {
|
||||
nand@1 {
|
||||
compatible = "brcm,nandcs";
|
||||
reg = <1>;
|
||||
nand-ecc-step-size = <512>;
|
||||
|
||||
@@ -149,7 +149,7 @@
|
||||
reg-names = "aon-ctrl", "aon-sram";
|
||||
};
|
||||
|
||||
nand: nand@3e2800 {
|
||||
nand_controller: nand-controller@3e2800 {
|
||||
status = "disabled";
|
||||
#address-cells = <1>;
|
||||
#size-cells = <0>;
|
||||
|
||||
@@ -29,10 +29,10 @@
|
||||
status = "okay";
|
||||
};
|
||||
|
||||
&nand {
|
||||
&nand_controller {
|
||||
status = "okay";
|
||||
|
||||
nandcs@0 {
|
||||
nand@0 {
|
||||
compatible = "brcm,nandcs";
|
||||
reg = <0>;
|
||||
nand-ecc-strength = <4>;
|
||||
|
||||
@@ -307,8 +307,8 @@
|
||||
fsl,pins = <
|
||||
MX6QDL_PAD_EIM_D24__UART3_TX_DATA 0x1b0b1
|
||||
MX6QDL_PAD_EIM_D25__UART3_RX_DATA 0x1b0b1
|
||||
MX6QDL_PAD_EIM_D30__UART3_RTS_B 0x1b0b1
|
||||
MX6QDL_PAD_EIM_D31__UART3_CTS_B 0x1b0b1
|
||||
MX6QDL_PAD_EIM_D31__UART3_RTS_B 0x1b0b1
|
||||
MX6QDL_PAD_EIM_D30__UART3_CTS_B 0x1b0b1
|
||||
>;
|
||||
};
|
||||
|
||||
@@ -395,6 +395,7 @@
|
||||
&uart3 {
|
||||
pinctrl-names = "default";
|
||||
pinctrl-0 = <&pinctrl_uart3>;
|
||||
uart-has-rtscts;
|
||||
status = "disabled";
|
||||
};
|
||||
|
||||
|
||||
@@ -426,7 +426,7 @@
|
||||
};
|
||||
};
|
||||
|
||||
sleep {
|
||||
suspend {
|
||||
global_pwroff: global-pwroff {
|
||||
rockchip,pins = <2 7 RK_FUNC_1 &pcfg_pull_none>;
|
||||
};
|
||||
|
||||
@@ -735,7 +735,7 @@
|
||||
* *_HDMI HDMI
|
||||
* *_MIPI_* MIPI
|
||||
*/
|
||||
pd_vio@RK3288_PD_VIO {
|
||||
power-domain@RK3288_PD_VIO {
|
||||
reg = <RK3288_PD_VIO>;
|
||||
clocks = <&cru ACLK_IEP>,
|
||||
<&cru ACLK_ISP>,
|
||||
@@ -768,7 +768,7 @@
|
||||
* Note: The following 3 are HEVC(H.265) clocks,
|
||||
* and on the ACLK_HEVC_NIU (NOC).
|
||||
*/
|
||||
pd_hevc@RK3288_PD_HEVC {
|
||||
power-domain@RK3288_PD_HEVC {
|
||||
reg = <RK3288_PD_HEVC>;
|
||||
clocks = <&cru ACLK_HEVC>,
|
||||
<&cru SCLK_HEVC_CABAC>,
|
||||
@@ -780,7 +780,7 @@
|
||||
* (video endecoder & decoder) clocks that on the
|
||||
* ACLK_VCODEC_NIU and HCLK_VCODEC_NIU (NOC).
|
||||
*/
|
||||
pd_video@RK3288_PD_VIDEO {
|
||||
power-domain@RK3288_PD_VIDEO {
|
||||
reg = <RK3288_PD_VIDEO>;
|
||||
clocks = <&cru ACLK_VCODEC>,
|
||||
<&cru HCLK_VCODEC>;
|
||||
@@ -790,7 +790,7 @@
|
||||
* Note: ACLK_GPU is the GPU clock,
|
||||
* and on the ACLK_GPU_NIU (NOC).
|
||||
*/
|
||||
pd_gpu@RK3288_PD_GPU {
|
||||
power-domain@RK3288_PD_GPU {
|
||||
reg = <RK3288_PD_GPU>;
|
||||
clocks = <&cru ACLK_GPU>;
|
||||
};
|
||||
@@ -1278,7 +1278,7 @@
|
||||
drive-strength = <12>;
|
||||
};
|
||||
|
||||
sleep {
|
||||
suspend {
|
||||
global_pwroff: global-pwroff {
|
||||
rockchip,pins = <0 0 RK_FUNC_1 &pcfg_pull_none>;
|
||||
};
|
||||
|
||||
@@ -334,7 +334,7 @@
|
||||
};
|
||||
};
|
||||
|
||||
rcc: rcc@40023810 {
|
||||
rcc: rcc@40023800 {
|
||||
#reset-cells = <1>;
|
||||
#clock-cells = <2>;
|
||||
compatible = "st,stm32f42xx-rcc", "st,stm32-rcc";
|
||||
|
||||
@@ -33,11 +33,11 @@
|
||||
* ^
|
||||
* ^
|
||||
* imx53_suspend code
|
||||
* PM_INFO structure(imx53_suspend_info)
|
||||
* PM_INFO structure(imx5_cpu_suspend_info)
|
||||
* ======================== low address =======================
|
||||
*/
|
||||
|
||||
/* Offsets of members of struct imx53_suspend_info */
|
||||
/* Offsets of members of struct imx5_cpu_suspend_info */
|
||||
#define SUSPEND_INFO_MX53_M4IF_V_OFFSET 0x0
|
||||
#define SUSPEND_INFO_MX53_IOMUXC_V_OFFSET 0x4
|
||||
#define SUSPEND_INFO_MX53_IO_COUNT_OFFSET 0x8
|
||||
|
||||
@@ -414,13 +414,13 @@
|
||||
clocks {
|
||||
compatible = "arm,scpi-clocks";
|
||||
|
||||
scpi_dvfs: scpi-dvfs {
|
||||
scpi_dvfs: clocks-0 {
|
||||
compatible = "arm,scpi-dvfs-clocks";
|
||||
#clock-cells = <1>;
|
||||
clock-indices = <0>, <1>, <2>;
|
||||
clock-output-names = "atlclk", "aplclk","gpuclk";
|
||||
};
|
||||
scpi_clk: scpi-clk {
|
||||
scpi_clk: clocks-1 {
|
||||
compatible = "arm,scpi-variable-clocks";
|
||||
#clock-cells = <1>;
|
||||
clock-indices = <3>;
|
||||
@@ -428,7 +428,7 @@
|
||||
};
|
||||
};
|
||||
|
||||
scpi_devpd: scpi-power-domains {
|
||||
scpi_devpd: power-controller {
|
||||
compatible = "arm,scpi-power-domains";
|
||||
num-domains = <2>;
|
||||
#power-domain-cells = <1>;
|
||||
|
||||
@@ -107,15 +107,11 @@ do { \
|
||||
|
||||
static inline pmd_t *pmd_alloc_one(struct mm_struct *mm, unsigned long address)
|
||||
{
|
||||
pmd_t *pmd = NULL;
|
||||
struct page *pg;
|
||||
pmd_t *pmd;
|
||||
|
||||
pg = alloc_pages(GFP_KERNEL | __GFP_ACCOUNT, PMD_ORDER);
|
||||
if (pg) {
|
||||
pgtable_pmd_page_ctor(pg);
|
||||
pmd = (pmd_t *)page_address(pg);
|
||||
pmd = (pmd_t *) __get_free_pages(GFP_KERNEL, PMD_ORDER);
|
||||
if (pmd)
|
||||
pmd_init((unsigned long)pmd, (unsigned long)invalid_pte_table);
|
||||
}
|
||||
return pmd;
|
||||
}
|
||||
|
||||
|
||||
@@ -230,6 +230,17 @@ int kvmppc_rtas_hcall(struct kvm_vcpu *vcpu)
|
||||
* value so we can restore it on the way out.
|
||||
*/
|
||||
orig_rets = args.rets;
|
||||
if (be32_to_cpu(args.nargs) >= ARRAY_SIZE(args.args)) {
|
||||
/*
|
||||
* Don't overflow our args array: ensure there is room for
|
||||
* at least rets[0] (even if the call specifies 0 nret).
|
||||
*
|
||||
* Each handler must then check for the correct nargs and nret
|
||||
* values, but they may always return failure in rets[0].
|
||||
*/
|
||||
rc = -EINVAL;
|
||||
goto fail;
|
||||
}
|
||||
args.rets = &args.args[be32_to_cpu(args.nargs)];
|
||||
|
||||
mutex_lock(&vcpu->kvm->arch.rtas_token_lock);
|
||||
@@ -257,9 +268,17 @@ int kvmppc_rtas_hcall(struct kvm_vcpu *vcpu)
|
||||
fail:
|
||||
/*
|
||||
* We only get here if the guest has called RTAS with a bogus
|
||||
* args pointer. That means we can't get to the args, and so we
|
||||
* can't fail the RTAS call. So fail right out to userspace,
|
||||
* which should kill the guest.
|
||||
* args pointer or nargs/nret values that would overflow the
|
||||
* array. That means we can't get to the args, and so we can't
|
||||
* fail the RTAS call. So fail right out to userspace, which
|
||||
* should kill the guest.
|
||||
*
|
||||
* SLOF should actually pass the hcall return value from the
|
||||
* rtas handler call in r3, so enter_rtas could be modified to
|
||||
* return a failure indication in r3 and we could return such
|
||||
* errors to the guest rather than failing to host userspace.
|
||||
* However old guests that don't test for failure could then
|
||||
* continue silently after errors, so for now we won't do this.
|
||||
*/
|
||||
return rc;
|
||||
}
|
||||
|
||||
@@ -19,6 +19,7 @@ void ftrace_caller(void);
|
||||
|
||||
extern char ftrace_graph_caller_end;
|
||||
extern unsigned long ftrace_plt;
|
||||
extern void *ftrace_func;
|
||||
|
||||
struct dyn_arch_ftrace { };
|
||||
|
||||
|
||||
@@ -55,6 +55,7 @@
|
||||
* > brasl %r0,ftrace_caller # offset 0
|
||||
*/
|
||||
|
||||
void *ftrace_func __read_mostly = ftrace_stub;
|
||||
unsigned long ftrace_plt;
|
||||
|
||||
static inline void ftrace_generate_orig_insn(struct ftrace_insn *insn)
|
||||
@@ -164,6 +165,7 @@ int ftrace_make_call(struct dyn_ftrace *rec, unsigned long addr)
|
||||
|
||||
int ftrace_update_ftrace_func(ftrace_func_t func)
|
||||
{
|
||||
ftrace_func = func;
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
||||
@@ -59,13 +59,13 @@ ENTRY(ftrace_caller)
|
||||
#ifdef CONFIG_HAVE_MARCH_Z196_FEATURES
|
||||
aghik %r2,%r0,-MCOUNT_INSN_SIZE
|
||||
lgrl %r4,function_trace_op
|
||||
lgrl %r1,ftrace_trace_function
|
||||
lgrl %r1,ftrace_func
|
||||
#else
|
||||
lgr %r2,%r0
|
||||
aghi %r2,-MCOUNT_INSN_SIZE
|
||||
larl %r4,function_trace_op
|
||||
lg %r4,0(%r4)
|
||||
larl %r1,ftrace_trace_function
|
||||
larl %r1,ftrace_func
|
||||
lg %r1,0(%r1)
|
||||
#endif
|
||||
lgr %r3,%r14
|
||||
|
||||
@@ -116,7 +116,7 @@ static inline void reg_set_seen(struct bpf_jit *jit, u32 b1)
|
||||
{
|
||||
u32 r1 = reg2hex[b1];
|
||||
|
||||
if (!jit->seen_reg[r1] && r1 >= 6 && r1 <= 15)
|
||||
if (r1 >= 6 && r1 <= 15 && !jit->seen_reg[r1])
|
||||
jit->seen_reg[r1] = 1;
|
||||
}
|
||||
|
||||
|
||||
@@ -49,7 +49,7 @@ struct bma180_part_info {
|
||||
|
||||
u8 int_reset_reg, int_reset_mask;
|
||||
u8 sleep_reg, sleep_mask;
|
||||
u8 bw_reg, bw_mask;
|
||||
u8 bw_reg, bw_mask, bw_offset;
|
||||
u8 scale_reg, scale_mask;
|
||||
u8 power_reg, power_mask, lowpower_val;
|
||||
u8 int_enable_reg, int_enable_mask;
|
||||
@@ -105,6 +105,7 @@ struct bma180_part_info {
|
||||
|
||||
#define BMA250_RANGE_MASK GENMASK(3, 0) /* Range of accel values */
|
||||
#define BMA250_BW_MASK GENMASK(4, 0) /* Accel bandwidth */
|
||||
#define BMA250_BW_OFFSET 8
|
||||
#define BMA250_SUSPEND_MASK BIT(7) /* chip will sleep */
|
||||
#define BMA250_LOWPOWER_MASK BIT(6)
|
||||
#define BMA250_DATA_INTEN_MASK BIT(4)
|
||||
@@ -242,7 +243,8 @@ static int bma180_set_bw(struct bma180_data *data, int val)
|
||||
for (i = 0; i < data->part_info->num_bw; ++i) {
|
||||
if (data->part_info->bw_table[i] == val) {
|
||||
ret = bma180_set_bits(data, data->part_info->bw_reg,
|
||||
data->part_info->bw_mask, i);
|
||||
data->part_info->bw_mask,
|
||||
i + data->part_info->bw_offset);
|
||||
if (ret) {
|
||||
dev_err(&data->client->dev,
|
||||
"failed to set bandwidth\n");
|
||||
@@ -625,32 +627,53 @@ static const struct iio_chan_spec bma250_channels[] = {
|
||||
|
||||
static const struct bma180_part_info bma180_part_info[] = {
|
||||
[BMA180] = {
|
||||
bma180_channels, ARRAY_SIZE(bma180_channels),
|
||||
bma180_scale_table, ARRAY_SIZE(bma180_scale_table),
|
||||
bma180_bw_table, ARRAY_SIZE(bma180_bw_table),
|
||||
BMA180_CTRL_REG0, BMA180_RESET_INT,
|
||||
BMA180_CTRL_REG0, BMA180_SLEEP,
|
||||
BMA180_BW_TCS, BMA180_BW,
|
||||
BMA180_OFFSET_LSB1, BMA180_RANGE,
|
||||
BMA180_TCO_Z, BMA180_MODE_CONFIG, BMA180_LOW_POWER,
|
||||
BMA180_CTRL_REG3, BMA180_NEW_DATA_INT,
|
||||
BMA180_RESET,
|
||||
bma180_chip_config,
|
||||
bma180_chip_disable,
|
||||
.channels = bma180_channels,
|
||||
.num_channels = ARRAY_SIZE(bma180_channels),
|
||||
.scale_table = bma180_scale_table,
|
||||
.num_scales = ARRAY_SIZE(bma180_scale_table),
|
||||
.bw_table = bma180_bw_table,
|
||||
.num_bw = ARRAY_SIZE(bma180_bw_table),
|
||||
.int_reset_reg = BMA180_CTRL_REG0,
|
||||
.int_reset_mask = BMA180_RESET_INT,
|
||||
.sleep_reg = BMA180_CTRL_REG0,
|
||||
.sleep_mask = BMA180_SLEEP,
|
||||
.bw_reg = BMA180_BW_TCS,
|
||||
.bw_mask = BMA180_BW,
|
||||
.scale_reg = BMA180_OFFSET_LSB1,
|
||||
.scale_mask = BMA180_RANGE,
|
||||
.power_reg = BMA180_TCO_Z,
|
||||
.power_mask = BMA180_MODE_CONFIG,
|
||||
.lowpower_val = BMA180_LOW_POWER,
|
||||
.int_enable_reg = BMA180_CTRL_REG3,
|
||||
.int_enable_mask = BMA180_NEW_DATA_INT,
|
||||
.softreset_reg = BMA180_RESET,
|
||||
.chip_config = bma180_chip_config,
|
||||
.chip_disable = bma180_chip_disable,
|
||||
},
|
||||
[BMA250] = {
|
||||
bma250_channels, ARRAY_SIZE(bma250_channels),
|
||||
bma250_scale_table, ARRAY_SIZE(bma250_scale_table),
|
||||
bma250_bw_table, ARRAY_SIZE(bma250_bw_table),
|
||||
BMA250_INT_RESET_REG, BMA250_INT_RESET_MASK,
|
||||
BMA250_POWER_REG, BMA250_SUSPEND_MASK,
|
||||
BMA250_BW_REG, BMA250_BW_MASK,
|
||||
BMA250_RANGE_REG, BMA250_RANGE_MASK,
|
||||
BMA250_POWER_REG, BMA250_LOWPOWER_MASK, 1,
|
||||
BMA250_INT_ENABLE_REG, BMA250_DATA_INTEN_MASK,
|
||||
BMA250_RESET_REG,
|
||||
bma250_chip_config,
|
||||
bma250_chip_disable,
|
||||
.channels = bma250_channels,
|
||||
.num_channels = ARRAY_SIZE(bma250_channels),
|
||||
.scale_table = bma250_scale_table,
|
||||
.num_scales = ARRAY_SIZE(bma250_scale_table),
|
||||
.bw_table = bma250_bw_table,
|
||||
.num_bw = ARRAY_SIZE(bma250_bw_table),
|
||||
.int_reset_reg = BMA250_INT_RESET_REG,
|
||||
.int_reset_mask = BMA250_INT_RESET_MASK,
|
||||
.sleep_reg = BMA250_POWER_REG,
|
||||
.sleep_mask = BMA250_SUSPEND_MASK,
|
||||
.bw_reg = BMA250_BW_REG,
|
||||
.bw_mask = BMA250_BW_MASK,
|
||||
.bw_offset = BMA250_BW_OFFSET,
|
||||
.scale_reg = BMA250_RANGE_REG,
|
||||
.scale_mask = BMA250_RANGE_MASK,
|
||||
.power_reg = BMA250_POWER_REG,
|
||||
.power_mask = BMA250_LOWPOWER_MASK,
|
||||
.lowpower_val = 1,
|
||||
.int_enable_reg = BMA250_INT_ENABLE_REG,
|
||||
.int_enable_mask = BMA250_DATA_INTEN_MASK,
|
||||
.softreset_reg = BMA250_RESET_REG,
|
||||
.chip_config = bma250_chip_config,
|
||||
.chip_disable = bma250_chip_disable,
|
||||
},
|
||||
};
|
||||
|
||||
|
||||
@@ -402,7 +402,7 @@ static int ngene_command_config_free_buf(struct ngene *dev, u8 *config)
|
||||
|
||||
com.cmd.hdr.Opcode = CMD_CONFIGURE_FREE_BUFFER;
|
||||
com.cmd.hdr.Length = 6;
|
||||
memcpy(&com.cmd.ConfigureBuffers.config, config, 6);
|
||||
memcpy(&com.cmd.ConfigureFreeBuffers.config, config, 6);
|
||||
com.in_len = 6;
|
||||
com.out_len = 0;
|
||||
|
||||
|
||||
@@ -407,12 +407,14 @@ enum _BUFFER_CONFIGS {
|
||||
|
||||
struct FW_CONFIGURE_FREE_BUFFERS {
|
||||
struct FW_HEADER hdr;
|
||||
u8 UVI1_BufferLength;
|
||||
u8 UVI2_BufferLength;
|
||||
u8 TVO_BufferLength;
|
||||
u8 AUD1_BufferLength;
|
||||
u8 AUD2_BufferLength;
|
||||
u8 TVA_BufferLength;
|
||||
struct {
|
||||
u8 UVI1_BufferLength;
|
||||
u8 UVI2_BufferLength;
|
||||
u8 TVO_BufferLength;
|
||||
u8 AUD1_BufferLength;
|
||||
u8 AUD2_BufferLength;
|
||||
u8 TVA_BufferLength;
|
||||
} __packed config;
|
||||
} __attribute__ ((__packed__));
|
||||
|
||||
struct FW_CONFIGURE_UART {
|
||||
|
||||
@@ -1094,7 +1094,7 @@ static void bcmgenet_power_up(struct bcmgenet_priv *priv,
|
||||
switch (mode) {
|
||||
case GENET_POWER_PASSIVE:
|
||||
reg &= ~(EXT_PWR_DOWN_DLL | EXT_PWR_DOWN_PHY |
|
||||
EXT_PWR_DOWN_BIAS);
|
||||
EXT_PWR_DOWN_BIAS | EXT_ENERGY_DET_MASK);
|
||||
/* fallthrough */
|
||||
case GENET_POWER_CABLE_SENSE:
|
||||
/* enable APD */
|
||||
@@ -2698,15 +2698,21 @@ static void bcmgenet_set_hw_addr(struct bcmgenet_priv *priv,
|
||||
/* Returns a reusable dma control register value */
|
||||
static u32 bcmgenet_dma_disable(struct bcmgenet_priv *priv)
|
||||
{
|
||||
unsigned int i;
|
||||
u32 reg;
|
||||
u32 dma_ctrl;
|
||||
|
||||
/* disable DMA */
|
||||
dma_ctrl = 1 << (DESC_INDEX + DMA_RING_BUF_EN_SHIFT) | DMA_EN;
|
||||
for (i = 0; i < priv->hw_params->tx_queues; i++)
|
||||
dma_ctrl |= (1 << (i + DMA_RING_BUF_EN_SHIFT));
|
||||
reg = bcmgenet_tdma_readl(priv, DMA_CTRL);
|
||||
reg &= ~dma_ctrl;
|
||||
bcmgenet_tdma_writel(priv, reg, DMA_CTRL);
|
||||
|
||||
dma_ctrl = 1 << (DESC_INDEX + DMA_RING_BUF_EN_SHIFT) | DMA_EN;
|
||||
for (i = 0; i < priv->hw_params->rx_queues; i++)
|
||||
dma_ctrl |= (1 << (i + DMA_RING_BUF_EN_SHIFT));
|
||||
reg = bcmgenet_rdma_readl(priv, DMA_CTRL);
|
||||
reg &= ~dma_ctrl;
|
||||
bcmgenet_rdma_writel(priv, reg, DMA_CTRL);
|
||||
@@ -2815,12 +2821,6 @@ static int bcmgenet_open(struct net_device *dev)
|
||||
|
||||
bcmgenet_set_hw_addr(priv, dev->dev_addr);
|
||||
|
||||
if (priv->internal_phy) {
|
||||
reg = bcmgenet_ext_readl(priv, EXT_EXT_PWR_MGMT);
|
||||
reg |= EXT_ENERGY_DET_MASK;
|
||||
bcmgenet_ext_writel(priv, reg, EXT_EXT_PWR_MGMT);
|
||||
}
|
||||
|
||||
/* Disable RX/TX DMA and flush TX queues */
|
||||
dma_ctrl = bcmgenet_dma_disable(priv);
|
||||
|
||||
@@ -3510,7 +3510,6 @@ static int bcmgenet_resume(struct device *d)
|
||||
struct bcmgenet_priv *priv = netdev_priv(dev);
|
||||
unsigned long dma_ctrl;
|
||||
int ret;
|
||||
u32 reg;
|
||||
|
||||
if (!netif_running(dev))
|
||||
return 0;
|
||||
@@ -3545,12 +3544,6 @@ static int bcmgenet_resume(struct device *d)
|
||||
|
||||
bcmgenet_set_hw_addr(priv, dev->dev_addr);
|
||||
|
||||
if (priv->internal_phy) {
|
||||
reg = bcmgenet_ext_readl(priv, EXT_EXT_PWR_MGMT);
|
||||
reg |= EXT_ENERGY_DET_MASK;
|
||||
bcmgenet_ext_writel(priv, reg, EXT_EXT_PWR_MGMT);
|
||||
}
|
||||
|
||||
if (priv->wolopts)
|
||||
bcmgenet_power_up(priv, GENET_POWER_WOL_MAGIC);
|
||||
|
||||
|
||||
@@ -167,12 +167,6 @@ int bcmgenet_wol_power_down_cfg(struct bcmgenet_priv *priv,
|
||||
reg |= CMD_RX_EN;
|
||||
bcmgenet_umac_writel(priv, reg, UMAC_CMD);
|
||||
|
||||
if (priv->hw_params->flags & GENET_HAS_EXT) {
|
||||
reg = bcmgenet_ext_readl(priv, EXT_EXT_PWR_MGMT);
|
||||
reg &= ~EXT_ENERGY_DET_MASK;
|
||||
bcmgenet_ext_writel(priv, reg, EXT_EXT_PWR_MGMT);
|
||||
}
|
||||
|
||||
/* Enable the MPD interrupt */
|
||||
cpu_mask_clear = UMAC_IRQ_MPD_R;
|
||||
|
||||
|
||||
@@ -7375,6 +7375,7 @@ err_flashmap:
|
||||
err_ioremap:
|
||||
free_netdev(netdev);
|
||||
err_alloc_etherdev:
|
||||
pci_disable_pcie_error_reporting(pdev);
|
||||
pci_release_mem_regions(pdev);
|
||||
err_pci_reg:
|
||||
err_dma:
|
||||
|
||||
@@ -2083,6 +2083,7 @@ err_sw_init:
|
||||
err_ioremap:
|
||||
free_netdev(netdev);
|
||||
err_alloc_netdev:
|
||||
pci_disable_pcie_error_reporting(pdev);
|
||||
pci_release_mem_regions(pdev);
|
||||
err_pci_reg:
|
||||
err_dma:
|
||||
|
||||
@@ -2641,6 +2641,7 @@ static int i40evf_probe(struct pci_dev *pdev, const struct pci_device_id *ent)
|
||||
err_ioremap:
|
||||
free_netdev(netdev);
|
||||
err_alloc_etherdev:
|
||||
pci_disable_pcie_error_reporting(pdev);
|
||||
pci_release_regions(pdev);
|
||||
err_pci_reg:
|
||||
err_dma:
|
||||
|
||||
@@ -948,6 +948,7 @@ static void igb_configure_msix(struct igb_adapter *adapter)
|
||||
**/
|
||||
static int igb_request_msix(struct igb_adapter *adapter)
|
||||
{
|
||||
unsigned int num_q_vectors = adapter->num_q_vectors;
|
||||
struct net_device *netdev = adapter->netdev;
|
||||
int i, err = 0, vector = 0, free_vector = 0;
|
||||
|
||||
@@ -956,7 +957,13 @@ static int igb_request_msix(struct igb_adapter *adapter)
|
||||
if (err)
|
||||
goto err_out;
|
||||
|
||||
for (i = 0; i < adapter->num_q_vectors; i++) {
|
||||
if (num_q_vectors > MAX_Q_VECTORS) {
|
||||
num_q_vectors = MAX_Q_VECTORS;
|
||||
dev_warn(&adapter->pdev->dev,
|
||||
"The number of queue vectors (%d) is higher than max allowed (%d)\n",
|
||||
adapter->num_q_vectors, MAX_Q_VECTORS);
|
||||
}
|
||||
for (i = 0; i < num_q_vectors; i++) {
|
||||
struct igb_q_vector *q_vector = adapter->q_vector[i];
|
||||
|
||||
vector++;
|
||||
@@ -2767,6 +2774,7 @@ err_sw_init:
|
||||
err_ioremap:
|
||||
free_netdev(netdev);
|
||||
err_alloc_etherdev:
|
||||
pci_disable_pcie_error_reporting(pdev);
|
||||
pci_release_mem_regions(pdev);
|
||||
err_pci_reg:
|
||||
err_dma:
|
||||
|
||||
@@ -9840,6 +9840,7 @@ err_ioremap:
|
||||
disable_dev = !test_and_set_bit(__IXGBE_DISABLED, &adapter->state);
|
||||
free_netdev(netdev);
|
||||
err_alloc_etherdev:
|
||||
pci_disable_pcie_error_reporting(pdev);
|
||||
pci_release_mem_regions(pdev);
|
||||
err_pci_reg:
|
||||
err_dma:
|
||||
|
||||
@@ -548,10 +548,8 @@ static int moxart_mac_probe(struct platform_device *pdev)
|
||||
SET_NETDEV_DEV(ndev, &pdev->dev);
|
||||
|
||||
ret = register_netdev(ndev);
|
||||
if (ret) {
|
||||
free_netdev(ndev);
|
||||
if (ret)
|
||||
goto init_fail;
|
||||
}
|
||||
|
||||
netdev_dbg(ndev, "%s: IRQ=%d address=%pM\n",
|
||||
__func__, ndev->irq, ndev->dev_addr);
|
||||
|
||||
@@ -746,12 +746,13 @@ static int emac_remove(struct platform_device *pdev)
|
||||
if (!has_acpi_companion(&pdev->dev))
|
||||
put_device(&adpt->phydev->mdio.dev);
|
||||
mdiobus_unregister(adpt->mii_bus);
|
||||
free_netdev(netdev);
|
||||
|
||||
if (adpt->phy.digital)
|
||||
iounmap(adpt->phy.digital);
|
||||
iounmap(adpt->phy.base);
|
||||
|
||||
free_netdev(netdev);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
||||
@@ -313,9 +313,8 @@ static void tlan_remove_one(struct pci_dev *pdev)
|
||||
pci_release_regions(pdev);
|
||||
#endif
|
||||
|
||||
free_netdev(dev);
|
||||
|
||||
cancel_work_sync(&priv->tlan_tqueue);
|
||||
free_netdev(dev);
|
||||
}
|
||||
|
||||
static void tlan_start(struct net_device *dev)
|
||||
|
||||
@@ -58,8 +58,8 @@ struct ti_syscon_reset_data {
|
||||
unsigned int nr_controls;
|
||||
};
|
||||
|
||||
#define to_ti_syscon_reset_data(rcdev) \
|
||||
container_of(rcdev, struct ti_syscon_reset_data, rcdev)
|
||||
#define to_ti_syscon_reset_data(_rcdev) \
|
||||
container_of(_rcdev, struct ti_syscon_reset_data, rcdev)
|
||||
|
||||
/**
|
||||
* ti_syscon_reset_assert() - assert device reset
|
||||
|
||||
@@ -718,8 +718,8 @@ static int max77686_init_rtc_regmap(struct max77686_rtc_info *info)
|
||||
|
||||
add_rtc_irq:
|
||||
ret = regmap_add_irq_chip(info->rtc_regmap, info->rtc_irq,
|
||||
IRQF_TRIGGER_FALLING | IRQF_ONESHOT |
|
||||
IRQF_SHARED, 0, info->drv_data->rtc_irq_chip,
|
||||
IRQF_ONESHOT | IRQF_SHARED,
|
||||
0, info->drv_data->rtc_irq_chip,
|
||||
&info->rtc_irq_data);
|
||||
if (ret < 0) {
|
||||
dev_err(info->dev, "Failed to add RTC irq chip: %d\n", ret);
|
||||
|
||||
@@ -500,7 +500,7 @@ ahc_inq(struct ahc_softc *ahc, u_int port)
|
||||
return ((ahc_inb(ahc, port))
|
||||
| (ahc_inb(ahc, port+1) << 8)
|
||||
| (ahc_inb(ahc, port+2) << 16)
|
||||
| (ahc_inb(ahc, port+3) << 24)
|
||||
| (((uint64_t)ahc_inb(ahc, port+3)) << 24)
|
||||
| (((uint64_t)ahc_inb(ahc, port+4)) << 32)
|
||||
| (((uint64_t)ahc_inb(ahc, port+5)) << 40)
|
||||
| (((uint64_t)ahc_inb(ahc, port+6)) << 48)
|
||||
|
||||
@@ -427,39 +427,10 @@ static umode_t iscsi_iface_attr_is_visible(struct kobject *kobj,
|
||||
struct device *dev = container_of(kobj, struct device, kobj);
|
||||
struct iscsi_iface *iface = iscsi_dev_to_iface(dev);
|
||||
struct iscsi_transport *t = iface->transport;
|
||||
int param;
|
||||
int param_type;
|
||||
int param = -1;
|
||||
|
||||
if (attr == &dev_attr_iface_enabled.attr)
|
||||
param = ISCSI_NET_PARAM_IFACE_ENABLE;
|
||||
else if (attr == &dev_attr_iface_vlan_id.attr)
|
||||
param = ISCSI_NET_PARAM_VLAN_ID;
|
||||
else if (attr == &dev_attr_iface_vlan_priority.attr)
|
||||
param = ISCSI_NET_PARAM_VLAN_PRIORITY;
|
||||
else if (attr == &dev_attr_iface_vlan_enabled.attr)
|
||||
param = ISCSI_NET_PARAM_VLAN_ENABLED;
|
||||
else if (attr == &dev_attr_iface_mtu.attr)
|
||||
param = ISCSI_NET_PARAM_MTU;
|
||||
else if (attr == &dev_attr_iface_port.attr)
|
||||
param = ISCSI_NET_PARAM_PORT;
|
||||
else if (attr == &dev_attr_iface_ipaddress_state.attr)
|
||||
param = ISCSI_NET_PARAM_IPADDR_STATE;
|
||||
else if (attr == &dev_attr_iface_delayed_ack_en.attr)
|
||||
param = ISCSI_NET_PARAM_DELAYED_ACK_EN;
|
||||
else if (attr == &dev_attr_iface_tcp_nagle_disable.attr)
|
||||
param = ISCSI_NET_PARAM_TCP_NAGLE_DISABLE;
|
||||
else if (attr == &dev_attr_iface_tcp_wsf_disable.attr)
|
||||
param = ISCSI_NET_PARAM_TCP_WSF_DISABLE;
|
||||
else if (attr == &dev_attr_iface_tcp_wsf.attr)
|
||||
param = ISCSI_NET_PARAM_TCP_WSF;
|
||||
else if (attr == &dev_attr_iface_tcp_timer_scale.attr)
|
||||
param = ISCSI_NET_PARAM_TCP_TIMER_SCALE;
|
||||
else if (attr == &dev_attr_iface_tcp_timestamp_en.attr)
|
||||
param = ISCSI_NET_PARAM_TCP_TIMESTAMP_EN;
|
||||
else if (attr == &dev_attr_iface_cache_id.attr)
|
||||
param = ISCSI_NET_PARAM_CACHE_ID;
|
||||
else if (attr == &dev_attr_iface_redirect_en.attr)
|
||||
param = ISCSI_NET_PARAM_REDIRECT_EN;
|
||||
else if (attr == &dev_attr_iface_def_taskmgmt_tmo.attr)
|
||||
param = ISCSI_IFACE_PARAM_DEF_TASKMGMT_TMO;
|
||||
else if (attr == &dev_attr_iface_header_digest.attr)
|
||||
@@ -496,6 +467,38 @@ static umode_t iscsi_iface_attr_is_visible(struct kobject *kobj,
|
||||
param = ISCSI_IFACE_PARAM_STRICT_LOGIN_COMP_EN;
|
||||
else if (attr == &dev_attr_iface_initiator_name.attr)
|
||||
param = ISCSI_IFACE_PARAM_INITIATOR_NAME;
|
||||
|
||||
if (param != -1)
|
||||
return t->attr_is_visible(ISCSI_IFACE_PARAM, param);
|
||||
|
||||
if (attr == &dev_attr_iface_vlan_id.attr)
|
||||
param = ISCSI_NET_PARAM_VLAN_ID;
|
||||
else if (attr == &dev_attr_iface_vlan_priority.attr)
|
||||
param = ISCSI_NET_PARAM_VLAN_PRIORITY;
|
||||
else if (attr == &dev_attr_iface_vlan_enabled.attr)
|
||||
param = ISCSI_NET_PARAM_VLAN_ENABLED;
|
||||
else if (attr == &dev_attr_iface_mtu.attr)
|
||||
param = ISCSI_NET_PARAM_MTU;
|
||||
else if (attr == &dev_attr_iface_port.attr)
|
||||
param = ISCSI_NET_PARAM_PORT;
|
||||
else if (attr == &dev_attr_iface_ipaddress_state.attr)
|
||||
param = ISCSI_NET_PARAM_IPADDR_STATE;
|
||||
else if (attr == &dev_attr_iface_delayed_ack_en.attr)
|
||||
param = ISCSI_NET_PARAM_DELAYED_ACK_EN;
|
||||
else if (attr == &dev_attr_iface_tcp_nagle_disable.attr)
|
||||
param = ISCSI_NET_PARAM_TCP_NAGLE_DISABLE;
|
||||
else if (attr == &dev_attr_iface_tcp_wsf_disable.attr)
|
||||
param = ISCSI_NET_PARAM_TCP_WSF_DISABLE;
|
||||
else if (attr == &dev_attr_iface_tcp_wsf.attr)
|
||||
param = ISCSI_NET_PARAM_TCP_WSF;
|
||||
else if (attr == &dev_attr_iface_tcp_timer_scale.attr)
|
||||
param = ISCSI_NET_PARAM_TCP_TIMER_SCALE;
|
||||
else if (attr == &dev_attr_iface_tcp_timestamp_en.attr)
|
||||
param = ISCSI_NET_PARAM_TCP_TIMESTAMP_EN;
|
||||
else if (attr == &dev_attr_iface_cache_id.attr)
|
||||
param = ISCSI_NET_PARAM_CACHE_ID;
|
||||
else if (attr == &dev_attr_iface_redirect_en.attr)
|
||||
param = ISCSI_NET_PARAM_REDIRECT_EN;
|
||||
else if (iface->iface_type == ISCSI_IFACE_TYPE_IPV4) {
|
||||
if (attr == &dev_attr_ipv4_iface_ipaddress.attr)
|
||||
param = ISCSI_NET_PARAM_IPV4_ADDR;
|
||||
@@ -586,32 +589,7 @@ static umode_t iscsi_iface_attr_is_visible(struct kobject *kobj,
|
||||
return 0;
|
||||
}
|
||||
|
||||
switch (param) {
|
||||
case ISCSI_IFACE_PARAM_DEF_TASKMGMT_TMO:
|
||||
case ISCSI_IFACE_PARAM_HDRDGST_EN:
|
||||
case ISCSI_IFACE_PARAM_DATADGST_EN:
|
||||
case ISCSI_IFACE_PARAM_IMM_DATA_EN:
|
||||
case ISCSI_IFACE_PARAM_INITIAL_R2T_EN:
|
||||
case ISCSI_IFACE_PARAM_DATASEQ_INORDER_EN:
|
||||
case ISCSI_IFACE_PARAM_PDU_INORDER_EN:
|
||||
case ISCSI_IFACE_PARAM_ERL:
|
||||
case ISCSI_IFACE_PARAM_MAX_RECV_DLENGTH:
|
||||
case ISCSI_IFACE_PARAM_FIRST_BURST:
|
||||
case ISCSI_IFACE_PARAM_MAX_R2T:
|
||||
case ISCSI_IFACE_PARAM_MAX_BURST:
|
||||
case ISCSI_IFACE_PARAM_CHAP_AUTH_EN:
|
||||
case ISCSI_IFACE_PARAM_BIDI_CHAP_EN:
|
||||
case ISCSI_IFACE_PARAM_DISCOVERY_AUTH_OPTIONAL:
|
||||
case ISCSI_IFACE_PARAM_DISCOVERY_LOGOUT_EN:
|
||||
case ISCSI_IFACE_PARAM_STRICT_LOGIN_COMP_EN:
|
||||
case ISCSI_IFACE_PARAM_INITIATOR_NAME:
|
||||
param_type = ISCSI_IFACE_PARAM;
|
||||
break;
|
||||
default:
|
||||
param_type = ISCSI_NET_PARAM;
|
||||
}
|
||||
|
||||
return t->attr_is_visible(param_type, param);
|
||||
return t->attr_is_visible(ISCSI_NET_PARAM, param);
|
||||
}
|
||||
|
||||
static struct attribute *iscsi_iface_attrs[] = {
|
||||
|
||||
@@ -338,13 +338,23 @@ static int mtk_spi_fifo_transfer(struct spi_master *master,
|
||||
mtk_spi_setup_packet(master);
|
||||
|
||||
cnt = xfer->len / 4;
|
||||
iowrite32_rep(mdata->base + SPI_TX_DATA_REG, xfer->tx_buf, cnt);
|
||||
if (xfer->tx_buf)
|
||||
iowrite32_rep(mdata->base + SPI_TX_DATA_REG, xfer->tx_buf, cnt);
|
||||
|
||||
if (xfer->rx_buf)
|
||||
ioread32_rep(mdata->base + SPI_RX_DATA_REG, xfer->rx_buf, cnt);
|
||||
|
||||
remainder = xfer->len % 4;
|
||||
if (remainder > 0) {
|
||||
reg_val = 0;
|
||||
memcpy(®_val, xfer->tx_buf + (cnt * 4), remainder);
|
||||
writel(reg_val, mdata->base + SPI_TX_DATA_REG);
|
||||
if (xfer->tx_buf) {
|
||||
memcpy(®_val, xfer->tx_buf + (cnt * 4), remainder);
|
||||
writel(reg_val, mdata->base + SPI_TX_DATA_REG);
|
||||
}
|
||||
if (xfer->rx_buf) {
|
||||
reg_val = readl(mdata->base + SPI_RX_DATA_REG);
|
||||
memcpy(xfer->rx_buf + (cnt * 4), ®_val, remainder);
|
||||
}
|
||||
}
|
||||
|
||||
mtk_spi_enable_transfer(master);
|
||||
|
||||
@@ -37,7 +37,7 @@
|
||||
#include "target_core_alua.h"
|
||||
|
||||
static sense_reason_t
|
||||
sbc_check_prot(struct se_device *, struct se_cmd *, unsigned char *, u32, bool);
|
||||
sbc_check_prot(struct se_device *, struct se_cmd *, unsigned char, u32, bool);
|
||||
static sense_reason_t sbc_execute_unmap(struct se_cmd *cmd);
|
||||
|
||||
static sense_reason_t
|
||||
@@ -319,14 +319,14 @@ static inline unsigned long long transport_lba_64_ext(unsigned char *cdb)
|
||||
}
|
||||
|
||||
static sense_reason_t
|
||||
sbc_setup_write_same(struct se_cmd *cmd, unsigned char *flags, struct sbc_ops *ops)
|
||||
sbc_setup_write_same(struct se_cmd *cmd, unsigned char flags, struct sbc_ops *ops)
|
||||
{
|
||||
struct se_device *dev = cmd->se_dev;
|
||||
sector_t end_lba = dev->transport->get_blocks(dev) + 1;
|
||||
unsigned int sectors = sbc_get_write_same_sectors(cmd);
|
||||
sense_reason_t ret;
|
||||
|
||||
if ((flags[0] & 0x04) || (flags[0] & 0x02)) {
|
||||
if ((flags & 0x04) || (flags & 0x02)) {
|
||||
pr_err("WRITE_SAME PBDATA and LBDATA"
|
||||
" bits not supported for Block Discard"
|
||||
" Emulation\n");
|
||||
@@ -348,7 +348,7 @@ sbc_setup_write_same(struct se_cmd *cmd, unsigned char *flags, struct sbc_ops *o
|
||||
}
|
||||
|
||||
/* We always have ANC_SUP == 0 so setting ANCHOR is always an error */
|
||||
if (flags[0] & 0x10) {
|
||||
if (flags & 0x10) {
|
||||
pr_warn("WRITE SAME with ANCHOR not supported\n");
|
||||
return TCM_INVALID_CDB_FIELD;
|
||||
}
|
||||
@@ -356,7 +356,7 @@ sbc_setup_write_same(struct se_cmd *cmd, unsigned char *flags, struct sbc_ops *o
|
||||
* Special case for WRITE_SAME w/ UNMAP=1 that ends up getting
|
||||
* translated into block discard requests within backend code.
|
||||
*/
|
||||
if (flags[0] & 0x08) {
|
||||
if (flags & 0x08) {
|
||||
if (!ops->execute_unmap)
|
||||
return TCM_UNSUPPORTED_SCSI_OPCODE;
|
||||
|
||||
@@ -371,7 +371,7 @@ sbc_setup_write_same(struct se_cmd *cmd, unsigned char *flags, struct sbc_ops *o
|
||||
if (!ops->execute_write_same)
|
||||
return TCM_UNSUPPORTED_SCSI_OPCODE;
|
||||
|
||||
ret = sbc_check_prot(dev, cmd, &cmd->t_task_cdb[0], sectors, true);
|
||||
ret = sbc_check_prot(dev, cmd, flags >> 5, sectors, true);
|
||||
if (ret)
|
||||
return ret;
|
||||
|
||||
@@ -729,10 +729,9 @@ sbc_set_prot_op_checks(u8 protect, bool fabric_prot, enum target_prot_type prot_
|
||||
}
|
||||
|
||||
static sense_reason_t
|
||||
sbc_check_prot(struct se_device *dev, struct se_cmd *cmd, unsigned char *cdb,
|
||||
sbc_check_prot(struct se_device *dev, struct se_cmd *cmd, unsigned char protect,
|
||||
u32 sectors, bool is_write)
|
||||
{
|
||||
u8 protect = cdb[1] >> 5;
|
||||
int sp_ops = cmd->se_sess->sup_prot_ops;
|
||||
int pi_prot_type = dev->dev_attrib.pi_prot_type;
|
||||
bool fabric_prot = false;
|
||||
@@ -780,7 +779,7 @@ sbc_check_prot(struct se_device *dev, struct se_cmd *cmd, unsigned char *cdb,
|
||||
/* Fallthrough */
|
||||
default:
|
||||
pr_err("Unable to determine pi_prot_type for CDB: 0x%02x "
|
||||
"PROTECT: 0x%02x\n", cdb[0], protect);
|
||||
"PROTECT: 0x%02x\n", cmd->t_task_cdb[0], protect);
|
||||
return TCM_INVALID_CDB_FIELD;
|
||||
}
|
||||
|
||||
@@ -855,7 +854,7 @@ sbc_parse_cdb(struct se_cmd *cmd, struct sbc_ops *ops)
|
||||
if (sbc_check_dpofua(dev, cmd, cdb))
|
||||
return TCM_INVALID_CDB_FIELD;
|
||||
|
||||
ret = sbc_check_prot(dev, cmd, cdb, sectors, false);
|
||||
ret = sbc_check_prot(dev, cmd, cdb[1] >> 5, sectors, false);
|
||||
if (ret)
|
||||
return ret;
|
||||
|
||||
@@ -869,7 +868,7 @@ sbc_parse_cdb(struct se_cmd *cmd, struct sbc_ops *ops)
|
||||
if (sbc_check_dpofua(dev, cmd, cdb))
|
||||
return TCM_INVALID_CDB_FIELD;
|
||||
|
||||
ret = sbc_check_prot(dev, cmd, cdb, sectors, false);
|
||||
ret = sbc_check_prot(dev, cmd, cdb[1] >> 5, sectors, false);
|
||||
if (ret)
|
||||
return ret;
|
||||
|
||||
@@ -883,7 +882,7 @@ sbc_parse_cdb(struct se_cmd *cmd, struct sbc_ops *ops)
|
||||
if (sbc_check_dpofua(dev, cmd, cdb))
|
||||
return TCM_INVALID_CDB_FIELD;
|
||||
|
||||
ret = sbc_check_prot(dev, cmd, cdb, sectors, false);
|
||||
ret = sbc_check_prot(dev, cmd, cdb[1] >> 5, sectors, false);
|
||||
if (ret)
|
||||
return ret;
|
||||
|
||||
@@ -904,7 +903,7 @@ sbc_parse_cdb(struct se_cmd *cmd, struct sbc_ops *ops)
|
||||
if (sbc_check_dpofua(dev, cmd, cdb))
|
||||
return TCM_INVALID_CDB_FIELD;
|
||||
|
||||
ret = sbc_check_prot(dev, cmd, cdb, sectors, true);
|
||||
ret = sbc_check_prot(dev, cmd, cdb[1] >> 5, sectors, true);
|
||||
if (ret)
|
||||
return ret;
|
||||
|
||||
@@ -918,7 +917,7 @@ sbc_parse_cdb(struct se_cmd *cmd, struct sbc_ops *ops)
|
||||
if (sbc_check_dpofua(dev, cmd, cdb))
|
||||
return TCM_INVALID_CDB_FIELD;
|
||||
|
||||
ret = sbc_check_prot(dev, cmd, cdb, sectors, true);
|
||||
ret = sbc_check_prot(dev, cmd, cdb[1] >> 5, sectors, true);
|
||||
if (ret)
|
||||
return ret;
|
||||
|
||||
@@ -932,7 +931,7 @@ sbc_parse_cdb(struct se_cmd *cmd, struct sbc_ops *ops)
|
||||
if (sbc_check_dpofua(dev, cmd, cdb))
|
||||
return TCM_INVALID_CDB_FIELD;
|
||||
|
||||
ret = sbc_check_prot(dev, cmd, cdb, sectors, true);
|
||||
ret = sbc_check_prot(dev, cmd, cdb[1] >> 5, sectors, true);
|
||||
if (ret)
|
||||
return ret;
|
||||
|
||||
@@ -991,7 +990,7 @@ sbc_parse_cdb(struct se_cmd *cmd, struct sbc_ops *ops)
|
||||
size = sbc_get_size(cmd, 1);
|
||||
cmd->t_task_lba = get_unaligned_be64(&cdb[12]);
|
||||
|
||||
ret = sbc_setup_write_same(cmd, &cdb[10], ops);
|
||||
ret = sbc_setup_write_same(cmd, cdb[10], ops);
|
||||
if (ret)
|
||||
return ret;
|
||||
break;
|
||||
@@ -1084,7 +1083,7 @@ sbc_parse_cdb(struct se_cmd *cmd, struct sbc_ops *ops)
|
||||
size = sbc_get_size(cmd, 1);
|
||||
cmd->t_task_lba = get_unaligned_be64(&cdb[2]);
|
||||
|
||||
ret = sbc_setup_write_same(cmd, &cdb[1], ops);
|
||||
ret = sbc_setup_write_same(cmd, cdb[1], ops);
|
||||
if (ret)
|
||||
return ret;
|
||||
break;
|
||||
@@ -1102,7 +1101,7 @@ sbc_parse_cdb(struct se_cmd *cmd, struct sbc_ops *ops)
|
||||
* Follow sbcr26 with WRITE_SAME (10) and check for the existence
|
||||
* of byte 1 bit 3 UNMAP instead of original reserved field
|
||||
*/
|
||||
ret = sbc_setup_write_same(cmd, &cdb[1], ops);
|
||||
ret = sbc_setup_write_same(cmd, cdb[1], ops);
|
||||
if (ret)
|
||||
return ret;
|
||||
break;
|
||||
|
||||
@@ -2112,7 +2112,7 @@ unregister:
|
||||
EXPORT_SYMBOL_GPL(thermal_zone_device_register);
|
||||
|
||||
/**
|
||||
* thermal_device_unregister - removes the registered thermal zone device
|
||||
* thermal_zone_device_unregister - removes the registered thermal zone device
|
||||
* @tz: the thermal zone device to remove
|
||||
*/
|
||||
void thermal_zone_device_unregister(struct thermal_zone_device *tz)
|
||||
|
||||
@@ -3896,6 +3896,47 @@ static int usb_set_lpm_timeout(struct usb_device *udev,
|
||||
return 0;
|
||||
}
|
||||
|
||||
/*
|
||||
* Don't allow device intiated U1/U2 if the system exit latency + one bus
|
||||
* interval is greater than the minimum service interval of any active
|
||||
* periodic endpoint. See USB 3.2 section 9.4.9
|
||||
*/
|
||||
static bool usb_device_may_initiate_lpm(struct usb_device *udev,
|
||||
enum usb3_link_state state)
|
||||
{
|
||||
unsigned int sel; /* us */
|
||||
int i, j;
|
||||
|
||||
if (state == USB3_LPM_U1)
|
||||
sel = DIV_ROUND_UP(udev->u1_params.sel, 1000);
|
||||
else if (state == USB3_LPM_U2)
|
||||
sel = DIV_ROUND_UP(udev->u2_params.sel, 1000);
|
||||
else
|
||||
return false;
|
||||
|
||||
for (i = 0; i < udev->actconfig->desc.bNumInterfaces; i++) {
|
||||
struct usb_interface *intf;
|
||||
struct usb_endpoint_descriptor *desc;
|
||||
unsigned int interval;
|
||||
|
||||
intf = udev->actconfig->interface[i];
|
||||
if (!intf)
|
||||
continue;
|
||||
|
||||
for (j = 0; j < intf->cur_altsetting->desc.bNumEndpoints; j++) {
|
||||
desc = &intf->cur_altsetting->endpoint[j].desc;
|
||||
|
||||
if (usb_endpoint_xfer_int(desc) ||
|
||||
usb_endpoint_xfer_isoc(desc)) {
|
||||
interval = (1 << (desc->bInterval - 1)) * 125;
|
||||
if (sel + 125 > interval)
|
||||
return false;
|
||||
}
|
||||
}
|
||||
}
|
||||
return true;
|
||||
}
|
||||
|
||||
/*
|
||||
* Enable the hub-initiated U1/U2 idle timeouts, and enable device-initiated
|
||||
* U1/U2 entry.
|
||||
@@ -3968,20 +4009,23 @@ static void usb_enable_link_state(struct usb_hcd *hcd, struct usb_device *udev,
|
||||
* U1/U2_ENABLE
|
||||
*/
|
||||
if (udev->actconfig &&
|
||||
usb_set_device_initiated_lpm(udev, state, true) == 0) {
|
||||
if (state == USB3_LPM_U1)
|
||||
udev->usb3_lpm_u1_enabled = 1;
|
||||
else if (state == USB3_LPM_U2)
|
||||
udev->usb3_lpm_u2_enabled = 1;
|
||||
} else {
|
||||
/* Don't request U1/U2 entry if the device
|
||||
* cannot transition to U1/U2.
|
||||
*/
|
||||
usb_set_lpm_timeout(udev, state, 0);
|
||||
hcd->driver->disable_usb3_lpm_timeout(hcd, udev, state);
|
||||
usb_device_may_initiate_lpm(udev, state)) {
|
||||
if (usb_set_device_initiated_lpm(udev, state, true)) {
|
||||
/*
|
||||
* Request to enable device initiated U1/U2 failed,
|
||||
* better to turn off lpm in this case.
|
||||
*/
|
||||
usb_set_lpm_timeout(udev, state, 0);
|
||||
hcd->driver->disable_usb3_lpm_timeout(hcd, udev, state);
|
||||
return;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
if (state == USB3_LPM_U1)
|
||||
udev->usb3_lpm_u1_enabled = 1;
|
||||
else if (state == USB3_LPM_U2)
|
||||
udev->usb3_lpm_u2_enabled = 1;
|
||||
}
|
||||
/*
|
||||
* Disable the hub-initiated U1/U2 idle timeouts, and disable device-initiated
|
||||
* U1/U2 entry.
|
||||
|
||||
@@ -325,10 +325,6 @@ static const struct usb_device_id usb_quirk_list[] = {
|
||||
/* DJI CineSSD */
|
||||
{ USB_DEVICE(0x2ca3, 0x0031), .driver_info = USB_QUIRK_NO_LPM },
|
||||
|
||||
/* Fibocom L850-GL LTE Modem */
|
||||
{ USB_DEVICE(0x2cb7, 0x0007), .driver_info =
|
||||
USB_QUIRK_IGNORE_REMOTE_WAKEUP },
|
||||
|
||||
/* INTEL VALUE SSD */
|
||||
{ USB_DEVICE(0x8086, 0xf1a5), .driver_info = USB_QUIRK_RESET_RESUME },
|
||||
|
||||
|
||||
@@ -149,8 +149,6 @@ struct max3421_hcd {
|
||||
*/
|
||||
struct urb *curr_urb;
|
||||
enum scheduling_pass sched_pass;
|
||||
struct usb_device *loaded_dev; /* dev that's loaded into the chip */
|
||||
int loaded_epnum; /* epnum whose toggles are loaded */
|
||||
int urb_done; /* > 0 -> no errors, < 0: errno */
|
||||
size_t curr_len;
|
||||
u8 hien;
|
||||
@@ -488,39 +486,17 @@ max3421_set_speed(struct usb_hcd *hcd, struct usb_device *dev)
|
||||
* Caller must NOT hold HCD spinlock.
|
||||
*/
|
||||
static void
|
||||
max3421_set_address(struct usb_hcd *hcd, struct usb_device *dev, int epnum,
|
||||
int force_toggles)
|
||||
max3421_set_address(struct usb_hcd *hcd, struct usb_device *dev, int epnum)
|
||||
{
|
||||
struct max3421_hcd *max3421_hcd = hcd_to_max3421(hcd);
|
||||
int old_epnum, same_ep, rcvtog, sndtog;
|
||||
struct usb_device *old_dev;
|
||||
int rcvtog, sndtog;
|
||||
u8 hctl;
|
||||
|
||||
old_dev = max3421_hcd->loaded_dev;
|
||||
old_epnum = max3421_hcd->loaded_epnum;
|
||||
|
||||
same_ep = (dev == old_dev && epnum == old_epnum);
|
||||
if (same_ep && !force_toggles)
|
||||
return;
|
||||
|
||||
if (old_dev && !same_ep) {
|
||||
/* save the old end-points toggles: */
|
||||
u8 hrsl = spi_rd8(hcd, MAX3421_REG_HRSL);
|
||||
|
||||
rcvtog = (hrsl >> MAX3421_HRSL_RCVTOGRD_BIT) & 1;
|
||||
sndtog = (hrsl >> MAX3421_HRSL_SNDTOGRD_BIT) & 1;
|
||||
|
||||
/* no locking: HCD (i.e., we) own toggles, don't we? */
|
||||
usb_settoggle(old_dev, old_epnum, 0, rcvtog);
|
||||
usb_settoggle(old_dev, old_epnum, 1, sndtog);
|
||||
}
|
||||
/* setup new endpoint's toggle bits: */
|
||||
rcvtog = usb_gettoggle(dev, epnum, 0);
|
||||
sndtog = usb_gettoggle(dev, epnum, 1);
|
||||
hctl = (BIT(rcvtog + MAX3421_HCTL_RCVTOG0_BIT) |
|
||||
BIT(sndtog + MAX3421_HCTL_SNDTOG0_BIT));
|
||||
|
||||
max3421_hcd->loaded_epnum = epnum;
|
||||
spi_wr8(hcd, MAX3421_REG_HCTL, hctl);
|
||||
|
||||
/*
|
||||
@@ -528,7 +504,6 @@ max3421_set_address(struct usb_hcd *hcd, struct usb_device *dev, int epnum,
|
||||
* address-assignment so it's best to just always load the
|
||||
* address whenever the end-point changed/was forced.
|
||||
*/
|
||||
max3421_hcd->loaded_dev = dev;
|
||||
spi_wr8(hcd, MAX3421_REG_PERADDR, dev->devnum);
|
||||
}
|
||||
|
||||
@@ -663,7 +638,7 @@ max3421_select_and_start_urb(struct usb_hcd *hcd)
|
||||
struct max3421_hcd *max3421_hcd = hcd_to_max3421(hcd);
|
||||
struct urb *urb, *curr_urb = NULL;
|
||||
struct max3421_ep *max3421_ep;
|
||||
int epnum, force_toggles = 0;
|
||||
int epnum;
|
||||
struct usb_host_endpoint *ep;
|
||||
struct list_head *pos;
|
||||
unsigned long flags;
|
||||
@@ -773,7 +748,6 @@ done:
|
||||
usb_settoggle(urb->dev, epnum, 0, 1);
|
||||
usb_settoggle(urb->dev, epnum, 1, 1);
|
||||
max3421_ep->pkt_state = PKT_STATE_SETUP;
|
||||
force_toggles = 1;
|
||||
} else
|
||||
max3421_ep->pkt_state = PKT_STATE_TRANSFER;
|
||||
}
|
||||
@@ -781,7 +755,7 @@ done:
|
||||
spin_unlock_irqrestore(&max3421_hcd->lock, flags);
|
||||
|
||||
max3421_ep->last_active = max3421_hcd->frame_number;
|
||||
max3421_set_address(hcd, urb->dev, epnum, force_toggles);
|
||||
max3421_set_address(hcd, urb->dev, epnum);
|
||||
max3421_set_speed(hcd, urb->dev);
|
||||
max3421_next_transfer(hcd, 0);
|
||||
return 1;
|
||||
@@ -1376,6 +1350,16 @@ max3421_urb_done(struct usb_hcd *hcd)
|
||||
status = 0;
|
||||
urb = max3421_hcd->curr_urb;
|
||||
if (urb) {
|
||||
/* save the old end-points toggles: */
|
||||
u8 hrsl = spi_rd8(hcd, MAX3421_REG_HRSL);
|
||||
int rcvtog = (hrsl >> MAX3421_HRSL_RCVTOGRD_BIT) & 1;
|
||||
int sndtog = (hrsl >> MAX3421_HRSL_SNDTOGRD_BIT) & 1;
|
||||
int epnum = usb_endpoint_num(&urb->ep->desc);
|
||||
|
||||
/* no locking: HCD (i.e., we) own toggles, don't we? */
|
||||
usb_settoggle(urb->dev, epnum, 0, rcvtog);
|
||||
usb_settoggle(urb->dev, epnum, 1, sndtog);
|
||||
|
||||
max3421_hcd->curr_urb = NULL;
|
||||
spin_lock_irqsave(&max3421_hcd->lock, flags);
|
||||
usb_hcd_unlink_urb_from_ep(hcd, urb);
|
||||
|
||||
@@ -1577,11 +1577,12 @@ int xhci_hub_status_data(struct usb_hcd *hcd, char *buf)
|
||||
* Inform the usbcore about resume-in-progress by returning
|
||||
* a non-zero value even if there are no status changes.
|
||||
*/
|
||||
spin_lock_irqsave(&xhci->lock, flags);
|
||||
|
||||
status = bus_state->resuming_ports;
|
||||
|
||||
mask = PORT_CSC | PORT_PEC | PORT_OCC | PORT_PLC | PORT_WRC | PORT_CEC;
|
||||
|
||||
spin_lock_irqsave(&xhci->lock, flags);
|
||||
/* For each port, did anything change? If so, set that bit in buf. */
|
||||
for (i = 0; i < max_ports; i++) {
|
||||
temp = readl(port_array[i]);
|
||||
|
||||
@@ -115,6 +115,8 @@ static struct dma_chan *usbhsf_dma_chan_get(struct usbhs_fifo *fifo,
|
||||
#define usbhsf_dma_map(p) __usbhsf_dma_map_ctrl(p, 1)
|
||||
#define usbhsf_dma_unmap(p) __usbhsf_dma_map_ctrl(p, 0)
|
||||
static int __usbhsf_dma_map_ctrl(struct usbhs_pkt *pkt, int map);
|
||||
static void usbhsf_tx_irq_ctrl(struct usbhs_pipe *pipe, int enable);
|
||||
static void usbhsf_rx_irq_ctrl(struct usbhs_pipe *pipe, int enable);
|
||||
struct usbhs_pkt *usbhs_pkt_pop(struct usbhs_pipe *pipe, struct usbhs_pkt *pkt)
|
||||
{
|
||||
struct usbhs_priv *priv = usbhs_pipe_to_priv(pipe);
|
||||
@@ -138,6 +140,11 @@ struct usbhs_pkt *usbhs_pkt_pop(struct usbhs_pipe *pipe, struct usbhs_pkt *pkt)
|
||||
dmaengine_terminate_all(chan);
|
||||
usbhsf_fifo_clear(pipe, fifo);
|
||||
usbhsf_dma_unmap(pkt);
|
||||
} else {
|
||||
if (usbhs_pipe_is_dir_in(pipe))
|
||||
usbhsf_rx_irq_ctrl(pipe, 0);
|
||||
else
|
||||
usbhsf_tx_irq_ctrl(pipe, 0);
|
||||
}
|
||||
|
||||
usbhs_pipe_running(pipe, 0);
|
||||
|
||||
@@ -153,6 +153,7 @@ static const struct usb_device_id id_table[] = {
|
||||
{ USB_DEVICE(0x10C4, 0x89A4) }, /* CESINEL FTBC Flexible Thyristor Bridge Controller */
|
||||
{ USB_DEVICE(0x10C4, 0x89FB) }, /* Qivicon ZigBee USB Radio Stick */
|
||||
{ USB_DEVICE(0x10C4, 0x8A2A) }, /* HubZ dual ZigBee and Z-Wave dongle */
|
||||
{ USB_DEVICE(0x10C4, 0x8A5B) }, /* CEL EM3588 ZigBee USB Stick */
|
||||
{ USB_DEVICE(0x10C4, 0x8A5E) }, /* CEL EM3588 ZigBee USB Stick Long Range */
|
||||
{ USB_DEVICE(0x10C4, 0x8B34) }, /* Qivicon ZigBee USB Radio Stick */
|
||||
{ USB_DEVICE(0x10C4, 0xEA60) }, /* Silicon Labs factory default */
|
||||
@@ -200,8 +201,8 @@ static const struct usb_device_id id_table[] = {
|
||||
{ USB_DEVICE(0x1901, 0x0194) }, /* GE Healthcare Remote Alarm Box */
|
||||
{ USB_DEVICE(0x1901, 0x0195) }, /* GE B850/B650/B450 CP2104 DP UART interface */
|
||||
{ USB_DEVICE(0x1901, 0x0196) }, /* GE B850 CP2105 DP UART interface */
|
||||
{ USB_DEVICE(0x1901, 0x0197) }, /* GE CS1000 Display serial interface */
|
||||
{ USB_DEVICE(0x1901, 0x0198) }, /* GE CS1000 M.2 Key E serial interface */
|
||||
{ USB_DEVICE(0x1901, 0x0197) }, /* GE CS1000 M.2 Key E serial interface */
|
||||
{ USB_DEVICE(0x1901, 0x0198) }, /* GE CS1000 Display serial interface */
|
||||
{ USB_DEVICE(0x199B, 0xBA30) }, /* LORD WSDA-200-USB */
|
||||
{ USB_DEVICE(0x19CF, 0x3000) }, /* Parrot NMEA GPS Flight Recorder */
|
||||
{ USB_DEVICE(0x1ADB, 0x0001) }, /* Schweitzer Engineering C662 Cable */
|
||||
|
||||
@@ -241,6 +241,7 @@ static void option_instat_callback(struct urb *urb);
|
||||
#define QUECTEL_PRODUCT_UC15 0x9090
|
||||
/* These u-blox products use Qualcomm's vendor ID */
|
||||
#define UBLOX_PRODUCT_R410M 0x90b2
|
||||
#define UBLOX_PRODUCT_R6XX 0x90fa
|
||||
/* These Yuga products use Qualcomm's vendor ID */
|
||||
#define YUGA_PRODUCT_CLM920_NC5 0x9625
|
||||
|
||||
@@ -1098,6 +1099,8 @@ static const struct usb_device_id option_ids[] = {
|
||||
/* u-blox products using Qualcomm vendor ID */
|
||||
{ USB_DEVICE(QUALCOMM_VENDOR_ID, UBLOX_PRODUCT_R410M),
|
||||
.driver_info = RSVD(1) | RSVD(3) },
|
||||
{ USB_DEVICE(QUALCOMM_VENDOR_ID, UBLOX_PRODUCT_R6XX),
|
||||
.driver_info = RSVD(3) },
|
||||
/* Quectel products using Quectel vendor ID */
|
||||
{ USB_DEVICE(QUECTEL_VENDOR_ID, QUECTEL_PRODUCT_EC21),
|
||||
.driver_info = RSVD(4) },
|
||||
|
||||
@@ -55,6 +55,13 @@ UNUSUAL_DEV(0x059f, 0x105f, 0x0000, 0x9999,
|
||||
USB_SC_DEVICE, USB_PR_DEVICE, NULL,
|
||||
US_FL_NO_REPORT_OPCODES),
|
||||
|
||||
/* Reported-by: Julian Sikorski <belegdol@gmail.com> */
|
||||
UNUSUAL_DEV(0x059f, 0x1061, 0x0000, 0x9999,
|
||||
"LaCie",
|
||||
"Rugged USB3-FW",
|
||||
USB_SC_DEVICE, USB_PR_DEVICE, NULL,
|
||||
US_FL_IGNORE_UAS),
|
||||
|
||||
/*
|
||||
* Apricorn USB3 dongle sometimes returns "USBSUSBSUSBS" in response to SCSI
|
||||
* commands in UAS mode. Observed with the 1.28 firmware; are there others?
|
||||
|
||||
@@ -484,7 +484,7 @@ again:
|
||||
* inode has not been flagged as nocompress. This flag can
|
||||
* change at any time if we discover bad compression ratios.
|
||||
*/
|
||||
if (inode_need_compress(inode)) {
|
||||
if (nr_pages > 1 && inode_need_compress(inode)) {
|
||||
WARN_ON(pages);
|
||||
pages = kcalloc(nr_pages, sizeof(struct page *), GFP_NOFS);
|
||||
if (!pages) {
|
||||
|
||||
@@ -868,7 +868,7 @@ static ssize_t mem_rw(struct file *file, char __user *buf,
|
||||
flags |= FOLL_WRITE;
|
||||
|
||||
while (count > 0) {
|
||||
int this_len = min_t(int, count, PAGE_SIZE);
|
||||
size_t this_len = min_t(size_t, count, PAGE_SIZE);
|
||||
|
||||
if (write && copy_from_user(page, buf, this_len)) {
|
||||
copied = -EFAULT;
|
||||
|
||||
@@ -31,7 +31,9 @@ static inline struct ip_tunnel_info *skb_tunnel_info(struct sk_buff *skb)
|
||||
return &md_dst->u.tun_info;
|
||||
|
||||
dst = skb_dst(skb);
|
||||
if (dst && dst->lwtstate)
|
||||
if (dst && dst->lwtstate &&
|
||||
(dst->lwtstate->type == LWTUNNEL_ENCAP_IP ||
|
||||
dst->lwtstate->type == LWTUNNEL_ENCAP_IP6))
|
||||
return lwt_tun_info(dst->lwtstate);
|
||||
|
||||
return NULL;
|
||||
|
||||
@@ -203,7 +203,7 @@ static inline bool ipv6_anycast_destination(const struct dst_entry *dst,
|
||||
int ip6_fragment(struct net *net, struct sock *sk, struct sk_buff *skb,
|
||||
int (*output)(struct net *, struct sock *, struct sk_buff *));
|
||||
|
||||
static inline int ip6_skb_dst_mtu(struct sk_buff *skb)
|
||||
static inline unsigned int ip6_skb_dst_mtu(struct sk_buff *skb)
|
||||
{
|
||||
struct ipv6_pinfo *np = skb->sk && !dev_recursion_level() ?
|
||||
inet6_sk(skb->sk) : NULL;
|
||||
|
||||
@@ -4454,7 +4454,7 @@ static const u64 cfs_bandwidth_slack_period = 5 * NSEC_PER_MSEC;
|
||||
static int runtime_refresh_within(struct cfs_bandwidth *cfs_b, u64 min_expire)
|
||||
{
|
||||
struct hrtimer *refresh_timer = &cfs_b->period_timer;
|
||||
u64 remaining;
|
||||
s64 remaining;
|
||||
|
||||
/* if the call-back is running a quota refresh is already occurring */
|
||||
if (hrtimer_callback_running(refresh_timer))
|
||||
@@ -4462,7 +4462,7 @@ static int runtime_refresh_within(struct cfs_bandwidth *cfs_b, u64 min_expire)
|
||||
|
||||
/* is a quota refresh about to occur? */
|
||||
remaining = ktime_to_ns(hrtimer_expires_remaining(refresh_timer));
|
||||
if (remaining < min_expire)
|
||||
if (remaining < (s64)min_expire)
|
||||
return 1;
|
||||
|
||||
return 0;
|
||||
|
||||
@@ -3081,10 +3081,30 @@ static bool rb_per_cpu_empty(struct ring_buffer_per_cpu *cpu_buffer)
|
||||
if (unlikely(!head))
|
||||
return true;
|
||||
|
||||
return reader->read == rb_page_commit(reader) &&
|
||||
(commit == reader ||
|
||||
(commit == head &&
|
||||
head->read == rb_page_commit(commit)));
|
||||
/* Reader should exhaust content in reader page */
|
||||
if (reader->read != rb_page_commit(reader))
|
||||
return false;
|
||||
|
||||
/*
|
||||
* If writers are committing on the reader page, knowing all
|
||||
* committed content has been read, the ring buffer is empty.
|
||||
*/
|
||||
if (commit == reader)
|
||||
return true;
|
||||
|
||||
/*
|
||||
* If writers are committing on a page other than reader page
|
||||
* and head page, there should always be content to read.
|
||||
*/
|
||||
if (commit != head)
|
||||
return false;
|
||||
|
||||
/*
|
||||
* Writers are committing on the head page, we just need
|
||||
* to care about there're committed data, and the reader will
|
||||
* swap reader page with head page when it is to read data.
|
||||
*/
|
||||
return rb_page_commit(commit) == 0;
|
||||
}
|
||||
|
||||
/**
|
||||
|
||||
@@ -486,7 +486,7 @@ int br_add_if(struct net_bridge *br, struct net_device *dev)
|
||||
struct net_bridge_port *p;
|
||||
int err = 0;
|
||||
unsigned br_hr, dev_hr;
|
||||
bool changed_addr;
|
||||
bool changed_addr, fdb_synced = false;
|
||||
|
||||
/* Don't allow bridging non-ethernet like devices, or DSA-enabled
|
||||
* master network devices since the bridge layer rx_handler prevents
|
||||
@@ -556,6 +556,19 @@ int br_add_if(struct net_bridge *br, struct net_device *dev)
|
||||
list_add_rcu(&p->list, &br->port_list);
|
||||
|
||||
nbp_update_port_count(br);
|
||||
if (!br_promisc_port(p) && (p->dev->priv_flags & IFF_UNICAST_FLT)) {
|
||||
/* When updating the port count we also update all ports'
|
||||
* promiscuous mode.
|
||||
* A port leaving promiscuous mode normally gets the bridge's
|
||||
* fdb synced to the unicast filter (if supported), however,
|
||||
* `br_port_clear_promisc` does not distinguish between
|
||||
* non-promiscuous ports and *new* ports, so we need to
|
||||
* sync explicitly here.
|
||||
*/
|
||||
fdb_synced = br_fdb_sync_static(br, p) == 0;
|
||||
if (!fdb_synced)
|
||||
netdev_err(dev, "failed to sync bridge static fdb addresses to this port\n");
|
||||
}
|
||||
|
||||
netdev_update_features(br->dev);
|
||||
|
||||
@@ -596,6 +609,8 @@ int br_add_if(struct net_bridge *br, struct net_device *dev)
|
||||
return 0;
|
||||
|
||||
err7:
|
||||
if (fdb_synced)
|
||||
br_fdb_unsync_static(br, p);
|
||||
list_del_rcu(&p->list);
|
||||
br_fdb_delete_by_port(br, p, 0, 1);
|
||||
nbp_update_port_count(br);
|
||||
|
||||
@@ -539,7 +539,8 @@ static int caif_seqpkt_sendmsg(struct socket *sock, struct msghdr *msg,
|
||||
goto err;
|
||||
|
||||
ret = -EINVAL;
|
||||
if (unlikely(msg->msg_iter.iov->iov_base == NULL))
|
||||
if (unlikely(msg->msg_iter.nr_segs == 0) ||
|
||||
unlikely(msg->msg_iter.iov->iov_base == NULL))
|
||||
goto err;
|
||||
noblock = msg->msg_flags & MSG_DONTWAIT;
|
||||
|
||||
|
||||
@@ -823,7 +823,7 @@ static int dn_auto_bind(struct socket *sock)
|
||||
static int dn_confirm_accept(struct sock *sk, long *timeo, gfp_t allocation)
|
||||
{
|
||||
struct dn_scp *scp = DN_SK(sk);
|
||||
DEFINE_WAIT(wait);
|
||||
DEFINE_WAIT_FUNC(wait, woken_wake_function);
|
||||
int err;
|
||||
|
||||
if (scp->state != DN_CR)
|
||||
@@ -833,11 +833,11 @@ static int dn_confirm_accept(struct sock *sk, long *timeo, gfp_t allocation)
|
||||
scp->segsize_loc = dst_metric_advmss(__sk_dst_get(sk));
|
||||
dn_send_conn_conf(sk, allocation);
|
||||
|
||||
prepare_to_wait(sk_sleep(sk), &wait, TASK_INTERRUPTIBLE);
|
||||
add_wait_queue(sk_sleep(sk), &wait);
|
||||
for(;;) {
|
||||
release_sock(sk);
|
||||
if (scp->state == DN_CC)
|
||||
*timeo = schedule_timeout(*timeo);
|
||||
*timeo = wait_woken(&wait, TASK_INTERRUPTIBLE, *timeo);
|
||||
lock_sock(sk);
|
||||
err = 0;
|
||||
if (scp->state == DN_RUN)
|
||||
@@ -851,9 +851,8 @@ static int dn_confirm_accept(struct sock *sk, long *timeo, gfp_t allocation)
|
||||
err = -EAGAIN;
|
||||
if (!*timeo)
|
||||
break;
|
||||
prepare_to_wait(sk_sleep(sk), &wait, TASK_INTERRUPTIBLE);
|
||||
}
|
||||
finish_wait(sk_sleep(sk), &wait);
|
||||
remove_wait_queue(sk_sleep(sk), &wait);
|
||||
if (err == 0) {
|
||||
sk->sk_socket->state = SS_CONNECTED;
|
||||
} else if (scp->state != DN_CC) {
|
||||
@@ -865,7 +864,7 @@ static int dn_confirm_accept(struct sock *sk, long *timeo, gfp_t allocation)
|
||||
static int dn_wait_run(struct sock *sk, long *timeo)
|
||||
{
|
||||
struct dn_scp *scp = DN_SK(sk);
|
||||
DEFINE_WAIT(wait);
|
||||
DEFINE_WAIT_FUNC(wait, woken_wake_function);
|
||||
int err = 0;
|
||||
|
||||
if (scp->state == DN_RUN)
|
||||
@@ -874,11 +873,11 @@ static int dn_wait_run(struct sock *sk, long *timeo)
|
||||
if (!*timeo)
|
||||
return -EALREADY;
|
||||
|
||||
prepare_to_wait(sk_sleep(sk), &wait, TASK_INTERRUPTIBLE);
|
||||
add_wait_queue(sk_sleep(sk), &wait);
|
||||
for(;;) {
|
||||
release_sock(sk);
|
||||
if (scp->state == DN_CI || scp->state == DN_CC)
|
||||
*timeo = schedule_timeout(*timeo);
|
||||
*timeo = wait_woken(&wait, TASK_INTERRUPTIBLE, *timeo);
|
||||
lock_sock(sk);
|
||||
err = 0;
|
||||
if (scp->state == DN_RUN)
|
||||
@@ -892,9 +891,8 @@ static int dn_wait_run(struct sock *sk, long *timeo)
|
||||
err = -ETIMEDOUT;
|
||||
if (!*timeo)
|
||||
break;
|
||||
prepare_to_wait(sk_sleep(sk), &wait, TASK_INTERRUPTIBLE);
|
||||
}
|
||||
finish_wait(sk_sleep(sk), &wait);
|
||||
remove_wait_queue(sk_sleep(sk), &wait);
|
||||
out:
|
||||
if (err == 0) {
|
||||
sk->sk_socket->state = SS_CONNECTED;
|
||||
@@ -1039,16 +1037,16 @@ static void dn_user_copy(struct sk_buff *skb, struct optdata_dn *opt)
|
||||
|
||||
static struct sk_buff *dn_wait_for_connect(struct sock *sk, long *timeo)
|
||||
{
|
||||
DEFINE_WAIT(wait);
|
||||
DEFINE_WAIT_FUNC(wait, woken_wake_function);
|
||||
struct sk_buff *skb = NULL;
|
||||
int err = 0;
|
||||
|
||||
prepare_to_wait(sk_sleep(sk), &wait, TASK_INTERRUPTIBLE);
|
||||
add_wait_queue(sk_sleep(sk), &wait);
|
||||
for(;;) {
|
||||
release_sock(sk);
|
||||
skb = skb_dequeue(&sk->sk_receive_queue);
|
||||
if (skb == NULL) {
|
||||
*timeo = schedule_timeout(*timeo);
|
||||
*timeo = wait_woken(&wait, TASK_INTERRUPTIBLE, *timeo);
|
||||
skb = skb_dequeue(&sk->sk_receive_queue);
|
||||
}
|
||||
lock_sock(sk);
|
||||
@@ -1063,9 +1061,8 @@ static struct sk_buff *dn_wait_for_connect(struct sock *sk, long *timeo)
|
||||
err = -EAGAIN;
|
||||
if (!*timeo)
|
||||
break;
|
||||
prepare_to_wait(sk_sleep(sk), &wait, TASK_INTERRUPTIBLE);
|
||||
}
|
||||
finish_wait(sk_sleep(sk), &wait);
|
||||
remove_wait_queue(sk_sleep(sk), &wait);
|
||||
|
||||
return skb == NULL ? ERR_PTR(err) : skb;
|
||||
}
|
||||
|
||||
@@ -280,7 +280,7 @@ void tcp_v4_mtu_reduced(struct sock *sk)
|
||||
|
||||
if ((1 << sk->sk_state) & (TCPF_LISTEN | TCPF_CLOSE))
|
||||
return;
|
||||
mtu = tcp_sk(sk)->mtu_info;
|
||||
mtu = READ_ONCE(tcp_sk(sk)->mtu_info);
|
||||
dst = inet_csk_update_pmtu(sk, mtu);
|
||||
if (!dst)
|
||||
return;
|
||||
@@ -447,7 +447,7 @@ void tcp_v4_err(struct sk_buff *icmp_skb, u32 info)
|
||||
if (sk->sk_state == TCP_LISTEN)
|
||||
goto out;
|
||||
|
||||
tp->mtu_info = info;
|
||||
WRITE_ONCE(tp->mtu_info, info);
|
||||
if (!sock_owned_by_user(sk)) {
|
||||
tcp_v4_mtu_reduced(sk);
|
||||
} else {
|
||||
|
||||
@@ -1383,6 +1383,7 @@ int tcp_mtu_to_mss(struct sock *sk, int pmtu)
|
||||
return __tcp_mtu_to_mss(sk, pmtu) -
|
||||
(tcp_sk(sk)->tcp_header_len - sizeof(struct tcphdr));
|
||||
}
|
||||
EXPORT_SYMBOL(tcp_mtu_to_mss);
|
||||
|
||||
/* Inverse of above */
|
||||
int tcp_mss_to_mtu(struct sock *sk, int mss)
|
||||
|
||||
@@ -312,11 +312,20 @@ failure:
|
||||
static void tcp_v6_mtu_reduced(struct sock *sk)
|
||||
{
|
||||
struct dst_entry *dst;
|
||||
u32 mtu;
|
||||
|
||||
if ((1 << sk->sk_state) & (TCPF_LISTEN | TCPF_CLOSE))
|
||||
return;
|
||||
|
||||
dst = inet6_csk_update_pmtu(sk, tcp_sk(sk)->mtu_info);
|
||||
mtu = READ_ONCE(tcp_sk(sk)->mtu_info);
|
||||
|
||||
/* Drop requests trying to increase our current mss.
|
||||
* Check done in __ip6_rt_update_pmtu() is too late.
|
||||
*/
|
||||
if (tcp_mtu_to_mss(sk, mtu) >= tcp_sk(sk)->mss_cache)
|
||||
return;
|
||||
|
||||
dst = inet6_csk_update_pmtu(sk, mtu);
|
||||
if (!dst)
|
||||
return;
|
||||
|
||||
@@ -395,6 +404,8 @@ static void tcp_v6_err(struct sk_buff *skb, struct inet6_skb_parm *opt,
|
||||
}
|
||||
|
||||
if (type == ICMPV6_PKT_TOOBIG) {
|
||||
u32 mtu = ntohl(info);
|
||||
|
||||
/* We are not interested in TCP_LISTEN and open_requests
|
||||
* (SYN-ACKs send out by Linux are always <576bytes so
|
||||
* they should go through unfragmented).
|
||||
@@ -405,7 +416,11 @@ static void tcp_v6_err(struct sk_buff *skb, struct inet6_skb_parm *opt,
|
||||
if (!ip6_sk_accept_pmtu(sk))
|
||||
goto out;
|
||||
|
||||
tp->mtu_info = ntohl(info);
|
||||
if (mtu < IPV6_MIN_MTU)
|
||||
goto out;
|
||||
|
||||
WRITE_ONCE(tp->mtu_info, mtu);
|
||||
|
||||
if (!sock_owned_by_user(sk))
|
||||
tcp_v6_mtu_reduced(sk);
|
||||
else if (!test_and_set_bit(TCP_MTU_REDUCED_DEFERRED,
|
||||
|
||||
@@ -141,7 +141,7 @@ static int __xfrm6_output(struct net *net, struct sock *sk, struct sk_buff *skb)
|
||||
{
|
||||
struct dst_entry *dst = skb_dst(skb);
|
||||
struct xfrm_state *x = dst->xfrm;
|
||||
int mtu;
|
||||
unsigned int mtu;
|
||||
bool toobig;
|
||||
|
||||
#ifdef CONFIG_NETFILTER
|
||||
|
||||
@@ -125,11 +125,9 @@ static void nr_heartbeat_expiry(unsigned long param)
|
||||
is accepted() it isn't 'dead' so doesn't get removed. */
|
||||
if (sock_flag(sk, SOCK_DESTROY) ||
|
||||
(sk->sk_state == TCP_LISTEN && sock_flag(sk, SOCK_DEAD))) {
|
||||
sock_hold(sk);
|
||||
bh_unlock_sock(sk);
|
||||
nr_destroy_socket(sk);
|
||||
sock_put(sk);
|
||||
return;
|
||||
goto out;
|
||||
}
|
||||
break;
|
||||
|
||||
@@ -150,6 +148,8 @@ static void nr_heartbeat_expiry(unsigned long param)
|
||||
|
||||
nr_start_heartbeat(sk);
|
||||
bh_unlock_sock(sk);
|
||||
out:
|
||||
sock_put(sk);
|
||||
}
|
||||
|
||||
static void nr_t2timer_expiry(unsigned long param)
|
||||
@@ -163,6 +163,7 @@ static void nr_t2timer_expiry(unsigned long param)
|
||||
nr_enquiry_response(sk);
|
||||
}
|
||||
bh_unlock_sock(sk);
|
||||
sock_put(sk);
|
||||
}
|
||||
|
||||
static void nr_t4timer_expiry(unsigned long param)
|
||||
@@ -172,6 +173,7 @@ static void nr_t4timer_expiry(unsigned long param)
|
||||
bh_lock_sock(sk);
|
||||
nr_sk(sk)->condition &= ~NR_COND_PEER_RX_BUSY;
|
||||
bh_unlock_sock(sk);
|
||||
sock_put(sk);
|
||||
}
|
||||
|
||||
static void nr_idletimer_expiry(unsigned long param)
|
||||
@@ -200,6 +202,7 @@ static void nr_idletimer_expiry(unsigned long param)
|
||||
sock_set_flag(sk, SOCK_DEAD);
|
||||
}
|
||||
bh_unlock_sock(sk);
|
||||
sock_put(sk);
|
||||
}
|
||||
|
||||
static void nr_t1timer_expiry(unsigned long param)
|
||||
@@ -212,8 +215,7 @@ static void nr_t1timer_expiry(unsigned long param)
|
||||
case NR_STATE_1:
|
||||
if (nr->n2count == nr->n2) {
|
||||
nr_disconnect(sk, ETIMEDOUT);
|
||||
bh_unlock_sock(sk);
|
||||
return;
|
||||
goto out;
|
||||
} else {
|
||||
nr->n2count++;
|
||||
nr_write_internal(sk, NR_CONNREQ);
|
||||
@@ -223,8 +225,7 @@ static void nr_t1timer_expiry(unsigned long param)
|
||||
case NR_STATE_2:
|
||||
if (nr->n2count == nr->n2) {
|
||||
nr_disconnect(sk, ETIMEDOUT);
|
||||
bh_unlock_sock(sk);
|
||||
return;
|
||||
goto out;
|
||||
} else {
|
||||
nr->n2count++;
|
||||
nr_write_internal(sk, NR_DISCREQ);
|
||||
@@ -234,8 +235,7 @@ static void nr_t1timer_expiry(unsigned long param)
|
||||
case NR_STATE_3:
|
||||
if (nr->n2count == nr->n2) {
|
||||
nr_disconnect(sk, ETIMEDOUT);
|
||||
bh_unlock_sock(sk);
|
||||
return;
|
||||
goto out;
|
||||
} else {
|
||||
nr->n2count++;
|
||||
nr_requeue_frames(sk);
|
||||
@@ -244,5 +244,7 @@ static void nr_t1timer_expiry(unsigned long param)
|
||||
}
|
||||
|
||||
nr_start_t1timer(sk);
|
||||
out:
|
||||
bh_unlock_sock(sk);
|
||||
sock_put(sk);
|
||||
}
|
||||
|
||||
@@ -82,15 +82,23 @@ UTS_TRUNCATE="cut -b -$UTS_LEN"
|
||||
# Only replace the real compile.h if the new one is different,
|
||||
# in order to preserve the timestamp and avoid unnecessary
|
||||
# recompilations.
|
||||
# We don't consider the file changed if only the date/time changed.
|
||||
# We don't consider the file changed if only the date/time changed,
|
||||
# unless KBUILD_BUILD_TIMESTAMP was explicitly set (e.g. for
|
||||
# reproducible builds with that value referring to a commit timestamp).
|
||||
# A kernel config change will increase the generation number, thus
|
||||
# causing compile.h to be updated (including date/time) due to the
|
||||
# changed comment in the
|
||||
# first line.
|
||||
|
||||
if [ -z "$KBUILD_BUILD_TIMESTAMP" ]; then
|
||||
IGNORE_PATTERN="UTS_VERSION"
|
||||
else
|
||||
IGNORE_PATTERN="NOT_A_PATTERN_TO_BE_MATCHED"
|
||||
fi
|
||||
|
||||
if [ -r $TARGET ] && \
|
||||
grep -v 'UTS_VERSION' $TARGET > .tmpver.1 && \
|
||||
grep -v 'UTS_VERSION' .tmpcompile > .tmpver.2 && \
|
||||
grep -v $IGNORE_PATTERN $TARGET > .tmpver.1 && \
|
||||
grep -v $IGNORE_PATTERN .tmpcompile > .tmpver.2 && \
|
||||
cmp -s .tmpver.1 .tmpver.2; then
|
||||
rm -f .tmpcompile
|
||||
else
|
||||
|
||||
@@ -828,6 +828,7 @@ static int snd_sb_csp_start(struct snd_sb_csp * p, int sample_width, int channel
|
||||
mixR = snd_sbmixer_read(p->chip, SB_DSP4_PCM_DEV + 1);
|
||||
snd_sbmixer_write(p->chip, SB_DSP4_PCM_DEV, mixL & 0x7);
|
||||
snd_sbmixer_write(p->chip, SB_DSP4_PCM_DEV + 1, mixR & 0x7);
|
||||
spin_unlock_irqrestore(&p->chip->mixer_lock, flags);
|
||||
|
||||
spin_lock(&p->chip->reg_lock);
|
||||
set_mode_register(p->chip, 0xc0); /* c0 = STOP */
|
||||
@@ -867,6 +868,7 @@ static int snd_sb_csp_start(struct snd_sb_csp * p, int sample_width, int channel
|
||||
spin_unlock(&p->chip->reg_lock);
|
||||
|
||||
/* restore PCM volume */
|
||||
spin_lock_irqsave(&p->chip->mixer_lock, flags);
|
||||
snd_sbmixer_write(p->chip, SB_DSP4_PCM_DEV, mixL);
|
||||
snd_sbmixer_write(p->chip, SB_DSP4_PCM_DEV + 1, mixR);
|
||||
spin_unlock_irqrestore(&p->chip->mixer_lock, flags);
|
||||
@@ -892,6 +894,7 @@ static int snd_sb_csp_stop(struct snd_sb_csp * p)
|
||||
mixR = snd_sbmixer_read(p->chip, SB_DSP4_PCM_DEV + 1);
|
||||
snd_sbmixer_write(p->chip, SB_DSP4_PCM_DEV, mixL & 0x7);
|
||||
snd_sbmixer_write(p->chip, SB_DSP4_PCM_DEV + 1, mixR & 0x7);
|
||||
spin_unlock_irqrestore(&p->chip->mixer_lock, flags);
|
||||
|
||||
spin_lock(&p->chip->reg_lock);
|
||||
if (p->running & SNDRV_SB_CSP_ST_QSOUND) {
|
||||
@@ -906,6 +909,7 @@ static int snd_sb_csp_stop(struct snd_sb_csp * p)
|
||||
spin_unlock(&p->chip->reg_lock);
|
||||
|
||||
/* restore PCM volume */
|
||||
spin_lock_irqsave(&p->chip->mixer_lock, flags);
|
||||
snd_sbmixer_write(p->chip, SB_DSP4_PCM_DEV, mixL);
|
||||
snd_sbmixer_write(p->chip, SB_DSP4_PCM_DEV + 1, mixR);
|
||||
spin_unlock_irqrestore(&p->chip->mixer_lock, flags);
|
||||
|
||||
@@ -1,4 +1,5 @@
|
||||
#include <stdio.h>
|
||||
#include <stdlib.h>
|
||||
#include <sys/epoll.h>
|
||||
#include <util/util.h>
|
||||
#include <util/bpf-loader.h>
|
||||
@@ -231,6 +232,7 @@ static int __test__bpf(int idx)
|
||||
bpf_testcase_table[idx].target_func,
|
||||
bpf_testcase_table[idx].expect_result);
|
||||
out:
|
||||
free(obj_buf);
|
||||
bpf__clear();
|
||||
return ret;
|
||||
}
|
||||
|
||||
@@ -61,7 +61,7 @@ int lzma_decompress_to_file(const char *input, int output_fd)
|
||||
|
||||
if (ferror(infile)) {
|
||||
pr_err("lzma: read error: %s\n", strerror(errno));
|
||||
goto err_fclose;
|
||||
goto err_lzma_end;
|
||||
}
|
||||
|
||||
if (feof(infile))
|
||||
@@ -75,7 +75,7 @@ int lzma_decompress_to_file(const char *input, int output_fd)
|
||||
|
||||
if (writen(output_fd, buf_out, write_size) != write_size) {
|
||||
pr_err("lzma: write error: %s\n", strerror(errno));
|
||||
goto err_fclose;
|
||||
goto err_lzma_end;
|
||||
}
|
||||
|
||||
strm.next_out = buf_out;
|
||||
@@ -87,11 +87,13 @@ int lzma_decompress_to_file(const char *input, int output_fd)
|
||||
break;
|
||||
|
||||
pr_err("lzma: failed %s\n", lzma_strerror(ret));
|
||||
goto err_fclose;
|
||||
goto err_lzma_end;
|
||||
}
|
||||
}
|
||||
|
||||
err = 0;
|
||||
err_lzma_end:
|
||||
lzma_end(&strm);
|
||||
err_fclose:
|
||||
fclose(infile);
|
||||
return err;
|
||||
|
||||
@@ -334,11 +334,11 @@ int probe_file__del_events(int fd, struct strfilter *filter)
|
||||
|
||||
ret = probe_file__get_events(fd, filter, namelist);
|
||||
if (ret < 0)
|
||||
return ret;
|
||||
goto out;
|
||||
|
||||
ret = probe_file__del_strlist(fd, namelist);
|
||||
out:
|
||||
strlist__delete(namelist);
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
|
||||
Reference in New Issue
Block a user