mirror of
https://github.com/hardkernel/linux.git
synced 2026-03-24 19:40:21 +09:00
Merge tag 'v6.6.101' of git://git.kernel.org/pub/scm/linux/kernel/git/stable/linux-stable into odroid-6.6.y
This is the 6.6.101 stable release Change-Id: I929f4b8ac20b7a7a37bff214d93e4a23f5c5c898
This commit is contained in:
2
Makefile
2
Makefile
@@ -1,7 +1,7 @@
|
||||
# SPDX-License-Identifier: GPL-2.0
|
||||
VERSION = 6
|
||||
PATCHLEVEL = 6
|
||||
SUBLEVEL = 100
|
||||
SUBLEVEL = 101
|
||||
EXTRAVERSION =
|
||||
NAME = Pinguïn Aangedreven
|
||||
|
||||
|
||||
@@ -142,7 +142,7 @@ endif
|
||||
# Need -Uarm for gcc < 3.x
|
||||
KBUILD_CPPFLAGS +=$(cpp-y)
|
||||
KBUILD_CFLAGS +=$(CFLAGS_ABI) $(CFLAGS_ISA) $(arch-y) $(tune-y) $(call cc-option,-mshort-load-bytes,$(call cc-option,-malignment-traps,)) -msoft-float -Uarm
|
||||
KBUILD_AFLAGS +=$(CFLAGS_ABI) $(AFLAGS_ISA) -Wa,$(arch-y) $(tune-y) -include asm/unified.h -msoft-float
|
||||
KBUILD_AFLAGS +=$(CFLAGS_ABI) $(AFLAGS_ISA) -Wa,$(arch-y) $(tune-y) -include $(srctree)/arch/arm/include/asm/unified.h -msoft-float
|
||||
|
||||
CHECKFLAGS += -D__arm__
|
||||
|
||||
|
||||
@@ -45,6 +45,11 @@
|
||||
/*
|
||||
* Save/restore interrupts.
|
||||
*/
|
||||
.macro save_and_disable_daif, flags
|
||||
mrs \flags, daif
|
||||
msr daifset, #0xf
|
||||
.endm
|
||||
|
||||
.macro save_and_disable_irq, flags
|
||||
mrs \flags, daif
|
||||
msr daifset, #3
|
||||
|
||||
@@ -364,6 +364,7 @@ static const struct arm64_ftr_bits ftr_id_aa64mmfr0[] = {
|
||||
};
|
||||
|
||||
static const struct arm64_ftr_bits ftr_id_aa64mmfr1[] = {
|
||||
ARM64_FTR_BITS(FTR_HIDDEN, FTR_STRICT, FTR_LOWER_SAFE, ID_AA64MMFR1_EL1_ECBHB_SHIFT, 4, 0),
|
||||
ARM64_FTR_BITS(FTR_HIDDEN, FTR_NONSTRICT, FTR_LOWER_SAFE, ID_AA64MMFR1_EL1_TIDCP1_SHIFT, 4, 0),
|
||||
ARM64_FTR_BITS(FTR_VISIBLE, FTR_STRICT, FTR_LOWER_SAFE, ID_AA64MMFR1_EL1_AFP_SHIFT, 4, 0),
|
||||
ARM64_FTR_BITS(FTR_HIDDEN, FTR_STRICT, FTR_LOWER_SAFE, ID_AA64MMFR1_EL1_HCX_SHIFT, 4, 0),
|
||||
|
||||
@@ -824,6 +824,7 @@ SYM_CODE_END(__bp_harden_el1_vectors)
|
||||
*
|
||||
*/
|
||||
SYM_FUNC_START(cpu_switch_to)
|
||||
save_and_disable_daif x11
|
||||
mov x10, #THREAD_CPU_CONTEXT
|
||||
add x8, x0, x10
|
||||
mov x9, sp
|
||||
@@ -847,6 +848,7 @@ SYM_FUNC_START(cpu_switch_to)
|
||||
ptrauth_keys_install_kernel x1, x8, x9, x10
|
||||
scs_save x0
|
||||
scs_load_current
|
||||
restore_irq x11
|
||||
ret
|
||||
SYM_FUNC_END(cpu_switch_to)
|
||||
NOKPROBE(cpu_switch_to)
|
||||
@@ -873,6 +875,7 @@ NOKPROBE(ret_from_fork)
|
||||
* Calls func(regs) using this CPU's irq stack and shadow irq stack.
|
||||
*/
|
||||
SYM_FUNC_START(call_on_irq_stack)
|
||||
save_and_disable_daif x9
|
||||
#ifdef CONFIG_SHADOW_CALL_STACK
|
||||
get_current_task x16
|
||||
scs_save x16
|
||||
@@ -887,8 +890,10 @@ SYM_FUNC_START(call_on_irq_stack)
|
||||
|
||||
/* Move to the new stack and call the function there */
|
||||
add sp, x16, #IRQ_STACK_SIZE
|
||||
restore_irq x9
|
||||
blr x1
|
||||
|
||||
save_and_disable_daif x9
|
||||
/*
|
||||
* Restore the SP from the FP, and restore the FP and LR from the frame
|
||||
* record.
|
||||
@@ -896,6 +901,7 @@ SYM_FUNC_START(call_on_irq_stack)
|
||||
mov sp, x29
|
||||
ldp x29, x30, [sp], #16
|
||||
scs_load_current
|
||||
restore_irq x9
|
||||
ret
|
||||
SYM_FUNC_END(call_on_irq_stack)
|
||||
NOKPROBE(call_on_irq_stack)
|
||||
|
||||
@@ -129,6 +129,7 @@ config CRYPTO_CHACHA20_P10
|
||||
config CRYPTO_POLY1305_P10
|
||||
tristate "Hash functions: Poly1305 (P10 or later)"
|
||||
depends on PPC64 && CPU_LITTLE_ENDIAN && VSX
|
||||
depends on BROKEN # Needs to be fixed to work in softirq context
|
||||
select CRYPTO_HASH
|
||||
select CRYPTO_LIB_POLY1305_GENERIC
|
||||
help
|
||||
|
||||
@@ -2734,7 +2734,7 @@ static void intel_pmu_read_event(struct perf_event *event)
|
||||
if (pmu_enabled)
|
||||
intel_pmu_disable_all();
|
||||
|
||||
if (is_topdown_event(event))
|
||||
if (is_topdown_count(event))
|
||||
static_call(intel_pmu_update_topdown_event)(event);
|
||||
else
|
||||
intel_pmu_drain_pebs_buffer();
|
||||
|
||||
@@ -192,7 +192,6 @@ static void hv_irq_compose_msi_msg(struct irq_data *data, struct msi_msg *msg)
|
||||
struct pci_dev *dev;
|
||||
struct hv_interrupt_entry out_entry, *stored_entry;
|
||||
struct irq_cfg *cfg = irqd_cfg(data);
|
||||
const cpumask_t *affinity;
|
||||
int cpu;
|
||||
u64 status;
|
||||
|
||||
@@ -204,8 +203,7 @@ static void hv_irq_compose_msi_msg(struct irq_data *data, struct msi_msg *msg)
|
||||
return;
|
||||
}
|
||||
|
||||
affinity = irq_data_get_effective_affinity_mask(data);
|
||||
cpu = cpumask_first_and(affinity, cpu_online_mask);
|
||||
cpu = cpumask_first(irq_data_get_effective_affinity_mask(data));
|
||||
|
||||
if (data->chip_data) {
|
||||
/*
|
||||
|
||||
@@ -549,6 +549,8 @@ static bool amd_check_tsa_microcode(void)
|
||||
p.model = c->x86_model;
|
||||
p.ext_model = c->x86_model >> 4;
|
||||
p.stepping = c->x86_stepping;
|
||||
/* reserved bits are expected to be 0 in test below */
|
||||
p.__reserved = 0;
|
||||
|
||||
if (cpu_has(c, X86_FEATURE_ZEN3) ||
|
||||
cpu_has(c, X86_FEATURE_ZEN4)) {
|
||||
|
||||
@@ -1173,6 +1173,8 @@ err_name:
|
||||
err_map:
|
||||
kfree(map);
|
||||
err:
|
||||
if (bus && bus->free_on_exit)
|
||||
kfree(bus);
|
||||
return ERR_PTR(ret);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(__regmap_init);
|
||||
|
||||
@@ -942,6 +942,7 @@ struct fsl_mc_device *fsl_mc_get_endpoint(struct fsl_mc_device *mc_dev,
|
||||
struct fsl_mc_obj_desc endpoint_desc = {{ 0 }};
|
||||
struct dprc_endpoint endpoint1 = {{ 0 }};
|
||||
struct dprc_endpoint endpoint2 = {{ 0 }};
|
||||
struct fsl_mc_bus *mc_bus;
|
||||
int state, err;
|
||||
|
||||
mc_bus_dev = to_fsl_mc_device(mc_dev->dev.parent);
|
||||
@@ -965,6 +966,8 @@ struct fsl_mc_device *fsl_mc_get_endpoint(struct fsl_mc_device *mc_dev,
|
||||
strcpy(endpoint_desc.type, endpoint2.type);
|
||||
endpoint_desc.id = endpoint2.id;
|
||||
endpoint = fsl_mc_device_lookup(&endpoint_desc, mc_bus_dev);
|
||||
if (endpoint)
|
||||
return endpoint;
|
||||
|
||||
/*
|
||||
* We know that the device has an endpoint because we verified by
|
||||
@@ -972,17 +975,13 @@ struct fsl_mc_device *fsl_mc_get_endpoint(struct fsl_mc_device *mc_dev,
|
||||
* yet discovered by the fsl-mc bus, thus the lookup returned NULL.
|
||||
* Force a rescan of the devices in this container and retry the lookup.
|
||||
*/
|
||||
if (!endpoint) {
|
||||
struct fsl_mc_bus *mc_bus = to_fsl_mc_bus(mc_bus_dev);
|
||||
|
||||
if (mutex_trylock(&mc_bus->scan_mutex)) {
|
||||
err = dprc_scan_objects(mc_bus_dev, true);
|
||||
mutex_unlock(&mc_bus->scan_mutex);
|
||||
}
|
||||
|
||||
if (err < 0)
|
||||
return ERR_PTR(err);
|
||||
mc_bus = to_fsl_mc_bus(mc_bus_dev);
|
||||
if (mutex_trylock(&mc_bus->scan_mutex)) {
|
||||
err = dprc_scan_objects(mc_bus_dev, true);
|
||||
mutex_unlock(&mc_bus->scan_mutex);
|
||||
}
|
||||
if (err < 0)
|
||||
return ERR_PTR(err);
|
||||
|
||||
endpoint = fsl_mc_device_lookup(&endpoint_desc, mc_bus_dev);
|
||||
/*
|
||||
|
||||
@@ -790,7 +790,7 @@ static void waveform_detach(struct comedi_device *dev)
|
||||
{
|
||||
struct waveform_private *devpriv = dev->private;
|
||||
|
||||
if (devpriv) {
|
||||
if (devpriv && dev->n_subdevices) {
|
||||
del_timer_sync(&devpriv->ai_timer);
|
||||
del_timer_sync(&devpriv->ao_timer);
|
||||
}
|
||||
|
||||
@@ -27,12 +27,14 @@ MODULE_DEVICE_TABLE(pci, adf_pci_tbl);
|
||||
|
||||
static int adf_probe(struct pci_dev *dev, const struct pci_device_id *ent);
|
||||
static void adf_remove(struct pci_dev *dev);
|
||||
static void adf_shutdown(struct pci_dev *dev);
|
||||
|
||||
static struct pci_driver adf_driver = {
|
||||
.id_table = adf_pci_tbl,
|
||||
.name = ADF_DH895XCC_DEVICE_NAME,
|
||||
.probe = adf_probe,
|
||||
.remove = adf_remove,
|
||||
.shutdown = adf_shutdown,
|
||||
.sriov_configure = adf_sriov_configure,
|
||||
.err_handler = &adf_err_handler,
|
||||
};
|
||||
@@ -227,6 +229,13 @@ static void adf_remove(struct pci_dev *pdev)
|
||||
kfree(accel_dev);
|
||||
}
|
||||
|
||||
static void adf_shutdown(struct pci_dev *pdev)
|
||||
{
|
||||
struct adf_accel_dev *accel_dev = adf_devmgr_pci_to_accel_dev(pdev);
|
||||
|
||||
adf_dev_down(accel_dev, false);
|
||||
}
|
||||
|
||||
static int __init adfdrv_init(void)
|
||||
{
|
||||
request_module("intel_qat");
|
||||
|
||||
@@ -1130,13 +1130,12 @@ svm_range_split_head(struct svm_range *prange,
|
||||
}
|
||||
|
||||
static void
|
||||
svm_range_add_child(struct svm_range *prange, struct mm_struct *mm,
|
||||
struct svm_range *pchild, enum svm_work_list_ops op)
|
||||
svm_range_add_child(struct svm_range *prange, struct svm_range *pchild, enum svm_work_list_ops op)
|
||||
{
|
||||
pr_debug("add child 0x%p [0x%lx 0x%lx] to prange 0x%p child list %d\n",
|
||||
pchild, pchild->start, pchild->last, prange, op);
|
||||
|
||||
pchild->work_item.mm = mm;
|
||||
pchild->work_item.mm = NULL;
|
||||
pchild->work_item.op = op;
|
||||
list_add_tail(&pchild->child_list, &prange->child_list);
|
||||
}
|
||||
@@ -1182,14 +1181,14 @@ svm_range_split_by_granularity(struct kfd_process *p, struct mm_struct *mm,
|
||||
r = svm_range_split(prange, start, prange->last, &head);
|
||||
if (r)
|
||||
return r;
|
||||
svm_range_add_child(parent, mm, head, SVM_OP_ADD_RANGE);
|
||||
svm_range_add_child(parent, head, SVM_OP_ADD_RANGE);
|
||||
}
|
||||
|
||||
if (last < prange->last) {
|
||||
r = svm_range_split(prange, prange->start, last, &tail);
|
||||
if (r)
|
||||
return r;
|
||||
svm_range_add_child(parent, mm, tail, SVM_OP_ADD_RANGE);
|
||||
svm_range_add_child(parent, tail, SVM_OP_ADD_RANGE);
|
||||
}
|
||||
|
||||
/* xnack on, update mapping on GPUs with ACCESS_IN_PLACE */
|
||||
@@ -2393,15 +2392,17 @@ svm_range_add_list_work(struct svm_range_list *svms, struct svm_range *prange,
|
||||
prange->work_item.op != SVM_OP_UNMAP_RANGE)
|
||||
prange->work_item.op = op;
|
||||
} else {
|
||||
prange->work_item.op = op;
|
||||
|
||||
/* Pairs with mmput in deferred_list_work */
|
||||
mmget(mm);
|
||||
prange->work_item.mm = mm;
|
||||
list_add_tail(&prange->deferred_list,
|
||||
&prange->svms->deferred_range_list);
|
||||
pr_debug("add prange 0x%p [0x%lx 0x%lx] to work list op %d\n",
|
||||
prange, prange->start, prange->last, op);
|
||||
/* Pairs with mmput in deferred_list_work.
|
||||
* If process is exiting and mm is gone, don't update mmu notifier.
|
||||
*/
|
||||
if (mmget_not_zero(mm)) {
|
||||
prange->work_item.mm = mm;
|
||||
prange->work_item.op = op;
|
||||
list_add_tail(&prange->deferred_list,
|
||||
&prange->svms->deferred_range_list);
|
||||
pr_debug("add prange 0x%p [0x%lx 0x%lx] to work list op %d\n",
|
||||
prange, prange->start, prange->last, op);
|
||||
}
|
||||
}
|
||||
spin_unlock(&svms->deferred_list_lock);
|
||||
}
|
||||
@@ -2415,8 +2416,7 @@ void schedule_deferred_list_work(struct svm_range_list *svms)
|
||||
}
|
||||
|
||||
static void
|
||||
svm_range_unmap_split(struct mm_struct *mm, struct svm_range *parent,
|
||||
struct svm_range *prange, unsigned long start,
|
||||
svm_range_unmap_split(struct svm_range *parent, struct svm_range *prange, unsigned long start,
|
||||
unsigned long last)
|
||||
{
|
||||
struct svm_range *head;
|
||||
@@ -2437,12 +2437,12 @@ svm_range_unmap_split(struct mm_struct *mm, struct svm_range *parent,
|
||||
svm_range_split(tail, last + 1, tail->last, &head);
|
||||
|
||||
if (head != prange && tail != prange) {
|
||||
svm_range_add_child(parent, mm, head, SVM_OP_UNMAP_RANGE);
|
||||
svm_range_add_child(parent, mm, tail, SVM_OP_ADD_RANGE);
|
||||
svm_range_add_child(parent, head, SVM_OP_UNMAP_RANGE);
|
||||
svm_range_add_child(parent, tail, SVM_OP_ADD_RANGE);
|
||||
} else if (tail != prange) {
|
||||
svm_range_add_child(parent, mm, tail, SVM_OP_UNMAP_RANGE);
|
||||
svm_range_add_child(parent, tail, SVM_OP_UNMAP_RANGE);
|
||||
} else if (head != prange) {
|
||||
svm_range_add_child(parent, mm, head, SVM_OP_UNMAP_RANGE);
|
||||
svm_range_add_child(parent, head, SVM_OP_UNMAP_RANGE);
|
||||
} else if (parent != prange) {
|
||||
prange->work_item.op = SVM_OP_UNMAP_RANGE;
|
||||
}
|
||||
@@ -2481,14 +2481,14 @@ svm_range_unmap_from_cpu(struct mm_struct *mm, struct svm_range *prange,
|
||||
l = min(last, pchild->last);
|
||||
if (l >= s)
|
||||
svm_range_unmap_from_gpus(pchild, s, l, trigger);
|
||||
svm_range_unmap_split(mm, prange, pchild, start, last);
|
||||
svm_range_unmap_split(prange, pchild, start, last);
|
||||
mutex_unlock(&pchild->lock);
|
||||
}
|
||||
s = max(start, prange->start);
|
||||
l = min(last, prange->last);
|
||||
if (l >= s)
|
||||
svm_range_unmap_from_gpus(prange, s, l, trigger);
|
||||
svm_range_unmap_split(mm, prange, prange, start, last);
|
||||
svm_range_unmap_split(prange, prange, start, last);
|
||||
|
||||
if (unmap_parent)
|
||||
svm_range_add_list_work(svms, prange, mm, SVM_OP_UNMAP_RANGE);
|
||||
@@ -2531,8 +2531,6 @@ svm_range_cpu_invalidate_pagetables(struct mmu_interval_notifier *mni,
|
||||
|
||||
if (range->event == MMU_NOTIFY_RELEASE)
|
||||
return true;
|
||||
if (!mmget_not_zero(mni->mm))
|
||||
return true;
|
||||
|
||||
start = mni->interval_tree.start;
|
||||
last = mni->interval_tree.last;
|
||||
@@ -2559,7 +2557,6 @@ svm_range_cpu_invalidate_pagetables(struct mmu_interval_notifier *mni,
|
||||
}
|
||||
|
||||
svm_range_unlock(prange);
|
||||
mmput(mni->mm);
|
||||
|
||||
return true;
|
||||
}
|
||||
|
||||
@@ -1352,7 +1352,7 @@ static int ti_sn_bridge_probe(struct auxiliary_device *adev,
|
||||
regmap_update_bits(pdata->regmap, SN_HPD_DISABLE_REG,
|
||||
HPD_DISABLE, 0);
|
||||
mutex_unlock(&pdata->comms_mutex);
|
||||
};
|
||||
}
|
||||
|
||||
drm_bridge_add(&pdata->bridge);
|
||||
|
||||
|
||||
@@ -1293,6 +1293,12 @@ int intel_dp_rate_select(struct intel_dp *intel_dp, int rate)
|
||||
void intel_dp_compute_rate(struct intel_dp *intel_dp, int port_clock,
|
||||
u8 *link_bw, u8 *rate_select)
|
||||
{
|
||||
struct drm_i915_private *i915 = dp_to_i915(intel_dp);
|
||||
|
||||
/* FIXME g4x can't generate an exact 2.7GHz with the 96MHz non-SSC refclk */
|
||||
if (IS_G4X(i915) && port_clock == 268800)
|
||||
port_clock = 270000;
|
||||
|
||||
/* eDP 1.4 rate select method. */
|
||||
if (intel_dp->use_rate_select) {
|
||||
*link_bw = 0;
|
||||
|
||||
@@ -346,20 +346,9 @@ void drm_sched_entity_destroy(struct drm_sched_entity *entity)
|
||||
}
|
||||
EXPORT_SYMBOL(drm_sched_entity_destroy);
|
||||
|
||||
/* drm_sched_entity_clear_dep - callback to clear the entities dependency */
|
||||
static void drm_sched_entity_clear_dep(struct dma_fence *f,
|
||||
struct dma_fence_cb *cb)
|
||||
{
|
||||
struct drm_sched_entity *entity =
|
||||
container_of(cb, struct drm_sched_entity, cb);
|
||||
|
||||
entity->dependency = NULL;
|
||||
dma_fence_put(f);
|
||||
}
|
||||
|
||||
/*
|
||||
* drm_sched_entity_clear_dep - callback to clear the entities dependency and
|
||||
* wake up scheduler
|
||||
* drm_sched_entity_wakeup - callback to clear the entity's dependency and
|
||||
* wake up the scheduler
|
||||
*/
|
||||
static void drm_sched_entity_wakeup(struct dma_fence *f,
|
||||
struct dma_fence_cb *cb)
|
||||
@@ -367,7 +356,8 @@ static void drm_sched_entity_wakeup(struct dma_fence *f,
|
||||
struct drm_sched_entity *entity =
|
||||
container_of(cb, struct drm_sched_entity, cb);
|
||||
|
||||
drm_sched_entity_clear_dep(f, cb);
|
||||
entity->dependency = NULL;
|
||||
dma_fence_put(f);
|
||||
drm_sched_wakeup_if_can_queue(entity->rq->sched);
|
||||
}
|
||||
|
||||
@@ -420,13 +410,6 @@ static bool drm_sched_entity_add_dependency_cb(struct drm_sched_entity *entity)
|
||||
fence = dma_fence_get(&s_fence->scheduled);
|
||||
dma_fence_put(entity->dependency);
|
||||
entity->dependency = fence;
|
||||
if (!dma_fence_add_callback(fence, &entity->cb,
|
||||
drm_sched_entity_clear_dep))
|
||||
return true;
|
||||
|
||||
/* Ignore it when it is already scheduled */
|
||||
dma_fence_put(fence);
|
||||
return false;
|
||||
}
|
||||
|
||||
if (!dma_fence_add_callback(entity->dependency, &entity->cb,
|
||||
|
||||
@@ -452,8 +452,10 @@ static int qup_i2c_bus_active(struct qup_i2c_dev *qup, int len)
|
||||
if (!(status & I2C_STATUS_BUS_ACTIVE))
|
||||
break;
|
||||
|
||||
if (time_after(jiffies, timeout))
|
||||
if (time_after(jiffies, timeout)) {
|
||||
ret = -ETIMEDOUT;
|
||||
break;
|
||||
}
|
||||
|
||||
usleep_range(len, len * 2);
|
||||
}
|
||||
|
||||
@@ -607,7 +607,6 @@ static int tegra_i2c_wait_for_config_load(struct tegra_i2c_dev *i2c_dev)
|
||||
static int tegra_i2c_init(struct tegra_i2c_dev *i2c_dev)
|
||||
{
|
||||
u32 val, clk_divisor, clk_multiplier, tsu_thd, tlow, thigh, non_hs_mode;
|
||||
acpi_handle handle = ACPI_HANDLE(i2c_dev->dev);
|
||||
struct i2c_timings *t = &i2c_dev->timings;
|
||||
int err;
|
||||
|
||||
@@ -619,11 +618,7 @@ static int tegra_i2c_init(struct tegra_i2c_dev *i2c_dev)
|
||||
* emit a noisy warning on error, which won't stay unnoticed and
|
||||
* won't hose machine entirely.
|
||||
*/
|
||||
if (handle)
|
||||
err = acpi_evaluate_object(handle, "_RST", NULL, NULL);
|
||||
else
|
||||
err = reset_control_reset(i2c_dev->rst);
|
||||
|
||||
err = device_reset(i2c_dev->dev);
|
||||
WARN_ON_ONCE(err);
|
||||
|
||||
if (IS_DVC(i2c_dev))
|
||||
@@ -1668,19 +1663,6 @@ static void tegra_i2c_parse_dt(struct tegra_i2c_dev *i2c_dev)
|
||||
i2c_dev->is_vi = true;
|
||||
}
|
||||
|
||||
static int tegra_i2c_init_reset(struct tegra_i2c_dev *i2c_dev)
|
||||
{
|
||||
if (ACPI_HANDLE(i2c_dev->dev))
|
||||
return 0;
|
||||
|
||||
i2c_dev->rst = devm_reset_control_get_exclusive(i2c_dev->dev, "i2c");
|
||||
if (IS_ERR(i2c_dev->rst))
|
||||
return dev_err_probe(i2c_dev->dev, PTR_ERR(i2c_dev->rst),
|
||||
"failed to get reset control\n");
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int tegra_i2c_init_clocks(struct tegra_i2c_dev *i2c_dev)
|
||||
{
|
||||
int err;
|
||||
@@ -1790,10 +1772,6 @@ static int tegra_i2c_probe(struct platform_device *pdev)
|
||||
|
||||
tegra_i2c_parse_dt(i2c_dev);
|
||||
|
||||
err = tegra_i2c_init_reset(i2c_dev);
|
||||
if (err)
|
||||
return err;
|
||||
|
||||
err = tegra_i2c_init_clocks(i2c_dev);
|
||||
if (err)
|
||||
return err;
|
||||
|
||||
@@ -116,15 +116,16 @@ static int virtio_i2c_complete_reqs(struct virtqueue *vq,
|
||||
for (i = 0; i < num; i++) {
|
||||
struct virtio_i2c_req *req = &reqs[i];
|
||||
|
||||
wait_for_completion(&req->completion);
|
||||
|
||||
if (!failed && req->in_hdr.status != VIRTIO_I2C_MSG_OK)
|
||||
failed = true;
|
||||
if (!failed) {
|
||||
if (wait_for_completion_interruptible(&req->completion))
|
||||
failed = true;
|
||||
else if (req->in_hdr.status != VIRTIO_I2C_MSG_OK)
|
||||
failed = true;
|
||||
else
|
||||
j++;
|
||||
}
|
||||
|
||||
i2c_put_dma_safe_msg_buf(reqs[i].buf, &msgs[i], !failed);
|
||||
|
||||
if (!failed)
|
||||
j++;
|
||||
}
|
||||
|
||||
return j;
|
||||
|
||||
@@ -308,7 +308,6 @@ static void ad7949_disable_reg(void *reg)
|
||||
|
||||
static int ad7949_spi_probe(struct spi_device *spi)
|
||||
{
|
||||
u32 spi_ctrl_mask = spi->controller->bits_per_word_mask;
|
||||
struct device *dev = &spi->dev;
|
||||
const struct ad7949_adc_spec *spec;
|
||||
struct ad7949_adc_chip *ad7949_adc;
|
||||
@@ -337,11 +336,11 @@ static int ad7949_spi_probe(struct spi_device *spi)
|
||||
ad7949_adc->resolution = spec->resolution;
|
||||
|
||||
/* Set SPI bits per word */
|
||||
if (spi_ctrl_mask & SPI_BPW_MASK(ad7949_adc->resolution)) {
|
||||
if (spi_is_bpw_supported(spi, ad7949_adc->resolution)) {
|
||||
spi->bits_per_word = ad7949_adc->resolution;
|
||||
} else if (spi_ctrl_mask == SPI_BPW_MASK(16)) {
|
||||
} else if (spi_is_bpw_supported(spi, 16)) {
|
||||
spi->bits_per_word = 16;
|
||||
} else if (spi_ctrl_mask == SPI_BPW_MASK(8)) {
|
||||
} else if (spi_is_bpw_supported(spi, 8)) {
|
||||
spi->bits_per_word = 8;
|
||||
} else {
|
||||
dev_err(dev, "unable to find common BPW with spi controller\n");
|
||||
|
||||
@@ -102,8 +102,7 @@ static int prox_read_raw(struct iio_dev *indio_dev,
|
||||
ret_type = prox_state->scale_precision;
|
||||
break;
|
||||
case IIO_CHAN_INFO_OFFSET:
|
||||
*val = hid_sensor_convert_exponent(
|
||||
prox_state->prox_attr.unit_expo);
|
||||
*val = 0;
|
||||
ret_type = IIO_VAL_INT;
|
||||
break;
|
||||
case IIO_CHAN_INFO_SAMP_FREQ:
|
||||
@@ -227,6 +226,11 @@ static int prox_parse_report(struct platform_device *pdev,
|
||||
dev_dbg(&pdev->dev, "prox %x:%x\n", st->prox_attr.index,
|
||||
st->prox_attr.report_id);
|
||||
|
||||
st->scale_precision = hid_sensor_format_scale(hsdev->usage,
|
||||
&st->prox_attr,
|
||||
&st->scale_pre_decml,
|
||||
&st->scale_post_decml);
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
|
||||
@@ -582,8 +582,8 @@ static int __ib_cache_gid_add(struct ib_device *ib_dev, u32 port,
|
||||
out_unlock:
|
||||
mutex_unlock(&table->lock);
|
||||
if (ret)
|
||||
pr_warn("%s: unable to add gid %pI6 error=%d\n",
|
||||
__func__, gid->raw, ret);
|
||||
pr_warn_ratelimited("%s: unable to add gid %pI6 error=%d\n",
|
||||
__func__, gid->raw, ret);
|
||||
return ret;
|
||||
}
|
||||
|
||||
|
||||
@@ -495,7 +495,7 @@ static irqreturn_t gpio_keys_irq_isr(int irq, void *dev_id)
|
||||
if (bdata->release_delay)
|
||||
hrtimer_start(&bdata->release_timer,
|
||||
ms_to_ktime(bdata->release_delay),
|
||||
HRTIMER_MODE_REL_HARD);
|
||||
HRTIMER_MODE_REL);
|
||||
out:
|
||||
spin_unlock_irqrestore(&bdata->lock, flags);
|
||||
return IRQ_HANDLED;
|
||||
@@ -632,7 +632,7 @@ static int gpio_keys_setup_key(struct platform_device *pdev,
|
||||
|
||||
bdata->release_delay = button->debounce_interval;
|
||||
hrtimer_init(&bdata->release_timer,
|
||||
CLOCK_REALTIME, HRTIMER_MODE_REL_HARD);
|
||||
CLOCK_REALTIME, HRTIMER_MODE_REL);
|
||||
bdata->release_timer.function = gpio_keys_irq_timer;
|
||||
|
||||
isr = gpio_keys_irq_isr;
|
||||
|
||||
@@ -165,6 +165,7 @@ static struct qcom_icc_node xm_pcie3_1 = {
|
||||
.id = SC7280_MASTER_PCIE_1,
|
||||
.channels = 1,
|
||||
.buswidth = 8,
|
||||
.num_links = 1,
|
||||
.links = { SC7280_SLAVE_ANOC_PCIE_GEM_NOC },
|
||||
};
|
||||
|
||||
|
||||
@@ -2858,7 +2858,12 @@ static int qcom_param_page_type_exec(struct nand_chip *chip, const struct nand_
|
||||
const struct nand_op_instr *instr = NULL;
|
||||
unsigned int op_id = 0;
|
||||
unsigned int len = 0;
|
||||
int ret;
|
||||
int ret, reg_base;
|
||||
|
||||
reg_base = NAND_READ_LOCATION_0;
|
||||
|
||||
if (nandc->props->qpic_v2)
|
||||
reg_base = NAND_READ_LOCATION_LAST_CW_0;
|
||||
|
||||
ret = qcom_parse_instructions(chip, subop, &q_op);
|
||||
if (ret)
|
||||
@@ -2910,7 +2915,10 @@ static int qcom_param_page_type_exec(struct nand_chip *chip, const struct nand_
|
||||
op_id = q_op.data_instr_idx;
|
||||
len = nand_subop_get_data_len(subop, op_id);
|
||||
|
||||
nandc_set_read_loc(chip, 0, 0, 0, len, 1);
|
||||
if (nandc->props->qpic_v2)
|
||||
nandc_set_read_loc_last(chip, reg_base, 0, len, 1);
|
||||
else
|
||||
nandc_set_read_loc_first(chip, reg_base, 0, len, 1);
|
||||
|
||||
if (!nandc->props->qpic_v2) {
|
||||
write_reg_dma(nandc, NAND_DEV_CMD_VLD, 1, 0);
|
||||
|
||||
@@ -125,13 +125,16 @@ void can_change_state(struct net_device *dev, struct can_frame *cf,
|
||||
EXPORT_SYMBOL_GPL(can_change_state);
|
||||
|
||||
/* CAN device restart for bus-off recovery */
|
||||
static void can_restart(struct net_device *dev)
|
||||
static int can_restart(struct net_device *dev)
|
||||
{
|
||||
struct can_priv *priv = netdev_priv(dev);
|
||||
struct sk_buff *skb;
|
||||
struct can_frame *cf;
|
||||
int err;
|
||||
|
||||
if (!priv->do_set_mode)
|
||||
return -EOPNOTSUPP;
|
||||
|
||||
if (netif_carrier_ok(dev))
|
||||
netdev_err(dev, "Attempt to restart for bus-off recovery, but carrier is OK?\n");
|
||||
|
||||
@@ -142,24 +145,25 @@ static void can_restart(struct net_device *dev)
|
||||
|
||||
/* send restart message upstream */
|
||||
skb = alloc_can_err_skb(dev, &cf);
|
||||
if (!skb)
|
||||
goto restart;
|
||||
|
||||
cf->can_id |= CAN_ERR_RESTARTED;
|
||||
|
||||
netif_rx(skb);
|
||||
|
||||
restart:
|
||||
netdev_dbg(dev, "restarted\n");
|
||||
priv->can_stats.restarts++;
|
||||
if (skb) {
|
||||
cf->can_id |= CAN_ERR_RESTARTED;
|
||||
netif_rx(skb);
|
||||
}
|
||||
|
||||
/* Now restart the device */
|
||||
netif_carrier_on(dev);
|
||||
err = priv->do_set_mode(dev, CAN_MODE_START);
|
||||
if (err) {
|
||||
netdev_err(dev, "Error %d during restart", err);
|
||||
netdev_err(dev, "Restart failed, error %pe\n", ERR_PTR(err));
|
||||
netif_carrier_off(dev);
|
||||
|
||||
return err;
|
||||
} else {
|
||||
netdev_dbg(dev, "Restarted\n");
|
||||
priv->can_stats.restarts++;
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void can_restart_work(struct work_struct *work)
|
||||
@@ -184,9 +188,8 @@ int can_restart_now(struct net_device *dev)
|
||||
return -EBUSY;
|
||||
|
||||
cancel_delayed_work_sync(&priv->restart_work);
|
||||
can_restart(dev);
|
||||
|
||||
return 0;
|
||||
return can_restart(dev);
|
||||
}
|
||||
|
||||
/* CAN bus-off
|
||||
|
||||
@@ -285,6 +285,12 @@ static int can_changelink(struct net_device *dev, struct nlattr *tb[],
|
||||
}
|
||||
|
||||
if (data[IFLA_CAN_RESTART_MS]) {
|
||||
if (!priv->do_set_mode) {
|
||||
NL_SET_ERR_MSG(extack,
|
||||
"Device doesn't support restart from Bus Off");
|
||||
return -EOPNOTSUPP;
|
||||
}
|
||||
|
||||
/* Do not allow changing restart delay while running */
|
||||
if (dev->flags & IFF_UP)
|
||||
return -EBUSY;
|
||||
@@ -292,6 +298,12 @@ static int can_changelink(struct net_device *dev, struct nlattr *tb[],
|
||||
}
|
||||
|
||||
if (data[IFLA_CAN_RESTART]) {
|
||||
if (!priv->do_set_mode) {
|
||||
NL_SET_ERR_MSG(extack,
|
||||
"Device doesn't support restart from Bus Off");
|
||||
return -EOPNOTSUPP;
|
||||
}
|
||||
|
||||
/* Do not allow a restart while not running */
|
||||
if (!(dev->flags & IFF_UP))
|
||||
return -EINVAL;
|
||||
|
||||
@@ -4654,12 +4654,19 @@ static int dpaa2_eth_connect_mac(struct dpaa2_eth_priv *priv)
|
||||
return PTR_ERR(dpmac_dev);
|
||||
}
|
||||
|
||||
if (IS_ERR(dpmac_dev) || dpmac_dev->dev.type != &fsl_mc_bus_dpmac_type)
|
||||
if (IS_ERR(dpmac_dev))
|
||||
return 0;
|
||||
|
||||
if (dpmac_dev->dev.type != &fsl_mc_bus_dpmac_type) {
|
||||
err = 0;
|
||||
goto out_put_device;
|
||||
}
|
||||
|
||||
mac = kzalloc(sizeof(struct dpaa2_mac), GFP_KERNEL);
|
||||
if (!mac)
|
||||
return -ENOMEM;
|
||||
if (!mac) {
|
||||
err = -ENOMEM;
|
||||
goto out_put_device;
|
||||
}
|
||||
|
||||
mac->mc_dev = dpmac_dev;
|
||||
mac->mc_io = priv->mc_io;
|
||||
@@ -4693,6 +4700,8 @@ err_close_mac:
|
||||
dpaa2_mac_close(mac);
|
||||
err_free_mac:
|
||||
kfree(mac);
|
||||
out_put_device:
|
||||
put_device(&dpmac_dev->dev);
|
||||
return err;
|
||||
}
|
||||
|
||||
|
||||
@@ -1447,12 +1447,19 @@ static int dpaa2_switch_port_connect_mac(struct ethsw_port_priv *port_priv)
|
||||
if (PTR_ERR(dpmac_dev) == -EPROBE_DEFER)
|
||||
return PTR_ERR(dpmac_dev);
|
||||
|
||||
if (IS_ERR(dpmac_dev) || dpmac_dev->dev.type != &fsl_mc_bus_dpmac_type)
|
||||
if (IS_ERR(dpmac_dev))
|
||||
return 0;
|
||||
|
||||
if (dpmac_dev->dev.type != &fsl_mc_bus_dpmac_type) {
|
||||
err = 0;
|
||||
goto out_put_device;
|
||||
}
|
||||
|
||||
mac = kzalloc(sizeof(*mac), GFP_KERNEL);
|
||||
if (!mac)
|
||||
return -ENOMEM;
|
||||
if (!mac) {
|
||||
err = -ENOMEM;
|
||||
goto out_put_device;
|
||||
}
|
||||
|
||||
mac->mc_dev = dpmac_dev;
|
||||
mac->mc_io = port_priv->ethsw_data->mc_io;
|
||||
@@ -1482,6 +1489,8 @@ err_close_mac:
|
||||
dpaa2_mac_close(mac);
|
||||
err_free_mac:
|
||||
kfree(mac);
|
||||
out_put_device:
|
||||
put_device(&dpmac_dev->dev);
|
||||
return err;
|
||||
}
|
||||
|
||||
|
||||
@@ -1801,49 +1801,56 @@ static void gve_turnup(struct gve_priv *priv)
|
||||
gve_set_napi_enabled(priv);
|
||||
}
|
||||
|
||||
static void gve_tx_timeout(struct net_device *dev, unsigned int txqueue)
|
||||
static struct gve_notify_block *gve_get_tx_notify_block(struct gve_priv *priv,
|
||||
unsigned int txqueue)
|
||||
{
|
||||
struct gve_notify_block *block;
|
||||
struct gve_tx_ring *tx = NULL;
|
||||
struct gve_priv *priv;
|
||||
u32 last_nic_done;
|
||||
u32 current_time;
|
||||
u32 ntfy_idx;
|
||||
|
||||
netdev_info(dev, "Timeout on tx queue, %d", txqueue);
|
||||
priv = netdev_priv(dev);
|
||||
if (txqueue > priv->tx_cfg.num_queues)
|
||||
goto reset;
|
||||
return NULL;
|
||||
|
||||
ntfy_idx = gve_tx_idx_to_ntfy(priv, txqueue);
|
||||
if (ntfy_idx >= priv->num_ntfy_blks)
|
||||
goto reset;
|
||||
return NULL;
|
||||
|
||||
block = &priv->ntfy_blocks[ntfy_idx];
|
||||
tx = block->tx;
|
||||
return &priv->ntfy_blocks[ntfy_idx];
|
||||
}
|
||||
|
||||
static bool gve_tx_timeout_try_q_kick(struct gve_priv *priv,
|
||||
unsigned int txqueue)
|
||||
{
|
||||
struct gve_notify_block *block;
|
||||
u32 current_time;
|
||||
|
||||
block = gve_get_tx_notify_block(priv, txqueue);
|
||||
|
||||
if (!block)
|
||||
return false;
|
||||
|
||||
current_time = jiffies_to_msecs(jiffies);
|
||||
if (tx->last_kick_msec + MIN_TX_TIMEOUT_GAP > current_time)
|
||||
goto reset;
|
||||
if (block->tx->last_kick_msec + MIN_TX_TIMEOUT_GAP > current_time)
|
||||
return false;
|
||||
|
||||
/* Check to see if there are missed completions, which will allow us to
|
||||
* kick the queue.
|
||||
*/
|
||||
last_nic_done = gve_tx_load_event_counter(priv, tx);
|
||||
if (last_nic_done - tx->done) {
|
||||
netdev_info(dev, "Kicking queue %d", txqueue);
|
||||
iowrite32be(GVE_IRQ_MASK, gve_irq_doorbell(priv, block));
|
||||
napi_schedule(&block->napi);
|
||||
tx->last_kick_msec = current_time;
|
||||
goto out;
|
||||
} // Else reset.
|
||||
netdev_info(priv->dev, "Kicking queue %d", txqueue);
|
||||
napi_schedule(&block->napi);
|
||||
block->tx->last_kick_msec = current_time;
|
||||
return true;
|
||||
}
|
||||
|
||||
reset:
|
||||
gve_schedule_reset(priv);
|
||||
static void gve_tx_timeout(struct net_device *dev, unsigned int txqueue)
|
||||
{
|
||||
struct gve_notify_block *block;
|
||||
struct gve_priv *priv;
|
||||
|
||||
out:
|
||||
if (tx)
|
||||
tx->queue_timeout++;
|
||||
netdev_info(dev, "Timeout on tx queue, %d", txqueue);
|
||||
priv = netdev_priv(dev);
|
||||
|
||||
if (!gve_tx_timeout_try_q_kick(priv, txqueue))
|
||||
gve_schedule_reset(priv);
|
||||
|
||||
block = gve_get_tx_notify_block(priv, txqueue);
|
||||
if (block)
|
||||
block->tx->queue_timeout++;
|
||||
priv->tx_timeo_cnt++;
|
||||
}
|
||||
|
||||
|
||||
@@ -11,6 +11,7 @@
|
||||
#include <linux/irq.h>
|
||||
#include <linux/ip.h>
|
||||
#include <linux/ipv6.h>
|
||||
#include <linux/iommu.h>
|
||||
#include <linux/module.h>
|
||||
#include <linux/pci.h>
|
||||
#include <linux/skbuff.h>
|
||||
@@ -1039,6 +1040,8 @@ static bool hns3_can_use_tx_sgl(struct hns3_enet_ring *ring,
|
||||
static void hns3_init_tx_spare_buffer(struct hns3_enet_ring *ring)
|
||||
{
|
||||
u32 alloc_size = ring->tqp->handle->kinfo.tx_spare_buf_size;
|
||||
struct net_device *netdev = ring_to_netdev(ring);
|
||||
struct hns3_nic_priv *priv = netdev_priv(netdev);
|
||||
struct hns3_tx_spare *tx_spare;
|
||||
struct page *page;
|
||||
dma_addr_t dma;
|
||||
@@ -1080,6 +1083,7 @@ static void hns3_init_tx_spare_buffer(struct hns3_enet_ring *ring)
|
||||
tx_spare->buf = page_address(page);
|
||||
tx_spare->len = PAGE_SIZE << order;
|
||||
ring->tx_spare = tx_spare;
|
||||
ring->tx_copybreak = priv->tx_copybreak;
|
||||
return;
|
||||
|
||||
dma_mapping_error:
|
||||
@@ -4879,6 +4883,30 @@ static void hns3_nic_dealloc_vector_data(struct hns3_nic_priv *priv)
|
||||
devm_kfree(&pdev->dev, priv->tqp_vector);
|
||||
}
|
||||
|
||||
static void hns3_update_tx_spare_buf_config(struct hns3_nic_priv *priv)
|
||||
{
|
||||
#define HNS3_MIN_SPARE_BUF_SIZE (2 * 1024 * 1024)
|
||||
#define HNS3_MAX_PACKET_SIZE (64 * 1024)
|
||||
|
||||
struct iommu_domain *domain = iommu_get_domain_for_dev(priv->dev);
|
||||
struct hnae3_ae_dev *ae_dev = hns3_get_ae_dev(priv->ae_handle);
|
||||
struct hnae3_handle *handle = priv->ae_handle;
|
||||
|
||||
if (ae_dev->dev_version < HNAE3_DEVICE_VERSION_V3)
|
||||
return;
|
||||
|
||||
if (!(domain && iommu_is_dma_domain(domain)))
|
||||
return;
|
||||
|
||||
priv->min_tx_copybreak = HNS3_MAX_PACKET_SIZE;
|
||||
priv->min_tx_spare_buf_size = HNS3_MIN_SPARE_BUF_SIZE;
|
||||
|
||||
if (priv->tx_copybreak < priv->min_tx_copybreak)
|
||||
priv->tx_copybreak = priv->min_tx_copybreak;
|
||||
if (handle->kinfo.tx_spare_buf_size < priv->min_tx_spare_buf_size)
|
||||
handle->kinfo.tx_spare_buf_size = priv->min_tx_spare_buf_size;
|
||||
}
|
||||
|
||||
static void hns3_ring_get_cfg(struct hnae3_queue *q, struct hns3_nic_priv *priv,
|
||||
unsigned int ring_type)
|
||||
{
|
||||
@@ -5113,6 +5141,7 @@ int hns3_init_all_ring(struct hns3_nic_priv *priv)
|
||||
int i, j;
|
||||
int ret;
|
||||
|
||||
hns3_update_tx_spare_buf_config(priv);
|
||||
for (i = 0; i < ring_num; i++) {
|
||||
ret = hns3_alloc_ring_memory(&priv->ring[i]);
|
||||
if (ret) {
|
||||
@@ -5317,6 +5346,8 @@ static int hns3_client_init(struct hnae3_handle *handle)
|
||||
priv->ae_handle = handle;
|
||||
priv->tx_timeout_count = 0;
|
||||
priv->max_non_tso_bd_num = ae_dev->dev_specs.max_non_tso_bd_num;
|
||||
priv->min_tx_copybreak = 0;
|
||||
priv->min_tx_spare_buf_size = 0;
|
||||
set_bit(HNS3_NIC_STATE_DOWN, &priv->state);
|
||||
|
||||
handle->msg_enable = netif_msg_init(debug, DEFAULT_MSG_LEVEL);
|
||||
|
||||
@@ -596,6 +596,8 @@ struct hns3_nic_priv {
|
||||
struct hns3_enet_coalesce rx_coal;
|
||||
u32 tx_copybreak;
|
||||
u32 rx_copybreak;
|
||||
u32 min_tx_copybreak;
|
||||
u32 min_tx_spare_buf_size;
|
||||
};
|
||||
|
||||
union l3_hdr_info {
|
||||
|
||||
@@ -9493,33 +9493,36 @@ static bool hclge_need_enable_vport_vlan_filter(struct hclge_vport *vport)
|
||||
return false;
|
||||
}
|
||||
|
||||
int hclge_enable_vport_vlan_filter(struct hclge_vport *vport, bool request_en)
|
||||
static int __hclge_enable_vport_vlan_filter(struct hclge_vport *vport,
|
||||
bool request_en)
|
||||
{
|
||||
struct hclge_dev *hdev = vport->back;
|
||||
bool need_en;
|
||||
int ret;
|
||||
|
||||
mutex_lock(&hdev->vport_lock);
|
||||
|
||||
vport->req_vlan_fltr_en = request_en;
|
||||
|
||||
need_en = hclge_need_enable_vport_vlan_filter(vport);
|
||||
if (need_en == vport->cur_vlan_fltr_en) {
|
||||
mutex_unlock(&hdev->vport_lock);
|
||||
if (need_en == vport->cur_vlan_fltr_en)
|
||||
return 0;
|
||||
}
|
||||
|
||||
ret = hclge_set_vport_vlan_filter(vport, need_en);
|
||||
if (ret) {
|
||||
mutex_unlock(&hdev->vport_lock);
|
||||
if (ret)
|
||||
return ret;
|
||||
}
|
||||
|
||||
vport->cur_vlan_fltr_en = need_en;
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
int hclge_enable_vport_vlan_filter(struct hclge_vport *vport, bool request_en)
|
||||
{
|
||||
struct hclge_dev *hdev = vport->back;
|
||||
int ret;
|
||||
|
||||
mutex_lock(&hdev->vport_lock);
|
||||
vport->req_vlan_fltr_en = request_en;
|
||||
ret = __hclge_enable_vport_vlan_filter(vport, request_en);
|
||||
mutex_unlock(&hdev->vport_lock);
|
||||
|
||||
return 0;
|
||||
return ret;
|
||||
}
|
||||
|
||||
static int hclge_enable_vlan_filter(struct hnae3_handle *handle, bool enable)
|
||||
@@ -10540,16 +10543,19 @@ static void hclge_sync_vlan_fltr_state(struct hclge_dev *hdev)
|
||||
&vport->state))
|
||||
continue;
|
||||
|
||||
ret = hclge_enable_vport_vlan_filter(vport,
|
||||
vport->req_vlan_fltr_en);
|
||||
mutex_lock(&hdev->vport_lock);
|
||||
ret = __hclge_enable_vport_vlan_filter(vport,
|
||||
vport->req_vlan_fltr_en);
|
||||
if (ret) {
|
||||
dev_err(&hdev->pdev->dev,
|
||||
"failed to sync vlan filter state for vport%u, ret = %d\n",
|
||||
vport->vport_id, ret);
|
||||
set_bit(HCLGE_VPORT_STATE_VLAN_FLTR_CHANGE,
|
||||
&vport->state);
|
||||
mutex_unlock(&hdev->vport_lock);
|
||||
return;
|
||||
}
|
||||
mutex_unlock(&hdev->vport_lock);
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
@@ -497,14 +497,14 @@ int hclge_ptp_init(struct hclge_dev *hdev)
|
||||
if (ret) {
|
||||
dev_err(&hdev->pdev->dev,
|
||||
"failed to init freq, ret = %d\n", ret);
|
||||
goto out;
|
||||
goto out_clear_int;
|
||||
}
|
||||
|
||||
ret = hclge_ptp_set_ts_mode(hdev, &hdev->ptp->ts_cfg);
|
||||
if (ret) {
|
||||
dev_err(&hdev->pdev->dev,
|
||||
"failed to init ts mode, ret = %d\n", ret);
|
||||
goto out;
|
||||
goto out_clear_int;
|
||||
}
|
||||
|
||||
ktime_get_real_ts64(&ts);
|
||||
@@ -512,7 +512,7 @@ int hclge_ptp_init(struct hclge_dev *hdev)
|
||||
if (ret) {
|
||||
dev_err(&hdev->pdev->dev,
|
||||
"failed to init ts time, ret = %d\n", ret);
|
||||
goto out;
|
||||
goto out_clear_int;
|
||||
}
|
||||
|
||||
set_bit(HCLGE_STATE_PTP_EN, &hdev->state);
|
||||
@@ -520,6 +520,9 @@ int hclge_ptp_init(struct hclge_dev *hdev)
|
||||
|
||||
return 0;
|
||||
|
||||
out_clear_int:
|
||||
clear_bit(HCLGE_PTP_FLAG_EN, &hdev->ptp->flags);
|
||||
hclge_ptp_int_en(hdev, false);
|
||||
out:
|
||||
hclge_ptp_destroy_clock(hdev);
|
||||
|
||||
|
||||
@@ -3029,11 +3029,7 @@ static void hclgevf_uninit_ae_dev(struct hnae3_ae_dev *ae_dev)
|
||||
|
||||
static u32 hclgevf_get_max_channels(struct hclgevf_dev *hdev)
|
||||
{
|
||||
struct hnae3_handle *nic = &hdev->nic;
|
||||
struct hnae3_knic_private_info *kinfo = &nic->kinfo;
|
||||
|
||||
return min_t(u32, hdev->rss_size_max,
|
||||
hdev->num_tqps / kinfo->tc_info.num_tc);
|
||||
return min(hdev->rss_size_max, hdev->num_tqps);
|
||||
}
|
||||
|
||||
/**
|
||||
|
||||
@@ -638,6 +638,9 @@
|
||||
/* For checksumming, the sum of all words in the NVM should equal 0xBABA. */
|
||||
#define NVM_SUM 0xBABA
|
||||
|
||||
/* Uninitialized ("empty") checksum word value */
|
||||
#define NVM_CHECKSUM_UNINITIALIZED 0xFFFF
|
||||
|
||||
/* PBA (printed board assembly) number words */
|
||||
#define NVM_PBA_OFFSET_0 8
|
||||
#define NVM_PBA_OFFSET_1 9
|
||||
|
||||
@@ -4274,6 +4274,8 @@ static s32 e1000_validate_nvm_checksum_ich8lan(struct e1000_hw *hw)
|
||||
ret_val = e1000e_update_nvm_checksum(hw);
|
||||
if (ret_val)
|
||||
return ret_val;
|
||||
} else if (hw->mac.type == e1000_pch_tgp) {
|
||||
return 0;
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
@@ -558,6 +558,12 @@ s32 e1000e_validate_nvm_checksum_generic(struct e1000_hw *hw)
|
||||
checksum += nvm_data;
|
||||
}
|
||||
|
||||
if (hw->mac.type == e1000_pch_tgp &&
|
||||
nvm_data == NVM_CHECKSUM_UNINITIALIZED) {
|
||||
e_dbg("Uninitialized NVM Checksum on TGP platform - ignoring\n");
|
||||
return 0;
|
||||
}
|
||||
|
||||
if (checksum != (u16)NVM_SUM) {
|
||||
e_dbg("NVM Checksum Invalid\n");
|
||||
return -E1000_ERR_NVM;
|
||||
|
||||
@@ -246,6 +246,7 @@ static const struct i40e_stats i40e_gstrings_net_stats[] = {
|
||||
I40E_NETDEV_STAT(rx_errors),
|
||||
I40E_NETDEV_STAT(tx_errors),
|
||||
I40E_NETDEV_STAT(rx_dropped),
|
||||
I40E_NETDEV_STAT(rx_missed_errors),
|
||||
I40E_NETDEV_STAT(tx_dropped),
|
||||
I40E_NETDEV_STAT(collisions),
|
||||
I40E_NETDEV_STAT(rx_length_errors),
|
||||
@@ -322,7 +323,7 @@ static const struct i40e_stats i40e_gstrings_stats[] = {
|
||||
I40E_PF_STAT("port.rx_broadcast", stats.eth.rx_broadcast),
|
||||
I40E_PF_STAT("port.tx_broadcast", stats.eth.tx_broadcast),
|
||||
I40E_PF_STAT("port.tx_errors", stats.eth.tx_errors),
|
||||
I40E_PF_STAT("port.rx_dropped", stats.eth.rx_discards),
|
||||
I40E_PF_STAT("port.rx_discards", stats.eth.rx_discards),
|
||||
I40E_PF_STAT("port.tx_dropped_link_down", stats.tx_dropped_link_down),
|
||||
I40E_PF_STAT("port.rx_crc_errors", stats.crc_errors),
|
||||
I40E_PF_STAT("port.illegal_bytes", stats.illegal_bytes),
|
||||
|
||||
@@ -509,6 +509,7 @@ static void i40e_get_netdev_stats_struct(struct net_device *netdev,
|
||||
stats->tx_dropped = vsi_stats->tx_dropped;
|
||||
stats->rx_errors = vsi_stats->rx_errors;
|
||||
stats->rx_dropped = vsi_stats->rx_dropped;
|
||||
stats->rx_missed_errors = vsi_stats->rx_missed_errors;
|
||||
stats->rx_crc_errors = vsi_stats->rx_crc_errors;
|
||||
stats->rx_length_errors = vsi_stats->rx_length_errors;
|
||||
}
|
||||
@@ -700,17 +701,13 @@ i40e_stats_update_rx_discards(struct i40e_vsi *vsi, struct i40e_hw *hw,
|
||||
struct i40e_eth_stats *stat_offset,
|
||||
struct i40e_eth_stats *stat)
|
||||
{
|
||||
u64 rx_rdpc, rx_rxerr;
|
||||
|
||||
i40e_stat_update32(hw, I40E_GLV_RDPC(stat_idx), offset_loaded,
|
||||
&stat_offset->rx_discards, &rx_rdpc);
|
||||
&stat_offset->rx_discards, &stat->rx_discards);
|
||||
i40e_stat_update64(hw,
|
||||
I40E_GL_RXERR1H(i40e_compute_pci_to_hw_id(vsi, hw)),
|
||||
I40E_GL_RXERR1L(i40e_compute_pci_to_hw_id(vsi, hw)),
|
||||
offset_loaded, &stat_offset->rx_discards_other,
|
||||
&rx_rxerr);
|
||||
|
||||
stat->rx_discards = rx_rdpc + rx_rxerr;
|
||||
&stat->rx_discards_other);
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -732,9 +729,6 @@ void i40e_update_eth_stats(struct i40e_vsi *vsi)
|
||||
i40e_stat_update32(hw, I40E_GLV_TEPC(stat_idx),
|
||||
vsi->stat_offsets_loaded,
|
||||
&oes->tx_errors, &es->tx_errors);
|
||||
i40e_stat_update32(hw, I40E_GLV_RDPC(stat_idx),
|
||||
vsi->stat_offsets_loaded,
|
||||
&oes->rx_discards, &es->rx_discards);
|
||||
i40e_stat_update32(hw, I40E_GLV_RUPP(stat_idx),
|
||||
vsi->stat_offsets_loaded,
|
||||
&oes->rx_unknown_protocol, &es->rx_unknown_protocol);
|
||||
@@ -991,8 +985,10 @@ static void i40e_update_vsi_stats(struct i40e_vsi *vsi)
|
||||
ns->tx_errors = es->tx_errors;
|
||||
ons->multicast = oes->rx_multicast;
|
||||
ns->multicast = es->rx_multicast;
|
||||
ons->rx_dropped = oes->rx_discards;
|
||||
ns->rx_dropped = es->rx_discards;
|
||||
ons->rx_dropped = oes->rx_discards_other;
|
||||
ns->rx_dropped = es->rx_discards_other;
|
||||
ons->rx_missed_errors = oes->rx_discards;
|
||||
ns->rx_missed_errors = es->rx_discards;
|
||||
ons->tx_dropped = oes->tx_discards;
|
||||
ns->tx_dropped = es->tx_discards;
|
||||
|
||||
|
||||
@@ -3143,10 +3143,10 @@ static int i40e_vc_del_mac_addr_msg(struct i40e_vf *vf, u8 *msg)
|
||||
const u8 *addr = al->list[i].addr;
|
||||
|
||||
/* Allow to delete VF primary MAC only if it was not set
|
||||
* administratively by PF or if VF is trusted.
|
||||
* administratively by PF.
|
||||
*/
|
||||
if (ether_addr_equal(addr, vf->default_lan_addr.addr)) {
|
||||
if (i40e_can_vf_change_mac(vf))
|
||||
if (!vf->pf_set_mac)
|
||||
was_unimac_deleted = true;
|
||||
else
|
||||
continue;
|
||||
@@ -5010,8 +5010,8 @@ int i40e_get_vf_stats(struct net_device *netdev, int vf_id,
|
||||
vf_stats->tx_bytes = stats->tx_bytes;
|
||||
vf_stats->broadcast = stats->rx_broadcast;
|
||||
vf_stats->multicast = stats->rx_multicast;
|
||||
vf_stats->rx_dropped = stats->rx_discards;
|
||||
vf_stats->tx_dropped = stats->tx_discards;
|
||||
vf_stats->rx_dropped = stats->rx_discards + stats->rx_discards_other;
|
||||
vf_stats->tx_dropped = stats->tx_errors;
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
@@ -1884,6 +1884,8 @@ enum ice_ddp_state ice_copy_and_init_pkg(struct ice_hw *hw, const u8 *buf,
|
||||
return ICE_DDP_PKG_ERR;
|
||||
|
||||
buf_copy = devm_kmemdup(ice_hw_to_dev(hw), buf, len, GFP_KERNEL);
|
||||
if (!buf_copy)
|
||||
return ICE_DDP_PKG_ERR;
|
||||
|
||||
state = ice_init_pkg(hw, buf_copy, len);
|
||||
if (!ice_is_init_pkg_successful(state)) {
|
||||
|
||||
@@ -1916,8 +1916,8 @@ static int cmd_exec(struct mlx5_core_dev *dev, void *in, int in_size, void *out,
|
||||
|
||||
err = mlx5_cmd_invoke(dev, inb, outb, out, out_size, callback, context,
|
||||
pages_queue, token, force_polling);
|
||||
if (callback)
|
||||
return err;
|
||||
if (callback && !err)
|
||||
return 0;
|
||||
|
||||
if (err > 0) /* Failed in FW, command didn't execute */
|
||||
err = deliv_status_to_err(err);
|
||||
|
||||
@@ -1170,19 +1170,19 @@ static void esw_set_peer_miss_rule_source_port(struct mlx5_eswitch *esw,
|
||||
static int esw_add_fdb_peer_miss_rules(struct mlx5_eswitch *esw,
|
||||
struct mlx5_core_dev *peer_dev)
|
||||
{
|
||||
struct mlx5_eswitch *peer_esw = peer_dev->priv.eswitch;
|
||||
struct mlx5_flow_destination dest = {};
|
||||
struct mlx5_flow_act flow_act = {0};
|
||||
struct mlx5_flow_handle **flows;
|
||||
/* total vports is the same for both e-switches */
|
||||
int nvports = esw->total_vports;
|
||||
struct mlx5_flow_handle *flow;
|
||||
struct mlx5_vport *peer_vport;
|
||||
struct mlx5_flow_spec *spec;
|
||||
struct mlx5_vport *vport;
|
||||
int err, pfindex;
|
||||
unsigned long i;
|
||||
void *misc;
|
||||
|
||||
if (!MLX5_VPORT_MANAGER(esw->dev) && !mlx5_core_is_ecpf_esw_manager(esw->dev))
|
||||
if (!MLX5_VPORT_MANAGER(peer_dev) &&
|
||||
!mlx5_core_is_ecpf_esw_manager(peer_dev))
|
||||
return 0;
|
||||
|
||||
spec = kvzalloc(sizeof(*spec), GFP_KERNEL);
|
||||
@@ -1191,7 +1191,7 @@ static int esw_add_fdb_peer_miss_rules(struct mlx5_eswitch *esw,
|
||||
|
||||
peer_miss_rules_setup(esw, peer_dev, spec, &dest);
|
||||
|
||||
flows = kvcalloc(nvports, sizeof(*flows), GFP_KERNEL);
|
||||
flows = kvcalloc(peer_esw->total_vports, sizeof(*flows), GFP_KERNEL);
|
||||
if (!flows) {
|
||||
err = -ENOMEM;
|
||||
goto alloc_flows_err;
|
||||
@@ -1201,10 +1201,10 @@ static int esw_add_fdb_peer_miss_rules(struct mlx5_eswitch *esw,
|
||||
misc = MLX5_ADDR_OF(fte_match_param, spec->match_value,
|
||||
misc_parameters);
|
||||
|
||||
if (mlx5_core_is_ecpf_esw_manager(esw->dev)) {
|
||||
vport = mlx5_eswitch_get_vport(esw, MLX5_VPORT_PF);
|
||||
esw_set_peer_miss_rule_source_port(esw, peer_dev->priv.eswitch,
|
||||
spec, MLX5_VPORT_PF);
|
||||
if (mlx5_core_is_ecpf_esw_manager(peer_dev)) {
|
||||
peer_vport = mlx5_eswitch_get_vport(peer_esw, MLX5_VPORT_PF);
|
||||
esw_set_peer_miss_rule_source_port(esw, peer_esw, spec,
|
||||
MLX5_VPORT_PF);
|
||||
|
||||
flow = mlx5_add_flow_rules(mlx5_eswitch_get_slow_fdb(esw),
|
||||
spec, &flow_act, &dest, 1);
|
||||
@@ -1212,11 +1212,11 @@ static int esw_add_fdb_peer_miss_rules(struct mlx5_eswitch *esw,
|
||||
err = PTR_ERR(flow);
|
||||
goto add_pf_flow_err;
|
||||
}
|
||||
flows[vport->index] = flow;
|
||||
flows[peer_vport->index] = flow;
|
||||
}
|
||||
|
||||
if (mlx5_ecpf_vport_exists(esw->dev)) {
|
||||
vport = mlx5_eswitch_get_vport(esw, MLX5_VPORT_ECPF);
|
||||
if (mlx5_ecpf_vport_exists(peer_dev)) {
|
||||
peer_vport = mlx5_eswitch_get_vport(peer_esw, MLX5_VPORT_ECPF);
|
||||
MLX5_SET(fte_match_set_misc, misc, source_port, MLX5_VPORT_ECPF);
|
||||
flow = mlx5_add_flow_rules(mlx5_eswitch_get_slow_fdb(esw),
|
||||
spec, &flow_act, &dest, 1);
|
||||
@@ -1224,13 +1224,14 @@ static int esw_add_fdb_peer_miss_rules(struct mlx5_eswitch *esw,
|
||||
err = PTR_ERR(flow);
|
||||
goto add_ecpf_flow_err;
|
||||
}
|
||||
flows[vport->index] = flow;
|
||||
flows[peer_vport->index] = flow;
|
||||
}
|
||||
|
||||
mlx5_esw_for_each_vf_vport(esw, i, vport, mlx5_core_max_vfs(esw->dev)) {
|
||||
mlx5_esw_for_each_vf_vport(peer_esw, i, peer_vport,
|
||||
mlx5_core_max_vfs(peer_dev)) {
|
||||
esw_set_peer_miss_rule_source_port(esw,
|
||||
peer_dev->priv.eswitch,
|
||||
spec, vport->vport);
|
||||
peer_esw,
|
||||
spec, peer_vport->vport);
|
||||
|
||||
flow = mlx5_add_flow_rules(mlx5_eswitch_get_slow_fdb(esw),
|
||||
spec, &flow_act, &dest, 1);
|
||||
@@ -1238,22 +1239,22 @@ static int esw_add_fdb_peer_miss_rules(struct mlx5_eswitch *esw,
|
||||
err = PTR_ERR(flow);
|
||||
goto add_vf_flow_err;
|
||||
}
|
||||
flows[vport->index] = flow;
|
||||
flows[peer_vport->index] = flow;
|
||||
}
|
||||
|
||||
if (mlx5_core_ec_sriov_enabled(esw->dev)) {
|
||||
mlx5_esw_for_each_ec_vf_vport(esw, i, vport, mlx5_core_max_ec_vfs(esw->dev)) {
|
||||
if (i >= mlx5_core_max_ec_vfs(peer_dev))
|
||||
break;
|
||||
esw_set_peer_miss_rule_source_port(esw, peer_dev->priv.eswitch,
|
||||
spec, vport->vport);
|
||||
if (mlx5_core_ec_sriov_enabled(peer_dev)) {
|
||||
mlx5_esw_for_each_ec_vf_vport(peer_esw, i, peer_vport,
|
||||
mlx5_core_max_ec_vfs(peer_dev)) {
|
||||
esw_set_peer_miss_rule_source_port(esw, peer_esw,
|
||||
spec,
|
||||
peer_vport->vport);
|
||||
flow = mlx5_add_flow_rules(esw->fdb_table.offloads.slow_fdb,
|
||||
spec, &flow_act, &dest, 1);
|
||||
if (IS_ERR(flow)) {
|
||||
err = PTR_ERR(flow);
|
||||
goto add_ec_vf_flow_err;
|
||||
}
|
||||
flows[vport->index] = flow;
|
||||
flows[peer_vport->index] = flow;
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1270,25 +1271,27 @@ static int esw_add_fdb_peer_miss_rules(struct mlx5_eswitch *esw,
|
||||
return 0;
|
||||
|
||||
add_ec_vf_flow_err:
|
||||
mlx5_esw_for_each_ec_vf_vport(esw, i, vport, mlx5_core_max_ec_vfs(esw->dev)) {
|
||||
if (!flows[vport->index])
|
||||
mlx5_esw_for_each_ec_vf_vport(peer_esw, i, peer_vport,
|
||||
mlx5_core_max_ec_vfs(peer_dev)) {
|
||||
if (!flows[peer_vport->index])
|
||||
continue;
|
||||
mlx5_del_flow_rules(flows[vport->index]);
|
||||
mlx5_del_flow_rules(flows[peer_vport->index]);
|
||||
}
|
||||
add_vf_flow_err:
|
||||
mlx5_esw_for_each_vf_vport(esw, i, vport, mlx5_core_max_vfs(esw->dev)) {
|
||||
if (!flows[vport->index])
|
||||
mlx5_esw_for_each_vf_vport(peer_esw, i, peer_vport,
|
||||
mlx5_core_max_vfs(peer_dev)) {
|
||||
if (!flows[peer_vport->index])
|
||||
continue;
|
||||
mlx5_del_flow_rules(flows[vport->index]);
|
||||
mlx5_del_flow_rules(flows[peer_vport->index]);
|
||||
}
|
||||
if (mlx5_ecpf_vport_exists(esw->dev)) {
|
||||
vport = mlx5_eswitch_get_vport(esw, MLX5_VPORT_ECPF);
|
||||
mlx5_del_flow_rules(flows[vport->index]);
|
||||
if (mlx5_ecpf_vport_exists(peer_dev)) {
|
||||
peer_vport = mlx5_eswitch_get_vport(peer_esw, MLX5_VPORT_ECPF);
|
||||
mlx5_del_flow_rules(flows[peer_vport->index]);
|
||||
}
|
||||
add_ecpf_flow_err:
|
||||
if (mlx5_core_is_ecpf_esw_manager(esw->dev)) {
|
||||
vport = mlx5_eswitch_get_vport(esw, MLX5_VPORT_PF);
|
||||
mlx5_del_flow_rules(flows[vport->index]);
|
||||
if (mlx5_core_is_ecpf_esw_manager(peer_dev)) {
|
||||
peer_vport = mlx5_eswitch_get_vport(peer_esw, MLX5_VPORT_PF);
|
||||
mlx5_del_flow_rules(flows[peer_vport->index]);
|
||||
}
|
||||
add_pf_flow_err:
|
||||
esw_warn(esw->dev, "FDB: Failed to add peer miss flow rule err %d\n", err);
|
||||
@@ -1301,37 +1304,34 @@ alloc_flows_err:
|
||||
static void esw_del_fdb_peer_miss_rules(struct mlx5_eswitch *esw,
|
||||
struct mlx5_core_dev *peer_dev)
|
||||
{
|
||||
struct mlx5_eswitch *peer_esw = peer_dev->priv.eswitch;
|
||||
u16 peer_index = mlx5_get_dev_index(peer_dev);
|
||||
struct mlx5_flow_handle **flows;
|
||||
struct mlx5_vport *vport;
|
||||
struct mlx5_vport *peer_vport;
|
||||
unsigned long i;
|
||||
|
||||
flows = esw->fdb_table.offloads.peer_miss_rules[peer_index];
|
||||
if (!flows)
|
||||
return;
|
||||
|
||||
if (mlx5_core_ec_sriov_enabled(esw->dev)) {
|
||||
mlx5_esw_for_each_ec_vf_vport(esw, i, vport, mlx5_core_max_ec_vfs(esw->dev)) {
|
||||
/* The flow for a particular vport could be NULL if the other ECPF
|
||||
* has fewer or no VFs enabled
|
||||
*/
|
||||
if (!flows[vport->index])
|
||||
continue;
|
||||
mlx5_del_flow_rules(flows[vport->index]);
|
||||
}
|
||||
if (mlx5_core_ec_sriov_enabled(peer_dev)) {
|
||||
mlx5_esw_for_each_ec_vf_vport(peer_esw, i, peer_vport,
|
||||
mlx5_core_max_ec_vfs(peer_dev))
|
||||
mlx5_del_flow_rules(flows[peer_vport->index]);
|
||||
}
|
||||
|
||||
mlx5_esw_for_each_vf_vport(esw, i, vport, mlx5_core_max_vfs(esw->dev))
|
||||
mlx5_del_flow_rules(flows[vport->index]);
|
||||
mlx5_esw_for_each_vf_vport(peer_esw, i, peer_vport,
|
||||
mlx5_core_max_vfs(peer_dev))
|
||||
mlx5_del_flow_rules(flows[peer_vport->index]);
|
||||
|
||||
if (mlx5_ecpf_vport_exists(esw->dev)) {
|
||||
vport = mlx5_eswitch_get_vport(esw, MLX5_VPORT_ECPF);
|
||||
mlx5_del_flow_rules(flows[vport->index]);
|
||||
if (mlx5_ecpf_vport_exists(peer_dev)) {
|
||||
peer_vport = mlx5_eswitch_get_vport(peer_esw, MLX5_VPORT_ECPF);
|
||||
mlx5_del_flow_rules(flows[peer_vport->index]);
|
||||
}
|
||||
|
||||
if (mlx5_core_is_ecpf_esw_manager(esw->dev)) {
|
||||
vport = mlx5_eswitch_get_vport(esw, MLX5_VPORT_PF);
|
||||
mlx5_del_flow_rules(flows[vport->index]);
|
||||
if (mlx5_core_is_ecpf_esw_manager(peer_dev)) {
|
||||
peer_vport = mlx5_eswitch_get_vport(peer_esw, MLX5_VPORT_PF);
|
||||
mlx5_del_flow_rules(flows[peer_vport->index]);
|
||||
}
|
||||
|
||||
kvfree(flows);
|
||||
|
||||
@@ -1087,6 +1087,9 @@ static void mt7921_sta_set_decap_offload(struct ieee80211_hw *hw,
|
||||
struct mt792x_sta *msta = (struct mt792x_sta *)sta->drv_priv;
|
||||
struct mt792x_dev *dev = mt792x_hw_dev(hw);
|
||||
|
||||
if (!msta->wcid.sta)
|
||||
return;
|
||||
|
||||
mt792x_mutex_acquire(dev);
|
||||
|
||||
if (enabled)
|
||||
|
||||
@@ -58,6 +58,8 @@ obj-$(CONFIG_X86_PLATFORM_DRIVERS_HP) += hp/
|
||||
# Hewlett Packard Enterprise
|
||||
obj-$(CONFIG_UV_SYSFS) += uv_sysfs.o
|
||||
|
||||
obj-$(CONFIG_FW_ATTR_CLASS) += firmware_attributes_class.o
|
||||
|
||||
# IBM Thinkpad and Lenovo
|
||||
obj-$(CONFIG_IBM_RTL) += ibm_rtl.o
|
||||
obj-$(CONFIG_IDEAPAD_LAPTOP) += ideapad-laptop.o
|
||||
@@ -108,7 +110,6 @@ obj-$(CONFIG_SYSTEM76_ACPI) += system76_acpi.o
|
||||
obj-$(CONFIG_TOPSTAR_LAPTOP) += topstar-laptop.o
|
||||
|
||||
# Platform drivers
|
||||
obj-$(CONFIG_FW_ATTR_CLASS) += firmware_attributes_class.o
|
||||
obj-$(CONFIG_SERIAL_MULTI_INSTANTIATE) += serial-multi-instantiate.o
|
||||
obj-$(CONFIG_MLX_PLATFORM) += mlx-platform.o
|
||||
obj-$(CONFIG_TOUCHSCREEN_DMI) += touchscreen_dmi.o
|
||||
|
||||
@@ -1600,7 +1600,7 @@ static int ideapad_kbd_bl_init(struct ideapad_private *priv)
|
||||
priv->kbd_bl.led.name = "platform::" LED_FUNCTION_KBD_BACKLIGHT;
|
||||
priv->kbd_bl.led.brightness_get = ideapad_kbd_bl_led_cdev_brightness_get;
|
||||
priv->kbd_bl.led.brightness_set_blocking = ideapad_kbd_bl_led_cdev_brightness_set;
|
||||
priv->kbd_bl.led.flags = LED_BRIGHT_HW_CHANGED;
|
||||
priv->kbd_bl.led.flags = LED_BRIGHT_HW_CHANGED | LED_RETAIN_AT_SHUTDOWN;
|
||||
|
||||
err = led_classdev_register(&priv->platform_device->dev, &priv->kbd_bl.led);
|
||||
if (err)
|
||||
|
||||
@@ -5422,6 +5422,7 @@ static void regulator_remove_coupling(struct regulator_dev *rdev)
|
||||
ERR_PTR(err));
|
||||
}
|
||||
|
||||
rdev->coupling_desc.n_coupled = 0;
|
||||
kfree(rdev->coupling_desc.coupled_rdevs);
|
||||
rdev->coupling_desc.coupled_rdevs = NULL;
|
||||
}
|
||||
|
||||
@@ -130,6 +130,7 @@ static int ism_cmd(struct ism_dev *ism, void *cmd)
|
||||
struct ism_req_hdr *req = cmd;
|
||||
struct ism_resp_hdr *resp = cmd;
|
||||
|
||||
spin_lock(&ism->cmd_lock);
|
||||
__ism_write_cmd(ism, req + 1, sizeof(*req), req->len - sizeof(*req));
|
||||
__ism_write_cmd(ism, req, 0, sizeof(*req));
|
||||
|
||||
@@ -143,6 +144,7 @@ static int ism_cmd(struct ism_dev *ism, void *cmd)
|
||||
}
|
||||
__ism_read_cmd(ism, resp + 1, sizeof(*resp), resp->len - sizeof(*resp));
|
||||
out:
|
||||
spin_unlock(&ism->cmd_lock);
|
||||
return resp->ret;
|
||||
}
|
||||
|
||||
@@ -630,6 +632,7 @@ static int ism_probe(struct pci_dev *pdev, const struct pci_device_id *id)
|
||||
return -ENOMEM;
|
||||
|
||||
spin_lock_init(&ism->lock);
|
||||
spin_lock_init(&ism->cmd_lock);
|
||||
dev_set_drvdata(&pdev->dev, ism);
|
||||
ism->pdev = pdev;
|
||||
ism->dev.parent = &pdev->dev;
|
||||
|
||||
@@ -1870,11 +1870,6 @@ static int cqspi_probe(struct platform_device *pdev)
|
||||
|
||||
pm_runtime_enable(dev);
|
||||
|
||||
if (cqspi->rx_chan) {
|
||||
dma_release_channel(cqspi->rx_chan);
|
||||
goto probe_setup_failed;
|
||||
}
|
||||
|
||||
ret = spi_register_controller(host);
|
||||
if (ret) {
|
||||
dev_err(&pdev->dev, "failed to register SPI ctlr %d\n", ret);
|
||||
|
||||
@@ -720,8 +720,7 @@ int vchiq_shutdown(struct vchiq_instance *instance)
|
||||
int status = 0;
|
||||
struct vchiq_state *state = instance->state;
|
||||
|
||||
if (mutex_lock_killable(&state->mutex))
|
||||
return -EAGAIN;
|
||||
mutex_lock(&state->mutex);
|
||||
|
||||
/* Remove all services */
|
||||
vchiq_shutdown_internal(state, instance);
|
||||
|
||||
@@ -1041,7 +1041,7 @@ static int tcpm_set_attached_state(struct tcpm_port *port, bool attached)
|
||||
port->data_role);
|
||||
}
|
||||
|
||||
static int tcpm_set_roles(struct tcpm_port *port, bool attached,
|
||||
static int tcpm_set_roles(struct tcpm_port *port, bool attached, int state,
|
||||
enum typec_role role, enum typec_data_role data)
|
||||
{
|
||||
enum typec_orientation orientation;
|
||||
@@ -1078,7 +1078,7 @@ static int tcpm_set_roles(struct tcpm_port *port, bool attached,
|
||||
}
|
||||
}
|
||||
|
||||
ret = tcpm_mux_set(port, TYPEC_STATE_USB, usb_role, orientation);
|
||||
ret = tcpm_mux_set(port, state, usb_role, orientation);
|
||||
if (ret < 0)
|
||||
return ret;
|
||||
|
||||
@@ -3622,16 +3622,6 @@ static int tcpm_src_attach(struct tcpm_port *port)
|
||||
|
||||
tcpm_enable_auto_vbus_discharge(port, true);
|
||||
|
||||
ret = tcpm_set_roles(port, true, TYPEC_SOURCE, tcpm_data_role_for_source(port));
|
||||
if (ret < 0)
|
||||
return ret;
|
||||
|
||||
if (port->pd_supported) {
|
||||
ret = port->tcpc->set_pd_rx(port->tcpc, true);
|
||||
if (ret < 0)
|
||||
goto out_disable_mux;
|
||||
}
|
||||
|
||||
/*
|
||||
* USB Type-C specification, version 1.2,
|
||||
* chapter 4.5.2.2.8.1 (Attached.SRC Requirements)
|
||||
@@ -3641,13 +3631,24 @@ static int tcpm_src_attach(struct tcpm_port *port)
|
||||
(polarity == TYPEC_POLARITY_CC2 && port->cc1 == TYPEC_CC_RA)) {
|
||||
ret = tcpm_set_vconn(port, true);
|
||||
if (ret < 0)
|
||||
goto out_disable_pd;
|
||||
return ret;
|
||||
}
|
||||
|
||||
ret = tcpm_set_vbus(port, true);
|
||||
if (ret < 0)
|
||||
goto out_disable_vconn;
|
||||
|
||||
ret = tcpm_set_roles(port, true, TYPEC_STATE_USB, TYPEC_SOURCE,
|
||||
tcpm_data_role_for_source(port));
|
||||
if (ret < 0)
|
||||
goto out_disable_vbus;
|
||||
|
||||
if (port->pd_supported) {
|
||||
ret = port->tcpc->set_pd_rx(port->tcpc, true);
|
||||
if (ret < 0)
|
||||
goto out_disable_mux;
|
||||
}
|
||||
|
||||
port->pd_capable = false;
|
||||
|
||||
port->partner = NULL;
|
||||
@@ -3657,14 +3658,14 @@ static int tcpm_src_attach(struct tcpm_port *port)
|
||||
|
||||
return 0;
|
||||
|
||||
out_disable_vconn:
|
||||
tcpm_set_vconn(port, false);
|
||||
out_disable_pd:
|
||||
if (port->pd_supported)
|
||||
port->tcpc->set_pd_rx(port->tcpc, false);
|
||||
out_disable_mux:
|
||||
tcpm_mux_set(port, TYPEC_STATE_SAFE, USB_ROLE_NONE,
|
||||
TYPEC_ORIENTATION_NONE);
|
||||
out_disable_vbus:
|
||||
tcpm_set_vbus(port, false);
|
||||
out_disable_vconn:
|
||||
tcpm_set_vconn(port, false);
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
@@ -3780,7 +3781,8 @@ static int tcpm_snk_attach(struct tcpm_port *port)
|
||||
|
||||
tcpm_enable_auto_vbus_discharge(port, true);
|
||||
|
||||
ret = tcpm_set_roles(port, true, TYPEC_SINK, tcpm_data_role_for_sink(port));
|
||||
ret = tcpm_set_roles(port, true, TYPEC_STATE_USB,
|
||||
TYPEC_SINK, tcpm_data_role_for_sink(port));
|
||||
if (ret < 0)
|
||||
return ret;
|
||||
|
||||
@@ -3802,12 +3804,24 @@ static void tcpm_snk_detach(struct tcpm_port *port)
|
||||
static int tcpm_acc_attach(struct tcpm_port *port)
|
||||
{
|
||||
int ret;
|
||||
enum typec_role role;
|
||||
enum typec_data_role data;
|
||||
int state = TYPEC_STATE_USB;
|
||||
|
||||
if (port->attached)
|
||||
return 0;
|
||||
|
||||
ret = tcpm_set_roles(port, true, TYPEC_SOURCE,
|
||||
tcpm_data_role_for_source(port));
|
||||
role = tcpm_port_is_sink(port) ? TYPEC_SINK : TYPEC_SOURCE;
|
||||
data = tcpm_port_is_sink(port) ? tcpm_data_role_for_sink(port)
|
||||
: tcpm_data_role_for_source(port);
|
||||
|
||||
if (tcpm_port_is_audio(port))
|
||||
state = TYPEC_MODE_AUDIO;
|
||||
|
||||
if (tcpm_port_is_debug(port))
|
||||
state = TYPEC_MODE_DEBUG;
|
||||
|
||||
ret = tcpm_set_roles(port, true, state, role, data);
|
||||
if (ret < 0)
|
||||
return ret;
|
||||
|
||||
@@ -4509,7 +4523,7 @@ static void run_state_machine(struct tcpm_port *port)
|
||||
*/
|
||||
tcpm_set_vconn(port, false);
|
||||
tcpm_set_vbus(port, false);
|
||||
tcpm_set_roles(port, port->self_powered, TYPEC_SOURCE,
|
||||
tcpm_set_roles(port, port->self_powered, TYPEC_STATE_USB, TYPEC_SOURCE,
|
||||
tcpm_data_role_for_source(port));
|
||||
/*
|
||||
* If tcpc fails to notify vbus off, TCPM will wait for PD_T_SAFE_0V +
|
||||
@@ -4541,7 +4555,7 @@ static void run_state_machine(struct tcpm_port *port)
|
||||
tcpm_set_vconn(port, false);
|
||||
if (port->pd_capable)
|
||||
tcpm_set_charge(port, false);
|
||||
tcpm_set_roles(port, port->self_powered, TYPEC_SINK,
|
||||
tcpm_set_roles(port, port->self_powered, TYPEC_STATE_USB, TYPEC_SINK,
|
||||
tcpm_data_role_for_sink(port));
|
||||
/*
|
||||
* VBUS may or may not toggle, depending on the adapter.
|
||||
@@ -4647,10 +4661,10 @@ static void run_state_machine(struct tcpm_port *port)
|
||||
case DR_SWAP_CHANGE_DR:
|
||||
tcpm_unregister_altmodes(port);
|
||||
if (port->data_role == TYPEC_HOST)
|
||||
tcpm_set_roles(port, true, port->pwr_role,
|
||||
tcpm_set_roles(port, true, TYPEC_STATE_USB, port->pwr_role,
|
||||
TYPEC_DEVICE);
|
||||
else
|
||||
tcpm_set_roles(port, true, port->pwr_role,
|
||||
tcpm_set_roles(port, true, TYPEC_STATE_USB, port->pwr_role,
|
||||
TYPEC_HOST);
|
||||
tcpm_ams_finish(port);
|
||||
tcpm_set_state(port, ready_state(port), 0);
|
||||
|
||||
@@ -2739,7 +2739,7 @@ int virtqueue_resize(struct virtqueue *_vq, u32 num,
|
||||
void (*recycle)(struct virtqueue *vq, void *buf))
|
||||
{
|
||||
struct vring_virtqueue *vq = to_vvq(_vq);
|
||||
int err;
|
||||
int err, err_reset;
|
||||
|
||||
if (num > vq->vq.num_max)
|
||||
return -E2BIG;
|
||||
@@ -2759,7 +2759,11 @@ int virtqueue_resize(struct virtqueue *_vq, u32 num,
|
||||
else
|
||||
err = virtqueue_resize_split(_vq, num);
|
||||
|
||||
return virtqueue_enable_after_reset(_vq);
|
||||
err_reset = virtqueue_enable_after_reset(_vq);
|
||||
if (err_reset)
|
||||
return err_reset;
|
||||
|
||||
return err;
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(virtqueue_resize);
|
||||
|
||||
|
||||
@@ -342,14 +342,12 @@ static int z_erofs_transform_plain(struct z_erofs_decompress_req *rq,
|
||||
|
||||
if (outpages > inpages) {
|
||||
DBG_BUGON(!rq->out[outpages - 1]);
|
||||
if (rq->out[outpages - 1] != rq->in[inpages - 1]) {
|
||||
if (rq->out[outpages - 1] != rq->in[inpages - 1])
|
||||
memcpy_to_page(rq->out[outpages - 1], 0, src +
|
||||
(interlaced_offset ? 0 : righthalf),
|
||||
lefthalf);
|
||||
} else if (!interlaced_offset) {
|
||||
else if (!interlaced_offset)
|
||||
memmove(src, src + righthalf, lefthalf);
|
||||
flush_dcache_page(rq->in[inpages - 1]);
|
||||
}
|
||||
}
|
||||
kunmap_local(src);
|
||||
return 0;
|
||||
|
||||
@@ -122,9 +122,11 @@ static inline unsigned int z_erofs_pclusterpages(struct z_erofs_pcluster *pcl)
|
||||
|
||||
/*
|
||||
* bit 30: I/O error occurred on this page
|
||||
* bit 29: CPU has dirty data in D-cache (needs aliasing handling);
|
||||
* bit 0 - 29: remaining parts to complete this page
|
||||
*/
|
||||
#define Z_EROFS_PAGE_EIO (1 << 30)
|
||||
#define Z_EROFS_ONLINEPAGE_EIO 30
|
||||
#define Z_EROFS_ONLINEPAGE_DIRTY 29
|
||||
|
||||
static inline void z_erofs_onlinepage_init(struct page *page)
|
||||
{
|
||||
@@ -143,7 +145,7 @@ static inline void z_erofs_onlinepage_split(struct page *page)
|
||||
atomic_inc((atomic_t *)&page->private);
|
||||
}
|
||||
|
||||
static void z_erofs_onlinepage_endio(struct page *page, int err)
|
||||
static void z_erofs_onlinepage_end(struct page *page, int err, bool dirty)
|
||||
{
|
||||
int orig, v;
|
||||
|
||||
@@ -151,16 +153,20 @@ static void z_erofs_onlinepage_endio(struct page *page, int err)
|
||||
|
||||
do {
|
||||
orig = atomic_read((atomic_t *)&page->private);
|
||||
v = (orig - 1) | (err ? Z_EROFS_PAGE_EIO : 0);
|
||||
DBG_BUGON(orig <= 0);
|
||||
v = dirty << Z_EROFS_ONLINEPAGE_DIRTY;
|
||||
v |= (orig - 1) | (!!err << Z_EROFS_ONLINEPAGE_EIO);
|
||||
} while (atomic_cmpxchg((atomic_t *)&page->private, orig, v) != orig);
|
||||
|
||||
if (!(v & ~Z_EROFS_PAGE_EIO)) {
|
||||
set_page_private(page, 0);
|
||||
ClearPagePrivate(page);
|
||||
if (!(v & Z_EROFS_PAGE_EIO))
|
||||
SetPageUptodate(page);
|
||||
unlock_page(page);
|
||||
}
|
||||
if (v & (BIT(Z_EROFS_ONLINEPAGE_DIRTY) - 1))
|
||||
return;
|
||||
set_page_private(page, 0);
|
||||
ClearPagePrivate(page);
|
||||
if (v & BIT(Z_EROFS_ONLINEPAGE_DIRTY))
|
||||
flush_dcache_page(page);
|
||||
if (!(v & BIT(Z_EROFS_ONLINEPAGE_EIO)))
|
||||
SetPageUptodate(page);
|
||||
unlock_page(page);
|
||||
}
|
||||
|
||||
#define Z_EROFS_ONSTACK_PAGES 32
|
||||
@@ -1060,7 +1066,7 @@ next_part:
|
||||
goto repeat;
|
||||
|
||||
out:
|
||||
z_erofs_onlinepage_endio(page, err);
|
||||
z_erofs_onlinepage_end(page, err, false);
|
||||
return err;
|
||||
}
|
||||
|
||||
@@ -1163,7 +1169,7 @@ static void z_erofs_fill_other_copies(struct z_erofs_decompress_backend *be,
|
||||
cur += len;
|
||||
}
|
||||
kunmap_local(dst);
|
||||
z_erofs_onlinepage_endio(bvi->bvec.page, err);
|
||||
z_erofs_onlinepage_end(bvi->bvec.page, err, true);
|
||||
list_del(p);
|
||||
kfree(bvi);
|
||||
}
|
||||
@@ -1333,7 +1339,7 @@ out:
|
||||
/* recycle all individual short-lived pages */
|
||||
if (z_erofs_put_shortlivedpage(be->pagepool, page))
|
||||
continue;
|
||||
z_erofs_onlinepage_endio(page, err);
|
||||
z_erofs_onlinepage_end(page, err, true);
|
||||
}
|
||||
|
||||
if (be->decompressed_pages != be->onstack_pages)
|
||||
|
||||
@@ -3029,14 +3029,23 @@ static void duplicateIXtree(struct super_block *sb, s64 blkno,
|
||||
*
|
||||
* RETURN VALUES:
|
||||
* 0 - success
|
||||
* -ENOMEM - insufficient memory
|
||||
* -EINVAL - unexpected inode type
|
||||
*/
|
||||
static int copy_from_dinode(struct dinode * dip, struct inode *ip)
|
||||
{
|
||||
struct jfs_inode_info *jfs_ip = JFS_IP(ip);
|
||||
struct jfs_sb_info *sbi = JFS_SBI(ip->i_sb);
|
||||
int fileset = le32_to_cpu(dip->di_fileset);
|
||||
|
||||
jfs_ip->fileset = le32_to_cpu(dip->di_fileset);
|
||||
switch (fileset) {
|
||||
case AGGR_RESERVED_I: case AGGREGATE_I: case BMAP_I:
|
||||
case LOG_I: case BADBLOCK_I: case FILESYSTEM_I:
|
||||
break;
|
||||
default:
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
jfs_ip->fileset = fileset;
|
||||
jfs_ip->mode2 = le32_to_cpu(dip->di_mode);
|
||||
jfs_set_inode_flags(ip);
|
||||
|
||||
|
||||
@@ -517,11 +517,18 @@ static int __nilfs_read_inode(struct super_block *sb,
|
||||
inode->i_op = &nilfs_symlink_inode_operations;
|
||||
inode_nohighmem(inode);
|
||||
inode->i_mapping->a_ops = &nilfs_aops;
|
||||
} else {
|
||||
} else if (S_ISCHR(inode->i_mode) || S_ISBLK(inode->i_mode) ||
|
||||
S_ISFIFO(inode->i_mode) || S_ISSOCK(inode->i_mode)) {
|
||||
inode->i_op = &nilfs_special_inode_operations;
|
||||
init_special_inode(
|
||||
inode, inode->i_mode,
|
||||
huge_decode_dev(le64_to_cpu(raw_inode->i_device_code)));
|
||||
} else {
|
||||
nilfs_error(sb,
|
||||
"invalid file type bits in mode 0%o for inode %lu",
|
||||
inode->i_mode, ino);
|
||||
err = -EIO;
|
||||
goto failed_unmap;
|
||||
}
|
||||
nilfs_ifile_unmap_inode(root->ifile, ino, bh);
|
||||
brelse(bh);
|
||||
|
||||
@@ -39,8 +39,10 @@ void ksmbd_conn_free(struct ksmbd_conn *conn)
|
||||
xa_destroy(&conn->sessions);
|
||||
kvfree(conn->request_buf);
|
||||
kfree(conn->preauth_info);
|
||||
if (atomic_dec_and_test(&conn->refcnt))
|
||||
if (atomic_dec_and_test(&conn->refcnt)) {
|
||||
conn->transport->ops->free_transport(conn->transport);
|
||||
kfree(conn);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
|
||||
@@ -132,6 +132,7 @@ struct ksmbd_transport_ops {
|
||||
void *buf, unsigned int len,
|
||||
struct smb2_buffer_desc_v1 *desc,
|
||||
unsigned int desc_len);
|
||||
void (*free_transport)(struct ksmbd_transport *kt);
|
||||
};
|
||||
|
||||
struct ksmbd_transport {
|
||||
|
||||
@@ -158,7 +158,8 @@ struct smb_direct_transport {
|
||||
};
|
||||
|
||||
#define KSMBD_TRANS(t) ((struct ksmbd_transport *)&((t)->transport))
|
||||
|
||||
#define SMBD_TRANS(t) ((struct smb_direct_transport *)container_of(t, \
|
||||
struct smb_direct_transport, transport))
|
||||
enum {
|
||||
SMB_DIRECT_MSG_NEGOTIATE_REQ = 0,
|
||||
SMB_DIRECT_MSG_DATA_TRANSFER
|
||||
@@ -409,6 +410,11 @@ err:
|
||||
return NULL;
|
||||
}
|
||||
|
||||
static void smb_direct_free_transport(struct ksmbd_transport *kt)
|
||||
{
|
||||
kfree(SMBD_TRANS(kt));
|
||||
}
|
||||
|
||||
static void free_transport(struct smb_direct_transport *t)
|
||||
{
|
||||
struct smb_direct_recvmsg *recvmsg;
|
||||
@@ -455,7 +461,6 @@ static void free_transport(struct smb_direct_transport *t)
|
||||
|
||||
smb_direct_destroy_pools(t);
|
||||
ksmbd_conn_free(KSMBD_TRANS(t)->conn);
|
||||
kfree(t);
|
||||
}
|
||||
|
||||
static struct smb_direct_sendmsg
|
||||
@@ -2301,4 +2306,5 @@ static struct ksmbd_transport_ops ksmbd_smb_direct_transport_ops = {
|
||||
.read = smb_direct_read,
|
||||
.rdma_read = smb_direct_rdma_read,
|
||||
.rdma_write = smb_direct_rdma_write,
|
||||
.free_transport = smb_direct_free_transport,
|
||||
};
|
||||
|
||||
@@ -93,15 +93,19 @@ static struct tcp_transport *alloc_transport(struct socket *client_sk)
|
||||
return t;
|
||||
}
|
||||
|
||||
static void ksmbd_tcp_free_transport(struct ksmbd_transport *kt)
|
||||
{
|
||||
struct tcp_transport *t = TCP_TRANS(kt);
|
||||
|
||||
sock_release(t->sock);
|
||||
kfree(t->iov);
|
||||
kfree(t);
|
||||
}
|
||||
|
||||
static void free_transport(struct tcp_transport *t)
|
||||
{
|
||||
kernel_sock_shutdown(t->sock, SHUT_RDWR);
|
||||
sock_release(t->sock);
|
||||
t->sock = NULL;
|
||||
|
||||
ksmbd_conn_free(KSMBD_TRANS(t)->conn);
|
||||
kfree(t->iov);
|
||||
kfree(t);
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -655,4 +659,5 @@ static struct ksmbd_transport_ops ksmbd_tcp_transport_ops = {
|
||||
.read = ksmbd_tcp_read,
|
||||
.writev = ksmbd_tcp_writev,
|
||||
.disconnect = ksmbd_tcp_disconnect,
|
||||
.free_transport = ksmbd_tcp_free_transport,
|
||||
};
|
||||
|
||||
@@ -7,6 +7,7 @@
|
||||
#define __KSMBD_TRANSPORT_TCP_H__
|
||||
|
||||
int ksmbd_tcp_set_interfaces(char *ifc_list, int ifc_list_sz);
|
||||
void ksmbd_free_transport(struct ksmbd_transport *kt);
|
||||
int ksmbd_tcp_init(void);
|
||||
void ksmbd_tcp_destroy(void);
|
||||
|
||||
|
||||
@@ -28,6 +28,7 @@ struct ism_dmb {
|
||||
|
||||
struct ism_dev {
|
||||
spinlock_t lock; /* protects the ism device */
|
||||
spinlock_t cmd_lock; /* serializes cmds */
|
||||
struct list_head list;
|
||||
struct pci_dev *pdev;
|
||||
|
||||
|
||||
@@ -4,6 +4,7 @@
|
||||
|
||||
#include <linux/compiler_attributes.h>
|
||||
#include <linux/types.h>
|
||||
#include <linux/stdarg.h>
|
||||
|
||||
int num_to_str(char *buf, int size, unsigned long long num, unsigned int width);
|
||||
|
||||
|
||||
@@ -1222,8 +1222,9 @@ static int __request_region_locked(struct resource *res, struct resource *parent
|
||||
* become unavailable to other users. Conflicts are
|
||||
* not expected. Warn to aid debugging if encountered.
|
||||
*/
|
||||
if (conflict->desc == IORES_DESC_DEVICE_PRIVATE_MEMORY) {
|
||||
pr_warn("Unaddressable device %s %pR conflicts with %pR",
|
||||
if (parent == &iomem_resource &&
|
||||
conflict->desc == IORES_DESC_DEVICE_PRIVATE_MEMORY) {
|
||||
pr_warn("Unaddressable device %s %pR conflicts with %pR\n",
|
||||
conflict->name, conflict, res);
|
||||
}
|
||||
if (conflict != parent) {
|
||||
|
||||
@@ -385,7 +385,9 @@ static void print_address_description(void *addr, u8 tag,
|
||||
}
|
||||
|
||||
if (is_vmalloc_addr(addr)) {
|
||||
pr_err("The buggy address %px belongs to a vmalloc virtual mapping\n", addr);
|
||||
pr_err("The buggy address belongs to a");
|
||||
if (!vmalloc_dump_obj(addr))
|
||||
pr_cont(" vmalloc virtual mapping\n");
|
||||
page = vmalloc_to_page(addr);
|
||||
}
|
||||
|
||||
|
||||
@@ -2387,7 +2387,7 @@ skip:
|
||||
VM_BUG_ON(khugepaged_scan.address < hstart ||
|
||||
khugepaged_scan.address + HPAGE_PMD_SIZE >
|
||||
hend);
|
||||
if (IS_ENABLED(CONFIG_SHMEM) && vma->vm_file) {
|
||||
if (IS_ENABLED(CONFIG_SHMEM) && !vma_is_anonymous(vma)) {
|
||||
struct file *file = get_file(vma->vm_file);
|
||||
pgoff_t pgoff = linear_page_index(vma,
|
||||
khugepaged_scan.address);
|
||||
@@ -2734,7 +2734,7 @@ int madvise_collapse(struct vm_area_struct *vma, struct vm_area_struct **prev,
|
||||
mmap_assert_locked(mm);
|
||||
memset(cc->node_load, 0, sizeof(cc->node_load));
|
||||
nodes_clear(cc->alloc_nmask);
|
||||
if (IS_ENABLED(CONFIG_SHMEM) && vma->vm_file) {
|
||||
if (IS_ENABLED(CONFIG_SHMEM) && !vma_is_anonymous(vma)) {
|
||||
struct file *file = get_file(vma->vm_file);
|
||||
pgoff_t pgoff = linear_page_index(vma, addr);
|
||||
|
||||
|
||||
@@ -988,6 +988,9 @@ static struct zspage *alloc_zspage(struct zs_pool *pool,
|
||||
if (!zspage)
|
||||
return NULL;
|
||||
|
||||
if (!IS_ENABLED(CONFIG_COMPACTION))
|
||||
gfp &= ~__GFP_MOVABLE;
|
||||
|
||||
zspage->magic = ZSPAGE_MAGIC;
|
||||
migrate_lock_init(zspage);
|
||||
|
||||
|
||||
@@ -35,6 +35,7 @@
|
||||
#include <linux/seq_file.h>
|
||||
#include <linux/export.h>
|
||||
#include <linux/etherdevice.h>
|
||||
#include <linux/refcount.h>
|
||||
|
||||
int sysctl_aarp_expiry_time = AARP_EXPIRY_TIME;
|
||||
int sysctl_aarp_tick_time = AARP_TICK_TIME;
|
||||
@@ -44,6 +45,7 @@ int sysctl_aarp_resolve_time = AARP_RESOLVE_TIME;
|
||||
/* Lists of aarp entries */
|
||||
/**
|
||||
* struct aarp_entry - AARP entry
|
||||
* @refcnt: Reference count
|
||||
* @last_sent: Last time we xmitted the aarp request
|
||||
* @packet_queue: Queue of frames wait for resolution
|
||||
* @status: Used for proxy AARP
|
||||
@@ -55,6 +57,7 @@ int sysctl_aarp_resolve_time = AARP_RESOLVE_TIME;
|
||||
* @next: Next entry in chain
|
||||
*/
|
||||
struct aarp_entry {
|
||||
refcount_t refcnt;
|
||||
/* These first two are only used for unresolved entries */
|
||||
unsigned long last_sent;
|
||||
struct sk_buff_head packet_queue;
|
||||
@@ -79,6 +82,17 @@ static DEFINE_RWLOCK(aarp_lock);
|
||||
/* Used to walk the list and purge/kick entries. */
|
||||
static struct timer_list aarp_timer;
|
||||
|
||||
static inline void aarp_entry_get(struct aarp_entry *a)
|
||||
{
|
||||
refcount_inc(&a->refcnt);
|
||||
}
|
||||
|
||||
static inline void aarp_entry_put(struct aarp_entry *a)
|
||||
{
|
||||
if (refcount_dec_and_test(&a->refcnt))
|
||||
kfree(a);
|
||||
}
|
||||
|
||||
/*
|
||||
* Delete an aarp queue
|
||||
*
|
||||
@@ -87,7 +101,7 @@ static struct timer_list aarp_timer;
|
||||
static void __aarp_expire(struct aarp_entry *a)
|
||||
{
|
||||
skb_queue_purge(&a->packet_queue);
|
||||
kfree(a);
|
||||
aarp_entry_put(a);
|
||||
}
|
||||
|
||||
/*
|
||||
@@ -380,9 +394,11 @@ static void aarp_purge(void)
|
||||
static struct aarp_entry *aarp_alloc(void)
|
||||
{
|
||||
struct aarp_entry *a = kmalloc(sizeof(*a), GFP_ATOMIC);
|
||||
if (!a)
|
||||
return NULL;
|
||||
|
||||
if (a)
|
||||
skb_queue_head_init(&a->packet_queue);
|
||||
refcount_set(&a->refcnt, 1);
|
||||
skb_queue_head_init(&a->packet_queue);
|
||||
return a;
|
||||
}
|
||||
|
||||
@@ -508,6 +524,7 @@ int aarp_proxy_probe_network(struct atalk_iface *atif, struct atalk_addr *sa)
|
||||
entry->dev = atif->dev;
|
||||
|
||||
write_lock_bh(&aarp_lock);
|
||||
aarp_entry_get(entry);
|
||||
|
||||
hash = sa->s_node % (AARP_HASH_SIZE - 1);
|
||||
entry->next = proxies[hash];
|
||||
@@ -533,6 +550,7 @@ int aarp_proxy_probe_network(struct atalk_iface *atif, struct atalk_addr *sa)
|
||||
retval = 1;
|
||||
}
|
||||
|
||||
aarp_entry_put(entry);
|
||||
write_unlock_bh(&aarp_lock);
|
||||
out:
|
||||
return retval;
|
||||
|
||||
@@ -979,8 +979,9 @@ static bool check_fully_established(struct mptcp_sock *msk, struct sock *ssk,
|
||||
if (subflow->mp_join)
|
||||
goto reset;
|
||||
subflow->mp_capable = 0;
|
||||
if (!mptcp_try_fallback(ssk))
|
||||
goto reset;
|
||||
pr_fallback(msk);
|
||||
mptcp_do_fallback(ssk);
|
||||
return false;
|
||||
}
|
||||
|
||||
|
||||
@@ -304,8 +304,14 @@ void mptcp_pm_mp_fail_received(struct sock *sk, u64 fail_seq)
|
||||
|
||||
pr_debug("fail_seq=%llu\n", fail_seq);
|
||||
|
||||
if (!READ_ONCE(msk->allow_infinite_fallback))
|
||||
/* After accepting the fail, we can't create any other subflows */
|
||||
spin_lock_bh(&msk->fallback_lock);
|
||||
if (!msk->allow_infinite_fallback) {
|
||||
spin_unlock_bh(&msk->fallback_lock);
|
||||
return;
|
||||
}
|
||||
msk->allow_subflows = false;
|
||||
spin_unlock_bh(&msk->fallback_lock);
|
||||
|
||||
if (!subflow->fail_tout) {
|
||||
pr_debug("send MP_FAIL response and infinite map\n");
|
||||
|
||||
@@ -623,10 +623,9 @@ static bool mptcp_check_data_fin(struct sock *sk)
|
||||
|
||||
static void mptcp_dss_corruption(struct mptcp_sock *msk, struct sock *ssk)
|
||||
{
|
||||
if (READ_ONCE(msk->allow_infinite_fallback)) {
|
||||
if (mptcp_try_fallback(ssk)) {
|
||||
MPTCP_INC_STATS(sock_net(ssk),
|
||||
MPTCP_MIB_DSSCORRUPTIONFALLBACK);
|
||||
mptcp_do_fallback(ssk);
|
||||
} else {
|
||||
MPTCP_INC_STATS(sock_net(ssk), MPTCP_MIB_DSSCORRUPTIONRESET);
|
||||
mptcp_subflow_reset(ssk);
|
||||
@@ -876,7 +875,7 @@ void mptcp_data_ready(struct sock *sk, struct sock *ssk)
|
||||
static void mptcp_subflow_joined(struct mptcp_sock *msk, struct sock *ssk)
|
||||
{
|
||||
mptcp_subflow_ctx(ssk)->map_seq = READ_ONCE(msk->ack_seq);
|
||||
WRITE_ONCE(msk->allow_infinite_fallback, false);
|
||||
msk->allow_infinite_fallback = false;
|
||||
mptcp_event(MPTCP_EVENT_SUB_ESTABLISHED, msk, ssk, GFP_ATOMIC);
|
||||
}
|
||||
|
||||
@@ -887,6 +886,14 @@ static bool __mptcp_finish_join(struct mptcp_sock *msk, struct sock *ssk)
|
||||
if (sk->sk_state != TCP_ESTABLISHED)
|
||||
return false;
|
||||
|
||||
spin_lock_bh(&msk->fallback_lock);
|
||||
if (!msk->allow_subflows) {
|
||||
spin_unlock_bh(&msk->fallback_lock);
|
||||
return false;
|
||||
}
|
||||
mptcp_subflow_joined(msk, ssk);
|
||||
spin_unlock_bh(&msk->fallback_lock);
|
||||
|
||||
/* attach to msk socket only after we are sure we will deal with it
|
||||
* at close time
|
||||
*/
|
||||
@@ -895,7 +902,6 @@ static bool __mptcp_finish_join(struct mptcp_sock *msk, struct sock *ssk)
|
||||
|
||||
mptcp_subflow_ctx(ssk)->subflow_id = msk->subflow_id++;
|
||||
mptcp_sockopt_sync_locked(msk, ssk);
|
||||
mptcp_subflow_joined(msk, ssk);
|
||||
mptcp_stop_tout_timer(sk);
|
||||
__mptcp_propagate_sndbuf(sk, ssk);
|
||||
return true;
|
||||
@@ -1231,10 +1237,14 @@ static void mptcp_update_infinite_map(struct mptcp_sock *msk,
|
||||
mpext->infinite_map = 1;
|
||||
mpext->data_len = 0;
|
||||
|
||||
if (!mptcp_try_fallback(ssk)) {
|
||||
mptcp_subflow_reset(ssk);
|
||||
return;
|
||||
}
|
||||
|
||||
MPTCP_INC_STATS(sock_net(ssk), MPTCP_MIB_INFINITEMAPTX);
|
||||
mptcp_subflow_ctx(ssk)->send_infinite_map = 0;
|
||||
pr_fallback(msk);
|
||||
mptcp_do_fallback(ssk);
|
||||
}
|
||||
|
||||
#define MPTCP_MAX_GSO_SIZE (GSO_LEGACY_MAX_SIZE - (MAX_TCP_HEADER + 1))
|
||||
@@ -2606,9 +2616,9 @@ static void mptcp_check_fastclose(struct mptcp_sock *msk)
|
||||
|
||||
static void __mptcp_retrans(struct sock *sk)
|
||||
{
|
||||
struct mptcp_sendmsg_info info = { .data_lock_held = true, };
|
||||
struct mptcp_sock *msk = mptcp_sk(sk);
|
||||
struct mptcp_subflow_context *subflow;
|
||||
struct mptcp_sendmsg_info info = {};
|
||||
struct mptcp_data_frag *dfrag;
|
||||
struct sock *ssk;
|
||||
int ret, err;
|
||||
@@ -2653,6 +2663,18 @@ static void __mptcp_retrans(struct sock *sk)
|
||||
info.sent = 0;
|
||||
info.limit = READ_ONCE(msk->csum_enabled) ? dfrag->data_len :
|
||||
dfrag->already_sent;
|
||||
|
||||
/*
|
||||
* make the whole retrans decision, xmit, disallow
|
||||
* fallback atomic
|
||||
*/
|
||||
spin_lock_bh(&msk->fallback_lock);
|
||||
if (__mptcp_check_fallback(msk)) {
|
||||
spin_unlock_bh(&msk->fallback_lock);
|
||||
release_sock(ssk);
|
||||
return;
|
||||
}
|
||||
|
||||
while (info.sent < info.limit) {
|
||||
ret = mptcp_sendmsg_frag(sk, ssk, dfrag, &info);
|
||||
if (ret <= 0)
|
||||
@@ -2666,8 +2688,9 @@ static void __mptcp_retrans(struct sock *sk)
|
||||
len = max(copied, len);
|
||||
tcp_push(ssk, 0, info.mss_now, tcp_sk(ssk)->nonagle,
|
||||
info.size_goal);
|
||||
WRITE_ONCE(msk->allow_infinite_fallback, false);
|
||||
msk->allow_infinite_fallback = false;
|
||||
}
|
||||
spin_unlock_bh(&msk->fallback_lock);
|
||||
|
||||
release_sock(ssk);
|
||||
}
|
||||
@@ -2796,11 +2819,13 @@ static void __mptcp_init_sock(struct sock *sk)
|
||||
WRITE_ONCE(msk->first, NULL);
|
||||
inet_csk(sk)->icsk_sync_mss = mptcp_sync_mss;
|
||||
WRITE_ONCE(msk->csum_enabled, mptcp_is_checksum_enabled(sock_net(sk)));
|
||||
WRITE_ONCE(msk->allow_infinite_fallback, true);
|
||||
msk->allow_infinite_fallback = true;
|
||||
msk->allow_subflows = true;
|
||||
msk->recovery = false;
|
||||
msk->subflow_id = 1;
|
||||
|
||||
mptcp_pm_data_init(msk);
|
||||
spin_lock_init(&msk->fallback_lock);
|
||||
|
||||
/* re-use the csk retrans timer for MPTCP-level retrans */
|
||||
timer_setup(&msk->sk.icsk_retransmit_timer, mptcp_retransmit_timer, 0);
|
||||
@@ -3183,7 +3208,16 @@ static int mptcp_disconnect(struct sock *sk, int flags)
|
||||
* subflow
|
||||
*/
|
||||
mptcp_destroy_common(msk, MPTCP_CF_FASTCLOSE);
|
||||
|
||||
/* The first subflow is already in TCP_CLOSE status, the following
|
||||
* can't overlap with a fallback anymore
|
||||
*/
|
||||
spin_lock_bh(&msk->fallback_lock);
|
||||
msk->allow_subflows = true;
|
||||
msk->allow_infinite_fallback = true;
|
||||
WRITE_ONCE(msk->flags, 0);
|
||||
spin_unlock_bh(&msk->fallback_lock);
|
||||
|
||||
msk->cb_flags = 0;
|
||||
msk->recovery = false;
|
||||
msk->can_ack = false;
|
||||
@@ -3599,7 +3633,13 @@ bool mptcp_finish_join(struct sock *ssk)
|
||||
|
||||
/* active subflow, already present inside the conn_list */
|
||||
if (!list_empty(&subflow->node)) {
|
||||
spin_lock_bh(&msk->fallback_lock);
|
||||
if (!msk->allow_subflows) {
|
||||
spin_unlock_bh(&msk->fallback_lock);
|
||||
return false;
|
||||
}
|
||||
mptcp_subflow_joined(msk, ssk);
|
||||
spin_unlock_bh(&msk->fallback_lock);
|
||||
mptcp_propagate_sndbuf(parent, ssk);
|
||||
return true;
|
||||
}
|
||||
@@ -3712,7 +3752,7 @@ static void mptcp_subflow_early_fallback(struct mptcp_sock *msk,
|
||||
struct mptcp_subflow_context *subflow)
|
||||
{
|
||||
subflow->request_mptcp = 0;
|
||||
__mptcp_do_fallback(msk);
|
||||
WARN_ON_ONCE(!__mptcp_try_fallback(msk));
|
||||
}
|
||||
|
||||
static int mptcp_connect(struct sock *sk, struct sockaddr *uaddr, int addr_len)
|
||||
|
||||
@@ -330,10 +330,16 @@ struct mptcp_sock {
|
||||
u64 rtt_us; /* last maximum rtt of subflows */
|
||||
} rcvq_space;
|
||||
u8 scaling_ratio;
|
||||
bool allow_subflows;
|
||||
|
||||
u32 subflow_id;
|
||||
u32 setsockopt_seq;
|
||||
char ca_name[TCP_CA_NAME_MAX];
|
||||
|
||||
spinlock_t fallback_lock; /* protects fallback,
|
||||
* allow_infinite_fallback and
|
||||
* allow_join
|
||||
*/
|
||||
};
|
||||
|
||||
#define mptcp_data_lock(sk) spin_lock_bh(&(sk)->sk_lock.slock)
|
||||
@@ -1097,25 +1103,33 @@ static inline bool mptcp_check_fallback(const struct sock *sk)
|
||||
return __mptcp_check_fallback(msk);
|
||||
}
|
||||
|
||||
static inline void __mptcp_do_fallback(struct mptcp_sock *msk)
|
||||
static inline bool __mptcp_try_fallback(struct mptcp_sock *msk)
|
||||
{
|
||||
if (test_bit(MPTCP_FALLBACK_DONE, &msk->flags)) {
|
||||
pr_debug("TCP fallback already done (msk=%p)\n", msk);
|
||||
return;
|
||||
return true;
|
||||
}
|
||||
if (WARN_ON_ONCE(!READ_ONCE(msk->allow_infinite_fallback)))
|
||||
return;
|
||||
spin_lock_bh(&msk->fallback_lock);
|
||||
if (!msk->allow_infinite_fallback) {
|
||||
spin_unlock_bh(&msk->fallback_lock);
|
||||
return false;
|
||||
}
|
||||
|
||||
msk->allow_subflows = false;
|
||||
set_bit(MPTCP_FALLBACK_DONE, &msk->flags);
|
||||
spin_unlock_bh(&msk->fallback_lock);
|
||||
return true;
|
||||
}
|
||||
|
||||
static inline void mptcp_do_fallback(struct sock *ssk)
|
||||
static inline bool mptcp_try_fallback(struct sock *ssk)
|
||||
{
|
||||
struct mptcp_subflow_context *subflow = mptcp_subflow_ctx(ssk);
|
||||
struct sock *sk = subflow->conn;
|
||||
struct mptcp_sock *msk;
|
||||
|
||||
msk = mptcp_sk(sk);
|
||||
__mptcp_do_fallback(msk);
|
||||
if (!__mptcp_try_fallback(msk))
|
||||
return false;
|
||||
if (READ_ONCE(msk->snd_data_fin_enable) && !(ssk->sk_shutdown & SEND_SHUTDOWN)) {
|
||||
gfp_t saved_allocation = ssk->sk_allocation;
|
||||
|
||||
@@ -1127,6 +1141,7 @@ static inline void mptcp_do_fallback(struct sock *ssk)
|
||||
tcp_shutdown(ssk, SEND_SHUTDOWN);
|
||||
ssk->sk_allocation = saved_allocation;
|
||||
}
|
||||
return true;
|
||||
}
|
||||
|
||||
#define pr_fallback(a) pr_debug("%s:fallback to TCP (msk=%p)\n", __func__, a)
|
||||
|
||||
@@ -524,9 +524,11 @@ static void subflow_finish_connect(struct sock *sk, const struct sk_buff *skb)
|
||||
mptcp_get_options(skb, &mp_opt);
|
||||
if (subflow->request_mptcp) {
|
||||
if (!(mp_opt.suboptions & OPTION_MPTCP_MPC_SYNACK)) {
|
||||
if (!mptcp_try_fallback(sk))
|
||||
goto do_reset;
|
||||
|
||||
MPTCP_INC_STATS(sock_net(sk),
|
||||
MPTCP_MIB_MPCAPABLEACTIVEFALLBACK);
|
||||
mptcp_do_fallback(sk);
|
||||
pr_fallback(msk);
|
||||
goto fallback;
|
||||
}
|
||||
@@ -1255,20 +1257,29 @@ static void subflow_sched_work_if_closed(struct mptcp_sock *msk, struct sock *ss
|
||||
mptcp_schedule_work(sk);
|
||||
}
|
||||
|
||||
static void mptcp_subflow_fail(struct mptcp_sock *msk, struct sock *ssk)
|
||||
static bool mptcp_subflow_fail(struct mptcp_sock *msk, struct sock *ssk)
|
||||
{
|
||||
struct mptcp_subflow_context *subflow = mptcp_subflow_ctx(ssk);
|
||||
unsigned long fail_tout;
|
||||
|
||||
/* we are really failing, prevent any later subflow join */
|
||||
spin_lock_bh(&msk->fallback_lock);
|
||||
if (!msk->allow_infinite_fallback) {
|
||||
spin_unlock_bh(&msk->fallback_lock);
|
||||
return false;
|
||||
}
|
||||
msk->allow_subflows = false;
|
||||
spin_unlock_bh(&msk->fallback_lock);
|
||||
|
||||
/* greceful failure can happen only on the MPC subflow */
|
||||
if (WARN_ON_ONCE(ssk != READ_ONCE(msk->first)))
|
||||
return;
|
||||
return false;
|
||||
|
||||
/* since the close timeout take precedence on the fail one,
|
||||
* no need to start the latter when the first is already set
|
||||
*/
|
||||
if (sock_flag((struct sock *)msk, SOCK_DEAD))
|
||||
return;
|
||||
return true;
|
||||
|
||||
/* we don't need extreme accuracy here, use a zero fail_tout as special
|
||||
* value meaning no fail timeout at all;
|
||||
@@ -1280,6 +1291,7 @@ static void mptcp_subflow_fail(struct mptcp_sock *msk, struct sock *ssk)
|
||||
tcp_send_ack(ssk);
|
||||
|
||||
mptcp_reset_tout_timer(msk, subflow->fail_tout);
|
||||
return true;
|
||||
}
|
||||
|
||||
static bool subflow_check_data_avail(struct sock *ssk)
|
||||
@@ -1340,17 +1352,16 @@ fallback:
|
||||
(subflow->mp_join || subflow->valid_csum_seen)) {
|
||||
subflow->send_mp_fail = 1;
|
||||
|
||||
if (!READ_ONCE(msk->allow_infinite_fallback)) {
|
||||
if (!mptcp_subflow_fail(msk, ssk)) {
|
||||
subflow->reset_transient = 0;
|
||||
subflow->reset_reason = MPTCP_RST_EMIDDLEBOX;
|
||||
goto reset;
|
||||
}
|
||||
mptcp_subflow_fail(msk, ssk);
|
||||
WRITE_ONCE(subflow->data_avail, MPTCP_SUBFLOW_DATA_AVAIL);
|
||||
return true;
|
||||
}
|
||||
|
||||
if (!READ_ONCE(msk->allow_infinite_fallback)) {
|
||||
if (!mptcp_try_fallback(ssk)) {
|
||||
/* fatal protocol error, close the socket.
|
||||
* subflow_error_report() will introduce the appropriate barriers
|
||||
*/
|
||||
@@ -1366,8 +1377,6 @@ reset:
|
||||
WRITE_ONCE(subflow->data_avail, MPTCP_SUBFLOW_NODATA);
|
||||
return false;
|
||||
}
|
||||
|
||||
mptcp_do_fallback(ssk);
|
||||
}
|
||||
|
||||
skb = skb_peek(&ssk->sk_receive_queue);
|
||||
@@ -1612,7 +1621,6 @@ int __mptcp_subflow_connect(struct sock *sk, const struct mptcp_addr_info *loc,
|
||||
/* discard the subflow socket */
|
||||
mptcp_sock_graft(ssk, sk->sk_socket);
|
||||
iput(SOCK_INODE(sf));
|
||||
WRITE_ONCE(msk->allow_infinite_fallback, false);
|
||||
mptcp_stop_tout_timer(sk);
|
||||
return 0;
|
||||
|
||||
@@ -1790,7 +1798,7 @@ static void subflow_state_change(struct sock *sk)
|
||||
|
||||
msk = mptcp_sk(parent);
|
||||
if (subflow_simultaneous_connect(sk)) {
|
||||
mptcp_do_fallback(sk);
|
||||
WARN_ON_ONCE(!mptcp_try_fallback(sk));
|
||||
pr_fallback(msk);
|
||||
subflow->conn_finished = 1;
|
||||
mptcp_propagate_state(parent, sk, subflow, NULL);
|
||||
|
||||
@@ -539,9 +539,6 @@ destroy_class:
|
||||
|
||||
static void qfq_destroy_class(struct Qdisc *sch, struct qfq_class *cl)
|
||||
{
|
||||
struct qfq_sched *q = qdisc_priv(sch);
|
||||
|
||||
qfq_rm_from_agg(q, cl);
|
||||
gen_kill_estimator(&cl->rate_est);
|
||||
qdisc_put(cl->qdisc);
|
||||
kfree(cl);
|
||||
@@ -562,10 +559,11 @@ static int qfq_delete_class(struct Qdisc *sch, unsigned long arg,
|
||||
|
||||
qdisc_purge_queue(cl->qdisc);
|
||||
qdisc_class_hash_remove(&q->clhash, &cl->common);
|
||||
qfq_destroy_class(sch, cl);
|
||||
qfq_rm_from_agg(q, cl);
|
||||
|
||||
sch_tree_unlock(sch);
|
||||
|
||||
qfq_destroy_class(sch, cl);
|
||||
return 0;
|
||||
}
|
||||
|
||||
@@ -1506,6 +1504,7 @@ static void qfq_destroy_qdisc(struct Qdisc *sch)
|
||||
for (i = 0; i < q->clhash.hashsize; i++) {
|
||||
hlist_for_each_entry_safe(cl, next, &q->clhash.hash[i],
|
||||
common.hnode) {
|
||||
qfq_rm_from_agg(q, cl);
|
||||
qfq_destroy_class(sch, cl);
|
||||
}
|
||||
}
|
||||
|
||||
@@ -880,7 +880,7 @@ static int xfrmi_changelink(struct net_device *dev, struct nlattr *tb[],
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
if (p.collect_md) {
|
||||
if (p.collect_md || xi->p.collect_md) {
|
||||
NL_SET_ERR_MSG(extack, "collect_md can't be changed");
|
||||
return -EINVAL;
|
||||
}
|
||||
@@ -891,11 +891,6 @@ static int xfrmi_changelink(struct net_device *dev, struct nlattr *tb[],
|
||||
} else {
|
||||
if (xi->dev != dev)
|
||||
return -EEXIST;
|
||||
if (xi->p.collect_md) {
|
||||
NL_SET_ERR_MSG(extack,
|
||||
"device can't be changed to collect_md");
|
||||
return -EINVAL;
|
||||
}
|
||||
}
|
||||
|
||||
return xfrmi_update(xi, &p);
|
||||
|
||||
@@ -72,6 +72,10 @@
|
||||
struct hda_tegra_soc {
|
||||
bool has_hda2codec_2x_reset;
|
||||
bool has_hda2hdmi;
|
||||
bool has_hda2codec_2x;
|
||||
bool input_stream;
|
||||
bool always_on;
|
||||
bool requires_init;
|
||||
};
|
||||
|
||||
struct hda_tegra {
|
||||
@@ -187,7 +191,9 @@ static int __maybe_unused hda_tegra_runtime_resume(struct device *dev)
|
||||
if (rc != 0)
|
||||
return rc;
|
||||
if (chip->running) {
|
||||
hda_tegra_init(hda);
|
||||
if (hda->soc->requires_init)
|
||||
hda_tegra_init(hda);
|
||||
|
||||
azx_init_chip(chip, 1);
|
||||
/* disable controller wake up event*/
|
||||
azx_writew(chip, WAKEEN, azx_readw(chip, WAKEEN) &
|
||||
@@ -252,7 +258,8 @@ static int hda_tegra_init_chip(struct azx *chip, struct platform_device *pdev)
|
||||
bus->remap_addr = hda->regs + HDA_BAR0;
|
||||
bus->addr = res->start + HDA_BAR0;
|
||||
|
||||
hda_tegra_init(hda);
|
||||
if (hda->soc->requires_init)
|
||||
hda_tegra_init(hda);
|
||||
|
||||
return 0;
|
||||
}
|
||||
@@ -325,7 +332,7 @@ static int hda_tegra_first_init(struct azx *chip, struct platform_device *pdev)
|
||||
* starts with offset 0 which is wrong as HW register for output stream
|
||||
* offset starts with 4.
|
||||
*/
|
||||
if (of_device_is_compatible(np, "nvidia,tegra234-hda"))
|
||||
if (!hda->soc->input_stream)
|
||||
chip->capture_streams = 4;
|
||||
|
||||
chip->playback_streams = (gcap >> 12) & 0x0f;
|
||||
@@ -421,7 +428,6 @@ static int hda_tegra_create(struct snd_card *card,
|
||||
chip->driver_caps = driver_caps;
|
||||
chip->driver_type = driver_caps & 0xff;
|
||||
chip->dev_index = 0;
|
||||
chip->jackpoll_interval = msecs_to_jiffies(5000);
|
||||
INIT_LIST_HEAD(&chip->pcm_list);
|
||||
|
||||
chip->codec_probe_mask = -1;
|
||||
@@ -438,7 +444,16 @@ static int hda_tegra_create(struct snd_card *card,
|
||||
chip->bus.core.sync_write = 0;
|
||||
chip->bus.core.needs_damn_long_delay = 1;
|
||||
chip->bus.core.aligned_mmio = 1;
|
||||
chip->bus.jackpoll_in_suspend = 1;
|
||||
|
||||
/*
|
||||
* HDA power domain and clocks are always on for Tegra264 and
|
||||
* the jack detection logic would work always, so no need of
|
||||
* jack polling mechanism running.
|
||||
*/
|
||||
if (!hda->soc->always_on) {
|
||||
chip->jackpoll_interval = msecs_to_jiffies(5000);
|
||||
chip->bus.jackpoll_in_suspend = 1;
|
||||
}
|
||||
|
||||
err = snd_device_new(card, SNDRV_DEV_LOWLEVEL, chip, &ops);
|
||||
if (err < 0) {
|
||||
@@ -452,22 +467,44 @@ static int hda_tegra_create(struct snd_card *card,
|
||||
static const struct hda_tegra_soc tegra30_data = {
|
||||
.has_hda2codec_2x_reset = true,
|
||||
.has_hda2hdmi = true,
|
||||
.has_hda2codec_2x = true,
|
||||
.input_stream = true,
|
||||
.always_on = false,
|
||||
.requires_init = true,
|
||||
};
|
||||
|
||||
static const struct hda_tegra_soc tegra194_data = {
|
||||
.has_hda2codec_2x_reset = false,
|
||||
.has_hda2hdmi = true,
|
||||
.has_hda2codec_2x = true,
|
||||
.input_stream = true,
|
||||
.always_on = false,
|
||||
.requires_init = true,
|
||||
};
|
||||
|
||||
static const struct hda_tegra_soc tegra234_data = {
|
||||
.has_hda2codec_2x_reset = true,
|
||||
.has_hda2hdmi = false,
|
||||
.has_hda2codec_2x = true,
|
||||
.input_stream = false,
|
||||
.always_on = false,
|
||||
.requires_init = true,
|
||||
};
|
||||
|
||||
static const struct hda_tegra_soc tegra264_data = {
|
||||
.has_hda2codec_2x_reset = true,
|
||||
.has_hda2hdmi = false,
|
||||
.has_hda2codec_2x = false,
|
||||
.input_stream = false,
|
||||
.always_on = true,
|
||||
.requires_init = false,
|
||||
};
|
||||
|
||||
static const struct of_device_id hda_tegra_match[] = {
|
||||
{ .compatible = "nvidia,tegra30-hda", .data = &tegra30_data },
|
||||
{ .compatible = "nvidia,tegra194-hda", .data = &tegra194_data },
|
||||
{ .compatible = "nvidia,tegra234-hda", .data = &tegra234_data },
|
||||
{ .compatible = "nvidia,tegra264-hda", .data = &tegra264_data },
|
||||
{},
|
||||
};
|
||||
MODULE_DEVICE_TABLE(of, hda_tegra_match);
|
||||
@@ -522,7 +559,9 @@ static int hda_tegra_probe(struct platform_device *pdev)
|
||||
hda->clocks[hda->nclocks++].id = "hda";
|
||||
if (hda->soc->has_hda2hdmi)
|
||||
hda->clocks[hda->nclocks++].id = "hda2hdmi";
|
||||
hda->clocks[hda->nclocks++].id = "hda2codec_2x";
|
||||
|
||||
if (hda->soc->has_hda2codec_2x)
|
||||
hda->clocks[hda->nclocks++].id = "hda2codec_2x";
|
||||
|
||||
err = devm_clk_bulk_get(&pdev->dev, hda->nclocks, hda->clocks);
|
||||
if (err < 0)
|
||||
|
||||
@@ -4559,6 +4559,9 @@ HDA_CODEC_ENTRY(0x10de002e, "Tegra186 HDMI/DP1", patch_tegra_hdmi),
|
||||
HDA_CODEC_ENTRY(0x10de002f, "Tegra194 HDMI/DP2", patch_tegra_hdmi),
|
||||
HDA_CODEC_ENTRY(0x10de0030, "Tegra194 HDMI/DP3", patch_tegra_hdmi),
|
||||
HDA_CODEC_ENTRY(0x10de0031, "Tegra234 HDMI/DP", patch_tegra234_hdmi),
|
||||
HDA_CODEC_ENTRY(0x10de0033, "SoC 33 HDMI/DP", patch_tegra234_hdmi),
|
||||
HDA_CODEC_ENTRY(0x10de0034, "Tegra264 HDMI/DP", patch_tegra234_hdmi),
|
||||
HDA_CODEC_ENTRY(0x10de0035, "SoC 35 HDMI/DP", patch_tegra234_hdmi),
|
||||
HDA_CODEC_ENTRY(0x10de0040, "GPU 40 HDMI/DP", patch_nvhdmi),
|
||||
HDA_CODEC_ENTRY(0x10de0041, "GPU 41 HDMI/DP", patch_nvhdmi),
|
||||
HDA_CODEC_ENTRY(0x10de0042, "GPU 42 HDMI/DP", patch_nvhdmi),
|
||||
@@ -4597,15 +4600,32 @@ HDA_CODEC_ENTRY(0x10de0097, "GPU 97 HDMI/DP", patch_nvhdmi),
|
||||
HDA_CODEC_ENTRY(0x10de0098, "GPU 98 HDMI/DP", patch_nvhdmi),
|
||||
HDA_CODEC_ENTRY(0x10de0099, "GPU 99 HDMI/DP", patch_nvhdmi),
|
||||
HDA_CODEC_ENTRY(0x10de009a, "GPU 9a HDMI/DP", patch_nvhdmi),
|
||||
HDA_CODEC_ENTRY(0x10de009b, "GPU 9b HDMI/DP", patch_nvhdmi),
|
||||
HDA_CODEC_ENTRY(0x10de009c, "GPU 9c HDMI/DP", patch_nvhdmi),
|
||||
HDA_CODEC_ENTRY(0x10de009d, "GPU 9d HDMI/DP", patch_nvhdmi),
|
||||
HDA_CODEC_ENTRY(0x10de009e, "GPU 9e HDMI/DP", patch_nvhdmi),
|
||||
HDA_CODEC_ENTRY(0x10de009f, "GPU 9f HDMI/DP", patch_nvhdmi),
|
||||
HDA_CODEC_ENTRY(0x10de00a0, "GPU a0 HDMI/DP", patch_nvhdmi),
|
||||
HDA_CODEC_ENTRY(0x10de00a1, "GPU a1 HDMI/DP", patch_nvhdmi),
|
||||
HDA_CODEC_ENTRY(0x10de00a3, "GPU a3 HDMI/DP", patch_nvhdmi),
|
||||
HDA_CODEC_ENTRY(0x10de00a4, "GPU a4 HDMI/DP", patch_nvhdmi),
|
||||
HDA_CODEC_ENTRY(0x10de00a5, "GPU a5 HDMI/DP", patch_nvhdmi),
|
||||
HDA_CODEC_ENTRY(0x10de00a6, "GPU a6 HDMI/DP", patch_nvhdmi),
|
||||
HDA_CODEC_ENTRY(0x10de00a7, "GPU a7 HDMI/DP", patch_nvhdmi),
|
||||
HDA_CODEC_ENTRY(0x10de00a8, "GPU a8 HDMI/DP", patch_nvhdmi),
|
||||
HDA_CODEC_ENTRY(0x10de00a9, "GPU a9 HDMI/DP", patch_nvhdmi),
|
||||
HDA_CODEC_ENTRY(0x10de00aa, "GPU aa HDMI/DP", patch_nvhdmi),
|
||||
HDA_CODEC_ENTRY(0x10de00ab, "GPU ab HDMI/DP", patch_nvhdmi),
|
||||
HDA_CODEC_ENTRY(0x10de00ad, "GPU ad HDMI/DP", patch_nvhdmi),
|
||||
HDA_CODEC_ENTRY(0x10de00ae, "GPU ae HDMI/DP", patch_nvhdmi),
|
||||
HDA_CODEC_ENTRY(0x10de00af, "GPU af HDMI/DP", patch_nvhdmi),
|
||||
HDA_CODEC_ENTRY(0x10de00b0, "GPU b0 HDMI/DP", patch_nvhdmi),
|
||||
HDA_CODEC_ENTRY(0x10de00b1, "GPU b1 HDMI/DP", patch_nvhdmi),
|
||||
HDA_CODEC_ENTRY(0x10de00c0, "GPU c0 HDMI/DP", patch_nvhdmi),
|
||||
HDA_CODEC_ENTRY(0x10de00c1, "GPU c1 HDMI/DP", patch_nvhdmi),
|
||||
HDA_CODEC_ENTRY(0x10de00c3, "GPU c3 HDMI/DP", patch_nvhdmi),
|
||||
HDA_CODEC_ENTRY(0x10de00c4, "GPU c4 HDMI/DP", patch_nvhdmi),
|
||||
HDA_CODEC_ENTRY(0x10de00c5, "GPU c5 HDMI/DP", patch_nvhdmi),
|
||||
HDA_CODEC_ENTRY(0x10de8001, "MCP73 HDMI", patch_nvhdmi_2ch),
|
||||
HDA_CODEC_ENTRY(0x10de8067, "MCP67/68 HDMI", patch_nvhdmi_2ch),
|
||||
HDA_CODEC_ENTRY(0x67663d82, "Arise 82 HDMI/DP", patch_gf_hdmi),
|
||||
|
||||
@@ -10103,6 +10103,7 @@ static const struct hda_quirk alc269_fixup_tbl[] = {
|
||||
SND_PCI_QUIRK(0x103c, 0x8788, "HP OMEN 15", ALC285_FIXUP_HP_MUTE_LED),
|
||||
SND_PCI_QUIRK(0x103c, 0x87b7, "HP Laptop 14-fq0xxx", ALC236_FIXUP_HP_MUTE_LED_COEFBIT2),
|
||||
SND_PCI_QUIRK(0x103c, 0x87c8, "HP", ALC287_FIXUP_HP_GPIO_LED),
|
||||
SND_PCI_QUIRK(0x103c, 0x87cc, "HP Pavilion 15-eg0xxx", ALC287_FIXUP_HP_GPIO_LED),
|
||||
SND_PCI_QUIRK(0x103c, 0x87d3, "HP Laptop 15-gw0xxx", ALC236_FIXUP_HP_MUTE_LED_COEFBIT2),
|
||||
SND_PCI_QUIRK(0x103c, 0x87df, "HP ProBook 430 G8 Notebook PC", ALC236_FIXUP_HP_GPIO_LED),
|
||||
SND_PCI_QUIRK(0x103c, 0x87e5, "HP ProBook 440 G8 Notebook PC", ALC236_FIXUP_HP_GPIO_LED),
|
||||
|
||||
@@ -12,7 +12,6 @@
|
||||
#include <sys/wait.h>
|
||||
#include <sys/mount.h>
|
||||
#include <fcntl.h>
|
||||
#include "network_helpers.h"
|
||||
|
||||
#define STACK_SIZE (1024 * 1024)
|
||||
static char child_stack[STACK_SIZE];
|
||||
@@ -75,50 +74,6 @@ cleanup:
|
||||
return ret;
|
||||
}
|
||||
|
||||
static int test_current_pid_tgid_cgrp(void *args)
|
||||
{
|
||||
struct test_ns_current_pid_tgid__bss *bss;
|
||||
struct test_ns_current_pid_tgid *skel;
|
||||
int server_fd = -1, ret = -1, err;
|
||||
int cgroup_fd = *(int *)args;
|
||||
pid_t tgid, pid;
|
||||
|
||||
skel = test_ns_current_pid_tgid__open();
|
||||
if (!ASSERT_OK_PTR(skel, "test_ns_current_pid_tgid__open"))
|
||||
return ret;
|
||||
|
||||
bpf_program__set_autoload(skel->progs.cgroup_bind4, true);
|
||||
|
||||
err = test_ns_current_pid_tgid__load(skel);
|
||||
if (!ASSERT_OK(err, "test_ns_current_pid_tgid__load"))
|
||||
goto cleanup;
|
||||
|
||||
bss = skel->bss;
|
||||
if (get_pid_tgid(&pid, &tgid, bss))
|
||||
goto cleanup;
|
||||
|
||||
skel->links.cgroup_bind4 = bpf_program__attach_cgroup(
|
||||
skel->progs.cgroup_bind4, cgroup_fd);
|
||||
if (!ASSERT_OK_PTR(skel->links.cgroup_bind4, "bpf_program__attach_cgroup"))
|
||||
goto cleanup;
|
||||
|
||||
server_fd = start_server(AF_INET, SOCK_STREAM, NULL, 0, 0);
|
||||
if (!ASSERT_GE(server_fd, 0, "start_server"))
|
||||
goto cleanup;
|
||||
|
||||
if (!ASSERT_EQ(bss->user_pid, pid, "pid"))
|
||||
goto cleanup;
|
||||
if (!ASSERT_EQ(bss->user_tgid, tgid, "tgid"))
|
||||
goto cleanup;
|
||||
ret = 0;
|
||||
|
||||
cleanup:
|
||||
if (server_fd >= 0)
|
||||
close(server_fd);
|
||||
test_ns_current_pid_tgid__destroy(skel);
|
||||
return ret;
|
||||
}
|
||||
|
||||
static void test_ns_current_pid_tgid_new_ns(int (*fn)(void *), void *arg)
|
||||
{
|
||||
int wstatus;
|
||||
@@ -140,25 +95,6 @@ static void test_ns_current_pid_tgid_new_ns(int (*fn)(void *), void *arg)
|
||||
return;
|
||||
}
|
||||
|
||||
static void test_in_netns(int (*fn)(void *), void *arg)
|
||||
{
|
||||
struct nstoken *nstoken = NULL;
|
||||
|
||||
SYS(cleanup, "ip netns add ns_current_pid_tgid");
|
||||
SYS(cleanup, "ip -net ns_current_pid_tgid link set dev lo up");
|
||||
|
||||
nstoken = open_netns("ns_current_pid_tgid");
|
||||
if (!ASSERT_OK_PTR(nstoken, "open_netns"))
|
||||
goto cleanup;
|
||||
|
||||
test_ns_current_pid_tgid_new_ns(fn, arg);
|
||||
|
||||
cleanup:
|
||||
if (nstoken)
|
||||
close_netns(nstoken);
|
||||
SYS_NOFAIL("ip netns del ns_current_pid_tgid");
|
||||
}
|
||||
|
||||
/* TODO: use a different tracepoint */
|
||||
void serial_test_ns_current_pid_tgid(void)
|
||||
{
|
||||
@@ -166,13 +102,4 @@ void serial_test_ns_current_pid_tgid(void)
|
||||
test_current_pid_tgid_tp(NULL);
|
||||
if (test__start_subtest("new_ns_tp"))
|
||||
test_ns_current_pid_tgid_new_ns(test_current_pid_tgid_tp, NULL);
|
||||
if (test__start_subtest("new_ns_cgrp")) {
|
||||
int cgroup_fd = -1;
|
||||
|
||||
cgroup_fd = test__join_cgroup("/sock_addr");
|
||||
if (ASSERT_GE(cgroup_fd, 0, "join_cgroup")) {
|
||||
test_in_netns(test_current_pid_tgid_cgrp, &cgroup_fd);
|
||||
close(cgroup_fd);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -28,11 +28,4 @@ int tp_handler(const void *ctx)
|
||||
return 0;
|
||||
}
|
||||
|
||||
SEC("?cgroup/bind4")
|
||||
int cgroup_bind4(struct bpf_sock_addr *ctx)
|
||||
{
|
||||
get_pid_tgid();
|
||||
return 1;
|
||||
}
|
||||
|
||||
char _license[] SEC("license") = "GPL";
|
||||
|
||||
@@ -4,7 +4,8 @@ top_srcdir = ../../../../..
|
||||
|
||||
CFLAGS = -Wall -Wl,--no-as-needed -O2 -g -I$(top_srcdir)/usr/include $(KHDR_INCLUDES)
|
||||
|
||||
TEST_PROGS := mptcp_connect.sh pm_netlink.sh mptcp_join.sh diag.sh \
|
||||
TEST_PROGS := mptcp_connect.sh mptcp_connect_mmap.sh mptcp_connect_sendfile.sh \
|
||||
mptcp_connect_checksum.sh pm_netlink.sh mptcp_join.sh diag.sh \
|
||||
simult_flows.sh mptcp_sockopt.sh userspace_pm.sh
|
||||
|
||||
TEST_GEN_FILES = mptcp_connect pm_nl_ctl mptcp_sockopt mptcp_inq
|
||||
|
||||
@@ -0,0 +1,5 @@
|
||||
#!/bin/bash
|
||||
# SPDX-License-Identifier: GPL-2.0
|
||||
|
||||
MPTCP_LIB_KSFT_TEST="$(basename "${0}" .sh)" \
|
||||
"$(dirname "${0}")/mptcp_connect.sh" -C "${@}"
|
||||
5
tools/testing/selftests/net/mptcp/mptcp_connect_mmap.sh
Normal file
5
tools/testing/selftests/net/mptcp/mptcp_connect_mmap.sh
Normal file
@@ -0,0 +1,5 @@
|
||||
#!/bin/bash
|
||||
# SPDX-License-Identifier: GPL-2.0
|
||||
|
||||
MPTCP_LIB_KSFT_TEST="$(basename "${0}" .sh)" \
|
||||
"$(dirname "${0}")/mptcp_connect.sh" -m mmap "${@}"
|
||||
@@ -0,0 +1,5 @@
|
||||
#!/bin/bash
|
||||
# SPDX-License-Identifier: GPL-2.0
|
||||
|
||||
MPTCP_LIB_KSFT_TEST="$(basename "${0}" .sh)" \
|
||||
"$(dirname "${0}")/mptcp_connect.sh" -m sendfile "${@}"
|
||||
Reference in New Issue
Block a user