0074f618ea
* aosp/android-4.19-stable:
ANDROID: GKI: fix crc issue with commit 6281beee5b
("block: don't merge across cgroup boundaries if blkcg is enabled")
Revert "PCI: Reduce warnings on possible RW1C corruption"
Linux 4.19.238
drm/amdkfd: Use drm_priv to pass VM from KFD to amdgpu
drm/amdgpu: Check if fd really is an amdgpu fd.
xfrm: policy: match with both mark and mask on user interfaces
selftests: cgroup: Test open-time cgroup namespace usage for migration checks
selftests: cgroup: Test open-time credential usage for migration checks
selftests: cgroup: Make cg_create() use 0755 for permission instead of 0644
cgroup: Use open-time cgroup namespace for process migration perm checks
cgroup: Allocate cgroup_file_ctx for kernfs_open_file->priv
cgroup: Use open-time credentials for process migraton perm checks
mm/sparsemem: fix 'mem_section' will never be NULL gcc 12 warning
arm64: module: remove (NOLOAD) from linker script
mm: don't skip swap entry even if zap_details specified
dmaengine: Revert "dmaengine: shdma: Fix runtime PM imbalance on error"
tools build: Use $(shell ) instead of `` to get embedded libperl's ccopts
tools build: Filter out options and warnings not supported by clang
irqchip/gic-v3: Fix GICR_CTLR.RWP polling
perf: qcom_l2_pmu: fix an incorrect NULL check on list iterator
ata: sata_dwc_460ex: Fix crash due to OOB write
arm64: patch_text: Fixup last cpu should be master
btrfs: fix qgroup reserve overflow the qgroup limit
x86/speculation: Restore speculation related MSRs during S3 resume
x86/pm: Save the MSR validity status at context setup
mm/mempolicy: fix mpol_new leak in shared_policy_replace
mmmremap.c: avoid pointless invalidate_range_start/end on mremap(old_size=0)
mmc: renesas_sdhi: don't overwrite TAP settings when HS400 tuning is complete
Revert "mmc: sdhci-xenon: fix annoying 1.8V regulator warning"
drbd: Fix five use after free bugs in get_initial_state
spi: bcm-qspi: fix MSPI only access with bcm_qspi_exec_mem_op()
qede: confirm skb is allocated before using
rxrpc: fix a race in rxrpc_exit_net()
net: openvswitch: don't send internal clone attribute to the userspace.
drm/imx: Fix memory leak in imx_pd_connector_get_modes
net: stmmac: Fix unset max_speed difference between DT and non-DT platforms
scsi: zorro7xx: Fix a resource leak in zorro7xx_remove_one()
Drivers: hv: vmbus: Fix potential crash on module unload
drm/amdgpu: fix off by one in amdgpu_gfx_kiq_acquire()
KVM: arm64: Check arm64_get_bp_hardening_data() didn't return NULL
mm: fix race between MADV_FREE reclaim and blkdev direct IO read
net: add missing SOF_TIMESTAMPING_OPT_ID support
parisc: Fix CPU affinity for Lasi, WAX and Dino chips
jfs: prevent NULL deref in diFree
virtio_console: eliminate anonymous module_init & module_exit
serial: samsung_tty: do not unlock port->lock for uart_write_wakeup()
NFS: swap-out must always use STABLE writes.
NFS: swap IO handling is slightly different for O_DIRECT IO
SUNRPC/call_alloc: async tasks mustn't block waiting for memory
clk: Enforce that disjoints limits are invalid
xen: delay xen_hvm_init_time_ops() if kdump is boot on vcpu>=32
NFSv4: Protect the state recovery thread against direct reclaim
w1: w1_therm: fixes w1_seq for ds28ea00 sensors
minix: fix bug when opening a file with O_DIRECT
init/main.c: return 1 from handled __setup() functions
Bluetooth: Fix use after free in hci_send_acl
xtensa: fix DTC warning unit_address_format
usb: dwc3: omap: fix "unbalanced disables for smps10_out1" on omap5evm
scsi: libfc: Fix use after free in fc_exch_abts_resp()
MIPS: fix fortify panic when copying asm exception handlers
bnxt_en: Eliminate unintended link toggle during FW reset
macvtap: advertise link netns via netlink
net/smc: correct settings of RMB window update limit
scsi: aha152x: Fix aha152x_setup() __setup handler return value
scsi: pm8001: Fix pm8001_mpi_task_abort_resp()
drm/amdkfd: make CRAT table missing message informational only
dm ioctl: prevent potential spectre v1 gadget
ipv4: Invalidate neighbour for broadcast address upon address addition
PCI: pciehp: Add Qualcomm quirk for Command Completed erratum
usb: ehci: add pci device support for Aspeed platforms
iommu/arm-smmu-v3: fix event handling soft lockup
PCI: aardvark: Fix support for MSI interrupts
powerpc: Set crashkernel offset to mid of RMA region
power: supply: axp20x_battery: properly report current when discharging
scsi: bfa: Replace snprintf() with sysfs_emit()
scsi: mvsas: Replace snprintf() with sysfs_emit()
powerpc: dts: t104xrdb: fix phy type for FMAN 4/5
ptp: replace snprintf with sysfs_emit
drm/amd/amdgpu/amdgpu_cs: fix refcount leak of a dma_fence obj
ath5k: fix OOB in ath5k_eeprom_read_pcal_info_5111
drm: Add orientation quirk for GPD Win Max
KVM: x86/svm: Clear reserved bits written to PerfEvtSeln MSRs
ARM: 9187/1: JIVE: fix return value of __setup handler
riscv module: remove (NOLOAD)
rtc: wm8350: Handle error for wm8350_register_irq
ubifs: Rectify space amount budget for mkdir/tmpfile operations
KVM: x86: Forbid VMM to set SYNIC/STIMER MSRs when SynIC wasn't activated
openvswitch: Fixed nd target mask field in the flow dump.
um: Fix uml_mconsole stop/go
ARM: dts: spear13xx: Update SPI dma properties
ARM: dts: spear1340: Update serial node properties
ASoC: topology: Allow TLV control to be either read or write
ubi: fastmap: Return error code if memory allocation fails in add_aeb()
bpf: Fix comment for helper bpf_current_task_under_cgroup()
mm/usercopy: return 1 from hardened_usercopy __setup() handler
mm/memcontrol: return 1 from cgroup.memory __setup() handler
mm/mmap: return 1 from stack_guard_gap __setup() handler
ACPI: CPPC: Avoid out of bounds access when parsing _CPC data
ubi: Fix race condition between ctrl_cdev_ioctl and ubi_cdev_ioctl
pinctrl: pinconf-generic: Print arguments for bias-pull-*
gfs2: Make sure FITRIM minlen is rounded up to fs block size
can: mcba_usb: properly check endpoint type
can: mcba_usb: mcba_usb_start_xmit(): fix double dev_kfree_skb in error path
ubifs: rename_whiteout: correct old_dir size computing
ubifs: Fix read out-of-bounds in ubifs_wbuf_write_nolock()
ubifs: setflags: Make dirtied_ino_d 8 bytes aligned
ubifs: Add missing iput if do_tmpfile() failed in rename whiteout
ubifs: Fix deadlock in concurrent rename whiteout and inode writeback
ubifs: rename_whiteout: Fix double free for whiteout_ui->data
KVM: x86: fix sending PV IPI
KVM: Prevent module exit until all VMs are freed
scsi: qla2xxx: Use correct feature type field during RFF_ID processing
scsi: qla2xxx: Reduce false trigger to login
scsi: qla2xxx: Fix hang due to session stuck
scsi: qla2xxx: Fix incorrect reporting of task management failure
scsi: qla2xxx: Suppress a kernel complaint in qla_create_qpair()
scsi: qla2xxx: Check for firmware dump already collected
scsi: qla2xxx: Fix warning for missing error code
scsi: qla2xxx: Fix stuck session in gpdb
powerpc: Fix build errors with newer binutils
powerpc/lib/sstep: Fix build errors with newer binutils
powerpc/lib/sstep: Fix 'sthcx' instruction
mmc: host: Return an error when ->enable_sdio_irq() ops is missing
media: hdpvr: initialize dev->worker at hdpvr_register_videodev
media: Revert "media: em28xx: add missing em28xx_close_extension"
video: fbdev: sm712fb: Fix crash in smtcfb_write()
ARM: mmp: Fix failure to remove sram device
ARM: tegra: tamonten: Fix I2C3 pad setting
media: cx88-mpeg: clear interrupt status register before streaming video
ASoC: soc-core: skip zero num_dai component in searching dai name
video: fbdev: udlfb: replace snprintf in show functions with sysfs_emit
video: fbdev: omapfb: panel-tpo-td043mtea1: Use sysfs_emit() instead of snprintf()
video: fbdev: omapfb: panel-dsi-cm: Use sysfs_emit() instead of snprintf()
ARM: dts: bcm2837: Add the missing L1/L2 cache information
ARM: dts: qcom: fix gic_irq_domain_translate warnings for msm8960
video: fbdev: omapfb: acx565akm: replace snprintf with sysfs_emit
video: fbdev: cirrusfb: check pixclock to avoid divide by zero
video: fbdev: w100fb: Reset global state
video: fbdev: nvidiafb: Use strscpy() to prevent buffer overflow
ntfs: add sanity check on allocation size
ext4: don't BUG if someone dirty pages without asking ext4 first
spi: tegra20: Use of_device_get_match_data()
PM: core: keep irq flags in device_pm_check_callbacks()
ACPI/APEI: Limit printable size of BERT table data
Revert "Revert "block, bfq: honor already-setup queue merges""
lib/raid6/test/Makefile: Use $(pound) instead of \# for Make 4.3
ACPICA: Avoid walking the ACPI Namespace if it is not there
bfq: fix use-after-free in bfq_dispatch_request
irqchip/nvic: Release nvic_base upon failure
irqchip/qcom-pdc: Fix broken locking
Fix incorrect type in assignment of ipv6 port for audit
loop: use sysfs_emit() in the sysfs xxx show()
selinux: use correct type for context length
lib/test: use after free in register_test_dev_kmod()
NFSv4/pNFS: Fix another issue with a list iterator pointing to the head
net/x25: Fix null-ptr-deref caused by x25_disconnect
qlcnic: dcb: default to returning -EOPNOTSUPP
net: phy: broadcom: Fix brcm_fet_config_init()
xen: fix is_xen_pmu()
clk: qcom: gcc-msm8994: Fix gpll4 width
netfilter: nf_conntrack_tcp: preserve liberal flag in tcp options
jfs: fix divide error in dbNextAG
kgdbts: fix return value of __setup handler
kgdboc: fix return value of __setup handler
tty: hvc: fix return value of __setup handler
pinctrl/rockchip: Add missing of_node_put() in rockchip_pinctrl_probe
pinctrl: nomadik: Add missing of_node_put() in nmk_pinctrl_probe
pinctrl: mediatek: Fix missing of_node_put() in mtk_pctrl_init
NFS: remove unneeded check in decode_devicenotify_args()
clk: tegra: tegra124-emc: Fix missing put_device() call in emc_ensure_emc_driver
clk: clps711x: Terminate clk_div_table with sentinel element
clk: loongson1: Terminate clk_div_table with sentinel element
clk: actions: Terminate clk_div_table with sentinel element
remoteproc: qcom_wcnss: Add missing of_node_put() in wcnss_alloc_memory_region
clk: qcom: clk-rcg2: Update the frac table for pixel clock
dma-debug: fix return value of __setup handlers
iio: adc: Add check for devm_request_threaded_irq
serial: 8250: Fix race condition in RTS-after-send handling
serial: 8250_mid: Balance reference count for PCI DMA device
clk: qcom: ipq8074: Use floor ops for SDCC1 clock
staging:iio:adc:ad7280a: Fix handing of device address bit reversing.
pwm: lpc18xx-sct: Initialize driver data and hardware before pwmchip_add()
mxser: fix xmit_buf leak in activate when LSR == 0xff
mfd: asic3: Add missing iounmap() on error asic3_mfd_probe
tcp: ensure PMTU updates are processed during fastopen
selftests/bpf/test_lirc_mode2.sh: Exit with proper code
i2c: mux: demux-pinctrl: do not deactivate a master that is not active
af_netlink: Fix shift out of bounds in group mask calculation
USB: storage: ums-realtek: fix error code in rts51x_read_mem()
mtd: rawnand: atmel: fix refcount issue in atmel_nand_controller_init
MIPS: RB532: fix return value of __setup handler
vxcan: enable local echo for sent CAN frames
mfd: mc13xxx: Add check for mc13xxx_irq_request
powerpc/sysdev: fix incorrect use to determine if list is empty
PCI: Reduce warnings on possible RW1C corruption
power: supply: wm8350-power: Add missing free in free_charger_irq
power: supply: wm8350-power: Handle error for wm8350_register_irq
i2c: xiic: Make bus names unique
hv_balloon: rate-limit "Unhandled message" warning
KVM: x86/emulator: Defer not-present segment check in __load_segment_descriptor()
KVM: x86: Fix emulation in writing cr8
powerpc/Makefile: Don't pass -mcpu=powerpc64 when building 32-bit
drm/bridge: cdns-dsi: Make sure to to create proper aliases for dt
power: supply: bq24190_charger: Fix bq24190_vbus_is_enabled() wrong false return
drm/tegra: Fix reference leak in tegra_dsi_ganged_probe
ext2: correct max file size computing
TOMOYO: fix __setup handlers return values
scsi: pm8001: Fix abort all task initialization
scsi: pm8001: Fix payload initialization in pm80xx_set_thermal_config()
scsi: pm8001: Fix command initialization in pm8001_chip_ssp_tm_req()
scsi: pm8001: Fix command initialization in pm80XX_send_read_log()
dm crypt: fix get_key_size compiler warning if !CONFIG_KEYS
iwlwifi: Fix -EIO error code that is never returned
HID: i2c-hid: fix GET/SET_REPORT for unnumbered reports
power: supply: ab8500: Fix memory leak in ab8500_fg_sysfs_init
ray_cs: Check ioremap return value
power: reset: gemini-poweroff: Fix IRQ check in gemini_poweroff_probe
KVM: PPC: Fix vmx/vsx mixup in mmio emulation
ath9k_htc: fix uninit value bugs
drm/amd/display: Fix a NULL pointer dereference in amdgpu_dm_connector_add_common_modes()
drm/edid: Don't clear formats if using deep color
mtd: onenand: Check for error irq
Bluetooth: hci_serdev: call init_rwsem() before p->open()
ath10k: fix memory overwrite of the WoWLAN wakeup packet pattern
drm/bridge: Fix free wrong object in sii8620_init_rcp_input_dev
mmc: davinci_mmc: Handle error for clk_enable
ASoC: msm8916-wcd-digital: Fix missing clk_disable_unprepare() in msm8916_wcd_digital_probe
ASoC: imx-es8328: Fix error return code in imx_es8328_probe()
ASoC: mxs: Fix error handling in mxs_sgtl5000_probe
ASoC: dmaengine: do not use a NULL prepare_slave_config() callback
video: fbdev: omapfb: Add missing of_node_put() in dvic_probe_of
ASoC: fsi: Add check for clk_enable
ASoC: wm8350: Handle error for wm8350_register_irq
ASoC: atmel: Add missing of_node_put() in at91sam9g20ek_audio_probe
media: stk1160: If start stream fails, return buffers with VB2_BUF_STATE_QUEUED
ALSA: firewire-lib: fix uninitialized flag for AV/C deferred transaction
memory: emif: check the pointer temp in get_device_details()
memory: emif: Add check for setup_interrupts
ASoC: atmel_ssc_dai: Handle errors for clk_enable
ASoC: mxs-saif: Handle errors for clk_enable
printk: fix return value of printk.devkmsg __setup handler
arm64: dts: broadcom: Fix sata nodename
arm64: dts: ns2: Fix spi-cpol and spi-cpha property
ALSA: spi: Add check for clk_enable()
ASoC: ti: davinci-i2s: Add check for clk_enable()
ASoC: rt5663: check the return value of devm_kzalloc() in rt5663_parse_dp()
media: usb: go7007: s2250-board: fix leak in probe()
media: em28xx: initialize refcount before kref_get
soc: ti: wkup_m3_ipc: Fix IRQ check in wkup_m3_ipc_probe
ARM: dts: qcom: ipq4019: fix sleep clock
video: fbdev: fbcvt.c: fix printing in fb_cvt_print_name()
video: fbdev: smscufx: Fix null-ptr-deref in ufx_usb_probe()
media: coda: Fix missing put_device() call in coda_get_vdoa_data
perf/x86/intel/pt: Fix address filter config for 32-bit kernel
perf/core: Fix address filter parser for multiple filters
sched/debug: Remove mpol_get/put and task_lock/unlock from sched_show_numa
clocksource: acpi_pm: fix return value of __setup handler
hwmon: (pmbus) Add Vin unit off handling
crypto: ccp - ccp_dmaengine_unregister release dma channels
ACPI: APEI: fix return value of __setup handlers
clocksource/drivers/timer-of: Check return value of of_iomap in timer_of_base_init()
crypto: vmx - add missing dependencies
hwrng: atmel - disable trng on failure path
PM: suspend: fix return value of __setup handler
PM: hibernate: fix __setup handler error handling
block: don't delete queue kobject before its children
hwmon: (sch56xx-common) Replace WDOG_ACTIVE with WDOG_HW_RUNNING
hwmon: (pmbus) Add mutex to regulator ops
spi: pxa2xx-pci: Balance reference count for PCI DMA device
selftests/x86: Add validity check and allow field splitting
spi: tegra114: Add missing IRQ check in tegra_spi_probe
crypto: mxs-dcp - Fix scatterlist processing
crypto: authenc - Fix sleep in atomic context in decrypt_tail
regulator: qcom_smd: fix for_each_child.cocci warnings
PCI: pciehp: Clear cmd_busy bit in polling mode
brcmfmac: pcie: Replace brcmf_pcie_copy_mem_todev with memcpy_toio
brcmfmac: firmware: Allocate space for default boardrev in nvram
media: davinci: vpif: fix unbalanced runtime PM get
DEC: Limit PMAX memory probing to R3k systems
lib/raid6/test: fix multiple definition linking error
thermal: int340x: Increase bitmap size
carl9170: fix missing bit-wise or operator for tx_params
ARM: dts: exynos: add missing HDMI supplies on SMDK5420
ARM: dts: exynos: add missing HDMI supplies on SMDK5250
ARM: dts: exynos: fix UART3 pins configuration in Exynos5250
ARM: dts: at91: sama5d2: Fix PMERRLOC resource size
video: fbdev: atari: Atari 2 bpp (STe) palette bugfix
video: fbdev: sm712fb: Fix crash in smtcfb_read()
drm/edid: check basic audio support on CEA extension block
block: don't merge across cgroup boundaries if blkcg is enabled
drivers: hamradio: 6pack: fix UAF bug caused by mod_timer()
ACPI: properties: Consistently return -ENOENT if there are no more references
powerpc/kvm: Fix kvm_use_magic_page
drbd: fix potential silent data corruption
mm,hwpoison: unmap poisoned page before invalidation
ALSA: hda/realtek: Fix audio regression on Mi Notebook Pro 2020
ALSA: cs4236: fix an incorrect NULL check on list iterator
Revert "Input: clear BTN_RIGHT/MIDDLE on buttonpads"
qed: validate and restrict untrusted VFs vlan promisc mode
qed: display VF trust config
scsi: libsas: Fix sas_ata_qc_issue() handling of NCQ NON DATA commands
mempolicy: mbind_range() set_policy() after vma_merge()
mm: invalidate hwpoison page cache page in fault path
mm/pages_alloc.c: don't create ZONE_MOVABLE beyond the end of a node
jffs2: fix memory leak in jffs2_scan_medium
jffs2: fix memory leak in jffs2_do_mount_fs
jffs2: fix use-after-free in jffs2_clear_xattr_subsystem
can: ems_usb: ems_usb_start_xmit(): fix double dev_kfree_skb() in error path
pinctrl: samsung: drop pin banks references on error paths
f2fs: fix to unlock page correctly in error path of is_alive()
NFSD: prevent integer overflow on 32 bit systems
NFSD: prevent underflow in nfssvc_decode_writeargs()
SUNRPC: avoid race between mod_timer() and del_timer_sync()
Documentation: update stable tree link
Documentation: add link to stable release candidate tree
ptrace: Check PTRACE_O_SUSPEND_SECCOMP permission on PTRACE_SEIZE
clk: uniphier: Fix fixed-rate initialization
iio: inkern: make a best effort on offset calculation
iio: inkern: apply consumer scale when no channel scale is available
iio: inkern: apply consumer scale on IIO_VAL_INT cases
iio: afe: rescale: use s64 for temporary scale calculations
coresight: Fix TRCCONFIGR.QE sysfs interface
xhci: make xhci_handshake timeout for xhci_reset() adjustable
USB: usb-storage: Fix use of bitfields for hardware data in ene_ub6250.c
virtio-blk: Use blk_validate_block_size() to validate block size
block: Add a helper to validate the block size
tpm: fix reference counting for struct tpm_chip
fuse: fix pipe buffer lifetime for direct_io
af_key: add __GFP_ZERO flag for compose_sadb_supported in function pfkey_register
spi: Fix erroneous sgs value with min_t()
net:mcf8390: Use platform_get_irq() to get the interrupt
spi: Fix invalid sgs value
ethernet: sun: Free the coherent when failing in probing
virtio_console: break out of buf poll on remove
xfrm: fix tunnel model fragmentation behavior
netdevice: add the case if dev is NULL
USB: serial: simple: add Nokia phone driver
USB: serial: pl2303: add IBM device IDs
ANDROID: incremental-fs: limit mount stack depth
UPSTREAM: binderfs: use __u32 for device numbers
Change-Id: I9b0e91bd0ad82b16888f9f7063ce346d520018f9
Signed-off-by: UtsavBalar1231 <utsavbalar1231@gmail.com>
Conflicts:
drivers/irqchip/irq-gic-v3.c
drivers/irqchip/qcom-pdc.c
drivers/mmc/core/host.c
drivers/usb/host/xhci.c
drivers/usb/host/xhci.h
874 lines
21 KiB
C
874 lines
21 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Functions related to segment and merge handling
|
|
*/
|
|
#include <linux/kernel.h>
|
|
#include <linux/module.h>
|
|
#include <linux/bio.h>
|
|
#include <linux/blkdev.h>
|
|
#include <linux/scatterlist.h>
|
|
#ifndef __GENKSYMS__
|
|
#include <linux/blkdev.h>
|
|
#include <linux/blk-cgroup.h>
|
|
#endif
|
|
|
|
#include <trace/events/block.h>
|
|
|
|
#include "blk.h"
|
|
|
|
static struct bio *blk_bio_discard_split(struct request_queue *q,
|
|
struct bio *bio,
|
|
struct bio_set *bs,
|
|
unsigned *nsegs)
|
|
{
|
|
unsigned int max_discard_sectors, granularity;
|
|
int alignment;
|
|
sector_t tmp;
|
|
unsigned split_sectors;
|
|
|
|
*nsegs = 1;
|
|
|
|
/* Zero-sector (unknown) and one-sector granularities are the same. */
|
|
granularity = max(q->limits.discard_granularity >> 9, 1U);
|
|
|
|
max_discard_sectors = min(q->limits.max_discard_sectors,
|
|
bio_allowed_max_sectors(q));
|
|
max_discard_sectors -= max_discard_sectors % granularity;
|
|
|
|
if (unlikely(!max_discard_sectors)) {
|
|
/* XXX: warn */
|
|
return NULL;
|
|
}
|
|
|
|
if (bio_sectors(bio) <= max_discard_sectors)
|
|
return NULL;
|
|
|
|
split_sectors = max_discard_sectors;
|
|
|
|
/*
|
|
* If the next starting sector would be misaligned, stop the discard at
|
|
* the previous aligned sector.
|
|
*/
|
|
alignment = (q->limits.discard_alignment >> 9) % granularity;
|
|
|
|
tmp = bio->bi_iter.bi_sector + split_sectors - alignment;
|
|
tmp = sector_div(tmp, granularity);
|
|
|
|
if (split_sectors > tmp)
|
|
split_sectors -= tmp;
|
|
|
|
return bio_split(bio, split_sectors, GFP_NOIO, bs);
|
|
}
|
|
|
|
static struct bio *blk_bio_write_zeroes_split(struct request_queue *q,
|
|
struct bio *bio, struct bio_set *bs, unsigned *nsegs)
|
|
{
|
|
*nsegs = 1;
|
|
|
|
if (!q->limits.max_write_zeroes_sectors)
|
|
return NULL;
|
|
|
|
if (bio_sectors(bio) <= q->limits.max_write_zeroes_sectors)
|
|
return NULL;
|
|
|
|
return bio_split(bio, q->limits.max_write_zeroes_sectors, GFP_NOIO, bs);
|
|
}
|
|
|
|
static struct bio *blk_bio_write_same_split(struct request_queue *q,
|
|
struct bio *bio,
|
|
struct bio_set *bs,
|
|
unsigned *nsegs)
|
|
{
|
|
*nsegs = 1;
|
|
|
|
if (!q->limits.max_write_same_sectors)
|
|
return NULL;
|
|
|
|
if (bio_sectors(bio) <= q->limits.max_write_same_sectors)
|
|
return NULL;
|
|
|
|
return bio_split(bio, q->limits.max_write_same_sectors, GFP_NOIO, bs);
|
|
}
|
|
|
|
static inline unsigned get_max_io_size(struct request_queue *q,
|
|
struct bio *bio)
|
|
{
|
|
unsigned sectors = blk_max_size_offset(q, bio->bi_iter.bi_sector);
|
|
unsigned mask = queue_logical_block_size(q) - 1;
|
|
|
|
/* aligned to logical block size */
|
|
sectors &= ~(mask >> 9);
|
|
|
|
return sectors;
|
|
}
|
|
|
|
static struct bio *blk_bio_segment_split(struct request_queue *q,
|
|
struct bio *bio,
|
|
struct bio_set *bs,
|
|
unsigned *segs)
|
|
{
|
|
struct bio_vec bv, bvprv, *bvprvp = NULL;
|
|
struct bvec_iter iter;
|
|
unsigned seg_size = 0, nsegs = 0, sectors = 0;
|
|
unsigned front_seg_size = bio->bi_seg_front_size;
|
|
bool do_split = true;
|
|
struct bio *new = NULL;
|
|
const unsigned max_sectors = get_max_io_size(q, bio);
|
|
|
|
bio_for_each_segment(bv, bio, iter) {
|
|
/*
|
|
* If the queue doesn't support SG gaps and adding this
|
|
* offset would create a gap, disallow it.
|
|
*/
|
|
if (bvprvp && bvec_gap_to_prev(q, bvprvp, bv.bv_offset))
|
|
goto split;
|
|
|
|
if (sectors + (bv.bv_len >> 9) > max_sectors) {
|
|
/*
|
|
* Consider this a new segment if we're splitting in
|
|
* the middle of this vector.
|
|
*/
|
|
if (nsegs < queue_max_segments(q) &&
|
|
sectors < max_sectors) {
|
|
nsegs++;
|
|
sectors = max_sectors;
|
|
}
|
|
goto split;
|
|
}
|
|
|
|
if (bvprvp && blk_queue_cluster(q)) {
|
|
if (seg_size + bv.bv_len > queue_max_segment_size(q))
|
|
goto new_segment;
|
|
if (!BIOVEC_PHYS_MERGEABLE(bvprvp, &bv))
|
|
goto new_segment;
|
|
if (!BIOVEC_SEG_BOUNDARY(q, bvprvp, &bv))
|
|
goto new_segment;
|
|
|
|
seg_size += bv.bv_len;
|
|
bvprv = bv;
|
|
bvprvp = &bvprv;
|
|
sectors += bv.bv_len >> 9;
|
|
|
|
continue;
|
|
}
|
|
new_segment:
|
|
if (nsegs == queue_max_segments(q))
|
|
goto split;
|
|
|
|
if (nsegs == 1 && seg_size > front_seg_size)
|
|
front_seg_size = seg_size;
|
|
|
|
nsegs++;
|
|
bvprv = bv;
|
|
bvprvp = &bvprv;
|
|
seg_size = bv.bv_len;
|
|
sectors += bv.bv_len >> 9;
|
|
|
|
}
|
|
|
|
do_split = false;
|
|
split:
|
|
*segs = nsegs;
|
|
|
|
if (do_split) {
|
|
new = bio_split(bio, sectors, GFP_NOIO, bs);
|
|
if (new)
|
|
bio = new;
|
|
}
|
|
|
|
if (nsegs == 1 && seg_size > front_seg_size)
|
|
front_seg_size = seg_size;
|
|
bio->bi_seg_front_size = front_seg_size;
|
|
if (seg_size > bio->bi_seg_back_size)
|
|
bio->bi_seg_back_size = seg_size;
|
|
|
|
return do_split ? new : NULL;
|
|
}
|
|
|
|
void blk_queue_split(struct request_queue *q, struct bio **bio)
|
|
{
|
|
struct bio *split, *res;
|
|
unsigned nsegs;
|
|
|
|
switch (bio_op(*bio)) {
|
|
case REQ_OP_DISCARD:
|
|
case REQ_OP_SECURE_ERASE:
|
|
split = blk_bio_discard_split(q, *bio, &q->bio_split, &nsegs);
|
|
break;
|
|
case REQ_OP_WRITE_ZEROES:
|
|
split = blk_bio_write_zeroes_split(q, *bio, &q->bio_split, &nsegs);
|
|
break;
|
|
case REQ_OP_WRITE_SAME:
|
|
split = blk_bio_write_same_split(q, *bio, &q->bio_split, &nsegs);
|
|
break;
|
|
default:
|
|
split = blk_bio_segment_split(q, *bio, &q->bio_split, &nsegs);
|
|
break;
|
|
}
|
|
|
|
/* physical segments can be figured out during splitting */
|
|
res = split ? split : *bio;
|
|
res->bi_phys_segments = nsegs;
|
|
bio_set_flag(res, BIO_SEG_VALID);
|
|
|
|
if (split) {
|
|
/* there isn't chance to merge the splitted bio */
|
|
split->bi_opf |= REQ_NOMERGE;
|
|
|
|
/*
|
|
* Since we're recursing into make_request here, ensure
|
|
* that we mark this bio as already having entered the queue.
|
|
* If not, and the queue is going away, we can get stuck
|
|
* forever on waiting for the queue reference to drop. But
|
|
* that will never happen, as we're already holding a
|
|
* reference to it.
|
|
*/
|
|
bio_set_flag(*bio, BIO_QUEUE_ENTERED);
|
|
|
|
bio_chain(split, *bio);
|
|
trace_block_split(q, split, (*bio)->bi_iter.bi_sector);
|
|
generic_make_request(*bio);
|
|
*bio = split;
|
|
}
|
|
}
|
|
EXPORT_SYMBOL(blk_queue_split);
|
|
|
|
static unsigned int __blk_recalc_rq_segments(struct request_queue *q,
|
|
struct bio *bio,
|
|
bool no_sg_merge)
|
|
{
|
|
struct bio_vec bv, bvprv = { NULL };
|
|
int cluster, prev = 0;
|
|
unsigned int seg_size, nr_phys_segs;
|
|
struct bio *fbio, *bbio;
|
|
struct bvec_iter iter;
|
|
|
|
if (!bio)
|
|
return 0;
|
|
|
|
switch (bio_op(bio)) {
|
|
case REQ_OP_DISCARD:
|
|
case REQ_OP_SECURE_ERASE:
|
|
case REQ_OP_WRITE_ZEROES:
|
|
return 0;
|
|
case REQ_OP_WRITE_SAME:
|
|
return 1;
|
|
}
|
|
|
|
fbio = bio;
|
|
cluster = blk_queue_cluster(q);
|
|
seg_size = 0;
|
|
nr_phys_segs = 0;
|
|
for_each_bio(bio) {
|
|
bio_for_each_segment(bv, bio, iter) {
|
|
/*
|
|
* If SG merging is disabled, each bio vector is
|
|
* a segment
|
|
*/
|
|
if (no_sg_merge)
|
|
goto new_segment;
|
|
|
|
if (prev && cluster) {
|
|
if (seg_size + bv.bv_len
|
|
> queue_max_segment_size(q))
|
|
goto new_segment;
|
|
if (!BIOVEC_PHYS_MERGEABLE(&bvprv, &bv))
|
|
goto new_segment;
|
|
if (!BIOVEC_SEG_BOUNDARY(q, &bvprv, &bv))
|
|
goto new_segment;
|
|
|
|
seg_size += bv.bv_len;
|
|
bvprv = bv;
|
|
continue;
|
|
}
|
|
new_segment:
|
|
if (nr_phys_segs == 1 && seg_size >
|
|
fbio->bi_seg_front_size)
|
|
fbio->bi_seg_front_size = seg_size;
|
|
|
|
nr_phys_segs++;
|
|
bvprv = bv;
|
|
prev = 1;
|
|
seg_size = bv.bv_len;
|
|
}
|
|
bbio = bio;
|
|
}
|
|
|
|
if (nr_phys_segs == 1 && seg_size > fbio->bi_seg_front_size)
|
|
fbio->bi_seg_front_size = seg_size;
|
|
if (seg_size > bbio->bi_seg_back_size)
|
|
bbio->bi_seg_back_size = seg_size;
|
|
|
|
return nr_phys_segs;
|
|
}
|
|
|
|
void blk_recalc_rq_segments(struct request *rq)
|
|
{
|
|
bool no_sg_merge = !!test_bit(QUEUE_FLAG_NO_SG_MERGE,
|
|
&rq->q->queue_flags);
|
|
|
|
rq->nr_phys_segments = __blk_recalc_rq_segments(rq->q, rq->bio,
|
|
no_sg_merge);
|
|
}
|
|
|
|
void blk_recount_segments(struct request_queue *q, struct bio *bio)
|
|
{
|
|
unsigned short seg_cnt = bio_segments(bio);
|
|
|
|
if (test_bit(QUEUE_FLAG_NO_SG_MERGE, &q->queue_flags) &&
|
|
(seg_cnt < queue_max_segments(q)))
|
|
bio->bi_phys_segments = seg_cnt;
|
|
else {
|
|
struct bio *nxt = bio->bi_next;
|
|
|
|
bio->bi_next = NULL;
|
|
bio->bi_phys_segments = __blk_recalc_rq_segments(q, bio, false);
|
|
bio->bi_next = nxt;
|
|
}
|
|
|
|
bio_set_flag(bio, BIO_SEG_VALID);
|
|
}
|
|
EXPORT_SYMBOL(blk_recount_segments);
|
|
|
|
static int blk_phys_contig_segment(struct request_queue *q, struct bio *bio,
|
|
struct bio *nxt)
|
|
{
|
|
struct bio_vec end_bv = { NULL }, nxt_bv;
|
|
|
|
if (!blk_queue_cluster(q))
|
|
return 0;
|
|
|
|
if (bio->bi_seg_back_size + nxt->bi_seg_front_size >
|
|
queue_max_segment_size(q))
|
|
return 0;
|
|
|
|
if (!bio_has_data(bio))
|
|
return 1;
|
|
|
|
bio_get_last_bvec(bio, &end_bv);
|
|
bio_get_first_bvec(nxt, &nxt_bv);
|
|
|
|
if (!BIOVEC_PHYS_MERGEABLE(&end_bv, &nxt_bv))
|
|
return 0;
|
|
|
|
/*
|
|
* bio and nxt are contiguous in memory; check if the queue allows
|
|
* these two to be merged into one
|
|
*/
|
|
if (BIOVEC_SEG_BOUNDARY(q, &end_bv, &nxt_bv))
|
|
return 1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static inline void
|
|
__blk_segment_map_sg(struct request_queue *q, struct bio_vec *bvec,
|
|
struct scatterlist *sglist, struct bio_vec *bvprv,
|
|
struct scatterlist **sg, int *nsegs, int *cluster)
|
|
{
|
|
|
|
int nbytes = bvec->bv_len;
|
|
|
|
if (*sg && *cluster) {
|
|
if ((*sg)->length + nbytes > queue_max_segment_size(q))
|
|
goto new_segment;
|
|
|
|
if (!BIOVEC_PHYS_MERGEABLE(bvprv, bvec))
|
|
goto new_segment;
|
|
if (!BIOVEC_SEG_BOUNDARY(q, bvprv, bvec))
|
|
goto new_segment;
|
|
|
|
(*sg)->length += nbytes;
|
|
} else {
|
|
new_segment:
|
|
if (!*sg)
|
|
*sg = sglist;
|
|
else {
|
|
/*
|
|
* If the driver previously mapped a shorter
|
|
* list, we could see a termination bit
|
|
* prematurely unless it fully inits the sg
|
|
* table on each mapping. We KNOW that there
|
|
* must be more entries here or the driver
|
|
* would be buggy, so force clear the
|
|
* termination bit to avoid doing a full
|
|
* sg_init_table() in drivers for each command.
|
|
*/
|
|
sg_unmark_end(*sg);
|
|
*sg = sg_next(*sg);
|
|
}
|
|
|
|
sg_set_page(*sg, bvec->bv_page, nbytes, bvec->bv_offset);
|
|
(*nsegs)++;
|
|
}
|
|
*bvprv = *bvec;
|
|
}
|
|
|
|
static inline int __blk_bvec_map_sg(struct request_queue *q, struct bio_vec bv,
|
|
struct scatterlist *sglist, struct scatterlist **sg)
|
|
{
|
|
*sg = sglist;
|
|
sg_set_page(*sg, bv.bv_page, bv.bv_len, bv.bv_offset);
|
|
return 1;
|
|
}
|
|
|
|
static int __blk_bios_map_sg(struct request_queue *q, struct bio *bio,
|
|
struct scatterlist *sglist,
|
|
struct scatterlist **sg)
|
|
{
|
|
struct bio_vec bvec, bvprv = { NULL };
|
|
struct bvec_iter iter;
|
|
int cluster = blk_queue_cluster(q), nsegs = 0;
|
|
|
|
for_each_bio(bio)
|
|
bio_for_each_segment(bvec, bio, iter)
|
|
__blk_segment_map_sg(q, &bvec, sglist, &bvprv, sg,
|
|
&nsegs, &cluster);
|
|
|
|
return nsegs;
|
|
}
|
|
|
|
/*
|
|
* map a request to scatterlist, return number of sg entries setup. Caller
|
|
* must make sure sg can hold rq->nr_phys_segments entries
|
|
*/
|
|
int blk_rq_map_sg(struct request_queue *q, struct request *rq,
|
|
struct scatterlist *sglist)
|
|
{
|
|
struct scatterlist *sg = NULL;
|
|
int nsegs = 0;
|
|
|
|
if (rq->rq_flags & RQF_SPECIAL_PAYLOAD)
|
|
nsegs = __blk_bvec_map_sg(q, rq->special_vec, sglist, &sg);
|
|
else if (rq->bio && bio_op(rq->bio) == REQ_OP_WRITE_SAME)
|
|
nsegs = __blk_bvec_map_sg(q, bio_iovec(rq->bio), sglist, &sg);
|
|
else if (rq->bio)
|
|
nsegs = __blk_bios_map_sg(q, rq->bio, sglist, &sg);
|
|
|
|
if (unlikely(rq->rq_flags & RQF_COPY_USER) &&
|
|
(blk_rq_bytes(rq) & q->dma_pad_mask)) {
|
|
unsigned int pad_len =
|
|
(q->dma_pad_mask & ~blk_rq_bytes(rq)) + 1;
|
|
|
|
sg->length += pad_len;
|
|
rq->extra_len += pad_len;
|
|
}
|
|
|
|
if (q->dma_drain_size && q->dma_drain_needed(rq)) {
|
|
if (op_is_write(req_op(rq)))
|
|
memset(q->dma_drain_buffer, 0, q->dma_drain_size);
|
|
|
|
sg_unmark_end(sg);
|
|
sg = sg_next(sg);
|
|
sg_set_page(sg, virt_to_page(q->dma_drain_buffer),
|
|
q->dma_drain_size,
|
|
((unsigned long)q->dma_drain_buffer) &
|
|
(PAGE_SIZE - 1));
|
|
nsegs++;
|
|
rq->extra_len += q->dma_drain_size;
|
|
}
|
|
|
|
if (sg)
|
|
sg_mark_end(sg);
|
|
|
|
/*
|
|
* Something must have been wrong if the figured number of
|
|
* segment is bigger than number of req's physical segments
|
|
*/
|
|
WARN_ON(nsegs > blk_rq_nr_phys_segments(rq));
|
|
|
|
return nsegs;
|
|
}
|
|
EXPORT_SYMBOL(blk_rq_map_sg);
|
|
|
|
static inline int ll_new_hw_segment(struct request_queue *q,
|
|
struct request *req,
|
|
struct bio *bio)
|
|
{
|
|
int nr_phys_segs = bio_phys_segments(q, bio);
|
|
|
|
if (req->nr_phys_segments + nr_phys_segs > queue_max_segments(q))
|
|
goto no_merge;
|
|
|
|
if (!blk_cgroup_mergeable(req, bio))
|
|
goto no_merge;
|
|
|
|
if (blk_integrity_merge_bio(q, req, bio) == false)
|
|
goto no_merge;
|
|
|
|
/*
|
|
* This will form the start of a new hw segment. Bump both
|
|
* counters.
|
|
*/
|
|
req->nr_phys_segments += nr_phys_segs;
|
|
return 1;
|
|
|
|
no_merge:
|
|
req_set_nomerge(q, req);
|
|
return 0;
|
|
}
|
|
|
|
int ll_back_merge_fn(struct request_queue *q, struct request *req,
|
|
struct bio *bio)
|
|
{
|
|
if (req_gap_back_merge(req, bio))
|
|
return 0;
|
|
if (blk_integrity_rq(req) &&
|
|
integrity_req_gap_back_merge(req, bio))
|
|
return 0;
|
|
if (blk_rq_sectors(req) + bio_sectors(bio) >
|
|
blk_rq_get_max_sectors(req, blk_rq_pos(req))) {
|
|
req_set_nomerge(q, req);
|
|
return 0;
|
|
}
|
|
if (!bio_crypt_ctx_mergeable(req->bio, blk_rq_bytes(req), bio))
|
|
return 0;
|
|
if (!bio_flagged(req->biotail, BIO_SEG_VALID))
|
|
blk_recount_segments(q, req->biotail);
|
|
if (!bio_flagged(bio, BIO_SEG_VALID))
|
|
blk_recount_segments(q, bio);
|
|
|
|
return ll_new_hw_segment(q, req, bio);
|
|
}
|
|
|
|
int ll_front_merge_fn(struct request_queue *q, struct request *req,
|
|
struct bio *bio)
|
|
{
|
|
|
|
if (req_gap_front_merge(req, bio))
|
|
return 0;
|
|
if (blk_integrity_rq(req) &&
|
|
integrity_req_gap_front_merge(req, bio))
|
|
return 0;
|
|
if (blk_rq_sectors(req) + bio_sectors(bio) >
|
|
blk_rq_get_max_sectors(req, bio->bi_iter.bi_sector)) {
|
|
req_set_nomerge(q, req);
|
|
return 0;
|
|
}
|
|
if (!bio_crypt_ctx_mergeable(bio, bio->bi_iter.bi_size, req->bio))
|
|
return 0;
|
|
if (!bio_flagged(bio, BIO_SEG_VALID))
|
|
blk_recount_segments(q, bio);
|
|
if (!bio_flagged(req->bio, BIO_SEG_VALID))
|
|
blk_recount_segments(q, req->bio);
|
|
|
|
return ll_new_hw_segment(q, req, bio);
|
|
}
|
|
|
|
/*
|
|
* blk-mq uses req->special to carry normal driver per-request payload, it
|
|
* does not indicate a prepared command that we cannot merge with.
|
|
*/
|
|
static bool req_no_special_merge(struct request *req)
|
|
{
|
|
struct request_queue *q = req->q;
|
|
|
|
return !q->mq_ops && req->special;
|
|
}
|
|
|
|
static bool req_attempt_discard_merge(struct request_queue *q, struct request *req,
|
|
struct request *next)
|
|
{
|
|
unsigned short segments = blk_rq_nr_discard_segments(req);
|
|
|
|
if (segments >= queue_max_discard_segments(q))
|
|
goto no_merge;
|
|
if (blk_rq_sectors(req) + bio_sectors(next->bio) >
|
|
blk_rq_get_max_sectors(req, blk_rq_pos(req)))
|
|
goto no_merge;
|
|
|
|
req->nr_phys_segments = segments + blk_rq_nr_discard_segments(next);
|
|
return true;
|
|
no_merge:
|
|
req_set_nomerge(q, req);
|
|
return false;
|
|
}
|
|
|
|
static int ll_merge_requests_fn(struct request_queue *q, struct request *req,
|
|
struct request *next)
|
|
{
|
|
int total_phys_segments;
|
|
unsigned int seg_size =
|
|
req->biotail->bi_seg_back_size + next->bio->bi_seg_front_size;
|
|
|
|
/*
|
|
* First check if the either of the requests are re-queued
|
|
* requests. Can't merge them if they are.
|
|
*/
|
|
if (req_no_special_merge(req) || req_no_special_merge(next))
|
|
return 0;
|
|
|
|
if (req_gap_back_merge(req, next->bio))
|
|
return 0;
|
|
|
|
/*
|
|
* Will it become too large?
|
|
*/
|
|
if ((blk_rq_sectors(req) + blk_rq_sectors(next)) >
|
|
blk_rq_get_max_sectors(req, blk_rq_pos(req)))
|
|
return 0;
|
|
|
|
total_phys_segments = req->nr_phys_segments + next->nr_phys_segments;
|
|
if (blk_phys_contig_segment(q, req->biotail, next->bio)) {
|
|
if (req->nr_phys_segments == 1)
|
|
req->bio->bi_seg_front_size = seg_size;
|
|
if (next->nr_phys_segments == 1)
|
|
next->biotail->bi_seg_back_size = seg_size;
|
|
total_phys_segments--;
|
|
}
|
|
|
|
if (total_phys_segments > queue_max_segments(q))
|
|
return 0;
|
|
|
|
if (!blk_cgroup_mergeable(req, next->bio))
|
|
return 0;
|
|
|
|
if (blk_integrity_merge_rq(q, req, next) == false)
|
|
return 0;
|
|
|
|
if (!bio_crypt_ctx_mergeable(req->bio, blk_rq_bytes(req), next->bio))
|
|
return 0;
|
|
|
|
/* Merge is OK... */
|
|
req->nr_phys_segments = total_phys_segments;
|
|
return 1;
|
|
}
|
|
|
|
/**
|
|
* blk_rq_set_mixed_merge - mark a request as mixed merge
|
|
* @rq: request to mark as mixed merge
|
|
*
|
|
* Description:
|
|
* @rq is about to be mixed merged. Make sure the attributes
|
|
* which can be mixed are set in each bio and mark @rq as mixed
|
|
* merged.
|
|
*/
|
|
void blk_rq_set_mixed_merge(struct request *rq)
|
|
{
|
|
unsigned int ff = rq->cmd_flags & REQ_FAILFAST_MASK;
|
|
struct bio *bio;
|
|
|
|
if (rq->rq_flags & RQF_MIXED_MERGE)
|
|
return;
|
|
|
|
/*
|
|
* @rq will no longer represent mixable attributes for all the
|
|
* contained bios. It will just track those of the first one.
|
|
* Distributes the attributs to each bio.
|
|
*/
|
|
for (bio = rq->bio; bio; bio = bio->bi_next) {
|
|
WARN_ON_ONCE((bio->bi_opf & REQ_FAILFAST_MASK) &&
|
|
(bio->bi_opf & REQ_FAILFAST_MASK) != ff);
|
|
bio->bi_opf |= ff;
|
|
}
|
|
rq->rq_flags |= RQF_MIXED_MERGE;
|
|
}
|
|
|
|
static void blk_account_io_merge(struct request *req)
|
|
{
|
|
if (blk_do_io_stat(req)) {
|
|
struct hd_struct *part;
|
|
int cpu;
|
|
|
|
cpu = part_stat_lock();
|
|
part = req->part;
|
|
|
|
part_round_stats(req->q, cpu, part);
|
|
part_dec_in_flight(req->q, part, rq_data_dir(req));
|
|
|
|
hd_struct_put(part);
|
|
part_stat_unlock();
|
|
}
|
|
}
|
|
|
|
/*
|
|
* For non-mq, this has to be called with the request spinlock acquired.
|
|
* For mq with scheduling, the appropriate queue wide lock should be held.
|
|
*/
|
|
static struct request *attempt_merge(struct request_queue *q,
|
|
struct request *req, struct request *next)
|
|
{
|
|
if (!q->mq_ops)
|
|
lockdep_assert_held(q->queue_lock);
|
|
|
|
if (!rq_mergeable(req) || !rq_mergeable(next))
|
|
return NULL;
|
|
|
|
if (req_op(req) != req_op(next))
|
|
return NULL;
|
|
|
|
/*
|
|
* not contiguous
|
|
*/
|
|
if (blk_rq_pos(req) + blk_rq_sectors(req) != blk_rq_pos(next))
|
|
return NULL;
|
|
|
|
if (rq_data_dir(req) != rq_data_dir(next)
|
|
|| req->rq_disk != next->rq_disk
|
|
|| req_no_special_merge(next))
|
|
return NULL;
|
|
|
|
if (req_op(req) == REQ_OP_WRITE_SAME &&
|
|
!blk_write_same_mergeable(req->bio, next->bio))
|
|
return NULL;
|
|
|
|
/*
|
|
* Don't allow merge of different write hints, or for a hint with
|
|
* non-hint IO.
|
|
*/
|
|
if (req->write_hint != next->write_hint)
|
|
return NULL;
|
|
|
|
/*
|
|
* If we are allowed to merge, then append bio list
|
|
* from next to rq and release next. merge_requests_fn
|
|
* will have updated segment counts, update sector
|
|
* counts here. Handle DISCARDs separately, as they
|
|
* have separate settings.
|
|
*/
|
|
if (req_op(req) == REQ_OP_DISCARD) {
|
|
if (!req_attempt_discard_merge(q, req, next))
|
|
return NULL;
|
|
} else if (!ll_merge_requests_fn(q, req, next))
|
|
return NULL;
|
|
|
|
/*
|
|
* If failfast settings disagree or any of the two is already
|
|
* a mixed merge, mark both as mixed before proceeding. This
|
|
* makes sure that all involved bios have mixable attributes
|
|
* set properly.
|
|
*/
|
|
if (((req->rq_flags | next->rq_flags) & RQF_MIXED_MERGE) ||
|
|
(req->cmd_flags & REQ_FAILFAST_MASK) !=
|
|
(next->cmd_flags & REQ_FAILFAST_MASK)) {
|
|
blk_rq_set_mixed_merge(req);
|
|
blk_rq_set_mixed_merge(next);
|
|
}
|
|
|
|
/*
|
|
* At this point we have either done a back merge or front merge. We
|
|
* need the smaller start_time_ns of the merged requests to be the
|
|
* current request for accounting purposes.
|
|
*/
|
|
if (next->start_time_ns < req->start_time_ns)
|
|
req->start_time_ns = next->start_time_ns;
|
|
|
|
req->biotail->bi_next = next->bio;
|
|
req->biotail = next->biotail;
|
|
|
|
req->__data_len += blk_rq_bytes(next);
|
|
|
|
if (req_op(req) != REQ_OP_DISCARD)
|
|
elv_merge_requests(q, req, next);
|
|
|
|
/*
|
|
* 'next' is going away, so update stats accordingly
|
|
*/
|
|
blk_account_io_merge(next);
|
|
|
|
req->ioprio = ioprio_best(req->ioprio, next->ioprio);
|
|
if (blk_rq_cpu_valid(next))
|
|
req->cpu = next->cpu;
|
|
|
|
/*
|
|
* ownership of bio passed from next to req, return 'next' for
|
|
* the caller to free
|
|
*/
|
|
next->bio = NULL;
|
|
return next;
|
|
}
|
|
|
|
struct request *attempt_back_merge(struct request_queue *q, struct request *rq)
|
|
{
|
|
struct request *next = elv_latter_request(q, rq);
|
|
|
|
if (next)
|
|
return attempt_merge(q, rq, next);
|
|
|
|
return NULL;
|
|
}
|
|
|
|
struct request *attempt_front_merge(struct request_queue *q, struct request *rq)
|
|
{
|
|
struct request *prev = elv_former_request(q, rq);
|
|
|
|
if (prev)
|
|
return attempt_merge(q, prev, rq);
|
|
|
|
return NULL;
|
|
}
|
|
|
|
int blk_attempt_req_merge(struct request_queue *q, struct request *rq,
|
|
struct request *next)
|
|
{
|
|
struct elevator_queue *e = q->elevator;
|
|
struct request *free;
|
|
|
|
if (!e->uses_mq && e->type->ops.sq.elevator_allow_rq_merge_fn)
|
|
if (!e->type->ops.sq.elevator_allow_rq_merge_fn(q, rq, next))
|
|
return 0;
|
|
|
|
free = attempt_merge(q, rq, next);
|
|
if (free) {
|
|
__blk_put_request(q, free);
|
|
return 1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
bool blk_rq_merge_ok(struct request *rq, struct bio *bio)
|
|
{
|
|
if (!rq_mergeable(rq) || !bio_mergeable(bio))
|
|
return false;
|
|
|
|
if (req_op(rq) != bio_op(bio))
|
|
return false;
|
|
|
|
/* different data direction or already started, don't merge */
|
|
if (bio_data_dir(bio) != rq_data_dir(rq))
|
|
return false;
|
|
|
|
/* must be same device and not a special request */
|
|
if (rq->rq_disk != bio->bi_disk || req_no_special_merge(rq))
|
|
return false;
|
|
|
|
/* don't merge across cgroup boundaries */
|
|
if (!blk_cgroup_mergeable(rq, bio))
|
|
return false;
|
|
|
|
/* only merge integrity protected bio into ditto rq */
|
|
if (blk_integrity_merge_bio(rq->q, rq, bio) == false)
|
|
return false;
|
|
|
|
/* must be using the same buffer */
|
|
if (req_op(rq) == REQ_OP_WRITE_SAME &&
|
|
!blk_write_same_mergeable(rq->bio, bio))
|
|
return false;
|
|
|
|
/*
|
|
* Don't allow merge of different write hints, or for a hint with
|
|
* non-hint IO.
|
|
*/
|
|
if (rq->write_hint != bio->bi_write_hint)
|
|
return false;
|
|
|
|
/* Only merge if the crypt contexts are compatible */
|
|
if (!bio_crypt_ctx_compatible(bio, rq->bio))
|
|
return false;
|
|
|
|
return true;
|
|
}
|
|
|
|
enum elv_merge blk_try_merge(struct request *rq, struct bio *bio)
|
|
{
|
|
if (req_op(rq) == REQ_OP_DISCARD &&
|
|
queue_max_discard_segments(rq->q) > 1)
|
|
return ELEVATOR_DISCARD_MERGE;
|
|
else if (blk_rq_pos(rq) + blk_rq_sectors(rq) == bio->bi_iter.bi_sector)
|
|
return ELEVATOR_BACK_MERGE;
|
|
else if (blk_rq_pos(rq) - bio_sectors(bio) == bio->bi_iter.bi_sector)
|
|
return ELEVATOR_FRONT_MERGE;
|
|
return ELEVATOR_NO_MERGE;
|
|
}
|