Changes in 5.15.46
binfmt_flat: do not stop relocating GOT entries prematurely on riscv
parisc/stifb: Implement fb_is_primary_device()
parisc/stifb: Keep track of hardware path of graphics card
RISC-V: Mark IORESOURCE_EXCLUSIVE for reserved mem instead of IORESOURCE_BUSY
riscv: Initialize thread pointer before calling C functions
riscv: Fix irq_work when SMP is disabled
riscv: Wire up memfd_secret in UAPI header
riscv: Move alternative length validation into subsection
ALSA: hda/realtek - Add new type for ALC245
ALSA: hda/realtek: Enable 4-speaker output for Dell XPS 15 9520 laptop
ALSA: hda/realtek - Fix microphone noise on ASUS TUF B550M-PLUS
ALSA: usb-audio: Cancel pending work at closing a MIDI substream
USB: serial: pl2303: fix type detection for odd device
USB: serial: option: add Quectel BG95 modem
USB: new quirk for Dell Gen 2 devices
usb: isp1760: Fix out-of-bounds array access
usb: dwc3: gadget: Move null pinter check to proper place
usb: core: hcd: Add support for deferring roothub registration
fs/ntfs3: Update valid size if -EIOCBQUEUED
fs/ntfs3: Fix fiemap + fix shrink file size (to remove preallocated space)
fs/ntfs3: Keep preallocated only if option prealloc enabled
fs/ntfs3: Check new size for limits
fs/ntfs3: In function ntfs_set_acl_ex do not change inode->i_mode if called from function ntfs_init_acl
fs/ntfs3: Fix some memory leaks in an error handling path of 'log_replay()'
fs/ntfs3: Update i_ctime when xattr is added
fs/ntfs3: Restore ntfs_xattr_get_acl and ntfs_xattr_set_acl functions
cifs: fix potential double free during failed mount
cifs: when extending a file with falloc we should make files not-sparse
xhci: Allow host runtime PM as default for Intel Alder Lake N xHCI
platform/x86: intel-hid: fix _DSM function index handling
x86/MCE/AMD: Fix memory leak when threshold_create_bank() fails
perf/x86/intel: Fix event constraints for ICL
x86/kexec: fix memory leak of elf header buffer
x86/sgx: Set active memcg prior to shmem allocation
ptrace/um: Replace PT_DTRACE with TIF_SINGLESTEP
ptrace/xtensa: Replace PT_SINGLESTEP with TIF_SINGLESTEP
ptrace: Reimplement PTRACE_KILL by always sending SIGKILL
btrfs: add "0x" prefix for unsupported optional features
btrfs: return correct error number for __extent_writepage_io()
btrfs: repair super block num_devices automatically
btrfs: fix the error handling for submit_extent_page() for btrfs_do_readpage()
iommu/vt-d: Add RPLS to quirk list to skip TE disabling
drm/vmwgfx: validate the screen formats
drm/virtio: fix NULL pointer dereference in virtio_gpu_conn_get_modes
selftests/bpf: Fix vfs_link kprobe definition
selftests/bpf: Fix parsing of prog types in UAPI hdr for bpftool sync
mwifiex: add mutex lock for call in mwifiex_dfs_chan_sw_work_queue
b43legacy: Fix assigning negative value to unsigned variable
b43: Fix assigning negative value to unsigned variable
ipw2x00: Fix potential NULL dereference in libipw_xmit()
ipv6: fix locking issues with loops over idev->addr_list
fbcon: Consistently protect deferred_takeover with console_lock()
x86/platform/uv: Update TSC sync state for UV5
ACPICA: Avoid cache flush inside virtual machines
mac80211: minstrel_ht: fix where rate stats are stored (fixes debugfs output)
drm/komeda: return early if drm_universal_plane_init() fails.
drm/amd/display: Disabling Z10 on DCN31
rcu-tasks: Fix race in schedule and flush work
rcu: Make TASKS_RUDE_RCU select IRQ_WORK
sfc: ef10: Fix assigning negative value to unsigned variable
ALSA: jack: Access input_dev under mutex
rtw88: 8821c: fix debugfs rssi value
spi: spi-rspi: Remove setting {src,dst}_{addr,addr_width} based on DMA direction
tools/power turbostat: fix ICX DRAM power numbers
scsi: lpfc: Move cfg_log_verbose check before calling lpfc_dmp_dbg()
scsi: lpfc: Fix SCSI I/O completion and abort handler deadlock
scsi: lpfc: Fix call trace observed during I/O with CMF enabled
cpuidle: PSCI: Improve support for suspend-to-RAM for PSCI OSI mode
drm/amd/pm: fix double free in si_parse_power_table()
ASoC: rsnd: care default case on rsnd_ssiu_busif_err_status_clear()
ASoC: rsnd: care return value from rsnd_node_fixed_index()
ath9k: fix QCA9561 PA bias level
media: venus: hfi: avoid null dereference in deinit
media: pci: cx23885: Fix the error handling in cx23885_initdev()
media: cx25821: Fix the warning when removing the module
md/bitmap: don't set sb values if can't pass sanity check
mmc: jz4740: Apply DMA engine limits to maximum segment size
drivers: mmc: sdhci_am654: Add the quirk to set TESTCD bit
scsi: megaraid: Fix error check return value of register_chrdev()
drm/amdgpu/sdma: Fix incorrect calculations of the wptr of the doorbells
scsi: ufs: Use pm_runtime_resume_and_get() instead of pm_runtime_get_sync()
scsi: lpfc: Fix resource leak in lpfc_sli4_send_seq_to_ulp()
ath11k: disable spectral scan during spectral deinit
ASoC: Intel: bytcr_rt5640: Add quirk for the HP Pro Tablet 408
drm/plane: Move range check for format_count earlier
drm/amd/pm: fix the compile warning
ath10k: skip ath10k_halt during suspend for driver state RESTARTING
arm64: compat: Do not treat syscall number as ESR_ELx for a bad syscall
drm: msm: fix error check return value of irq_of_parse_and_map()
scsi: target: tcmu: Fix possible data corruption
ipv6: Don't send rs packets to the interface of ARPHRD_TUNNEL
net/mlx5: fs, delete the FTE when there are no rules attached to it
ASoC: dapm: Don't fold register value changes into notifications
mlxsw: spectrum_dcb: Do not warn about priority changes
mlxsw: Treat LLDP packets as control
drm/amdgpu/psp: move PSP memory alloc from hw_init to sw_init
drm/amdgpu/ucode: Remove firmware load type check in amdgpu_ucode_free_bo
regulator: mt6315: Enforce regulator-compatible, not name
HID: bigben: fix slab-out-of-bounds Write in bigben_probe
of: Support more than one crash kernel regions for kexec -s
ASoC: tscs454: Add endianness flag in snd_soc_component_driver
scsi: lpfc: Alter FPIN stat accounting logic
net: remove two BUG() from skb_checksum_help()
s390/preempt: disable __preempt_count_add() optimization for PROFILE_ALL_BRANCHES
perf/amd/ibs: Cascade pmu init functions' return value
sched/core: Avoid obvious double update_rq_clock warning
spi: stm32-qspi: Fix wait_cmd timeout in APM mode
dma-debug: change allocation mode from GFP_NOWAIT to GFP_ATIOMIC
ACPI: PM: Block ASUS B1400CEAE from suspend to idle by default
ipmi:ssif: Check for NULL msg when handling events and messages
ipmi: Fix pr_fmt to avoid compilation issues
rtlwifi: Use pr_warn instead of WARN_ONCE
mt76: mt7921: accept rx frames with non-standard VHT MCS10-11
mt76: fix encap offload ethernet type check
media: rga: fix possible memory leak in rga_probe
media: coda: limit frame interval enumeration to supported encoder frame sizes
media: hantro: HEVC: unconditionnaly set pps_{cb/cr}_qp_offset values
media: ccs-core.c: fix failure to call clk_disable_unprepare
media: imon: reorganize serialization
media: cec-adap.c: fix is_configuring state
usbnet: Run unregister_netdev() before unbind() again
openrisc: start CPU timer early in boot
nvme-pci: fix a NULL pointer dereference in nvme_alloc_admin_tags
ASoC: rt5645: Fix errorenous cleanup order
nbd: Fix hung on disconnect request if socket is closed before
drm/amd/pm: update smartshift powerboost calc for smu12
drm/amd/pm: update smartshift powerboost calc for smu13
net: phy: micrel: Allow probing without .driver_data
media: exynos4-is: Fix compile warning
media: hantro: Stop using H.264 parameter pic_num
ASoC: max98357a: remove dependency on GPIOLIB
ASoC: rt1015p: remove dependency on GPIOLIB
ACPI: CPPC: Assume no transition latency if no PCCT
nvme: set non-mdts limits in nvme_scan_work
can: mcp251xfd: silence clang's -Wunaligned-access warning
x86/microcode: Add explicit CPU vendor dependency
net: ipa: ignore endianness if there is no header
m68k: atari: Make Atari ROM port I/O write macros return void
rxrpc: Return an error to sendmsg if call failed
rxrpc, afs: Fix selection of abort codes
afs: Adjust ACK interpretation to try and cope with NAT
eth: tg3: silence the GCC 12 array-bounds warning
char: tpm: cr50_i2c: Suppress duplicated error message in .remove()
selftests/bpf: fix btf_dump/btf_dump due to recent clang change
gfs2: use i_lock spin_lock for inode qadata
scsi: target: tcmu: Avoid holding XArray lock when calling lock_page
IB/rdmavt: add missing locks in rvt_ruc_loopback
ARM: dts: ox820: align interrupt controller node name with dtschema
ARM: dts: socfpga: align interrupt controller node name with dtschema
ARM: dts: s5pv210: align DMA channels with dtschema
arm64: dts: qcom: msm8994: Fix the cont_splash_mem address
arm64: dts: qcom: msm8994: Fix BLSP[12]_DMA channels count
PM / devfreq: rk3399_dmc: Disable edev on remove()
crypto: ccree - use fine grained DMA mapping dir
soc: ti: ti_sci_pm_domains: Check for null return of devm_kcalloc
fs: jfs: fix possible NULL pointer dereference in dbFree()
arm64: dts: qcom: sdm845-xiaomi-beryllium: fix typo in panel's vddio-supply property
ALSA: usb-audio: Add quirk bits for enabling/disabling generic implicit fb
ALSA: usb-audio: Move generic implicit fb quirk entries into quirks.c
ARM: OMAP1: clock: Fix UART rate reporting algorithm
powerpc/fadump: Fix fadump to work with a different endian capture kernel
fat: add ratelimit to fat*_ent_bread()
pinctrl: renesas: rzn1: Fix possible null-ptr-deref in sh_pfc_map_resources()
ARM: versatile: Add missing of_node_put in dcscb_init
ARM: dts: exynos: add atmel,24c128 fallback to Samsung EEPROM
ARM: hisi: Add missing of_node_put after of_find_compatible_node
cpufreq: Avoid unnecessary frequency updates due to mismatch
powerpc/rtas: Keep MSR[RI] set when calling RTAS
PCI: Avoid pci_dev_lock() AB/BA deadlock with sriov_numvfs_store()
KVM: PPC: Book3S HV Nested: L2 LPCR should inherit L1 LPES setting
alpha: fix alloc_zeroed_user_highpage_movable()
tracing: incorrect isolate_mote_t cast in mm_vmscan_lru_isolate
powerpc/powernv/vas: Assign real address to rx_fifo in vas_rx_win_attr
powerpc/xics: fix refcount leak in icp_opal_init()
powerpc/powernv: fix missing of_node_put in uv_init()
macintosh/via-pmu: Fix build failure when CONFIG_INPUT is disabled
powerpc/iommu: Add missing of_node_put in iommu_init_early_dart
smb3: check for null tcon
RDMA/hfi1: Prevent panic when SDMA is disabled
Input: gpio-keys - cancel delayed work only in case of GPIO
drm: fix EDID struct for old ARM OABI format
drm/bridge_connector: enable HPD by default if supported
dt-bindings: display: sitronix, st7735r: Fix backlight in example
drm/vmwgfx: Fix an invalid read
ath11k: acquire ab->base_lock in unassign when finding the peer by addr
drm: bridge: it66121: Fix the register page length
ath9k: fix ar9003_get_eepmisc
drm/edid: fix invalid EDID extension block filtering
drm/bridge: adv7511: clean up CEC adapter when probe fails
drm: bridge: icn6211: Fix register layout
drm: bridge: icn6211: Fix HFP_HSW_HBP_HI and HFP_MIN handling
mtd: spinand: gigadevice: fix Quad IO for GD5F1GQ5UExxG
spi: qcom-qspi: Add minItems to interconnect-names
ASoC: mediatek: Fix error handling in mt8173_max98090_dev_probe
ASoC: mediatek: Fix missing of_node_put in mt2701_wm8960_machine_probe
x86/delay: Fix the wrong asm constraint in delay_loop()
drm/vc4: hvs: Fix frame count register readout
drm/mediatek: Fix mtk_cec_mask()
drm/vc4: hvs: Reset muxes at probe time
drm/vc4: txp: Don't set TXP_VSTART_AT_EOF
drm/vc4: txp: Force alpha to be 0xff if it's disabled
libbpf: Don't error out on CO-RE relos for overriden weak subprogs
x86/PCI: Fix ALi M1487 (IBC) PIRQ router link value interpretation
mptcp: reset the packet scheduler on PRIO change
nl80211: show SSID for P2P_GO interfaces
drm/komeda: Fix an undefined behavior bug in komeda_plane_add()
drm: mali-dp: potential dereference of null pointer
spi: spi-ti-qspi: Fix return value handling of wait_for_completion_timeout
scftorture: Fix distribution of short handler delays
net: dsa: mt7530: 1G can also support 1000BASE-X link mode
ixp4xx_eth: fix error check return value of platform_get_irq()
NFC: NULL out the dev->rfkill to prevent UAF
efi: Add missing prototype for efi_capsule_setup_info
device property: Check fwnode->secondary when finding properties
device property: Allow error pointer to be passed to fwnode APIs
target: remove an incorrect unmap zeroes data deduction
drbd: fix duplicate array initializer
EDAC/dmc520: Don't print an error for each unconfigured interrupt line
mtd: rawnand: denali: Use managed device resources
HID: hid-led: fix maximum brightness for Dream Cheeky
HID: elan: Fix potential double free in elan_input_configured
drm/bridge: Fix error handling in analogix_dp_probe
regulator: da9121: Fix uninit-value in da9121_assign_chip_model()
drm/mediatek: dpi: Use mt8183 output formats for mt8192
signal: Deliver SIGTRAP on perf event asynchronously if blocked
sched/fair: Fix cfs_rq_clock_pelt() for throttled cfs_rq
sched/psi: report zeroes for CPU full at the system level
spi: img-spfi: Fix pm_runtime_get_sync() error checking
cpufreq: Fix possible race in cpufreq online error path
printk: use atomic updates for klogd work
printk: add missing memory barrier to wake_up_klogd()
printk: wake waiters for safe and NMI contexts
ath9k_htc: fix potential out of bounds access with invalid rxstatus->rs_keyix
media: i2c: max9286: Use dev_err_probe() helper
media: i2c: max9286: Use "maxim,gpio-poc" property
media: i2c: max9286: fix kernel oops when removing module
media: hantro: Empty encoder capture buffers by default
drm/panel: simple: Add missing bus flags for Innolux G070Y2-L01
ALSA: pcm: Check for null pointer of pointer substream before dereferencing it
mtdblock: warn if opened on NAND
inotify: show inotify mask flags in proc fdinfo
fsnotify: fix wrong lockdep annotations
spi: rockchip: Stop spi slave dma receiver when cs inactive
spi: rockchip: Preset cs-high and clk polarity in setup progress
spi: rockchip: fix missing error on unsupported SPI_CS_HIGH
of: overlay: do not break notify on NOTIFY_{OK|STOP}
selftests/damon: add damon to selftests root Makefile
drm/msm/dp: Modify prototype of encoder based API
drm/msm/hdmi: switch to drm_bridge_connector
drm/msm/dpu: adjust display_v_end for eDP and DP
scsi: iscsi: Fix harmless double shift bug
scsi: ufs: qcom: Fix ufs_qcom_resume()
scsi: ufs: core: Exclude UECxx from SFR dump list
drm/v3d: Fix null pointer dereference of pointer perfmon
selftests/resctrl: Fix null pointer dereference on open failed
libbpf: Fix logic for finding matching program for CO-RE relocation
mtd: spi-nor: core: Check written SR value in spi_nor_write_16bit_sr_and_check()
x86/pm: Fix false positive kmemleak report in msr_build_context()
mtd: rawnand: cadence: fix possible null-ptr-deref in cadence_nand_dt_probe()
mtd: rawnand: intel: fix possible null-ptr-deref in ebu_nand_probe()
x86/speculation: Add missing prototype for unpriv_ebpf_notify()
ASoC: rk3328: fix disabling mclk on pclk probe failure
perf tools: Add missing headers needed by util/data.h
drm/msm/disp/dpu1: set vbif hw config to NULL to avoid use after memory free during pm runtime resume
drm/msm/dp: stop event kernel thread when DP unbind
drm/msm/dp: fix error check return value of irq_of_parse_and_map()
drm/msm/dp: reset DP controller before transmit phy test pattern
drm/msm/dp: do not stop transmitting phy test pattern during DP phy compliance test
drm/msm/dsi: fix error checks and return values for DSI xmit functions
drm/msm/hdmi: check return value after calling platform_get_resource_byname()
drm/msm/hdmi: fix error check return value of irq_of_parse_and_map()
drm/msm: add missing include to msm_drv.c
drm/panel: panel-simple: Fix proper bpc for AM-1280800N3TZQW-T00H
kunit: fix debugfs code to use enum kunit_status, not bool
drm/rockchip: vop: fix possible null-ptr-deref in vop_bind()
spi: cadence-quadspi: fix Direct Access Mode disable for SoCFPGA
perf tools: Use Python devtools for version autodetection rather than runtime
virtio_blk: fix the discard_granularity and discard_alignment queue limits
nl80211: don't hold RTNL in color change request
x86: Fix return value of __setup handlers
irqchip/exiu: Fix acknowledgment of edge triggered interrupts
irqchip/aspeed-i2c-ic: Fix irq_of_parse_and_map() return value
irqchip/aspeed-scu-ic: Fix irq_of_parse_and_map() return value
x86/mm: Cleanup the control_va_addr_alignment() __setup handler
arm64: fix types in copy_highpage()
regulator: core: Fix enable_count imbalance with EXCLUSIVE_GET
drm/msm/dsi: fix address for second DSI PHY on SDM660
drm/msm/dp: fix event thread stuck in wait_event after kthread_stop()
drm/msm/mdp5: Return error code in mdp5_pipe_release when deadlock is detected
drm/msm/mdp5: Return error code in mdp5_mixer_release when deadlock is detected
drm/msm: return an error pointer in msm_gem_prime_get_sg_table()
media: uvcvideo: Fix missing check to determine if element is found in list
arm64: stackleak: fix current_top_of_stack()
iomap: iomap_write_failed fix
spi: spi-fsl-qspi: check return value after calling platform_get_resource_byname()
Revert "cpufreq: Fix possible race in cpufreq online error path"
regulator: qcom_smd: Fix up PM8950 regulator configuration
samples: bpf: Don't fail for a missing VMLINUX_BTF when VMLINUX_H is provided
perf/amd/ibs: Use interrupt regs ip for stack unwinding
ath11k: Don't check arvif->is_started before sending management frames
wilc1000: fix crash observed in AP mode with cfg80211_register_netdevice()
HID: amd_sfh: Modify the bus name
HID: amd_sfh: Modify the hid name
ASoC: fsl: Use dev_err_probe() helper
ASoC: fsl: Fix refcount leak in imx_sgtl5000_probe
ASoC: imx-hdmi: Fix refcount leak in imx_hdmi_probe
ASoC: mxs-saif: Fix refcount leak in mxs_saif_probe
regulator: pfuze100: Fix refcount leak in pfuze_parse_regulators_dt
dma-direct: factor out a helper for DMA_ATTR_NO_KERNEL_MAPPING allocations
dma-direct: don't fail on highmem CMA pages in dma_direct_alloc_pages
ASoC: samsung: Use dev_err_probe() helper
ASoC: samsung: Fix refcount leak in aries_audio_probe
block: Fix the bio.bi_opf comment
kselftest/cgroup: fix test_stress.sh to use OUTPUT dir
scripts/faddr2line: Fix overlapping text section failures
media: aspeed: Fix an error handling path in aspeed_video_probe()
media: exynos4-is: Fix PM disable depth imbalance in fimc_is_probe
mt76: mt7921: Fix the error handling path of mt7921_pci_probe()
mt76: do not attempt to reorder received 802.3 packets without agg session
media: st-delta: Fix PM disable depth imbalance in delta_probe
media: atmel: atmel-isc: Fix PM disable depth imbalance in atmel_isc_probe
media: i2c: rdacm2x: properly set subdev entity function
media: exynos4-is: Change clk_disable to clk_disable_unprepare
media: pvrusb2: fix array-index-out-of-bounds in pvr2_i2c_core_init
media: vsp1: Fix offset calculation for plane cropping
media: atmel: atmel-sama5d2-isc: fix wrong mask in YUYV format check
media: hantro: HEVC: Fix tile info buffer value computation
Bluetooth: fix dangling sco_conn and use-after-free in sco_sock_timeout
Bluetooth: use hdev lock in activate_scan for hci_is_adv_monitoring
Bluetooth: use hdev lock for accept_list and reject_list in conn req
nvme: set dma alignment to dword
m68k: math-emu: Fix dependencies of math emulation support
sctp: read sk->sk_bound_dev_if once in sctp_rcv()
net: hinic: add missing destroy_workqueue in hinic_pf_to_mgmt_init
ASoC: ti: j721e-evm: Fix refcount leak in j721e_soc_probe_*
kselftest/arm64: bti: force static linking
media: ov7670: remove ov7670_power_off from ov7670_remove
media: i2c: ov5648: fix wrong pointer passed to IS_ERR() and PTR_ERR()
media: staging: media: rkvdec: Make use of the helper function devm_platform_ioremap_resource()
media: rkvdec: h264: Fix dpb_valid implementation
media: rkvdec: h264: Fix bit depth wrap in pps packet
regulator: scmi: Fix refcount leak in scmi_regulator_probe
ext4: reject the 'commit' option on ext2 filesystems
drm/msm/a6xx: Fix refcount leak in a6xx_gpu_init
drm: msm: fix possible memory leak in mdp5_crtc_cursor_set()
x86/sev: Annotate stack change in the #VC handler
drm/msm: don't free the IRQ if it was not requested
selftests/bpf: Add missed ima_setup.sh in Makefile
drm/msm/dpu: handle pm_runtime_get_sync() errors in bind path
drm/i915: Fix CFI violation with show_dynamic_id()
thermal/drivers/bcm2711: Don't clamp temperature at zero
thermal/drivers/broadcom: Fix potential NULL dereference in sr_thermal_probe
thermal/core: Fix memory leak in __thermal_cooling_device_register()
thermal/drivers/imx_sc_thermal: Fix refcount leak in imx_sc_thermal_probe
bfq: Relax waker detection for shared queues
bfq: Allow current waker to defend against a tentative one
ASoC: wm2000: fix missing clk_disable_unprepare() on error in wm2000_anc_transition()
PM: domains: Fix initialization of genpd's next_wakeup
net: macb: Fix PTP one step sync support
NFC: hci: fix sleep in atomic context bugs in nfc_hci_hcp_message_tx
ASoC: max98090: Move check for invalid values before casting in max98090_put_enab_tlv()
net: stmmac: selftests: Use kcalloc() instead of kzalloc()
net: stmmac: fix out-of-bounds access in a selftest
hv_netvsc: Fix potential dereference of NULL pointer
hwmon: (pmbus) Check PEC support before reading other registers
rxrpc: Fix listen() setting the bar too high for the prealloc rings
rxrpc: Don't try to resend the request if we're receiving the reply
rxrpc: Fix overlapping ACK accounting
rxrpc: Don't let ack.previousPacket regress
rxrpc: Fix decision on when to generate an IDLE ACK
net: huawei: hinic: Use devm_kcalloc() instead of devm_kzalloc()
hinic: Avoid some over memory allocation
net: dsa: restrict SMSC_LAN9303_I2C kconfig
net/smc: postpone sk_refcnt increment in connect()
dma-direct: factor out dma_set_{de,en}crypted helpers
dma-direct: don't call dma_set_decrypted for remapped allocations
dma-direct: always leak memory that can't be re-encrypted
dma-direct: don't over-decrypt memory
arm64: dts: rockchip: Move drive-impedance-ohm to emmc phy on rk3399
arm64: dts: mt8192: Fix nor_flash status disable typo
PCI/ACPI: Allow D3 only if Root Port can signal and wake from D3
memory: samsung: exynos5422-dmc: Avoid some over memory allocation
ARM: dts: BCM5301X: update CRU block description
ARM: dts: BCM5301X: Update pin controller node name
ARM: dts: suniv: F1C100: fix watchdog compatible
soc: qcom: smp2p: Fix missing of_node_put() in smp2p_parse_ipc
soc: qcom: smsm: Fix missing of_node_put() in smsm_parse_ipc
PCI: cadence: Fix find_first_zero_bit() limit
PCI: rockchip: Fix find_first_zero_bit() limit
PCI: mediatek: Fix refcount leak in mtk_pcie_subsys_powerup()
PCI: dwc: Fix setting error return on MSI DMA mapping failure
ARM: dts: ci4x10: Adapt to changes in imx6qdl.dtsi regarding fec clocks
soc: qcom: llcc: Add MODULE_DEVICE_TABLE()
KVM: nVMX: Leave most VM-Exit info fields unmodified on failed VM-Entry
KVM: nVMX: Clear IDT vectoring on nested VM-Exit for double/triple fault
crypto: qat - set CIPHER capability for QAT GEN2
crypto: qat - set COMPRESSION capability for QAT GEN2
crypto: qat - set CIPHER capability for DH895XCC
crypto: qat - set COMPRESSION capability for DH895XCC
platform/chrome: cros_ec: fix error handling in cros_ec_register()
ARM: dts: imx6dl-colibri: Fix I2C pinmuxing
platform/chrome: Re-introduce cros_ec_cmd_xfer and use it for ioctls
can: xilinx_can: mark bit timing constants as const
ARM: dts: stm32: Fix PHY post-reset delay on Avenger96
ARM: dts: bcm2835-rpi-zero-w: Fix GPIO line name for Wifi/BT
ARM: dts: bcm2837-rpi-cm3-io3: Fix GPIO line names for SMPS I2C
ARM: dts: bcm2837-rpi-3-b-plus: Fix GPIO line name of power LED
ARM: dts: bcm2835-rpi-b: Fix GPIO line names
misc: ocxl: fix possible double free in ocxl_file_register_afu
crypto: marvell/cesa - ECB does not IV
gpiolib: of: Introduce hook for missing gpio-ranges
pinctrl: bcm2835: implement hook for missing gpio-ranges
arm: mediatek: select arch timer for mt7629
pinctrl/rockchip: support deferring other gpio params
pinctrl: mediatek: mt8195: enable driver on mtk platforms
arm64: dts: qcom: qrb5165-rb5: Fix can-clock node name
Drivers: hv: vmbus: Fix handling of messages with transaction ID of zero
powerpc/fadump: fix PT_LOAD segment for boot memory area
mfd: ipaq-micro: Fix error check return value of platform_get_irq()
scsi: fcoe: Fix Wstringop-overflow warnings in fcoe_wwn_from_mac()
soc: bcm: Check for NULL return of devm_kzalloc()
arm64: dts: ti: k3-am64-mcu: remove incorrect UART base clock rates
ASoC: sh: rz-ssi: Check return value of pm_runtime_resume_and_get()
ASoC: sh: rz-ssi: Propagate error codes returned from platform_get_irq_byname()
ASoC: sh: rz-ssi: Release the DMA channels in rz_ssi_probe() error path
firmware: arm_scmi: Fix list protocols enumeration in the base protocol
nvdimm: Fix firmware activation deadlock scenarios
nvdimm: Allow overwrite in the presence of disabled dimms
pinctrl: mvebu: Fix irq_of_parse_and_map() return value
drivers/base/node.c: fix compaction sysfs file leak
dax: fix cache flush on PMD-mapped pages
drivers/base/memory: fix an unlikely reference counting issue in __add_memory_block()
firmware: arm_ffa: Fix uuid parameter to ffa_partition_probe
firmware: arm_ffa: Remove incorrect assignment of driver_data
list: introduce list_is_head() helper and re-use it in list.h
list: fix a data-race around ep->rdllist
drm/msm/dpu: fix error check return value of irq_of_parse_and_map()
powerpc/8xx: export 'cpm_setbrg' for modules
pinctrl: renesas: r8a779a0: Fix GPIO function on I2C-capable pins
pinctrl: renesas: core: Fix possible null-ptr-deref in sh_pfc_map_resources()
powerpc/idle: Fix return value of __setup() handler
powerpc/4xx/cpm: Fix return value of __setup() handler
RDMA/hns: Add the detection for CMDQ status in the device initialization process
arm64: dts: marvell: espressobin-ultra: fix SPI-NOR config
arm64: dts: marvell: espressobin-ultra: enable front USB3 port
ASoC: atmel-pdmic: Remove endianness flag on pdmic component
ASoC: atmel-classd: Remove endianness flag on class d component
proc: fix dentry/inode overinstantiating under /proc/${pid}/net
ipc/mqueue: use get_tree_nodev() in mqueue_get_tree()
PCI: imx6: Fix PERST# start-up sequence
tty: fix deadlock caused by calling printk() under tty_port->lock
crypto: sun8i-ss - rework handling of IV
crypto: sun8i-ss - handle zero sized sg
crypto: cryptd - Protect per-CPU resource by disabling BH.
ARM: dts: at91: sama7g5: remove interrupt-parent from gic node
hugetlbfs: fix hugetlbfs_statfs() locking
Input: sparcspkr - fix refcount leak in bbc_beep_probe
PCI/AER: Clear MULTI_ERR_COR/UNCOR_RCV bits
PCI: microchip: Fix potential race in interrupt handling
hwrng: omap3-rom - fix using wrong clk_disable() in omap_rom_rng_runtime_resume()
powerpc/64: Only WARN if __pa()/__va() called with bad addresses
powerpc/perf: Fix the threshold compare group constraint for power10
powerpc/perf: Fix the threshold compare group constraint for power9
macintosh: via-pmu and via-cuda need RTC_LIB
powerpc/xive: Add some error handling code to 'xive_spapr_init()'
powerpc/xive: Fix refcount leak in xive_spapr_init
powerpc/fsl_rio: Fix refcount leak in fsl_rio_setup
mfd: davinci_voicecodec: Fix possible null-ptr-deref davinci_vc_probe()
nfsd: destroy percpu stats counters after reply cache shutdown
mailbox: forward the hrtimer if not queued and under a lock
RDMA/hfi1: Prevent use of lock before it is initialized
KVM: LAPIC: Drop pending LAPIC timer injection when canceling the timer
Input: stmfts - do not leave device disabled in stmfts_input_open
OPP: call of_node_put() on error path in _bandwidth_supported()
f2fs: support fault injection for dquot_initialize()
f2fs: fix to do sanity check on inline_dots inode
f2fs: fix dereference of stale list iterator after loop body
iommu/amd: Enable swiotlb in all cases
iommu/mediatek: Fix 2 HW sharing pgtable issue
iommu/mediatek: Add list_del in mtk_iommu_remove
iommu/mediatek: Remove clk_disable in mtk_iommu_remove
iommu/mediatek: Add mutex for m4u_group and m4u_dom in data
i2c: at91: use dma safe buffers
cpufreq: mediatek: Use module_init and add module_exit
cpufreq: mediatek: Unregister platform device on exit
iommu/arm-smmu-v3-sva: Fix mm use-after-free
MIPS: Loongson: Use hwmon_device_register_with_groups() to register hwmon
iommu/mediatek: Fix NULL pointer dereference when printing dev_name
i2c: at91: Initialize dma_buf in at91_twi_xfer()
dmaengine: idxd: Fix the error handling path in idxd_cdev_register()
NFS: Do not report EINTR/ERESTARTSYS as mapping errors
NFS: fsync() should report filesystem errors over EINTR/ERESTARTSYS
NFS: Don't report ENOSPC write errors twice
NFS: Do not report flush errors in nfs_write_end()
NFS: Don't report errors from nfs_pageio_complete() more than once
NFSv4/pNFS: Do not fail I/O when we fail to allocate the pNFS layout
NFS: Further fixes to the writeback error handling
video: fbdev: clcdfb: Fix refcount leak in clcdfb_of_vram_setup
dmaengine: stm32-mdma: remove GISR1 register
dmaengine: stm32-mdma: fix chan initialization in stm32_mdma_irq_handler()
iommu/amd: Increase timeout waiting for GA log enablement
i2c: npcm: Fix timeout calculation
i2c: npcm: Correct register access width
i2c: npcm: Handle spurious interrupts
i2c: rcar: fix PM ref counts in probe error paths
perf build: Fix btf__load_from_kernel_by_id() feature check
perf c2c: Use stdio interface if slang is not supported
perf jevents: Fix event syntax error caused by ExtSel
video: fbdev: vesafb: Fix a use-after-free due early fb_info cleanup
NFS: Always initialise fattr->label in nfs_fattr_alloc()
NFS: Create a new nfs_alloc_fattr_with_label() function
NFS: Convert GFP_NOFS to GFP_KERNEL
NFSv4.1 mark qualified async operations as MOVEABLE tasks
f2fs: fix to avoid f2fs_bug_on() in dec_valid_node_count()
f2fs: fix to do sanity check on block address in f2fs_do_zero_range()
f2fs: fix to clear dirty inode in f2fs_evict_inode()
f2fs: fix deadloop in foreground GC
f2fs: don't need inode lock for system hidden quota
f2fs: fix to do sanity check on total_data_blocks
f2fs: don't use casefolded comparison for "." and ".."
f2fs: fix fallocate to use file_modified to update permissions consistently
f2fs: fix to do sanity check for inline inode
objtool: Fix objtool regression on x32 systems
objtool: Fix symbol creation
wifi: mac80211: fix use-after-free in chanctx code
iwlwifi: mvm: fix assert 1F04 upon reconfig
fs-writeback: writeback_sb_inodes:Recalculate 'wrote' according skipped pages
efi: Do not import certificates from UEFI Secure Boot for T2 Macs
bfq: Avoid false marking of bic as stably merged
bfq: Avoid merging queues with different parents
bfq: Split shared queues on move between cgroups
bfq: Update cgroup information before merging bio
bfq: Drop pointless unlock-lock pair
bfq: Remove pointless bfq_init_rq() calls
bfq: Track whether bfq_group is still online
bfq: Get rid of __bio_blkcg() usage
bfq: Make sure bfqg for which we are queueing requests is online
ext4: mark group as trimmed only if it was fully scanned
ext4: fix use-after-free in ext4_rename_dir_prepare
ext4: fix race condition between ext4_write and ext4_convert_inline_data
ext4: fix warning in ext4_handle_inode_extension
ext4: fix bug_on in ext4_writepages
ext4: filter out EXT4_FC_REPLAY from on-disk superblock field s_state
ext4: fix bug_on in __es_tree_search
ext4: verify dir block before splitting it
ext4: avoid cycles in directory h-tree
ACPI: property: Release subnode properties with data nodes
tty: goldfish: Introduce gf_ioread32()/gf_iowrite32()
tracing: Fix potential double free in create_var_ref()
tracing: Initialize integer variable to prevent garbage return value
drm/amdgpu: add beige goby PCI ID
PCI/PM: Fix bridge_d3_blacklist[] Elo i2 overwrite of Gigabyte X299
PCI: qcom: Fix runtime PM imbalance on probe errors
PCI: qcom: Fix unbalanced PHY init on probe errors
staging: r8188eu: prevent ->Ssid overflow in rtw_wx_set_scan()
mm, compaction: fast_find_migrateblock() should return pfn in the target zone
s390/perf: obtain sie_block from the right address
s390/stp: clock_delta should be signed
dlm: fix plock invalid read
dlm: uninitialized variable on error in dlm_listen_for_all()
dlm: fix missing lkb refcount handling
ocfs2: dlmfs: fix error handling of user_dlm_destroy_lock
scsi: dc395x: Fix a missing check on list iterator
scsi: ufs: qcom: Add a readl() to make sure ref_clk gets enabled
landlock: Add clang-format exceptions
landlock: Format with clang-format
selftests/landlock: Add clang-format exceptions
selftests/landlock: Normalize array assignment
selftests/landlock: Format with clang-format
samples/landlock: Add clang-format exceptions
samples/landlock: Format with clang-format
landlock: Fix landlock_add_rule(2) documentation
selftests/landlock: Make tests build with old libc
selftests/landlock: Extend tests for minimal valid attribute size
selftests/landlock: Add tests for unknown access rights
selftests/landlock: Extend access right tests to directories
selftests/landlock: Fully test file rename with "remove" access
selftests/landlock: Add tests for O_PATH
landlock: Change landlock_add_rule(2) argument check ordering
landlock: Change landlock_restrict_self(2) check ordering
selftests/landlock: Test landlock_create_ruleset(2) argument check ordering
landlock: Define access_mask_t to enforce a consistent access mask size
landlock: Reduce the maximum number of layers to 16
landlock: Create find_rule() from unmask_layers()
landlock: Fix same-layer rule unions
drm/amdgpu/cs: make commands with 0 chunks illegal behaviour.
drm/nouveau/subdev/bus: Ratelimit logging for fault errors
drm/etnaviv: check for reaped mapping in etnaviv_iommu_unmap_gem
drm/nouveau/clk: Fix an incorrect NULL check on list iterator
drm/nouveau/kms/nv50-: atom: fix an incorrect NULL check on list iterator
drm/bridge: analogix_dp: Grab runtime PM reference for DP-AUX
drm/i915/dsi: fix VBT send packet port selection for ICL+
md: fix an incorrect NULL check in does_sb_need_changing
md: fix an incorrect NULL check in md_reload_sb
mtd: cfi_cmdset_0002: Move and rename chip_check/chip_ready/chip_good_for_write
mtd: cfi_cmdset_0002: Use chip_ready() for write on S29GL064N
media: coda: Fix reported H264 profile
media: coda: Add more H264 levels for CODA960
ima: remove the IMA_TEMPLATE Kconfig option
Kconfig: Add option for asm goto w/ tied outputs to workaround clang-13 bug
RDMA/hfi1: Fix potential integer multiplication overflow errors
mmc: core: Allows to override the timeout value for ioctl() path
csky: patch_text: Fixup last cpu should be master
irqchip/armada-370-xp: Do not touch Performance Counter Overflow on A375, A38x, A39x
irqchip: irq-xtensa-mx: fix initial IRQ affinity
thermal: devfreq_cooling: use local ops instead of global ops
cfg80211: declare MODULE_FIRMWARE for regulatory.db
mac80211: upgrade passive scan to active scan on DFS channels after beacon rx
um: Use asm-generic/dma-mapping.h
um: chan_user: Fix winch_tramp() return value
um: Fix out-of-bounds read in LDT setup
kexec_file: drop weak attribute from arch_kexec_apply_relocations[_add]
ftrace: Clean up hash direct_functions on register failures
ksmbd: fix outstanding credits related bugs
iommu/msm: Fix an incorrect NULL check on list iterator
iommu/dma: Fix iova map result check bug
Revert "mm/cma.c: remove redundant cma_mutex lock"
mm/page_alloc: always attempt to allocate at least one page during bulk allocation
nodemask.h: fix compilation error with GCC12
hugetlb: fix huge_pmd_unshare address update
mm/memremap: fix missing call to untrack_pfn() in pagemap_range()
xtensa/simdisk: fix proc_read_simdisk()
rtl818x: Prevent using not initialized queues
ASoC: rt5514: Fix event generation for "DSP Voice Wake Up" control
carl9170: tx: fix an incorrect use of list iterator
stm: ltdc: fix two incorrect NULL checks on list iterator
bcache: improve multithreaded bch_btree_check()
bcache: improve multithreaded bch_sectors_dirty_init()
bcache: remove incremental dirty sector counting for bch_sectors_dirty_init()
bcache: avoid journal no-space deadlock by reserving 1 journal bucket
serial: pch: don't overwrite xmit->buf[0] by x_char
tilcdc: tilcdc_external: fix an incorrect NULL check on list iterator
gma500: fix an incorrect NULL check on list iterator
arm64: dts: qcom: ipq8074: fix the sleep clock frequency
arm64: tegra: Add missing DFLL reset on Tegra210
clk: tegra: Add missing reset deassertion
phy: qcom-qmp: fix struct clk leak on probe errors
ARM: dts: s5pv210: Remove spi-cs-high on panel in Aries
ARM: pxa: maybe fix gpio lookup tables
SMB3: EBADF/EIO errors in rename/open caused by race condition in smb2_compound_op
docs/conf.py: Cope with removal of language=None in Sphinx 5.0.0
dt-bindings: gpio: altera: correct interrupt-cells
vdpasim: allow to enable a vq repeatedly
blk-iolatency: Fix inflight count imbalances and IO hangs on offline
coresight: core: Fix coresight device probe failure issue
phy: qcom-qmp: fix reset-controller leak on probe errors
net: ipa: fix page free in ipa_endpoint_trans_release()
net: ipa: fix page free in ipa_endpoint_replenish_one()
kseltest/cgroup: Make test_stress.sh work if run interactively
list: test: Add a test for list_is_head()
Revert "random: use static branch for crng_ready()"
staging: r8188eu: delete rtw_wx_read/write32()
RDMA/hns: Remove the num_cqc_timer variable
RDMA/rxe: Generate a completion for unsupported/invalid opcode
MIPS: IP27: Remove incorrect `cpu_has_fpu' override
MIPS: IP30: Remove incorrect `cpu_has_fpu' override
ext4: only allow test_dummy_encryption when supported
interconnect: qcom: sc7180: Drop IP0 interconnects
interconnect: qcom: icc-rpmh: Add BCMs to commit list in pre_aggregate
fs: add two trivial lookup helpers
exportfs: support idmapped mounts
fs/ntfs3: Fix invalid free in log_replay
md: Don't set mddev private to NULL in raid0 pers->free
md: fix double free of io_acct_set bioset
md: bcache: check the return value of kzalloc() in detached_dev_do_request()
pinctrl/rockchip: support setting input-enable param
block: fix bio_clone_blkg_association() to associate with proper blkcg_gq
Linux 5.15.46
Signed-off-by: Greg Kroah-Hartman <gregkh@google.com>
Change-Id: I7b65df29c22a01b81a94cd844867a18e73098a15
1461 lines
40 KiB
C
1461 lines
40 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/*
|
|
* A fairly generic DMA-API to IOMMU-API glue layer.
|
|
*
|
|
* Copyright (C) 2014-2015 ARM Ltd.
|
|
*
|
|
* based in part on arch/arm/mm/dma-mapping.c:
|
|
* Copyright (C) 2000-2004 Russell King
|
|
*/
|
|
|
|
#include <linux/acpi_iort.h>
|
|
#include <linux/device.h>
|
|
#include <linux/dma-map-ops.h>
|
|
#include <linux/dma-iommu.h>
|
|
#include <linux/gfp.h>
|
|
#include <linux/huge_mm.h>
|
|
#include <linux/iommu.h>
|
|
#include <linux/iova.h>
|
|
#include <linux/irq.h>
|
|
#include <linux/mm.h>
|
|
#include <linux/mutex.h>
|
|
#include <linux/pci.h>
|
|
#include <linux/swiotlb.h>
|
|
#include <linux/scatterlist.h>
|
|
#include <linux/vmalloc.h>
|
|
#include <linux/crash_dump.h>
|
|
#include <linux/dma-direct.h>
|
|
#include <trace/hooks/iommu.h>
|
|
|
|
struct iommu_dma_msi_page {
|
|
struct list_head list;
|
|
dma_addr_t iova;
|
|
phys_addr_t phys;
|
|
};
|
|
|
|
enum iommu_dma_cookie_type {
|
|
IOMMU_DMA_IOVA_COOKIE,
|
|
IOMMU_DMA_MSI_COOKIE,
|
|
};
|
|
|
|
struct iommu_dma_cookie {
|
|
enum iommu_dma_cookie_type type;
|
|
union {
|
|
/* Full allocator for IOMMU_DMA_IOVA_COOKIE */
|
|
struct iova_domain iovad;
|
|
/* Trivial linear page allocator for IOMMU_DMA_MSI_COOKIE */
|
|
dma_addr_t msi_iova;
|
|
};
|
|
struct list_head msi_page_list;
|
|
|
|
/* Domain for flush queue callback; NULL if flush queue not in use */
|
|
struct iommu_domain *fq_domain;
|
|
};
|
|
|
|
static DEFINE_STATIC_KEY_FALSE(iommu_deferred_attach_enabled);
|
|
bool iommu_dma_forcedac __read_mostly;
|
|
|
|
static int __init iommu_dma_forcedac_setup(char *str)
|
|
{
|
|
int ret = kstrtobool(str, &iommu_dma_forcedac);
|
|
|
|
if (!ret && iommu_dma_forcedac)
|
|
pr_info("Forcing DAC for PCI devices\n");
|
|
return ret;
|
|
}
|
|
early_param("iommu.forcedac", iommu_dma_forcedac_setup);
|
|
|
|
static void iommu_dma_entry_dtor(unsigned long data)
|
|
{
|
|
struct page *freelist = (struct page *)data;
|
|
|
|
while (freelist) {
|
|
unsigned long p = (unsigned long)page_address(freelist);
|
|
|
|
freelist = freelist->freelist;
|
|
free_page(p);
|
|
}
|
|
}
|
|
|
|
static inline size_t cookie_msi_granule(struct iommu_dma_cookie *cookie)
|
|
{
|
|
if (cookie->type == IOMMU_DMA_IOVA_COOKIE)
|
|
return cookie->iovad.granule;
|
|
return PAGE_SIZE;
|
|
}
|
|
|
|
static struct iommu_dma_cookie *cookie_alloc(enum iommu_dma_cookie_type type)
|
|
{
|
|
struct iommu_dma_cookie *cookie;
|
|
|
|
cookie = kzalloc(sizeof(*cookie), GFP_KERNEL);
|
|
if (cookie) {
|
|
INIT_LIST_HEAD(&cookie->msi_page_list);
|
|
cookie->type = type;
|
|
}
|
|
return cookie;
|
|
}
|
|
|
|
/**
|
|
* iommu_get_dma_cookie - Acquire DMA-API resources for a domain
|
|
* @domain: IOMMU domain to prepare for DMA-API usage
|
|
*
|
|
* IOMMU drivers should normally call this from their domain_alloc
|
|
* callback when domain->type == IOMMU_DOMAIN_DMA.
|
|
*/
|
|
int iommu_get_dma_cookie(struct iommu_domain *domain)
|
|
{
|
|
if (domain->iova_cookie)
|
|
return -EEXIST;
|
|
|
|
domain->iova_cookie = cookie_alloc(IOMMU_DMA_IOVA_COOKIE);
|
|
if (!domain->iova_cookie)
|
|
return -ENOMEM;
|
|
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(iommu_get_dma_cookie);
|
|
|
|
/**
|
|
* iommu_get_msi_cookie - Acquire just MSI remapping resources
|
|
* @domain: IOMMU domain to prepare
|
|
* @base: Start address of IOVA region for MSI mappings
|
|
*
|
|
* Users who manage their own IOVA allocation and do not want DMA API support,
|
|
* but would still like to take advantage of automatic MSI remapping, can use
|
|
* this to initialise their own domain appropriately. Users should reserve a
|
|
* contiguous IOVA region, starting at @base, large enough to accommodate the
|
|
* number of PAGE_SIZE mappings necessary to cover every MSI doorbell address
|
|
* used by the devices attached to @domain.
|
|
*/
|
|
int iommu_get_msi_cookie(struct iommu_domain *domain, dma_addr_t base)
|
|
{
|
|
struct iommu_dma_cookie *cookie;
|
|
|
|
if (domain->type != IOMMU_DOMAIN_UNMANAGED)
|
|
return -EINVAL;
|
|
|
|
if (domain->iova_cookie)
|
|
return -EEXIST;
|
|
|
|
cookie = cookie_alloc(IOMMU_DMA_MSI_COOKIE);
|
|
if (!cookie)
|
|
return -ENOMEM;
|
|
|
|
cookie->msi_iova = base;
|
|
domain->iova_cookie = cookie;
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(iommu_get_msi_cookie);
|
|
|
|
/**
|
|
* iommu_put_dma_cookie - Release a domain's DMA mapping resources
|
|
* @domain: IOMMU domain previously prepared by iommu_get_dma_cookie() or
|
|
* iommu_get_msi_cookie()
|
|
*
|
|
* IOMMU drivers should normally call this from their domain_free callback.
|
|
*/
|
|
void iommu_put_dma_cookie(struct iommu_domain *domain)
|
|
{
|
|
struct iommu_dma_cookie *cookie = domain->iova_cookie;
|
|
struct iommu_dma_msi_page *msi, *tmp;
|
|
|
|
if (!cookie)
|
|
return;
|
|
|
|
if (cookie->type == IOMMU_DMA_IOVA_COOKIE && cookie->iovad.granule)
|
|
put_iova_domain(&cookie->iovad);
|
|
|
|
list_for_each_entry_safe(msi, tmp, &cookie->msi_page_list, list) {
|
|
list_del(&msi->list);
|
|
kfree(msi);
|
|
}
|
|
kfree(cookie);
|
|
domain->iova_cookie = NULL;
|
|
}
|
|
EXPORT_SYMBOL(iommu_put_dma_cookie);
|
|
|
|
/**
|
|
* iommu_dma_get_resv_regions - Reserved region driver helper
|
|
* @dev: Device from iommu_get_resv_regions()
|
|
* @list: Reserved region list from iommu_get_resv_regions()
|
|
*
|
|
* IOMMU drivers can use this to implement their .get_resv_regions callback
|
|
* for general non-IOMMU-specific reservations. Currently, this covers GICv3
|
|
* ITS region reservation on ACPI based ARM platforms that may require HW MSI
|
|
* reservation.
|
|
*/
|
|
void iommu_dma_get_resv_regions(struct device *dev, struct list_head *list)
|
|
{
|
|
|
|
if (!is_of_node(dev_iommu_fwspec_get(dev)->iommu_fwnode))
|
|
iort_iommu_msi_get_resv_regions(dev, list);
|
|
|
|
}
|
|
EXPORT_SYMBOL(iommu_dma_get_resv_regions);
|
|
|
|
static int cookie_init_hw_msi_region(struct iommu_dma_cookie *cookie,
|
|
phys_addr_t start, phys_addr_t end)
|
|
{
|
|
struct iova_domain *iovad = &cookie->iovad;
|
|
struct iommu_dma_msi_page *msi_page;
|
|
int i, num_pages;
|
|
|
|
start -= iova_offset(iovad, start);
|
|
num_pages = iova_align(iovad, end - start) >> iova_shift(iovad);
|
|
|
|
for (i = 0; i < num_pages; i++) {
|
|
msi_page = kmalloc(sizeof(*msi_page), GFP_KERNEL);
|
|
if (!msi_page)
|
|
return -ENOMEM;
|
|
|
|
msi_page->phys = start;
|
|
msi_page->iova = start;
|
|
INIT_LIST_HEAD(&msi_page->list);
|
|
list_add(&msi_page->list, &cookie->msi_page_list);
|
|
start += iovad->granule;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int iova_reserve_pci_windows(struct pci_dev *dev,
|
|
struct iova_domain *iovad)
|
|
{
|
|
struct pci_host_bridge *bridge = pci_find_host_bridge(dev->bus);
|
|
struct resource_entry *window;
|
|
unsigned long lo, hi;
|
|
phys_addr_t start = 0, end;
|
|
|
|
resource_list_for_each_entry(window, &bridge->windows) {
|
|
if (resource_type(window->res) != IORESOURCE_MEM)
|
|
continue;
|
|
|
|
lo = iova_pfn(iovad, window->res->start - window->offset);
|
|
hi = iova_pfn(iovad, window->res->end - window->offset);
|
|
reserve_iova(iovad, lo, hi);
|
|
}
|
|
|
|
/* Get reserved DMA windows from host bridge */
|
|
resource_list_for_each_entry(window, &bridge->dma_ranges) {
|
|
end = window->res->start - window->offset;
|
|
resv_iova:
|
|
if (end > start) {
|
|
lo = iova_pfn(iovad, start);
|
|
hi = iova_pfn(iovad, end);
|
|
reserve_iova(iovad, lo, hi);
|
|
} else if (end < start) {
|
|
/* dma_ranges list should be sorted */
|
|
dev_err(&dev->dev,
|
|
"Failed to reserve IOVA [%pa-%pa]\n",
|
|
&start, &end);
|
|
return -EINVAL;
|
|
}
|
|
|
|
start = window->res->end - window->offset + 1;
|
|
/* If window is last entry */
|
|
if (window->node.next == &bridge->dma_ranges &&
|
|
end != ~(phys_addr_t)0) {
|
|
end = ~(phys_addr_t)0;
|
|
goto resv_iova;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int iova_reserve_iommu_regions(struct device *dev,
|
|
struct iommu_domain *domain)
|
|
{
|
|
struct iommu_dma_cookie *cookie = domain->iova_cookie;
|
|
struct iova_domain *iovad = &cookie->iovad;
|
|
struct iommu_resv_region *region;
|
|
LIST_HEAD(resv_regions);
|
|
int ret = 0;
|
|
|
|
if (dev_is_pci(dev)) {
|
|
ret = iova_reserve_pci_windows(to_pci_dev(dev), iovad);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
iommu_get_resv_regions(dev, &resv_regions);
|
|
list_for_each_entry(region, &resv_regions, list) {
|
|
unsigned long lo, hi;
|
|
|
|
/* We ARE the software that manages these! */
|
|
if (region->type == IOMMU_RESV_SW_MSI)
|
|
continue;
|
|
|
|
lo = iova_pfn(iovad, region->start);
|
|
hi = iova_pfn(iovad, region->start + region->length - 1);
|
|
reserve_iova(iovad, lo, hi);
|
|
|
|
if (region->type == IOMMU_RESV_MSI)
|
|
ret = cookie_init_hw_msi_region(cookie, region->start,
|
|
region->start + region->length);
|
|
if (ret)
|
|
break;
|
|
}
|
|
iommu_put_resv_regions(dev, &resv_regions);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void iommu_dma_flush_iotlb_all(struct iova_domain *iovad)
|
|
{
|
|
struct iommu_dma_cookie *cookie;
|
|
struct iommu_domain *domain;
|
|
|
|
cookie = container_of(iovad, struct iommu_dma_cookie, iovad);
|
|
domain = cookie->fq_domain;
|
|
|
|
domain->ops->flush_iotlb_all(domain);
|
|
}
|
|
|
|
static bool dev_is_untrusted(struct device *dev)
|
|
{
|
|
return dev_is_pci(dev) && to_pci_dev(dev)->untrusted;
|
|
}
|
|
|
|
static bool dev_use_swiotlb(struct device *dev)
|
|
{
|
|
return IS_ENABLED(CONFIG_SWIOTLB) && dev_is_untrusted(dev);
|
|
}
|
|
|
|
/* sysfs updates are serialised by the mutex of the group owning @domain */
|
|
int iommu_dma_init_fq(struct iommu_domain *domain)
|
|
{
|
|
struct iommu_dma_cookie *cookie = domain->iova_cookie;
|
|
int ret;
|
|
|
|
if (cookie->fq_domain)
|
|
return 0;
|
|
|
|
ret = init_iova_flush_queue(&cookie->iovad, iommu_dma_flush_iotlb_all,
|
|
iommu_dma_entry_dtor);
|
|
if (ret) {
|
|
pr_warn("iova flush queue initialization failed\n");
|
|
return ret;
|
|
}
|
|
/*
|
|
* Prevent incomplete iovad->fq being observable. Pairs with path from
|
|
* __iommu_dma_unmap() through iommu_dma_free_iova() to queue_iova()
|
|
*/
|
|
smp_wmb();
|
|
WRITE_ONCE(cookie->fq_domain, domain);
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* iommu_dma_init_domain - Initialise a DMA mapping domain
|
|
* @domain: IOMMU domain previously prepared by iommu_get_dma_cookie()
|
|
* @base: IOVA at which the mappable address space starts
|
|
* @limit: Last address of the IOVA space
|
|
* @dev: Device the domain is being initialised for
|
|
*
|
|
* @base and @limit + 1 should be exact multiples of IOMMU page granularity to
|
|
* avoid rounding surprises. If necessary, we reserve the page at address 0
|
|
* to ensure it is an invalid IOVA. It is safe to reinitialise a domain, but
|
|
* any change which could make prior IOVAs invalid will fail.
|
|
*/
|
|
static int iommu_dma_init_domain(struct iommu_domain *domain, dma_addr_t base,
|
|
dma_addr_t limit, struct device *dev)
|
|
{
|
|
struct iommu_dma_cookie *cookie = domain->iova_cookie;
|
|
unsigned long order, base_pfn;
|
|
struct iova_domain *iovad;
|
|
|
|
if (!cookie || cookie->type != IOMMU_DMA_IOVA_COOKIE)
|
|
return -EINVAL;
|
|
|
|
iovad = &cookie->iovad;
|
|
|
|
/* Use the smallest supported page size for IOVA granularity */
|
|
order = __ffs(domain->pgsize_bitmap);
|
|
base_pfn = max_t(unsigned long, 1, base >> order);
|
|
|
|
/* Check the domain allows at least some access to the device... */
|
|
if (domain->geometry.force_aperture) {
|
|
if (base > domain->geometry.aperture_end ||
|
|
limit < domain->geometry.aperture_start) {
|
|
pr_warn("specified DMA range outside IOMMU capability\n");
|
|
return -EFAULT;
|
|
}
|
|
/* ...then finally give it a kicking to make sure it fits */
|
|
base_pfn = max_t(unsigned long, base_pfn,
|
|
domain->geometry.aperture_start >> order);
|
|
}
|
|
|
|
/* start_pfn is always nonzero for an already-initialised domain */
|
|
if (iovad->start_pfn) {
|
|
if (1UL << order != iovad->granule ||
|
|
base_pfn != iovad->start_pfn) {
|
|
pr_warn("Incompatible range for DMA domain\n");
|
|
return -EFAULT;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
init_iova_domain(iovad, 1UL << order, base_pfn);
|
|
|
|
/* If the FQ fails we can simply fall back to strict mode */
|
|
if (domain->type == IOMMU_DOMAIN_DMA_FQ && iommu_dma_init_fq(domain))
|
|
domain->type = IOMMU_DOMAIN_DMA;
|
|
|
|
return iova_reserve_iommu_regions(dev, domain);
|
|
}
|
|
|
|
/*
|
|
* Should be called prior to using dma-apis.
|
|
*/
|
|
int iommu_dma_enable_best_fit_algo(struct device *dev)
|
|
{
|
|
struct iommu_domain *domain;
|
|
struct iova_domain *iovad;
|
|
|
|
domain = iommu_get_domain_for_dev(dev);
|
|
if (!domain || !domain->iova_cookie)
|
|
return -EINVAL;
|
|
|
|
iovad = &((struct iommu_dma_cookie *)domain->iova_cookie)->iovad;
|
|
iovad->best_fit = true;
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(iommu_dma_enable_best_fit_algo);
|
|
|
|
/**
|
|
* dma_info_to_prot - Translate DMA API directions and attributes to IOMMU API
|
|
* page flags.
|
|
* @dir: Direction of DMA transfer
|
|
* @coherent: Is the DMA master cache-coherent?
|
|
* @attrs: DMA attributes for the mapping
|
|
*
|
|
* Return: corresponding IOMMU API page protection flags
|
|
*/
|
|
static int dma_info_to_prot(enum dma_data_direction dir, bool coherent,
|
|
unsigned long attrs)
|
|
{
|
|
int prot = coherent ? IOMMU_CACHE : 0;
|
|
|
|
if (attrs & DMA_ATTR_PRIVILEGED)
|
|
prot |= IOMMU_PRIV;
|
|
if (attrs & DMA_ATTR_SYS_CACHE_ONLY)
|
|
prot |= IOMMU_SYS_CACHE;
|
|
if (attrs & DMA_ATTR_SYS_CACHE_ONLY_NWA)
|
|
prot |= IOMMU_SYS_CACHE_NWA;
|
|
|
|
switch (dir) {
|
|
case DMA_BIDIRECTIONAL:
|
|
return prot | IOMMU_READ | IOMMU_WRITE;
|
|
case DMA_TO_DEVICE:
|
|
return prot | IOMMU_READ;
|
|
case DMA_FROM_DEVICE:
|
|
return prot | IOMMU_WRITE;
|
|
default:
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
static dma_addr_t iommu_dma_alloc_iova(struct iommu_domain *domain,
|
|
size_t size, u64 dma_limit, struct device *dev)
|
|
{
|
|
struct iommu_dma_cookie *cookie = domain->iova_cookie;
|
|
struct iova_domain *iovad = &cookie->iovad;
|
|
unsigned long shift, iova_len, iova = 0;
|
|
|
|
if (cookie->type == IOMMU_DMA_MSI_COOKIE) {
|
|
cookie->msi_iova += size;
|
|
return cookie->msi_iova - size;
|
|
}
|
|
|
|
shift = iova_shift(iovad);
|
|
iova_len = size >> shift;
|
|
/*
|
|
* Freeing non-power-of-two-sized allocations back into the IOVA caches
|
|
* will come back to bite us badly, so we have to waste a bit of space
|
|
* rounding up anything cacheable to make sure that can't happen. The
|
|
* order of the unadjusted size will still match upon freeing.
|
|
*/
|
|
if (iova_len < (1 << (IOVA_RANGE_CACHE_MAX_SIZE - 1)))
|
|
iova_len = roundup_pow_of_two(iova_len);
|
|
|
|
dma_limit = min_not_zero(dma_limit, dev->bus_dma_limit);
|
|
|
|
if (domain->geometry.force_aperture)
|
|
dma_limit = min(dma_limit, (u64)domain->geometry.aperture_end);
|
|
|
|
/* Try to get PCI devices a SAC address */
|
|
if (dma_limit > DMA_BIT_MASK(32) && !iommu_dma_forcedac && dev_is_pci(dev))
|
|
iova = alloc_iova_fast(iovad, iova_len,
|
|
DMA_BIT_MASK(32) >> shift, false);
|
|
|
|
if (!iova)
|
|
iova = alloc_iova_fast(iovad, iova_len, dma_limit >> shift,
|
|
true);
|
|
|
|
trace_android_vh_iommu_iovad_alloc_iova(dev, iovad, (dma_addr_t)iova << shift, size);
|
|
|
|
return (dma_addr_t)iova << shift;
|
|
}
|
|
|
|
static void iommu_dma_free_iova(struct iommu_dma_cookie *cookie,
|
|
dma_addr_t iova, size_t size, struct iommu_iotlb_gather *gather)
|
|
{
|
|
struct iova_domain *iovad = &cookie->iovad;
|
|
|
|
/* The MSI case is only ever cleaning up its most recent allocation */
|
|
if (cookie->type == IOMMU_DMA_MSI_COOKIE)
|
|
cookie->msi_iova -= size;
|
|
else if (gather && gather->queued)
|
|
queue_iova(iovad, iova_pfn(iovad, iova),
|
|
size >> iova_shift(iovad),
|
|
(unsigned long)gather->freelist);
|
|
else
|
|
free_iova_fast(iovad, iova_pfn(iovad, iova),
|
|
size >> iova_shift(iovad));
|
|
|
|
trace_android_vh_iommu_iovad_free_iova(iovad, iova, size);
|
|
}
|
|
|
|
static void __iommu_dma_unmap(struct device *dev, dma_addr_t dma_addr,
|
|
size_t size)
|
|
{
|
|
struct iommu_domain *domain = iommu_get_dma_domain(dev);
|
|
struct iommu_dma_cookie *cookie = domain->iova_cookie;
|
|
struct iova_domain *iovad = &cookie->iovad;
|
|
size_t iova_off = iova_offset(iovad, dma_addr);
|
|
struct iommu_iotlb_gather iotlb_gather;
|
|
size_t unmapped;
|
|
|
|
dma_addr -= iova_off;
|
|
size = iova_align(iovad, size + iova_off);
|
|
iommu_iotlb_gather_init(&iotlb_gather);
|
|
iotlb_gather.queued = READ_ONCE(cookie->fq_domain);
|
|
|
|
unmapped = iommu_unmap_fast(domain, dma_addr, size, &iotlb_gather);
|
|
WARN_ON(unmapped != size);
|
|
|
|
if (!iotlb_gather.queued)
|
|
iommu_iotlb_sync(domain, &iotlb_gather);
|
|
iommu_dma_free_iova(cookie, dma_addr, size, &iotlb_gather);
|
|
}
|
|
|
|
static dma_addr_t __iommu_dma_map(struct device *dev, phys_addr_t phys,
|
|
size_t size, int prot, u64 dma_mask)
|
|
{
|
|
struct iommu_domain *domain = iommu_get_dma_domain(dev);
|
|
struct iommu_dma_cookie *cookie = domain->iova_cookie;
|
|
struct iova_domain *iovad = &cookie->iovad;
|
|
size_t iova_off = iova_offset(iovad, phys);
|
|
dma_addr_t iova;
|
|
|
|
if (static_branch_unlikely(&iommu_deferred_attach_enabled) &&
|
|
iommu_deferred_attach(dev, domain))
|
|
return DMA_MAPPING_ERROR;
|
|
|
|
size = iova_align(iovad, size + iova_off);
|
|
|
|
iova = iommu_dma_alloc_iova(domain, size, dma_mask, dev);
|
|
if (!iova)
|
|
return DMA_MAPPING_ERROR;
|
|
|
|
if (iommu_map_atomic(domain, iova, phys - iova_off, size, prot)) {
|
|
iommu_dma_free_iova(cookie, iova, size, NULL);
|
|
return DMA_MAPPING_ERROR;
|
|
}
|
|
return iova + iova_off;
|
|
}
|
|
|
|
static void __iommu_dma_free_pages(struct page **pages, int count)
|
|
{
|
|
while (count--)
|
|
__free_page(pages[count]);
|
|
kvfree(pages);
|
|
}
|
|
|
|
static struct page **__iommu_dma_alloc_pages(struct device *dev,
|
|
unsigned int count, unsigned long order_mask, gfp_t gfp)
|
|
{
|
|
struct page **pages;
|
|
unsigned int i = 0, nid = dev_to_node(dev);
|
|
|
|
order_mask &= (2U << MAX_ORDER) - 1;
|
|
if (!order_mask)
|
|
return NULL;
|
|
|
|
pages = kvzalloc(count * sizeof(*pages), GFP_KERNEL);
|
|
if (!pages)
|
|
return NULL;
|
|
|
|
/* IOMMU can map any pages, so himem can also be used here */
|
|
gfp |= __GFP_NOWARN | __GFP_HIGHMEM;
|
|
|
|
/* It makes no sense to muck about with huge pages */
|
|
gfp &= ~__GFP_COMP;
|
|
|
|
while (count) {
|
|
struct page *page = NULL;
|
|
unsigned int order_size;
|
|
|
|
/*
|
|
* Higher-order allocations are a convenience rather
|
|
* than a necessity, hence using __GFP_NORETRY until
|
|
* falling back to minimum-order allocations.
|
|
*/
|
|
for (order_mask &= (2U << __fls(count)) - 1;
|
|
order_mask; order_mask &= ~order_size) {
|
|
unsigned int order = __fls(order_mask);
|
|
gfp_t alloc_flags = gfp;
|
|
|
|
order_size = 1U << order;
|
|
if (order_mask > order_size)
|
|
alloc_flags |= __GFP_NORETRY;
|
|
page = alloc_pages_node(nid, alloc_flags, order);
|
|
if (!page)
|
|
continue;
|
|
if (order)
|
|
split_page(page, order);
|
|
break;
|
|
}
|
|
if (!page) {
|
|
__iommu_dma_free_pages(pages, i);
|
|
return NULL;
|
|
}
|
|
count -= order_size;
|
|
while (order_size--)
|
|
pages[i++] = page++;
|
|
}
|
|
return pages;
|
|
}
|
|
|
|
/*
|
|
* If size is less than PAGE_SIZE, then a full CPU page will be allocated,
|
|
* but an IOMMU which supports smaller pages might not map the whole thing.
|
|
*/
|
|
static struct page **__iommu_dma_alloc_noncontiguous(struct device *dev,
|
|
size_t size, struct sg_table *sgt, gfp_t gfp, pgprot_t prot,
|
|
unsigned long attrs)
|
|
{
|
|
struct iommu_domain *domain = iommu_get_dma_domain(dev);
|
|
struct iommu_dma_cookie *cookie = domain->iova_cookie;
|
|
struct iova_domain *iovad = &cookie->iovad;
|
|
bool coherent = dev_is_dma_coherent(dev);
|
|
int ioprot = dma_info_to_prot(DMA_BIDIRECTIONAL, coherent, attrs);
|
|
unsigned int count, min_size, alloc_sizes = domain->pgsize_bitmap;
|
|
struct page **pages;
|
|
dma_addr_t iova;
|
|
ssize_t ret;
|
|
|
|
if (static_branch_unlikely(&iommu_deferred_attach_enabled) &&
|
|
iommu_deferred_attach(dev, domain))
|
|
return NULL;
|
|
|
|
min_size = alloc_sizes & -alloc_sizes;
|
|
if (min_size < PAGE_SIZE) {
|
|
min_size = PAGE_SIZE;
|
|
alloc_sizes |= PAGE_SIZE;
|
|
} else {
|
|
size = ALIGN(size, min_size);
|
|
}
|
|
if (attrs & DMA_ATTR_ALLOC_SINGLE_PAGES)
|
|
alloc_sizes = min_size;
|
|
|
|
count = PAGE_ALIGN(size) >> PAGE_SHIFT;
|
|
pages = __iommu_dma_alloc_pages(dev, count, alloc_sizes >> PAGE_SHIFT,
|
|
gfp);
|
|
if (!pages)
|
|
return NULL;
|
|
|
|
size = iova_align(iovad, size);
|
|
iova = iommu_dma_alloc_iova(domain, size, dev->coherent_dma_mask, dev);
|
|
if (!iova)
|
|
goto out_free_pages;
|
|
|
|
if (sg_alloc_table_from_pages(sgt, pages, count, 0, size, GFP_KERNEL))
|
|
goto out_free_iova;
|
|
|
|
if (!(ioprot & IOMMU_CACHE)) {
|
|
struct scatterlist *sg;
|
|
int i;
|
|
|
|
for_each_sg(sgt->sgl, sg, sgt->orig_nents, i)
|
|
arch_dma_prep_coherent(sg_page(sg), sg->length);
|
|
}
|
|
|
|
ret = iommu_map_sg_atomic(domain, iova, sgt->sgl, sgt->orig_nents, ioprot);
|
|
if (ret < 0 || ret < size)
|
|
goto out_free_sg;
|
|
|
|
sgt->sgl->dma_address = iova;
|
|
sgt->sgl->dma_length = size;
|
|
return pages;
|
|
|
|
out_free_sg:
|
|
sg_free_table(sgt);
|
|
out_free_iova:
|
|
iommu_dma_free_iova(cookie, iova, size, NULL);
|
|
out_free_pages:
|
|
__iommu_dma_free_pages(pages, count);
|
|
return NULL;
|
|
}
|
|
|
|
static void *iommu_dma_alloc_remap(struct device *dev, size_t size,
|
|
dma_addr_t *dma_handle, gfp_t gfp, pgprot_t prot,
|
|
unsigned long attrs)
|
|
{
|
|
struct page **pages;
|
|
struct sg_table sgt;
|
|
void *vaddr;
|
|
|
|
pages = __iommu_dma_alloc_noncontiguous(dev, size, &sgt, gfp, prot,
|
|
attrs);
|
|
if (!pages)
|
|
return NULL;
|
|
*dma_handle = sgt.sgl->dma_address;
|
|
sg_free_table(&sgt);
|
|
vaddr = dma_common_pages_remap(pages, size, prot,
|
|
__builtin_return_address(0));
|
|
if (!vaddr)
|
|
goto out_unmap;
|
|
return vaddr;
|
|
|
|
out_unmap:
|
|
__iommu_dma_unmap(dev, *dma_handle, size);
|
|
__iommu_dma_free_pages(pages, PAGE_ALIGN(size) >> PAGE_SHIFT);
|
|
return NULL;
|
|
}
|
|
|
|
#ifdef CONFIG_DMA_REMAP
|
|
static struct sg_table *iommu_dma_alloc_noncontiguous(struct device *dev,
|
|
size_t size, enum dma_data_direction dir, gfp_t gfp,
|
|
unsigned long attrs)
|
|
{
|
|
struct dma_sgt_handle *sh;
|
|
|
|
sh = kmalloc(sizeof(*sh), gfp);
|
|
if (!sh)
|
|
return NULL;
|
|
|
|
sh->pages = __iommu_dma_alloc_noncontiguous(dev, size, &sh->sgt, gfp,
|
|
PAGE_KERNEL, attrs);
|
|
if (!sh->pages) {
|
|
kfree(sh);
|
|
return NULL;
|
|
}
|
|
return &sh->sgt;
|
|
}
|
|
|
|
static void iommu_dma_free_noncontiguous(struct device *dev, size_t size,
|
|
struct sg_table *sgt, enum dma_data_direction dir)
|
|
{
|
|
struct dma_sgt_handle *sh = sgt_handle(sgt);
|
|
|
|
__iommu_dma_unmap(dev, sgt->sgl->dma_address, size);
|
|
__iommu_dma_free_pages(sh->pages, PAGE_ALIGN(size) >> PAGE_SHIFT);
|
|
sg_free_table(&sh->sgt);
|
|
kfree(sh);
|
|
}
|
|
#endif /* CONFIG_DMA_REMAP */
|
|
|
|
static void iommu_dma_sync_single_for_cpu(struct device *dev,
|
|
dma_addr_t dma_handle, size_t size, enum dma_data_direction dir)
|
|
{
|
|
phys_addr_t phys;
|
|
|
|
if (dev_is_dma_coherent(dev) && !dev_use_swiotlb(dev))
|
|
return;
|
|
|
|
phys = iommu_iova_to_phys(iommu_get_dma_domain(dev), dma_handle);
|
|
if (!dev_is_dma_coherent(dev))
|
|
arch_sync_dma_for_cpu(phys, size, dir);
|
|
|
|
if (is_swiotlb_buffer(dev, phys))
|
|
swiotlb_sync_single_for_cpu(dev, phys, size, dir);
|
|
}
|
|
|
|
static void iommu_dma_sync_single_for_device(struct device *dev,
|
|
dma_addr_t dma_handle, size_t size, enum dma_data_direction dir)
|
|
{
|
|
phys_addr_t phys;
|
|
|
|
if (dev_is_dma_coherent(dev) && !dev_use_swiotlb(dev))
|
|
return;
|
|
|
|
phys = iommu_iova_to_phys(iommu_get_dma_domain(dev), dma_handle);
|
|
if (is_swiotlb_buffer(dev, phys))
|
|
swiotlb_sync_single_for_device(dev, phys, size, dir);
|
|
|
|
if (!dev_is_dma_coherent(dev))
|
|
arch_sync_dma_for_device(phys, size, dir);
|
|
}
|
|
|
|
static void iommu_dma_sync_sg_for_cpu(struct device *dev,
|
|
struct scatterlist *sgl, int nelems,
|
|
enum dma_data_direction dir)
|
|
{
|
|
struct scatterlist *sg;
|
|
int i;
|
|
|
|
if (dev_use_swiotlb(dev))
|
|
for_each_sg(sgl, sg, nelems, i)
|
|
iommu_dma_sync_single_for_cpu(dev, sg_dma_address(sg),
|
|
sg->length, dir);
|
|
else if (!dev_is_dma_coherent(dev))
|
|
for_each_sg(sgl, sg, nelems, i)
|
|
arch_sync_dma_for_cpu(sg_phys(sg), sg->length, dir);
|
|
}
|
|
|
|
static void iommu_dma_sync_sg_for_device(struct device *dev,
|
|
struct scatterlist *sgl, int nelems,
|
|
enum dma_data_direction dir)
|
|
{
|
|
struct scatterlist *sg;
|
|
int i;
|
|
|
|
if (dev_use_swiotlb(dev))
|
|
for_each_sg(sgl, sg, nelems, i)
|
|
iommu_dma_sync_single_for_device(dev,
|
|
sg_dma_address(sg),
|
|
sg->length, dir);
|
|
else if (!dev_is_dma_coherent(dev))
|
|
for_each_sg(sgl, sg, nelems, i)
|
|
arch_sync_dma_for_device(sg_phys(sg), sg->length, dir);
|
|
}
|
|
|
|
static dma_addr_t iommu_dma_map_page(struct device *dev, struct page *page,
|
|
unsigned long offset, size_t size, enum dma_data_direction dir,
|
|
unsigned long attrs)
|
|
{
|
|
phys_addr_t phys = page_to_phys(page) + offset;
|
|
bool coherent = dev_is_dma_coherent(dev);
|
|
int prot = dma_info_to_prot(dir, coherent, attrs);
|
|
struct iommu_domain *domain = iommu_get_dma_domain(dev);
|
|
struct iommu_dma_cookie *cookie = domain->iova_cookie;
|
|
struct iova_domain *iovad = &cookie->iovad;
|
|
dma_addr_t iova, dma_mask = dma_get_mask(dev);
|
|
|
|
/*
|
|
* If both the physical buffer start address and size are
|
|
* page aligned, we don't need to use a bounce page.
|
|
*/
|
|
if (dev_use_swiotlb(dev) && iova_offset(iovad, phys | size)) {
|
|
void *padding_start;
|
|
size_t padding_size, aligned_size;
|
|
|
|
aligned_size = iova_align(iovad, size);
|
|
phys = swiotlb_tbl_map_single(dev, phys, size, aligned_size,
|
|
iova_mask(iovad), dir, attrs);
|
|
|
|
if (phys == DMA_MAPPING_ERROR)
|
|
return DMA_MAPPING_ERROR;
|
|
|
|
/* Cleanup the padding area. */
|
|
padding_start = phys_to_virt(phys);
|
|
padding_size = aligned_size;
|
|
|
|
if (!(attrs & DMA_ATTR_SKIP_CPU_SYNC) &&
|
|
(dir == DMA_TO_DEVICE || dir == DMA_BIDIRECTIONAL)) {
|
|
padding_start += size;
|
|
padding_size -= size;
|
|
}
|
|
|
|
memset(padding_start, 0, padding_size);
|
|
}
|
|
|
|
if (!coherent && !(attrs & DMA_ATTR_SKIP_CPU_SYNC))
|
|
arch_sync_dma_for_device(phys, size, dir);
|
|
|
|
iova = __iommu_dma_map(dev, phys, size, prot, dma_mask);
|
|
if (iova == DMA_MAPPING_ERROR && is_swiotlb_buffer(dev, phys))
|
|
swiotlb_tbl_unmap_single(dev, phys, size, dir, attrs);
|
|
return iova;
|
|
}
|
|
|
|
static void iommu_dma_unmap_page(struct device *dev, dma_addr_t dma_handle,
|
|
size_t size, enum dma_data_direction dir, unsigned long attrs)
|
|
{
|
|
struct iommu_domain *domain = iommu_get_dma_domain(dev);
|
|
phys_addr_t phys;
|
|
|
|
phys = iommu_iova_to_phys(domain, dma_handle);
|
|
if (WARN_ON(!phys))
|
|
return;
|
|
|
|
if (!(attrs & DMA_ATTR_SKIP_CPU_SYNC) && !dev_is_dma_coherent(dev))
|
|
arch_sync_dma_for_cpu(phys, size, dir);
|
|
|
|
__iommu_dma_unmap(dev, dma_handle, size);
|
|
|
|
if (unlikely(is_swiotlb_buffer(dev, phys)))
|
|
swiotlb_tbl_unmap_single(dev, phys, size, dir, attrs);
|
|
}
|
|
|
|
/*
|
|
* Prepare a successfully-mapped scatterlist to give back to the caller.
|
|
*
|
|
* At this point the segments are already laid out by iommu_dma_map_sg() to
|
|
* avoid individually crossing any boundaries, so we merely need to check a
|
|
* segment's start address to avoid concatenating across one.
|
|
*/
|
|
static int __finalise_sg(struct device *dev, struct scatterlist *sg, int nents,
|
|
dma_addr_t dma_addr)
|
|
{
|
|
struct scatterlist *s, *cur = sg;
|
|
unsigned long seg_mask = dma_get_seg_boundary(dev);
|
|
unsigned int cur_len = 0, max_len = dma_get_max_seg_size(dev);
|
|
int i, count = 0;
|
|
|
|
for_each_sg(sg, s, nents, i) {
|
|
/* Restore this segment's original unaligned fields first */
|
|
unsigned int s_iova_off = sg_dma_address(s);
|
|
unsigned int s_length = sg_dma_len(s);
|
|
unsigned int s_iova_len = s->length;
|
|
|
|
s->offset += s_iova_off;
|
|
s->length = s_length;
|
|
sg_dma_address(s) = DMA_MAPPING_ERROR;
|
|
sg_dma_len(s) = 0;
|
|
|
|
/*
|
|
* Now fill in the real DMA data. If...
|
|
* - there is a valid output segment to append to
|
|
* - and this segment starts on an IOVA page boundary
|
|
* - but doesn't fall at a segment boundary
|
|
* - and wouldn't make the resulting output segment too long
|
|
*/
|
|
if (cur_len && !s_iova_off && (dma_addr & seg_mask) &&
|
|
(max_len - cur_len >= s_length)) {
|
|
/* ...then concatenate it with the previous one */
|
|
cur_len += s_length;
|
|
} else {
|
|
/* Otherwise start the next output segment */
|
|
if (i > 0)
|
|
cur = sg_next(cur);
|
|
cur_len = s_length;
|
|
count++;
|
|
|
|
sg_dma_address(cur) = dma_addr + s_iova_off;
|
|
}
|
|
|
|
sg_dma_len(cur) = cur_len;
|
|
dma_addr += s_iova_len;
|
|
|
|
if (s_length + s_iova_off < s_iova_len)
|
|
cur_len = 0;
|
|
}
|
|
return count;
|
|
}
|
|
|
|
/*
|
|
* If mapping failed, then just restore the original list,
|
|
* but making sure the DMA fields are invalidated.
|
|
*/
|
|
static void __invalidate_sg(struct scatterlist *sg, int nents)
|
|
{
|
|
struct scatterlist *s;
|
|
int i;
|
|
|
|
for_each_sg(sg, s, nents, i) {
|
|
if (sg_dma_address(s) != DMA_MAPPING_ERROR)
|
|
s->offset += sg_dma_address(s);
|
|
if (sg_dma_len(s))
|
|
s->length = sg_dma_len(s);
|
|
sg_dma_address(s) = DMA_MAPPING_ERROR;
|
|
sg_dma_len(s) = 0;
|
|
}
|
|
}
|
|
|
|
static void iommu_dma_unmap_sg_swiotlb(struct device *dev, struct scatterlist *sg,
|
|
int nents, enum dma_data_direction dir, unsigned long attrs)
|
|
{
|
|
struct scatterlist *s;
|
|
int i;
|
|
|
|
for_each_sg(sg, s, nents, i)
|
|
iommu_dma_unmap_page(dev, sg_dma_address(s),
|
|
sg_dma_len(s), dir, attrs);
|
|
}
|
|
|
|
static int iommu_dma_map_sg_swiotlb(struct device *dev, struct scatterlist *sg,
|
|
int nents, enum dma_data_direction dir, unsigned long attrs)
|
|
{
|
|
struct scatterlist *s;
|
|
int i;
|
|
|
|
for_each_sg(sg, s, nents, i) {
|
|
sg_dma_address(s) = iommu_dma_map_page(dev, sg_page(s),
|
|
s->offset, s->length, dir, attrs);
|
|
if (sg_dma_address(s) == DMA_MAPPING_ERROR)
|
|
goto out_unmap;
|
|
sg_dma_len(s) = s->length;
|
|
}
|
|
|
|
return nents;
|
|
|
|
out_unmap:
|
|
iommu_dma_unmap_sg_swiotlb(dev, sg, i, dir, attrs | DMA_ATTR_SKIP_CPU_SYNC);
|
|
return -EIO;
|
|
}
|
|
|
|
/*
|
|
* The DMA API client is passing in a scatterlist which could describe
|
|
* any old buffer layout, but the IOMMU API requires everything to be
|
|
* aligned to IOMMU pages. Hence the need for this complicated bit of
|
|
* impedance-matching, to be able to hand off a suitably-aligned list,
|
|
* but still preserve the original offsets and sizes for the caller.
|
|
*/
|
|
static int iommu_dma_map_sg(struct device *dev, struct scatterlist *sg,
|
|
int nents, enum dma_data_direction dir, unsigned long attrs)
|
|
{
|
|
struct iommu_domain *domain = iommu_get_dma_domain(dev);
|
|
struct iommu_dma_cookie *cookie = domain->iova_cookie;
|
|
struct iova_domain *iovad = &cookie->iovad;
|
|
struct scatterlist *s, *prev = NULL;
|
|
int prot = dma_info_to_prot(dir, dev_is_dma_coherent(dev), attrs);
|
|
dma_addr_t iova;
|
|
size_t iova_len = 0;
|
|
unsigned long mask = dma_get_seg_boundary(dev);
|
|
ssize_t ret;
|
|
int i;
|
|
|
|
if (static_branch_unlikely(&iommu_deferred_attach_enabled)) {
|
|
ret = iommu_deferred_attach(dev, domain);
|
|
if (ret)
|
|
goto out;
|
|
}
|
|
|
|
if (dev_use_swiotlb(dev))
|
|
return iommu_dma_map_sg_swiotlb(dev, sg, nents, dir, attrs);
|
|
|
|
if (!(attrs & DMA_ATTR_SKIP_CPU_SYNC))
|
|
iommu_dma_sync_sg_for_device(dev, sg, nents, dir);
|
|
|
|
/*
|
|
* Work out how much IOVA space we need, and align the segments to
|
|
* IOVA granules for the IOMMU driver to handle. With some clever
|
|
* trickery we can modify the list in-place, but reversibly, by
|
|
* stashing the unaligned parts in the as-yet-unused DMA fields.
|
|
*/
|
|
for_each_sg(sg, s, nents, i) {
|
|
size_t s_iova_off = iova_offset(iovad, s->offset);
|
|
size_t s_length = s->length;
|
|
size_t pad_len = (mask - iova_len + 1) & mask;
|
|
|
|
sg_dma_address(s) = s_iova_off;
|
|
sg_dma_len(s) = s_length;
|
|
s->offset -= s_iova_off;
|
|
s_length = iova_align(iovad, s_length + s_iova_off);
|
|
s->length = s_length;
|
|
|
|
/*
|
|
* Due to the alignment of our single IOVA allocation, we can
|
|
* depend on these assumptions about the segment boundary mask:
|
|
* - If mask size >= IOVA size, then the IOVA range cannot
|
|
* possibly fall across a boundary, so we don't care.
|
|
* - If mask size < IOVA size, then the IOVA range must start
|
|
* exactly on a boundary, therefore we can lay things out
|
|
* based purely on segment lengths without needing to know
|
|
* the actual addresses beforehand.
|
|
* - The mask must be a power of 2, so pad_len == 0 if
|
|
* iova_len == 0, thus we cannot dereference prev the first
|
|
* time through here (i.e. before it has a meaningful value).
|
|
*/
|
|
if (pad_len && pad_len < s_length - 1) {
|
|
prev->length += pad_len;
|
|
iova_len += pad_len;
|
|
}
|
|
|
|
iova_len += s_length;
|
|
prev = s;
|
|
}
|
|
|
|
iova = iommu_dma_alloc_iova(domain, iova_len, dma_get_mask(dev), dev);
|
|
if (!iova) {
|
|
ret = -ENOMEM;
|
|
goto out_restore_sg;
|
|
}
|
|
|
|
/*
|
|
* We'll leave any physical concatenation to the IOMMU driver's
|
|
* implementation - it knows better than we do.
|
|
*/
|
|
ret = iommu_map_sg_atomic(domain, iova, sg, nents, prot);
|
|
if (ret < 0 || ret < iova_len)
|
|
goto out_free_iova;
|
|
|
|
return __finalise_sg(dev, sg, nents, iova);
|
|
|
|
out_free_iova:
|
|
iommu_dma_free_iova(cookie, iova, iova_len, NULL);
|
|
out_restore_sg:
|
|
__invalidate_sg(sg, nents);
|
|
out:
|
|
if (ret != -ENOMEM)
|
|
return -EINVAL;
|
|
return ret;
|
|
}
|
|
|
|
static void iommu_dma_unmap_sg(struct device *dev, struct scatterlist *sg,
|
|
int nents, enum dma_data_direction dir, unsigned long attrs)
|
|
{
|
|
dma_addr_t start, end;
|
|
struct scatterlist *tmp;
|
|
int i;
|
|
|
|
if (dev_use_swiotlb(dev)) {
|
|
iommu_dma_unmap_sg_swiotlb(dev, sg, nents, dir, attrs);
|
|
return;
|
|
}
|
|
|
|
if (!(attrs & DMA_ATTR_SKIP_CPU_SYNC))
|
|
iommu_dma_sync_sg_for_cpu(dev, sg, nents, dir);
|
|
|
|
/*
|
|
* The scatterlist segments are mapped into a single
|
|
* contiguous IOVA allocation, so this is incredibly easy.
|
|
*/
|
|
start = sg_dma_address(sg);
|
|
for_each_sg(sg_next(sg), tmp, nents - 1, i) {
|
|
if (sg_dma_len(tmp) == 0)
|
|
break;
|
|
sg = tmp;
|
|
}
|
|
end = sg_dma_address(sg) + sg_dma_len(sg);
|
|
__iommu_dma_unmap(dev, start, end - start);
|
|
}
|
|
|
|
static dma_addr_t iommu_dma_map_resource(struct device *dev, phys_addr_t phys,
|
|
size_t size, enum dma_data_direction dir, unsigned long attrs)
|
|
{
|
|
return __iommu_dma_map(dev, phys, size,
|
|
dma_info_to_prot(dir, false, attrs) | IOMMU_MMIO,
|
|
dma_get_mask(dev));
|
|
}
|
|
|
|
static void iommu_dma_unmap_resource(struct device *dev, dma_addr_t handle,
|
|
size_t size, enum dma_data_direction dir, unsigned long attrs)
|
|
{
|
|
__iommu_dma_unmap(dev, handle, size);
|
|
}
|
|
|
|
static void __iommu_dma_free(struct device *dev, size_t size, void *cpu_addr)
|
|
{
|
|
size_t alloc_size = PAGE_ALIGN(size);
|
|
int count = alloc_size >> PAGE_SHIFT;
|
|
struct page *page = NULL, **pages = NULL;
|
|
|
|
/* Non-coherent atomic allocation? Easy */
|
|
if (IS_ENABLED(CONFIG_DMA_DIRECT_REMAP) &&
|
|
dma_free_from_pool(dev, cpu_addr, alloc_size))
|
|
return;
|
|
|
|
if (IS_ENABLED(CONFIG_DMA_REMAP) && is_vmalloc_addr(cpu_addr)) {
|
|
/*
|
|
* If it the address is remapped, then it's either non-coherent
|
|
* or highmem CMA, or an iommu_dma_alloc_remap() construction.
|
|
*/
|
|
pages = dma_common_find_pages(cpu_addr);
|
|
if (!pages)
|
|
page = vmalloc_to_page(cpu_addr);
|
|
dma_common_free_remap(cpu_addr, alloc_size);
|
|
} else {
|
|
/* Lowmem means a coherent atomic or CMA allocation */
|
|
page = virt_to_page(cpu_addr);
|
|
}
|
|
|
|
if (pages)
|
|
__iommu_dma_free_pages(pages, count);
|
|
if (page)
|
|
dma_free_contiguous(dev, page, alloc_size);
|
|
}
|
|
|
|
static void iommu_dma_free(struct device *dev, size_t size, void *cpu_addr,
|
|
dma_addr_t handle, unsigned long attrs)
|
|
{
|
|
__iommu_dma_unmap(dev, handle, size);
|
|
__iommu_dma_free(dev, size, cpu_addr);
|
|
}
|
|
|
|
static void *iommu_dma_alloc_pages(struct device *dev, size_t size,
|
|
struct page **pagep, gfp_t gfp, unsigned long attrs)
|
|
{
|
|
bool coherent = dev_is_dma_coherent(dev);
|
|
size_t alloc_size = PAGE_ALIGN(size);
|
|
int node = dev_to_node(dev);
|
|
struct page *page = NULL;
|
|
void *cpu_addr;
|
|
|
|
page = dma_alloc_contiguous(dev, alloc_size, gfp);
|
|
if (!page)
|
|
page = alloc_pages_node(node, gfp, get_order(alloc_size));
|
|
if (!page)
|
|
return NULL;
|
|
|
|
if (IS_ENABLED(CONFIG_DMA_REMAP) && (!coherent || PageHighMem(page))) {
|
|
pgprot_t prot = dma_pgprot(dev, PAGE_KERNEL, attrs);
|
|
|
|
cpu_addr = dma_common_contiguous_remap(page, alloc_size,
|
|
prot, __builtin_return_address(0));
|
|
if (!cpu_addr)
|
|
goto out_free_pages;
|
|
|
|
if (!coherent)
|
|
arch_dma_prep_coherent(page, size);
|
|
} else {
|
|
cpu_addr = page_address(page);
|
|
}
|
|
|
|
*pagep = page;
|
|
memset(cpu_addr, 0, alloc_size);
|
|
return cpu_addr;
|
|
out_free_pages:
|
|
dma_free_contiguous(dev, page, alloc_size);
|
|
return NULL;
|
|
}
|
|
|
|
static void *iommu_dma_alloc(struct device *dev, size_t size,
|
|
dma_addr_t *handle, gfp_t gfp, unsigned long attrs)
|
|
{
|
|
bool coherent = dev_is_dma_coherent(dev);
|
|
int ioprot = dma_info_to_prot(DMA_BIDIRECTIONAL, coherent, attrs);
|
|
struct page *page = NULL;
|
|
void *cpu_addr;
|
|
|
|
gfp |= __GFP_ZERO;
|
|
|
|
if (IS_ENABLED(CONFIG_DMA_REMAP) && gfpflags_allow_blocking(gfp) &&
|
|
!(attrs & DMA_ATTR_FORCE_CONTIGUOUS)) {
|
|
return iommu_dma_alloc_remap(dev, size, handle, gfp,
|
|
dma_pgprot(dev, PAGE_KERNEL, attrs), attrs);
|
|
}
|
|
|
|
if (IS_ENABLED(CONFIG_DMA_DIRECT_REMAP) &&
|
|
!gfpflags_allow_blocking(gfp) && !coherent)
|
|
page = dma_alloc_from_pool(dev, PAGE_ALIGN(size), &cpu_addr,
|
|
gfp, NULL);
|
|
else
|
|
cpu_addr = iommu_dma_alloc_pages(dev, size, &page, gfp, attrs);
|
|
if (!cpu_addr)
|
|
return NULL;
|
|
|
|
*handle = __iommu_dma_map(dev, page_to_phys(page), size, ioprot,
|
|
dev->coherent_dma_mask);
|
|
if (*handle == DMA_MAPPING_ERROR) {
|
|
__iommu_dma_free(dev, size, cpu_addr);
|
|
return NULL;
|
|
}
|
|
|
|
return cpu_addr;
|
|
}
|
|
|
|
static int iommu_dma_mmap(struct device *dev, struct vm_area_struct *vma,
|
|
void *cpu_addr, dma_addr_t dma_addr, size_t size,
|
|
unsigned long attrs)
|
|
{
|
|
unsigned long nr_pages = PAGE_ALIGN(size) >> PAGE_SHIFT;
|
|
unsigned long pfn, off = vma->vm_pgoff;
|
|
int ret;
|
|
|
|
vma->vm_page_prot = dma_pgprot(dev, vma->vm_page_prot, attrs);
|
|
|
|
if (dma_mmap_from_dev_coherent(dev, vma, cpu_addr, size, &ret))
|
|
return ret;
|
|
|
|
if (off >= nr_pages || vma_pages(vma) > nr_pages - off)
|
|
return -ENXIO;
|
|
|
|
if (IS_ENABLED(CONFIG_DMA_REMAP) && is_vmalloc_addr(cpu_addr)) {
|
|
struct page **pages = dma_common_find_pages(cpu_addr);
|
|
|
|
if (pages)
|
|
return vm_map_pages(vma, pages, nr_pages);
|
|
pfn = vmalloc_to_pfn(cpu_addr);
|
|
} else {
|
|
pfn = page_to_pfn(virt_to_page(cpu_addr));
|
|
}
|
|
|
|
return remap_pfn_range(vma, vma->vm_start, pfn + off,
|
|
vma->vm_end - vma->vm_start,
|
|
vma->vm_page_prot);
|
|
}
|
|
|
|
static int iommu_dma_get_sgtable(struct device *dev, struct sg_table *sgt,
|
|
void *cpu_addr, dma_addr_t dma_addr, size_t size,
|
|
unsigned long attrs)
|
|
{
|
|
struct page *page;
|
|
int ret;
|
|
|
|
if (IS_ENABLED(CONFIG_DMA_REMAP) && is_vmalloc_addr(cpu_addr)) {
|
|
struct page **pages = dma_common_find_pages(cpu_addr);
|
|
|
|
if (pages) {
|
|
return sg_alloc_table_from_pages(sgt, pages,
|
|
PAGE_ALIGN(size) >> PAGE_SHIFT,
|
|
0, size, GFP_KERNEL);
|
|
}
|
|
|
|
page = vmalloc_to_page(cpu_addr);
|
|
} else {
|
|
page = virt_to_page(cpu_addr);
|
|
}
|
|
|
|
ret = sg_alloc_table(sgt, 1, GFP_KERNEL);
|
|
if (!ret)
|
|
sg_set_page(sgt->sgl, page, PAGE_ALIGN(size), 0);
|
|
return ret;
|
|
}
|
|
|
|
static unsigned long iommu_dma_get_merge_boundary(struct device *dev)
|
|
{
|
|
struct iommu_domain *domain = iommu_get_dma_domain(dev);
|
|
|
|
return (1UL << __ffs(domain->pgsize_bitmap)) - 1;
|
|
}
|
|
|
|
static const struct dma_map_ops iommu_dma_ops = {
|
|
.alloc = iommu_dma_alloc,
|
|
.free = iommu_dma_free,
|
|
.alloc_pages = dma_common_alloc_pages,
|
|
.free_pages = dma_common_free_pages,
|
|
#ifdef CONFIG_DMA_REMAP
|
|
.alloc_noncontiguous = iommu_dma_alloc_noncontiguous,
|
|
.free_noncontiguous = iommu_dma_free_noncontiguous,
|
|
#endif
|
|
.mmap = iommu_dma_mmap,
|
|
.get_sgtable = iommu_dma_get_sgtable,
|
|
.map_page = iommu_dma_map_page,
|
|
.unmap_page = iommu_dma_unmap_page,
|
|
.map_sg = iommu_dma_map_sg,
|
|
.unmap_sg = iommu_dma_unmap_sg,
|
|
.sync_single_for_cpu = iommu_dma_sync_single_for_cpu,
|
|
.sync_single_for_device = iommu_dma_sync_single_for_device,
|
|
.sync_sg_for_cpu = iommu_dma_sync_sg_for_cpu,
|
|
.sync_sg_for_device = iommu_dma_sync_sg_for_device,
|
|
.map_resource = iommu_dma_map_resource,
|
|
.unmap_resource = iommu_dma_unmap_resource,
|
|
.get_merge_boundary = iommu_dma_get_merge_boundary,
|
|
};
|
|
|
|
/*
|
|
* The IOMMU core code allocates the default DMA domain, which the underlying
|
|
* IOMMU driver needs to support via the dma-iommu layer.
|
|
*/
|
|
void iommu_setup_dma_ops(struct device *dev, u64 dma_base, u64 dma_limit)
|
|
{
|
|
struct iommu_domain *domain = iommu_get_domain_for_dev(dev);
|
|
|
|
if (!domain)
|
|
goto out_err;
|
|
|
|
/*
|
|
* The IOMMU core code allocates the default DMA domain, which the
|
|
* underlying IOMMU driver needs to support via the dma-iommu layer.
|
|
*/
|
|
if (iommu_is_dma_domain(domain)) {
|
|
if (iommu_dma_init_domain(domain, dma_base, dma_limit, dev))
|
|
goto out_err;
|
|
dev->dma_ops = &iommu_dma_ops;
|
|
}
|
|
|
|
return;
|
|
out_err:
|
|
pr_warn("Failed to set up IOMMU for device %s; retaining platform DMA ops\n",
|
|
dev_name(dev));
|
|
}
|
|
EXPORT_SYMBOL_GPL(iommu_setup_dma_ops);
|
|
|
|
static struct iommu_dma_msi_page *iommu_dma_get_msi_page(struct device *dev,
|
|
phys_addr_t msi_addr, struct iommu_domain *domain)
|
|
{
|
|
struct iommu_dma_cookie *cookie = domain->iova_cookie;
|
|
struct iommu_dma_msi_page *msi_page;
|
|
dma_addr_t iova;
|
|
int prot = IOMMU_WRITE | IOMMU_NOEXEC | IOMMU_MMIO;
|
|
size_t size = cookie_msi_granule(cookie);
|
|
|
|
msi_addr &= ~(phys_addr_t)(size - 1);
|
|
list_for_each_entry(msi_page, &cookie->msi_page_list, list)
|
|
if (msi_page->phys == msi_addr)
|
|
return msi_page;
|
|
|
|
msi_page = kzalloc(sizeof(*msi_page), GFP_KERNEL);
|
|
if (!msi_page)
|
|
return NULL;
|
|
|
|
iova = iommu_dma_alloc_iova(domain, size, dma_get_mask(dev), dev);
|
|
if (!iova)
|
|
goto out_free_page;
|
|
|
|
if (iommu_map(domain, iova, msi_addr, size, prot))
|
|
goto out_free_iova;
|
|
|
|
INIT_LIST_HEAD(&msi_page->list);
|
|
msi_page->phys = msi_addr;
|
|
msi_page->iova = iova;
|
|
list_add(&msi_page->list, &cookie->msi_page_list);
|
|
return msi_page;
|
|
|
|
out_free_iova:
|
|
iommu_dma_free_iova(cookie, iova, size, NULL);
|
|
out_free_page:
|
|
kfree(msi_page);
|
|
return NULL;
|
|
}
|
|
|
|
int iommu_dma_prepare_msi(struct msi_desc *desc, phys_addr_t msi_addr)
|
|
{
|
|
struct device *dev = msi_desc_to_dev(desc);
|
|
struct iommu_domain *domain = iommu_get_domain_for_dev(dev);
|
|
struct iommu_dma_msi_page *msi_page;
|
|
static DEFINE_MUTEX(msi_prepare_lock); /* see below */
|
|
|
|
if (!domain || !domain->iova_cookie) {
|
|
desc->iommu_cookie = NULL;
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* In fact the whole prepare operation should already be serialised by
|
|
* irq_domain_mutex further up the callchain, but that's pretty subtle
|
|
* on its own, so consider this locking as failsafe documentation...
|
|
*/
|
|
mutex_lock(&msi_prepare_lock);
|
|
msi_page = iommu_dma_get_msi_page(dev, msi_addr, domain);
|
|
mutex_unlock(&msi_prepare_lock);
|
|
|
|
msi_desc_set_iommu_cookie(desc, msi_page);
|
|
|
|
if (!msi_page)
|
|
return -ENOMEM;
|
|
return 0;
|
|
}
|
|
|
|
void iommu_dma_compose_msi_msg(struct msi_desc *desc,
|
|
struct msi_msg *msg)
|
|
{
|
|
struct device *dev = msi_desc_to_dev(desc);
|
|
const struct iommu_domain *domain = iommu_get_domain_for_dev(dev);
|
|
const struct iommu_dma_msi_page *msi_page;
|
|
|
|
msi_page = msi_desc_get_iommu_cookie(desc);
|
|
|
|
if (!domain || !domain->iova_cookie || WARN_ON(!msi_page))
|
|
return;
|
|
|
|
msg->address_hi = upper_32_bits(msi_page->iova);
|
|
msg->address_lo &= cookie_msi_granule(domain->iova_cookie) - 1;
|
|
msg->address_lo += lower_32_bits(msi_page->iova);
|
|
}
|
|
|
|
static int iommu_dma_init(void)
|
|
{
|
|
if (is_kdump_kernel())
|
|
static_branch_enable(&iommu_deferred_attach_enabled);
|
|
|
|
return iova_cache_get();
|
|
}
|
|
arch_initcall(iommu_dma_init);
|