123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756 |
- // SPDX-License-Identifier: GPL-2.0
- /*
- * intel-pasid.c - PASID idr, table and entry manipulation
- *
- * Copyright (C) 2018 Intel Corporation
- *
- * Author: Lu Baolu <[email protected]>
- */
- #define pr_fmt(fmt) "DMAR: " fmt
- #include <linux/bitops.h>
- #include <linux/cpufeature.h>
- #include <linux/dmar.h>
- #include <linux/iommu.h>
- #include <linux/memory.h>
- #include <linux/pci.h>
- #include <linux/pci-ats.h>
- #include <linux/spinlock.h>
- #include "iommu.h"
- #include "pasid.h"
- /*
- * Intel IOMMU system wide PASID name space:
- */
- u32 intel_pasid_max_id = PASID_MAX;
- int vcmd_alloc_pasid(struct intel_iommu *iommu, u32 *pasid)
- {
- unsigned long flags;
- u8 status_code;
- int ret = 0;
- u64 res;
- raw_spin_lock_irqsave(&iommu->register_lock, flags);
- dmar_writeq(iommu->reg + DMAR_VCMD_REG, VCMD_CMD_ALLOC);
- IOMMU_WAIT_OP(iommu, DMAR_VCRSP_REG, dmar_readq,
- !(res & VCMD_VRSP_IP), res);
- raw_spin_unlock_irqrestore(&iommu->register_lock, flags);
- status_code = VCMD_VRSP_SC(res);
- switch (status_code) {
- case VCMD_VRSP_SC_SUCCESS:
- *pasid = VCMD_VRSP_RESULT_PASID(res);
- break;
- case VCMD_VRSP_SC_NO_PASID_AVAIL:
- pr_info("IOMMU: %s: No PASID available\n", iommu->name);
- ret = -ENOSPC;
- break;
- default:
- ret = -ENODEV;
- pr_warn("IOMMU: %s: Unexpected error code %d\n",
- iommu->name, status_code);
- }
- return ret;
- }
- void vcmd_free_pasid(struct intel_iommu *iommu, u32 pasid)
- {
- unsigned long flags;
- u8 status_code;
- u64 res;
- raw_spin_lock_irqsave(&iommu->register_lock, flags);
- dmar_writeq(iommu->reg + DMAR_VCMD_REG,
- VCMD_CMD_OPERAND(pasid) | VCMD_CMD_FREE);
- IOMMU_WAIT_OP(iommu, DMAR_VCRSP_REG, dmar_readq,
- !(res & VCMD_VRSP_IP), res);
- raw_spin_unlock_irqrestore(&iommu->register_lock, flags);
- status_code = VCMD_VRSP_SC(res);
- switch (status_code) {
- case VCMD_VRSP_SC_SUCCESS:
- break;
- case VCMD_VRSP_SC_INVALID_PASID:
- pr_info("IOMMU: %s: Invalid PASID\n", iommu->name);
- break;
- default:
- pr_warn("IOMMU: %s: Unexpected error code %d\n",
- iommu->name, status_code);
- }
- }
- /*
- * Per device pasid table management:
- */
- /*
- * Allocate a pasid table for @dev. It should be called in a
- * single-thread context.
- */
- int intel_pasid_alloc_table(struct device *dev)
- {
- struct device_domain_info *info;
- struct pasid_table *pasid_table;
- struct page *pages;
- u32 max_pasid = 0;
- int order, size;
- might_sleep();
- info = dev_iommu_priv_get(dev);
- if (WARN_ON(!info || !dev_is_pci(dev) || info->pasid_table))
- return -EINVAL;
- pasid_table = kzalloc(sizeof(*pasid_table), GFP_KERNEL);
- if (!pasid_table)
- return -ENOMEM;
- if (info->pasid_supported)
- max_pasid = min_t(u32, pci_max_pasids(to_pci_dev(dev)),
- intel_pasid_max_id);
- size = max_pasid >> (PASID_PDE_SHIFT - 3);
- order = size ? get_order(size) : 0;
- pages = alloc_pages_node(info->iommu->node,
- GFP_KERNEL | __GFP_ZERO, order);
- if (!pages) {
- kfree(pasid_table);
- return -ENOMEM;
- }
- pasid_table->table = page_address(pages);
- pasid_table->order = order;
- pasid_table->max_pasid = 1 << (order + PAGE_SHIFT + 3);
- info->pasid_table = pasid_table;
- if (!ecap_coherent(info->iommu->ecap))
- clflush_cache_range(pasid_table->table, (1 << order) * PAGE_SIZE);
- return 0;
- }
- void intel_pasid_free_table(struct device *dev)
- {
- struct device_domain_info *info;
- struct pasid_table *pasid_table;
- struct pasid_dir_entry *dir;
- struct pasid_entry *table;
- int i, max_pde;
- info = dev_iommu_priv_get(dev);
- if (!info || !dev_is_pci(dev) || !info->pasid_table)
- return;
- pasid_table = info->pasid_table;
- info->pasid_table = NULL;
- /* Free scalable mode PASID directory tables: */
- dir = pasid_table->table;
- max_pde = pasid_table->max_pasid >> PASID_PDE_SHIFT;
- for (i = 0; i < max_pde; i++) {
- table = get_pasid_table_from_pde(&dir[i]);
- free_pgtable_page(table);
- }
- free_pages((unsigned long)pasid_table->table, pasid_table->order);
- kfree(pasid_table);
- }
- struct pasid_table *intel_pasid_get_table(struct device *dev)
- {
- struct device_domain_info *info;
- info = dev_iommu_priv_get(dev);
- if (!info)
- return NULL;
- return info->pasid_table;
- }
- static int intel_pasid_get_dev_max_id(struct device *dev)
- {
- struct device_domain_info *info;
- info = dev_iommu_priv_get(dev);
- if (!info || !info->pasid_table)
- return 0;
- return info->pasid_table->max_pasid;
- }
- static struct pasid_entry *intel_pasid_get_entry(struct device *dev, u32 pasid)
- {
- struct device_domain_info *info;
- struct pasid_table *pasid_table;
- struct pasid_dir_entry *dir;
- struct pasid_entry *entries;
- int dir_index, index;
- pasid_table = intel_pasid_get_table(dev);
- if (WARN_ON(!pasid_table || pasid >= intel_pasid_get_dev_max_id(dev)))
- return NULL;
- dir = pasid_table->table;
- info = dev_iommu_priv_get(dev);
- dir_index = pasid >> PASID_PDE_SHIFT;
- index = pasid & PASID_PTE_MASK;
- retry:
- entries = get_pasid_table_from_pde(&dir[dir_index]);
- if (!entries) {
- entries = alloc_pgtable_page(info->iommu->node);
- if (!entries)
- return NULL;
- /*
- * The pasid directory table entry won't be freed after
- * allocation. No worry about the race with free and
- * clear. However, this entry might be populated by others
- * while we are preparing it. Use theirs with a retry.
- */
- if (cmpxchg64(&dir[dir_index].val, 0ULL,
- (u64)virt_to_phys(entries) | PASID_PTE_PRESENT)) {
- free_pgtable_page(entries);
- goto retry;
- }
- if (!ecap_coherent(info->iommu->ecap)) {
- clflush_cache_range(entries, VTD_PAGE_SIZE);
- clflush_cache_range(&dir[dir_index].val, sizeof(*dir));
- }
- }
- return &entries[index];
- }
- /*
- * Interfaces for PASID table entry manipulation:
- */
- static inline void pasid_clear_entry(struct pasid_entry *pe)
- {
- WRITE_ONCE(pe->val[0], 0);
- WRITE_ONCE(pe->val[1], 0);
- WRITE_ONCE(pe->val[2], 0);
- WRITE_ONCE(pe->val[3], 0);
- WRITE_ONCE(pe->val[4], 0);
- WRITE_ONCE(pe->val[5], 0);
- WRITE_ONCE(pe->val[6], 0);
- WRITE_ONCE(pe->val[7], 0);
- }
- static inline void pasid_clear_entry_with_fpd(struct pasid_entry *pe)
- {
- WRITE_ONCE(pe->val[0], PASID_PTE_FPD);
- WRITE_ONCE(pe->val[1], 0);
- WRITE_ONCE(pe->val[2], 0);
- WRITE_ONCE(pe->val[3], 0);
- WRITE_ONCE(pe->val[4], 0);
- WRITE_ONCE(pe->val[5], 0);
- WRITE_ONCE(pe->val[6], 0);
- WRITE_ONCE(pe->val[7], 0);
- }
- static void
- intel_pasid_clear_entry(struct device *dev, u32 pasid, bool fault_ignore)
- {
- struct pasid_entry *pe;
- pe = intel_pasid_get_entry(dev, pasid);
- if (WARN_ON(!pe))
- return;
- if (fault_ignore && pasid_pte_is_present(pe))
- pasid_clear_entry_with_fpd(pe);
- else
- pasid_clear_entry(pe);
- }
- static inline void pasid_set_bits(u64 *ptr, u64 mask, u64 bits)
- {
- u64 old;
- old = READ_ONCE(*ptr);
- WRITE_ONCE(*ptr, (old & ~mask) | bits);
- }
- /*
- * Setup the DID(Domain Identifier) field (Bit 64~79) of scalable mode
- * PASID entry.
- */
- static inline void
- pasid_set_domain_id(struct pasid_entry *pe, u64 value)
- {
- pasid_set_bits(&pe->val[1], GENMASK_ULL(15, 0), value);
- }
- /*
- * Get domain ID value of a scalable mode PASID entry.
- */
- static inline u16
- pasid_get_domain_id(struct pasid_entry *pe)
- {
- return (u16)(READ_ONCE(pe->val[1]) & GENMASK_ULL(15, 0));
- }
- /*
- * Setup the SLPTPTR(Second Level Page Table Pointer) field (Bit 12~63)
- * of a scalable mode PASID entry.
- */
- static inline void
- pasid_set_slptr(struct pasid_entry *pe, u64 value)
- {
- pasid_set_bits(&pe->val[0], VTD_PAGE_MASK, value);
- }
- /*
- * Setup the AW(Address Width) field (Bit 2~4) of a scalable mode PASID
- * entry.
- */
- static inline void
- pasid_set_address_width(struct pasid_entry *pe, u64 value)
- {
- pasid_set_bits(&pe->val[0], GENMASK_ULL(4, 2), value << 2);
- }
- /*
- * Setup the PGTT(PASID Granular Translation Type) field (Bit 6~8)
- * of a scalable mode PASID entry.
- */
- static inline void
- pasid_set_translation_type(struct pasid_entry *pe, u64 value)
- {
- pasid_set_bits(&pe->val[0], GENMASK_ULL(8, 6), value << 6);
- }
- /*
- * Enable fault processing by clearing the FPD(Fault Processing
- * Disable) field (Bit 1) of a scalable mode PASID entry.
- */
- static inline void pasid_set_fault_enable(struct pasid_entry *pe)
- {
- pasid_set_bits(&pe->val[0], 1 << 1, 0);
- }
- /*
- * Setup the SRE(Supervisor Request Enable) field (Bit 128) of a
- * scalable mode PASID entry.
- */
- static inline void pasid_set_sre(struct pasid_entry *pe)
- {
- pasid_set_bits(&pe->val[2], 1 << 0, 1);
- }
- /*
- * Setup the WPE(Write Protect Enable) field (Bit 132) of a
- * scalable mode PASID entry.
- */
- static inline void pasid_set_wpe(struct pasid_entry *pe)
- {
- pasid_set_bits(&pe->val[2], 1 << 4, 1 << 4);
- }
- /*
- * Setup the P(Present) field (Bit 0) of a scalable mode PASID
- * entry.
- */
- static inline void pasid_set_present(struct pasid_entry *pe)
- {
- pasid_set_bits(&pe->val[0], 1 << 0, 1);
- }
- /*
- * Setup Page Walk Snoop bit (Bit 87) of a scalable mode PASID
- * entry.
- */
- static inline void pasid_set_page_snoop(struct pasid_entry *pe, bool value)
- {
- pasid_set_bits(&pe->val[1], 1 << 23, value << 23);
- }
- /*
- * Setup No Execute Enable bit (Bit 133) of a scalable mode PASID
- * entry. It is required when XD bit of the first level page table
- * entry is about to be set.
- */
- static inline void pasid_set_nxe(struct pasid_entry *pe)
- {
- pasid_set_bits(&pe->val[2], 1 << 5, 1 << 5);
- }
- /*
- * Setup the Page Snoop (PGSNP) field (Bit 88) of a scalable mode
- * PASID entry.
- */
- static inline void
- pasid_set_pgsnp(struct pasid_entry *pe)
- {
- pasid_set_bits(&pe->val[1], 1ULL << 24, 1ULL << 24);
- }
- /*
- * Setup the First Level Page table Pointer field (Bit 140~191)
- * of a scalable mode PASID entry.
- */
- static inline void
- pasid_set_flptr(struct pasid_entry *pe, u64 value)
- {
- pasid_set_bits(&pe->val[2], VTD_PAGE_MASK, value);
- }
- /*
- * Setup the First Level Paging Mode field (Bit 130~131) of a
- * scalable mode PASID entry.
- */
- static inline void
- pasid_set_flpm(struct pasid_entry *pe, u64 value)
- {
- pasid_set_bits(&pe->val[2], GENMASK_ULL(3, 2), value << 2);
- }
- static void
- pasid_cache_invalidation_with_pasid(struct intel_iommu *iommu,
- u16 did, u32 pasid)
- {
- struct qi_desc desc;
- desc.qw0 = QI_PC_DID(did) | QI_PC_GRAN(QI_PC_PASID_SEL) |
- QI_PC_PASID(pasid) | QI_PC_TYPE;
- desc.qw1 = 0;
- desc.qw2 = 0;
- desc.qw3 = 0;
- qi_submit_sync(iommu, &desc, 1, 0);
- }
- static void
- devtlb_invalidation_with_pasid(struct intel_iommu *iommu,
- struct device *dev, u32 pasid)
- {
- struct device_domain_info *info;
- u16 sid, qdep, pfsid;
- info = dev_iommu_priv_get(dev);
- if (!info || !info->ats_enabled)
- return;
- sid = info->bus << 8 | info->devfn;
- qdep = info->ats_qdep;
- pfsid = info->pfsid;
- /*
- * When PASID 0 is used, it indicates RID2PASID(DMA request w/o PASID),
- * devTLB flush w/o PASID should be used. For non-zero PASID under
- * SVA usage, device could do DMA with multiple PASIDs. It is more
- * efficient to flush devTLB specific to the PASID.
- */
- if (pasid == PASID_RID2PASID)
- qi_flush_dev_iotlb(iommu, sid, pfsid, qdep, 0, 64 - VTD_PAGE_SHIFT);
- else
- qi_flush_dev_iotlb_pasid(iommu, sid, pfsid, pasid, qdep, 0, 64 - VTD_PAGE_SHIFT);
- }
- void intel_pasid_tear_down_entry(struct intel_iommu *iommu, struct device *dev,
- u32 pasid, bool fault_ignore)
- {
- struct pasid_entry *pte;
- u16 did, pgtt;
- spin_lock(&iommu->lock);
- pte = intel_pasid_get_entry(dev, pasid);
- if (WARN_ON(!pte) || !pasid_pte_is_present(pte)) {
- spin_unlock(&iommu->lock);
- return;
- }
- did = pasid_get_domain_id(pte);
- pgtt = pasid_pte_get_pgtt(pte);
- intel_pasid_clear_entry(dev, pasid, fault_ignore);
- spin_unlock(&iommu->lock);
- if (!ecap_coherent(iommu->ecap))
- clflush_cache_range(pte, sizeof(*pte));
- pasid_cache_invalidation_with_pasid(iommu, did, pasid);
- if (pgtt == PASID_ENTRY_PGTT_PT || pgtt == PASID_ENTRY_PGTT_FL_ONLY)
- qi_flush_piotlb(iommu, did, pasid, 0, -1, 0);
- else
- iommu->flush.flush_iotlb(iommu, did, 0, 0, DMA_TLB_DSI_FLUSH);
- /* Device IOTLB doesn't need to be flushed in caching mode. */
- if (!cap_caching_mode(iommu->cap))
- devtlb_invalidation_with_pasid(iommu, dev, pasid);
- }
- /*
- * This function flushes cache for a newly setup pasid table entry.
- * Caller of it should not modify the in-use pasid table entries.
- */
- static void pasid_flush_caches(struct intel_iommu *iommu,
- struct pasid_entry *pte,
- u32 pasid, u16 did)
- {
- if (!ecap_coherent(iommu->ecap))
- clflush_cache_range(pte, sizeof(*pte));
- if (cap_caching_mode(iommu->cap)) {
- pasid_cache_invalidation_with_pasid(iommu, did, pasid);
- qi_flush_piotlb(iommu, did, pasid, 0, -1, 0);
- } else {
- iommu_flush_write_buffer(iommu);
- }
- }
- /*
- * Set up the scalable mode pasid table entry for first only
- * translation type.
- */
- int intel_pasid_setup_first_level(struct intel_iommu *iommu,
- struct device *dev, pgd_t *pgd,
- u32 pasid, u16 did, int flags)
- {
- struct pasid_entry *pte;
- if (!ecap_flts(iommu->ecap)) {
- pr_err("No first level translation support on %s\n",
- iommu->name);
- return -EINVAL;
- }
- if (flags & PASID_FLAG_SUPERVISOR_MODE) {
- #ifdef CONFIG_X86
- unsigned long cr0 = read_cr0();
- /* CR0.WP is normally set but just to be sure */
- if (unlikely(!(cr0 & X86_CR0_WP))) {
- pr_err("No CPU write protect!\n");
- return -EINVAL;
- }
- #endif
- if (!ecap_srs(iommu->ecap)) {
- pr_err("No supervisor request support on %s\n",
- iommu->name);
- return -EINVAL;
- }
- }
- if ((flags & PASID_FLAG_FL5LP) && !cap_fl5lp_support(iommu->cap)) {
- pr_err("No 5-level paging support for first-level on %s\n",
- iommu->name);
- return -EINVAL;
- }
- spin_lock(&iommu->lock);
- pte = intel_pasid_get_entry(dev, pasid);
- if (!pte) {
- spin_unlock(&iommu->lock);
- return -ENODEV;
- }
- if (pasid_pte_is_present(pte)) {
- spin_unlock(&iommu->lock);
- return -EBUSY;
- }
- pasid_clear_entry(pte);
- /* Setup the first level page table pointer: */
- pasid_set_flptr(pte, (u64)__pa(pgd));
- if (flags & PASID_FLAG_SUPERVISOR_MODE) {
- pasid_set_sre(pte);
- pasid_set_wpe(pte);
- }
- if (flags & PASID_FLAG_FL5LP)
- pasid_set_flpm(pte, 1);
- if (flags & PASID_FLAG_PAGE_SNOOP)
- pasid_set_pgsnp(pte);
- pasid_set_domain_id(pte, did);
- pasid_set_address_width(pte, iommu->agaw);
- pasid_set_page_snoop(pte, !!ecap_smpwc(iommu->ecap));
- pasid_set_nxe(pte);
- /* Setup Present and PASID Granular Transfer Type: */
- pasid_set_translation_type(pte, PASID_ENTRY_PGTT_FL_ONLY);
- pasid_set_present(pte);
- spin_unlock(&iommu->lock);
- pasid_flush_caches(iommu, pte, pasid, did);
- return 0;
- }
- /*
- * Skip top levels of page tables for iommu which has less agaw
- * than default. Unnecessary for PT mode.
- */
- static inline int iommu_skip_agaw(struct dmar_domain *domain,
- struct intel_iommu *iommu,
- struct dma_pte **pgd)
- {
- int agaw;
- for (agaw = domain->agaw; agaw > iommu->agaw; agaw--) {
- *pgd = phys_to_virt(dma_pte_addr(*pgd));
- if (!dma_pte_present(*pgd))
- return -EINVAL;
- }
- return agaw;
- }
- /*
- * Set up the scalable mode pasid entry for second only translation type.
- */
- int intel_pasid_setup_second_level(struct intel_iommu *iommu,
- struct dmar_domain *domain,
- struct device *dev, u32 pasid)
- {
- struct pasid_entry *pte;
- struct dma_pte *pgd;
- u64 pgd_val;
- int agaw;
- u16 did;
- /*
- * If hardware advertises no support for second level
- * translation, return directly.
- */
- if (!ecap_slts(iommu->ecap)) {
- pr_err("No second level translation support on %s\n",
- iommu->name);
- return -EINVAL;
- }
- pgd = domain->pgd;
- agaw = iommu_skip_agaw(domain, iommu, &pgd);
- if (agaw < 0) {
- dev_err(dev, "Invalid domain page table\n");
- return -EINVAL;
- }
- pgd_val = virt_to_phys(pgd);
- did = domain_id_iommu(domain, iommu);
- spin_lock(&iommu->lock);
- pte = intel_pasid_get_entry(dev, pasid);
- if (!pte) {
- spin_unlock(&iommu->lock);
- return -ENODEV;
- }
- if (pasid_pte_is_present(pte)) {
- spin_unlock(&iommu->lock);
- return -EBUSY;
- }
- pasid_clear_entry(pte);
- pasid_set_domain_id(pte, did);
- pasid_set_slptr(pte, pgd_val);
- pasid_set_address_width(pte, agaw);
- pasid_set_translation_type(pte, PASID_ENTRY_PGTT_SL_ONLY);
- pasid_set_fault_enable(pte);
- pasid_set_page_snoop(pte, !!ecap_smpwc(iommu->ecap));
- /*
- * Since it is a second level only translation setup, we should
- * set SRE bit as well (addresses are expected to be GPAs).
- */
- if (pasid != PASID_RID2PASID && ecap_srs(iommu->ecap))
- pasid_set_sre(pte);
- pasid_set_present(pte);
- spin_unlock(&iommu->lock);
- pasid_flush_caches(iommu, pte, pasid, did);
- return 0;
- }
- /*
- * Set up the scalable mode pasid entry for passthrough translation type.
- */
- int intel_pasid_setup_pass_through(struct intel_iommu *iommu,
- struct dmar_domain *domain,
- struct device *dev, u32 pasid)
- {
- u16 did = FLPT_DEFAULT_DID;
- struct pasid_entry *pte;
- spin_lock(&iommu->lock);
- pte = intel_pasid_get_entry(dev, pasid);
- if (!pte) {
- spin_unlock(&iommu->lock);
- return -ENODEV;
- }
- if (pasid_pte_is_present(pte)) {
- spin_unlock(&iommu->lock);
- return -EBUSY;
- }
- pasid_clear_entry(pte);
- pasid_set_domain_id(pte, did);
- pasid_set_address_width(pte, iommu->agaw);
- pasid_set_translation_type(pte, PASID_ENTRY_PGTT_PT);
- pasid_set_fault_enable(pte);
- pasid_set_page_snoop(pte, !!ecap_smpwc(iommu->ecap));
- /*
- * We should set SRE bit as well since the addresses are expected
- * to be GPAs.
- */
- if (ecap_srs(iommu->ecap))
- pasid_set_sre(pte);
- pasid_set_present(pte);
- spin_unlock(&iommu->lock);
- pasid_flush_caches(iommu, pte, pasid, did);
- return 0;
- }
- /*
- * Set the page snoop control for a pasid entry which has been set up.
- */
- void intel_pasid_setup_page_snoop_control(struct intel_iommu *iommu,
- struct device *dev, u32 pasid)
- {
- struct pasid_entry *pte;
- u16 did;
- spin_lock(&iommu->lock);
- pte = intel_pasid_get_entry(dev, pasid);
- if (WARN_ON(!pte || !pasid_pte_is_present(pte))) {
- spin_unlock(&iommu->lock);
- return;
- }
- pasid_set_pgsnp(pte);
- did = pasid_get_domain_id(pte);
- spin_unlock(&iommu->lock);
- if (!ecap_coherent(iommu->ecap))
- clflush_cache_range(pte, sizeof(*pte));
- /*
- * VT-d spec 3.4 table23 states guides for cache invalidation:
- *
- * - PASID-selective-within-Domain PASID-cache invalidation
- * - PASID-selective PASID-based IOTLB invalidation
- * - If (pasid is RID_PASID)
- * - Global Device-TLB invalidation to affected functions
- * Else
- * - PASID-based Device-TLB invalidation (with S=1 and
- * Addr[63:12]=0x7FFFFFFF_FFFFF) to affected functions
- */
- pasid_cache_invalidation_with_pasid(iommu, did, pasid);
- qi_flush_piotlb(iommu, did, pasid, 0, -1, 0);
- /* Device IOTLB doesn't need to be flushed in caching mode. */
- if (!cap_caching_mode(iommu->cap))
- devtlb_invalidation_with_pasid(iommu, dev, pasid);
- }
|