Loading arch/riscv/include/asm/cacheflush.h +0 −8 Original line number Diff line number Diff line Loading @@ -42,16 +42,8 @@ void flush_icache_mm(struct mm_struct *mm, bool local); #endif /* CONFIG_SMP */ /* * The T-Head CMO errata internally probe the CBOM block size, but otherwise * don't depend on Zicbom. */ extern unsigned int riscv_cbom_block_size; #ifdef CONFIG_RISCV_ISA_ZICBOM void riscv_init_cbom_blocksize(void); #else static inline void riscv_init_cbom_blocksize(void) { } #endif #ifdef CONFIG_RISCV_DMA_NONCOHERENT void riscv_noncoherent_supported(void); Loading arch/riscv/include/asm/kvm_vcpu_timer.h +1 −0 Original line number Diff line number Diff line Loading @@ -45,6 +45,7 @@ int kvm_riscv_vcpu_timer_deinit(struct kvm_vcpu *vcpu); int kvm_riscv_vcpu_timer_reset(struct kvm_vcpu *vcpu); void kvm_riscv_vcpu_timer_restore(struct kvm_vcpu *vcpu); void kvm_riscv_guest_timer_init(struct kvm *kvm); void kvm_riscv_vcpu_timer_sync(struct kvm_vcpu *vcpu); void kvm_riscv_vcpu_timer_save(struct kvm_vcpu *vcpu); bool kvm_riscv_vcpu_timer_pending(struct kvm_vcpu *vcpu); Loading arch/riscv/kvm/vcpu.c +3 −0 Original line number Diff line number Diff line Loading @@ -708,6 +708,9 @@ void kvm_riscv_vcpu_sync_interrupts(struct kvm_vcpu *vcpu) clear_bit(IRQ_VS_SOFT, &v->irqs_pending); } } /* Sync-up timer CSRs */ kvm_riscv_vcpu_timer_sync(vcpu); } int kvm_riscv_vcpu_set_interrupt(struct kvm_vcpu *vcpu, unsigned int irq) Loading arch/riscv/kvm/vcpu_timer.c +15 −2 Original line number Diff line number Diff line Loading @@ -320,20 +320,33 @@ void kvm_riscv_vcpu_timer_restore(struct kvm_vcpu *vcpu) kvm_riscv_vcpu_timer_unblocking(vcpu); } void kvm_riscv_vcpu_timer_save(struct kvm_vcpu *vcpu) void kvm_riscv_vcpu_timer_sync(struct kvm_vcpu *vcpu) { struct kvm_vcpu_timer *t = &vcpu->arch.timer; if (!t->sstc_enabled) return; t = &vcpu->arch.timer; #if defined(CONFIG_32BIT) t->next_cycles = csr_read(CSR_VSTIMECMP); t->next_cycles |= (u64)csr_read(CSR_VSTIMECMPH) << 32; #else t->next_cycles = csr_read(CSR_VSTIMECMP); #endif } void kvm_riscv_vcpu_timer_save(struct kvm_vcpu *vcpu) { struct kvm_vcpu_timer *t = &vcpu->arch.timer; if (!t->sstc_enabled) return; /* * The vstimecmp CSRs are saved by kvm_riscv_vcpu_timer_sync() * upon every VM exit so no need to save here. */ /* timer should be enabled for the remaining operations */ if (unlikely(!t->init_done)) return; Loading arch/riscv/mm/cacheflush.c +38 −0 Original line number Diff line number Diff line Loading @@ -3,6 +3,7 @@ * Copyright (C) 2017 SiFive */ #include <linux/of.h> #include <asm/cacheflush.h> #ifdef CONFIG_SMP Loading Loading @@ -86,3 +87,40 @@ void flush_icache_pte(pte_t pte) flush_icache_all(); } #endif /* CONFIG_MMU */ unsigned int riscv_cbom_block_size; EXPORT_SYMBOL_GPL(riscv_cbom_block_size); void riscv_init_cbom_blocksize(void) { struct device_node *node; unsigned long cbom_hartid; u32 val, probed_block_size; int ret; probed_block_size = 0; for_each_of_cpu_node(node) { unsigned long hartid; ret = riscv_of_processor_hartid(node, &hartid); if (ret) continue; /* set block-size for cbom extension if available */ ret = of_property_read_u32(node, "riscv,cbom-block-size", &val); if (ret) continue; if (!probed_block_size) { probed_block_size = val; cbom_hartid = hartid; } else { if (probed_block_size != val) pr_warn("cbom-block-size mismatched between harts %lu and %lu\n", cbom_hartid, hartid); } } if (probed_block_size) riscv_cbom_block_size = probed_block_size; } Loading
arch/riscv/include/asm/cacheflush.h +0 −8 Original line number Diff line number Diff line Loading @@ -42,16 +42,8 @@ void flush_icache_mm(struct mm_struct *mm, bool local); #endif /* CONFIG_SMP */ /* * The T-Head CMO errata internally probe the CBOM block size, but otherwise * don't depend on Zicbom. */ extern unsigned int riscv_cbom_block_size; #ifdef CONFIG_RISCV_ISA_ZICBOM void riscv_init_cbom_blocksize(void); #else static inline void riscv_init_cbom_blocksize(void) { } #endif #ifdef CONFIG_RISCV_DMA_NONCOHERENT void riscv_noncoherent_supported(void); Loading
arch/riscv/include/asm/kvm_vcpu_timer.h +1 −0 Original line number Diff line number Diff line Loading @@ -45,6 +45,7 @@ int kvm_riscv_vcpu_timer_deinit(struct kvm_vcpu *vcpu); int kvm_riscv_vcpu_timer_reset(struct kvm_vcpu *vcpu); void kvm_riscv_vcpu_timer_restore(struct kvm_vcpu *vcpu); void kvm_riscv_guest_timer_init(struct kvm *kvm); void kvm_riscv_vcpu_timer_sync(struct kvm_vcpu *vcpu); void kvm_riscv_vcpu_timer_save(struct kvm_vcpu *vcpu); bool kvm_riscv_vcpu_timer_pending(struct kvm_vcpu *vcpu); Loading
arch/riscv/kvm/vcpu.c +3 −0 Original line number Diff line number Diff line Loading @@ -708,6 +708,9 @@ void kvm_riscv_vcpu_sync_interrupts(struct kvm_vcpu *vcpu) clear_bit(IRQ_VS_SOFT, &v->irqs_pending); } } /* Sync-up timer CSRs */ kvm_riscv_vcpu_timer_sync(vcpu); } int kvm_riscv_vcpu_set_interrupt(struct kvm_vcpu *vcpu, unsigned int irq) Loading
arch/riscv/kvm/vcpu_timer.c +15 −2 Original line number Diff line number Diff line Loading @@ -320,20 +320,33 @@ void kvm_riscv_vcpu_timer_restore(struct kvm_vcpu *vcpu) kvm_riscv_vcpu_timer_unblocking(vcpu); } void kvm_riscv_vcpu_timer_save(struct kvm_vcpu *vcpu) void kvm_riscv_vcpu_timer_sync(struct kvm_vcpu *vcpu) { struct kvm_vcpu_timer *t = &vcpu->arch.timer; if (!t->sstc_enabled) return; t = &vcpu->arch.timer; #if defined(CONFIG_32BIT) t->next_cycles = csr_read(CSR_VSTIMECMP); t->next_cycles |= (u64)csr_read(CSR_VSTIMECMPH) << 32; #else t->next_cycles = csr_read(CSR_VSTIMECMP); #endif } void kvm_riscv_vcpu_timer_save(struct kvm_vcpu *vcpu) { struct kvm_vcpu_timer *t = &vcpu->arch.timer; if (!t->sstc_enabled) return; /* * The vstimecmp CSRs are saved by kvm_riscv_vcpu_timer_sync() * upon every VM exit so no need to save here. */ /* timer should be enabled for the remaining operations */ if (unlikely(!t->init_done)) return; Loading
arch/riscv/mm/cacheflush.c +38 −0 Original line number Diff line number Diff line Loading @@ -3,6 +3,7 @@ * Copyright (C) 2017 SiFive */ #include <linux/of.h> #include <asm/cacheflush.h> #ifdef CONFIG_SMP Loading Loading @@ -86,3 +87,40 @@ void flush_icache_pte(pte_t pte) flush_icache_all(); } #endif /* CONFIG_MMU */ unsigned int riscv_cbom_block_size; EXPORT_SYMBOL_GPL(riscv_cbom_block_size); void riscv_init_cbom_blocksize(void) { struct device_node *node; unsigned long cbom_hartid; u32 val, probed_block_size; int ret; probed_block_size = 0; for_each_of_cpu_node(node) { unsigned long hartid; ret = riscv_of_processor_hartid(node, &hartid); if (ret) continue; /* set block-size for cbom extension if available */ ret = of_property_read_u32(node, "riscv,cbom-block-size", &val); if (ret) continue; if (!probed_block_size) { probed_block_size = val; cbom_hartid = hartid; } else { if (probed_block_size != val) pr_warn("cbom-block-size mismatched between harts %lu and %lu\n", cbom_hartid, hartid); } } if (probed_block_size) riscv_cbom_block_size = probed_block_size; }