Loading...
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 | /* SPDX-License-Identifier: GPL-2.0-or-later */ /* * Copyright 2016,2017 IBM Corporation. */ #ifndef _ASM_POWERPC_XIVE_H #define _ASM_POWERPC_XIVE_H #define XIVE_INVALID_VP 0xffffffff #ifdef CONFIG_PPC_XIVE /* * Thread Interrupt Management Area (TIMA) * * This is a global MMIO region divided in 4 pages of varying access * permissions, providing access to per-cpu interrupt management * functions. It always identifies the CPU doing the access based * on the PowerBus initiator ID, thus we always access via the * same offset regardless of where the code is executing */ extern void __iomem *xive_tima; extern unsigned long xive_tima_os; /* * Offset in the TM area of our current execution level (provided by * the backend) */ extern u32 xive_tima_offset; /* * Per-irq data (irq_get_handler_data for normal IRQs), IPIs * have it stored in the xive_cpu structure. We also cache * for normal interrupts the current target CPU. * * This structure is setup by the backend for each interrupt. */ struct xive_irq_data { u64 flags; u64 eoi_page; void __iomem *eoi_mmio; u64 trig_page; void __iomem *trig_mmio; u32 esb_shift; int src_chip; u32 hw_irq; /* Setup/used by frontend */ int target; /* * saved_p means that there is a queue entry for this interrupt * in some CPU's queue (not including guest vcpu queues), even * if P is not set in the source ESB. * stale_p means that there is no queue entry for this interrupt * in some CPU's queue, even if P is set in the source ESB. */ bool saved_p; bool stale_p; }; #define XIVE_IRQ_FLAG_STORE_EOI 0x01 #define XIVE_IRQ_FLAG_LSI 0x02 #define XIVE_IRQ_FLAG_SHIFT_BUG 0x04 #define XIVE_IRQ_FLAG_MASK_FW 0x08 #define XIVE_IRQ_FLAG_EOI_FW 0x10 #define XIVE_IRQ_FLAG_H_INT_ESB 0x20 /* Special flag set by KVM for excalation interrupts */ #define XIVE_IRQ_NO_EOI 0x80 #define XIVE_INVALID_CHIP_ID -1 /* A queue tracking structure in a CPU */ struct xive_q { __be32 *qpage; u32 msk; u32 idx; u32 toggle; u64 eoi_phys; u32 esc_irq; atomic_t count; atomic_t pending_count; u64 guest_qaddr; u32 guest_qshift; }; /* Global enable flags for the XIVE support */ extern bool __xive_enabled; static inline bool xive_enabled(void) { return __xive_enabled; } extern bool xive_spapr_init(void); extern bool xive_native_init(void); extern void xive_smp_probe(void); extern int xive_smp_prepare_cpu(unsigned int cpu); extern void xive_smp_setup_cpu(void); extern void xive_smp_disable_cpu(void); extern void xive_teardown_cpu(void); extern void xive_shutdown(void); extern void xive_flush_interrupt(void); /* xmon hook */ extern void xmon_xive_do_dump(int cpu); extern int xmon_xive_get_irq_config(u32 hw_irq, struct irq_data *d); /* APIs used by KVM */ extern u32 xive_native_default_eq_shift(void); extern u32 xive_native_alloc_vp_block(u32 max_vcpus); extern void xive_native_free_vp_block(u32 vp_base); extern int xive_native_populate_irq_data(u32 hw_irq, struct xive_irq_data *data); extern void xive_cleanup_irq_data(struct xive_irq_data *xd); extern u32 xive_native_alloc_irq(void); extern void xive_native_free_irq(u32 irq); extern int xive_native_configure_irq(u32 hw_irq, u32 target, u8 prio, u32 sw_irq); extern int xive_native_configure_queue(u32 vp_id, struct xive_q *q, u8 prio, __be32 *qpage, u32 order, bool can_escalate); extern void xive_native_disable_queue(u32 vp_id, struct xive_q *q, u8 prio); extern void xive_native_sync_source(u32 hw_irq); extern void xive_native_sync_queue(u32 hw_irq); extern bool is_xive_irq(struct irq_chip *chip); extern int xive_native_enable_vp(u32 vp_id, bool single_escalation); extern int xive_native_disable_vp(u32 vp_id); extern int xive_native_get_vp_info(u32 vp_id, u32 *out_cam_id, u32 *out_chip_id); extern bool xive_native_has_single_escalation(void); extern int xive_native_get_queue_info(u32 vp_id, uint32_t prio, u64 *out_qpage, u64 *out_qsize, u64 *out_qeoi_page, u32 *out_escalate_irq, u64 *out_qflags); extern int xive_native_get_queue_state(u32 vp_id, uint32_t prio, u32 *qtoggle, u32 *qindex); extern int xive_native_set_queue_state(u32 vp_id, uint32_t prio, u32 qtoggle, u32 qindex); extern int xive_native_get_vp_state(u32 vp_id, u64 *out_state); extern bool xive_native_has_queue_state_support(void); #else static inline bool xive_enabled(void) { return false; } static inline bool xive_spapr_init(void) { return false; } static inline bool xive_native_init(void) { return false; } static inline void xive_smp_probe(void) { } static inline int xive_smp_prepare_cpu(unsigned int cpu) { return -EINVAL; } static inline void xive_smp_setup_cpu(void) { } static inline void xive_smp_disable_cpu(void) { } static inline void xive_kexec_teardown_cpu(int secondary) { } static inline void xive_shutdown(void) { } static inline void xive_flush_interrupt(void) { } static inline u32 xive_native_alloc_vp_block(u32 max_vcpus) { return XIVE_INVALID_VP; } static inline void xive_native_free_vp_block(u32 vp_base) { } #endif #endif /* _ASM_POWERPC_XIVE_H */ |