/* SPDX-License-Identifier: GPL-2.0-only OR MIT */ /* Copyright (c) 2023 Imagination Technologies Ltd. */ #ifndef PVR_ROGUE_FWIF_H #define PVR_ROGUE_FWIF_H #include #include #include #include #include #include "pvr_rogue_defs.h" #include "pvr_rogue_fwif_common.h" #include "pvr_rogue_fwif_shared.h" /* **************************************************************************** * Logging type **************************************************************************** */ #define ROGUE_FWIF_LOG_TYPE_NONE 0x00000000U #define ROGUE_FWIF_LOG_TYPE_TRACE 0x00000001U #define ROGUE_FWIF_LOG_TYPE_GROUP_MAIN 0x00000002U #define ROGUE_FWIF_LOG_TYPE_GROUP_MTS 0x00000004U #define ROGUE_FWIF_LOG_TYPE_GROUP_CLEANUP 0x00000008U #define ROGUE_FWIF_LOG_TYPE_GROUP_CSW 0x00000010U #define ROGUE_FWIF_LOG_TYPE_GROUP_BIF 0x00000020U #define ROGUE_FWIF_LOG_TYPE_GROUP_PM 0x00000040U #define ROGUE_FWIF_LOG_TYPE_GROUP_RTD 0x00000080U #define ROGUE_FWIF_LOG_TYPE_GROUP_SPM 0x00000100U #define ROGUE_FWIF_LOG_TYPE_GROUP_POW 0x00000200U #define ROGUE_FWIF_LOG_TYPE_GROUP_HWR 0x00000400U #define ROGUE_FWIF_LOG_TYPE_GROUP_HWP 0x00000800U #define ROGUE_FWIF_LOG_TYPE_GROUP_RPM 0x00001000U #define ROGUE_FWIF_LOG_TYPE_GROUP_DMA 0x00002000U #define ROGUE_FWIF_LOG_TYPE_GROUP_MISC 0x00004000U #define ROGUE_FWIF_LOG_TYPE_GROUP_DEBUG 0x80000000U #define ROGUE_FWIF_LOG_TYPE_GROUP_MASK 0x80007FFEU #define ROGUE_FWIF_LOG_TYPE_MASK 0x80007FFFU /* String used in pvrdebug -h output */ #define ROGUE_FWIF_LOG_GROUPS_STRING_LIST \ "main,mts,cleanup,csw,bif,pm,rtd,spm,pow,hwr,hwp,rpm,dma,misc,debug" /* Table entry to map log group strings to log type value */ struct rogue_fwif_log_group_map_entry { const char *log_group_name; u32 log_group_type; }; /* **************************************************************************** * ROGUE FW signature checks **************************************************************************** */ #define ROGUE_FW_SIG_BUFFER_SIZE_MIN (8192) #define ROGUE_FWIF_TIMEDIFF_ID ((0x1UL << 28) | ROGUE_CR_TIMER) /* **************************************************************************** * Trace Buffer **************************************************************************** */ /* Default size of ROGUE_FWIF_TRACEBUF_SPACE in DWords */ #define ROGUE_FW_TRACE_BUF_DEFAULT_SIZE_IN_DWORDS 12000U #define ROGUE_FW_TRACE_BUFFER_ASSERT_SIZE 200U #define ROGUE_FW_THREAD_NUM 1U #define ROGUE_FW_THREAD_MAX 2U #define ROGUE_FW_POLL_TYPE_SET 0x80000000U struct rogue_fwif_file_info_buf { char path[ROGUE_FW_TRACE_BUFFER_ASSERT_SIZE]; char info[ROGUE_FW_TRACE_BUFFER_ASSERT_SIZE]; u32 line_num; u32 padding; } __aligned(8); struct rogue_fwif_tracebuf_space { u32 trace_pointer; u32 trace_buffer_fw_addr; /* To be used by host when reading from trace buffer */ u32 *trace_buffer; struct rogue_fwif_file_info_buf assert_buf; } __aligned(8); /* Total number of FW fault logs stored */ #define ROGUE_FWIF_FWFAULTINFO_MAX (8U) struct rogue_fw_fault_info { aligned_u64 cr_timer; aligned_u64 os_timer; u32 data __aligned(8); u32 reserved; struct rogue_fwif_file_info_buf fault_buf; } __aligned(8); enum rogue_fwif_pow_state { ROGUE_FWIF_POW_OFF, /* idle and ready to full power down */ ROGUE_FWIF_POW_ON, /* running HW commands */ ROGUE_FWIF_POW_FORCED_IDLE, /* forced idle */ ROGUE_FWIF_POW_IDLE, /* idle waiting for host handshake */ }; /* Firmware HWR states */ /* The HW state is ok or locked up */ #define ROGUE_FWIF_HWR_HARDWARE_OK BIT(0) /* Tells if a HWR reset is in progress */ #define ROGUE_FWIF_HWR_RESET_IN_PROGRESS BIT(1) /* A DM unrelated lockup has been detected */ #define ROGUE_FWIF_HWR_GENERAL_LOCKUP BIT(3) /* At least one DM is running without being close to a lockup */ #define ROGUE_FWIF_HWR_DM_RUNNING_OK BIT(4) /* At least one DM is close to lockup */ #define ROGUE_FWIF_HWR_DM_STALLING BIT(5) /* The FW has faulted and needs to restart */ #define ROGUE_FWIF_HWR_FW_FAULT BIT(6) /* The FW has requested the host to restart it */ #define ROGUE_FWIF_HWR_RESTART_REQUESTED BIT(7) #define ROGUE_FWIF_PHR_STATE_SHIFT (8U) /* The FW has requested the host to restart it, per PHR configuration */ #define ROGUE_FWIF_PHR_RESTART_REQUESTED ((1) << ROGUE_FWIF_PHR_STATE_SHIFT) /* A PHR triggered GPU reset has just finished */ #define ROGUE_FWIF_PHR_RESTART_FINISHED ((2) << ROGUE_FWIF_PHR_STATE_SHIFT) #define ROGUE_FWIF_PHR_RESTART_MASK \ (ROGUE_FWIF_PHR_RESTART_REQUESTED | ROGUE_FWIF_PHR_RESTART_FINISHED) #define ROGUE_FWIF_PHR_MODE_OFF (0UL) #define ROGUE_FWIF_PHR_MODE_RD_RESET (1UL) #define ROGUE_FWIF_PHR_MODE_FULL_RESET (2UL) /* Firmware per-DM HWR states */ /* DM is working if all flags are cleared */ #define ROGUE_FWIF_DM_STATE_WORKING (0) /* DM is idle and ready for HWR */ #define ROGUE_FWIF_DM_STATE_READY_FOR_HWR BIT(0) /* DM need to skip to next cmd before resuming processing */ #define ROGUE_FWIF_DM_STATE_NEEDS_SKIP BIT(2) /* DM need partial render cleanup before resuming processing */ #define ROGUE_FWIF_DM_STATE_NEEDS_PR_CLEANUP BIT(3) /* DM need to increment Recovery Count once fully recovered */ #define ROGUE_FWIF_DM_STATE_NEEDS_TRACE_CLEAR BIT(4) /* DM was identified as locking up and causing HWR */ #define ROGUE_FWIF_DM_STATE_GUILTY_LOCKUP BIT(5) /* DM was innocently affected by another lockup which caused HWR */ #define ROGUE_FWIF_DM_STATE_INNOCENT_LOCKUP BIT(6) /* DM was identified as over-running and causing HWR */ #define ROGUE_FWIF_DM_STATE_GUILTY_OVERRUNING BIT(7) /* DM was innocently affected by another DM over-running which caused HWR */ #define ROGUE_FWIF_DM_STATE_INNOCENT_OVERRUNING BIT(8) /* DM was forced into HWR as it delayed more important workloads */ #define ROGUE_FWIF_DM_STATE_HARD_CONTEXT_SWITCH BIT(9) /* DM was forced into HWR due to an uncorrected GPU ECC error */ #define ROGUE_FWIF_DM_STATE_GPU_ECC_HWR BIT(10) /* Firmware's connection state */ enum rogue_fwif_connection_fw_state { /* Firmware is offline */ ROGUE_FW_CONNECTION_FW_OFFLINE = 0, /* Firmware is initialised */ ROGUE_FW_CONNECTION_FW_READY, /* Firmware connection is fully established */ ROGUE_FW_CONNECTION_FW_ACTIVE, /* Firmware is clearing up connection data*/ ROGUE_FW_CONNECTION_FW_OFFLOADING, ROGUE_FW_CONNECTION_FW_STATE_COUNT }; /* OS' connection state */ enum rogue_fwif_connection_os_state { /* OS is offline */ ROGUE_FW_CONNECTION_OS_OFFLINE = 0, /* OS's KM driver is setup and waiting */ ROGUE_FW_CONNECTION_OS_READY, /* OS connection is fully established */ ROGUE_FW_CONNECTION_OS_ACTIVE, ROGUE_FW_CONNECTION_OS_STATE_COUNT }; struct rogue_fwif_os_runtime_flags { unsigned int os_state : 3; unsigned int fl_ok : 1; unsigned int fl_grow_pending : 1; unsigned int isolated_os : 1; unsigned int reserved : 26; }; #define PVR_SLR_LOG_ENTRIES 10 /* MAX_CLIENT_CCB_NAME not visible to this header */ #define PVR_SLR_LOG_STRLEN 30 struct rogue_fwif_slr_entry { aligned_u64 timestamp; u32 fw_ctx_addr; u32 num_ufos; char ccb_name[PVR_SLR_LOG_STRLEN]; char padding[2]; } __aligned(8); #define MAX_THREAD_NUM 2 /* firmware trace control data */ struct rogue_fwif_tracebuf { u32 log_type; struct rogue_fwif_tracebuf_space tracebuf[MAX_THREAD_NUM]; /* * Member initialised only when sTraceBuf is actually allocated (in * ROGUETraceBufferInitOnDemandResources) */ u32 tracebuf_size_in_dwords; /* Compatibility and other flags */ u32 tracebuf_flags; } __aligned(8); /* firmware system data shared with the Host driver */ struct rogue_fwif_sysdata { /* Configuration flags from host */ u32 config_flags; /* Extended configuration flags from host */ u32 config_flags_ext; enum rogue_fwif_pow_state pow_state; u32 hw_perf_ridx; u32 hw_perf_widx; u32 hw_perf_wrap_count; /* Constant after setup, needed in FW */ u32 hw_perf_size; /* The number of times the FW drops a packet due to buffer full */ u32 hw_perf_drop_count; /* * ui32HWPerfUt, ui32FirstDropOrdinal, ui32LastDropOrdinal only valid * when FW is built with ROGUE_HWPERF_UTILIZATION & * ROGUE_HWPERF_DROP_TRACKING defined in rogue_fw_hwperf.c */ /* Buffer utilisation, high watermark of bytes in use */ u32 hw_perf_ut; /* The ordinal of the first packet the FW dropped */ u32 first_drop_ordinal; /* The ordinal of the last packet the FW dropped */ u32 last_drop_ordinal; /* State flags for each Operating System mirrored from Fw coremem */ struct rogue_fwif_os_runtime_flags os_runtime_flags_mirror[ROGUE_FW_MAX_NUM_OS]; struct rogue_fw_fault_info fault_info[ROGUE_FWIF_FWFAULTINFO_MAX]; u32 fw_faults; u32 cr_poll_addr[MAX_THREAD_NUM]; u32 cr_poll_mask[MAX_THREAD_NUM]; u32 cr_poll_count[MAX_THREAD_NUM]; aligned_u64 start_idle_time; #if defined(SUPPORT_ROGUE_FW_STATS_FRAMEWORK) # define ROGUE_FWIF_STATS_FRAMEWORK_LINESIZE (8) # define ROGUE_FWIF_STATS_FRAMEWORK_MAX \ (2048 * ROGUE_FWIF_STATS_FRAMEWORK_LINESIZE) u32 fw_stats_buf[ROGUE_FWIF_STATS_FRAMEWORK_MAX] __aligned(8); #endif u32 hwr_state_flags; u32 hwr_recovery_flags[PVR_FWIF_DM_MAX]; /* Compatibility and other flags */ u32 fw_sys_data_flags; /* Identify whether MC config is P-P or P-S */ u32 mc_config; } __aligned(8); /* per-os firmware shared data */ struct rogue_fwif_osdata { /* Configuration flags from an OS */ u32 fw_os_config_flags; /* Markers to signal that the host should perform a full sync check */ u32 fw_sync_check_mark; u32 host_sync_check_mark; u32 forced_updates_requested; u8 slr_log_wp; struct rogue_fwif_slr_entry slr_log_first; struct rogue_fwif_slr_entry slr_log[PVR_SLR_LOG_ENTRIES]; aligned_u64 last_forced_update_time; /* Interrupt count from Threads > */ u32 interrupt_count[MAX_THREAD_NUM]; u32 kccb_cmds_executed; u32 power_sync_fw_addr; /* Compatibility and other flags */ u32 fw_os_data_flags; u32 padding; } __aligned(8); /* Firmware trace time-stamp field breakup */ /* ROGUE_CR_TIMER register read (48 bits) value*/ #define ROGUE_FWT_TIMESTAMP_TIME_SHIFT (0U) #define ROGUE_FWT_TIMESTAMP_TIME_CLRMSK (0xFFFF000000000000ull) /* Extra debug-info (16 bits) */ #define ROGUE_FWT_TIMESTAMP_DEBUG_INFO_SHIFT (48U) #define ROGUE_FWT_TIMESTAMP_DEBUG_INFO_CLRMSK ~ROGUE_FWT_TIMESTAMP_TIME_CLRMSK /* Debug-info sub-fields */ /* * Bit 0: ROGUE_CR_EVENT_STATUS_MMU_PAGE_FAULT bit from ROGUE_CR_EVENT_STATUS * register */ #define ROGUE_FWT_DEBUG_INFO_MMU_PAGE_FAULT_SHIFT (0U) #define ROGUE_FWT_DEBUG_INFO_MMU_PAGE_FAULT_SET \ BIT(ROGUE_FWT_DEBUG_INFO_MMU_PAGE_FAULT_SHIFT) /* Bit 1: ROGUE_CR_BIF_MMU_ENTRY_PENDING bit from ROGUE_CR_BIF_MMU_ENTRY register */ #define ROGUE_FWT_DEBUG_INFO_MMU_ENTRY_PENDING_SHIFT (1U) #define ROGUE_FWT_DEBUG_INFO_MMU_ENTRY_PENDING_SET \ BIT(ROGUE_FWT_DEBUG_INFO_MMU_ENTRY_PENDING_SHIFT) /* Bit 2: ROGUE_CR_SLAVE_EVENT register is non-zero */ #define ROGUE_FWT_DEBUG_INFO_SLAVE_EVENTS_SHIFT (2U) #define ROGUE_FWT_DEBUG_INFO_SLAVE_EVENTS_SET \ BIT(ROGUE_FWT_DEBUG_INFO_SLAVE_EVENTS_SHIFT) /* Bit 3-15: Unused bits */ #define ROGUE_FWT_DEBUG_INFO_STR_MAXLEN 64 #define ROGUE_FWT_DEBUG_INFO_STR_PREPEND " (debug info: " #define ROGUE_FWT_DEBUG_INFO_STR_APPEND ")" /* ****************************************************************************** * HWR Data ****************************************************************************** */ enum rogue_hwrtype { ROGUE_HWRTYPE_UNKNOWNFAILURE = 0, ROGUE_HWRTYPE_OVERRUN = 1, ROGUE_HWRTYPE_POLLFAILURE = 2, ROGUE_HWRTYPE_BIF0FAULT = 3, ROGUE_HWRTYPE_BIF1FAULT = 4, ROGUE_HWRTYPE_TEXASBIF0FAULT = 5, ROGUE_HWRTYPE_MMUFAULT = 6, ROGUE_HWRTYPE_MMUMETAFAULT = 7, ROGUE_HWRTYPE_MIPSTLBFAULT = 8, ROGUE_HWRTYPE_ECCFAULT = 9, ROGUE_HWRTYPE_MMURISCVFAULT = 10, }; #define ROGUE_FWIF_HWRTYPE_BIF_BANK_GET(hwr_type) \ (((hwr_type) == ROGUE_HWRTYPE_BIF0FAULT) ? 0 : 1) #define ROGUE_FWIF_HWRTYPE_PAGE_FAULT_GET(hwr_type) \ ((((hwr_type) == ROGUE_HWRTYPE_BIF0FAULT) || \ ((hwr_type) == ROGUE_HWRTYPE_BIF1FAULT) || \ ((hwr_type) == ROGUE_HWRTYPE_TEXASBIF0FAULT) || \ ((hwr_type) == ROGUE_HWRTYPE_MMUFAULT) || \ ((hwr_type) == ROGUE_HWRTYPE_MMUMETAFAULT) || \ ((hwr_type) == ROGUE_HWRTYPE_MIPSTLBFAULT) || \ ((hwr_type) == ROGUE_HWRTYPE_MMURISCVFAULT)) \ ? true \ : false) struct rogue_bifinfo { aligned_u64 bif_req_status; aligned_u64 bif_mmu_status; aligned_u64 pc_address; /* phys address of the page catalogue */ aligned_u64 reserved; }; struct rogue_eccinfo { u32 fault_gpu; }; struct rogue_mmuinfo { aligned_u64 mmu_status[2]; aligned_u64 pc_address; /* phys address of the page catalogue */ aligned_u64 reserved; }; struct rogue_pollinfo { u32 thread_num; u32 cr_poll_addr; u32 cr_poll_mask; u32 cr_poll_last_value; aligned_u64 reserved; } __aligned(8); struct rogue_tlbinfo { u32 bad_addr; u32 entry_lo; }; struct rogue_hwrinfo { union { struct rogue_bifinfo bif_info; struct rogue_mmuinfo mmu_info; struct rogue_pollinfo poll_info; struct rogue_tlbinfo tlb_info; struct rogue_eccinfo ecc_info; } hwr_data; aligned_u64 cr_timer; aligned_u64 os_timer; u32 frame_num; u32 pid; u32 active_hwrt_data; u32 hwr_number; u32 event_status; u32 hwr_recovery_flags; enum rogue_hwrtype hwr_type; u32 dm; u32 core_id; aligned_u64 cr_time_of_kick; aligned_u64 cr_time_hw_reset_start; aligned_u64 cr_time_hw_reset_finish; aligned_u64 cr_time_freelist_ready; aligned_u64 reserved[2]; } __aligned(8); /* Number of first HWR logs recorded (never overwritten by newer logs) */ #define ROGUE_FWIF_HWINFO_MAX_FIRST 8U /* Number of latest HWR logs (older logs are overwritten by newer logs) */ #define ROGUE_FWIF_HWINFO_MAX_LAST 8U /* Total number of HWR logs stored in a buffer */ #define ROGUE_FWIF_HWINFO_MAX \ (ROGUE_FWIF_HWINFO_MAX_FIRST + ROGUE_FWIF_HWINFO_MAX_LAST) /* Index of the last log in the HWR log buffer */ #define ROGUE_FWIF_HWINFO_LAST_INDEX (ROGUE_FWIF_HWINFO_MAX - 1U) struct rogue_fwif_hwrinfobuf { struct rogue_hwrinfo hwr_info[ROGUE_FWIF_HWINFO_MAX]; u32 hwr_counter; u32 write_index; u32 dd_req_count; u32 hwr_info_buf_flags; /* Compatibility and other flags */ u32 hwr_dm_locked_up_count[PVR_FWIF_DM_MAX]; u32 hwr_dm_overran_count[PVR_FWIF_DM_MAX]; u32 hwr_dm_recovered_count[PVR_FWIF_DM_MAX]; u32 hwr_dm_false_detect_count[PVR_FWIF_DM_MAX]; } __aligned(8); #define ROGUE_FWIF_CTXSWITCH_PROFILE_FAST_EN (1) #define ROGUE_FWIF_CTXSWITCH_PROFILE_MEDIUM_EN (2) #define ROGUE_FWIF_CTXSWITCH_PROFILE_SLOW_EN (3) #define ROGUE_FWIF_CTXSWITCH_PROFILE_NODELAY_EN (4) #define ROGUE_FWIF_CDM_ARBITRATION_TASK_DEMAND_EN (1) #define ROGUE_FWIF_CDM_ARBITRATION_ROUND_ROBIN_EN (2) #define ROGUE_FWIF_ISP_SCHEDMODE_VER1_IPP (1) #define ROGUE_FWIF_ISP_SCHEDMODE_VER2_ISP (2) /* ****************************************************************************** * ROGUE firmware Init Config Data ****************************************************************************** */ /* Flag definitions affecting the firmware globally */ #define ROGUE_FWIF_INICFG_CTXSWITCH_MODE_RAND BIT(0) #define ROGUE_FWIF_INICFG_CTXSWITCH_SRESET_EN BIT(1) #define ROGUE_FWIF_INICFG_HWPERF_EN BIT(2) #define ROGUE_FWIF_INICFG_DM_KILL_MODE_RAND_EN BIT(3) #define ROGUE_FWIF_INICFG_POW_RASCALDUST BIT(4) /* Bit 5 is reserved. */ #define ROGUE_FWIF_INICFG_FBCDC_V3_1_EN BIT(6) #define ROGUE_FWIF_INICFG_CHECK_MLIST_EN BIT(7) #define ROGUE_FWIF_INICFG_DISABLE_CLKGATING_EN BIT(8) /* Bit 9 is reserved. */ /* Bit 10 is reserved. */ /* Bit 11 is reserved. */ #define ROGUE_FWIF_INICFG_REGCONFIG_EN BIT(12) #define ROGUE_FWIF_INICFG_ASSERT_ON_OUTOFMEMORY BIT(13) #define ROGUE_FWIF_INICFG_HWP_DISABLE_FILTER BIT(14) /* Bit 15 is reserved. */ #define ROGUE_FWIF_INICFG_CTXSWITCH_PROFILE_SHIFT (16) #define ROGUE_FWIF_INICFG_CTXSWITCH_PROFILE_FAST \ (ROGUE_FWIF_CTXSWITCH_PROFILE_FAST_EN \ << ROGUE_FWIF_INICFG_CTXSWITCH_PROFILE_SHIFT) #define ROGUE_FWIF_INICFG_CTXSWITCH_PROFILE_MEDIUM \ (ROGUE_FWIF_CTXSWITCH_PROFILE_MEDIUM_EN \ << ROGUE_FWIF_INICFG_CTXSWITCH_PROFILE_SHIFT) #define ROGUE_FWIF_INICFG_CTXSWITCH_PROFILE_SLOW \ (ROGUE_FWIF_CTXSWITCH_PROFILE_SLOW_EN \ << ROGUE_FWIF_INICFG_CTXSWITCH_PROFILE_SHIFT) #define ROGUE_FWIF_INICFG_CTXSWITCH_PROFILE_NODELAY \ (ROGUE_FWIF_CTXSWITCH_PROFILE_NODELAY_EN \ << ROGUE_FWIF_INICFG_CTXSWITCH_PROFILE_SHIFT) #define ROGUE_FWIF_INICFG_CTXSWITCH_PROFILE_MASK \ (7 << ROGUE_FWIF_INICFG_CTXSWITCH_PROFILE_SHIFT) #define ROGUE_FWIF_INICFG_DISABLE_DM_OVERLAP BIT(19) #define ROGUE_FWIF_INICFG_ASSERT_ON_HWR_TRIGGER BIT(20) #define ROGUE_FWIF_INICFG_FABRIC_COHERENCY_ENABLED BIT(21) #define ROGUE_FWIF_INICFG_VALIDATE_IRQ BIT(22) #define ROGUE_FWIF_INICFG_DISABLE_PDP_EN BIT(23) #define ROGUE_FWIF_INICFG_SPU_POWER_STATE_MASK_CHANGE_EN BIT(24) #define ROGUE_FWIF_INICFG_WORKEST BIT(25) #define ROGUE_FWIF_INICFG_PDVFS BIT(26) #define ROGUE_FWIF_INICFG_CDM_ARBITRATION_SHIFT (27) #define ROGUE_FWIF_INICFG_CDM_ARBITRATION_TASK_DEMAND \ (ROGUE_FWIF_CDM_ARBITRATION_TASK_DEMAND_EN \ << ROGUE_FWIF_INICFG_CDM_ARBITRATION_SHIFT) #define ROGUE_FWIF_INICFG_CDM_ARBITRATION_ROUND_ROBIN \ (ROGUE_FWIF_CDM_ARBITRATION_ROUND_ROBIN_EN \ << ROGUE_FWIF_INICFG_CDM_ARBITRATION_SHIFT) #define ROGUE_FWIF_INICFG_CDM_ARBITRATION_MASK \ (3 << ROGUE_FWIF_INICFG_CDM_ARBITRATION_SHIFT) #define ROGUE_FWIF_INICFG_ISPSCHEDMODE_SHIFT (29) #define ROGUE_FWIF_INICFG_ISPSCHEDMODE_NONE (0) #define ROGUE_FWIF_INICFG_ISPSCHEDMODE_VER1_IPP \ (ROGUE_FWIF_ISP_SCHEDMODE_VER1_IPP \ << ROGUE_FWIF_INICFG_ISPSCHEDMODE_SHIFT) #define ROGUE_FWIF_INICFG_ISPSCHEDMODE_VER2_ISP \ (ROGUE_FWIF_ISP_SCHEDMODE_VER2_ISP \ << ROGUE_FWIF_INICFG_ISPSCHEDMODE_SHIFT) #define ROGUE_FWIF_INICFG_ISPSCHEDMODE_MASK \ (ROGUE_FWIF_INICFG_ISPSCHEDMODE_VER1_IPP | \ ROGUE_FWIF_INICFG_ISPSCHEDMODE_VER2_ISP) #define ROGUE_FWIF_INICFG_VALIDATE_SOCUSC_TIMER BIT(31) #define ROGUE_FWIF_INICFG_ALL (0xFFFFFFFFU) /* Extended Flag definitions affecting the firmware globally */ #define ROGUE_FWIF_INICFG_EXT_TFBC_CONTROL_SHIFT (0) /* [7] YUV10 override * [6:4] Quality * [3] Quality enable * [2:1] Compression scheme * [0] Lossy group */ #define ROGUE_FWIF_INICFG_EXT_TFBC_CONTROL_MASK (0xFF) #define ROGUE_FWIF_INICFG_EXT_ALL (ROGUE_FWIF_INICFG_EXT_TFBC_CONTROL_MASK) /* Flag definitions affecting only workloads submitted by a particular OS */ #define ROGUE_FWIF_INICFG_OS_CTXSWITCH_TDM_EN BIT(0) #define ROGUE_FWIF_INICFG_OS_CTXSWITCH_GEOM_EN BIT(1) #define ROGUE_FWIF_INICFG_OS_CTXSWITCH_FRAG_EN BIT(2) #define ROGUE_FWIF_INICFG_OS_CTXSWITCH_CDM_EN BIT(3) #define ROGUE_FWIF_INICFG_OS_LOW_PRIO_CS_TDM BIT(4) #define ROGUE_FWIF_INICFG_OS_LOW_PRIO_CS_GEOM BIT(5) #define ROGUE_FWIF_INICFG_OS_LOW_PRIO_CS_FRAG BIT(6) #define ROGUE_FWIF_INICFG_OS_LOW_PRIO_CS_CDM BIT(7) #define ROGUE_FWIF_INICFG_OS_ALL (0xFF) #define ROGUE_FWIF_INICFG_OS_CTXSWITCH_DM_ALL \ (ROGUE_FWIF_INICFG_OS_CTXSWITCH_TDM_EN | \ ROGUE_FWIF_INICFG_OS_CTXSWITCH_GEOM_EN | \ ROGUE_FWIF_INICFG_OS_CTXSWITCH_FRAG_EN | \ ROGUE_FWIF_INICFG_OS_CTXSWITCH_CDM_EN) #define ROGUE_FWIF_INICFG_OS_CTXSWITCH_CLRMSK \ ~(ROGUE_FWIF_INICFG_OS_CTXSWITCH_DM_ALL) #define ROGUE_FWIF_FILTCFG_TRUNCATE_HALF BIT(3) #define ROGUE_FWIF_FILTCFG_TRUNCATE_INT BIT(2) #define ROGUE_FWIF_FILTCFG_NEW_FILTER_MODE BIT(1) enum rogue_activepm_conf { ROGUE_ACTIVEPM_FORCE_OFF = 0, ROGUE_ACTIVEPM_FORCE_ON = 1, ROGUE_ACTIVEPM_DEFAULT = 2 }; enum rogue_rd_power_island_conf { ROGUE_RD_POWER_ISLAND_FORCE_OFF = 0, ROGUE_RD_POWER_ISLAND_FORCE_ON = 1, ROGUE_RD_POWER_ISLAND_DEFAULT = 2 }; struct rogue_fw_register_list { /* Register number */ u16 reg_num; /* Indirect register number (or 0 if not used) */ u16 indirect_reg_num; /* Start value for indirect register */ u16 indirect_start_val; /* End value for indirect register */ u16 indirect_end_val; }; struct rogue_fwif_dllist_node { u32 p; u32 n; }; /* * This number is used to represent an invalid page catalogue physical address */ #define ROGUE_FWIF_INVALID_PC_PHYADDR 0xFFFFFFFFFFFFFFFFLLU /* This number is used to represent unallocated page catalog base register */ #define ROGUE_FW_BIF_INVALID_PCSET 0xFFFFFFFFU /* Firmware memory context. */ struct rogue_fwif_fwmemcontext { /* device physical address of context's page catalogue */ aligned_u64 pc_dev_paddr; /* * associated page catalog base register (ROGUE_FW_BIF_INVALID_PCSET == * unallocated) */ u32 page_cat_base_reg_set; /* breakpoint address */ u32 breakpoint_addr; /* breakpoint handler address */ u32 bp_handler_addr; /* DM and enable control for BP */ u32 breakpoint_ctl; /* Compatibility and other flags */ u32 fw_mem_ctx_flags; u32 padding; } __aligned(8); /* * FW context state flags */ #define ROGUE_FWIF_CONTEXT_FLAGS_NEED_RESUME (0x00000001U) #define ROGUE_FWIF_CONTEXT_FLAGS_MC_NEED_RESUME_MASKFULL (0x000000FFU) #define ROGUE_FWIF_CONTEXT_FLAGS_TDM_HEADER_STALE (0x00000100U) #define ROGUE_FWIF_CONTEXT_FLAGS_LAST_KICK_SECURE (0x00000200U) #define ROGUE_NUM_GEOM_CORES_MAX 4 /* * FW-accessible TA state which must be written out to memory on context store */ struct rogue_fwif_geom_ctx_state_per_geom { /* To store in mid-TA */ aligned_u64 geom_reg_vdm_call_stack_pointer; /* Initial value (in case is 'lost' due to a lock-up */ aligned_u64 geom_reg_vdm_call_stack_pointer_init; u32 geom_reg_vbs_so_prim[4]; u16 geom_current_idx; u16 padding[3]; } __aligned(8); struct rogue_fwif_geom_ctx_state { /* FW-accessible TA state which must be written out to memory on context store */ struct rogue_fwif_geom_ctx_state_per_geom geom_core[ROGUE_NUM_GEOM_CORES_MAX]; } __aligned(8); /* * FW-accessible ISP state which must be written out to memory on context store */ struct rogue_fwif_frag_ctx_state { u32 frag_reg_pm_deallocated_mask_status; u32 frag_reg_dm_pds_mtilefree_status; /* Compatibility and other flags */ u32 ctx_state_flags; /* * frag_reg_isp_store should be the last element of the structure as this * is an array whose size is determined at runtime after detecting the * ROGUE core */ u32 frag_reg_isp_store[]; } __aligned(8); #define ROGUE_FWIF_CTX_USING_BUFFER_A (0) #define ROGUE_FWIF_CTX_USING_BUFFER_B (1U) struct rogue_fwif_compute_ctx_state { u32 ctx_state_flags; /* Target buffer and other flags */ }; struct rogue_fwif_fwcommoncontext { /* CCB details for this firmware context */ u32 ccbctl_fw_addr; /* CCB control */ u32 ccb_fw_addr; /* CCB base */ struct rogue_fwif_dma_addr ccb_meta_dma_addr; /* Context suspend state */ /* geom/frag context suspend state, read/written by FW */ u32 context_state_addr __aligned(8); /* Flags e.g. for context switching */ u32 fw_com_ctx_flags; u32 priority; u32 priority_seq_num; /* Framework state */ /* Register updates for Framework */ u32 rf_cmd_addr __aligned(8); /* Statistic updates waiting to be passed back to the host... */ /* True when some stats are pending */ bool stats_pending __aligned(4); /* Number of stores on this context since last update */ s32 stats_num_stores; /* Number of OOMs on this context since last update */ s32 stats_num_out_of_memory; /* Number of PRs on this context since last update */ s32 stats_num_partial_renders; /* Data Master type */ u32 dm; /* Device Virtual Address of the signal the context is waiting on */ aligned_u64 wait_signal_address; /* List entry for the wait-signal list */ struct rogue_fwif_dllist_node wait_signal_node __aligned(8); /* List entry for the buffer stalled list */ struct rogue_fwif_dllist_node buf_stalled_node __aligned(8); /* Address of the circular buffer queue pointers */ aligned_u64 cbuf_queue_ctrl_addr; aligned_u64 robustness_address; /* Max HWR deadline limit in ms */ u32 max_deadline_ms; /* Following HWR circular buffer read-offset needs resetting */ bool read_offset_needs_reset; /* List entry for the waiting list */ struct rogue_fwif_dllist_node waiting_node __aligned(8); /* List entry for the run list */ struct rogue_fwif_dllist_node run_node __aligned(8); /* UFO that last failed (or NULL) */ struct rogue_fwif_ufo last_failed_ufo; /* Memory context */ u32 fw_mem_context_fw_addr; /* References to the host side originators */ /* the Server Common Context */ u32 server_common_context_id; /* associated process ID */ u32 pid; /* True when Geom DM OOM is not allowed */ bool geom_oom_disabled __aligned(4); } __aligned(8); /* Firmware render context. */ struct rogue_fwif_fwrendercontext { /* Geometry firmware context. */ struct rogue_fwif_fwcommoncontext geom_context; /* Fragment firmware context. */ struct rogue_fwif_fwcommoncontext frag_context; struct rogue_fwif_static_rendercontext_state static_render_context_state; /* Number of commands submitted to the WorkEst FW CCB */ u32 work_est_ccb_submitted; /* Compatibility and other flags */ u32 fw_render_ctx_flags; } __aligned(8); /* Firmware compute context. */ struct rogue_fwif_fwcomputecontext { /* Firmware context for the CDM */ struct rogue_fwif_fwcommoncontext cdm_context; struct rogue_fwif_static_computecontext_state static_compute_context_state; /* Number of commands submitted to the WorkEst FW CCB */ u32 work_est_ccb_submitted; /* Compatibility and other flags */ u32 compute_ctx_flags; u32 wgp_state; u32 wgp_checksum; u32 core_mask_a; u32 core_mask_b; } __aligned(8); /* Firmware TDM context. */ struct rogue_fwif_fwtdmcontext { /* Firmware context for the TDM */ struct rogue_fwif_fwcommoncontext tdm_context; /* Number of commands submitted to the WorkEst FW CCB */ u32 work_est_ccb_submitted; } __aligned(8); /* Firmware TQ3D context. */ struct rogue_fwif_fwtransfercontext { /* Firmware context for TQ3D. */ struct rogue_fwif_fwcommoncontext tq_context; } __aligned(8); /* ****************************************************************************** * Defines for CMD_TYPE corruption detection and forward compatibility check ****************************************************************************** */ /* * CMD_TYPE 32bit contains: * 31:16 Reserved for magic value to detect corruption (16 bits) * 15 Reserved for ROGUE_CCB_TYPE_TASK (1 bit) * 14:0 Bits available for CMD_TYPEs (15 bits) */ /* Magic value to detect corruption */ #define ROGUE_CMD_MAGIC_DWORD (0x2ABC) #define ROGUE_CMD_MAGIC_DWORD_MASK (0xFFFF0000U) #define ROGUE_CMD_MAGIC_DWORD_SHIFT (16U) #define ROGUE_CMD_MAGIC_DWORD_SHIFTED \ (ROGUE_CMD_MAGIC_DWORD << ROGUE_CMD_MAGIC_DWORD_SHIFT) /* Kernel CCB control for ROGUE */ struct rogue_fwif_ccb_ctl { /* write offset into array of commands (MUST be aligned to 16 bytes!) */ u32 write_offset; /* Padding to ensure read and write offsets are in separate cache lines. */ u8 padding[128 - sizeof(u32)]; /* read offset into array of commands */ u32 read_offset; /* Offset wrapping mask (Total capacity of the CCB - 1) */ u32 wrap_mask; /* size of each command in bytes */ u32 cmd_size; u32 padding2; } __aligned(8); /* Kernel CCB command structure for ROGUE */ #define ROGUE_FWIF_MMUCACHEDATA_FLAGS_PT (0x1U) /* MMU_CTRL_INVAL_PT_EN */ #define ROGUE_FWIF_MMUCACHEDATA_FLAGS_PD (0x2U) /* MMU_CTRL_INVAL_PD_EN */ #define ROGUE_FWIF_MMUCACHEDATA_FLAGS_PC (0x4U) /* MMU_CTRL_INVAL_PC_EN */ /* * can't use PM_TLB0 bit from BIFPM_CTRL reg because it collides with PT * bit from BIF_CTRL reg */ #define ROGUE_FWIF_MMUCACHEDATA_FLAGS_PMTLB (0x10) /* BIF_CTRL_INVAL_TLB1_EN */ #define ROGUE_FWIF_MMUCACHEDATA_FLAGS_TLB \ (ROGUE_FWIF_MMUCACHEDATA_FLAGS_PMTLB | 0x8) /* MMU_CTRL_INVAL_ALL_CONTEXTS_EN */ #define ROGUE_FWIF_MMUCACHEDATA_FLAGS_CTX_ALL (0x800) /* indicates FW should interrupt the host */ #define ROGUE_FWIF_MMUCACHEDATA_FLAGS_INTERRUPT (0x4000000U) struct rogue_fwif_mmucachedata { u32 cache_flags; u32 mmu_cache_sync_fw_addr; u32 mmu_cache_sync_update_value; }; #define ROGUE_FWIF_BPDATA_FLAGS_ENABLE BIT(0) #define ROGUE_FWIF_BPDATA_FLAGS_WRITE BIT(1) #define ROGUE_FWIF_BPDATA_FLAGS_CTL BIT(2) #define ROGUE_FWIF_BPDATA_FLAGS_REGS BIT(3) struct rogue_fwif_bpdata { /* Memory context */ u32 fw_mem_context_fw_addr; /* Breakpoint address */ u32 bp_addr; /* Breakpoint handler */ u32 bp_handler_addr; /* Breakpoint control */ u32 bp_dm; u32 bp_data_flags; /* Number of temporary registers to overallocate */ u32 temp_regs; /* Number of shared registers to overallocate */ u32 shared_regs; /* DM associated with the breakpoint */ u32 dm; }; #define ROGUE_FWIF_KCCB_CMD_KICK_DATA_MAX_NUM_CLEANUP_CTLS \ (ROGUE_FWIF_PRBUFFER_MAXSUPPORTED + 1U) /* +1 is RTDATASET cleanup */ struct rogue_fwif_kccb_cmd_kick_data { /* address of the firmware context */ u32 context_fw_addr; /* Client CCB woff update */ u32 client_woff_update; /* Client CCB wrap mask update after CCCB growth */ u32 client_wrap_mask_update; /* number of CleanupCtl pointers attached */ u32 num_cleanup_ctl; /* CleanupCtl structures associated with command */ u32 cleanup_ctl_fw_addr [ROGUE_FWIF_KCCB_CMD_KICK_DATA_MAX_NUM_CLEANUP_CTLS]; /* * offset to the CmdHeader which houses the workload estimation kick * data. */ u32 work_est_cmd_header_offset; }; struct rogue_fwif_kccb_cmd_combined_geom_frag_kick_data { struct rogue_fwif_kccb_cmd_kick_data geom_cmd_kick_data; struct rogue_fwif_kccb_cmd_kick_data frag_cmd_kick_data; }; struct rogue_fwif_kccb_cmd_force_update_data { /* address of the firmware context */ u32 context_fw_addr; /* Client CCB fence offset */ u32 ccb_fence_offset; }; enum rogue_fwif_cleanup_type { /* FW common context cleanup */ ROGUE_FWIF_CLEANUP_FWCOMMONCONTEXT, /* FW HW RT data cleanup */ ROGUE_FWIF_CLEANUP_HWRTDATA, /* FW freelist cleanup */ ROGUE_FWIF_CLEANUP_FREELIST, /* FW ZS Buffer cleanup */ ROGUE_FWIF_CLEANUP_ZSBUFFER, }; struct rogue_fwif_cleanup_request { /* Cleanup type */ enum rogue_fwif_cleanup_type cleanup_type; union { /* FW common context to cleanup */ u32 context_fw_addr; /* HW RT to cleanup */ u32 hwrt_data_fw_addr; /* Freelist to cleanup */ u32 freelist_fw_addr; /* ZS Buffer to cleanup */ u32 zs_buffer_fw_addr; } cleanup_data; }; enum rogue_fwif_power_type { ROGUE_FWIF_POW_OFF_REQ = 1, ROGUE_FWIF_POW_FORCED_IDLE_REQ, ROGUE_FWIF_POW_NUM_UNITS_CHANGE, ROGUE_FWIF_POW_APM_LATENCY_CHANGE }; enum rogue_fwif_power_force_idle_type { ROGUE_FWIF_POWER_FORCE_IDLE = 1, ROGUE_FWIF_POWER_CANCEL_FORCED_IDLE, ROGUE_FWIF_POWER_HOST_TIMEOUT, }; struct rogue_fwif_power_request { /* Type of power request */ enum rogue_fwif_power_type pow_type; union { /* Number of active Dusts */ u32 num_of_dusts; /* If the operation is mandatory */ bool forced __aligned(4); /* * Type of Request. Consolidating Force Idle, Cancel Forced * Idle, Host Timeout */ enum rogue_fwif_power_force_idle_type pow_request_type; } power_req_data; }; struct rogue_fwif_slcflushinvaldata { /* Context to fence on (only useful when bDMContext == TRUE) */ u32 context_fw_addr; /* Invalidate the cache as well as flushing */ bool inval __aligned(4); /* The data to flush/invalidate belongs to a specific DM context */ bool dm_context __aligned(4); /* Optional address of range (only useful when bDMContext == FALSE) */ aligned_u64 address; /* Optional size of range (only useful when bDMContext == FALSE) */ aligned_u64 size; }; enum rogue_fwif_hwperf_update_config { ROGUE_FWIF_HWPERF_CTRL_TOGGLE = 0, ROGUE_FWIF_HWPERF_CTRL_SET = 1, ROGUE_FWIF_HWPERF_CTRL_EMIT_FEATURES_EV = 2 }; struct rogue_fwif_hwperf_ctrl { enum rogue_fwif_hwperf_update_config opcode; /* Control operation code */ aligned_u64 mask; /* Mask of events to toggle */ }; struct rogue_fwif_hwperf_config_enable_blks { /* Number of ROGUE_HWPERF_CONFIG_MUX_CNTBLK in the array */ u32 num_blocks; /* Address of the ROGUE_HWPERF_CONFIG_MUX_CNTBLK array */ u32 block_configs_fw_addr; }; struct rogue_fwif_hwperf_config_da_blks { /* Number of ROGUE_HWPERF_CONFIG_CNTBLK in the array */ u32 num_blocks; /* Address of the ROGUE_HWPERF_CONFIG_CNTBLK array */ u32 block_configs_fw_addr; }; struct rogue_fwif_coreclkspeedchange_data { u32 new_clock_speed; /* New clock speed */ }; #define ROGUE_FWIF_HWPERF_CTRL_BLKS_MAX 16 struct rogue_fwif_hwperf_ctrl_blks { bool enable; /* Number of block IDs in the array */ u32 num_blocks; /* Array of ROGUE_HWPERF_CNTBLK_ID values */ u16 block_ids[ROGUE_FWIF_HWPERF_CTRL_BLKS_MAX]; }; struct rogue_fwif_hwperf_select_custom_cntrs { u16 custom_block; u16 num_counters; u32 custom_counter_ids_fw_addr; }; struct rogue_fwif_zsbuffer_backing_data { u32 zs_buffer_fw_addr; /* ZS-Buffer FW address */ bool done __aligned(4); /* action backing/unbacking succeeded */ }; struct rogue_fwif_freelist_gs_data { /* Freelist FW address */ u32 freelist_fw_addr; /* Amount of the Freelist change */ u32 delta_pages; /* New amount of pages on the freelist (including ready pages) */ u32 new_pages; /* Number of ready pages to be held in reserve until OOM */ u32 ready_pages; }; #define MAX_FREELISTS_SIZE 3 #define MAX_HW_GEOM_FRAG_CONTEXTS_SIZE 3 #define ROGUE_FWIF_MAX_FREELISTS_TO_RECONSTRUCT \ (MAX_HW_GEOM_FRAG_CONTEXTS_SIZE * MAX_FREELISTS_SIZE * 2U) #define ROGUE_FWIF_FREELISTS_RECONSTRUCTION_FAILED_FLAG 0x80000000U struct rogue_fwif_freelists_reconstruction_data { u32 freelist_count; u32 freelist_ids[ROGUE_FWIF_MAX_FREELISTS_TO_RECONSTRUCT]; }; struct rogue_fwif_write_offset_update_data { /* * Context to that may need to be resumed following write offset update */ u32 context_fw_addr; } __aligned(8); /* ****************************************************************************** * Proactive DVFS Structures ****************************************************************************** */ #define NUM_OPP_VALUES 16 struct pdvfs_opp { u32 volt; /* V */ u32 freq; /* Hz */ } __aligned(8); struct rogue_fwif_pdvfs_opp { struct pdvfs_opp opp_values[NUM_OPP_VALUES]; u32 min_opp_point; u32 max_opp_point; } __aligned(8); struct rogue_fwif_pdvfs_max_freq_data { u32 max_opp_point; } __aligned(8); struct rogue_fwif_pdvfs_min_freq_data { u32 min_opp_point; } __aligned(8); /* ****************************************************************************** * Register configuration structures ****************************************************************************** */ #define ROGUE_FWIF_REG_CFG_MAX_SIZE 512 enum rogue_fwif_regdata_cmd_type { ROGUE_FWIF_REGCFG_CMD_ADD = 101, ROGUE_FWIF_REGCFG_CMD_CLEAR = 102, ROGUE_FWIF_REGCFG_CMD_ENABLE = 103, ROGUE_FWIF_REGCFG_CMD_DISABLE = 104 }; enum rogue_fwif_reg_cfg_type { /* Sidekick power event */ ROGUE_FWIF_REG_CFG_TYPE_PWR_ON = 0, /* Rascal / dust power event */ ROGUE_FWIF_REG_CFG_TYPE_DUST_CHANGE, /* Geometry kick */ ROGUE_FWIF_REG_CFG_TYPE_GEOM, /* Fragment kick */ ROGUE_FWIF_REG_CFG_TYPE_FRAG, /* Compute kick */ ROGUE_FWIF_REG_CFG_TYPE_CDM, /* TLA kick */ ROGUE_FWIF_REG_CFG_TYPE_TLA, /* TDM kick */ ROGUE_FWIF_REG_CFG_TYPE_TDM, /* Applies to all types. Keep as last element */ ROGUE_FWIF_REG_CFG_TYPE_ALL }; struct rogue_fwif_reg_cfg_rec { u64 sddr; u64 mask; u64 value; }; struct rogue_fwif_regconfig_data { enum rogue_fwif_regdata_cmd_type cmd_type; enum rogue_fwif_reg_cfg_type reg_config_type; struct rogue_fwif_reg_cfg_rec reg_config __aligned(8); }; struct rogue_fwif_reg_cfg { /* * PDump WRW command write granularity is 32 bits. * Add padding to ensure array size is 32 bit granular. */ u8 num_regs_type[ALIGN((u32)ROGUE_FWIF_REG_CFG_TYPE_ALL, sizeof(u32))] __aligned(8); struct rogue_fwif_reg_cfg_rec reg_configs[ROGUE_FWIF_REG_CFG_MAX_SIZE] __aligned(8); } __aligned(8); enum rogue_fwif_os_state_change { ROGUE_FWIF_OS_ONLINE = 1, ROGUE_FWIF_OS_OFFLINE }; struct rogue_fwif_os_state_change_data { u32 osid; enum rogue_fwif_os_state_change new_os_state; } __aligned(8); enum rogue_fwif_counter_dump_request { ROGUE_FWIF_PWR_COUNTER_DUMP_START = 1, ROGUE_FWIF_PWR_COUNTER_DUMP_STOP, ROGUE_FWIF_PWR_COUNTER_DUMP_SAMPLE, }; struct rogue_fwif_counter_dump_data { enum rogue_fwif_counter_dump_request counter_dump_request; } __aligned(8); enum rogue_fwif_kccb_cmd_type { /* Common commands */ ROGUE_FWIF_KCCB_CMD_KICK = 101U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, ROGUE_FWIF_KCCB_CMD_MMUCACHE = 102U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, ROGUE_FWIF_KCCB_CMD_BP = 103U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* SLC flush and invalidation request */ ROGUE_FWIF_KCCB_CMD_SLCFLUSHINVAL = 105U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* * Requests cleanup of a FW resource (type specified in the command * data) */ ROGUE_FWIF_KCCB_CMD_CLEANUP = 106U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* Power request */ ROGUE_FWIF_KCCB_CMD_POW = 107U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* Backing for on-demand ZS-Buffer done */ ROGUE_FWIF_KCCB_CMD_ZSBUFFER_BACKING_UPDATE = 108U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* Unbacking for on-demand ZS-Buffer done */ ROGUE_FWIF_KCCB_CMD_ZSBUFFER_UNBACKING_UPDATE = 109U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* Freelist Grow done */ ROGUE_FWIF_KCCB_CMD_FREELIST_GROW_UPDATE = 110U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* Freelists Reconstruction done */ ROGUE_FWIF_KCCB_CMD_FREELISTS_RECONSTRUCTION_UPDATE = 112U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* * Informs the firmware that the host has added more data to a CDM2 * Circular Buffer */ ROGUE_FWIF_KCCB_CMD_NOTIFY_WRITE_OFFSET_UPDATE = 114U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* Health check request */ ROGUE_FWIF_KCCB_CMD_HEALTH_CHECK = 115U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* Forcing signalling of all unmet UFOs for a given CCB offset */ ROGUE_FWIF_KCCB_CMD_FORCE_UPDATE = 116U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* There is a geometry and a fragment command in this single kick */ ROGUE_FWIF_KCCB_CMD_COMBINED_GEOM_FRAG_KICK = 117U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* Informs the FW that a Guest OS has come online / offline. */ ROGUE_FWIF_KCCB_CMD_OS_ONLINE_STATE_CONFIGURE = 118U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* Commands only permitted to the native or host OS */ ROGUE_FWIF_KCCB_CMD_REGCONFIG = 200U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* Configure HWPerf events (to be generated) and HWPerf buffer address (if required) */ ROGUE_FWIF_KCCB_CMD_HWPERF_UPDATE_CONFIG = 201U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* Enable or disable multiple HWPerf blocks (reusing existing configuration) */ ROGUE_FWIF_KCCB_CMD_HWPERF_CTRL_BLKS = 203U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* Core clock speed change event */ ROGUE_FWIF_KCCB_CMD_CORECLKSPEEDCHANGE = 204U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* * Ask the firmware to update its cached ui32LogType value from the (shared) * tracebuf control structure */ ROGUE_FWIF_KCCB_CMD_LOGTYPE_UPDATE = 206U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* Set a maximum frequency/OPP point */ ROGUE_FWIF_KCCB_CMD_PDVFS_LIMIT_MAX_FREQ = 207U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* * Changes the relative scheduling priority for a particular OSid. It can * only be serviced for the Host DDK */ ROGUE_FWIF_KCCB_CMD_OSID_PRIORITY_CHANGE = 208U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* Set or clear firmware state flags */ ROGUE_FWIF_KCCB_CMD_STATEFLAGS_CTRL = 209U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* Set a minimum frequency/OPP point */ ROGUE_FWIF_KCCB_CMD_PDVFS_LIMIT_MIN_FREQ = 212U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* Configure Periodic Hardware Reset behaviour */ ROGUE_FWIF_KCCB_CMD_PHR_CFG = 213U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* Configure Safety Firmware Watchdog */ ROGUE_FWIF_KCCB_CMD_WDG_CFG = 215U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* Controls counter dumping in the FW */ ROGUE_FWIF_KCCB_CMD_COUNTER_DUMP = 216U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* Configure, clear and enable multiple HWPerf blocks */ ROGUE_FWIF_KCCB_CMD_HWPERF_CONFIG_ENABLE_BLKS = 217U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* Configure the custom counters for HWPerf */ ROGUE_FWIF_KCCB_CMD_HWPERF_SELECT_CUSTOM_CNTRS = 218U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* Configure directly addressable counters for HWPerf */ ROGUE_FWIF_KCCB_CMD_HWPERF_CONFIG_BLKS = 220U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, }; #define ROGUE_FWIF_LAST_ALLOWED_GUEST_KCCB_CMD \ (ROGUE_FWIF_KCCB_CMD_REGCONFIG - 1) /* Kernel CCB command packet */ struct rogue_fwif_kccb_cmd { /* Command type */ enum rogue_fwif_kccb_cmd_type cmd_type; /* Compatibility and other flags */ u32 kccb_flags; /* * NOTE: Make sure that uCmdData is the last member of this struct * This is to calculate actual command size for device mem copy. * (Refer ROGUEGetCmdMemCopySize()) */ union { /* Data for Kick command */ struct rogue_fwif_kccb_cmd_kick_data cmd_kick_data; /* Data for combined geom/frag Kick command */ struct rogue_fwif_kccb_cmd_combined_geom_frag_kick_data combined_geom_frag_cmd_kick_data; /* Data for MMU cache command */ struct rogue_fwif_mmucachedata mmu_cache_data; /* Data for Breakpoint Commands */ struct rogue_fwif_bpdata bp_data; /* Data for SLC Flush/Inval commands */ struct rogue_fwif_slcflushinvaldata slc_flush_inval_data; /* Data for cleanup commands */ struct rogue_fwif_cleanup_request cleanup_data; /* Data for power request commands */ struct rogue_fwif_power_request pow_data; /* Data for HWPerf control command */ struct rogue_fwif_hwperf_ctrl hw_perf_ctrl; /* * Data for HWPerf configure, clear and enable performance * counter block command */ struct rogue_fwif_hwperf_config_enable_blks hw_perf_cfg_enable_blks; /* * Data for HWPerf enable or disable performance counter block * commands */ struct rogue_fwif_hwperf_ctrl_blks hw_perf_ctrl_blks; /* Data for HWPerf configure the custom counters to read */ struct rogue_fwif_hwperf_select_custom_cntrs hw_perf_select_cstm_cntrs; /* Data for HWPerf configure Directly Addressable blocks */ struct rogue_fwif_hwperf_config_da_blks hw_perf_cfg_da_blks; /* Data for core clock speed change */ struct rogue_fwif_coreclkspeedchange_data core_clk_speed_change_data; /* Feedback for Z/S Buffer backing/unbacking */ struct rogue_fwif_zsbuffer_backing_data zs_buffer_backing_data; /* Feedback for Freelist grow/shrink */ struct rogue_fwif_freelist_gs_data free_list_gs_data; /* Feedback for Freelists reconstruction*/ struct rogue_fwif_freelists_reconstruction_data free_lists_reconstruction_data; /* Data for custom register configuration */ struct rogue_fwif_regconfig_data reg_config_data; /* Data for informing the FW about the write offset update */ struct rogue_fwif_write_offset_update_data write_offset_update_data; /* Data for setting the max frequency/OPP */ struct rogue_fwif_pdvfs_max_freq_data pdvfs_max_freq_data; /* Data for setting the min frequency/OPP */ struct rogue_fwif_pdvfs_min_freq_data pdvfs_min_freq_data; /* Data for updating the Guest Online states */ struct rogue_fwif_os_state_change_data cmd_os_online_state_data; /* Dev address for TBI buffer allocated on demand */ u32 tbi_buffer_fw_addr; /* Data for dumping of register ranges */ struct rogue_fwif_counter_dump_data counter_dump_config_data; /* Data for signalling all unmet fences for a given CCB */ struct rogue_fwif_kccb_cmd_force_update_data force_update_data; } cmd_data __aligned(8); } __aligned(8); PVR_FW_STRUCT_SIZE_ASSERT(struct rogue_fwif_kccb_cmd); /* ****************************************************************************** * Firmware CCB command structure for ROGUE ****************************************************************************** */ struct rogue_fwif_fwccb_cmd_zsbuffer_backing_data { u32 zs_buffer_id; }; struct rogue_fwif_fwccb_cmd_freelist_gs_data { u32 freelist_id; }; struct rogue_fwif_fwccb_cmd_freelists_reconstruction_data { u32 freelist_count; u32 hwr_counter; u32 freelist_ids[ROGUE_FWIF_MAX_FREELISTS_TO_RECONSTRUCT]; }; /* 1 if a page fault happened */ #define ROGUE_FWIF_FWCCB_CMD_CONTEXT_RESET_FLAG_PF BIT(0) /* 1 if applicable to all contexts */ #define ROGUE_FWIF_FWCCB_CMD_CONTEXT_RESET_FLAG_ALL_CTXS BIT(1) struct rogue_fwif_fwccb_cmd_context_reset_data { /* Context affected by the reset */ u32 server_common_context_id; /* Reason for reset */ enum rogue_context_reset_reason reset_reason; /* Data Master affected by the reset */ u32 dm; /* Job ref running at the time of reset */ u32 reset_job_ref; /* ROGUE_FWIF_FWCCB_CMD_CONTEXT_RESET_FLAG bitfield */ u32 flags; /* At what page catalog address */ aligned_u64 pc_address; /* Page fault address (only when applicable) */ aligned_u64 fault_address; }; struct rogue_fwif_fwccb_cmd_fw_pagefault_data { /* Page fault address */ u64 fw_fault_addr; }; enum rogue_fwif_fwccb_cmd_type { /* Requests ZSBuffer to be backed with physical pages */ ROGUE_FWIF_FWCCB_CMD_ZSBUFFER_BACKING = 101U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* Requests ZSBuffer to be unbacked */ ROGUE_FWIF_FWCCB_CMD_ZSBUFFER_UNBACKING = 102U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* Requests an on-demand freelist grow/shrink */ ROGUE_FWIF_FWCCB_CMD_FREELIST_GROW = 103U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* Requests freelists reconstruction */ ROGUE_FWIF_FWCCB_CMD_FREELISTS_RECONSTRUCTION = 104U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* Notifies host of a HWR event on a context */ ROGUE_FWIF_FWCCB_CMD_CONTEXT_RESET_NOTIFICATION = 105U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* Requests an on-demand debug dump */ ROGUE_FWIF_FWCCB_CMD_DEBUG_DUMP = 106U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* Requests an on-demand update on process stats */ ROGUE_FWIF_FWCCB_CMD_UPDATE_STATS = 107U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, ROGUE_FWIF_FWCCB_CMD_CORE_CLK_RATE_CHANGE = 108U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, ROGUE_FWIF_FWCCB_CMD_REQUEST_GPU_RESTART = 109U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, /* Notifies host of a FW pagefault */ ROGUE_FWIF_FWCCB_CMD_CONTEXT_FW_PF_NOTIFICATION = 112U | ROGUE_CMD_MAGIC_DWORD_SHIFTED, }; enum rogue_fwif_fwccb_cmd_update_stats_type { /* * PVRSRVStatsUpdateRenderContextStats should increase the value of the * ui32TotalNumPartialRenders stat */ ROGUE_FWIF_FWCCB_CMD_UPDATE_NUM_PARTIAL_RENDERS = 1, /* * PVRSRVStatsUpdateRenderContextStats should increase the value of the * ui32TotalNumOutOfMemory stat */ ROGUE_FWIF_FWCCB_CMD_UPDATE_NUM_OUT_OF_MEMORY, /* * PVRSRVStatsUpdateRenderContextStats should increase the value of the * ui32NumGeomStores stat */ ROGUE_FWIF_FWCCB_CMD_UPDATE_NUM_GEOM_STORES, /* * PVRSRVStatsUpdateRenderContextStats should increase the value of the * ui32NumFragStores stat */ ROGUE_FWIF_FWCCB_CMD_UPDATE_NUM_FRAG_STORES, /* * PVRSRVStatsUpdateRenderContextStats should increase the value of the * ui32NumCDMStores stat */ ROGUE_FWIF_FWCCB_CMD_UPDATE_NUM_CDM_STORES, /* * PVRSRVStatsUpdateRenderContextStats should increase the value of the * ui32NumTDMStores stat */ ROGUE_FWIF_FWCCB_CMD_UPDATE_NUM_TDM_STORES }; struct rogue_fwif_fwccb_cmd_update_stats_data { /* Element to update */ enum rogue_fwif_fwccb_cmd_update_stats_type element_to_update; /* The pid of the process whose stats are being updated */ u32 pid_owner; /* Adjustment to be made to the statistic */ s32 adjustment_value; }; struct rogue_fwif_fwccb_cmd_core_clk_rate_change_data { u32 core_clk_rate; } __aligned(8); struct rogue_fwif_fwccb_cmd { /* Command type */ enum rogue_fwif_fwccb_cmd_type cmd_type; /* Compatibility and other flags */ u32 fwccb_flags; union { /* Data for Z/S-Buffer on-demand (un)backing*/ struct rogue_fwif_fwccb_cmd_zsbuffer_backing_data cmd_zs_buffer_backing; /* Data for on-demand freelist grow/shrink */ struct rogue_fwif_fwccb_cmd_freelist_gs_data cmd_free_list_gs; /* Data for freelists reconstruction */ struct rogue_fwif_fwccb_cmd_freelists_reconstruction_data cmd_freelists_reconstruction; /* Data for context reset notification */ struct rogue_fwif_fwccb_cmd_context_reset_data cmd_context_reset_notification; /* Data for updating process stats */ struct rogue_fwif_fwccb_cmd_update_stats_data cmd_update_stats_data; struct rogue_fwif_fwccb_cmd_core_clk_rate_change_data cmd_core_clk_rate_change; struct rogue_fwif_fwccb_cmd_fw_pagefault_data cmd_fw_pagefault; } cmd_data __aligned(8); } __aligned(8); PVR_FW_STRUCT_SIZE_ASSERT(struct rogue_fwif_fwccb_cmd); /* ****************************************************************************** * Workload estimation Firmware CCB command structure for ROGUE ****************************************************************************** */ struct rogue_fwif_workest_fwccb_cmd { /* Index for return data array */ u16 return_data_index; /* The cycles the workload took on the hardware */ u32 cycles_taken; }; /* ****************************************************************************** * Client CCB commands for ROGUE ****************************************************************************** */ /* * Required memory alignment for 64-bit variables accessible by Meta * (The gcc meta aligns 64-bit variables to 64-bit; therefore, memory shared * between the host and meta that contains 64-bit variables has to maintain * this alignment) */ #define ROGUE_FWIF_FWALLOC_ALIGN sizeof(u64) #define ROGUE_CCB_TYPE_TASK BIT(15) #define ROGUE_CCB_FWALLOC_ALIGN(size) \ (((size) + (ROGUE_FWIF_FWALLOC_ALIGN - 1)) & \ ~(ROGUE_FWIF_FWALLOC_ALIGN - 1)) #define ROGUE_FWIF_CCB_CMD_TYPE_GEOM \ (201U | ROGUE_CMD_MAGIC_DWORD_SHIFTED | ROGUE_CCB_TYPE_TASK) #define ROGUE_FWIF_CCB_CMD_TYPE_TQ_3D \ (202U | ROGUE_CMD_MAGIC_DWORD_SHIFTED | ROGUE_CCB_TYPE_TASK) #define ROGUE_FWIF_CCB_CMD_TYPE_FRAG \ (203U | ROGUE_CMD_MAGIC_DWORD_SHIFTED | ROGUE_CCB_TYPE_TASK) #define ROGUE_FWIF_CCB_CMD_TYPE_FRAG_PR \ (204U | ROGUE_CMD_MAGIC_DWORD_SHIFTED | ROGUE_CCB_TYPE_TASK) #define ROGUE_FWIF_CCB_CMD_TYPE_CDM \ (205U | ROGUE_CMD_MAGIC_DWORD_SHIFTED | ROGUE_CCB_TYPE_TASK) #define ROGUE_FWIF_CCB_CMD_TYPE_TQ_TDM \ (206U | ROGUE_CMD_MAGIC_DWORD_SHIFTED | ROGUE_CCB_TYPE_TASK) #define ROGUE_FWIF_CCB_CMD_TYPE_FBSC_INVALIDATE \ (207U | ROGUE_CMD_MAGIC_DWORD_SHIFTED | ROGUE_CCB_TYPE_TASK) #define ROGUE_FWIF_CCB_CMD_TYPE_TQ_2D \ (208U | ROGUE_CMD_MAGIC_DWORD_SHIFTED | ROGUE_CCB_TYPE_TASK) #define ROGUE_FWIF_CCB_CMD_TYPE_PRE_TIMESTAMP \ (209U | ROGUE_CMD_MAGIC_DWORD_SHIFTED | ROGUE_CCB_TYPE_TASK) #define ROGUE_FWIF_CCB_CMD_TYPE_NULL \ (210U | ROGUE_CMD_MAGIC_DWORD_SHIFTED | ROGUE_CCB_TYPE_TASK) #define ROGUE_FWIF_CCB_CMD_TYPE_ABORT \ (211U | ROGUE_CMD_MAGIC_DWORD_SHIFTED | ROGUE_CCB_TYPE_TASK) /* Leave a gap between CCB specific commands and generic commands */ #define ROGUE_FWIF_CCB_CMD_TYPE_FENCE (212U | ROGUE_CMD_MAGIC_DWORD_SHIFTED) #define ROGUE_FWIF_CCB_CMD_TYPE_UPDATE (213U | ROGUE_CMD_MAGIC_DWORD_SHIFTED) #define ROGUE_FWIF_CCB_CMD_TYPE_RMW_UPDATE \ (214U | ROGUE_CMD_MAGIC_DWORD_SHIFTED) #define ROGUE_FWIF_CCB_CMD_TYPE_FENCE_PR (215U | ROGUE_CMD_MAGIC_DWORD_SHIFTED) #define ROGUE_FWIF_CCB_CMD_TYPE_PRIORITY (216U | ROGUE_CMD_MAGIC_DWORD_SHIFTED) /* * Pre and Post timestamp commands are supposed to sandwich the DM cmd. The * padding code with the CCB wrap upsets the FW if we don't have the task type * bit cleared for POST_TIMESTAMPs. That's why we have 2 different cmd types. */ #define ROGUE_FWIF_CCB_CMD_TYPE_POST_TIMESTAMP \ (217U | ROGUE_CMD_MAGIC_DWORD_SHIFTED) #define ROGUE_FWIF_CCB_CMD_TYPE_UNFENCED_UPDATE \ (218U | ROGUE_CMD_MAGIC_DWORD_SHIFTED) #define ROGUE_FWIF_CCB_CMD_TYPE_UNFENCED_RMW_UPDATE \ (219U | ROGUE_CMD_MAGIC_DWORD_SHIFTED) #define ROGUE_FWIF_CCB_CMD_TYPE_PADDING (221U | ROGUE_CMD_MAGIC_DWORD_SHIFTED) struct rogue_fwif_workest_kick_data { /* Index for the KM Workload estimation return data array */ u16 return_data_index __aligned(8); /* Predicted time taken to do the work in cycles */ u32 cycles_prediction __aligned(8); /* Deadline for the workload */ aligned_u64 deadline; }; struct rogue_fwif_ccb_cmd_header { u32 cmd_type; u32 cmd_size; /* * external job reference - provided by client and used in debug for * tracking submitted work */ u32 ext_job_ref; /* * internal job reference - generated by services and used in debug for * tracking submitted work */ u32 int_job_ref; /* Workload Estimation - Workload Estimation Data */ struct rogue_fwif_workest_kick_data work_est_kick_data __aligned(8); }; /* ****************************************************************************** * Client CCB commands which are only required by the kernel ****************************************************************************** */ struct rogue_fwif_cmd_priority { s32 priority; }; /* ****************************************************************************** * Signature and Checksums Buffer ****************************************************************************** */ struct rogue_fwif_sigbuf_ctl { /* Ptr to Signature Buffer memory */ u32 buffer_fw_addr; /* Amount of space left for storing regs in the buffer */ u32 left_size_in_regs; } __aligned(8); struct rogue_fwif_counter_dump_ctl { /* Ptr to counter dump buffer */ u32 buffer_fw_addr; /* Amount of space for storing in the buffer */ u32 size_in_dwords; } __aligned(8); struct rogue_fwif_firmware_gcov_ctl { /* Ptr to firmware gcov buffer */ u32 buffer_fw_addr; /* Amount of space for storing in the buffer */ u32 size; } __aligned(8); /* ***************************************************************************** * ROGUE Compatibility checks ***************************************************************************** */ /* * WARNING: Whenever the layout of ROGUE_FWIF_COMPCHECKS_BVNC changes, the * following define should be increased by 1 to indicate to the compatibility * logic that layout has changed. */ #define ROGUE_FWIF_COMPCHECKS_LAYOUT_VERSION 3 struct rogue_fwif_compchecks_bvnc { /* WARNING: This field must be defined as first one in this structure */ u32 layout_version; aligned_u64 bvnc; } __aligned(8); struct rogue_fwif_init_options { u8 os_count_support; u8 padding[7]; } __aligned(8); #define ROGUE_FWIF_COMPCHECKS_BVNC_DECLARE_AND_INIT(name) \ struct rogue_fwif_compchecks_bvnc(name) = { \ ROGUE_FWIF_COMPCHECKS_LAYOUT_VERSION, \ 0, \ } static inline void rogue_fwif_compchecks_bvnc_init(struct rogue_fwif_compchecks_bvnc *compchecks) { compchecks->layout_version = ROGUE_FWIF_COMPCHECKS_LAYOUT_VERSION; compchecks->bvnc = 0; } struct rogue_fwif_compchecks { /* hardware BVNC (from the ROGUE registers) */ struct rogue_fwif_compchecks_bvnc hw_bvnc; /* firmware BVNC */ struct rogue_fwif_compchecks_bvnc fw_bvnc; /* identifier of the FW processor version */ u32 fw_processor_version; /* software DDK version */ u32 ddk_version; /* software DDK build no. */ u32 ddk_build; /* build options bit-field */ u32 build_options; /* initialisation options bit-field */ struct rogue_fwif_init_options init_options; /* Information is valid */ bool updated __aligned(4); u32 padding; } __aligned(8); /* ****************************************************************************** * Updated configuration post FW data init. ****************************************************************************** */ struct rogue_fwif_runtime_cfg { /* APM latency in ms before signalling IDLE to the host */ u32 active_pm_latency_ms; /* Compatibility and other flags */ u32 runtime_cfg_flags; /* * If set, APM latency does not reset to system default each GPU power * transition */ bool active_pm_latency_persistant __aligned(4); /* Core clock speed, currently only used to calculate timer ticks */ u32 core_clock_speed; /* Last number of dusts change requested by the host */ u32 default_dusts_num_init; /* Periodic Hardware Reset configuration values */ u32 phr_mode; /* New number of milliseconds C/S is allowed to last */ u32 hcs_deadline_ms; /* The watchdog period in microseconds */ u32 wdg_period_us; /* Array of priorities per OS */ u32 osid_priority[ROGUE_FW_MAX_NUM_OS]; /* On-demand allocated HWPerf buffer address, to be passed to the FW */ u32 hwperf_buf_fw_addr; bool padding __aligned(4); }; /* ***************************************************************************** * Control data for ROGUE ***************************************************************************** */ #define ROGUE_FWIF_HWR_DEBUG_DUMP_ALL (99999U) enum rogue_fwif_tpu_dm { ROGUE_FWIF_TPU_DM_PDM = 0, ROGUE_FWIF_TPU_DM_VDM = 1, ROGUE_FWIF_TPU_DM_CDM = 2, ROGUE_FWIF_TPU_DM_TDM = 3, ROGUE_FWIF_TPU_DM_LAST }; enum rogue_fwif_gpio_val_mode { /* No GPIO validation */ ROGUE_FWIF_GPIO_VAL_OFF = 0, /* * Simple test case that initiates by sending data via the GPIO and then * sends back any data received over the GPIO */ ROGUE_FWIF_GPIO_VAL_GENERAL = 1, /* * More complex test case that writes and reads data across the entire * GPIO AP address range. */ ROGUE_FWIF_GPIO_VAL_AP = 2, /* Validates the GPIO Testbench. */ ROGUE_FWIF_GPIO_VAL_TESTBENCH = 5, /* Send and then receive each byte in the range 0-255. */ ROGUE_FWIF_GPIO_VAL_LOOPBACK = 6, /* Send and then receive each power-of-2 byte in the range 0-255. */ ROGUE_FWIF_GPIO_VAL_LOOPBACK_LITE = 7, ROGUE_FWIF_GPIO_VAL_LAST }; enum fw_perf_conf { FW_PERF_CONF_NONE = 0, FW_PERF_CONF_ICACHE = 1, FW_PERF_CONF_DCACHE = 2, FW_PERF_CONF_JTLB_INSTR = 5, FW_PERF_CONF_INSTRUCTIONS = 6 }; enum fw_boot_stage { FW_BOOT_STAGE_TLB_INIT_FAILURE = -2, FW_BOOT_STAGE_NOT_AVAILABLE = -1, FW_BOOT_NOT_STARTED = 0, FW_BOOT_BLDR_STARTED = 1, FW_BOOT_CACHE_DONE, FW_BOOT_TLB_DONE, FW_BOOT_MAIN_STARTED, FW_BOOT_ALIGNCHECKS_DONE, FW_BOOT_INIT_DONE, }; /* * Kernel CCB return slot responses. Usage of bit-fields instead of bare * integers allows FW to possibly pack-in several responses for each single kCCB * command. */ /* Command executed (return status from FW) */ #define ROGUE_FWIF_KCCB_RTN_SLOT_CMD_EXECUTED BIT(0) /* A cleanup was requested but resource busy */ #define ROGUE_FWIF_KCCB_RTN_SLOT_CLEANUP_BUSY BIT(1) /* Poll failed in FW for a HW operation to complete */ #define ROGUE_FWIF_KCCB_RTN_SLOT_POLL_FAILURE BIT(2) /* Reset value of a kCCB return slot (set by host) */ #define ROGUE_FWIF_KCCB_RTN_SLOT_NO_RESPONSE 0x0U struct rogue_fwif_connection_ctl { /* Fw-Os connection states */ enum rogue_fwif_connection_fw_state connection_fw_state; enum rogue_fwif_connection_os_state connection_os_state; u32 alive_fw_token; u32 alive_os_token; } __aligned(8); struct rogue_fwif_osinit { /* Kernel CCB */ u32 kernel_ccbctl_fw_addr; u32 kernel_ccb_fw_addr; u32 kernel_ccb_rtn_slots_fw_addr; /* Firmware CCB */ u32 firmware_ccbctl_fw_addr; u32 firmware_ccb_fw_addr; /* Workload Estimation Firmware CCB */ u32 work_est_firmware_ccbctl_fw_addr; u32 work_est_firmware_ccb_fw_addr; u32 rogue_fwif_hwr_info_buf_ctl_fw_addr; u32 hwr_debug_dump_limit; u32 fw_os_data_fw_addr; /* Compatibility checks to be populated by the Firmware */ struct rogue_fwif_compchecks rogue_comp_checks; } __aligned(8); /* BVNC Features */ struct rogue_hwperf_bvnc_block { /* Counter block ID, see ROGUE_HWPERF_CNTBLK_ID */ u16 block_id; /* Number of counters in this block type */ u16 num_counters; /* Number of blocks of this type */ u16 num_blocks; u16 reserved; }; #define ROGUE_HWPERF_MAX_BVNC_LEN (24) #define ROGUE_HWPERF_MAX_BVNC_BLOCK_LEN (16U) /* BVNC Features */ struct rogue_hwperf_bvnc { /* BVNC string */ char bvnc_string[ROGUE_HWPERF_MAX_BVNC_LEN]; /* See ROGUE_HWPERF_FEATURE_FLAGS */ u32 bvnc_km_feature_flags; /* Number of blocks described in aBvncBlocks */ u16 num_bvnc_blocks; /* Number of GPU cores present */ u16 bvnc_gpu_cores; /* Supported Performance Blocks for BVNC */ struct rogue_hwperf_bvnc_block bvnc_blocks[ROGUE_HWPERF_MAX_BVNC_BLOCK_LEN]; }; PVR_FW_STRUCT_SIZE_ASSERT(struct rogue_hwperf_bvnc); struct rogue_fwif_sysinit { /* Fault read address */ aligned_u64 fault_phys_addr; /* PDS execution base */ aligned_u64 pds_exec_base; /* UCS execution base */ aligned_u64 usc_exec_base; /* FBCDC bindless texture state table base */ aligned_u64 fbcdc_state_table_base; aligned_u64 fbcdc_large_state_table_base; /* Texture state base */ aligned_u64 texture_heap_base; /* Event filter for Firmware events */ u64 hw_perf_filter; aligned_u64 slc3_fence_dev_addr; u32 tpu_trilinear_frac_mask[ROGUE_FWIF_TPU_DM_LAST] __aligned(8); /* Signature and Checksum Buffers for DMs */ struct rogue_fwif_sigbuf_ctl sigbuf_ctl[PVR_FWIF_DM_MAX]; struct rogue_fwif_pdvfs_opp pdvfs_opp_info; struct rogue_fwif_dma_addr coremem_data_store; struct rogue_fwif_counter_dump_ctl counter_dump_ctl; u32 filter_flags; u32 runtime_cfg_fw_addr; u32 trace_buf_ctl_fw_addr; u32 fw_sys_data_fw_addr; u32 gpu_util_fw_cb_ctl_fw_addr; u32 reg_cfg_fw_addr; u32 hwperf_ctl_fw_addr; u32 align_checks; /* Core clock speed at FW boot time */ u32 initial_core_clock_speed; /* APM latency in ms before signalling IDLE to the host */ u32 active_pm_latency_ms; /* Flag to be set by the Firmware after successful start */ bool firmware_started __aligned(4); /* Host/FW Trace synchronisation Partition Marker */ u32 marker_val; /* Firmware initialization complete time */ u32 firmware_started_timestamp; u32 jones_disable_mask; /* Firmware performance counter config */ enum fw_perf_conf firmware_perf; /* * FW Pointer to memory containing core clock rate in Hz. * Firmware (PDVFS) updates the memory when running on non primary FW * thread to communicate to host driver. */ u32 core_clock_rate_fw_addr; enum rogue_fwif_gpio_val_mode gpio_validation_mode; /* Used in HWPerf for decoding BVNC Features */ struct rogue_hwperf_bvnc bvnc_km_feature_flags; /* Value to write into ROGUE_CR_TFBC_COMPRESSION_CONTROL */ u32 tfbc_compression_control; } __aligned(8); /* ***************************************************************************** * Timer correlation shared data and defines ***************************************************************************** */ struct rogue_fwif_time_corr { aligned_u64 os_timestamp; aligned_u64 os_mono_timestamp; aligned_u64 cr_timestamp; /* * Utility variable used to convert CR timer deltas to OS timer deltas * (nS), where the deltas are relative to the timestamps above: * deltaOS = (deltaCR * K) >> decimal_shift, see full explanation below */ aligned_u64 cr_delta_to_os_delta_kns; u32 core_clock_speed; u32 reserved; } __aligned(8); /* * The following macros are used to help converting FW timestamps to the Host * time domain. On the FW the ROGUE_CR_TIMER counter is used to keep track of * time; it increments by 1 every 256 GPU clock ticks, so the general * formula to perform the conversion is: * * [ GPU clock speed in Hz, if (scale == 10^9) then deltaOS is in nS, * otherwise if (scale == 10^6) then deltaOS is in uS ] * * deltaCR * 256 256 * scale * deltaOS = --------------- * scale = deltaCR * K [ K = --------------- ] * GPUclockspeed GPUclockspeed * * The actual K is multiplied by 2^20 (and deltaCR * K is divided by 2^20) * to get some better accuracy and to avoid returning 0 in the integer * division 256000000/GPUfreq if GPUfreq is greater than 256MHz. * This is the same as keeping K as a decimal number. * * The maximum deltaOS is slightly more than 5hrs for all GPU frequencies * (deltaCR * K is more or less a constant), and it's relative to the base * OS timestamp sampled as a part of the timer correlation data. * This base is refreshed on GPU power-on, DVFS transition and periodic * frequency calibration (executed every few seconds if the FW is doing * some work), so as long as the GPU is doing something and one of these * events is triggered then deltaCR * K will not overflow and deltaOS will be * correct. */ #define ROGUE_FWIF_CRDELTA_TO_OSDELTA_ACCURACY_SHIFT (20) #define ROGUE_FWIF_GET_DELTA_OSTIME_NS(delta_cr, k) \ (((delta_cr) * (k)) >> ROGUE_FWIF_CRDELTA_TO_OSDELTA_ACCURACY_SHIFT) /* ****************************************************************************** * GPU Utilisation ****************************************************************************** */ /* See rogue_common.h for a list of GPU states */ #define ROGUE_FWIF_GPU_UTIL_TIME_MASK \ (0xFFFFFFFFFFFFFFFFull & ~ROGUE_FWIF_GPU_UTIL_STATE_MASK) #define ROGUE_FWIF_GPU_UTIL_GET_TIME(word) \ ((word)(&ROGUE_FWIF_GPU_UTIL_TIME_MASK)) #define ROGUE_FWIF_GPU_UTIL_GET_STATE(word) \ ((word)(&ROGUE_FWIF_GPU_UTIL_STATE_MASK)) /* * The OS timestamps computed by the FW are approximations of the real time, * which means they could be slightly behind or ahead the real timer on the * Host. In some cases we can perform subtractions between FW approximated * timestamps and real OS timestamps, so we need a form of protection against * negative results if for instance the FW one is a bit ahead of time. */ #define ROGUE_FWIF_GPU_UTIL_GET_PERIOD(newtime, oldtime) \ (((newtime) > (oldtime)) ? ((newtime) - (oldtime)) : 0U) #define ROGUE_FWIF_GPU_UTIL_MAKE_WORD(time, state) \ (ROGUE_FWIF_GPU_UTIL_GET_TIME(time) | \ ROGUE_FWIF_GPU_UTIL_GET_STATE(state)) /* * The timer correlation array must be big enough to ensure old entries won't be * overwritten before all the HWPerf events linked to those entries are * processed by the MISR. The update frequency of this array depends on how fast * the system can change state (basically how small the APM latency is) and * perform DVFS transitions. * * The minimum size is 2 (not 1) to avoid race conditions between the FW reading * an entry while the Host is updating it. With 2 entries in the worst case the * FW will read old data, which is still quite ok if the Host is updating the * timer correlation at that time. */ #define ROGUE_FWIF_TIME_CORR_ARRAY_SIZE 256U #define ROGUE_FWIF_TIME_CORR_CURR_INDEX(seqcount) \ ((seqcount) % ROGUE_FWIF_TIME_CORR_ARRAY_SIZE) /* Make sure the timer correlation array size is a power of 2 */ static_assert((ROGUE_FWIF_TIME_CORR_ARRAY_SIZE & (ROGUE_FWIF_TIME_CORR_ARRAY_SIZE - 1U)) == 0U, "ROGUE_FWIF_TIME_CORR_ARRAY_SIZE must be a power of two"); struct rogue_fwif_gpu_util_fwcb { struct rogue_fwif_time_corr time_corr[ROGUE_FWIF_TIME_CORR_ARRAY_SIZE]; u32 time_corr_seq_count; /* Compatibility and other flags */ u32 gpu_util_flags; /* Last GPU state + OS time of the last state update */ aligned_u64 last_word; /* Counters for the amount of time the GPU was active/idle/blocked */ aligned_u64 stats_counters[PVR_FWIF_GPU_UTIL_STATE_NUM]; } __aligned(8); struct rogue_fwif_rta_ctl { /* Render number */ u32 render_target_index; /* index in RTA */ u32 current_render_target; /* total active RTs */ u32 active_render_targets; /* total active RTs from the first TA kick, for OOM */ u32 cumul_active_render_targets; /* Array of valid RT indices */ u32 valid_render_targets_fw_addr; /* Array of number of occurred partial renders per render target */ u32 rta_num_partial_renders_fw_addr; /* Number of render targets in the array */ u32 max_rts; /* Compatibility and other flags */ u32 rta_ctl_flags; } __aligned(8); struct rogue_fwif_freelist { aligned_u64 freelist_dev_addr; aligned_u64 current_dev_addr; u32 current_stack_top; u32 max_pages; u32 grow_pages; /* HW pages */ u32 current_pages; u32 allocated_page_count; u32 allocated_mmu_page_count; u32 freelist_id; bool grow_pending __aligned(4); /* Pages that should be used only when OOM is reached */ u32 ready_pages; /* Compatibility and other flags */ u32 freelist_flags; /* PM Global PB on which Freelist is loaded */ u32 pm_global_pb; u32 padding; } __aligned(8); /* ****************************************************************************** * HWRTData ****************************************************************************** */ /* HWRTData flags */ /* Deprecated flags 1:0 */ #define HWRTDATA_HAS_LAST_GEOM BIT(2) #define HWRTDATA_PARTIAL_RENDERED BIT(3) #define HWRTDATA_DISABLE_TILE_REORDERING BIT(4) #define HWRTDATA_NEED_BRN65101_BLIT BIT(5) #define HWRTDATA_FIRST_BRN65101_STRIP BIT(6) #define HWRTDATA_NEED_BRN67182_2ND_RENDER BIT(7) enum rogue_fwif_rtdata_state { ROGUE_FWIF_RTDATA_STATE_NONE = 0, ROGUE_FWIF_RTDATA_STATE_KICK_GEOM, ROGUE_FWIF_RTDATA_STATE_KICK_GEOM_FIRST, ROGUE_FWIF_RTDATA_STATE_GEOM_FINISHED, ROGUE_FWIF_RTDATA_STATE_KICK_FRAG, ROGUE_FWIF_RTDATA_STATE_FRAG_FINISHED, ROGUE_FWIF_RTDATA_STATE_FRAG_CONTEXT_STORED, ROGUE_FWIF_RTDATA_STATE_GEOM_OUTOFMEM, ROGUE_FWIF_RTDATA_STATE_PARTIALRENDERFINISHED, /* * In case of HWR, we can't set the RTDATA state to NONE, as this will * cause any TA to become a first TA. To ensure all related TA's are * skipped, we use the HWR state */ ROGUE_FWIF_RTDATA_STATE_HWR, ROGUE_FWIF_RTDATA_STATE_UNKNOWN = 0x7FFFFFFFU }; struct rogue_fwif_hwrtdata_common { bool geom_caches_need_zeroing __aligned(4); u32 screen_pixel_max; aligned_u64 multi_sample_ctl; u64 flipped_multi_sample_ctl; u32 tpc_stride; u32 tpc_size; u32 te_screen; u32 mtile_stride; u32 teaa; u32 te_mtile1; u32 te_mtile2; u32 isp_merge_lower_x; u32 isp_merge_lower_y; u32 isp_merge_upper_x; u32 isp_merge_upper_y; u32 isp_merge_scale_x; u32 isp_merge_scale_y; u32 rgn_header_size; u32 isp_mtile_size; u32 padding; } __aligned(8); struct rogue_fwif_hwrtdata { /* MList Data Store */ aligned_u64 pm_mlist_dev_addr; aligned_u64 vce_cat_base[4]; aligned_u64 vce_last_cat_base[4]; aligned_u64 te_cat_base[4]; aligned_u64 te_last_cat_base[4]; aligned_u64 alist_cat_base; aligned_u64 alist_last_cat_base; aligned_u64 pm_alist_stack_pointer; u32 pm_mlist_stack_pointer; u32 hwrt_data_common_fw_addr; u32 hwrt_data_flags; enum rogue_fwif_rtdata_state state; u32 freelists_fw_addr[MAX_FREELISTS_SIZE] __aligned(8); u32 freelist_hwr_snapshot[MAX_FREELISTS_SIZE]; aligned_u64 vheap_table_dev_addr; struct rogue_fwif_rta_ctl rta_ctl; aligned_u64 tail_ptrs_dev_addr; aligned_u64 macrotile_array_dev_addr; aligned_u64 rgn_header_dev_addr; aligned_u64 rtc_dev_addr; u32 owner_geom_not_used_by_host __aligned(8); bool geom_caches_need_zeroing __aligned(4); struct rogue_fwif_cleanup_ctl cleanup_state __aligned(64); } __aligned(8); /* ****************************************************************************** * Sync checkpoints ****************************************************************************** */ #define PVR_SYNC_CHECKPOINT_UNDEF 0x000 #define PVR_SYNC_CHECKPOINT_ACTIVE 0xac1 /* Checkpoint has not signaled. */ #define PVR_SYNC_CHECKPOINT_SIGNALED 0x519 /* Checkpoint has signaled. */ #define PVR_SYNC_CHECKPOINT_ERRORED 0xeff /* Checkpoint has been errored. */ #include "pvr_rogue_fwif_check.h" #endif /* PVR_ROGUE_FWIF_H */