/* SPDX-License-Identifier: GPL-2.0-only * * Copyright (c) 2019-2021, The Linux Foundation. All rights reserved. * Copyright (c) 2021-2023 Qualcomm Innovation Center, Inc. All rights reserved. */ #ifndef _QAIC_H_ #define _QAIC_H_ #include #include #include #include #include #include #include #include #include #include #include #define QAIC_DBC_BASE SZ_128K #define QAIC_DBC_SIZE SZ_4K #define QAIC_NO_PARTITION -1 #define QAIC_DBC_OFF(i) ((i) * QAIC_DBC_SIZE + QAIC_DBC_BASE) #define to_qaic_bo(obj) container_of(obj, struct qaic_bo, base) #define to_qaic_drm_device(dev) container_of(dev, struct qaic_drm_device, drm) #define to_drm(qddev) (&(qddev)->drm) #define to_accel_kdev(qddev) (to_drm(qddev)->accel->kdev) /* Return Linux device of accel node */ #define to_qaic_device(dev) (to_qaic_drm_device((dev))->qdev) enum __packed dev_states { /* Device is offline or will be very soon */ QAIC_OFFLINE, /* Device is booting, not clear if it's in a usable state */ QAIC_BOOT, /* Device is fully operational */ QAIC_ONLINE, }; extern bool datapath_polling; struct qaic_user { /* Uniquely identifies this user for the device */ int handle; struct kref ref_count; /* Char device opened by this user */ struct qaic_drm_device *qddev; /* Node in list of users that opened this drm device */ struct list_head node; /* SRCU used to synchronize this user during cleanup */ struct srcu_struct qddev_lock; atomic_t chunk_id; }; struct dma_bridge_chan { /* Pointer to device strcut maintained by driver */ struct qaic_device *qdev; /* ID of this DMA bridge channel(DBC) */ unsigned int id; /* Synchronizes access to xfer_list */ spinlock_t xfer_lock; /* Base address of request queue */ void *req_q_base; /* Base address of response queue */ void *rsp_q_base; /* * Base bus address of request queue. Response queue bus address can be * calculated by adding request queue size to this variable */ dma_addr_t dma_addr; /* Total size of request and response queue in byte */ u32 total_size; /* Capacity of request/response queue */ u32 nelem; /* The user that opened this DBC */ struct qaic_user *usr; /* * Request ID of next memory handle that goes in request queue. One * memory handle can enqueue more than one request elements, all * this requests that belong to same memory handle have same request ID */ u16 next_req_id; /* true: DBC is in use; false: DBC not in use */ bool in_use; /* * Base address of device registers. Used to read/write request and * response queue's head and tail pointer of this DBC. */ void __iomem *dbc_base; /* Head of list where each node is a memory handle queued in request queue */ struct list_head xfer_list; /* Synchronizes DBC readers during cleanup */ struct srcu_struct ch_lock; /* * When this DBC is released, any thread waiting on this wait queue is * woken up */ wait_queue_head_t dbc_release; /* Head of list where each node is a bo associated with this DBC */ struct list_head bo_lists; /* The irq line for this DBC. Used for polling */ unsigned int irq; /* Polling work item to simulate interrupts */ struct work_struct poll_work; }; struct qaic_device { /* Pointer to base PCI device struct of our physical device */ struct pci_dev *pdev; /* Req. ID of request that will be queued next in MHI control device */ u32 next_seq_num; /* Base address of bar 0 */ void __iomem *bar_0; /* Base address of bar 2 */ void __iomem *bar_2; /* Controller structure for MHI devices */ struct mhi_controller *mhi_cntrl; /* MHI control channel device */ struct mhi_device *cntl_ch; /* List of requests queued in MHI control device */ struct list_head cntl_xfer_list; /* Synchronizes MHI control device transactions and its xfer list */ struct mutex cntl_mutex; /* Array of DBC struct of this device */ struct dma_bridge_chan *dbc; /* Work queue for tasks related to MHI control device */ struct workqueue_struct *cntl_wq; /* Synchronizes all the users of device during cleanup */ struct srcu_struct dev_lock; /* Track the state of the device during resets */ enum dev_states dev_state; /* true: single MSI is used to operate device */ bool single_msi; /* * true: A tx MHI transaction has failed and a rx buffer is still queued * in control device. Such a buffer is considered lost rx buffer * false: No rx buffer is lost in control device */ bool cntl_lost_buf; /* Maximum number of DBC supported by this device */ u32 num_dbc; /* Reference to the drm_device for this device when it is created */ struct qaic_drm_device *qddev; /* Generate the CRC of a control message */ u32 (*gen_crc)(void *msg); /* Validate the CRC of a control message */ bool (*valid_crc)(void *msg); /* MHI "QAIC_TIMESYNC" channel device */ struct mhi_device *qts_ch; /* Work queue for tasks related to MHI "QAIC_TIMESYNC" channel */ struct workqueue_struct *qts_wq; }; struct qaic_drm_device { /* The drm device struct of this drm device */ struct drm_device drm; /* Pointer to the root device struct driven by this driver */ struct qaic_device *qdev; /* * The physical device can be partition in number of logical devices. * And each logical device is given a partition id. This member stores * that id. QAIC_NO_PARTITION is a sentinel used to mark that this drm * device is the actual physical device */ s32 partition_id; /* Head in list of users who have opened this drm device */ struct list_head users; /* Synchronizes access to users list */ struct mutex users_mutex; }; struct qaic_bo { struct drm_gem_object base; /* Scatter/gather table for allocate/imported BO */ struct sg_table *sgt; /* Head in list of slices of this BO */ struct list_head slices; /* Total nents, for all slices of this BO */ int total_slice_nents; /* * Direction of transfer. It can assume only two value DMA_TO_DEVICE and * DMA_FROM_DEVICE. */ int dir; /* The pointer of the DBC which operates on this BO */ struct dma_bridge_chan *dbc; /* Number of slice that belongs to this buffer */ u32 nr_slice; /* Number of slice that have been transferred by DMA engine */ u32 nr_slice_xfer_done; /* * If true then user has attached slicing information to this BO by * calling DRM_IOCTL_QAIC_ATTACH_SLICE_BO ioctl. */ bool sliced; /* Request ID of this BO if it is queued for execution */ u16 req_id; /* Handle assigned to this BO */ u32 handle; /* Wait on this for completion of DMA transfer of this BO */ struct completion xfer_done; /* * Node in linked list where head is dbc->xfer_list. * This link list contain BO's that are queued for DMA transfer. */ struct list_head xfer_list; /* * Node in linked list where head is dbc->bo_lists. * This link list contain BO's that are associated with the DBC it is * linked to. */ struct list_head bo_list; struct { /* * Latest timestamp(ns) at which kernel received a request to * execute this BO */ u64 req_received_ts; /* * Latest timestamp(ns) at which kernel enqueued requests of * this BO for execution in DMA queue */ u64 req_submit_ts; /* * Latest timestamp(ns) at which kernel received a completion * interrupt for requests of this BO */ u64 req_processed_ts; /* * Number of elements already enqueued in DMA queue before * enqueuing requests of this BO */ u32 queue_level_before; } perf_stats; /* Synchronizes BO operations */ struct mutex lock; }; struct bo_slice { /* Mapped pages */ struct sg_table *sgt; /* Number of requests required to queue in DMA queue */ int nents; /* See enum dma_data_direction */ int dir; /* Actual requests that will be copied in DMA queue */ struct dbc_req *reqs; struct kref ref_count; /* true: No DMA transfer required */ bool no_xfer; /* Pointer to the parent BO handle */ struct qaic_bo *bo; /* Node in list of slices maintained by parent BO */ struct list_head slice; /* Size of this slice in bytes */ u64 size; /* Offset of this slice in buffer */ u64 offset; }; int get_dbc_req_elem_size(void); int get_dbc_rsp_elem_size(void); int get_cntl_version(struct qaic_device *qdev, struct qaic_user *usr, u16 *major, u16 *minor); int qaic_manage_ioctl(struct drm_device *dev, void *data, struct drm_file *file_priv); void qaic_mhi_ul_xfer_cb(struct mhi_device *mhi_dev, struct mhi_result *mhi_result); void qaic_mhi_dl_xfer_cb(struct mhi_device *mhi_dev, struct mhi_result *mhi_result); int qaic_control_open(struct qaic_device *qdev); void qaic_control_close(struct qaic_device *qdev); void qaic_release_usr(struct qaic_device *qdev, struct qaic_user *usr); irqreturn_t dbc_irq_threaded_fn(int irq, void *data); irqreturn_t dbc_irq_handler(int irq, void *data); int disable_dbc(struct qaic_device *qdev, u32 dbc_id, struct qaic_user *usr); void enable_dbc(struct qaic_device *qdev, u32 dbc_id, struct qaic_user *usr); void wakeup_dbc(struct qaic_device *qdev, u32 dbc_id); void release_dbc(struct qaic_device *qdev, u32 dbc_id); void wake_all_cntl(struct qaic_device *qdev); void qaic_dev_reset_clean_local_state(struct qaic_device *qdev); struct drm_gem_object *qaic_gem_prime_import(struct drm_device *dev, struct dma_buf *dma_buf); int qaic_create_bo_ioctl(struct drm_device *dev, void *data, struct drm_file *file_priv); int qaic_mmap_bo_ioctl(struct drm_device *dev, void *data, struct drm_file *file_priv); int qaic_attach_slice_bo_ioctl(struct drm_device *dev, void *data, struct drm_file *file_priv); int qaic_execute_bo_ioctl(struct drm_device *dev, void *data, struct drm_file *file_priv); int qaic_partial_execute_bo_ioctl(struct drm_device *dev, void *data, struct drm_file *file_priv); int qaic_wait_bo_ioctl(struct drm_device *dev, void *data, struct drm_file *file_priv); int qaic_perf_stats_bo_ioctl(struct drm_device *dev, void *data, struct drm_file *file_priv); int qaic_detach_slice_bo_ioctl(struct drm_device *dev, void *data, struct drm_file *file_priv); void irq_polling_work(struct work_struct *work); #endif /* _QAIC_H_ */