301 lines
10 KiB
C
301 lines
10 KiB
C
/* SPDX-License-Identifier: GPL-2.0-only
|
|
*
|
|
* Copyright (c) 2019-2021, The Linux Foundation. All rights reserved.
|
|
* Copyright (c) 2021-2023 Qualcomm Innovation Center, Inc. All rights reserved.
|
|
*/
|
|
|
|
#ifndef _QAIC_H_
|
|
#define _QAIC_H_
|
|
|
|
#include <linux/interrupt.h>
|
|
#include <linux/kref.h>
|
|
#include <linux/mhi.h>
|
|
#include <linux/mutex.h>
|
|
#include <linux/pci.h>
|
|
#include <linux/spinlock.h>
|
|
#include <linux/srcu.h>
|
|
#include <linux/wait.h>
|
|
#include <linux/workqueue.h>
|
|
#include <drm/drm_device.h>
|
|
#include <drm/drm_gem.h>
|
|
|
|
#define QAIC_DBC_BASE SZ_128K
|
|
#define QAIC_DBC_SIZE SZ_4K
|
|
|
|
#define QAIC_NO_PARTITION -1
|
|
|
|
#define QAIC_DBC_OFF(i) ((i) * QAIC_DBC_SIZE + QAIC_DBC_BASE)
|
|
|
|
#define to_qaic_bo(obj) container_of(obj, struct qaic_bo, base)
|
|
#define to_qaic_drm_device(dev) container_of(dev, struct qaic_drm_device, drm)
|
|
#define to_drm(qddev) (&(qddev)->drm)
|
|
#define to_accel_kdev(qddev) (to_drm(qddev)->accel->kdev) /* Return Linux device of accel node */
|
|
|
|
enum __packed dev_states {
|
|
/* Device is offline or will be very soon */
|
|
QAIC_OFFLINE,
|
|
/* Device is booting, not clear if it's in a usable state */
|
|
QAIC_BOOT,
|
|
/* Device is fully operational */
|
|
QAIC_ONLINE,
|
|
};
|
|
|
|
extern bool datapath_polling;
|
|
|
|
struct qaic_user {
|
|
/* Uniquely identifies this user for the device */
|
|
int handle;
|
|
struct kref ref_count;
|
|
/* Char device opened by this user */
|
|
struct qaic_drm_device *qddev;
|
|
/* Node in list of users that opened this drm device */
|
|
struct list_head node;
|
|
/* SRCU used to synchronize this user during cleanup */
|
|
struct srcu_struct qddev_lock;
|
|
atomic_t chunk_id;
|
|
};
|
|
|
|
struct dma_bridge_chan {
|
|
/* Pointer to device strcut maintained by driver */
|
|
struct qaic_device *qdev;
|
|
/* ID of this DMA bridge channel(DBC) */
|
|
unsigned int id;
|
|
/* Synchronizes access to xfer_list */
|
|
spinlock_t xfer_lock;
|
|
/* Base address of request queue */
|
|
void *req_q_base;
|
|
/* Base address of response queue */
|
|
void *rsp_q_base;
|
|
/*
|
|
* Base bus address of request queue. Response queue bus address can be
|
|
* calculated by adding request queue size to this variable
|
|
*/
|
|
dma_addr_t dma_addr;
|
|
/* Total size of request and response queue in byte */
|
|
u32 total_size;
|
|
/* Capacity of request/response queue */
|
|
u32 nelem;
|
|
/* The user that opened this DBC */
|
|
struct qaic_user *usr;
|
|
/*
|
|
* Request ID of next memory handle that goes in request queue. One
|
|
* memory handle can enqueue more than one request elements, all
|
|
* this requests that belong to same memory handle have same request ID
|
|
*/
|
|
u16 next_req_id;
|
|
/* true: DBC is in use; false: DBC not in use */
|
|
bool in_use;
|
|
/*
|
|
* Base address of device registers. Used to read/write request and
|
|
* response queue's head and tail pointer of this DBC.
|
|
*/
|
|
void __iomem *dbc_base;
|
|
/* Head of list where each node is a memory handle queued in request queue */
|
|
struct list_head xfer_list;
|
|
/* Synchronizes DBC readers during cleanup */
|
|
struct srcu_struct ch_lock;
|
|
/*
|
|
* When this DBC is released, any thread waiting on this wait queue is
|
|
* woken up
|
|
*/
|
|
wait_queue_head_t dbc_release;
|
|
/* Head of list where each node is a bo associated with this DBC */
|
|
struct list_head bo_lists;
|
|
/* The irq line for this DBC. Used for polling */
|
|
unsigned int irq;
|
|
/* Polling work item to simulate interrupts */
|
|
struct work_struct poll_work;
|
|
};
|
|
|
|
struct qaic_device {
|
|
/* Pointer to base PCI device struct of our physical device */
|
|
struct pci_dev *pdev;
|
|
/* Req. ID of request that will be queued next in MHI control device */
|
|
u32 next_seq_num;
|
|
/* Base address of bar 0 */
|
|
void __iomem *bar_0;
|
|
/* Base address of bar 2 */
|
|
void __iomem *bar_2;
|
|
/* Controller structure for MHI devices */
|
|
struct mhi_controller *mhi_cntrl;
|
|
/* MHI control channel device */
|
|
struct mhi_device *cntl_ch;
|
|
/* List of requests queued in MHI control device */
|
|
struct list_head cntl_xfer_list;
|
|
/* Synchronizes MHI control device transactions and its xfer list */
|
|
struct mutex cntl_mutex;
|
|
/* Array of DBC struct of this device */
|
|
struct dma_bridge_chan *dbc;
|
|
/* Work queue for tasks related to MHI control device */
|
|
struct workqueue_struct *cntl_wq;
|
|
/* Synchronizes all the users of device during cleanup */
|
|
struct srcu_struct dev_lock;
|
|
/* Track the state of the device during resets */
|
|
enum dev_states dev_state;
|
|
/* true: single MSI is used to operate device */
|
|
bool single_msi;
|
|
/*
|
|
* true: A tx MHI transaction has failed and a rx buffer is still queued
|
|
* in control device. Such a buffer is considered lost rx buffer
|
|
* false: No rx buffer is lost in control device
|
|
*/
|
|
bool cntl_lost_buf;
|
|
/* Maximum number of DBC supported by this device */
|
|
u32 num_dbc;
|
|
/* Reference to the drm_device for this device when it is created */
|
|
struct qaic_drm_device *qddev;
|
|
/* Generate the CRC of a control message */
|
|
u32 (*gen_crc)(void *msg);
|
|
/* Validate the CRC of a control message */
|
|
bool (*valid_crc)(void *msg);
|
|
/* MHI "QAIC_TIMESYNC" channel device */
|
|
struct mhi_device *qts_ch;
|
|
/* Work queue for tasks related to MHI "QAIC_TIMESYNC" channel */
|
|
struct workqueue_struct *qts_wq;
|
|
};
|
|
|
|
struct qaic_drm_device {
|
|
/* The drm device struct of this drm device */
|
|
struct drm_device drm;
|
|
/* Pointer to the root device struct driven by this driver */
|
|
struct qaic_device *qdev;
|
|
/*
|
|
* The physical device can be partition in number of logical devices.
|
|
* And each logical device is given a partition id. This member stores
|
|
* that id. QAIC_NO_PARTITION is a sentinel used to mark that this drm
|
|
* device is the actual physical device
|
|
*/
|
|
s32 partition_id;
|
|
/* Head in list of users who have opened this drm device */
|
|
struct list_head users;
|
|
/* Synchronizes access to users list */
|
|
struct mutex users_mutex;
|
|
};
|
|
|
|
struct qaic_bo {
|
|
struct drm_gem_object base;
|
|
/* Scatter/gather table for allocate/imported BO */
|
|
struct sg_table *sgt;
|
|
/* Head in list of slices of this BO */
|
|
struct list_head slices;
|
|
/* Total nents, for all slices of this BO */
|
|
int total_slice_nents;
|
|
/*
|
|
* Direction of transfer. It can assume only two value DMA_TO_DEVICE and
|
|
* DMA_FROM_DEVICE.
|
|
*/
|
|
int dir;
|
|
/* The pointer of the DBC which operates on this BO */
|
|
struct dma_bridge_chan *dbc;
|
|
/* Number of slice that belongs to this buffer */
|
|
u32 nr_slice;
|
|
/* Number of slice that have been transferred by DMA engine */
|
|
u32 nr_slice_xfer_done;
|
|
/* true = BO is queued for execution, true = BO is not queued */
|
|
bool queued;
|
|
/*
|
|
* If true then user has attached slicing information to this BO by
|
|
* calling DRM_IOCTL_QAIC_ATTACH_SLICE_BO ioctl.
|
|
*/
|
|
bool sliced;
|
|
/* Request ID of this BO if it is queued for execution */
|
|
u16 req_id;
|
|
/* Handle assigned to this BO */
|
|
u32 handle;
|
|
/* Wait on this for completion of DMA transfer of this BO */
|
|
struct completion xfer_done;
|
|
/*
|
|
* Node in linked list where head is dbc->xfer_list.
|
|
* This link list contain BO's that are queued for DMA transfer.
|
|
*/
|
|
struct list_head xfer_list;
|
|
/*
|
|
* Node in linked list where head is dbc->bo_lists.
|
|
* This link list contain BO's that are associated with the DBC it is
|
|
* linked to.
|
|
*/
|
|
struct list_head bo_list;
|
|
struct {
|
|
/*
|
|
* Latest timestamp(ns) at which kernel received a request to
|
|
* execute this BO
|
|
*/
|
|
u64 req_received_ts;
|
|
/*
|
|
* Latest timestamp(ns) at which kernel enqueued requests of
|
|
* this BO for execution in DMA queue
|
|
*/
|
|
u64 req_submit_ts;
|
|
/*
|
|
* Latest timestamp(ns) at which kernel received a completion
|
|
* interrupt for requests of this BO
|
|
*/
|
|
u64 req_processed_ts;
|
|
/*
|
|
* Number of elements already enqueued in DMA queue before
|
|
* enqueuing requests of this BO
|
|
*/
|
|
u32 queue_level_before;
|
|
} perf_stats;
|
|
/* Synchronizes BO operations */
|
|
struct mutex lock;
|
|
};
|
|
|
|
struct bo_slice {
|
|
/* Mapped pages */
|
|
struct sg_table *sgt;
|
|
/* Number of requests required to queue in DMA queue */
|
|
int nents;
|
|
/* See enum dma_data_direction */
|
|
int dir;
|
|
/* Actual requests that will be copied in DMA queue */
|
|
struct dbc_req *reqs;
|
|
struct kref ref_count;
|
|
/* true: No DMA transfer required */
|
|
bool no_xfer;
|
|
/* Pointer to the parent BO handle */
|
|
struct qaic_bo *bo;
|
|
/* Node in list of slices maintained by parent BO */
|
|
struct list_head slice;
|
|
/* Size of this slice in bytes */
|
|
u64 size;
|
|
/* Offset of this slice in buffer */
|
|
u64 offset;
|
|
};
|
|
|
|
int get_dbc_req_elem_size(void);
|
|
int get_dbc_rsp_elem_size(void);
|
|
int get_cntl_version(struct qaic_device *qdev, struct qaic_user *usr, u16 *major, u16 *minor);
|
|
int qaic_manage_ioctl(struct drm_device *dev, void *data, struct drm_file *file_priv);
|
|
void qaic_mhi_ul_xfer_cb(struct mhi_device *mhi_dev, struct mhi_result *mhi_result);
|
|
|
|
void qaic_mhi_dl_xfer_cb(struct mhi_device *mhi_dev, struct mhi_result *mhi_result);
|
|
|
|
int qaic_control_open(struct qaic_device *qdev);
|
|
void qaic_control_close(struct qaic_device *qdev);
|
|
void qaic_release_usr(struct qaic_device *qdev, struct qaic_user *usr);
|
|
|
|
irqreturn_t dbc_irq_threaded_fn(int irq, void *data);
|
|
irqreturn_t dbc_irq_handler(int irq, void *data);
|
|
int disable_dbc(struct qaic_device *qdev, u32 dbc_id, struct qaic_user *usr);
|
|
void enable_dbc(struct qaic_device *qdev, u32 dbc_id, struct qaic_user *usr);
|
|
void wakeup_dbc(struct qaic_device *qdev, u32 dbc_id);
|
|
void release_dbc(struct qaic_device *qdev, u32 dbc_id);
|
|
|
|
void wake_all_cntl(struct qaic_device *qdev);
|
|
void qaic_dev_reset_clean_local_state(struct qaic_device *qdev);
|
|
|
|
struct drm_gem_object *qaic_gem_prime_import(struct drm_device *dev, struct dma_buf *dma_buf);
|
|
|
|
int qaic_create_bo_ioctl(struct drm_device *dev, void *data, struct drm_file *file_priv);
|
|
int qaic_mmap_bo_ioctl(struct drm_device *dev, void *data, struct drm_file *file_priv);
|
|
int qaic_attach_slice_bo_ioctl(struct drm_device *dev, void *data, struct drm_file *file_priv);
|
|
int qaic_execute_bo_ioctl(struct drm_device *dev, void *data, struct drm_file *file_priv);
|
|
int qaic_partial_execute_bo_ioctl(struct drm_device *dev, void *data, struct drm_file *file_priv);
|
|
int qaic_wait_bo_ioctl(struct drm_device *dev, void *data, struct drm_file *file_priv);
|
|
int qaic_perf_stats_bo_ioctl(struct drm_device *dev, void *data, struct drm_file *file_priv);
|
|
int qaic_detach_slice_bo_ioctl(struct drm_device *dev, void *data, struct drm_file *file_priv);
|
|
void irq_polling_work(struct work_struct *work);
|
|
|
|
#endif /* _QAIC_H_ */
|