163 lines
		
	
	
		
			3.9 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			163 lines
		
	
	
		
			3.9 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| // SPDX-License-Identifier: GPL-2.0+
 | |
| /* Copyright (c) 2015-2016 Quantenna Communications. All rights reserved. */
 | |
| 
 | |
| #include <linux/types.h>
 | |
| #include <linux/io.h>
 | |
| 
 | |
| #include "shm_ipc.h"
 | |
| 
 | |
| #undef pr_fmt
 | |
| #define pr_fmt(fmt)	"qtnfmac shm_ipc: %s: " fmt, __func__
 | |
| 
 | |
| static bool qtnf_shm_ipc_has_new_data(struct qtnf_shm_ipc *ipc)
 | |
| {
 | |
| 	const u32 flags = readl(&ipc->shm_region->headroom.hdr.flags);
 | |
| 
 | |
| 	return (flags & QTNF_SHM_IPC_NEW_DATA);
 | |
| }
 | |
| 
 | |
| static void qtnf_shm_handle_new_data(struct qtnf_shm_ipc *ipc)
 | |
| {
 | |
| 	size_t size;
 | |
| 	bool rx_buff_ok = true;
 | |
| 	struct qtnf_shm_ipc_region_header __iomem *shm_reg_hdr;
 | |
| 
 | |
| 	shm_reg_hdr = &ipc->shm_region->headroom.hdr;
 | |
| 
 | |
| 	size = readw(&shm_reg_hdr->data_len);
 | |
| 
 | |
| 	if (unlikely(size == 0 || size > QTN_IPC_MAX_DATA_SZ)) {
 | |
| 		pr_err("wrong rx packet size: %zu\n", size);
 | |
| 		rx_buff_ok = false;
 | |
| 	}
 | |
| 
 | |
| 	if (likely(rx_buff_ok)) {
 | |
| 		ipc->rx_packet_count++;
 | |
| 		ipc->rx_callback.fn(ipc->rx_callback.arg,
 | |
| 				    ipc->shm_region->data, size);
 | |
| 	}
 | |
| 
 | |
| 	writel(QTNF_SHM_IPC_ACK, &shm_reg_hdr->flags);
 | |
| 	readl(&shm_reg_hdr->flags); /* flush PCIe write */
 | |
| 
 | |
| 	ipc->interrupt.fn(ipc->interrupt.arg);
 | |
| }
 | |
| 
 | |
| static void qtnf_shm_ipc_irq_work(struct work_struct *work)
 | |
| {
 | |
| 	struct qtnf_shm_ipc *ipc = container_of(work, struct qtnf_shm_ipc,
 | |
| 						irq_work);
 | |
| 
 | |
| 	while (qtnf_shm_ipc_has_new_data(ipc))
 | |
| 		qtnf_shm_handle_new_data(ipc);
 | |
| }
 | |
| 
 | |
| static void qtnf_shm_ipc_irq_inbound_handler(struct qtnf_shm_ipc *ipc)
 | |
| {
 | |
| 	u32 flags;
 | |
| 
 | |
| 	flags = readl(&ipc->shm_region->headroom.hdr.flags);
 | |
| 
 | |
| 	if (flags & QTNF_SHM_IPC_NEW_DATA)
 | |
| 		queue_work(ipc->workqueue, &ipc->irq_work);
 | |
| }
 | |
| 
 | |
| static void qtnf_shm_ipc_irq_outbound_handler(struct qtnf_shm_ipc *ipc)
 | |
| {
 | |
| 	u32 flags;
 | |
| 
 | |
| 	if (!READ_ONCE(ipc->waiting_for_ack))
 | |
| 		return;
 | |
| 
 | |
| 	flags = readl(&ipc->shm_region->headroom.hdr.flags);
 | |
| 
 | |
| 	if (flags & QTNF_SHM_IPC_ACK) {
 | |
| 		WRITE_ONCE(ipc->waiting_for_ack, 0);
 | |
| 		complete(&ipc->tx_completion);
 | |
| 	}
 | |
| }
 | |
| 
 | |
| int qtnf_shm_ipc_init(struct qtnf_shm_ipc *ipc,
 | |
| 		      enum qtnf_shm_ipc_direction direction,
 | |
| 		      struct qtnf_shm_ipc_region __iomem *shm_region,
 | |
| 		      struct workqueue_struct *workqueue,
 | |
| 		      const struct qtnf_shm_ipc_int *interrupt,
 | |
| 		      const struct qtnf_shm_ipc_rx_callback *rx_callback)
 | |
| {
 | |
| 	BUILD_BUG_ON(offsetof(struct qtnf_shm_ipc_region, data) !=
 | |
| 		     QTN_IPC_REG_HDR_SZ);
 | |
| 	BUILD_BUG_ON(sizeof(struct qtnf_shm_ipc_region) > QTN_IPC_REG_SZ);
 | |
| 
 | |
| 	ipc->shm_region = shm_region;
 | |
| 	ipc->direction = direction;
 | |
| 	ipc->interrupt = *interrupt;
 | |
| 	ipc->rx_callback = *rx_callback;
 | |
| 	ipc->tx_packet_count = 0;
 | |
| 	ipc->rx_packet_count = 0;
 | |
| 	ipc->workqueue = workqueue;
 | |
| 	ipc->waiting_for_ack = 0;
 | |
| 	ipc->tx_timeout_count = 0;
 | |
| 
 | |
| 	switch (direction) {
 | |
| 	case QTNF_SHM_IPC_OUTBOUND:
 | |
| 		ipc->irq_handler = qtnf_shm_ipc_irq_outbound_handler;
 | |
| 		break;
 | |
| 	case QTNF_SHM_IPC_INBOUND:
 | |
| 		ipc->irq_handler = qtnf_shm_ipc_irq_inbound_handler;
 | |
| 		break;
 | |
| 	default:
 | |
| 		return -EINVAL;
 | |
| 	}
 | |
| 
 | |
| 	INIT_WORK(&ipc->irq_work, qtnf_shm_ipc_irq_work);
 | |
| 	init_completion(&ipc->tx_completion);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| void qtnf_shm_ipc_free(struct qtnf_shm_ipc *ipc)
 | |
| {
 | |
| 	complete_all(&ipc->tx_completion);
 | |
| }
 | |
| 
 | |
| int qtnf_shm_ipc_send(struct qtnf_shm_ipc *ipc, const u8 *buf, size_t size)
 | |
| {
 | |
| 	int ret = 0;
 | |
| 	struct qtnf_shm_ipc_region_header __iomem *shm_reg_hdr;
 | |
| 
 | |
| 	shm_reg_hdr = &ipc->shm_region->headroom.hdr;
 | |
| 
 | |
| 	if (unlikely(size > QTN_IPC_MAX_DATA_SZ))
 | |
| 		return -E2BIG;
 | |
| 
 | |
| 	ipc->tx_packet_count++;
 | |
| 
 | |
| 	writew(size, &shm_reg_hdr->data_len);
 | |
| 	memcpy_toio(ipc->shm_region->data, buf, size);
 | |
| 
 | |
| 	/* sync previous writes before proceeding */
 | |
| 	dma_wmb();
 | |
| 
 | |
| 	WRITE_ONCE(ipc->waiting_for_ack, 1);
 | |
| 
 | |
| 	/* sync previous memory write before announcing new data ready */
 | |
| 	wmb();
 | |
| 
 | |
| 	writel(QTNF_SHM_IPC_NEW_DATA, &shm_reg_hdr->flags);
 | |
| 	readl(&shm_reg_hdr->flags); /* flush PCIe write */
 | |
| 
 | |
| 	ipc->interrupt.fn(ipc->interrupt.arg);
 | |
| 
 | |
| 	if (!wait_for_completion_timeout(&ipc->tx_completion,
 | |
| 					 QTN_SHM_IPC_ACK_TIMEOUT)) {
 | |
| 		ret = -ETIMEDOUT;
 | |
| 		ipc->tx_timeout_count++;
 | |
| 		pr_err("TX ACK timeout\n");
 | |
| 	}
 | |
| 
 | |
| 	/* now we're not waiting for ACK even in case of timeout */
 | |
| 	WRITE_ONCE(ipc->waiting_for_ack, 0);
 | |
| 
 | |
| 	return ret;
 | |
| }
 |