From 98877ed5fe34e65878cdbc716f602e4bcef47336 Mon Sep 17 00:00:00 2001 From: Michal Wilczynski Date: Thu, 22 Aug 2024 10:21:27 +0200 Subject: serial: pxa: Add SpacemiT K1-X serial driver This driver seems to be required for the bash shell to appear, without it the shell never works. Port it from the vendor kernel [1]. [1] - https://github.com/BPI-SINOVOIP/pi-linux.git Change-Id: Ic12731d4ccb4a72daeaee582476ab04962556452 Signed-off-by: Michal Wilczynski --- drivers/tty/serial/Kconfig | 9 +- drivers/tty/serial/Makefile | 1 + drivers/tty/serial/pxa_k1x.c | 2351 ++++++++++++++++++++++++++++++++++++++++++ 3 files changed, 2359 insertions(+), 2 deletions(-) create mode 100644 drivers/tty/serial/pxa_k1x.c diff --git a/drivers/tty/serial/Kconfig b/drivers/tty/serial/Kconfig index bdc568a4ab66..954ffb8f29bb 100644 --- a/drivers/tty/serial/Kconfig +++ b/drivers/tty/serial/Kconfig @@ -408,10 +408,9 @@ config SERIAL_21285_CONSOLE config SERIAL_PXA bool "PXA serial port support (DEPRECATED)" - depends on ARCH_PXA || ARCH_MMP + depends on ARCH_PXA || ARCH_MMP || ARCH_SPACEMIT_K1X select SERIAL_CORE select SERIAL_8250_PXA if SERIAL_8250=y - select SERIAL_PXA_NON8250 if !SERIAL_8250=y help If you have a machine based on an Intel XScale PXA2xx CPU you can enable its onboard serial ports by enabling this option. @@ -423,6 +422,12 @@ config SERIAL_PXA_NON8250 bool depends on !SERIAL_8250 || COMPILE_TEST +config SERIAL_PXA_SPACEMIT_K1X + bool "Spacemit PXA driver suppor" + depends on !SERIAL_8250 || COMPILE_TEST + help + Spacemit PXA serial driver + config SERIAL_PXA_CONSOLE bool "Console on PXA serial port (DEPRECATED)" depends on SERIAL_PXA diff --git a/drivers/tty/serial/Makefile b/drivers/tty/serial/Makefile index 138abbc89738..ac89487c6152 100644 --- a/drivers/tty/serial/Makefile +++ b/drivers/tty/serial/Makefile @@ -30,6 +30,7 @@ obj-$(CONFIG_SERIAL_CLPS711X) += clps711x.o obj-$(CONFIG_SERIAL_PXA_NON8250) += pxa.o obj-$(CONFIG_SERIAL_SA1100) += sa1100.o obj-$(CONFIG_SERIAL_BCM63XX) += bcm63xx_uart.o +obj-$(CONFIG_SERIAL_PXA_SPACEMIT_K1X) += pxa_k1x.o obj-$(CONFIG_SERIAL_SAMSUNG) += samsung_tty.o obj-$(CONFIG_SERIAL_MAX3100) += max3100.o obj-$(CONFIG_SERIAL_MAX310X) += max310x.o diff --git a/drivers/tty/serial/pxa_k1x.c b/drivers/tty/serial/pxa_k1x.c new file mode 100644 index 000000000000..767fe217c6a3 --- /dev/null +++ b/drivers/tty/serial/pxa_k1x.c @@ -0,0 +1,2351 @@ +// SPDX-License-Identifier: GPL-2.0+ +/* + * Based on drivers/serial/8250.c by Russell King. + * + * Author: Nicolas Pitre + * Created: Feb 20, 2003 + * Copyright: (C) 2003 Monta Vista Software, Inc. + * + * Note 1: This driver is made separate from the already too overloaded + * 8250.c because it needs some kirks of its own and that'll make it + * easier to add DMA support. + * + * Note 2: I'm too sick of device allocation policies for serial ports. + * If someone else wants to request an "official" allocation of major/minor + * for this driver please be my guest. And don't forget that new hardware + * to come from Intel might have more than 3 or 4 of those UARTs. Let's + * hope for a better port registration and dynamic device allocation scheme + * with the serial core maintainer satisfaction to appear soon. + */ + + +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include + +#define DMA_BLOCK UART_XMIT_SIZE +#define DMA_BURST_SIZE (8) +#define DMA_FIFO_THRESHOLD (32) +#define DMA_RX_BLOCK_SIZE DMA_BLOCK +#define DMA_BUF_POLLING_SWITCH (1) + +#define PXA_UART_TX (0) +#define PXA_UART_RX (1) + +#define UARTCLK_FPGA (14750000) + +#define NUM_UART_PORTS (10) +#define BT_UART_PORT (2) + +#define UART_FCR_PXA_BUS32 (0x20) /* 32-Bit Peripheral Bus */ +#define UART_FCR_PXA_TRAIL (0x10) /* Trailing Bytes */ +#define UART_FOR (9) /* receive FIFO count register */ + +#define PXA_NAME_LEN (8) + +#define SUPPORT_POWER_QOS (1) + +#define TX_DMA_RUNNING (1 << 0) +#define RX_DMA_RUNNING (1 << 1) + +#define PXA_TIMER_TIMEOUT (3*HZ) +#define BLOCK_SUSPEND_TIMEOUT (3000) + + +/* + * DMA related data is stored in this struct, + * making it separated from non-DMA mode. + */ +struct uart_pxa_dma { + unsigned int dma_status; + struct dma_chan *txdma_chan; + struct dma_chan *rxdma_chan; + struct dma_async_tx_descriptor *rx_desc; + struct dma_async_tx_descriptor *tx_desc; + void *txdma_addr; + void *rxdma_addr; + dma_addr_t txdma_addr_phys; + dma_addr_t rxdma_addr_phys; + int tx_stop; + int rx_stop; + dma_cookie_t rx_cookie; + dma_cookie_t tx_cookie; + int tx_size; /* size of last transmit bytes */ + struct tasklet_struct tklet; + +#ifdef CONFIG_PM + /* We needn't save rx dma register because we + * just restart the dma totallly after resume + */ + void *tx_buf_save; + int tx_saved_len; +#endif + + bool dma_init; + +#if (DMA_BUF_POLLING_SWITCH == 1) + int dma_poll_timeout; + int dma_poll_max_time; +#endif +}; + +struct uart_pxa_port { + struct uart_port port; + unsigned char ier; + unsigned char lcr; + unsigned char mcr; + unsigned int lsr_break_flag; + struct clk *fclk; + struct clk *gclk; + struct reset_control *resets; + char name[PXA_NAME_LEN]; + + struct timer_list pxa_timer; + int edge_wakeup_gpio; + struct work_struct uart_tx_lpm_work; + int dma_enable; + struct uart_pxa_dma uart_dma; + unsigned long flags; + unsigned int cons_udelay; /* us */ + bool from_resume; + bool device_ctrl_rts; + bool in_resume; + unsigned int current_baud; + unsigned int clk_fpga; +}; + +static void pxa_uart_transmit_dma_cb(void *data); +static void pxa_uart_receive_dma_cb(void *data); +static void pxa_uart_transmit_dma_start(struct uart_pxa_port *up, int count); +static void pxa_uart_receive_dma_start(struct uart_pxa_port *up); +static inline void wait_for_xmitr(struct uart_pxa_port *up); +static unsigned int serial_pxa_tx_empty(struct uart_port *port); +#ifdef CONFIG_PM +static void _pxa_timer_handler(struct uart_pxa_port *up); +#endif + +static inline void stop_dma(struct uart_pxa_port *up, int read) +{ + unsigned long flags; + struct uart_pxa_dma *pxa_dma = &up->uart_dma; + struct dma_chan *channel; + + if (!pxa_dma->dma_init) { + return; + } + + channel = read ? pxa_dma->rxdma_chan : pxa_dma->txdma_chan; + + dmaengine_terminate_all(channel); + spin_lock_irqsave(&up->port.lock, flags); + if (read) { + pxa_dma->dma_status &= ~RX_DMA_RUNNING; + } else { + pxa_dma->dma_status &= ~TX_DMA_RUNNING; + } + spin_unlock_irqrestore(&up->port.lock, flags); +} + +static inline unsigned int serial_in(struct uart_pxa_port *up, int offset) +{ + offset <<= 2; + return readl(up->port.membase + offset); +} + +static inline void serial_out(struct uart_pxa_port *up, int offset, int value) +{ + offset <<= 2; + writel(value, up->port.membase + offset); +} + +static void serial_pxa_enable_ms(struct uart_port *port) +{ + struct uart_pxa_port *up = (struct uart_pxa_port *)port; + + if (up->dma_enable) { + return; + } + + up->ier |= UART_IER_MSI; + serial_out(up, UART_IER, up->ier); +} + +static void serial_pxa_stop_tx(struct uart_port *port) +{ + struct uart_pxa_port *up = (struct uart_pxa_port *)port; + unsigned int timeout = 0x100000 / up->cons_udelay; + + if (up->dma_enable) { + up->uart_dma.tx_stop = 1; + + if (up->ier & UART_IER_DMAE && up->uart_dma.dma_init) { + /* + * Here we cannot use dma_status to determine + * whether dma has been transfer completed + * As there when this function is being caled, + * it would hold a spinlock and possible with irqsave, + * If this function is being called over core0, + * we may cannot get status change from the flag, + * As irq handler is being blocked to be called yet. + */ + while (dma_async_is_tx_complete(up->uart_dma.txdma_chan, + up->uart_dma.tx_cookie, NULL, NULL) + != DMA_COMPLETE && (timeout-- > 0)) + udelay(up->cons_udelay); + + BUG_ON(timeout == 0); + } + } else { + if (up->ier & UART_IER_THRI) { + up->ier &= ~UART_IER_THRI; + serial_out(up, UART_IER, up->ier); + } + } +} + +static void serial_pxa_stop_rx(struct uart_port *port) +{ + struct uart_pxa_port *up = (struct uart_pxa_port *)port; + + if (up->dma_enable) { + if (up->ier & UART_IER_DMAE) { + spin_unlock_irqrestore(&up->port.lock, up->flags); + stop_dma(up, PXA_UART_RX); + spin_lock_irqsave(&up->port.lock, up->flags); + } + up->uart_dma.rx_stop = 1; + } else { + up->ier &= ~UART_IER_RLSI; + up->port.read_status_mask &= ~UART_LSR_DR; + serial_out(up, UART_IER, up->ier); + } +} + +static inline void receive_chars(struct uart_pxa_port *up, int *status) +{ + unsigned int ch, flag; + int max_count = 256; + + do { + /* work around Errata #20 according to + * Intel(R) PXA27x Processor Family + * Specification Update (May 2005) + * + * Step 2 + * Disable the Reciever Time Out Interrupt via IER[RTOEI] + */ + spin_lock_irqsave(&up->port.lock, up->flags); + up->ier &= ~UART_IER_RTOIE; + serial_out(up, UART_IER, up->ier); + spin_unlock_irqrestore(&up->port.lock, up->flags); + + ch = serial_in(up, UART_RX); + flag = TTY_NORMAL; + up->port.icount.rx++; + + if (unlikely(*status & (UART_LSR_BI | UART_LSR_PE | + UART_LSR_FE | UART_LSR_OE))) { + /* + * For statistics only + */ + if (*status & UART_LSR_BI) { + *status &= ~(UART_LSR_FE | UART_LSR_PE); + up->port.icount.brk++; + /* + * We do the SysRQ and SAK checking + * here because otherwise the break + * may get masked by ignore_status_mask + * or read_status_mask. + */ + if (uart_handle_break(&up->port)) + goto ignore_char; + } else if (*status & UART_LSR_PE) { + up->port.icount.parity++; + } else if (*status & UART_LSR_FE) { + up->port.icount.frame++; + } + + if (*status & UART_LSR_OE) { + up->port.icount.overrun++; + } + + /* + * Mask off conditions which should be ignored. + */ + *status &= up->port.read_status_mask; + +#ifdef CONFIG_SERIAL_PXA_CONSOLE + if (up->port.line == up->port.cons->index) { + /* Recover the break flag from console xmit */ + *status |= up->lsr_break_flag; + up->lsr_break_flag = 0; + } +#endif + if (*status & UART_LSR_BI) { + flag = TTY_BREAK; + } else if (*status & UART_LSR_PE) { + flag = TTY_PARITY; + } else if (*status & UART_LSR_FE) { + flag = TTY_FRAME; + } + } + if (!uart_handle_sysrq_char(&up->port, ch)) + uart_insert_char(&up->port, *status, UART_LSR_OE, ch, flag); + + ignore_char: + *status = serial_in(up, UART_LSR); + } while ((*status & UART_LSR_DR) && (max_count-- > 0)); + tty_flip_buffer_push(&up->port.state->port); + + /* work around Errata #20 according to + * Intel(R) PXA27x Processor Family + * Specification Update (May 2005) + * + * Step 6: + * No more data in FIFO: Re-enable RTO interrupt via IER[RTOIE] + */ + spin_lock_irqsave(&up->port.lock, up->flags); + up->ier |= UART_IER_RTOIE; + serial_out(up, UART_IER, up->ier); + spin_unlock_irqrestore(&up->port.lock, up->flags); +} + +static void transmit_chars(struct uart_pxa_port *up) +{ + struct circ_buf *xmit = &up->port.state->xmit; + int count; + + if (up->port.x_char) { + serial_out(up, UART_TX, up->port.x_char); + up->port.icount.tx++; + up->port.x_char = 0; + return; + } + if (uart_circ_empty(xmit) || uart_tx_stopped(&up->port)) { + spin_lock_irqsave(&up->port.lock, up->flags); + serial_pxa_stop_tx(&up->port); + spin_unlock_irqrestore(&up->port.lock, up->flags); + return; + } + + count = up->port.fifosize / 2; + do { + serial_out(up, UART_TX, xmit->buf[xmit->tail]); + xmit->tail = (xmit->tail + 1) & (UART_XMIT_SIZE - 1); + up->port.icount.tx++; + if (uart_circ_empty(xmit)) + break; + } while (--count > 0); + + if (uart_circ_chars_pending(xmit) < WAKEUP_CHARS) + uart_write_wakeup(&up->port); + + + if (uart_circ_empty(xmit)) + { + spin_lock_irqsave(&up->port.lock, up->flags); + serial_pxa_stop_tx(&up->port); + spin_unlock_irqrestore(&up->port.lock, up->flags); + } +} + +static inline void dma_receive_chars(struct uart_pxa_port *up, int *status) +{ + struct tty_port *port = &up->port.state->port; + unsigned char ch; + int max_count = 256; + int count = 0; + unsigned char *tmp; + unsigned int flag = TTY_NORMAL; + struct uart_pxa_dma *pxa_dma = &up->uart_dma; + struct dma_tx_state dma_state; + + if (!pxa_dma->dma_init) + return; + + /* + * Pause DMA channel and deal with the bytes received by DMA + */ + dmaengine_pause(pxa_dma->rxdma_chan); + dmaengine_tx_status(pxa_dma->rxdma_chan, pxa_dma->rx_cookie, + &dma_state); + count = DMA_RX_BLOCK_SIZE - dma_state.residue; + tmp = pxa_dma->rxdma_addr; + if (up->port.sysrq) { + while (count > 0) { + if (!uart_handle_sysrq_char(&up->port, *tmp)) { + uart_insert_char(&up->port, *status, 0, *tmp, flag); + up->port.icount.rx++; + } + tmp++; + count--; + } + } else { + tty_insert_flip_string(port, tmp, count); + up->port.icount.rx += count; + } + + /* deal with the bytes in rx FIFO */ + do { + ch = serial_in(up, UART_RX); + flag = TTY_NORMAL; + up->port.icount.rx++; + + if (unlikely(*status & (UART_LSR_BI | UART_LSR_PE | + UART_LSR_FE | UART_LSR_OE))) { + /* + * For statistics only + */ + if (*status & UART_LSR_BI) { + *status &= ~(UART_LSR_FE | UART_LSR_PE); + up->port.icount.brk++; + /* + * We do the SysRQ and SAK checking + * here because otherwise the break + * may get masked by ignore_status_mask + * or read_status_mask. + */ + if (uart_handle_break(&up->port)) + goto ignore_char2; + } else if (*status & UART_LSR_PE) { + up->port.icount.parity++; + } else if (*status & UART_LSR_FE) { + up->port.icount.frame++; + } + + if (*status & UART_LSR_OE) { + up->port.icount.overrun++; + } + + /* + * Mask off conditions which should be ignored. + */ + *status &= up->port.read_status_mask; + +#ifdef CONFIG_SERIAL_PXA_CONSOLE + if (up->port.line == up->port.cons->index) { + /* Recover the break flag from console xmit */ + *status |= up->lsr_break_flag; + up->lsr_break_flag = 0; + } +#endif /* #ifdef CONFIG_SERIAL_PXA_CONSOLE */ + if (*status & UART_LSR_BI) { + flag = TTY_BREAK; + } else if (*status & UART_LSR_PE) { + flag = TTY_PARITY; + } else if (*status & UART_LSR_FE) { + flag = TTY_FRAME; + } + } + if (!uart_handle_sysrq_char(&up->port, ch)) + uart_insert_char(&up->port, *status, UART_LSR_OE, + ch, flag); + ignore_char2: + *status = serial_in(up, UART_LSR); + } while ((*status & UART_LSR_DR) && (max_count-- > 0)); + + tty_flip_buffer_push(port); + stop_dma(up, 1); + if (pxa_dma->rx_stop) + return; + pxa_uart_receive_dma_start(up); +} + +static void serial_pxa_start_tx(struct uart_port *port) +{ + struct uart_pxa_port *up = (struct uart_pxa_port *)port; + + if (up->dma_enable) { + up->uart_dma.tx_stop = 0; + tasklet_schedule(&up->uart_dma.tklet); + } else { + if (!(up->ier & UART_IER_THRI)) { + up->ier |= UART_IER_THRI; + serial_out(up, UART_IER, up->ier); + } + } +} + +/* should hold up->port.lock */ +static inline void check_modem_status(struct uart_pxa_port *up) +{ + int status; + + status = serial_in(up, UART_MSR); + + if ((status & UART_MSR_ANY_DELTA) == 0) + return; + + if (status & UART_MSR_TERI) + up->port.icount.rng++; + if (status & UART_MSR_DDSR) + up->port.icount.dsr++; + if (status & UART_MSR_DDCD) + uart_handle_dcd_change(&up->port, status & UART_MSR_DCD); + if (status & UART_MSR_DCTS) + uart_handle_cts_change(&up->port, status & UART_MSR_CTS); + + wake_up_interruptible(&up->port.state->port.delta_msr_wait); +} + +static int serial_pxa_is_open(struct uart_pxa_port *up); + +/* + * This handles the interrupt from one port. + */ +static inline irqreturn_t serial_pxa_irq(int irq, void *dev_id) +{ + struct uart_pxa_port *up = dev_id; + unsigned int iir, lsr; + + /* + * If FCR[4] is set, we may receive EOC interrupt when: + * 1) current descritor of DMA finishes successfully; + * 2) there are still trailing bytes in UART FIFO. + * This interrupt alway comes along with UART_IIR_NO_INT. + * So this interrupt is just ignored by us. + */ + iir = serial_in(up, UART_IIR); + if (iir & UART_IIR_NO_INT) + return IRQ_NONE; + + /* in case the clock is disabled */ + if (!serial_pxa_is_open(up)) + return IRQ_HANDLED; + +#ifdef CONFIG_PM +#if SUPPORT_POWER_QOS + /* timer is not active */ + if (!mod_timer(&up->pxa_timer, jiffies + PXA_TIMER_TIMEOUT)) + pm_runtime_get_sync(up->port.dev); +#endif +#endif + + lsr = serial_in(up, UART_LSR); + if (up->dma_enable) { + /* we only need to deal with FIFOE here */ + if (lsr & UART_LSR_FIFOE) + dma_receive_chars(up, &lsr); + } else { + if (lsr & UART_LSR_DR) { + receive_chars(up, &lsr); + /* Avoid suspend within 3 seconds. */ + if (up->edge_wakeup_gpio >= 0) + pm_wakeup_event(up->port.dev, BLOCK_SUSPEND_TIMEOUT); + } + + check_modem_status(up); + if (lsr & UART_LSR_THRE) { + transmit_chars(up); + /* wait Tx empty */ + while (!serial_pxa_tx_empty((struct uart_port *)dev_id)) + ; + } + } + + return IRQ_HANDLED; +} + +static unsigned int serial_pxa_tx_empty(struct uart_port *port) +{ + struct uart_pxa_port *up = (struct uart_pxa_port *)port; + unsigned long flags; + unsigned int ret; + + spin_lock_irqsave(&up->port.lock, flags); + if (up->dma_enable) { + if (up->ier & UART_IER_DMAE) { + if (up->uart_dma.dma_status & TX_DMA_RUNNING) { + spin_unlock_irqrestore(&up->port.lock, flags); + return 0; + } + } + } + ret = serial_in(up, UART_LSR) & UART_LSR_TEMT ? TIOCSER_TEMT : 0; + spin_unlock_irqrestore(&up->port.lock, flags); + + return ret; +} + +static unsigned int serial_pxa_get_mctrl(struct uart_port *port) +{ + struct uart_pxa_port *up = (struct uart_pxa_port *)port; + unsigned char status; + unsigned int ret; + + status = serial_in(up, UART_MSR); + + ret = 0; + if (status & UART_MSR_DCD) + ret |= TIOCM_CAR; + if (status & UART_MSR_RI) + ret |= TIOCM_RNG; + if (status & UART_MSR_DSR) + ret |= TIOCM_DSR; + if (status & UART_MSR_CTS) + ret |= TIOCM_CTS; + return ret; +} + +#if 0 +extern int bluesleep_hostwake_is_active(int port); +#endif + +static void serial_pxa_set_mctrl(struct uart_port *port, unsigned int mctrl) +{ + struct uart_pxa_port *up = (struct uart_pxa_port *)port; + unsigned char mcr = 0; + int hostwake = 0; + + if (mctrl & TIOCM_RTS) + mcr |= UART_MCR_RTS; + if (mctrl & TIOCM_DTR) + mcr |= UART_MCR_DTR; + if (mctrl & TIOCM_OUT1) + mcr |= UART_MCR_OUT1; + if (mctrl & TIOCM_OUT2) + mcr |= UART_MCR_OUT2; + if (mctrl & TIOCM_LOOP) + mcr |= UART_MCR_LOOP; + + /* + * If the RTS should be controlled conditionally, such as UART2 RTS + * should be controlled by BT HOST_WAKE handler, here we should + * block the serial_core to assert RTS if BT HOST_WAKE is disactive. + */ + if (up->device_ctrl_rts) { + #if 0 + hostwake = (bluesleep_hostwake_is_active(BT_UART_PORT) == 0); + #endif + if ((hostwake || up->in_resume) && (mctrl & TIOCM_RTS)) + mcr &= ~UART_MCR_RTS; + } + + mcr |= up->mcr; + + serial_out(up, UART_MCR, mcr); + +#ifdef CONFIG_BT + if (up->port.line == BT_UART_PORT) + pr_info("%s: rts: 0x%x\n", __func__, mcr & UART_MCR_RTS); +#endif +} + +static void serial_pxa_break_ctl(struct uart_port *port, int break_state) +{ + struct uart_pxa_port *up = (struct uart_pxa_port *)port; + unsigned long flags; + + spin_lock_irqsave(&up->port.lock, flags); + if (break_state == -1) + up->lcr |= UART_LCR_SBC; + else + up->lcr &= ~UART_LCR_SBC; + serial_out(up, UART_LCR, up->lcr); + spin_unlock_irqrestore(&up->port.lock, flags); +} + +static void pxa_uart_transmit_dma_start(struct uart_pxa_port *up, int count) +{ + struct uart_pxa_dma *pxa_dma = &up->uart_dma; + struct dma_slave_config slave_config; + int ret; + + if (!pxa_dma->txdma_chan) { + dev_err(up->port.dev, "tx dma channel is not initialized\n"); + return; + } + + slave_config.direction = DMA_MEM_TO_DEV; + slave_config.dst_addr = up->port.mapbase; + slave_config.dst_maxburst = DMA_BURST_SIZE; + slave_config.dst_addr_width = DMA_SLAVE_BUSWIDTH_1_BYTE; + + ret = dmaengine_slave_config(pxa_dma->txdma_chan, &slave_config); + if (ret) { + dev_err(up->port.dev, + "%s: dmaengine slave config err.\n", __func__); + return; + } + + pxa_dma->tx_size = count; + pxa_dma->tx_desc = dmaengine_prep_slave_single(pxa_dma->txdma_chan, + pxa_dma->txdma_addr_phys, count, DMA_MEM_TO_DEV, 0); + if(pxa_dma->tx_desc == NULL) { + dev_err(up->port.dev, + "%s: Unable to get desc for Tx\n", __func__); + return; + } + pxa_dma->tx_desc->callback = pxa_uart_transmit_dma_cb; + pxa_dma->tx_desc->callback_param = up; + + pxa_dma->tx_cookie = dmaengine_submit(pxa_dma->tx_desc); +#ifdef CONFIG_PM +#if SUPPORT_POWER_QOS + pm_runtime_get_sync(up->port.dev); +#endif +#endif + + dma_async_issue_pending(pxa_dma->txdma_chan); +} + +static void pxa_uart_receive_dma_start(struct uart_pxa_port *up) +{ + unsigned long flags; + struct uart_pxa_dma *uart_dma = &up->uart_dma; + struct dma_slave_config slave_config; + int ret; + + if (!uart_dma->rxdma_chan) { + dev_err(up->port.dev, "rx dma channel is not initialized\n"); + return; + } + + spin_lock_irqsave(&up->port.lock, flags); + if (uart_dma->dma_status & RX_DMA_RUNNING) { + spin_unlock_irqrestore(&up->port.lock, flags); + return; + } + uart_dma->dma_status |= RX_DMA_RUNNING; + spin_unlock_irqrestore(&up->port.lock, flags); + + slave_config.direction = DMA_DEV_TO_MEM; + slave_config.src_addr = up->port.mapbase; + slave_config.src_maxburst = DMA_BURST_SIZE; + slave_config.src_addr_width = DMA_SLAVE_BUSWIDTH_1_BYTE; + + ret = dmaengine_slave_config(uart_dma->rxdma_chan, &slave_config); + if (ret) { + dev_err(up->port.dev, "%s: dmaengine slave config err.\n", __func__); + return; + } + + uart_dma->rx_desc = dmaengine_prep_slave_single(uart_dma->rxdma_chan, + uart_dma->rxdma_addr_phys, DMA_RX_BLOCK_SIZE, DMA_DEV_TO_MEM, 0); + if(uart_dma->rx_desc == NULL) { + dev_err(up->port.dev, "%s: Unable to get desc for Rx\n", __func__); + return; + } + uart_dma->rx_desc->callback = pxa_uart_receive_dma_cb; + uart_dma->rx_desc->callback_param = up; + + uart_dma->rx_cookie = dmaengine_submit(uart_dma->rx_desc); + dma_async_issue_pending(uart_dma->rxdma_chan); +} + +static void pxa_uart_receive_dma_cb(void *data) +{ + unsigned long flags; + struct uart_pxa_port *up = (struct uart_pxa_port *)data; + struct uart_pxa_dma *pxa_dma = &up->uart_dma; + struct tty_port *port = &up->port.state->port; + unsigned int count; + unsigned char *tmp = pxa_dma->rxdma_addr; + struct dma_tx_state dma_state; +#if (DMA_BUF_POLLING_SWITCH == 1) + unsigned int buf_used, trail_cnt = 0; + unsigned char *trail_addr, *last_burst_addr; + u8 mark_1bytes = 0xff; + int timeout = 5000, cycle = 2; /* us */ + int times_1 = 0, times_2 = 0, duration_time_us; +#endif + +#ifdef CONFIG_PM +#if SUPPORT_POWER_QOS + if (!mod_timer(&up->pxa_timer, jiffies + PXA_TIMER_TIMEOUT)) + pm_runtime_get_sync(up->port.dev); +#endif +#endif + + dmaengine_tx_status(pxa_dma->rxdma_chan, pxa_dma->rx_cookie, &dma_state); + count = DMA_RX_BLOCK_SIZE - dma_state.residue; + + /* + * Workaround for DMA memory update delay: + * When receiving the trailing bytes (current threshold is 32 bytes), + * if the first byte is still the marked value 0xff and not be updated + * by DMA, we will polling until it done, in case the received real + * data is 0xff, we will timeout after 10ms and get a warning info. + */ +#if (DMA_BUF_POLLING_SWITCH == 1) + buf_used = count; + /* DMA trailing bytes receive */ + if ((count > 0) && (count < DMA_FIFO_THRESHOLD)) { + trail_cnt = count; + trail_addr = tmp; + times_1 = timeout / cycle; + times_2 = timeout / cycle; + + while ((*trail_addr == mark_1bytes) && (times_1-- >= 0)) { + udelay(cycle); + } + + if (trail_cnt > 1) { + trail_addr = trail_addr + trail_cnt - 1; + while ((*trail_addr == mark_1bytes) && (times_2-- >= 0)) { + udelay(cycle); + } + } + + if ((times_1 <= 0) || (times_2 <= 0)) { + pxa_dma->dma_poll_timeout++; + } + } + + /* DMA burst receive + DMA trailing bytes receive */ + if ((count >= DMA_FIFO_THRESHOLD) && (count < DMA_RX_BLOCK_SIZE)) { + /* + * calculate out the number of trailing bytes: + * count = burst size * n + trailling bytes, + * trailling bytes >= threshold - burst size + * trailling bytes < threshold + */ + trail_cnt = (count % DMA_BURST_SIZE) + (DMA_FIFO_THRESHOLD - DMA_BURST_SIZE); + + /* polling the trailing bytes */ + trail_addr = tmp + count - trail_cnt; + /* + * if the burst size equals to the threshold, and the receive + * bytes is also the integral multiples of burst size, the trailing + * bytes will be zero. + */ + if ((DMA_FIFO_THRESHOLD == DMA_BURST_SIZE) && (trail_cnt == 0)) { + trail_addr = tmp + count - DMA_BURST_SIZE; + trail_cnt = DMA_BURST_SIZE; + } + + times_1 = timeout / cycle; + times_2 = timeout / cycle; + + while ((*trail_addr == mark_1bytes) && (times_1-- >= 0)) { + udelay(cycle); + } + + if (trail_cnt > 1) { + trail_addr = trail_addr + trail_cnt - 1; + while ((*trail_addr == mark_1bytes) && (times_2-- >= 0)) { + udelay(cycle); + } + } + + if ((times_1 <= 0) || (times_2 <= 0)) { + pxa_dma->dma_poll_timeout++; + } + } + + /* DMA burst receive */ + if (count == DMA_RX_BLOCK_SIZE) { + /* polling the last burst bytes */ + last_burst_addr = tmp + DMA_RX_BLOCK_SIZE - DMA_BURST_SIZE; + trail_cnt = DMA_BURST_SIZE; + times_1 = timeout / cycle; + times_2 = timeout / cycle; + + while ((*last_burst_addr == mark_1bytes) && (times_1-- >= 0)) { + udelay(cycle); + } + + if (trail_cnt > 1) { + last_burst_addr = tmp + DMA_RX_BLOCK_SIZE - 1; + while ((*last_burst_addr == mark_1bytes) && (times_2-- >= 0)) { + udelay(cycle); + } + } + + if ((times_1 <= 0) || (times_2 <= 0)) { + pxa_dma->dma_poll_timeout++; + } + } +#endif /* #if (DMA_BUF_POLLING_SWITCH == 1) */ + + if (up->port.sysrq) { + while (count > 0) { + if (!uart_handle_sysrq_char(&up->port, *tmp)) { + tty_insert_flip_char(port, *tmp, TTY_NORMAL); + up->port.icount.rx++; + } + tmp++; + count--; + } + } else { + tty_insert_flip_string(port, tmp, count); + up->port.icount.rx += count; + } + tty_flip_buffer_push(port); + + spin_lock_irqsave(&up->port.lock, flags); + /* + * DMA_RUNNING flag should be clear only after + * all dma interface operation completed + */ + pxa_dma->dma_status &= ~RX_DMA_RUNNING; + spin_unlock_irqrestore(&up->port.lock, flags); + + /* Mark the DMA buf[0] as 0xff for next checking and polling. */ +#if (DMA_BUF_POLLING_SWITCH == 1) + if (buf_used > 0) { + tmp = pxa_dma->rxdma_addr; + /* mark the already overrided buf */ + memset(tmp, mark_1bytes, buf_used); + } + + if (times_1 > 0) { + duration_time_us = (timeout / cycle - times_1) * cycle; + if (pxa_dma->dma_poll_max_time < duration_time_us) { + pxa_dma->dma_poll_max_time = duration_time_us; + } + } + if (times_2 > 0) { + duration_time_us = (timeout / cycle - times_2) * cycle; + if (pxa_dma->dma_poll_max_time < duration_time_us) { + pxa_dma->dma_poll_max_time = duration_time_us; + } + } + if ((times_1 > 0) && (times_2 > 0)) { + duration_time_us = (2 * timeout / cycle - times_1 - times_2) * cycle; + if (pxa_dma->dma_poll_max_time < duration_time_us) { + pxa_dma->dma_poll_max_time = duration_time_us; + } + } +#endif /* #if (DMA_BUF_POLLING_SWITCH == 1) */ + + if (pxa_dma->rx_stop || !serial_pxa_is_open(up)) { + return; + } + pxa_uart_receive_dma_start(up); + /* Avoid suspend within 3 seconds. */ + if (up->edge_wakeup_gpio >= 0) { + pm_wakeup_event(up->port.dev, BLOCK_SUSPEND_TIMEOUT); + } +} + +static void pxa_uart_transmit_dma_cb(void *data) +{ + struct uart_pxa_port *up = (struct uart_pxa_port *)data; + struct uart_pxa_dma *pxa_dma = &up->uart_dma; + struct circ_buf *xmit = &up->port.state->xmit; + + if (up->from_resume) { + up->from_resume = false; + } + + if (dma_async_is_tx_complete(pxa_dma->txdma_chan, pxa_dma->tx_cookie, + NULL, NULL) == DMA_COMPLETE) { + schedule_work(&up->uart_tx_lpm_work); + } + + spin_lock_irqsave(&up->port.lock, up->flags); + /* + * DMA_RUNNING flag should be clear only after + * all dma interface operation completed + */ + pxa_dma->dma_status &= ~TX_DMA_RUNNING; + spin_unlock_irqrestore(&up->port.lock, up->flags); + + /* if tx stop, stop transmit DMA and return */ + if (pxa_dma->tx_stop || !serial_pxa_is_open(up)) { + return; + } + + if (up->port.x_char) { + serial_out(up, UART_TX, up->port.x_char); + up->port.icount.tx++; + up->port.x_char = 0; + } + + if (uart_circ_chars_pending(xmit) < WAKEUP_CHARS) { + uart_write_wakeup(&up->port); + } + + if (!uart_circ_empty(xmit)) { + tasklet_schedule(&pxa_dma->tklet); + } +} + +static void pxa_uart_dma_init(struct uart_pxa_port *up) +{ + struct uart_pxa_dma *pxa_dma = &up->uart_dma; + dma_cap_mask_t mask; + + dma_cap_zero(mask); + dma_cap_set(DMA_SLAVE, mask); + + if (NULL == pxa_dma->rxdma_chan) { + pxa_dma->rxdma_chan = dma_request_slave_channel(up->port.dev, "rx"); + if (IS_ERR_OR_NULL(pxa_dma->rxdma_chan)) { + dev_WARN_ONCE(up->port.dev, 1, "failed to request rx dma channel\n"); + goto out; + } + } + + if (NULL == pxa_dma->txdma_chan) { + pxa_dma->txdma_chan = dma_request_slave_channel(up->port.dev, "tx"); + if (IS_ERR_OR_NULL(pxa_dma->txdma_chan)) { + dev_WARN_ONCE(up->port.dev, 1, "failed to request tx dma channel\n"); + goto err_txdma; + } + } + + if (NULL == pxa_dma->txdma_addr) { + pxa_dma->txdma_addr = dma_direct_alloc(up->port.dev, DMA_BLOCK, + &pxa_dma->txdma_addr_phys, GFP_KERNEL, + DMA_ATTR_FORCE_CONTIGUOUS); + if (!pxa_dma->txdma_addr) { + dev_WARN_ONCE(up->port.dev, 1, "failed to allocate tx dma memory\n"); + goto txdma_err_alloc; + } + } + + if (NULL == pxa_dma->rxdma_addr) { + pxa_dma->rxdma_addr = dma_direct_alloc(up->port.dev, + DMA_RX_BLOCK_SIZE, + &pxa_dma->rxdma_addr_phys, GFP_KERNEL, + DMA_ATTR_FORCE_CONTIGUOUS); + if (!pxa_dma->rxdma_addr) { + dev_WARN_ONCE(up->port.dev, 1, "failed to allocate rx dma memory\n"); + goto rxdma_err_alloc; + } + } + + pxa_dma->dma_status = 0; + pxa_dma->dma_init = true; + return; + +rxdma_err_alloc: + dma_direct_free(up->port.dev, DMA_BLOCK, pxa_dma->txdma_addr, + pxa_dma->txdma_addr_phys, DMA_ATTR_FORCE_CONTIGUOUS); + pxa_dma->txdma_addr = NULL; +txdma_err_alloc: + dma_release_channel(pxa_dma->txdma_chan); + pxa_dma->txdma_chan = NULL; +err_txdma: + dma_release_channel(pxa_dma->rxdma_chan); + pxa_dma->rxdma_chan = NULL; +out: + pxa_dma->dma_init = false; + return; +} + +static void pxa_uart_dma_uninit(struct uart_pxa_port *up) +{ + struct uart_pxa_dma *pxa_dma; + pxa_dma = &up->uart_dma; + + stop_dma(up, PXA_UART_TX); + stop_dma(up, PXA_UART_RX); + + pxa_dma->dma_init = false; + + if (pxa_dma->txdma_chan != NULL) { + dma_release_channel(pxa_dma->txdma_chan); + pxa_dma->txdma_chan = NULL; + } + + if (pxa_dma->txdma_addr != NULL) { + dma_direct_free(up->port.dev, DMA_BLOCK, pxa_dma->txdma_addr, + pxa_dma->txdma_addr_phys, + DMA_ATTR_FORCE_CONTIGUOUS); + pxa_dma->txdma_addr = NULL; + } + + if (pxa_dma->rxdma_chan != NULL) { + dma_release_channel(pxa_dma->rxdma_chan); + pxa_dma->rxdma_chan = NULL; + } + + if (pxa_dma->rxdma_addr != NULL) { + dma_direct_free(up->port.dev, DMA_RX_BLOCK_SIZE, pxa_dma->rxdma_addr, + pxa_dma->rxdma_addr_phys, + DMA_ATTR_FORCE_CONTIGUOUS); + pxa_dma->rxdma_addr = NULL; + } +} + +static void uart_task_action(unsigned long data) +{ + struct uart_pxa_port *up = (struct uart_pxa_port *)data; + struct circ_buf *xmit = &up->port.state->xmit; + unsigned char *tmp = up->uart_dma.txdma_addr; + unsigned long flags; + int count = 0, c; + + /* if the tx is stop or the uart device is suspended, just return. */ + /* if port is shutdown,just return. */ + if (up->uart_dma.tx_stop || up->port.suspended || + !serial_pxa_is_open(up) || up->from_resume) { + return; + } + + spin_lock_irqsave(&up->port.lock, flags); + if (up->uart_dma.dma_status & TX_DMA_RUNNING) { + spin_unlock_irqrestore(&up->port.lock, flags); + return; + } + + up->uart_dma.dma_status |= TX_DMA_RUNNING; + while (1) { + c = CIRC_CNT_TO_END(xmit->head, xmit->tail, UART_XMIT_SIZE); + if (c <= 0) { + break; + } + + memcpy(tmp, xmit->buf + xmit->tail, c); + xmit->tail = (xmit->tail + c) & (UART_XMIT_SIZE - 1); + tmp += c; + count += c; + up->port.icount.tx += c; + } + spin_unlock_irqrestore(&up->port.lock, flags); + + pr_debug("count =%d", count); + pxa_uart_transmit_dma_start(up, count); +} + +static int serial_pxa_startup(struct uart_port *port) +{ + struct uart_pxa_port *up = (struct uart_pxa_port *)port; + unsigned long flags; + int tmp = 0; + + if (port->line == 3) { + up->mcr |= UART_MCR_AFE; + } else { + up->mcr = 0; + } + + #ifdef CONFIG_SOC_SPACEMIT_K1_FPGA + if (up->clk_fpga) { + up->port.uartclk = up->clk_fpga; + } else { + up->port.uartclk = UARTCLK_FPGA; + } + #else + up->port.uartclk = clk_get_rate(up->fclk); + #endif + + /* + * Allocate the IRQ + */ + enable_irq(up->port.irq); + + /* + * Clear the FIFO buffers and disable them. + * (they will be reenabled in set_termios()) + */ + serial_out(up, UART_FCR, UART_FCR_ENABLE_FIFO); + serial_out(up, UART_FCR, UART_FCR_ENABLE_FIFO | + UART_FCR_CLEAR_RCVR | UART_FCR_CLEAR_XMIT); + serial_out(up, UART_FCR, 0); + + /* + * Clear the interrupt registers. + */ + (void) serial_in(up, UART_LSR); + (void) serial_in(up, UART_RX); + (void) serial_in(up, UART_IIR); + (void) serial_in(up, UART_MSR); + + /* + * Now, initialize the UART + */ + serial_out(up, UART_LCR, UART_LCR_WLEN8); + + spin_lock_irqsave(&up->port.lock, flags); + up->port.mctrl |= TIOCM_OUT2; + tmp = serial_in(up, UART_MCR); + tmp |= TIOCM_OUT2; + serial_out(up, UART_MCR, tmp); + spin_unlock_irqrestore(&up->port.lock, flags); + + /* + * Finally, enable interrupts. Note: Modem status interrupts + * are set via set_termios(), which will be occurring imminently + * anyway, so we don't enable them here. + */ + if (up->dma_enable) { + pxa_uart_dma_init(up); + up->uart_dma.rx_stop = 0; + pxa_uart_receive_dma_start(up); + tasklet_init(&up->uart_dma.tklet, uart_task_action, (unsigned long)up); + } + + spin_lock_irqsave(&up->port.lock, flags); + if (up->dma_enable) { + up->ier = UART_IER_DMAE | UART_IER_UUE; + } else { + up->ier = UART_IER_RLSI | UART_IER_RDI | UART_IER_RTOIE | UART_IER_UUE; + } + serial_out(up, UART_IER, up->ier); + spin_unlock_irqrestore(&up->port.lock, flags); + + /* + * And clear the interrupt registers again for luck. + */ + (void) serial_in(up, UART_LSR); + (void) serial_in(up, UART_RX); + (void) serial_in(up, UART_IIR); + (void) serial_in(up, UART_MSR); + + return 0; +} + +static void serial_pxa_shutdown(struct uart_port *port) +{ + struct uart_pxa_port *up = (struct uart_pxa_port *)port; + unsigned long flags; + unsigned int tmp = 0; + + disable_irq(up->port.irq); + if (up->dma_enable) { + tasklet_kill(&up->uart_dma.tklet); + up->uart_dma.tx_stop = 1; + up->uart_dma.rx_stop = 1; + pxa_uart_dma_uninit(up); + } + + flush_work(&up->uart_tx_lpm_work); + + /* + * Disable interrupts from this port + */ + spin_lock_irqsave(&up->port.lock, flags); + up->ier = 0; + serial_out(up, UART_IER, 0); + + up->port.mctrl &= ~TIOCM_OUT2; + tmp = serial_in(up, UART_MCR); + tmp &= ~TIOCM_OUT2; + serial_out(up, UART_MCR, tmp); + spin_unlock_irqrestore(&up->port.lock, flags); + + /* + * Disable break condition and FIFOs + */ + serial_out(up, UART_LCR, serial_in(up, UART_LCR) & ~UART_LCR_SBC); + serial_out(up, UART_FCR, UART_FCR_ENABLE_FIFO | + UART_FCR_CLEAR_RCVR | + UART_FCR_CLEAR_XMIT); + serial_out(up, UART_FCR, 0); +} + +static int pxa_set_baudrate_clk(struct uart_port *port, unsigned int baud) +{ + struct uart_pxa_port *up = (struct uart_pxa_port *)port; + unsigned long rate; + int ret; + + #ifdef CONFIG_SOC_SPACEMIT_K1_FPGA + if (uart_console(port)) { + return 0; + } + #endif + + if (up->current_baud == baud) { + return 0; + } + + switch (baud) { + case 500000: + case 1000000: + case 1500000: + case 3000000: + rate = 48000000; + break; + case 576000: + case 1152000: + case 2500000: + case 4000000: + rate = 73000000; + break; + case 2000000: + case 3500000: + rate = 58000000; + break; + default: + rate = 14700000; + break; + } + + /* If some platforms have no 48M or 73M, the clk + * driver should choose the closest 58M clk source. + * For one target baudrate, the quot is figured out + * by formula [quot] = clk_rate / 16 / baudrate, and + * choose the closest integral value above zero. + * So for different clk source, the real baudrate is + * baudrate = clk_rate / 16 / [quot]. */ + ret = clk_set_rate(up->fclk, rate); + if (ret < 0) { + dev_err(port->dev, + "Failed to set clk rate %lu\n", rate); + return ret; + } + + up->port.uartclk = clk_get_rate(up->fclk); + up->current_baud = baud; + + return 0; +} + +static void +serial_pxa_set_termios(struct uart_port *port, struct ktermios *termios, + const struct ktermios *old) +{ + struct uart_pxa_port *up = (struct uart_pxa_port *)port; + unsigned char cval, fcr = 0; + unsigned long flags; + unsigned int baud, quot; + unsigned int dll; + int ret; + + if (up->dma_enable && up->uart_dma.dma_init) + stop_dma(up, PXA_UART_RX); + + cval = UART_LCR_WLEN(tty_get_char_size(termios->c_cflag)); + + if (termios->c_cflag & CSTOPB) + cval |= UART_LCR_STOP; + if (termios->c_cflag & PARENB) + cval |= UART_LCR_PARITY; + if (!(termios->c_cflag & PARODD)) + cval |= UART_LCR_EPAR; + + /* + * Ask the core to calculate the divisor for us. + */ + baud = uart_get_baud_rate(port, termios, old, 0, 4000000); + if (!baud) + baud = 9600; + ret = pxa_set_baudrate_clk(port, baud); + if (ret < 0) { + dev_err(port->dev, "Failed to set baud rate clk: %d\n", ret); + return; + } + if (tty_termios_baud_rate(termios)) + tty_termios_encode_baud_rate(termios, baud, baud); + + quot = uart_get_divisor(port, baud); + + if (!quot) + quot = 1; + if (up->dma_enable) { + fcr = UART_FCR_ENABLE_FIFO | UART_FCR_PXAR32 | + UART_FCR_PXA_TRAIL; + fcr &= ~UART_FCR_PXA_BUS32; + } else { + if ((up->port.uartclk / quot) < (2400 * 16)) + fcr = UART_FCR_ENABLE_FIFO | UART_FCR_PXAR1; + else if ((up->port.uartclk / quot) < (230400 * 16)) + fcr = UART_FCR_ENABLE_FIFO | UART_FCR_PXAR8; + else + fcr = UART_FCR_ENABLE_FIFO | UART_FCR_PXAR32; + } + + /* + * Ok, we're now changing the port state. Do it with + * interrupts disabled. + */ + spin_lock_irqsave(&up->port.lock, flags); + + /* + * Ensure the port will be enabled. + * This is required especially for serial console. + */ + up->ier |= UART_IER_UUE; + + /* + * Update the per-port timeout. + */ + uart_update_timeout(port, termios->c_cflag, baud); + + up->port.read_status_mask = UART_LSR_OE | UART_LSR_THRE | UART_LSR_DR; + if (termios->c_iflag & INPCK) + up->port.read_status_mask |= UART_LSR_FE | UART_LSR_PE; + if (termios->c_iflag & (IGNBRK | BRKINT | PARMRK)) + up->port.read_status_mask |= UART_LSR_BI; + + /* + * Characters to ignore + */ + up->port.ignore_status_mask = 0; + if (termios->c_iflag & IGNPAR) + up->port.ignore_status_mask |= UART_LSR_PE | UART_LSR_FE; + if (termios->c_iflag & IGNBRK) { + up->port.ignore_status_mask |= UART_LSR_BI; + /* + * If we're ignoring parity and break indicators, + * ignore overruns too (for real raw support). + */ + if (termios->c_iflag & IGNPAR) + up->port.ignore_status_mask |= UART_LSR_OE; + } + + /* + * ignore all characters if CREAD is not set + */ + if ((termios->c_cflag & CREAD) == 0) + up->port.ignore_status_mask |= UART_LSR_DR; + + /* + * CTS flow control flag and modem status interrupts + */ + if (!up->dma_enable) { + /* Don't enable modem status interrupt if DMA is enabled. + * Inherited from the old code. + * Please also refer to serial_pxa_enable_ms(). + */ + up->ier &= ~UART_IER_MSI; + if (UART_ENABLE_MS(&up->port, termios->c_cflag)) + up->ier |= UART_IER_MSI; + } + + serial_out(up, UART_IER, up->ier); + + if (termios->c_cflag & CRTSCTS) { + up->mcr |= UART_MCR_AFE; + } else { + up->mcr &= ~UART_MCR_AFE; + } + +// #ifndef CONFIG_SOC_SPACEMIT_K1_FPGA +// if (uart_console(port)) { +// quot = 1; +// } +// #endif + + serial_out(up, UART_LCR, cval | UART_LCR_DLAB); /* set DLAB */ + + /* + * the right DLL/DLH setting sequence is: + * write DLH --> read DLH --> write DLL + */ + serial_out(up, UART_DLM, (quot >> 8) & 0xff); /* MS of divisor */ + (void) serial_in(up, UART_DLM); + serial_out(up, UART_DLL, quot & 0xff); /* LS of divisor */ + + /* + * work around Errata #75 according to Intel(R) PXA27x Processor Family + * Specification Update (Nov 2005) + */ + + /* + * read DLL twice in case the uart semi-stable state to trigger this warning. + */ + (void) serial_in(up, UART_DLL); + dll = serial_in(up, UART_DLL); + WARN(dll != (quot & 0xff), + "uart %d baud %d target 0x%x real 0x%x\n", + up->port.line, baud, quot & 0xff, dll); + + serial_out(up, UART_LCR, cval); /* reset DLAB */ + up->lcr = cval; /* Save LCR */ + serial_pxa_set_mctrl(&up->port, up->port.mctrl); + serial_out(up, UART_FCR, fcr); + spin_unlock_irqrestore(&up->port.lock, flags); + + /* + * when console polling the xfer status, we use one-eighth of xfer + * one byte time as an udelay cycle, and clamp to between 1us and + * 20us, for example, if the baud rate is 115200, the caculated out + * udelay cycle is 10us. + */ + if (uart_console(&up->port)) { + up->cons_udelay = 1000000000 / baud * 10 / 8 / 1000; + if (up->cons_udelay <= 0) + up->cons_udelay = 1; + if (up->cons_udelay > 20) + up->cons_udelay = 20; + } + + if (up->dma_enable && up->uart_dma.dma_init) + pxa_uart_receive_dma_start(up); +} + +static void serial_pxa_pm(struct uart_port *port, unsigned int state, unsigned int oldstate) +{ + struct uart_pxa_port *up = (struct uart_pxa_port *)port; + + if (!state) { + clk_prepare_enable(up->gclk); + clk_prepare_enable(up->fclk); + } else { + clk_disable_unprepare(up->fclk); + clk_disable_unprepare(up->gclk); + } +} + +static void serial_pxa_release_port(struct uart_port *port) +{ +} + +static int serial_pxa_request_port(struct uart_port *port) +{ + return 0; +} + +static void serial_pxa_config_port(struct uart_port *port, int flags) +{ + struct uart_pxa_port *up = (struct uart_pxa_port *)port; + up->port.type = PORT_PXA; +} + +static int +serial_pxa_verify_port(struct uart_port *port, struct serial_struct *ser) +{ + /* we don't want the core code to modify any port params */ + return -EINVAL; +} + +static const char * +serial_pxa_type(struct uart_port *port) +{ + struct uart_pxa_port *up = (struct uart_pxa_port *)port; + return up->name; +} + +static struct uart_pxa_port *serial_pxa_ports[NUM_UART_PORTS]; +static struct uart_driver serial_pxa_reg; + +#ifdef CONFIG_PM +void serial_pxa_get_qos(int port) +{ + struct uart_pxa_port *up; + + if ((port < 0) || (port >= NUM_UART_PORTS)) { + pr_err("%s: wrong uart port %d\n", __func__, port); + return; + } + + up = serial_pxa_ports[port]; + if (!mod_timer(&up->pxa_timer, jiffies + PXA_TIMER_TIMEOUT)) { + pr_info("bluesleep: %s: get qos\n", __func__); + pm_runtime_get_sync(up->port.dev); + } + + return; +} + +EXPORT_SYMBOL_GPL(serial_pxa_get_qos); +#endif /* #ifdef CONFIG_PM */ + +void serial_pxa_assert_rts(int port) +{ + struct uart_pxa_port *up; + unsigned long flags; + + if ((port < 0) || (port >= NUM_UART_PORTS)) { + pr_err("%s: wrong uart port %d\n", __func__, port); + return; + } + + up = serial_pxa_ports[port]; + + spin_lock_irqsave(&up->port.lock, flags); + if (!serial_pxa_is_open(up)) { + spin_unlock_irqrestore(&up->port.lock, flags); + pr_info("%s: uart %d is shutdown\n", __func__, port); + return; + } + serial_pxa_set_mctrl(&up->port, up->port.mctrl | TIOCM_RTS); + /* clear serial_core hw_stopped when BT not assert RTS yet */ + uart_handle_cts_change(&up->port, UART_MSR_CTS); + spin_unlock_irqrestore(&up->port.lock, flags); + + return; +} + +EXPORT_SYMBOL_GPL(serial_pxa_assert_rts); + +void serial_pxa_deassert_rts(int port) +{ + struct uart_pxa_port *up; + unsigned long flags; + + if ((port < 0) || (port >= NUM_UART_PORTS)) { + pr_err("%s: wrong uart port %d\n", __func__, port); + return; + } + + up = serial_pxa_ports[port]; + + spin_lock_irqsave(&up->port.lock, flags); + if (!serial_pxa_is_open(up)) { + spin_unlock_irqrestore(&up->port.lock, flags); + pr_info("%s: uart %d is shutdown\n", __func__, port); + return; + } + serial_pxa_set_mctrl(&up->port, up->port.mctrl & ~TIOCM_RTS); + spin_unlock_irqrestore(&up->port.lock, flags); + + return; +} + +EXPORT_SYMBOL_GPL(serial_pxa_deassert_rts); + + +#ifdef CONFIG_SERIAL_PXA_CONSOLE + +#define BOTH_EMPTY (UART_LSR_TEMT | UART_LSR_THRE) + +/* + * Wait for transmitter & holding register to empty + */ +static void wait_for_xmitr(struct uart_pxa_port *up) +{ + unsigned int status, tmout = 10000; + unsigned int cycle; + + /* + * for early console, the cons_udelay is not initialized yet, + * we use 1us udelay cycle. + */ + if (uart_console(&up->port)) + cycle = up->cons_udelay; + else + cycle = 1; + + tmout = 10000 / cycle; + + /* Wait up to 10ms for the character(s) to be sent. */ + do { + status = serial_in(up, UART_LSR); + + if (status & UART_LSR_BI) + up->lsr_break_flag = UART_LSR_BI; + + if ((status & BOTH_EMPTY) == BOTH_EMPTY) + break; + udelay(cycle); + } while (--tmout); + + /* Wait up to 1s for flow control if necessary */ + if (up->port.flags & UPF_CONS_FLOW) { + tmout = 1000000; + while (--tmout && + ((serial_in(up, UART_MSR) & UART_MSR_CTS) == 0)) + udelay(cycle); + } + + /* + * From the opinion of uart, there is no reason not to return + * the status. When it happened, do panic to debug. + */ + if (!tmout) { + if (up->port.flags & UPF_CONS_FLOW) + status = serial_in(up, UART_MSR); + else + status = serial_in(up, UART_LSR); + panic("failed to read uart status, status:0x%08x\n", status); + } +} + +static void serial_pxa_console_putchar(struct uart_port *port, unsigned char ch) +{ + struct uart_pxa_port *up = (struct uart_pxa_port *)port; + + wait_for_xmitr(up); + serial_out(up, UART_TX, ch); +} + +/* + * Print a string to the serial port trying not to disturb + * any possible real use of the port... + * + * The console_lock must be held when we get here. + */ +static void +serial_pxa_console_write(struct console *co, const char *s, unsigned int count) +{ + struct uart_pxa_port *up = serial_pxa_ports[co->index]; + unsigned int ier; + unsigned long flags; + int locked = 1; + + clk_enable(up->gclk); + clk_enable(up->fclk); + + local_irq_save(flags); + if (up->port.sysrq) + locked = 0; + else if (oops_in_progress) + locked = spin_trylock(&up->port.lock); + else + spin_lock(&up->port.lock); + + /* + * First save the IER then disable the interrupts + */ + ier = serial_in(up, UART_IER); + serial_out(up, UART_IER, UART_IER_UUE); + + uart_console_write(&up->port, s, count, serial_pxa_console_putchar); + + /* + * Finally, wait for transmitter to become empty + * and restore the IER + */ + wait_for_xmitr(up); + serial_out(up, UART_IER, ier); + + if (locked) + spin_unlock(&up->port.lock); + local_irq_restore(flags); + clk_disable(up->fclk); + clk_disable(up->gclk); +} + +#ifdef CONFIG_CONSOLE_POLL +/* + * Console polling routines for writing and reading from the uart while + * in an interrupt or debug context. + */ + +static int serial_pxa_get_poll_char(struct uart_port *port) +{ + struct uart_pxa_port *up = (struct uart_pxa_port *)port; + unsigned char lsr = serial_in(up, UART_LSR); + + while (!(lsr & UART_LSR_DR)) + lsr = serial_in(up, UART_LSR); + + return serial_in(up, UART_RX); +} + + +static void serial_pxa_put_poll_char(struct uart_port *port, + unsigned char c) +{ + unsigned int ier; + struct uart_pxa_port *up = (struct uart_pxa_port *)port; + + /* + * First save the IER then disable the interrupts + */ + ier = serial_in(up, UART_IER); + serial_out(up, UART_IER, UART_IER_UUE); + + wait_for_xmitr(up); + /* + * Send the character out. + */ + serial_out(up, UART_TX, c); + /* If a LF, also do CR... */ + if (c == 10) { + wait_for_xmitr(up); + serial_out(up, UART_TX, 13); + } + + /* + * Finally, wait for transmitter to become empty + * and restore the IER + */ + wait_for_xmitr(up); + serial_out(up, UART_IER, ier); +} + +#endif /* CONFIG_CONSOLE_POLL */ + +static int __init +serial_pxa_console_setup(struct console *co, char *options) +{ + struct uart_pxa_port *up; + int baud = 9600; + int bits = 8; + int parity = 'n'; + int flow = 'n'; + + if (co->index == -1 || co->index >= serial_pxa_reg.nr) + co->index = 0; + up = serial_pxa_ports[co->index]; + if (!up) + return -ENODEV; + + if (options) + uart_parse_options(options, &baud, &parity, &bits, &flow); + + return uart_set_options(&up->port, co, baud, parity, bits, flow); +} + +static struct console serial_pxa_console = { + .name = "ttyS", + .write = serial_pxa_console_write, + .device = uart_console_device, + .setup = serial_pxa_console_setup, + .flags = CON_PRINTBUFFER, + .index = -1, + .data = &serial_pxa_reg, +}; + +/* Support for earlycon */ +static void pxa_early_write(struct console *con, const char *s, + unsigned n) +{ + struct earlycon_device *dev = con->data; + + uart_console_write(&dev->port, s, n, serial_pxa_console_putchar); +} + +static int __init pxa_early_console_setup(struct earlycon_device *device, const char *opt) +{ + if (!device->port.membase) { + return -ENODEV; + } + + device->con->write = pxa_early_write; + return 0; +} + +EARLYCON_DECLARE(pxa_serial, pxa_early_console_setup); +OF_EARLYCON_DECLARE(pxa_serial, "spacemit,pxa-uart", pxa_early_console_setup); + +#define PXA_CONSOLE &serial_pxa_console +#else +#define PXA_CONSOLE NULL +#endif /* CONFIG_SERIAL_PXA_CONSOLE */ + +static const struct uart_ops serial_pxa_pops = { + .tx_empty = serial_pxa_tx_empty, + .set_mctrl = serial_pxa_set_mctrl, + .get_mctrl = serial_pxa_get_mctrl, + .stop_tx = serial_pxa_stop_tx, + .start_tx = serial_pxa_start_tx, + .stop_rx = serial_pxa_stop_rx, + .enable_ms = serial_pxa_enable_ms, + .break_ctl = serial_pxa_break_ctl, + .startup = serial_pxa_startup, + .shutdown = serial_pxa_shutdown, + .set_termios = serial_pxa_set_termios, + .pm = serial_pxa_pm, + .type = serial_pxa_type, + .release_port = serial_pxa_release_port, + .request_port = serial_pxa_request_port, + .config_port = serial_pxa_config_port, + .verify_port = serial_pxa_verify_port, +#if defined(CONFIG_CONSOLE_POLL) && defined(CONFIG_SERIAL_PXA_CONSOLE) + .poll_get_char = serial_pxa_get_poll_char, + .poll_put_char = serial_pxa_put_poll_char, +#endif +}; + +static struct uart_driver serial_pxa_reg = { + .owner = THIS_MODULE, + .driver_name = "PXA serial", + .dev_name = "ttyS", + .major = TTY_MAJOR, + .minor = 64, + .nr = NUM_UART_PORTS, + .cons = PXA_CONSOLE, +}; + +static int serial_pxa_is_open(struct uart_pxa_port *up) +{ + struct uart_state *state; + struct uart_pxa_dma *pxa_dma; + + if (!up) { + return 0; + } + + state = serial_pxa_reg.state + up->port.line; + pxa_dma = &up->uart_dma; + + /* clock is enabled, DMA is ready */ + if (up->dma_enable) { + return ((up->ier & UART_IER_DMAE) && pxa_dma->dma_init && + (state->pm_state == UART_PM_STATE_ON)); + } else { + return (state->pm_state == UART_PM_STATE_ON); + } +} + +#ifdef CONFIG_PM +static int serial_pxa_suspend(struct device *dev) +{ + struct uart_pxa_port *sport = dev_get_drvdata(dev); + struct uart_pxa_dma *pxa_dma = &sport->uart_dma; + struct dma_tx_state dma_state; + unsigned char tmp[256]; + int fifo_cnt, cnt = 0; + + if (!console_suspend_enabled || !sport) { + return 0; + } + + if (serial_pxa_is_open(sport) && sport->dma_enable) { + int sent = 0; + unsigned long flags; + + local_irq_save(flags); + /* + * tx stop and suspend and when resume, + * tx startup would be called and set it to 0 + */ + pxa_dma->tx_stop = 1; + pxa_dma->rx_stop = 1; + pxa_dma->tx_saved_len = 0; + if (dma_async_is_tx_complete(pxa_dma->txdma_chan, \ + pxa_dma->tx_cookie, NULL, NULL)!= DMA_COMPLETE) { + dmaengine_pause(pxa_dma->txdma_chan); + udelay(100); + dmaengine_tx_status(pxa_dma->txdma_chan, + pxa_dma->tx_cookie, &dma_state); + sent = pxa_dma->tx_size - dma_state.residue; + pxa_dma->tx_saved_len = dma_state.residue; + memcpy(pxa_dma->tx_buf_save, pxa_dma->txdma_addr + sent, + dma_state.residue); + stop_dma(sport, PXA_UART_TX); + } + + if (dma_async_is_tx_complete(pxa_dma->rxdma_chan, + pxa_dma->rx_cookie, NULL, NULL) != DMA_COMPLETE) { + /* before stop receive, de-assert RTS */ + spin_lock(&sport->port.lock); + serial_pxa_set_mctrl(&sport->port, sport->port.mctrl & ~TIOCM_RTS); + spin_unlock(&sport->port.lock); + udelay(100); + dmaengine_pause(pxa_dma->rxdma_chan); + udelay(100); + pxa_uart_receive_dma_cb(sport); + stop_dma(sport, PXA_UART_RX); + + /* fetch the trailing bytes from FIFO */ + fifo_cnt = serial_in(sport, UART_FOR); + while (fifo_cnt > 0) { + *(tmp + cnt) = serial_in(sport, UART_RX) & 0xff; + cnt++; + fifo_cnt = serial_in(sport, UART_FOR); + } + + if (cnt > 0) { + tty_insert_flip_string(&sport->port.state->port, tmp, cnt); + sport->port.icount.rx += cnt; + tty_flip_buffer_push(&sport->port.state->port); + } + } + local_irq_restore(flags); + } + + if (sport) { + uart_suspend_port(&serial_pxa_reg, &sport->port); + } + +#ifdef CONFIG_PM + /* Remove uart rx constraint which will block system entering low power state. */ + if (del_timer_sync(&sport->pxa_timer)) { + _pxa_timer_handler(sport); + } +#endif + + return 0; +} + +static int serial_pxa_resume(struct device *dev) +{ + struct uart_pxa_port *sport = dev_get_drvdata(dev); + struct uart_pxa_dma *pxa_dma = &sport->uart_dma; + + if (!console_suspend_enabled || !sport) { + return 0; + } + + sport->in_resume = true; + uart_resume_port(&serial_pxa_reg, &sport->port); + + if (serial_pxa_is_open(sport) && sport->dma_enable) { + if (pxa_dma->tx_saved_len > 0) { + /* + * if this transmit is not finished, we should stop + * new transmit from user space. + */ + sport->from_resume = true; + memcpy(pxa_dma->txdma_addr, pxa_dma->tx_buf_save, + pxa_dma->tx_saved_len); + pxa_uart_transmit_dma_start(sport, + pxa_dma->tx_saved_len); + } else { + tasklet_schedule(&pxa_dma->tklet); + } + + pxa_uart_receive_dma_start(sport); + } + sport->in_resume = false; + + return 0; +} + +static const struct dev_pm_ops serial_pxa_pm_ops = { + .suspend = serial_pxa_suspend, + .resume = serial_pxa_resume, +}; +#endif + +#ifdef CONFIG_PM +static void _pxa_timer_handler(struct uart_pxa_port *up) +{ +#if SUPPORT_POWER_QOS + pm_runtime_put_sync(up->port.dev); +#endif + if (up->port.line == BT_UART_PORT) { + pr_info("bluesleep: %s: release qos\n", __func__); + } +} + +static void pxa_timer_handler(struct timer_list *t) +{ + struct uart_pxa_port *up = from_timer(up, t, pxa_timer); + + _pxa_timer_handler(up); +} + +static void __maybe_unused uart_edge_wakeup_handler(int gpio, void *data) +{ + struct uart_pxa_port *up = (struct uart_pxa_port *)data; + + if (!mod_timer(&up->pxa_timer, jiffies + PXA_TIMER_TIMEOUT)) { +#if SUPPORT_POWER_QOS + pm_runtime_get_sync(up->port.dev); +#endif + } + + /* Avoid suspend within 3 seconds. */ + pm_wakeup_event(up->port.dev, BLOCK_SUSPEND_TIMEOUT); +} + +static void uart_tx_lpm_handler(struct work_struct *work) +{ + struct uart_pxa_port *up = container_of(work, + struct uart_pxa_port, uart_tx_lpm_work); + + /* Polling until TX FIFO is empty */ + while (!(serial_in(up, UART_LSR) & UART_LSR_TEMT)) { + usleep_range(1000, 2000); + } +#if SUPPORT_POWER_QOS + pm_runtime_put_sync(up->port.dev); +#endif +} +#endif + +static const struct of_device_id serial_pxa_dt_ids[] = { + { .compatible = "spacemit,pxa-uart", }, + {} +}; + +static int serial_pxa_probe_dt(struct platform_device *pdev, struct uart_pxa_port *sport) +{ + struct device_node *np = pdev->dev.of_node; + int ret; + + if (!np) { + return 1; + } + + /* device tree is used */ + if (of_get_property(np, "dmas", NULL)) { + sport->dma_enable = 1; + } + ret = of_alias_get_id(np, "serial"); + if (ret < 0) { + dev_err(&pdev->dev, "failed to get alias id, errno %d\n", ret); + return ret; + } + sport->port.line = ret; + +#ifdef CONFIG_PM + if (of_property_read_u32(np, "edge-wakeup-pin", &sport->edge_wakeup_gpio)) { + dev_info(&pdev->dev, "no edge-wakeup-pin defined\n"); + } +#endif + sport->device_ctrl_rts = of_property_read_bool(np, "device-control-rts"); + + return 0; +} + +static int serial_pxa_probe(struct platform_device *dev) +{ + struct uart_pxa_port *sport; + struct resource *mmres; + int ret; + int irq; + struct resource *dmares; + struct uart_pxa_dma *pxa_dma; +#ifdef CONFIG_SOC_SPACEMIT_K1_FPGA + struct device_node *np = dev->dev.of_node; +#endif + mmres = platform_get_resource(dev, IORESOURCE_MEM, 0); + if (!mmres) { + return -ENODEV; + } + + irq = platform_get_irq(dev, 0); + if (irq < 0) { + return irq; + } + + sport = kzalloc(sizeof(struct uart_pxa_port), GFP_KERNEL); + if (!sport) { + return -ENOMEM; + } + +#ifdef CONFIG_PM + sport->uart_dma.tx_buf_save = kmalloc(DMA_BLOCK, GFP_KERNEL); + if (!sport->uart_dma.tx_buf_save) { + kfree(sport); + return -ENOMEM; + } +#endif + sport->gclk = devm_clk_get(&dev->dev, "gate"); + if (IS_ERR(sport->gclk)) { + ret = PTR_ERR(sport->gclk); + goto err_free; + } + + sport->fclk = devm_clk_get(&dev->dev, "func"); + if (IS_ERR(sport->fclk)) { + ret = PTR_ERR(sport->fclk); + goto err_free; + } + + ret = clk_prepare(sport->gclk); + if (ret) { + clk_put(sport->gclk); + goto err_free; + } + + ret = clk_prepare(sport->fclk); + if (ret) { + clk_put(sport->fclk); + goto err_free; + } + + sport->port.type = PORT_PXA; + sport->port.iotype = UPIO_MEM; + sport->port.mapbase = mmres->start; + sport->port.irq = irq; + sport->port.fifosize = 64; + sport->port.ops = &serial_pxa_pops; + sport->port.dev = &dev->dev; + sport->port.flags = UPF_IOREMAP | UPF_BOOT_AUTOCONF; + +#ifdef CONFIG_SOC_SPACEMIT_K1_FPGA + of_property_read_u32(np, "clk-fpga", &sport->clk_fpga); + if (sport->clk_fpga) { + sport->port.uartclk = sport->clk_fpga; + } else { + sport->port.uartclk = UARTCLK_FPGA; + } +#else + sport->port.uartclk = clk_get_rate(sport->fclk); +#endif + sport->resets = devm_reset_control_get_optional(&dev->dev, NULL); + if(IS_ERR(sport->resets)) { + ret = PTR_ERR(sport->resets); + goto err_clk; + } + reset_control_deassert(sport->resets); + + sport->port.has_sysrq = IS_ENABLED(CONFIG_SERIAL_PXA_CONSOLE); + + sport->edge_wakeup_gpio = -1; + + pxa_dma = &sport->uart_dma; + pxa_dma->txdma_chan = NULL; + pxa_dma->rxdma_chan = NULL; + pxa_dma->txdma_addr = NULL; + pxa_dma->rxdma_addr = NULL; + pxa_dma->dma_init = false; + sport->dma_enable = 0; + sport->cons_udelay = 1; + sport->in_resume = false; + + ret = serial_pxa_probe_dt(dev, sport); + if (ret > 0) { + sport->port.line = dev->id; + } else if (ret < 0) { + goto err_rst; + } + if (sport->port.line >= ARRAY_SIZE(serial_pxa_ports)) { + dev_err(&dev->dev, "serial%d out of range\n", sport->port.line); + ret = -EINVAL; + goto err_rst; + } + snprintf(sport->name, PXA_NAME_LEN - 1, "UART%d", sport->port.line + 1); + + dma_set_mask(&dev->dev, DMA_BIT_MASK(64)); + dma_set_coherent_mask(&dev->dev, DMA_BIT_MASK(64)); + if (ret > 0 && sport->dma_enable) { + /* Get Rx DMA mapping value */ + dmares = platform_get_resource(dev, IORESOURCE_DMA, 0); + if (dmares) { + /* Get Tx DMA mapping value */ + dmares = platform_get_resource(dev, IORESOURCE_DMA, 1); + if (dmares) { + sport->dma_enable = 1; + } + } + } + + /* + * Allocate the IRQ in probe, otherwise if move suspend/resume to + * noirq stage will cause a unbalance irq enable warning. + */ + ret = request_irq(sport->port.irq, serial_pxa_irq, 0, sport->name, sport); + if (ret) { + goto err_rst; + } + disable_irq(sport->port.irq); + +#ifdef CONFIG_PM +#if SUPPORT_POWER_QOS + pm_runtime_enable(&dev->dev); + pm_runtime_set_active(&dev->dev); + pm_runtime_irq_safe(&dev->dev); +#endif +#endif + + sport->port.membase = ioremap(mmres->start, resource_size(mmres)); + if (!sport->port.membase) { + ret = -ENOMEM; + goto err_qos; + } + +#ifdef CONFIG_PM + INIT_WORK(&sport->uart_tx_lpm_work, uart_tx_lpm_handler); + + timer_setup(&sport->pxa_timer, pxa_timer_handler, 0); +#endif + + serial_pxa_ports[sport->port.line] = sport; + uart_add_one_port(&serial_pxa_reg, &sport->port); + dev_info(&dev->dev, "uart clk_rate: %lu\n", clk_get_rate(sport->fclk)); + platform_set_drvdata(dev, sport); + +#ifdef CONFIG_PM +/* if (sport->edge_wakeup_gpio >= 0) { + device_init_wakeup(&dev->dev, 1); + ret = request_mfp_edge_wakeup(sport->edge_wakeup_gpio, + uart_edge_wakeup_handler, + sport, &dev->dev); + if (ret) { + dev_err(&dev->dev, "failed to request edge wakeup.\n"); + goto err_edge; + } + }*/ +#endif + + return 0; + +#ifdef CONFIG_PM +//err_edge: + uart_remove_one_port(&serial_pxa_reg, &sport->port); + iounmap(sport->port.membase); +#endif +err_qos: +#ifdef CONFIG_PM + pm_runtime_disable(&dev->dev); +#endif + free_irq(sport->port.irq, sport); +err_rst: + reset_control_assert(sport->resets); +err_clk: + clk_unprepare(sport->fclk); + clk_unprepare(sport->gclk); + clk_put(sport->fclk); + clk_put(sport->gclk); +err_free: + kfree(sport); + return ret; +} + +static int serial_pxa_remove(struct platform_device *dev) +{ + struct uart_pxa_port *sport = platform_get_drvdata(dev); + +#ifdef CONFIG_PM + pm_runtime_disable(&dev->dev); +#endif + + uart_remove_one_port(&serial_pxa_reg, &sport->port); + + reset_control_assert(sport->resets); + free_irq(sport->port.irq, sport); + clk_unprepare(sport->fclk); + clk_unprepare(sport->gclk); + clk_put(sport->fclk); + clk_put(sport->gclk); + +#ifdef CONFIG_PM +/* if (sport->edge_wakeup_gpio >= 0) { + remove_mfp_edge_wakeup(sport->edge_wakeup_gpio); + } +*/ + kfree(sport->uart_dma.tx_buf_save); +#endif + kfree(sport); + serial_pxa_ports[dev->id] = NULL; + + return 0; +} + + +static struct platform_driver serial_pxa_driver = { + .probe = serial_pxa_probe, + .remove = serial_pxa_remove, + .driver = { + .name = "pxa2xx-uart", +#ifdef CONFIG_PM + .pm = &serial_pxa_pm_ops, +#endif + .suppress_bind_attrs = true, + .of_match_table = serial_pxa_dt_ids, + }, +}; + + +/* 8250 driver for PXA serial ports should be used */ +static int __init serial_pxa_init(void) +{ + int ret; + + ret = uart_register_driver(&serial_pxa_reg); + if (ret != 0) { + return ret; + } + + ret = platform_driver_register(&serial_pxa_driver); + if (ret != 0) { + uart_unregister_driver(&serial_pxa_reg); + } + + return ret; +} + +static void __exit serial_pxa_exit(void) +{ + platform_driver_unregister(&serial_pxa_driver); + uart_unregister_driver(&serial_pxa_reg); +} +module_init(serial_pxa_init); +module_exit(serial_pxa_exit); -- cgit v1.2.3