else
areq_ctx->assoc_buff_type = SSI_DMA_BUF_MLLI;
- if ((do_chain) || areq_ctx->assoc_buff_type == SSI_DMA_BUF_MLLI) {
+ if (do_chain || areq_ctx->assoc_buff_type == SSI_DMA_BUF_MLLI) {
dev_dbg(dev, "Chain assoc: buff_type=%s nents=%u\n",
cc_dma_buf_type(areq_ctx->assoc_buff_type),
areq_ctx->assoc.nents);
areq_ctx->mlli_params.mlli_dma_addr);
}
- if ((src) && areq_ctx->in_nents) {
+ if (src && areq_ctx->in_nents) {
dev_dbg(dev, "Unmapped sg src: virt=%pK dma=%pad len=0x%X\n",
sg_virt(src), &sg_dma_address(src), sg_dma_len(src));
dma_unmap_sg(dev, src,
drvdata->irq = irr;
/* Completion interrupt - most probable */
- if ((irr & SSI_COMP_IRQ_MASK)) {
+ if (irr & SSI_COMP_IRQ_MASK) {
/* Mask AXI completion interrupt - will be unmasked in
* Deferred service handler
*/
}
#ifdef CC_SUPPORT_FIPS
/* TEE FIPS interrupt */
- if ((irr & SSI_GPR0_IRQ_MASK)) {
+ if (irr & SSI_GPR0_IRQ_MASK) {
/* Mask interrupt - will be unmasked in Deferred service
* handler
*/
}
#endif
/* AXI error interrupt */
- if ((irr & SSI_AXI_ERR_IRQ_MASK)) {
+ if (irr & SSI_AXI_ERR_IRQ_MASK) {
u32 axi_err;
/* Read the AXI error ID */
return -EBUSY;
}
- if ((req_mgr_h->q_free_slots >= total_seq_len))
+ if (req_mgr_h->q_free_slots >= total_seq_len)
return 0;
/* Wait for space in HW queue. Poll constant num of iterations. */