/* * Copyright (c) by Jaroslav Kysela * Copyright (c) by Takashi Iwai * Copyright (c) by Scott McNab * * Trident 4DWave-NX memory page allocation (TLB area) * Trident chip can handle only 16MByte of the memory at the same time. * * * This program is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation; either version 2 of the License, or * (at your option) any later version. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program; if not, write to the Free Software * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA * */ #include #include #include #include #include #include /* page arguments of these two macros are Trident page (4096 bytes), not like * aligned pages in others */ #define __set_tlb_bus(trident,page,ptr,addr) \ do { (trident)->tlb.entries[page] = cpu_to_le32((addr) & ~(SNDRV_TRIDENT_PAGE_SIZE-1)); \ (trident)->tlb.shadow_entries[page] = (ptr); } while (0) #define __tlb_to_ptr(trident,page) \ (void*)((trident)->tlb.shadow_entries[page]) #define __tlb_to_addr(trident,page) \ (dma_addr_t)le32_to_cpu((trident->tlb.entries[page]) & ~(SNDRV_TRIDENT_PAGE_SIZE - 1)) #if PAGE_SIZE == 4096 /* page size == SNDRV_TRIDENT_PAGE_SIZE */ #define ALIGN_PAGE_SIZE PAGE_SIZE /* minimum page size for allocation */ #define MAX_ALIGN_PAGES SNDRV_TRIDENT_MAX_PAGES /* maxmium aligned pages */ /* fill TLB entrie(s) corresponding to page with ptr */ #define set_tlb_bus(trident,page,ptr,addr) __set_tlb_bus(trident,page,ptr,addr) /* fill TLB entrie(s) corresponding to page with silence pointer */ #define set_silent_tlb(trident,page) __set_tlb_bus(trident, page, (unsigned long)trident->tlb.silent_page.area, trident->tlb.silent_page.addr) /* get aligned page from offset address */ #define get_aligned_page(offset) ((offset) >> 12) /* get offset address from aligned page */ #define aligned_page_offset(page) ((page) << 12) /* get buffer address from aligned page */ #define page_to_ptr(trident,page) __tlb_to_ptr(trident, page) /* get PCI physical address from aligned page */ #define page_to_addr(trident,page) __tlb_to_addr(trident, page) #elif PAGE_SIZE == 8192 /* page size == SNDRV_TRIDENT_PAGE_SIZE x 2*/ #define ALIGN_PAGE_SIZE PAGE_SIZE #define MAX_ALIGN_PAGES (SNDRV_TRIDENT_MAX_PAGES / 2) #define get_aligned_page(offset) ((offset) >> 13) #define aligned_page_offset(page) ((page) << 13) #define page_to_ptr(trident,page) __tlb_to_ptr(trident, (page) << 1) #define page_to_addr(trident,page) __tlb_to_addr(trident, (page) << 1) /* fill TLB entries -- we need to fill two entries */ static inline void set_tlb_bus(struct snd_trident *trident, int page, unsigned long ptr, dma_addr_t addr) { page <<= 1; __set_tlb_bus(trident, page, ptr, addr); __set_tlb_bus(trident, page+1, ptr + SNDRV_TRIDENT_PAGE_SIZE, addr + SNDRV_TRIDENT_PAGE_SIZE); } static inline void set_silent_tlb(struct snd_trident *trident, int page) { page <<= 1; __set_tlb_bus(trident, page, (unsigned long)trident->tlb.silent_page.area, trident->tlb.silent_page.addr); __set_tlb_bus(trident, page+1, (unsigned long)trident->tlb.silent_page.area, trident->tlb.silent_page.addr); } #else /* arbitrary size */ #define UNIT_PAGES (PAGE_SIZE / SNDRV_TRIDENT_PAGE_SIZE) #define ALIGN_PAGE_SIZE (SNDRV_TRIDENT_PAGE_SIZE * UNIT_PAGES) #define MAX_ALIGN_PAGES (SNDRV_TRIDENT_MAX_PAGES / UNIT_PAGES) /* Note: if alignment doesn't match to the maximum size, the last few blocks * become unusable. To use such blocks, you'll need to check the validity * of accessing page in set_tlb_bus and set_silent_tlb. search_empty() * should also check it, too. */ #define get_aligned_page(offset) ((offset) / ALIGN_PAGE_SIZE) #define aligned_page_offset(page) ((page) * ALIGN_PAGE_SIZE) #define page_to_ptr(trident,page) __tlb_to_ptr(trident, (page) * UNIT_PAGES) #define page_to_addr(trident,page) __tlb_to_addr(trident, (page) * UNIT_PAGES) /* fill TLB entries -- UNIT_PAGES entries must be filled */ static inline void set_tlb_bus(struct snd_trident *trident, int page, unsigned long ptr, dma_addr_t addr) { int i; page *= UNIT_PAGES; for (i = 0; i < UNIT_PAGES; i++, page++) { __set_tlb_bus(trident, page, ptr, addr); ptr += SNDRV_TRIDENT_PAGE_SIZE; addr += SNDRV_TRIDENT_PAGE_SIZE; } } static inline void set_silent_tlb(struct snd_trident *trident, int page) { int i; page *= UNIT_PAGES; for (i = 0; i < UNIT_PAGES; i++, page++) __set_tlb_bus(trident, page, (unsigned long)trident->tlb.silent_page.area, trident->tlb.silent_page.addr); } #endif /* PAGE_SIZE */ /* calculate buffer pointer from offset address */ static inline void *offset_ptr(struct snd_trident *trident, int offset) { char *ptr; ptr = page_to_ptr(trident, get_aligned_page(offset)); ptr += offset % ALIGN_PAGE_SIZE; return (void*)ptr; } /* first and last (aligned) pages of memory block */ #define firstpg(blk) (((struct snd_trident_memblk_arg *)snd_util_memblk_argptr(blk))->first_page) #define lastpg(blk) (((struct snd_trident_memblk_arg *)snd_util_memblk_argptr(blk))->last_page) /* * search empty pages which may contain given size */ static struct snd_util_memblk * search_empty(struct snd_util_memhdr *hdr, int size) { struct snd_util_memblk *blk, *prev; int page, psize; struct list_head *p; psize = get_aligned_page(size + ALIGN_PAGE_SIZE -1); prev = NULL; page = 0; list_for_each(p, &hdr->block) { blk = list_entry(p, struct snd_util_memblk, list); if (page + psize <= firstpg(blk)) goto __found_pages; page = lastpg(blk) + 1; } if (page + psize > MAX_ALIGN_PAGES) return NULL; __found_pages: /* create a new memory block */ blk = __snd_util_memblk_new(hdr, psize * ALIGN_PAGE_SIZE, p->prev); if (blk == NULL) return NULL; blk->offset = aligned_page_offset(page); /* set aligned offset */ firstpg(blk) = page; lastpg(blk) = page + psize - 1; return blk; } /* * check if the given pointer is valid for pages */ static int is_valid_page(unsigned long ptr) { if (ptr & ~0x3fffffffUL) { snd_printk(KERN_ERR "max memory size is 1GB!!\n"); return 0; } if (ptr & (SNDRV_TRIDENT_PAGE_SIZE-1)) { snd_printk(KERN_ERR "page is not aligned\n"); return 0; } return 1; } /* * page allocation for DMA (Scatter-Gather version) */ static struct snd_util_memblk * snd_trident_alloc_sg_pages(struct snd_trident *trident, struct snd_pcm_substream *substream) { struct snd_util_memhdr *hdr; struct snd_util_memblk *blk; struct snd_pcm_runtime *runtime = substream->runtime; int idx, page; struct snd_sg_buf *sgbuf = snd_pcm_substream_sgbuf(substream); snd_assert(runtime->dma_bytes > 0 && runtime->dma_bytes <= SNDRV_TRIDENT_MAX_PAGES * SNDRV_TRIDENT_PAGE_SIZE, return NULL); hdr = trident->tlb.memhdr; snd_assert(hdr != NULL, return NULL); down(&hdr->block_mutex); blk = search_empty(hdr, runtime->dma_bytes); if (blk == NULL) { up(&hdr->block_mutex); return NULL; } if (lastpg(blk) - firstpg(blk) >= sgbuf->pages) { snd_printk(KERN_ERR "page calculation doesn't match: allocated pages = %d, trident = %d/%d\n", sgbuf->pages, firstpg(blk), lastpg(blk)); __snd_util_mem_free(hdr, blk); up(&hdr->block_mutex); return NULL; } /* set TLB entries */ idx = 0; for (page = firstpg(blk); page <= lastpg(blk); page++, idx++) { dma_addr_t addr = sgbuf->table[idx].addr; unsigned long ptr = (unsigned long)sgbuf->table[idx].buf; if (! is_valid_page(addr)) { __snd_util_mem_free(hdr, blk); up(&hdr->block_mutex); return NULL; } set_tlb_bus(trident, page, ptr, addr); } up(&hdr->block_mutex); return blk; } /* * page allocation for DMA (contiguous version) */ static struct snd_util_memblk * snd_trident_alloc_cont_pages(struct snd_trident *trident, struct snd_pcm_substream *substream) { struct snd_util_memhdr *hdr; struct snd_util_memblk *blk; int page; struct snd_pcm_runtime *runtime = substream->runtime; dma_addr_t addr; unsigned long ptr; snd_assert(runtime->dma_bytes> 0 && runtime->dma_bytes <= SNDRV_TRIDENT_MAX_PAGES * SNDRV_TRIDENT_PAGE_SIZE, return NULL); hdr = trident->tlb.memhdr; snd_assert(hdr != NULL, return NULL); down(&hdr->block_mutex); blk = search_empty(hdr, runtime->dma_bytes); if (blk == NULL) { up(&hdr->block_mutex); return NULL; } /* set TLB entries */ addr = runtime->dma_addr; ptr = (unsigned long)runtime->dma_area; for (page = firstpg(blk); page <= lastpg(blk); page++, ptr += SNDRV_TRIDENT_PAGE_SIZE, addr += SNDRV_TRIDENT_PAGE_SIZE) { if (! is_valid_page(addr)) { __snd_util_mem_free(hdr, blk); up(&hdr->block_mutex); return NULL; } set_tlb_bus(trident, page, ptr, addr); } up(&hdr->block_mutex); return blk; } /* * page allocation for DMA */ struct snd_util_memblk * snd_trident_alloc_pages(struct snd_trident *trident, struct snd_pcm_substream *substream) { snd_assert(trident != NULL, return NULL); snd_assert(substream != NULL, return NULL); if (substream->dma_buffer.dev.type == SNDRV_DMA_TYPE_DEV_SG) return snd_trident_alloc_sg_pages(trident, substream); else return snd_trident_alloc_cont_pages(trident, substream); } /* * release DMA buffer from page table */ int snd_trident_free_pages(struct snd_trident *trident, struct snd_util_memblk *blk) { struct snd_util_memhdr *hdr; int page; snd_assert(trident != NULL, return -EINVAL); snd_assert(blk != NULL, return -EINVAL); hdr = trident->tlb.memhdr; down(&hdr->block_mutex); /* reset TLB entries */ for (page = firstpg(blk); page <= lastpg(blk); page++) set_silent_tlb(trident, page); /* free memory block */ __snd_util_mem_free(hdr, blk); up(&hdr->block_mutex); return 0; } /*---------------------------------------------------------------- * memory allocation using multiple pages (for synth) *---------------------------------------------------------------- * Unlike the DMA allocation above, non-contiguous pages are * assigned to TLB. *----------------------------------------------------------------*/ /* */ static int synth_alloc_pages(struct snd_trident *hw, struct snd_util_memblk *blk); static int synth_free_pages(struct snd_trident *hw, struct snd_util_memblk *blk); /* * allocate a synth sample area */ struct snd_util_memblk * snd_trident_synth_alloc(struct snd_trident *hw, unsigned int size) { struct snd_util_memblk *blk; struct snd_util_memhdr *hdr = hw->tlb.memhdr; down(&hdr->block_mutex); blk = __snd_util_mem_alloc(hdr, size); if (blk == NULL) { up(&hdr->block_mutex); return NULL; } if (synth_alloc_pages(hw, blk)) { __snd_util_mem_free(hdr, blk); up(&hdr->block_mutex); return NULL; } up(&hdr->block_mutex); return blk; } /* * free a synth sample area */ int snd_trident_synth_free(struct snd_trident *hw, struct snd_util_memblk *blk) { struct snd_util_memhdr *hdr = hw->tlb.memhdr; down(&hdr->block_mutex); synth_free_pages(hw, blk); __snd_util_mem_free(hdr, blk); up(&hdr->block_mutex); return 0; } /* * reset TLB entry and free kernel page */ static void clear_tlb(struct snd_trident *trident, int page) { void *ptr = page_to_ptr(trident, page); dma_addr_t addr = page_to_addr(trident, page); set_silent_tlb(trident, page); if (ptr) { struct snd_dma_buffer dmab; dmab.dev.type = SNDRV_DMA_TYPE_DEV; dmab.dev.dev = snd_dma_pci_data(trident->pci); dmab.area = ptr; dmab.addr = addr; dmab.bytes = ALIGN_PAGE_SIZE; snd_dma_free_pages(&dmab); } } /* check new allocation range */ static void get_single_page_range(struct snd_util_memhdr *hdr, struct snd_util_memblk *blk, int *first_page_ret, int *last_page_ret) { struct list_head *p; struct snd_util_memblk *q; int first_page, last_page; first_page = firstpg(blk); if ((p = blk->list.prev) != &hdr->block) { q = list_entry(p, struct snd_util_memblk, list); if (lastpg(q) == first_page) first_page++; /* first page was already allocated */ } last_page = lastpg(blk); if ((p = blk->list.next) != &hdr->block) { q = list_entry(p, struct snd_util_memblk, list); if (firstpg(q) == last_page) last_page--; /* last page was already allocated */ } *first_page_ret = first_page; *last_page_ret = last_page; } /* * allocate kernel pages and assign them to TLB */ static int synth_alloc_pages(struct snd_trident *hw, struct snd_util_memblk *blk) { int page, first_page, last_page; struct snd_dma_buffer dmab; firstpg(blk) = get_aligned_page(blk->offset); lastpg(blk) = get_aligned_page(blk->offset + blk->size - 1); get_single_page_range(hw->tlb.memhdr, blk, &first_page, &last_page); /* allocate a kernel page for each Trident page - * fortunately Trident page size and kernel PAGE_SIZE is identical! */ for (page = first_page; page <= last_page; page++) { if (snd_dma_alloc_pages(SNDRV_DMA_TYPE_DEV, snd_dma_pci_data(hw->pci), ALIGN_PAGE_SIZE, &dmab) < 0) goto __fail; if (! is_valid_page(dmab.addr)) { snd_dma_free_pages(&dmab); goto __fail; } set_tlb_bus(hw, page, (unsigned long)dmab.area, dmab.addr); } return 0; __fail: /* release allocated pages */ last_page = page - 1; for (page = first_page; page <= last_page; page++) clear_tlb(hw, page); return -ENOMEM; } /* * free pages */ static int synth_free_pages(struct snd_trident *trident, struct snd_util_memblk *blk) { int page, first_page, last_page; get_single_page_range(trident->tlb.memhdr, blk, &first_page, &last_page); for (page = first_page; page <= last_page; page++) clear_tlb(trident, page); return 0; } /* * copy_from_user(blk + offset, data, size) */ int snd_trident_synth_copy_from_user(struct snd_trident *trident, struct snd_util_memblk *blk, int offset, const char __user *data, int size) { int page, nextofs, end_offset, temp, temp1; offset += blk->offset; end_offset = offset + size; page = get_aligned_page(offset) + 1; do { nextofs = aligned_page_offset(page); temp = nextofs - offset; temp1 = end_offset - offset; if (temp1 < temp) temp = temp1; if (copy_from_user(offset_ptr(trident, offset), data, temp)) return -EFAULT; offset = nextofs; data += temp; page++; } while (offset < end_offset); return 0; }