]> bbs.cooldavid.org Git - net-next-2.6.git/blame - sound/pci/trident/trident_memory.c
[ALSA] Changed Jaroslav Kysela's e-mail from perex@suse.cz to perex@perex.cz
[net-next-2.6.git] / sound / pci / trident / trident_memory.c
CommitLineData
1da177e4 1/*
c1017a4c 2 * Copyright (c) by Jaroslav Kysela <perex@perex.cz>
1da177e4
LT
3 * Copyright (c) by Takashi Iwai <tiwai@suse.de>
4 * Copyright (c) by Scott McNab <sdm@fractalgraphics.com.au>
5 *
6 * Trident 4DWave-NX memory page allocation (TLB area)
7 * Trident chip can handle only 16MByte of the memory at the same time.
8 *
9 *
10 * This program is free software; you can redistribute it and/or modify
11 * it under the terms of the GNU General Public License as published by
12 * the Free Software Foundation; either version 2 of the License, or
13 * (at your option) any later version.
14 *
15 * This program is distributed in the hope that it will be useful,
16 * but WITHOUT ANY WARRANTY; without even the implied warranty of
17 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
18 * GNU General Public License for more details.
19 *
20 * You should have received a copy of the GNU General Public License
21 * along with this program; if not, write to the Free Software
22 * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
23 *
24 */
25
26#include <sound/driver.h>
27#include <asm/io.h>
28#include <linux/pci.h>
29#include <linux/time.h>
62932df8
IM
30#include <linux/mutex.h>
31
1da177e4
LT
32#include <sound/core.h>
33#include <sound/trident.h>
34
35/* page arguments of these two macros are Trident page (4096 bytes), not like
36 * aligned pages in others
37 */
38#define __set_tlb_bus(trident,page,ptr,addr) \
39 do { (trident)->tlb.entries[page] = cpu_to_le32((addr) & ~(SNDRV_TRIDENT_PAGE_SIZE-1)); \
40 (trident)->tlb.shadow_entries[page] = (ptr); } while (0)
41#define __tlb_to_ptr(trident,page) \
42 (void*)((trident)->tlb.shadow_entries[page])
43#define __tlb_to_addr(trident,page) \
44 (dma_addr_t)le32_to_cpu((trident->tlb.entries[page]) & ~(SNDRV_TRIDENT_PAGE_SIZE - 1))
45
46#if PAGE_SIZE == 4096
47/* page size == SNDRV_TRIDENT_PAGE_SIZE */
48#define ALIGN_PAGE_SIZE PAGE_SIZE /* minimum page size for allocation */
49#define MAX_ALIGN_PAGES SNDRV_TRIDENT_MAX_PAGES /* maxmium aligned pages */
50/* fill TLB entrie(s) corresponding to page with ptr */
51#define set_tlb_bus(trident,page,ptr,addr) __set_tlb_bus(trident,page,ptr,addr)
52/* fill TLB entrie(s) corresponding to page with silence pointer */
53#define set_silent_tlb(trident,page) __set_tlb_bus(trident, page, (unsigned long)trident->tlb.silent_page.area, trident->tlb.silent_page.addr)
54/* get aligned page from offset address */
55#define get_aligned_page(offset) ((offset) >> 12)
56/* get offset address from aligned page */
57#define aligned_page_offset(page) ((page) << 12)
58/* get buffer address from aligned page */
59#define page_to_ptr(trident,page) __tlb_to_ptr(trident, page)
60/* get PCI physical address from aligned page */
61#define page_to_addr(trident,page) __tlb_to_addr(trident, page)
62
63#elif PAGE_SIZE == 8192
64/* page size == SNDRV_TRIDENT_PAGE_SIZE x 2*/
65#define ALIGN_PAGE_SIZE PAGE_SIZE
66#define MAX_ALIGN_PAGES (SNDRV_TRIDENT_MAX_PAGES / 2)
67#define get_aligned_page(offset) ((offset) >> 13)
68#define aligned_page_offset(page) ((page) << 13)
69#define page_to_ptr(trident,page) __tlb_to_ptr(trident, (page) << 1)
70#define page_to_addr(trident,page) __tlb_to_addr(trident, (page) << 1)
71
72/* fill TLB entries -- we need to fill two entries */
bee1a5be
TI
73static inline void set_tlb_bus(struct snd_trident *trident, int page,
74 unsigned long ptr, dma_addr_t addr)
1da177e4
LT
75{
76 page <<= 1;
77 __set_tlb_bus(trident, page, ptr, addr);
78 __set_tlb_bus(trident, page+1, ptr + SNDRV_TRIDENT_PAGE_SIZE, addr + SNDRV_TRIDENT_PAGE_SIZE);
79}
bee1a5be 80static inline void set_silent_tlb(struct snd_trident *trident, int page)
1da177e4
LT
81{
82 page <<= 1;
83 __set_tlb_bus(trident, page, (unsigned long)trident->tlb.silent_page.area, trident->tlb.silent_page.addr);
84 __set_tlb_bus(trident, page+1, (unsigned long)trident->tlb.silent_page.area, trident->tlb.silent_page.addr);
85}
86
87#else
88/* arbitrary size */
89#define UNIT_PAGES (PAGE_SIZE / SNDRV_TRIDENT_PAGE_SIZE)
90#define ALIGN_PAGE_SIZE (SNDRV_TRIDENT_PAGE_SIZE * UNIT_PAGES)
91#define MAX_ALIGN_PAGES (SNDRV_TRIDENT_MAX_PAGES / UNIT_PAGES)
92/* Note: if alignment doesn't match to the maximum size, the last few blocks
93 * become unusable. To use such blocks, you'll need to check the validity
94 * of accessing page in set_tlb_bus and set_silent_tlb. search_empty()
95 * should also check it, too.
96 */
97#define get_aligned_page(offset) ((offset) / ALIGN_PAGE_SIZE)
98#define aligned_page_offset(page) ((page) * ALIGN_PAGE_SIZE)
99#define page_to_ptr(trident,page) __tlb_to_ptr(trident, (page) * UNIT_PAGES)
100#define page_to_addr(trident,page) __tlb_to_addr(trident, (page) * UNIT_PAGES)
101
102/* fill TLB entries -- UNIT_PAGES entries must be filled */
bee1a5be
TI
103static inline void set_tlb_bus(struct snd_trident *trident, int page,
104 unsigned long ptr, dma_addr_t addr)
1da177e4
LT
105{
106 int i;
107 page *= UNIT_PAGES;
108 for (i = 0; i < UNIT_PAGES; i++, page++) {
109 __set_tlb_bus(trident, page, ptr, addr);
110 ptr += SNDRV_TRIDENT_PAGE_SIZE;
111 addr += SNDRV_TRIDENT_PAGE_SIZE;
112 }
113}
bee1a5be 114static inline void set_silent_tlb(struct snd_trident *trident, int page)
1da177e4
LT
115{
116 int i;
117 page *= UNIT_PAGES;
118 for (i = 0; i < UNIT_PAGES; i++, page++)
119 __set_tlb_bus(trident, page, (unsigned long)trident->tlb.silent_page.area, trident->tlb.silent_page.addr);
120}
121
122#endif /* PAGE_SIZE */
123
124/* calculate buffer pointer from offset address */
bee1a5be 125static inline void *offset_ptr(struct snd_trident *trident, int offset)
1da177e4
LT
126{
127 char *ptr;
128 ptr = page_to_ptr(trident, get_aligned_page(offset));
129 ptr += offset % ALIGN_PAGE_SIZE;
130 return (void*)ptr;
131}
132
133/* first and last (aligned) pages of memory block */
bee1a5be
TI
134#define firstpg(blk) (((struct snd_trident_memblk_arg *)snd_util_memblk_argptr(blk))->first_page)
135#define lastpg(blk) (((struct snd_trident_memblk_arg *)snd_util_memblk_argptr(blk))->last_page)
1da177e4
LT
136
137/*
138 * search empty pages which may contain given size
139 */
bee1a5be
TI
140static struct snd_util_memblk *
141search_empty(struct snd_util_memhdr *hdr, int size)
1da177e4 142{
bee1a5be 143 struct snd_util_memblk *blk, *prev;
1da177e4
LT
144 int page, psize;
145 struct list_head *p;
146
147 psize = get_aligned_page(size + ALIGN_PAGE_SIZE -1);
148 prev = NULL;
149 page = 0;
150 list_for_each(p, &hdr->block) {
bee1a5be 151 blk = list_entry(p, struct snd_util_memblk, list);
1da177e4
LT
152 if (page + psize <= firstpg(blk))
153 goto __found_pages;
154 page = lastpg(blk) + 1;
155 }
156 if (page + psize > MAX_ALIGN_PAGES)
157 return NULL;
158
159__found_pages:
160 /* create a new memory block */
161 blk = __snd_util_memblk_new(hdr, psize * ALIGN_PAGE_SIZE, p->prev);
162 if (blk == NULL)
163 return NULL;
164 blk->offset = aligned_page_offset(page); /* set aligned offset */
165 firstpg(blk) = page;
166 lastpg(blk) = page + psize - 1;
167 return blk;
168}
169
170
171/*
172 * check if the given pointer is valid for pages
173 */
174static int is_valid_page(unsigned long ptr)
175{
176 if (ptr & ~0x3fffffffUL) {
99b359ba 177 snd_printk(KERN_ERR "max memory size is 1GB!!\n");
1da177e4
LT
178 return 0;
179 }
180 if (ptr & (SNDRV_TRIDENT_PAGE_SIZE-1)) {
99b359ba 181 snd_printk(KERN_ERR "page is not aligned\n");
1da177e4
LT
182 return 0;
183 }
184 return 1;
185}
186
187/*
188 * page allocation for DMA (Scatter-Gather version)
189 */
bee1a5be
TI
190static struct snd_util_memblk *
191snd_trident_alloc_sg_pages(struct snd_trident *trident,
192 struct snd_pcm_substream *substream)
1da177e4 193{
bee1a5be
TI
194 struct snd_util_memhdr *hdr;
195 struct snd_util_memblk *blk;
196 struct snd_pcm_runtime *runtime = substream->runtime;
1da177e4
LT
197 int idx, page;
198 struct snd_sg_buf *sgbuf = snd_pcm_substream_sgbuf(substream);
199
200 snd_assert(runtime->dma_bytes > 0 && runtime->dma_bytes <= SNDRV_TRIDENT_MAX_PAGES * SNDRV_TRIDENT_PAGE_SIZE, return NULL);
201 hdr = trident->tlb.memhdr;
202 snd_assert(hdr != NULL, return NULL);
203
204
205
62932df8 206 mutex_lock(&hdr->block_mutex);
1da177e4
LT
207 blk = search_empty(hdr, runtime->dma_bytes);
208 if (blk == NULL) {
62932df8 209 mutex_unlock(&hdr->block_mutex);
1da177e4
LT
210 return NULL;
211 }
212 if (lastpg(blk) - firstpg(blk) >= sgbuf->pages) {
213 snd_printk(KERN_ERR "page calculation doesn't match: allocated pages = %d, trident = %d/%d\n", sgbuf->pages, firstpg(blk), lastpg(blk));
214 __snd_util_mem_free(hdr, blk);
62932df8 215 mutex_unlock(&hdr->block_mutex);
1da177e4
LT
216 return NULL;
217 }
218
219 /* set TLB entries */
220 idx = 0;
221 for (page = firstpg(blk); page <= lastpg(blk); page++, idx++) {
222 dma_addr_t addr = sgbuf->table[idx].addr;
223 unsigned long ptr = (unsigned long)sgbuf->table[idx].buf;
224 if (! is_valid_page(addr)) {
225 __snd_util_mem_free(hdr, blk);
62932df8 226 mutex_unlock(&hdr->block_mutex);
1da177e4
LT
227 return NULL;
228 }
229 set_tlb_bus(trident, page, ptr, addr);
230 }
62932df8 231 mutex_unlock(&hdr->block_mutex);
1da177e4
LT
232 return blk;
233}
234
235/*
236 * page allocation for DMA (contiguous version)
237 */
bee1a5be
TI
238static struct snd_util_memblk *
239snd_trident_alloc_cont_pages(struct snd_trident *trident,
240 struct snd_pcm_substream *substream)
1da177e4 241{
bee1a5be
TI
242 struct snd_util_memhdr *hdr;
243 struct snd_util_memblk *blk;
1da177e4 244 int page;
bee1a5be 245 struct snd_pcm_runtime *runtime = substream->runtime;
1da177e4
LT
246 dma_addr_t addr;
247 unsigned long ptr;
248
249 snd_assert(runtime->dma_bytes> 0 && runtime->dma_bytes <= SNDRV_TRIDENT_MAX_PAGES * SNDRV_TRIDENT_PAGE_SIZE, return NULL);
250 hdr = trident->tlb.memhdr;
251 snd_assert(hdr != NULL, return NULL);
252
62932df8 253 mutex_lock(&hdr->block_mutex);
1da177e4
LT
254 blk = search_empty(hdr, runtime->dma_bytes);
255 if (blk == NULL) {
62932df8 256 mutex_unlock(&hdr->block_mutex);
1da177e4
LT
257 return NULL;
258 }
259
260 /* set TLB entries */
261 addr = runtime->dma_addr;
262 ptr = (unsigned long)runtime->dma_area;
263 for (page = firstpg(blk); page <= lastpg(blk); page++,
264 ptr += SNDRV_TRIDENT_PAGE_SIZE, addr += SNDRV_TRIDENT_PAGE_SIZE) {
265 if (! is_valid_page(addr)) {
266 __snd_util_mem_free(hdr, blk);
62932df8 267 mutex_unlock(&hdr->block_mutex);
1da177e4
LT
268 return NULL;
269 }
270 set_tlb_bus(trident, page, ptr, addr);
271 }
62932df8 272 mutex_unlock(&hdr->block_mutex);
1da177e4
LT
273 return blk;
274}
275
276/*
277 * page allocation for DMA
278 */
bee1a5be
TI
279struct snd_util_memblk *
280snd_trident_alloc_pages(struct snd_trident *trident,
281 struct snd_pcm_substream *substream)
1da177e4
LT
282{
283 snd_assert(trident != NULL, return NULL);
284 snd_assert(substream != NULL, return NULL);
285 if (substream->dma_buffer.dev.type == SNDRV_DMA_TYPE_DEV_SG)
286 return snd_trident_alloc_sg_pages(trident, substream);
287 else
288 return snd_trident_alloc_cont_pages(trident, substream);
289}
290
291
292/*
293 * release DMA buffer from page table
294 */
bee1a5be
TI
295int snd_trident_free_pages(struct snd_trident *trident,
296 struct snd_util_memblk *blk)
1da177e4 297{
bee1a5be 298 struct snd_util_memhdr *hdr;
1da177e4
LT
299 int page;
300
301 snd_assert(trident != NULL, return -EINVAL);
302 snd_assert(blk != NULL, return -EINVAL);
303
304 hdr = trident->tlb.memhdr;
62932df8 305 mutex_lock(&hdr->block_mutex);
1da177e4
LT
306 /* reset TLB entries */
307 for (page = firstpg(blk); page <= lastpg(blk); page++)
308 set_silent_tlb(trident, page);
309 /* free memory block */
310 __snd_util_mem_free(hdr, blk);
62932df8 311 mutex_unlock(&hdr->block_mutex);
1da177e4
LT
312 return 0;
313}
314
315
316/*----------------------------------------------------------------
317 * memory allocation using multiple pages (for synth)
318 *----------------------------------------------------------------
319 * Unlike the DMA allocation above, non-contiguous pages are
320 * assigned to TLB.
321 *----------------------------------------------------------------*/
322
323/*
324 */
bee1a5be
TI
325static int synth_alloc_pages(struct snd_trident *hw, struct snd_util_memblk *blk);
326static int synth_free_pages(struct snd_trident *hw, struct snd_util_memblk *blk);
1da177e4
LT
327
328/*
329 * allocate a synth sample area
330 */
bee1a5be
TI
331struct snd_util_memblk *
332snd_trident_synth_alloc(struct snd_trident *hw, unsigned int size)
1da177e4 333{
bee1a5be
TI
334 struct snd_util_memblk *blk;
335 struct snd_util_memhdr *hdr = hw->tlb.memhdr;
1da177e4 336
62932df8 337 mutex_lock(&hdr->block_mutex);
1da177e4
LT
338 blk = __snd_util_mem_alloc(hdr, size);
339 if (blk == NULL) {
62932df8 340 mutex_unlock(&hdr->block_mutex);
1da177e4
LT
341 return NULL;
342 }
343 if (synth_alloc_pages(hw, blk)) {
344 __snd_util_mem_free(hdr, blk);
62932df8 345 mutex_unlock(&hdr->block_mutex);
1da177e4
LT
346 return NULL;
347 }
62932df8 348 mutex_unlock(&hdr->block_mutex);
1da177e4
LT
349 return blk;
350}
351
cbef55f3 352EXPORT_SYMBOL(snd_trident_synth_alloc);
1da177e4
LT
353
354/*
355 * free a synth sample area
356 */
357int
bee1a5be 358snd_trident_synth_free(struct snd_trident *hw, struct snd_util_memblk *blk)
1da177e4 359{
bee1a5be 360 struct snd_util_memhdr *hdr = hw->tlb.memhdr;
1da177e4 361
62932df8 362 mutex_lock(&hdr->block_mutex);
1da177e4
LT
363 synth_free_pages(hw, blk);
364 __snd_util_mem_free(hdr, blk);
62932df8 365 mutex_unlock(&hdr->block_mutex);
1da177e4
LT
366 return 0;
367}
368
cbef55f3 369EXPORT_SYMBOL(snd_trident_synth_free);
1da177e4
LT
370
371/*
372 * reset TLB entry and free kernel page
373 */
bee1a5be 374static void clear_tlb(struct snd_trident *trident, int page)
1da177e4
LT
375{
376 void *ptr = page_to_ptr(trident, page);
377 dma_addr_t addr = page_to_addr(trident, page);
378 set_silent_tlb(trident, page);
379 if (ptr) {
380 struct snd_dma_buffer dmab;
381 dmab.dev.type = SNDRV_DMA_TYPE_DEV;
382 dmab.dev.dev = snd_dma_pci_data(trident->pci);
383 dmab.area = ptr;
384 dmab.addr = addr;
385 dmab.bytes = ALIGN_PAGE_SIZE;
386 snd_dma_free_pages(&dmab);
387 }
388}
389
390/* check new allocation range */
bee1a5be
TI
391static void get_single_page_range(struct snd_util_memhdr *hdr,
392 struct snd_util_memblk *blk,
393 int *first_page_ret, int *last_page_ret)
1da177e4
LT
394{
395 struct list_head *p;
bee1a5be 396 struct snd_util_memblk *q;
1da177e4
LT
397 int first_page, last_page;
398 first_page = firstpg(blk);
399 if ((p = blk->list.prev) != &hdr->block) {
bee1a5be 400 q = list_entry(p, struct snd_util_memblk, list);
1da177e4
LT
401 if (lastpg(q) == first_page)
402 first_page++; /* first page was already allocated */
403 }
404 last_page = lastpg(blk);
405 if ((p = blk->list.next) != &hdr->block) {
bee1a5be 406 q = list_entry(p, struct snd_util_memblk, list);
1da177e4
LT
407 if (firstpg(q) == last_page)
408 last_page--; /* last page was already allocated */
409 }
410 *first_page_ret = first_page;
411 *last_page_ret = last_page;
412}
413
414/*
415 * allocate kernel pages and assign them to TLB
416 */
bee1a5be 417static int synth_alloc_pages(struct snd_trident *hw, struct snd_util_memblk *blk)
1da177e4
LT
418{
419 int page, first_page, last_page;
420 struct snd_dma_buffer dmab;
421
422 firstpg(blk) = get_aligned_page(blk->offset);
423 lastpg(blk) = get_aligned_page(blk->offset + blk->size - 1);
424 get_single_page_range(hw->tlb.memhdr, blk, &first_page, &last_page);
425
426 /* allocate a kernel page for each Trident page -
427 * fortunately Trident page size and kernel PAGE_SIZE is identical!
428 */
429 for (page = first_page; page <= last_page; page++) {
430 if (snd_dma_alloc_pages(SNDRV_DMA_TYPE_DEV, snd_dma_pci_data(hw->pci),
431 ALIGN_PAGE_SIZE, &dmab) < 0)
432 goto __fail;
433 if (! is_valid_page(dmab.addr)) {
434 snd_dma_free_pages(&dmab);
435 goto __fail;
436 }
437 set_tlb_bus(hw, page, (unsigned long)dmab.area, dmab.addr);
438 }
439 return 0;
440
441__fail:
442 /* release allocated pages */
443 last_page = page - 1;
444 for (page = first_page; page <= last_page; page++)
445 clear_tlb(hw, page);
446
447 return -ENOMEM;
448}
449
450/*
451 * free pages
452 */
bee1a5be 453static int synth_free_pages(struct snd_trident *trident, struct snd_util_memblk *blk)
1da177e4
LT
454{
455 int page, first_page, last_page;
456
457 get_single_page_range(trident->tlb.memhdr, blk, &first_page, &last_page);
458 for (page = first_page; page <= last_page; page++)
459 clear_tlb(trident, page);
460
461 return 0;
462}
463
464/*
465 * copy_from_user(blk + offset, data, size)
466 */
bee1a5be
TI
467int snd_trident_synth_copy_from_user(struct snd_trident *trident,
468 struct snd_util_memblk *blk,
469 int offset, const char __user *data, int size)
1da177e4
LT
470{
471 int page, nextofs, end_offset, temp, temp1;
472
473 offset += blk->offset;
474 end_offset = offset + size;
475 page = get_aligned_page(offset) + 1;
476 do {
477 nextofs = aligned_page_offset(page);
478 temp = nextofs - offset;
479 temp1 = end_offset - offset;
480 if (temp1 < temp)
481 temp = temp1;
482 if (copy_from_user(offset_ptr(trident, offset), data, temp))
483 return -EFAULT;
484 offset = nextofs;
485 data += temp;
486 page++;
487 } while (offset < end_offset);
488 return 0;
489}
490
cbef55f3 491EXPORT_SYMBOL(snd_trident_synth_copy_from_user);