1
2
3
4
5
6
7
8
9
10#include <linux/interrupt.h>
11#include <linux/mutex.h>
12#include <linux/pci.h>
13#include <linux/io.h>
14
15#include <sound/core.h>
16#include "mixart.h"
17#include "mixart_hwdep.h"
18#include "mixart_core.h"
19
20
21#define MSG_TIMEOUT_JIFFIES (400 * HZ) / 1000
22
23#define MSG_DESCRIPTOR_SIZE 0x24
24#define MSG_HEADER_SIZE (MSG_DESCRIPTOR_SIZE + 4)
25
26#define MSG_DEFAULT_SIZE 512
27
28#define MSG_TYPE_MASK 0x00000003
29#define MSG_TYPE_NOTIFY 0
30#define MSG_TYPE_COMMAND 1
31#define MSG_TYPE_REQUEST 2
32#define MSG_TYPE_ANSWER 3
33#define MSG_CANCEL_NOTIFY_MASK 0x80000000
34
35
36static int retrieve_msg_frame(struct mixart_mgr *mgr, u32 *msg_frame)
37{
38
39 u32 headptr, tailptr;
40
41 tailptr = readl_be(MIXART_MEM(mgr, MSG_OUTBOUND_POST_TAIL));
42 headptr = readl_be(MIXART_MEM(mgr, MSG_OUTBOUND_POST_HEAD));
43
44 if (tailptr == headptr)
45 return 0;
46
47 if (tailptr < MSG_OUTBOUND_POST_STACK)
48 return 0;
49 if (tailptr >= MSG_OUTBOUND_POST_STACK + MSG_BOUND_STACK_SIZE)
50 return 0;
51
52 *msg_frame = readl_be(MIXART_MEM(mgr, tailptr));
53
54
55 tailptr += 4;
56 if( tailptr >= (MSG_OUTBOUND_POST_STACK+MSG_BOUND_STACK_SIZE) )
57 tailptr = MSG_OUTBOUND_POST_STACK;
58 writel_be(tailptr, MIXART_MEM(mgr, MSG_OUTBOUND_POST_TAIL));
59
60 return 1;
61}
62
63static int get_msg(struct mixart_mgr *mgr, struct mixart_msg *resp,
64 u32 msg_frame_address )
65{
66 u32 headptr;
67 u32 size;
68 int err;
69#ifndef __BIG_ENDIAN
70 unsigned int i;
71#endif
72
73 mutex_lock(&mgr->msg_lock);
74 err = 0;
75
76
77 size = readl_be(MIXART_MEM(mgr, msg_frame_address));
78 resp->message_id = readl_be(MIXART_MEM(mgr, msg_frame_address + 4));
79 resp->uid.object_id = readl_be(MIXART_MEM(mgr, msg_frame_address + 8));
80 resp->uid.desc = readl_be(MIXART_MEM(mgr, msg_frame_address + 12));
81
82 if( (size < MSG_DESCRIPTOR_SIZE) || (resp->size < (size - MSG_DESCRIPTOR_SIZE))) {
83 err = -EINVAL;
84 dev_err(&mgr->pci->dev,
85 "problem with response size = %d\n", size);
86 goto _clean_exit;
87 }
88 size -= MSG_DESCRIPTOR_SIZE;
89
90 memcpy_fromio(resp->data, MIXART_MEM(mgr, msg_frame_address + MSG_HEADER_SIZE ), size);
91 resp->size = size;
92
93
94#ifndef __BIG_ENDIAN
95 size /= 4;
96 for(i=0; i < size; i++) {
97 ((u32*)resp->data)[i] = be32_to_cpu(((__be32*)resp->data)[i]);
98 }
99#endif
100
101
102
103
104 headptr = readl_be(MIXART_MEM(mgr, MSG_OUTBOUND_FREE_HEAD));
105
106 if( (headptr < MSG_OUTBOUND_FREE_STACK) || ( headptr >= (MSG_OUTBOUND_FREE_STACK+MSG_BOUND_STACK_SIZE))) {
107 err = -EINVAL;
108 goto _clean_exit;
109 }
110
111
112 writel_be(msg_frame_address, MIXART_MEM(mgr, headptr));
113
114
115 headptr += 4;
116 if( headptr >= (MSG_OUTBOUND_FREE_STACK+MSG_BOUND_STACK_SIZE) )
117 headptr = MSG_OUTBOUND_FREE_STACK;
118
119 writel_be(headptr, MIXART_MEM(mgr, MSG_OUTBOUND_FREE_HEAD));
120
121 _clean_exit:
122 mutex_unlock(&mgr->msg_lock);
123
124 return err;
125}
126
127
128
129
130
131
132static int send_msg( struct mixart_mgr *mgr,
133 struct mixart_msg *msg,
134 int max_answersize,
135 int mark_pending,
136 u32 *msg_event)
137{
138 u32 headptr, tailptr;
139 u32 msg_frame_address;
140 int i;
141
142 if (snd_BUG_ON(msg->size % 4))
143 return -EINVAL;
144
145
146 tailptr = readl_be(MIXART_MEM(mgr, MSG_INBOUND_FREE_TAIL));
147 headptr = readl_be(MIXART_MEM(mgr, MSG_INBOUND_FREE_HEAD));
148
149 if (tailptr == headptr) {
150 dev_err(&mgr->pci->dev, "error: no message frame available\n");
151 return -EBUSY;
152 }
153
154 if( (tailptr < MSG_INBOUND_FREE_STACK) || (tailptr >= (MSG_INBOUND_FREE_STACK+MSG_BOUND_STACK_SIZE))) {
155 return -EINVAL;
156 }
157
158 msg_frame_address = readl_be(MIXART_MEM(mgr, tailptr));
159 writel(0, MIXART_MEM(mgr, tailptr));
160
161
162 tailptr += 4;
163 if( tailptr >= (MSG_INBOUND_FREE_STACK+MSG_BOUND_STACK_SIZE) )
164 tailptr = MSG_INBOUND_FREE_STACK;
165
166 writel_be(tailptr, MIXART_MEM(mgr, MSG_INBOUND_FREE_TAIL));
167
168
169
170
171 writel_be( msg->size + MSG_DESCRIPTOR_SIZE, MIXART_MEM(mgr, msg_frame_address) );
172 writel_be( msg->message_id , MIXART_MEM(mgr, msg_frame_address + 4) );
173 writel_be( msg->uid.object_id, MIXART_MEM(mgr, msg_frame_address + 8) );
174 writel_be( msg->uid.desc, MIXART_MEM(mgr, msg_frame_address + 12) );
175 writel_be( MSG_DESCRIPTOR_SIZE, MIXART_MEM(mgr, msg_frame_address + 16) );
176 writel_be( MSG_DESCRIPTOR_SIZE, MIXART_MEM(mgr, msg_frame_address + 20) );
177 writel_be( msg->size, MIXART_MEM(mgr, msg_frame_address + 24) );
178 writel_be( MSG_DESCRIPTOR_SIZE, MIXART_MEM(mgr, msg_frame_address + 28) );
179 writel_be( 0, MIXART_MEM(mgr, msg_frame_address + 32) );
180 writel_be( MSG_DESCRIPTOR_SIZE + max_answersize, MIXART_MEM(mgr, msg_frame_address + 36) );
181
182
183 for( i=0; i < msg->size; i+=4 ) {
184 writel_be( *(u32*)(msg->data + i), MIXART_MEM(mgr, MSG_HEADER_SIZE + msg_frame_address + i) );
185 }
186
187 if( mark_pending ) {
188 if( *msg_event ) {
189
190 mgr->pending_event = *msg_event;
191 }
192 else {
193
194 mgr->pending_event = msg_frame_address;
195
196
197 *msg_event = msg_frame_address;
198 }
199 }
200
201
202 msg_frame_address |= MSG_TYPE_REQUEST;
203
204
205 headptr = readl_be(MIXART_MEM(mgr, MSG_INBOUND_POST_HEAD));
206
207 if( (headptr < MSG_INBOUND_POST_STACK) || (headptr >= (MSG_INBOUND_POST_STACK+MSG_BOUND_STACK_SIZE))) {
208 return -EINVAL;
209 }
210
211 writel_be(msg_frame_address, MIXART_MEM(mgr, headptr));
212
213
214 headptr += 4;
215 if( headptr >= (MSG_INBOUND_POST_STACK+MSG_BOUND_STACK_SIZE) )
216 headptr = MSG_INBOUND_POST_STACK;
217
218 writel_be(headptr, MIXART_MEM(mgr, MSG_INBOUND_POST_HEAD));
219
220 return 0;
221}
222
223
224int snd_mixart_send_msg(struct mixart_mgr *mgr, struct mixart_msg *request, int max_resp_size, void *resp_data)
225{
226 struct mixart_msg resp;
227 u32 msg_frame = 0;
228 int err;
229 wait_queue_entry_t wait;
230 long timeout;
231
232 init_waitqueue_entry(&wait, current);
233
234 mutex_lock(&mgr->msg_lock);
235
236 err = send_msg(mgr, request, max_resp_size, 1, &msg_frame);
237 if (err) {
238 mutex_unlock(&mgr->msg_lock);
239 return err;
240 }
241
242 set_current_state(TASK_UNINTERRUPTIBLE);
243 add_wait_queue(&mgr->msg_sleep, &wait);
244 mutex_unlock(&mgr->msg_lock);
245 timeout = schedule_timeout(MSG_TIMEOUT_JIFFIES);
246 remove_wait_queue(&mgr->msg_sleep, &wait);
247
248 if (! timeout) {
249
250 dev_err(&mgr->pci->dev,
251 "error: no response on msg %x\n", msg_frame);
252 return -EIO;
253 }
254
255
256 resp.message_id = 0;
257 resp.uid = (struct mixart_uid){0,0};
258 resp.data = resp_data;
259 resp.size = max_resp_size;
260
261 err = get_msg(mgr, &resp, msg_frame);
262
263 if( request->message_id != resp.message_id )
264 dev_err(&mgr->pci->dev, "RESPONSE ERROR!\n");
265
266 return err;
267}
268
269
270int snd_mixart_send_msg_wait_notif(struct mixart_mgr *mgr,
271 struct mixart_msg *request, u32 notif_event)
272{
273 int err;
274 wait_queue_entry_t wait;
275 long timeout;
276
277 if (snd_BUG_ON(!notif_event))
278 return -EINVAL;
279 if (snd_BUG_ON((notif_event & MSG_TYPE_MASK) != MSG_TYPE_NOTIFY))
280 return -EINVAL;
281 if (snd_BUG_ON(notif_event & MSG_CANCEL_NOTIFY_MASK))
282 return -EINVAL;
283
284 init_waitqueue_entry(&wait, current);
285
286 mutex_lock(&mgr->msg_lock);
287
288 err = send_msg(mgr, request, MSG_DEFAULT_SIZE, 1, ¬if_event);
289 if(err) {
290 mutex_unlock(&mgr->msg_lock);
291 return err;
292 }
293
294 set_current_state(TASK_UNINTERRUPTIBLE);
295 add_wait_queue(&mgr->msg_sleep, &wait);
296 mutex_unlock(&mgr->msg_lock);
297 timeout = schedule_timeout(MSG_TIMEOUT_JIFFIES);
298 remove_wait_queue(&mgr->msg_sleep, &wait);
299
300 if (! timeout) {
301
302 dev_err(&mgr->pci->dev,
303 "error: notification %x not received\n", notif_event);
304 return -EIO;
305 }
306
307 return 0;
308}
309
310
311int snd_mixart_send_msg_nonblock(struct mixart_mgr *mgr, struct mixart_msg *request)
312{
313 u32 message_frame;
314 int err;
315
316
317 mutex_lock(&mgr->msg_lock);
318 err = send_msg(mgr, request, MSG_DEFAULT_SIZE, 0, &message_frame);
319 mutex_unlock(&mgr->msg_lock);
320
321
322 atomic_inc(&mgr->msg_processed);
323
324 return err;
325}
326
327
328
329static u32 mixart_msg_data[MSG_DEFAULT_SIZE / 4];
330
331
332static void snd_mixart_process_msg(struct mixart_mgr *mgr)
333{
334 struct mixart_msg resp;
335 u32 msg, addr, type;
336 int err;
337
338 while (mgr->msg_fifo_readptr != mgr->msg_fifo_writeptr) {
339 msg = mgr->msg_fifo[mgr->msg_fifo_readptr];
340 mgr->msg_fifo_readptr++;
341 mgr->msg_fifo_readptr %= MSG_FIFO_SIZE;
342
343
344 addr = msg & ~MSG_TYPE_MASK;
345 type = msg & MSG_TYPE_MASK;
346
347 switch (type) {
348 case MSG_TYPE_ANSWER:
349
350 resp.message_id = 0;
351 resp.data = mixart_msg_data;
352 resp.size = sizeof(mixart_msg_data);
353 err = get_msg(mgr, &resp, addr);
354 if( err < 0 ) {
355 dev_err(&mgr->pci->dev,
356 "error(%d) reading mf %x\n",
357 err, msg);
358 break;
359 }
360
361 switch(resp.message_id) {
362 case MSG_STREAM_START_INPUT_STAGE_PACKET:
363 case MSG_STREAM_START_OUTPUT_STAGE_PACKET:
364 case MSG_STREAM_STOP_INPUT_STAGE_PACKET:
365 case MSG_STREAM_STOP_OUTPUT_STAGE_PACKET:
366 if(mixart_msg_data[0])
367 dev_err(&mgr->pci->dev,
368 "error MSG_STREAM_ST***_***PUT_STAGE_PACKET status=%x\n",
369 mixart_msg_data[0]);
370 break;
371 default:
372 dev_dbg(&mgr->pci->dev,
373 "received mf(%x) : msg_id(%x) uid(%x, %x) size(%zd)\n",
374 msg, resp.message_id, resp.uid.object_id, resp.uid.desc, resp.size);
375 break;
376 }
377 break;
378 case MSG_TYPE_NOTIFY:
379
380 case MSG_TYPE_COMMAND:
381
382 default:
383 dev_err(&mgr->pci->dev,
384 "doesn't know what to do with message %x\n",
385 msg);
386 }
387
388
389 atomic_dec(&mgr->msg_processed);
390
391 }
392}
393
394
395irqreturn_t snd_mixart_interrupt(int irq, void *dev_id)
396{
397 struct mixart_mgr *mgr = dev_id;
398 u32 it_reg;
399
400 it_reg = readl_le(MIXART_REG(mgr, MIXART_PCI_OMISR_OFFSET));
401 if( !(it_reg & MIXART_OIDI) ) {
402
403 return IRQ_NONE;
404 }
405
406
407 writel_le(MIXART_HOST_ALL_INTERRUPT_MASKED, MIXART_REG(mgr, MIXART_PCI_OMIMR_OFFSET));
408
409
410 it_reg = readl(MIXART_REG(mgr, MIXART_PCI_ODBR_OFFSET));
411 writel(it_reg, MIXART_REG(mgr, MIXART_PCI_ODBR_OFFSET));
412
413
414 writel_le( MIXART_OIDI, MIXART_REG(mgr, MIXART_PCI_OMISR_OFFSET) );
415
416 return IRQ_WAKE_THREAD;
417}
418
419irqreturn_t snd_mixart_threaded_irq(int irq, void *dev_id)
420{
421 struct mixart_mgr *mgr = dev_id;
422 int err;
423 struct mixart_msg resp;
424 u32 msg;
425
426 mutex_lock(&mgr->lock);
427
428 while (retrieve_msg_frame(mgr, &msg)) {
429
430 switch (msg & MSG_TYPE_MASK) {
431 case MSG_TYPE_COMMAND:
432 resp.message_id = 0;
433 resp.data = mixart_msg_data;
434 resp.size = sizeof(mixart_msg_data);
435 err = get_msg(mgr, &resp, msg & ~MSG_TYPE_MASK);
436 if( err < 0 ) {
437 dev_err(&mgr->pci->dev,
438 "interrupt: error(%d) reading mf %x\n",
439 err, msg);
440 break;
441 }
442
443 if(resp.message_id == MSG_SERVICES_TIMER_NOTIFY) {
444 int i;
445 struct mixart_timer_notify *notify;
446 notify = (struct mixart_timer_notify *)mixart_msg_data;
447
448 for(i=0; i<notify->stream_count; i++) {
449
450 u32 buffer_id = notify->streams[i].buffer_id;
451 unsigned int chip_number = (buffer_id & MIXART_NOTIFY_CARD_MASK) >> MIXART_NOTIFY_CARD_OFFSET;
452 unsigned int pcm_number = (buffer_id & MIXART_NOTIFY_PCM_MASK ) >> MIXART_NOTIFY_PCM_OFFSET;
453 unsigned int sub_number = buffer_id & MIXART_NOTIFY_SUBS_MASK;
454 unsigned int is_capture = ((buffer_id & MIXART_NOTIFY_CAPT_MASK) != 0);
455
456 struct snd_mixart *chip = mgr->chip[chip_number];
457 struct mixart_stream *stream;
458
459 if ((chip_number >= mgr->num_cards) || (pcm_number >= MIXART_PCM_TOTAL) || (sub_number >= MIXART_PLAYBACK_STREAMS)) {
460 dev_err(&mgr->pci->dev,
461 "error MSG_SERVICES_TIMER_NOTIFY buffer_id (%x) pos(%d)\n",
462 buffer_id, notify->streams[i].sample_pos_low_part);
463 break;
464 }
465
466 if (is_capture)
467 stream = &chip->capture_stream[pcm_number];
468 else
469 stream = &chip->playback_stream[pcm_number][sub_number];
470
471 if (stream->substream && (stream->status == MIXART_STREAM_STATUS_RUNNING)) {
472 struct snd_pcm_runtime *runtime = stream->substream->runtime;
473 int elapsed = 0;
474 u64 sample_count = ((u64)notify->streams[i].sample_pos_high_part) << 32;
475 sample_count |= notify->streams[i].sample_pos_low_part;
476
477 while (1) {
478 u64 new_elapse_pos = stream->abs_period_elapsed + runtime->period_size;
479
480 if (new_elapse_pos > sample_count) {
481 break;
482 }
483 else {
484 elapsed = 1;
485 stream->buf_periods++;
486 if (stream->buf_periods >= runtime->periods)
487 stream->buf_periods = 0;
488
489 stream->abs_period_elapsed = new_elapse_pos;
490 }
491 }
492 stream->buf_period_frag = (u32)( sample_count - stream->abs_period_elapsed );
493
494 if(elapsed) {
495 mutex_unlock(&mgr->lock);
496 snd_pcm_period_elapsed(stream->substream);
497 mutex_lock(&mgr->lock);
498 }
499 }
500 }
501 break;
502 }
503 if(resp.message_id == MSG_SERVICES_REPORT_TRACES) {
504 if(resp.size > 1) {
505#ifndef __BIG_ENDIAN
506
507 int i;
508 for(i=0; i<(resp.size/4); i++) {
509 ((__be32*)mixart_msg_data)[i] = cpu_to_be32((mixart_msg_data)[i]);
510 }
511#endif
512 ((char*)mixart_msg_data)[resp.size - 1] = 0;
513 dev_dbg(&mgr->pci->dev,
514 "MIXART TRACE : %s\n",
515 (char *)mixart_msg_data);
516 }
517 break;
518 }
519
520 dev_dbg(&mgr->pci->dev, "command %x not handled\n",
521 resp.message_id);
522 break;
523
524 case MSG_TYPE_NOTIFY:
525 if(msg & MSG_CANCEL_NOTIFY_MASK) {
526 msg &= ~MSG_CANCEL_NOTIFY_MASK;
527 dev_err(&mgr->pci->dev,
528 "canceled notification %x !\n", msg);
529 }
530
531 case MSG_TYPE_ANSWER:
532
533 mutex_lock(&mgr->msg_lock);
534 if( (msg & ~MSG_TYPE_MASK) == mgr->pending_event ) {
535 wake_up(&mgr->msg_sleep);
536 mgr->pending_event = 0;
537 }
538
539 else {
540 mgr->msg_fifo[mgr->msg_fifo_writeptr] = msg;
541 mgr->msg_fifo_writeptr++;
542 mgr->msg_fifo_writeptr %= MSG_FIFO_SIZE;
543 snd_mixart_process_msg(mgr);
544 }
545 mutex_unlock(&mgr->msg_lock);
546 break;
547 case MSG_TYPE_REQUEST:
548 default:
549 dev_dbg(&mgr->pci->dev,
550 "interrupt received request %x\n", msg);
551
552 break;
553 }
554 }
555
556
557 writel_le( MIXART_ALLOW_OUTBOUND_DOORBELL, MIXART_REG( mgr, MIXART_PCI_OMIMR_OFFSET));
558
559 mutex_unlock(&mgr->lock);
560
561 return IRQ_HANDLED;
562}
563
564
565void snd_mixart_init_mailbox(struct mixart_mgr *mgr)
566{
567 writel( 0, MIXART_MEM( mgr, MSG_HOST_RSC_PROTECTION ) );
568 writel( 0, MIXART_MEM( mgr, MSG_AGENT_RSC_PROTECTION ) );
569
570
571 if(mgr->irq >= 0) {
572 writel_le( MIXART_ALLOW_OUTBOUND_DOORBELL, MIXART_REG( mgr, MIXART_PCI_OMIMR_OFFSET));
573 }
574 return;
575}
576
577void snd_mixart_exit_mailbox(struct mixart_mgr *mgr)
578{
579
580 writel_le( MIXART_HOST_ALL_INTERRUPT_MASKED, MIXART_REG( mgr, MIXART_PCI_OMIMR_OFFSET));
581 return;
582}
583
584void snd_mixart_reset_board(struct mixart_mgr *mgr)
585{
586
587 writel_be( 1, MIXART_REG(mgr, MIXART_BA1_BRUTAL_RESET_OFFSET) );
588 return;
589}
590