1
2
3
4
5
6
7
8
9
10
11
12
13#define KMSG_COMPONENT "appldata"
14#define pr_fmt(fmt) KMSG_COMPONENT ": " fmt
15
16#include <linux/module.h>
17#include <linux/init.h>
18#include <linux/slab.h>
19#include <linux/errno.h>
20#include <linux/interrupt.h>
21#include <linux/proc_fs.h>
22#include <linux/mm.h>
23#include <linux/swap.h>
24#include <linux/pagemap.h>
25#include <linux/sysctl.h>
26#include <linux/notifier.h>
27#include <linux/cpu.h>
28#include <linux/workqueue.h>
29#include <linux/suspend.h>
30#include <linux/platform_device.h>
31#include <asm/appldata.h>
32#include <asm/timer.h>
33#include <asm/uaccess.h>
34#include <asm/io.h>
35#include <asm/smp.h>
36
37#include "appldata.h"
38
39
40#define APPLDATA_CPU_INTERVAL 10000
41
42
43
44#define TOD_MICRO 0x01000
45
46
47static struct platform_device *appldata_pdev;
48
49
50
51
52static const char appldata_proc_name[APPLDATA_PROC_NAME_LENGTH] = "appldata";
53static int appldata_timer_handler(ctl_table *ctl, int write,
54 void __user *buffer, size_t *lenp, loff_t *ppos);
55static int appldata_interval_handler(ctl_table *ctl, int write,
56 void __user *buffer,
57 size_t *lenp, loff_t *ppos);
58
59static struct ctl_table_header *appldata_sysctl_header;
60static struct ctl_table appldata_table[] = {
61 {
62 .procname = "timer",
63 .mode = S_IRUGO | S_IWUSR,
64 .proc_handler = &appldata_timer_handler,
65 },
66 {
67 .procname = "interval",
68 .mode = S_IRUGO | S_IWUSR,
69 .proc_handler = &appldata_interval_handler,
70 },
71 { },
72};
73
74static struct ctl_table appldata_dir_table[] = {
75 {
76 .procname = appldata_proc_name,
77 .maxlen = 0,
78 .mode = S_IRUGO | S_IXUGO,
79 .child = appldata_table,
80 },
81 { },
82};
83
84
85
86
87static DEFINE_PER_CPU(struct vtimer_list, appldata_timer);
88static atomic_t appldata_expire_count = ATOMIC_INIT(0);
89
90static DEFINE_SPINLOCK(appldata_timer_lock);
91static int appldata_interval = APPLDATA_CPU_INTERVAL;
92static int appldata_timer_active;
93static int appldata_timer_suspended = 0;
94
95
96
97
98static struct workqueue_struct *appldata_wq;
99static void appldata_work_fn(struct work_struct *work);
100static DECLARE_WORK(appldata_work, appldata_work_fn);
101
102
103
104
105
106static DEFINE_MUTEX(appldata_ops_mutex);
107static LIST_HEAD(appldata_ops_list);
108
109
110
111
112
113
114
115
116static void appldata_timer_function(unsigned long data)
117{
118 if (atomic_dec_and_test(&appldata_expire_count)) {
119 atomic_set(&appldata_expire_count, num_online_cpus());
120 queue_work(appldata_wq, (struct work_struct *) data);
121 }
122}
123
124
125
126
127
128
129static void appldata_work_fn(struct work_struct *work)
130{
131 struct list_head *lh;
132 struct appldata_ops *ops;
133 int i;
134
135 i = 0;
136 get_online_cpus();
137 mutex_lock(&appldata_ops_mutex);
138 list_for_each(lh, &appldata_ops_list) {
139 ops = list_entry(lh, struct appldata_ops, list);
140 if (ops->active == 1) {
141 ops->callback(ops->data);
142 }
143 }
144 mutex_unlock(&appldata_ops_mutex);
145 put_online_cpus();
146}
147
148
149
150
151
152
153int appldata_diag(char record_nr, u16 function, unsigned long buffer,
154 u16 length, char *mod_lvl)
155{
156 struct appldata_product_id id = {
157 .prod_nr = {0xD3, 0xC9, 0xD5, 0xE4,
158 0xE7, 0xD2, 0xD9},
159 .prod_fn = 0xD5D3,
160 .version_nr = 0xF2F6,
161 .release_nr = 0xF0F1,
162 };
163
164 id.record_nr = record_nr;
165 id.mod_lvl = (mod_lvl[0]) << 8 | mod_lvl[1];
166 return appldata_asm(&id, function, (void *) buffer, length);
167}
168
169
170
171
172
173
174
175
176
177
178
179static void __appldata_mod_vtimer_wrap(void *p) {
180 struct {
181 struct vtimer_list *timer;
182 u64 expires;
183 } *args = p;
184 mod_virt_timer_periodic(args->timer, args->expires);
185}
186
187#define APPLDATA_ADD_TIMER 0
188#define APPLDATA_DEL_TIMER 1
189#define APPLDATA_MOD_TIMER 2
190
191
192
193
194
195
196
197static void
198__appldata_vtimer_setup(int cmd)
199{
200 u64 per_cpu_interval;
201 int i;
202
203 switch (cmd) {
204 case APPLDATA_ADD_TIMER:
205 if (appldata_timer_active)
206 break;
207 per_cpu_interval = (u64) (appldata_interval*1000 /
208 num_online_cpus()) * TOD_MICRO;
209 for_each_online_cpu(i) {
210 per_cpu(appldata_timer, i).expires = per_cpu_interval;
211 smp_call_function_single(i, add_virt_timer_periodic,
212 &per_cpu(appldata_timer, i),
213 1);
214 }
215 appldata_timer_active = 1;
216 break;
217 case APPLDATA_DEL_TIMER:
218 for_each_online_cpu(i)
219 del_virt_timer(&per_cpu(appldata_timer, i));
220 if (!appldata_timer_active)
221 break;
222 appldata_timer_active = 0;
223 atomic_set(&appldata_expire_count, num_online_cpus());
224 break;
225 case APPLDATA_MOD_TIMER:
226 per_cpu_interval = (u64) (appldata_interval*1000 /
227 num_online_cpus()) * TOD_MICRO;
228 if (!appldata_timer_active)
229 break;
230 for_each_online_cpu(i) {
231 struct {
232 struct vtimer_list *timer;
233 u64 expires;
234 } args;
235 args.timer = &per_cpu(appldata_timer, i);
236 args.expires = per_cpu_interval;
237 smp_call_function_single(i, __appldata_mod_vtimer_wrap,
238 &args, 1);
239 }
240 }
241}
242
243
244
245
246
247
248static int
249appldata_timer_handler(ctl_table *ctl, int write,
250 void __user *buffer, size_t *lenp, loff_t *ppos)
251{
252 int len;
253 char buf[2];
254
255 if (!*lenp || *ppos) {
256 *lenp = 0;
257 return 0;
258 }
259 if (!write) {
260 len = sprintf(buf, appldata_timer_active ? "1\n" : "0\n");
261 if (len > *lenp)
262 len = *lenp;
263 if (copy_to_user(buffer, buf, len))
264 return -EFAULT;
265 goto out;
266 }
267 len = *lenp;
268 if (copy_from_user(buf, buffer, len > sizeof(buf) ? sizeof(buf) : len))
269 return -EFAULT;
270 get_online_cpus();
271 spin_lock(&appldata_timer_lock);
272 if (buf[0] == '1')
273 __appldata_vtimer_setup(APPLDATA_ADD_TIMER);
274 else if (buf[0] == '0')
275 __appldata_vtimer_setup(APPLDATA_DEL_TIMER);
276 spin_unlock(&appldata_timer_lock);
277 put_online_cpus();
278out:
279 *lenp = len;
280 *ppos += len;
281 return 0;
282}
283
284
285
286
287
288
289
290static int
291appldata_interval_handler(ctl_table *ctl, int write,
292 void __user *buffer, size_t *lenp, loff_t *ppos)
293{
294 int len, interval;
295 char buf[16];
296
297 if (!*lenp || *ppos) {
298 *lenp = 0;
299 return 0;
300 }
301 if (!write) {
302 len = sprintf(buf, "%i\n", appldata_interval);
303 if (len > *lenp)
304 len = *lenp;
305 if (copy_to_user(buffer, buf, len))
306 return -EFAULT;
307 goto out;
308 }
309 len = *lenp;
310 if (copy_from_user(buf, buffer, len > sizeof(buf) ? sizeof(buf) : len)) {
311 return -EFAULT;
312 }
313 interval = 0;
314 sscanf(buf, "%i", &interval);
315 if (interval <= 0)
316 return -EINVAL;
317
318 get_online_cpus();
319 spin_lock(&appldata_timer_lock);
320 appldata_interval = interval;
321 __appldata_vtimer_setup(APPLDATA_MOD_TIMER);
322 spin_unlock(&appldata_timer_lock);
323 put_online_cpus();
324out:
325 *lenp = len;
326 *ppos += len;
327 return 0;
328}
329
330
331
332
333
334
335
336static int
337appldata_generic_handler(ctl_table *ctl, int write,
338 void __user *buffer, size_t *lenp, loff_t *ppos)
339{
340 struct appldata_ops *ops = NULL, *tmp_ops;
341 int rc, len, found;
342 char buf[2];
343 struct list_head *lh;
344
345 found = 0;
346 mutex_lock(&appldata_ops_mutex);
347 list_for_each(lh, &appldata_ops_list) {
348 tmp_ops = list_entry(lh, struct appldata_ops, list);
349 if (&tmp_ops->ctl_table[2] == ctl) {
350 found = 1;
351 }
352 }
353 if (!found) {
354 mutex_unlock(&appldata_ops_mutex);
355 return -ENODEV;
356 }
357 ops = ctl->data;
358 if (!try_module_get(ops->owner)) {
359 mutex_unlock(&appldata_ops_mutex);
360 return -ENODEV;
361 }
362 mutex_unlock(&appldata_ops_mutex);
363
364 if (!*lenp || *ppos) {
365 *lenp = 0;
366 module_put(ops->owner);
367 return 0;
368 }
369 if (!write) {
370 len = sprintf(buf, ops->active ? "1\n" : "0\n");
371 if (len > *lenp)
372 len = *lenp;
373 if (copy_to_user(buffer, buf, len)) {
374 module_put(ops->owner);
375 return -EFAULT;
376 }
377 goto out;
378 }
379 len = *lenp;
380 if (copy_from_user(buf, buffer,
381 len > sizeof(buf) ? sizeof(buf) : len)) {
382 module_put(ops->owner);
383 return -EFAULT;
384 }
385
386 mutex_lock(&appldata_ops_mutex);
387 if ((buf[0] == '1') && (ops->active == 0)) {
388
389 if (!try_module_get(ops->owner)) {
390 mutex_unlock(&appldata_ops_mutex);
391 module_put(ops->owner);
392 return -ENODEV;
393 }
394 ops->callback(ops->data);
395 rc = appldata_diag(ops->record_nr,
396 APPLDATA_START_INTERVAL_REC,
397 (unsigned long) ops->data, ops->size,
398 ops->mod_lvl);
399 if (rc != 0) {
400 pr_err("Starting the data collection for %s "
401 "failed with rc=%d\n", ops->name, rc);
402 module_put(ops->owner);
403 } else
404 ops->active = 1;
405 } else if ((buf[0] == '0') && (ops->active == 1)) {
406 ops->active = 0;
407 rc = appldata_diag(ops->record_nr, APPLDATA_STOP_REC,
408 (unsigned long) ops->data, ops->size,
409 ops->mod_lvl);
410 if (rc != 0)
411 pr_err("Stopping the data collection for %s "
412 "failed with rc=%d\n", ops->name, rc);
413 module_put(ops->owner);
414 }
415 mutex_unlock(&appldata_ops_mutex);
416out:
417 *lenp = len;
418 *ppos += len;
419 module_put(ops->owner);
420 return 0;
421}
422
423
424
425
426
427
428
429
430
431
432int appldata_register_ops(struct appldata_ops *ops)
433{
434 if (ops->size > APPLDATA_MAX_REC_SIZE)
435 return -EINVAL;
436
437 ops->ctl_table = kzalloc(4 * sizeof(struct ctl_table), GFP_KERNEL);
438 if (!ops->ctl_table)
439 return -ENOMEM;
440
441 mutex_lock(&appldata_ops_mutex);
442 list_add(&ops->list, &appldata_ops_list);
443 mutex_unlock(&appldata_ops_mutex);
444
445 ops->ctl_table[0].procname = appldata_proc_name;
446 ops->ctl_table[0].maxlen = 0;
447 ops->ctl_table[0].mode = S_IRUGO | S_IXUGO;
448 ops->ctl_table[0].child = &ops->ctl_table[2];
449
450 ops->ctl_table[2].procname = ops->name;
451 ops->ctl_table[2].mode = S_IRUGO | S_IWUSR;
452 ops->ctl_table[2].proc_handler = appldata_generic_handler;
453 ops->ctl_table[2].data = ops;
454
455 ops->sysctl_header = register_sysctl_table(ops->ctl_table);
456 if (!ops->sysctl_header)
457 goto out;
458 return 0;
459out:
460 mutex_lock(&appldata_ops_mutex);
461 list_del(&ops->list);
462 mutex_unlock(&appldata_ops_mutex);
463 kfree(ops->ctl_table);
464 return -ENOMEM;
465}
466
467
468
469
470
471
472void appldata_unregister_ops(struct appldata_ops *ops)
473{
474 mutex_lock(&appldata_ops_mutex);
475 list_del(&ops->list);
476 mutex_unlock(&appldata_ops_mutex);
477 unregister_sysctl_table(ops->sysctl_header);
478 kfree(ops->ctl_table);
479}
480
481
482
483
484static int appldata_freeze(struct device *dev)
485{
486 struct appldata_ops *ops;
487 int rc;
488 struct list_head *lh;
489
490 get_online_cpus();
491 spin_lock(&appldata_timer_lock);
492 if (appldata_timer_active) {
493 __appldata_vtimer_setup(APPLDATA_DEL_TIMER);
494 appldata_timer_suspended = 1;
495 }
496 spin_unlock(&appldata_timer_lock);
497 put_online_cpus();
498
499 mutex_lock(&appldata_ops_mutex);
500 list_for_each(lh, &appldata_ops_list) {
501 ops = list_entry(lh, struct appldata_ops, list);
502 if (ops->active == 1) {
503 rc = appldata_diag(ops->record_nr, APPLDATA_STOP_REC,
504 (unsigned long) ops->data, ops->size,
505 ops->mod_lvl);
506 if (rc != 0)
507 pr_err("Stopping the data collection for %s "
508 "failed with rc=%d\n", ops->name, rc);
509 }
510 }
511 mutex_unlock(&appldata_ops_mutex);
512 return 0;
513}
514
515static int appldata_restore(struct device *dev)
516{
517 struct appldata_ops *ops;
518 int rc;
519 struct list_head *lh;
520
521 get_online_cpus();
522 spin_lock(&appldata_timer_lock);
523 if (appldata_timer_suspended) {
524 __appldata_vtimer_setup(APPLDATA_ADD_TIMER);
525 appldata_timer_suspended = 0;
526 }
527 spin_unlock(&appldata_timer_lock);
528 put_online_cpus();
529
530 mutex_lock(&appldata_ops_mutex);
531 list_for_each(lh, &appldata_ops_list) {
532 ops = list_entry(lh, struct appldata_ops, list);
533 if (ops->active == 1) {
534 ops->callback(ops->data);
535 rc = appldata_diag(ops->record_nr,
536 APPLDATA_START_INTERVAL_REC,
537 (unsigned long) ops->data, ops->size,
538 ops->mod_lvl);
539 if (rc != 0) {
540 pr_err("Starting the data collection for %s "
541 "failed with rc=%d\n", ops->name, rc);
542 }
543 }
544 }
545 mutex_unlock(&appldata_ops_mutex);
546 return 0;
547}
548
549static int appldata_thaw(struct device *dev)
550{
551 return appldata_restore(dev);
552}
553
554static struct dev_pm_ops appldata_pm_ops = {
555 .freeze = appldata_freeze,
556 .thaw = appldata_thaw,
557 .restore = appldata_restore,
558};
559
560static struct platform_driver appldata_pdrv = {
561 .driver = {
562 .name = "appldata",
563 .owner = THIS_MODULE,
564 .pm = &appldata_pm_ops,
565 },
566};
567
568
569
570
571
572static void __cpuinit appldata_online_cpu(int cpu)
573{
574 init_virt_timer(&per_cpu(appldata_timer, cpu));
575 per_cpu(appldata_timer, cpu).function = appldata_timer_function;
576 per_cpu(appldata_timer, cpu).data = (unsigned long)
577 &appldata_work;
578 atomic_inc(&appldata_expire_count);
579 spin_lock(&appldata_timer_lock);
580 __appldata_vtimer_setup(APPLDATA_MOD_TIMER);
581 spin_unlock(&appldata_timer_lock);
582}
583
584static void __cpuinit appldata_offline_cpu(int cpu)
585{
586 del_virt_timer(&per_cpu(appldata_timer, cpu));
587 if (atomic_dec_and_test(&appldata_expire_count)) {
588 atomic_set(&appldata_expire_count, num_online_cpus());
589 queue_work(appldata_wq, &appldata_work);
590 }
591 spin_lock(&appldata_timer_lock);
592 __appldata_vtimer_setup(APPLDATA_MOD_TIMER);
593 spin_unlock(&appldata_timer_lock);
594}
595
596static int __cpuinit appldata_cpu_notify(struct notifier_block *self,
597 unsigned long action,
598 void *hcpu)
599{
600 switch (action) {
601 case CPU_ONLINE:
602 case CPU_ONLINE_FROZEN:
603 appldata_online_cpu((long) hcpu);
604 break;
605 case CPU_DEAD:
606 case CPU_DEAD_FROZEN:
607 appldata_offline_cpu((long) hcpu);
608 break;
609 default:
610 break;
611 }
612 return NOTIFY_OK;
613}
614
615static struct notifier_block __cpuinitdata appldata_nb = {
616 .notifier_call = appldata_cpu_notify,
617};
618
619
620
621
622
623
624static int __init appldata_init(void)
625{
626 int i, rc;
627
628 rc = platform_driver_register(&appldata_pdrv);
629 if (rc)
630 return rc;
631
632 appldata_pdev = platform_device_register_simple("appldata", -1, NULL,
633 0);
634 if (IS_ERR(appldata_pdev)) {
635 rc = PTR_ERR(appldata_pdev);
636 goto out_driver;
637 }
638 appldata_wq = create_singlethread_workqueue("appldata");
639 if (!appldata_wq) {
640 rc = -ENOMEM;
641 goto out_device;
642 }
643
644 get_online_cpus();
645 for_each_online_cpu(i)
646 appldata_online_cpu(i);
647 put_online_cpus();
648
649
650 register_hotcpu_notifier(&appldata_nb);
651
652 appldata_sysctl_header = register_sysctl_table(appldata_dir_table);
653 return 0;
654
655out_device:
656 platform_device_unregister(appldata_pdev);
657out_driver:
658 platform_driver_unregister(&appldata_pdrv);
659 return rc;
660}
661
662__initcall(appldata_init);
663
664
665
666EXPORT_SYMBOL_GPL(appldata_register_ops);
667EXPORT_SYMBOL_GPL(appldata_unregister_ops);
668EXPORT_SYMBOL_GPL(appldata_diag);
669
670#ifdef CONFIG_SWAP
671EXPORT_SYMBOL_GPL(si_swapinfo);
672#endif
673EXPORT_SYMBOL_GPL(nr_threads);
674EXPORT_SYMBOL_GPL(nr_running);
675EXPORT_SYMBOL_GPL(nr_iowait);
676