1
2
3
4
5
6#include <stdio.h>
7#include <stdlib.h>
8#include <stdint.h>
9#include <string.h>
10#include <stdarg.h>
11#include <unistd.h>
12#include <pthread.h>
13#include <syslog.h>
14#include <getopt.h>
15#include <sys/file.h>
16#include <stddef.h>
17#include <errno.h>
18#include <limits.h>
19#include <sys/mman.h>
20#include <sys/queue.h>
21#include <sys/stat.h>
22
23#include <rte_compat.h>
24#include <rte_common.h>
25#include <rte_debug.h>
26#include <rte_memory.h>
27#include <rte_launch.h>
28#include <rte_eal.h>
29#include <rte_errno.h>
30#include <rte_per_lcore.h>
31#include <rte_lcore.h>
32#include <rte_service_component.h>
33#include <rte_log.h>
34#include <rte_random.h>
35#include <rte_cycles.h>
36#include <rte_string_fns.h>
37#include <rte_cpuflags.h>
38#include <rte_interrupts.h>
39#include <rte_bus.h>
40#include <rte_dev.h>
41#include <rte_devargs.h>
42#include <rte_version.h>
43#include <rte_vfio.h>
44#include <malloc_heap.h>
45#include <telemetry_internal.h>
46
47#include "eal_private.h"
48#include "eal_thread.h"
49#include "eal_internal_cfg.h"
50#include "eal_filesystem.h"
51#include "eal_hugepages.h"
52#include "eal_options.h"
53#include "eal_memcfg.h"
54#include "eal_trace.h"
55
56#define MEMSIZE_IF_NO_HUGE_PAGE (64ULL * 1024ULL * 1024ULL)
57
58
59
60static int mem_cfg_fd = -1;
61
62static struct flock wr_lock = {
63 .l_type = F_WRLCK,
64 .l_whence = SEEK_SET,
65 .l_start = offsetof(struct rte_mem_config, memsegs),
66 .l_len = RTE_SIZEOF_FIELD(struct rte_mem_config, memsegs),
67};
68
69
70struct lcore_config lcore_config[RTE_MAX_LCORE];
71
72
73int rte_cycles_vmware_tsc_map;
74
75static const char *default_runtime_dir = "/var/run";
76
77int
78eal_create_runtime_dir(void)
79{
80 const char *directory = default_runtime_dir;
81 const char *xdg_runtime_dir = getenv("XDG_RUNTIME_DIR");
82 const char *fallback = "/tmp";
83 char run_dir[PATH_MAX];
84 char tmp[PATH_MAX];
85 int ret;
86
87 if (getuid() != 0) {
88
89 if (xdg_runtime_dir != NULL)
90 directory = xdg_runtime_dir;
91 else
92 directory = fallback;
93 }
94
95 ret = snprintf(tmp, sizeof(tmp), "%s/dpdk", directory);
96 if (ret < 0 || ret == sizeof(tmp)) {
97 RTE_LOG(ERR, EAL, "Error creating DPDK runtime path name\n");
98 return -1;
99 }
100
101
102 ret = snprintf(run_dir, sizeof(run_dir), "%s/%s",
103 tmp, eal_get_hugefile_prefix());
104 if (ret < 0 || ret == sizeof(run_dir)) {
105 RTE_LOG(ERR, EAL, "Error creating prefix-specific runtime path name\n");
106 return -1;
107 }
108
109
110
111
112 ret = mkdir(tmp, 0700);
113 if (ret < 0 && errno != EEXIST) {
114 RTE_LOG(ERR, EAL, "Error creating '%s': %s\n",
115 tmp, strerror(errno));
116 return -1;
117 }
118
119 ret = mkdir(run_dir, 0700);
120 if (ret < 0 && errno != EEXIST) {
121 RTE_LOG(ERR, EAL, "Error creating '%s': %s\n",
122 run_dir, strerror(errno));
123 return -1;
124 }
125
126 if (eal_set_runtime_dir(run_dir, sizeof(run_dir)))
127 return -1;
128
129 return 0;
130}
131
132int
133eal_clean_runtime_dir(void)
134{
135
136
137
138 return 0;
139}
140
141
142int
143eal_parse_sysfs_value(const char *filename, unsigned long *val)
144{
145 FILE *f;
146 char buf[BUFSIZ];
147 char *end = NULL;
148
149 if ((f = fopen(filename, "r")) == NULL) {
150 RTE_LOG(ERR, EAL, "%s(): cannot open sysfs value %s\n",
151 __func__, filename);
152 return -1;
153 }
154
155 if (fgets(buf, sizeof(buf), f) == NULL) {
156 RTE_LOG(ERR, EAL, "%s(): cannot read sysfs value %s\n",
157 __func__, filename);
158 fclose(f);
159 return -1;
160 }
161 *val = strtoul(buf, &end, 0);
162 if ((buf[0] == '\0') || (end == NULL) || (*end != '\n')) {
163 RTE_LOG(ERR, EAL, "%s(): cannot parse sysfs value %s\n",
164 __func__, filename);
165 fclose(f);
166 return -1;
167 }
168 fclose(f);
169 return 0;
170}
171
172
173
174
175
176
177
178
179static int
180rte_eal_config_create(void)
181{
182 struct rte_config *config = rte_eal_get_configuration();
183 const struct internal_config *internal_conf =
184 eal_get_internal_configuration();
185 size_t page_sz = sysconf(_SC_PAGE_SIZE);
186 size_t cfg_len = sizeof(struct rte_mem_config);
187 size_t cfg_len_aligned = RTE_ALIGN(cfg_len, page_sz);
188 void *rte_mem_cfg_addr, *mapped_mem_cfg_addr;
189 int retval;
190
191 const char *pathname = eal_runtime_config_path();
192
193 if (internal_conf->no_shconf)
194 return 0;
195
196
197 if (internal_conf->base_virtaddr != 0)
198 rte_mem_cfg_addr = (void *)
199 RTE_ALIGN_FLOOR(internal_conf->base_virtaddr -
200 sizeof(struct rte_mem_config), page_sz);
201 else
202 rte_mem_cfg_addr = NULL;
203
204 if (mem_cfg_fd < 0){
205 mem_cfg_fd = open(pathname, O_RDWR | O_CREAT, 0600);
206 if (mem_cfg_fd < 0) {
207 RTE_LOG(ERR, EAL, "Cannot open '%s' for rte_mem_config\n",
208 pathname);
209 return -1;
210 }
211 }
212
213 retval = ftruncate(mem_cfg_fd, cfg_len);
214 if (retval < 0){
215 close(mem_cfg_fd);
216 mem_cfg_fd = -1;
217 RTE_LOG(ERR, EAL, "Cannot resize '%s' for rte_mem_config\n",
218 pathname);
219 return -1;
220 }
221
222 retval = fcntl(mem_cfg_fd, F_SETLK, &wr_lock);
223 if (retval < 0){
224 close(mem_cfg_fd);
225 mem_cfg_fd = -1;
226 RTE_LOG(ERR, EAL, "Cannot create lock on '%s'. Is another primary "
227 "process running?\n", pathname);
228 return -1;
229 }
230
231
232 rte_mem_cfg_addr = eal_get_virtual_area(rte_mem_cfg_addr,
233 &cfg_len_aligned, page_sz, 0, 0);
234 if (rte_mem_cfg_addr == NULL) {
235 RTE_LOG(ERR, EAL, "Cannot mmap memory for rte_config\n");
236 close(mem_cfg_fd);
237 mem_cfg_fd = -1;
238 return -1;
239 }
240
241
242 mapped_mem_cfg_addr = mmap(rte_mem_cfg_addr,
243 cfg_len_aligned, PROT_READ | PROT_WRITE,
244 MAP_SHARED | MAP_FIXED, mem_cfg_fd, 0);
245 if (mapped_mem_cfg_addr == MAP_FAILED) {
246 RTE_LOG(ERR, EAL, "Cannot remap memory for rte_config\n");
247 munmap(rte_mem_cfg_addr, cfg_len);
248 close(mem_cfg_fd);
249 mem_cfg_fd = -1;
250 return -1;
251 }
252
253 memcpy(rte_mem_cfg_addr, config->mem_config, sizeof(struct rte_mem_config));
254 config->mem_config = rte_mem_cfg_addr;
255
256
257
258
259 config->mem_config->mem_cfg_addr = (uintptr_t) rte_mem_cfg_addr;
260 return 0;
261}
262
263
264static int
265rte_eal_config_attach(void)
266{
267 void *rte_mem_cfg_addr;
268 const char *pathname = eal_runtime_config_path();
269 struct rte_config *config = rte_eal_get_configuration();
270 const struct internal_config *internal_conf =
271 eal_get_internal_configuration();
272
273
274 if (internal_conf->no_shconf)
275 return 0;
276
277 if (mem_cfg_fd < 0){
278 mem_cfg_fd = open(pathname, O_RDWR);
279 if (mem_cfg_fd < 0) {
280 RTE_LOG(ERR, EAL, "Cannot open '%s' for rte_mem_config\n",
281 pathname);
282 return -1;
283 }
284 }
285
286 rte_mem_cfg_addr = mmap(NULL, sizeof(*config->mem_config),
287 PROT_READ, MAP_SHARED, mem_cfg_fd, 0);
288
289 if (rte_mem_cfg_addr == MAP_FAILED) {
290 close(mem_cfg_fd);
291 mem_cfg_fd = -1;
292 RTE_LOG(ERR, EAL, "Cannot mmap memory for rte_config! error %i (%s)\n",
293 errno, strerror(errno));
294 return -1;
295 }
296
297 config->mem_config = rte_mem_cfg_addr;
298
299 return 0;
300}
301
302
303static int
304rte_eal_config_reattach(void)
305{
306 struct rte_mem_config *mem_config;
307 void *rte_mem_cfg_addr;
308 struct rte_config *config = rte_eal_get_configuration();
309 const struct internal_config *internal_conf =
310 eal_get_internal_configuration();
311
312 if (internal_conf->no_shconf)
313 return 0;
314
315
316 rte_mem_cfg_addr =
317 (void *)(uintptr_t)config->mem_config->mem_cfg_addr;
318
319
320 munmap(config->mem_config, sizeof(struct rte_mem_config));
321
322
323 mem_config = (struct rte_mem_config *) mmap(rte_mem_cfg_addr,
324 sizeof(*mem_config), PROT_READ | PROT_WRITE, MAP_SHARED,
325 mem_cfg_fd, 0);
326 close(mem_cfg_fd);
327 mem_cfg_fd = -1;
328
329 if (mem_config == MAP_FAILED || mem_config != rte_mem_cfg_addr) {
330 if (mem_config != MAP_FAILED) {
331
332 RTE_LOG(ERR, EAL, "Cannot mmap memory for rte_config at [%p], got [%p]"
333 " - please use '--" OPT_BASE_VIRTADDR
334 "' option\n",
335 rte_mem_cfg_addr, mem_config);
336 munmap(mem_config, sizeof(struct rte_mem_config));
337 return -1;
338 }
339 RTE_LOG(ERR, EAL, "Cannot mmap memory for rte_config! error %i (%s)\n",
340 errno, strerror(errno));
341 return -1;
342 }
343
344 config->mem_config = mem_config;
345
346 return 0;
347}
348
349
350enum rte_proc_type_t
351eal_proc_type_detect(void)
352{
353 enum rte_proc_type_t ptype = RTE_PROC_PRIMARY;
354 const char *pathname = eal_runtime_config_path();
355 const struct internal_config *internal_conf =
356 eal_get_internal_configuration();
357
358
359 if (!internal_conf->no_shconf) {
360
361
362
363
364
365 if (((mem_cfg_fd = open(pathname, O_RDWR)) >= 0) &&
366 (fcntl(mem_cfg_fd, F_SETLK, &wr_lock) < 0))
367 ptype = RTE_PROC_SECONDARY;
368 }
369
370 RTE_LOG(INFO, EAL, "Auto-detected process type: %s\n",
371 ptype == RTE_PROC_PRIMARY ? "PRIMARY" : "SECONDARY");
372
373 return ptype;
374}
375
376
377static int
378rte_config_init(void)
379{
380 struct rte_config *config = rte_eal_get_configuration();
381 const struct internal_config *internal_conf =
382 eal_get_internal_configuration();
383
384 config->process_type = internal_conf->process_type;
385
386 switch (config->process_type) {
387 case RTE_PROC_PRIMARY:
388 if (rte_eal_config_create() < 0)
389 return -1;
390 eal_mcfg_update_from_internal();
391 break;
392 case RTE_PROC_SECONDARY:
393 if (rte_eal_config_attach() < 0)
394 return -1;
395 eal_mcfg_wait_complete();
396 if (eal_mcfg_check_version() < 0) {
397 RTE_LOG(ERR, EAL, "Primary and secondary process DPDK version mismatch\n");
398 return -1;
399 }
400 if (rte_eal_config_reattach() < 0)
401 return -1;
402 if (!__rte_mp_enable()) {
403 RTE_LOG(ERR, EAL, "Primary process refused secondary attachment\n");
404 return -1;
405 }
406 eal_mcfg_update_internal();
407 break;
408 case RTE_PROC_AUTO:
409 case RTE_PROC_INVALID:
410 RTE_LOG(ERR, EAL, "Invalid process type %d\n",
411 config->process_type);
412 return -1;
413 }
414
415 return 0;
416}
417
418
419static void
420eal_usage(const char *prgname)
421{
422 rte_usage_hook_t hook = eal_get_application_usage_hook();
423
424 printf("\nUsage: %s ", prgname);
425 eal_common_usage();
426
427 if (hook) {
428 printf("===== Application Usage =====\n\n");
429 (hook)(prgname);
430 }
431}
432
433static inline size_t
434eal_get_hugepage_mem_size(void)
435{
436 uint64_t size = 0;
437 unsigned i, j;
438 struct internal_config *internal_conf =
439 eal_get_internal_configuration();
440
441 for (i = 0; i < internal_conf->num_hugepage_sizes; i++) {
442 struct hugepage_info *hpi = &internal_conf->hugepage_info[i];
443 if (strnlen(hpi->hugedir, sizeof(hpi->hugedir)) != 0) {
444 for (j = 0; j < RTE_MAX_NUMA_NODES; j++) {
445 size += hpi->hugepage_sz * hpi->num_pages[j];
446 }
447 }
448 }
449
450 return (size < SIZE_MAX) ? (size_t)(size) : SIZE_MAX;
451}
452
453
454static void
455eal_log_level_parse(int argc, char **argv)
456{
457 int opt;
458 char **argvopt;
459 int option_index;
460 const int old_optind = optind;
461 const int old_optopt = optopt;
462 const int old_optreset = optreset;
463 char * const old_optarg = optarg;
464 struct internal_config *internal_conf =
465 eal_get_internal_configuration();
466
467 argvopt = argv;
468 optind = 1;
469 optreset = 1;
470
471 while ((opt = getopt_long(argc, argvopt, eal_short_options,
472 eal_long_options, &option_index)) != EOF) {
473
474 int ret;
475
476
477 if (opt == '?')
478 break;
479
480 ret = (opt == OPT_LOG_LEVEL_NUM) ?
481 eal_parse_common_option(opt, optarg, internal_conf) : 0;
482
483
484 if (ret < 0)
485 break;
486 }
487
488
489 optind = old_optind;
490 optopt = old_optopt;
491 optreset = old_optreset;
492 optarg = old_optarg;
493}
494
495
496static int
497eal_parse_args(int argc, char **argv)
498{
499 int opt, ret;
500 char **argvopt;
501 int option_index;
502 char *prgname = argv[0];
503 const int old_optind = optind;
504 const int old_optopt = optopt;
505 const int old_optreset = optreset;
506 char * const old_optarg = optarg;
507 struct internal_config *internal_conf =
508 eal_get_internal_configuration();
509
510 argvopt = argv;
511 optind = 1;
512 optreset = 1;
513
514 while ((opt = getopt_long(argc, argvopt, eal_short_options,
515 eal_long_options, &option_index)) != EOF) {
516
517
518 if (opt == '?') {
519 eal_usage(prgname);
520 ret = -1;
521 goto out;
522 }
523
524
525 if (opt == OPT_LOG_LEVEL_NUM)
526 continue;
527
528 ret = eal_parse_common_option(opt, optarg, internal_conf);
529
530 if (ret < 0) {
531 eal_usage(prgname);
532 ret = -1;
533 goto out;
534 }
535
536 if (ret == 0)
537 continue;
538
539 switch (opt) {
540 case OPT_MBUF_POOL_OPS_NAME_NUM:
541 {
542 char *ops_name = strdup(optarg);
543 if (ops_name == NULL)
544 RTE_LOG(ERR, EAL, "Could not store mbuf pool ops name\n");
545 else {
546
547 if (internal_conf->user_mbuf_pool_ops_name !=
548 NULL)
549 free(internal_conf->user_mbuf_pool_ops_name);
550
551 internal_conf->user_mbuf_pool_ops_name =
552 ops_name;
553 }
554 break;
555 }
556 case 'h':
557 eal_usage(prgname);
558 exit(EXIT_SUCCESS);
559 default:
560 if (opt < OPT_LONG_MIN_NUM && isprint(opt)) {
561 RTE_LOG(ERR, EAL, "Option %c is not supported "
562 "on FreeBSD\n", opt);
563 } else if (opt >= OPT_LONG_MIN_NUM &&
564 opt < OPT_LONG_MAX_NUM) {
565 RTE_LOG(ERR, EAL, "Option %s is not supported "
566 "on FreeBSD\n",
567 eal_long_options[option_index].name);
568 } else {
569 RTE_LOG(ERR, EAL, "Option %d is not supported "
570 "on FreeBSD\n", opt);
571 }
572 eal_usage(prgname);
573 ret = -1;
574 goto out;
575 }
576 }
577
578
579 if (eal_create_runtime_dir() < 0) {
580 if (internal_conf->no_shconf == 0) {
581 RTE_LOG(ERR, EAL, "Cannot create runtime directory\n");
582 ret = -1;
583 goto out;
584 } else
585 RTE_LOG(WARNING, EAL, "No DPDK runtime directory created\n");
586 }
587
588 if (eal_adjust_config(internal_conf) != 0) {
589 ret = -1;
590 goto out;
591 }
592
593
594 if (eal_check_common_options(internal_conf) != 0) {
595 eal_usage(prgname);
596 ret = -1;
597 goto out;
598 }
599
600 if (optind >= 0)
601 argv[optind-1] = prgname;
602 ret = optind-1;
603
604out:
605
606 optind = old_optind;
607 optopt = old_optopt;
608 optreset = old_optreset;
609 optarg = old_optarg;
610
611 return ret;
612}
613
614static int
615check_socket(const struct rte_memseg_list *msl, void *arg)
616{
617 int *socket_id = arg;
618
619 if (msl->external)
620 return 0;
621
622 if (msl->socket_id == *socket_id && msl->memseg_arr.count != 0)
623 return 1;
624
625 return 0;
626}
627
628static void
629eal_check_mem_on_local_socket(void)
630{
631 int socket_id;
632 const struct rte_config *config = rte_eal_get_configuration();
633
634 socket_id = rte_lcore_to_socket_id(config->main_lcore);
635
636 if (rte_memseg_list_walk(check_socket, &socket_id) == 0)
637 RTE_LOG(WARNING, EAL, "WARNING: Main core has no memory on local socket!\n");
638}
639
640
641static int
642sync_func(__rte_unused void *arg)
643{
644 return 0;
645}
646
647int
648rte_eal_iopl_init(void)
649{
650 static int fd = -1;
651
652 if (fd < 0)
653 fd = open("/dev/io", O_RDWR);
654
655 if (fd < 0)
656 return -1;
657
658 return 0;
659}
660
661static void rte_eal_init_alert(const char *msg)
662{
663 fprintf(stderr, "EAL: FATAL: %s\n", msg);
664 RTE_LOG(ERR, EAL, "%s\n", msg);
665}
666
667
668int
669rte_eal_init(int argc, char **argv)
670{
671 int i, fctret, ret;
672 pthread_t thread_id;
673 static uint32_t run_once;
674 uint32_t has_run = 0;
675 char cpuset[RTE_CPU_AFFINITY_STR_LEN];
676 char thread_name[RTE_MAX_THREAD_NAME_LEN];
677 const struct rte_config *config = rte_eal_get_configuration();
678 struct internal_config *internal_conf =
679 eal_get_internal_configuration();
680
681
682 if (!rte_cpu_is_supported()) {
683 rte_eal_init_alert("unsupported cpu type.");
684 rte_errno = ENOTSUP;
685 return -1;
686 }
687
688 if (!__atomic_compare_exchange_n(&run_once, &has_run, 1, 0,
689 __ATOMIC_RELAXED, __ATOMIC_RELAXED)) {
690 rte_eal_init_alert("already called initialization.");
691 rte_errno = EALREADY;
692 return -1;
693 }
694
695 thread_id = pthread_self();
696
697 eal_reset_internal_config(internal_conf);
698
699
700 eal_save_args(argc, argv);
701
702
703 eal_log_level_parse(argc, argv);
704
705 if (rte_eal_cpu_init() < 0) {
706 rte_eal_init_alert("Cannot detect lcores.");
707 rte_errno = ENOTSUP;
708 return -1;
709 }
710
711 fctret = eal_parse_args(argc, argv);
712 if (fctret < 0) {
713 rte_eal_init_alert("Invalid 'command line' arguments.");
714 rte_errno = EINVAL;
715 __atomic_store_n(&run_once, 0, __ATOMIC_RELAXED);
716 return -1;
717 }
718
719
720 internal_conf->legacy_mem = true;
721
722 if (eal_plugins_init() < 0) {
723 rte_eal_init_alert("Cannot init plugins");
724 rte_errno = EINVAL;
725 __atomic_store_n(&run_once, 0, __ATOMIC_RELAXED);
726 return -1;
727 }
728
729 if (eal_trace_init() < 0) {
730 rte_eal_init_alert("Cannot init trace");
731 rte_errno = EFAULT;
732 __atomic_store_n(&run_once, 0, __ATOMIC_RELAXED);
733 return -1;
734 }
735
736 if (eal_option_device_parse()) {
737 rte_errno = ENODEV;
738 __atomic_store_n(&run_once, 0, __ATOMIC_RELAXED);
739 return -1;
740 }
741
742 if (rte_config_init() < 0) {
743 rte_eal_init_alert("Cannot init config");
744 return -1;
745 }
746
747 if (rte_eal_intr_init() < 0) {
748 rte_eal_init_alert("Cannot init interrupt-handling thread");
749 return -1;
750 }
751
752 if (rte_eal_alarm_init() < 0) {
753 rte_eal_init_alert("Cannot init alarm");
754
755 return -1;
756 }
757
758
759
760
761 if (rte_mp_channel_init() < 0 && rte_errno != ENOTSUP) {
762 rte_eal_init_alert("failed to init mp channel");
763 if (rte_eal_process_type() == RTE_PROC_PRIMARY) {
764 rte_errno = EFAULT;
765 return -1;
766 }
767 }
768
769 if (rte_bus_scan()) {
770 rte_eal_init_alert("Cannot scan the buses for devices");
771 rte_errno = ENODEV;
772 __atomic_store_n(&run_once, 0, __ATOMIC_RELAXED);
773 return -1;
774 }
775
776
777 if (internal_conf->iova_mode == RTE_IOVA_DC) {
778
779 enum rte_iova_mode iova_mode = rte_bus_get_iommu_class();
780
781 if (iova_mode == RTE_IOVA_DC)
782 iova_mode = RTE_IOVA_PA;
783 rte_eal_get_configuration()->iova_mode = iova_mode;
784 } else {
785 rte_eal_get_configuration()->iova_mode =
786 internal_conf->iova_mode;
787 }
788
789 RTE_LOG(INFO, EAL, "Selected IOVA mode '%s'\n",
790 rte_eal_iova_mode() == RTE_IOVA_PA ? "PA" : "VA");
791
792 if (internal_conf->no_hugetlbfs == 0) {
793
794 ret = internal_conf->process_type == RTE_PROC_PRIMARY ?
795 eal_hugepage_info_init() :
796 eal_hugepage_info_read();
797 if (ret < 0) {
798 rte_eal_init_alert("Cannot get hugepage information.");
799 rte_errno = EACCES;
800 __atomic_store_n(&run_once, 0, __ATOMIC_RELAXED);
801 return -1;
802 }
803 }
804
805 if (internal_conf->memory == 0 && internal_conf->force_sockets == 0) {
806 if (internal_conf->no_hugetlbfs)
807 internal_conf->memory = MEMSIZE_IF_NO_HUGE_PAGE;
808 else
809 internal_conf->memory = eal_get_hugepage_mem_size();
810 }
811
812 if (internal_conf->vmware_tsc_map == 1) {
813#ifdef RTE_LIBRTE_EAL_VMWARE_TSC_MAP_SUPPORT
814 rte_cycles_vmware_tsc_map = 1;
815 RTE_LOG (DEBUG, EAL, "Using VMWARE TSC MAP, "
816 "you must have monitor_control.pseudo_perfctr = TRUE\n");
817#else
818 RTE_LOG (WARNING, EAL, "Ignoring --vmware-tsc-map because "
819 "RTE_LIBRTE_EAL_VMWARE_TSC_MAP_SUPPORT is not set\n");
820#endif
821 }
822
823
824
825
826
827 if (rte_eal_memzone_init() < 0) {
828 rte_eal_init_alert("Cannot init memzone");
829 rte_errno = ENODEV;
830 return -1;
831 }
832
833 if (rte_eal_memory_init() < 0) {
834 rte_eal_init_alert("Cannot init memory");
835 rte_errno = ENOMEM;
836 return -1;
837 }
838
839 if (rte_eal_malloc_heap_init() < 0) {
840 rte_eal_init_alert("Cannot init malloc heap");
841 rte_errno = ENODEV;
842 return -1;
843 }
844
845 if (rte_eal_tailqs_init() < 0) {
846 rte_eal_init_alert("Cannot init tail queues for objects");
847 rte_errno = EFAULT;
848 return -1;
849 }
850
851 if (rte_eal_timer_init() < 0) {
852 rte_eal_init_alert("Cannot init HPET or TSC timers");
853 rte_errno = ENOTSUP;
854 return -1;
855 }
856
857 eal_check_mem_on_local_socket();
858
859 if (pthread_setaffinity_np(pthread_self(), sizeof(rte_cpuset_t),
860 &lcore_config[config->main_lcore].cpuset) != 0) {
861 rte_eal_init_alert("Cannot set affinity");
862 rte_errno = EINVAL;
863 return -1;
864 }
865 __rte_thread_init(config->main_lcore,
866 &lcore_config[config->main_lcore].cpuset);
867
868 ret = eal_thread_dump_current_affinity(cpuset, sizeof(cpuset));
869
870 RTE_LOG(DEBUG, EAL, "Main lcore %u is ready (tid=%p;cpuset=[%s%s])\n",
871 config->main_lcore, thread_id, cpuset,
872 ret == 0 ? "" : "...");
873
874 RTE_LCORE_FOREACH_WORKER(i) {
875
876
877
878
879
880 if (pipe(lcore_config[i].pipe_main2worker) < 0)
881 rte_panic("Cannot create pipe\n");
882 if (pipe(lcore_config[i].pipe_worker2main) < 0)
883 rte_panic("Cannot create pipe\n");
884
885 lcore_config[i].state = WAIT;
886
887
888 ret = pthread_create(&lcore_config[i].thread_id, NULL,
889 eal_thread_loop, NULL);
890 if (ret != 0)
891 rte_panic("Cannot create thread\n");
892
893
894 snprintf(thread_name, sizeof(thread_name),
895 "lcore-worker-%d", i);
896 rte_thread_setname(lcore_config[i].thread_id, thread_name);
897
898 ret = pthread_setaffinity_np(lcore_config[i].thread_id,
899 sizeof(rte_cpuset_t), &lcore_config[i].cpuset);
900 if (ret != 0)
901 rte_panic("Cannot set affinity\n");
902 }
903
904
905
906
907
908 rte_eal_mp_remote_launch(sync_func, NULL, SKIP_MAIN);
909 rte_eal_mp_wait_lcore();
910
911
912 ret = rte_service_init();
913 if (ret) {
914 rte_eal_init_alert("rte_service_init() failed");
915 rte_errno = -ret;
916 return -1;
917 }
918
919
920 if (rte_bus_probe()) {
921 rte_eal_init_alert("Cannot probe devices");
922 rte_errno = ENOTSUP;
923 return -1;
924 }
925
926
927
928
929 ret = rte_service_start_with_defaults();
930 if (ret < 0 && ret != -ENOTSUP) {
931 rte_errno = -ret;
932 return -1;
933 }
934
935
936
937
938
939
940
941
942
943
944
945 if (!internal_conf->no_shconf && eal_clean_runtime_dir() < 0) {
946 rte_eal_init_alert("Cannot clear runtime directory");
947 return -1;
948 }
949 if (!internal_conf->no_telemetry) {
950 int tlog = rte_log_register_type_and_pick_level(
951 "lib.telemetry", RTE_LOG_WARNING);
952 if (tlog < 0)
953 tlog = RTE_LOGTYPE_EAL;
954 if (rte_telemetry_init(rte_eal_get_runtime_dir(),
955 rte_version(),
956 &internal_conf->ctrl_cpuset, rte_log, tlog) != 0)
957 return -1;
958 }
959
960 eal_mcfg_complete();
961
962 return fctret;
963}
964
965int
966rte_eal_cleanup(void)
967{
968 struct internal_config *internal_conf =
969 eal_get_internal_configuration();
970 rte_service_finalize();
971 rte_mp_channel_cleanup();
972
973 rte_eal_memory_detach();
974 rte_trace_save();
975 eal_trace_fini();
976 eal_cleanup_config(internal_conf);
977 return 0;
978}
979
980int rte_eal_create_uio_dev(void)
981{
982 const struct internal_config *internal_conf =
983 eal_get_internal_configuration();
984 return internal_conf->create_uio_dev;
985}
986
987enum rte_intr_mode
988rte_eal_vfio_intr_mode(void)
989{
990 return RTE_INTR_MODE_NONE;
991}
992
993void
994rte_eal_vfio_get_vf_token(__rte_unused rte_uuid_t vf_token)
995{
996}
997
998int rte_vfio_setup_device(__rte_unused const char *sysfs_base,
999 __rte_unused const char *dev_addr,
1000 __rte_unused int *vfio_dev_fd,
1001 __rte_unused struct vfio_device_info *device_info)
1002{
1003 return -1;
1004}
1005
1006int rte_vfio_release_device(__rte_unused const char *sysfs_base,
1007 __rte_unused const char *dev_addr,
1008 __rte_unused int fd)
1009{
1010 return -1;
1011}
1012
1013int rte_vfio_enable(__rte_unused const char *modname)
1014{
1015 return -1;
1016}
1017
1018int rte_vfio_is_enabled(__rte_unused const char *modname)
1019{
1020 return 0;
1021}
1022
1023int rte_vfio_noiommu_is_enabled(void)
1024{
1025 return 0;
1026}
1027
1028int rte_vfio_clear_group(__rte_unused int vfio_group_fd)
1029{
1030 return 0;
1031}
1032
1033int
1034rte_vfio_get_group_num(__rte_unused const char *sysfs_base,
1035 __rte_unused const char *dev_addr,
1036 __rte_unused int *iommu_group_num)
1037{
1038 return -1;
1039}
1040
1041int
1042rte_vfio_get_container_fd(void)
1043{
1044 return -1;
1045}
1046
1047int
1048rte_vfio_get_group_fd(__rte_unused int iommu_group_num)
1049{
1050 return -1;
1051}
1052
1053int
1054rte_vfio_container_create(void)
1055{
1056 return -1;
1057}
1058
1059int
1060rte_vfio_container_destroy(__rte_unused int container_fd)
1061{
1062 return -1;
1063}
1064
1065int
1066rte_vfio_container_group_bind(__rte_unused int container_fd,
1067 __rte_unused int iommu_group_num)
1068{
1069 return -1;
1070}
1071
1072int
1073rte_vfio_container_group_unbind(__rte_unused int container_fd,
1074 __rte_unused int iommu_group_num)
1075{
1076 return -1;
1077}
1078
1079int
1080rte_vfio_container_dma_map(__rte_unused int container_fd,
1081 __rte_unused uint64_t vaddr,
1082 __rte_unused uint64_t iova,
1083 __rte_unused uint64_t len)
1084{
1085 return -1;
1086}
1087
1088int
1089rte_vfio_container_dma_unmap(__rte_unused int container_fd,
1090 __rte_unused uint64_t vaddr,
1091 __rte_unused uint64_t iova,
1092 __rte_unused uint64_t len)
1093{
1094 return -1;
1095}
1096