Please put output for: cat /proc/meminfo cat /proc/sys/vm/nr_hugepages Thank you, Maxim.
On 10 May 2016 at 08:36, nousi <[email protected]> wrote: > mmap is failing in "odp_shm_reserve" function (odp_queue_init_global() > ---> odp_shm_reserve() ---> odp_shm_reserve()) > > > debug logs: > > root@ubuntu-15-10:/home/linaro/linaro/odp/example/classifier# > ./odp_classifier -i eno1 -m 0 -p > "ODP_PMR_SIP_ADDR:192.168.10.11:FFFFFFFF:queue1" > -p "ODP_PMR_SIP_ADDR:10.130.69.0:000000FF:queue2" -p > "ODP_PMR_SIP_ADDR:10.130.68.0:FFFFFE00:queue3" > odp_pool.c:104:odp_pool_init_global(): > Pool init global > odp_pool.c:105:odp_pool_init_global(): pool_entry_s size 8512 > odp_pool.c:106:odp_pool_init_global(): pool_entry_t size 8512 > odp_pool.c:107:odp_pool_init_global(): odp_buffer_hdr_t size 216 > odp_pool.c:108:odp_pool_init_global(): > odp_queue.c:130:odp_queue_init_global():Queue init ... > odp_shared_memory.c:296:odp_shm_reserve():odp_queues: > :: No huge pages, fall back to normal pages, > check: /proc/sys/vm/nr_hugepages. > odp_queue.c:154:odp_queue_init_global():done > odp_queue.c:155:odp_queue_init_global():Queue init global > odp_queue.c:157:odp_queue_init_global(): struct queue_entry_s size 320 > odp_queue.c:159:odp_queue_init_global(): queue_entry_t size 320 > odp_queue.c:160:odp_queue_init_global(): > odp_schedule.c:145:odp_schedule_init_global():Schedule init ... > odp_shared_memory.c:296:odp_shm_reserve():odp_scheduler: > :: No huge pages, fall back to normal pages, > check: /proc/sys/vm/nr_hugepages. > odp_shared_memory.c:296:odp_shm_reserve():odp_sched_pool: > :: No huge pages, fall back to normal pages, > check: /proc/sys/vm/nr_hugepages. > odp_schedule.c:226:odp_schedule_init_global():done > odp_shared_memory.c:296:odp_shm_reserve():odp_pktio_entries: > :: No huge pages, fall back to normal pages, > check: /proc/sys/vm/nr_hugepages. > odp_shared_memory.c:296:odp_shm_reserve():crypto_pool: > :: No huge pages, fall back to normal pages, > check: /proc/sys/vm/nr_hugepages. > odp_shared_memory.c:296:odp_shm_reserve():shm_odp_cos_tbl: > :: No huge pages, fall back to normal pages, > check: /proc/sys/vm/nr_hugepages. > odp_shared_memory.c:296:odp_shm_reserve():shm_odp_pmr_tbl: > :: No huge pages, fall back to normal pages, > check: /proc/sys/vm/nr_hugepages. > main :: odp_init_global done! > odp_classifier.c:500:main():main :: odp_init_local done! > odp_classifier.c:505:main():main :: odp_shm_reserve done! > > ODP system info > --------------- > ODP API version: 1.10.0 > CPU model: Intel(R) Core(TM) i7-5600U CPU > CPU freq (hz): 2600000000 > Cache line size: 64 > CPU count: 4 > > Running ODP appl: "odp_classifier" > ----------------- > Using IF:eno1 > > num worker threads: 2 > first CPU: 2 > cpu mask: 0xC > odp_shared_memory.c:296:odp_shm_reserve():packet_pool: > :: No huge pages, fall back to normal pages, > check: /proc/sys/vm/nr_hugepages. > odp_pool.c:759:odp_pool_print():Pool info > odp_pool.c:760:odp_pool_print():--------- > odp_pool.c:762:odp_pool_print(): pool 1 > odp_pool.c:764:odp_pool_print(): name packet_pool > odp_pool.c:769:odp_pool_print(): pool type packet > odp_pool.c:771:odp_pool_print(): pool storage ODP managed shm handle 11 > odp_pool.c:773:odp_pool_print(): pool status active > odp_pool.c:777:odp_pool_print(): pool opts segmented, non-zeroized, > created > odp_pool.c:778:odp_pool_print(): pool base 0x7f5091aab000 > odp_pool.c:780:odp_pool_print(): pool size 1310720 (320 pages) > odp_pool.c:781:odp_pool_print(): pool mdata base 0x7f5091bb5940 > odp_pool.c:782:odp_pool_print(): udata size 0 > odp_pool.c:783:odp_pool_print(): headroom 66 > odp_pool.c:784:odp_pool_print(): tailroom 0 > odp_pool.c:791:odp_pool_print(): seg length 1856 requested, 1936 used > odp_pool.c:793:odp_pool_print(): pkt length 1856 requested, 1936 used > odp_pool.c:795:odp_pool_print(): num bufs 564 > odp_pool.c:797:odp_pool_print(): bufs available 564 > odp_pool.c:798:odp_pool_print(): bufs in use 0 > odp_pool.c:799:odp_pool_print(): buf allocs 0 > odp_pool.c:800:odp_pool_print(): buf frees 0 > odp_pool.c:801:odp_pool_print(): buf empty 0 > odp_pool.c:803:odp_pool_print(): blk size 1936 > odp_pool.c:805:odp_pool_print(): blks available 564 > odp_pool.c:806:odp_pool_print(): blk allocs 0 > odp_pool.c:807:odp_pool_print(): blk frees 0 > odp_pool.c:808:odp_pool_print(): blk empty 0 > odp_pool.c:809:odp_pool_print(): buf high wm value 282 > odp_pool.c:810:odp_pool_print(): buf high wm count 0 > odp_pool.c:811:odp_pool_print(): buf low wm value 141 > odp_pool.c:812:odp_pool_print(): buf low wm count 0 > odp_pool.c:813:odp_pool_print(): blk high wm value 282 > odp_pool.c:814:odp_pool_print(): blk high wm count 0 > odp_pool.c:815:odp_pool_print(): blk low wm value 141 > odp_pool.c:816:odp_pool_print(): blk low wm count 0 > main :: odp_pool_print done! > odp_packet_io.c:230:setup_pktio_entry():eno1 uses socket_mmap > created pktio:01, dev:eno1, queue mode (ATOMIC queues) > default pktio01 > odp_shared_memory.c:296:odp_shm_reserve(): DefaultPool: > :: No huge pages, fall back to normal pages, > check: /proc/sys/vm/nr_hugepages. > odp_shared_memory.c:296:odp_shm_reserve(): queue1Pool0: > :: No huge pages, fall back to normal pages, > check: /proc/sys/vm/nr_hugepages. > odp_shared_memory.c:296:odp_shm_reserve(): queue2Pool1: > :: No huge pages, fall back to normal pages, > check: /proc/sys/vm/nr_hugepages. > odp_shared_memory.c:296:odp_shm_reserve(): queue3Pool2: > :: No huge pages, fall back to normal pages, > check: /proc/sys/vm/nr_hugepages. > > ---------------------------------------- > CLASSIFIER EXAMPLE STATISTICS > ---------------------------------------- > CONFIGURATION > > COS VALUE MASK > ---------------------------------------- > queue1 192.168.10.11 FFFFFFFF > queue2 10.130.69.0 000000FF > queue3 10.130.68.0 FFFFFE00 > > RECEIVED PACKETS > ---------------------------------------- > queue1 |queue2 |queue3 |DefaultCos |Total Packets > queue pool |queue pool |queue pool |queue pool | > 845 845 |0 0 |0 0 |38 38 |883 > > > On Mon, May 9, 2016 at 9:00 PM, Bill Fischofer <[email protected]> > wrote: > >> >> >> On Mon, May 9, 2016 at 6:57 AM, nousi <[email protected]> wrote: >> >>> >>> Hi All, >>> >>> Pleas help me in running ODP classifier example with huge pages. >>> In ubuntu 15.10 by default interface naming as "eno1" and value in >>> "/proc/sys/vm/nr_hugepages." is 1024. >>> classifier example program could not able to use huge pages even though >>> nr_hugepages value is non zero. >>> I could able to run the classier example, but it is not using huge pages. >>> >>> >>> >>> *console log is pasted below for you reference.*root@odp/example/classifier# >>> ./odp_classifier -i eno1 -m 0 -p >>> "ODP_PMR_SIP_ADDR:192.168.10.11:FFFFFFFF:queue1" >>> -p "ODP_PMR_SIP_ADDR:10.130.69.0:000000FF:queue2" -p >>> "ODP_PMR_SIP_ADDR:10.130.68.0:FFFFFE00:queue3" >>> odp_pool.c:104:odp_pool_init_global(): >>> Pool init global >>> odp_pool.c:105:odp_pool_init_global(): pool_entry_s size 8512 >>> odp_pool.c:106:odp_pool_init_global(): pool_entry_t size 8512 >>> odp_pool.c:107:odp_pool_init_global(): odp_buffer_hdr_t size 216 >>> odp_pool.c:108:odp_pool_init_global(): >>> odp_queue.c:130:odp_queue_init_global():Queue init ... >>> odp_shared_memory.c:296:odp_shm_reserve(): odp_queues: >>> No huge pages, fall back to normal pages, >>> check: /proc/sys/vm/nr_hugepages. >>> >> >> This is an informational message saying that the linux-generic >> implementation was unable to allocate huge pages so it's falling back to >> normal pages. I'm not sure why you're seeing that except that it seems >> that some allocations may have been successful (those in odp_pool.c) while >> those for queue initialization were not. >> >> I'll let others who are more expert in this area chime in with some >> additional thoughts. >> >> >>> >>> >>> >>> Thanks & Regards, >>> >>> *B.Nousilal,* >>> >>> _______________________________________________ >>> lng-odp mailing list >>> [email protected] >>> https://lists.linaro.org/mailman/listinfo/lng-odp >>> >>> >> > > > -- > > > *Thanks & Regards,B.Nousilal,* > > _______________________________________________ > lng-odp mailing list > [email protected] > https://lists.linaro.org/mailman/listinfo/lng-odp > >
_______________________________________________ lng-odp mailing list [email protected] https://lists.linaro.org/mailman/listinfo/lng-odp
