mmap is failing in "odp_shm_reserve" function (odp_queue_init_global() --->
odp_shm_reserve() ---> odp_shm_reserve())


debug logs:

root@ubuntu-15-10:/home/linaro/linaro/odp/example/classifier#
./odp_classifier -i eno1 -m 0 -p
"ODP_PMR_SIP_ADDR:192.168.10.11:FFFFFFFF:queue1"
-p "ODP_PMR_SIP_ADDR:10.130.69.0:000000FF:queue2" -p
"ODP_PMR_SIP_ADDR:10.130.68.0:FFFFFE00:queue3"
odp_pool.c:104:odp_pool_init_global():
Pool init global
odp_pool.c:105:odp_pool_init_global():  pool_entry_s size     8512
odp_pool.c:106:odp_pool_init_global():  pool_entry_t size     8512
odp_pool.c:107:odp_pool_init_global():  odp_buffer_hdr_t size 216
odp_pool.c:108:odp_pool_init_global():
odp_queue.c:130:odp_queue_init_global():Queue init ...
odp_shared_memory.c:296:odp_shm_reserve():odp_queues:
    :: No huge pages, fall back to normal pages,
    check: /proc/sys/vm/nr_hugepages.
odp_queue.c:154:odp_queue_init_global():done
odp_queue.c:155:odp_queue_init_global():Queue init global
odp_queue.c:157:odp_queue_init_global():  struct queue_entry_s size 320
odp_queue.c:159:odp_queue_init_global():  queue_entry_t size        320
odp_queue.c:160:odp_queue_init_global():
odp_schedule.c:145:odp_schedule_init_global():Schedule init ...
odp_shared_memory.c:296:odp_shm_reserve():odp_scheduler:
    :: No huge pages, fall back to normal pages,
    check: /proc/sys/vm/nr_hugepages.
odp_shared_memory.c:296:odp_shm_reserve():odp_sched_pool:
    :: No huge pages, fall back to normal pages,
    check: /proc/sys/vm/nr_hugepages.
odp_schedule.c:226:odp_schedule_init_global():done
odp_shared_memory.c:296:odp_shm_reserve():odp_pktio_entries:
    :: No huge pages, fall back to normal pages,
    check: /proc/sys/vm/nr_hugepages.
odp_shared_memory.c:296:odp_shm_reserve():crypto_pool:
    :: No huge pages, fall back to normal pages,
    check: /proc/sys/vm/nr_hugepages.
odp_shared_memory.c:296:odp_shm_reserve():shm_odp_cos_tbl:
    :: No huge pages, fall back to normal pages,
    check: /proc/sys/vm/nr_hugepages.
odp_shared_memory.c:296:odp_shm_reserve():shm_odp_pmr_tbl:
    :: No huge pages, fall back to normal pages,
    check: /proc/sys/vm/nr_hugepages.
main :: odp_init_global done!
odp_classifier.c:500:main():main :: odp_init_local done!
odp_classifier.c:505:main():main :: odp_shm_reserve done!

ODP system info
---------------
ODP API version: 1.10.0
CPU model:       Intel(R) Core(TM) i7-5600U CPU
CPU freq (hz):   2600000000
Cache line size: 64
CPU count:       4

Running ODP appl: "odp_classifier"
-----------------
Using IF:eno1

num worker threads: 2
first CPU:          2
cpu mask:           0xC
odp_shared_memory.c:296:odp_shm_reserve():packet_pool:
    :: No huge pages, fall back to normal pages,
    check: /proc/sys/vm/nr_hugepages.
odp_pool.c:759:odp_pool_print():Pool info
odp_pool.c:760:odp_pool_print():---------
odp_pool.c:762:odp_pool_print(): pool            1
odp_pool.c:764:odp_pool_print(): name            packet_pool
odp_pool.c:769:odp_pool_print(): pool type       packet
odp_pool.c:771:odp_pool_print(): pool storage    ODP managed shm handle 11
odp_pool.c:773:odp_pool_print(): pool status     active
odp_pool.c:777:odp_pool_print(): pool opts       segmented, non-zeroized,
created
odp_pool.c:778:odp_pool_print(): pool base       0x7f5091aab000
odp_pool.c:780:odp_pool_print(): pool size       1310720 (320 pages)
odp_pool.c:781:odp_pool_print(): pool mdata base 0x7f5091bb5940
odp_pool.c:782:odp_pool_print(): udata size      0
odp_pool.c:783:odp_pool_print(): headroom        66
odp_pool.c:784:odp_pool_print(): tailroom        0
odp_pool.c:791:odp_pool_print(): seg length      1856 requested, 1936 used
odp_pool.c:793:odp_pool_print(): pkt length      1856 requested, 1936 used
odp_pool.c:795:odp_pool_print(): num bufs        564
odp_pool.c:797:odp_pool_print(): bufs available  564
odp_pool.c:798:odp_pool_print(): bufs in use     0
odp_pool.c:799:odp_pool_print(): buf allocs      0
odp_pool.c:800:odp_pool_print(): buf frees       0
odp_pool.c:801:odp_pool_print(): buf empty       0
odp_pool.c:803:odp_pool_print(): blk size        1936
odp_pool.c:805:odp_pool_print(): blks available  564
odp_pool.c:806:odp_pool_print(): blk allocs      0
odp_pool.c:807:odp_pool_print(): blk frees       0
odp_pool.c:808:odp_pool_print(): blk empty       0
odp_pool.c:809:odp_pool_print(): buf high wm value   282
odp_pool.c:810:odp_pool_print(): buf high wm count   0
odp_pool.c:811:odp_pool_print(): buf low wm value    141
odp_pool.c:812:odp_pool_print(): buf low wm count    0
odp_pool.c:813:odp_pool_print(): blk high wm value   282
odp_pool.c:814:odp_pool_print(): blk high wm count   0
odp_pool.c:815:odp_pool_print(): blk low wm value    141
odp_pool.c:816:odp_pool_print(): blk low wm count    0
main :: odp_pool_print done!
odp_packet_io.c:230:setup_pktio_entry():eno1 uses socket_mmap
  created pktio:01, dev:eno1, queue mode (ATOMIC queues)
      default pktio01
odp_shared_memory.c:296:odp_shm_reserve():  DefaultPool:
    :: No huge pages, fall back to normal pages,
    check: /proc/sys/vm/nr_hugepages.
odp_shared_memory.c:296:odp_shm_reserve():  queue1Pool0:
    :: No huge pages, fall back to normal pages,
    check: /proc/sys/vm/nr_hugepages.
odp_shared_memory.c:296:odp_shm_reserve():  queue2Pool1:
    :: No huge pages, fall back to normal pages,
    check: /proc/sys/vm/nr_hugepages.
odp_shared_memory.c:296:odp_shm_reserve():  queue3Pool2:
    :: No huge pages, fall back to normal pages,
    check: /proc/sys/vm/nr_hugepages.

----------------------------------------
CLASSIFIER EXAMPLE STATISTICS
----------------------------------------
CONFIGURATION

COS    VALUE        MASK
----------------------------------------
queue1    192.168.10.11    FFFFFFFF
queue2    10.130.69.0    000000FF
queue3    10.130.68.0    FFFFFE00

RECEIVED PACKETS
----------------------------------------
queue1       |queue2       |queue3       |DefaultCos   |Total Packets
queue  pool  |queue  pool  |queue  pool  |queue  pool  |
845    845   |0      0     |0      0     |38     38    |883


On Mon, May 9, 2016 at 9:00 PM, Bill Fischofer <[email protected]>
wrote:

>
>
> On Mon, May 9, 2016 at 6:57 AM, nousi <[email protected]> wrote:
>
>>
>> Hi All,
>>
>> Pleas help me in running ODP classifier example with huge pages.
>> In ubuntu 15.10 by default interface naming as "eno1" and value in
>> "/proc/sys/vm/nr_hugepages." is 1024.
>> classifier example program could not able to use huge pages even though
>> nr_hugepages value is non zero.
>> I could able to run the classier example, but it is not using huge pages.
>>
>>
>>
>> *console log is pasted below for you reference.*root@odp/example/classifier#
>> ./odp_classifier -i eno1 -m 0 -p 
>> "ODP_PMR_SIP_ADDR:192.168.10.11:FFFFFFFF:queue1"
>> -p "ODP_PMR_SIP_ADDR:10.130.69.0:000000FF:queue2" -p
>> "ODP_PMR_SIP_ADDR:10.130.68.0:FFFFFE00:queue3"
>> odp_pool.c:104:odp_pool_init_global():
>> Pool init global
>> odp_pool.c:105:odp_pool_init_global():  pool_entry_s size     8512
>> odp_pool.c:106:odp_pool_init_global():  pool_entry_t size     8512
>> odp_pool.c:107:odp_pool_init_global():  odp_buffer_hdr_t size 216
>> odp_pool.c:108:odp_pool_init_global():
>> odp_queue.c:130:odp_queue_init_global():Queue init ...
>> odp_shared_memory.c:296:odp_shm_reserve(): odp_queues:
>>     No huge pages, fall back to normal pages,
>>     check: /proc/sys/vm/nr_hugepages.
>>
>
> This is an informational message saying that the linux-generic
> implementation was unable to allocate huge pages so it's falling back to
> normal pages.  I'm not sure why you're seeing that except that it seems
> that some allocations may have been successful (those in odp_pool.c) while
> those for queue initialization were not.
>
> I'll let others who are more expert in this area chime in with some
> additional thoughts.
>
>
>>
>>
>>
>> Thanks & Regards,
>>
>> *B.Nousilal,*
>>
>> _______________________________________________
>> lng-odp mailing list
>> [email protected]
>> https://lists.linaro.org/mailman/listinfo/lng-odp
>>
>>
>


-- 


*Thanks & Regards,B.Nousilal,*
_______________________________________________
lng-odp mailing list
[email protected]
https://lists.linaro.org/mailman/listinfo/lng-odp

Reply via email to