Does: mount -t hugetlbfs none /mnt/hugetlbfs help?
Maxim. On 10 May 2016 at 13:16, Maxim Uvarov <[email protected]> wrote: > looks like you have enough free HP. Which error returns mmap() on try > with HP? > > On 10 May 2016 at 11:57, nousi <[email protected]> wrote: > >> linaro@ubuntu-15-10:~$ cat /proc/sys/vm/nr_hugepages >> 1024 >> linaro@ubuntu-15-10:~$ >> >> linaro@ubuntu-15-10:~$ cat /proc/meminfo >> MemTotal: 8061836 kB >> MemFree: 470516 kB >> MemAvailable: 1901932 kB >> Buffers: 92600 kB >> Cached: 1939696 kB >> SwapCached: 7516 kB >> Active: 3238960 kB >> Inactive: 1804492 kB >> Active(anon): 2712440 kB >> Inactive(anon): 1069492 kB >> Active(file): 526520 kB >> Inactive(file): 735000 kB >> Unevictable: 76 kB >> Mlocked: 76 kB >> SwapTotal: 16547836 kB >> SwapFree: 16370160 kB >> Dirty: 19816 kB >> Writeback: 0 kB >> AnonPages: 3004784 kB >> Mapped: 679960 kB >> Shmem: 770776 kB >> Slab: 264692 kB >> SReclaimable: 212172 kB >> SUnreclaim: 52520 kB >> KernelStack: 11952 kB >> PageTables: 65780 kB >> NFS_Unstable: 0 kB >> Bounce: 0 kB >> WritebackTmp: 0 kB >> CommitLimit: 19530176 kB >> Committed_AS: 11165432 kB >> VmallocTotal: 34359738367 kB >> VmallocUsed: 410416 kB >> VmallocChunk: 34358947836 kB >> HardwareCorrupted: 0 kB >> AnonHugePages: 583680 kB >> CmaTotal: 0 kB >> CmaFree: 0 kB >> HugePages_Total: 1024 >> HugePages_Free: 1022 >> HugePages_Rsvd: 1022 >> HugePages_Surp: 0 >> Hugepagesize: 2048 kB >> DirectMap4k: 234400 kB >> DirectMap2M: 6991872 kB >> DirectMap1G: 2097152 kB >> linaro@ubuntu-15-10:~$ >> >> >> On Tue, May 10, 2016 at 12:38 PM, Maxim Uvarov <[email protected]> >> wrote: >> >>> Please put output for: >>> cat /proc/meminfo >>> cat /proc/sys/vm/nr_hugepages >>> >>> Thank you, >>> Maxim. >>> >>> >>> >>> On 10 May 2016 at 08:36, nousi <[email protected]> wrote: >>> >>>> mmap is failing in "odp_shm_reserve" function (odp_queue_init_global() >>>> ---> odp_shm_reserve() ---> odp_shm_reserve()) >>>> >>>> >>>> debug logs: >>>> >>>> root@ubuntu-15-10:/home/linaro/linaro/odp/example/classifier# >>>> ./odp_classifier -i eno1 -m 0 -p >>>> "ODP_PMR_SIP_ADDR:192.168.10.11:FFFFFFFF:queue1" >>>> -p "ODP_PMR_SIP_ADDR:10.130.69.0:000000FF:queue2" -p >>>> "ODP_PMR_SIP_ADDR:10.130.68.0:FFFFFE00:queue3" >>>> odp_pool.c:104:odp_pool_init_global(): >>>> Pool init global >>>> odp_pool.c:105:odp_pool_init_global(): pool_entry_s size 8512 >>>> odp_pool.c:106:odp_pool_init_global(): pool_entry_t size 8512 >>>> odp_pool.c:107:odp_pool_init_global(): odp_buffer_hdr_t size 216 >>>> odp_pool.c:108:odp_pool_init_global(): >>>> odp_queue.c:130:odp_queue_init_global():Queue init ... >>>> odp_shared_memory.c:296:odp_shm_reserve():odp_queues: >>>> :: No huge pages, fall back to normal pages, >>>> check: /proc/sys/vm/nr_hugepages. >>>> odp_queue.c:154:odp_queue_init_global():done >>>> odp_queue.c:155:odp_queue_init_global():Queue init global >>>> odp_queue.c:157:odp_queue_init_global(): struct queue_entry_s size 320 >>>> odp_queue.c:159:odp_queue_init_global(): queue_entry_t size 320 >>>> odp_queue.c:160:odp_queue_init_global(): >>>> odp_schedule.c:145:odp_schedule_init_global():Schedule init ... >>>> odp_shared_memory.c:296:odp_shm_reserve():odp_scheduler: >>>> :: No huge pages, fall back to normal pages, >>>> check: /proc/sys/vm/nr_hugepages. >>>> odp_shared_memory.c:296:odp_shm_reserve():odp_sched_pool: >>>> :: No huge pages, fall back to normal pages, >>>> check: /proc/sys/vm/nr_hugepages. >>>> odp_schedule.c:226:odp_schedule_init_global():done >>>> odp_shared_memory.c:296:odp_shm_reserve():odp_pktio_entries: >>>> :: No huge pages, fall back to normal pages, >>>> check: /proc/sys/vm/nr_hugepages. >>>> odp_shared_memory.c:296:odp_shm_reserve():crypto_pool: >>>> :: No huge pages, fall back to normal pages, >>>> check: /proc/sys/vm/nr_hugepages. >>>> odp_shared_memory.c:296:odp_shm_reserve():shm_odp_cos_tbl: >>>> :: No huge pages, fall back to normal pages, >>>> check: /proc/sys/vm/nr_hugepages. >>>> odp_shared_memory.c:296:odp_shm_reserve():shm_odp_pmr_tbl: >>>> :: No huge pages, fall back to normal pages, >>>> check: /proc/sys/vm/nr_hugepages. >>>> main :: odp_init_global done! >>>> odp_classifier.c:500:main():main :: odp_init_local done! >>>> odp_classifier.c:505:main():main :: odp_shm_reserve done! >>>> >>>> ODP system info >>>> --------------- >>>> ODP API version: 1.10.0 >>>> CPU model: Intel(R) Core(TM) i7-5600U CPU >>>> CPU freq (hz): 2600000000 >>>> Cache line size: 64 >>>> CPU count: 4 >>>> >>>> Running ODP appl: "odp_classifier" >>>> ----------------- >>>> Using IF:eno1 >>>> >>>> num worker threads: 2 >>>> first CPU: 2 >>>> cpu mask: 0xC >>>> odp_shared_memory.c:296:odp_shm_reserve():packet_pool: >>>> :: No huge pages, fall back to normal pages, >>>> check: /proc/sys/vm/nr_hugepages. >>>> odp_pool.c:759:odp_pool_print():Pool info >>>> odp_pool.c:760:odp_pool_print():--------- >>>> odp_pool.c:762:odp_pool_print(): pool 1 >>>> odp_pool.c:764:odp_pool_print(): name packet_pool >>>> odp_pool.c:769:odp_pool_print(): pool type packet >>>> odp_pool.c:771:odp_pool_print(): pool storage ODP managed shm handle >>>> 11 >>>> odp_pool.c:773:odp_pool_print(): pool status active >>>> odp_pool.c:777:odp_pool_print(): pool opts segmented, >>>> non-zeroized, created >>>> odp_pool.c:778:odp_pool_print(): pool base 0x7f5091aab000 >>>> odp_pool.c:780:odp_pool_print(): pool size 1310720 (320 pages) >>>> odp_pool.c:781:odp_pool_print(): pool mdata base 0x7f5091bb5940 >>>> odp_pool.c:782:odp_pool_print(): udata size 0 >>>> odp_pool.c:783:odp_pool_print(): headroom 66 >>>> odp_pool.c:784:odp_pool_print(): tailroom 0 >>>> odp_pool.c:791:odp_pool_print(): seg length 1856 requested, 1936 >>>> used >>>> odp_pool.c:793:odp_pool_print(): pkt length 1856 requested, 1936 >>>> used >>>> odp_pool.c:795:odp_pool_print(): num bufs 564 >>>> odp_pool.c:797:odp_pool_print(): bufs available 564 >>>> odp_pool.c:798:odp_pool_print(): bufs in use 0 >>>> odp_pool.c:799:odp_pool_print(): buf allocs 0 >>>> odp_pool.c:800:odp_pool_print(): buf frees 0 >>>> odp_pool.c:801:odp_pool_print(): buf empty 0 >>>> odp_pool.c:803:odp_pool_print(): blk size 1936 >>>> odp_pool.c:805:odp_pool_print(): blks available 564 >>>> odp_pool.c:806:odp_pool_print(): blk allocs 0 >>>> odp_pool.c:807:odp_pool_print(): blk frees 0 >>>> odp_pool.c:808:odp_pool_print(): blk empty 0 >>>> odp_pool.c:809:odp_pool_print(): buf high wm value 282 >>>> odp_pool.c:810:odp_pool_print(): buf high wm count 0 >>>> odp_pool.c:811:odp_pool_print(): buf low wm value 141 >>>> odp_pool.c:812:odp_pool_print(): buf low wm count 0 >>>> odp_pool.c:813:odp_pool_print(): blk high wm value 282 >>>> odp_pool.c:814:odp_pool_print(): blk high wm count 0 >>>> odp_pool.c:815:odp_pool_print(): blk low wm value 141 >>>> odp_pool.c:816:odp_pool_print(): blk low wm count 0 >>>> main :: odp_pool_print done! >>>> odp_packet_io.c:230:setup_pktio_entry():eno1 uses socket_mmap >>>> created pktio:01, dev:eno1, queue mode (ATOMIC queues) >>>> default pktio01 >>>> odp_shared_memory.c:296:odp_shm_reserve(): DefaultPool: >>>> :: No huge pages, fall back to normal pages, >>>> check: /proc/sys/vm/nr_hugepages. >>>> odp_shared_memory.c:296:odp_shm_reserve(): queue1Pool0: >>>> :: No huge pages, fall back to normal pages, >>>> check: /proc/sys/vm/nr_hugepages. >>>> odp_shared_memory.c:296:odp_shm_reserve(): queue2Pool1: >>>> :: No huge pages, fall back to normal pages, >>>> check: /proc/sys/vm/nr_hugepages. >>>> odp_shared_memory.c:296:odp_shm_reserve(): queue3Pool2: >>>> :: No huge pages, fall back to normal pages, >>>> check: /proc/sys/vm/nr_hugepages. >>>> >>>> ---------------------------------------- >>>> CLASSIFIER EXAMPLE STATISTICS >>>> ---------------------------------------- >>>> CONFIGURATION >>>> >>>> COS VALUE MASK >>>> ---------------------------------------- >>>> queue1 192.168.10.11 FFFFFFFF >>>> queue2 10.130.69.0 000000FF >>>> queue3 10.130.68.0 FFFFFE00 >>>> >>>> RECEIVED PACKETS >>>> ---------------------------------------- >>>> queue1 |queue2 |queue3 |DefaultCos |Total Packets >>>> queue pool |queue pool |queue pool |queue pool | >>>> 845 845 |0 0 |0 0 |38 38 |883 >>>> >>>> >>>> On Mon, May 9, 2016 at 9:00 PM, Bill Fischofer < >>>> [email protected]> wrote: >>>> >>>>> >>>>> >>>>> On Mon, May 9, 2016 at 6:57 AM, nousi <[email protected]> wrote: >>>>> >>>>>> >>>>>> Hi All, >>>>>> >>>>>> Pleas help me in running ODP classifier example with huge pages. >>>>>> In ubuntu 15.10 by default interface naming as "eno1" and value in >>>>>> "/proc/sys/vm/nr_hugepages." is 1024. >>>>>> classifier example program could not able to use huge pages even >>>>>> though nr_hugepages value is non zero. >>>>>> I could able to run the classier example, but it is not using huge >>>>>> pages. >>>>>> >>>>>> >>>>>> >>>>>> *console log is pasted below for you >>>>>> reference.*root@odp/example/classifier# >>>>>> ./odp_classifier -i eno1 -m 0 -p >>>>>> "ODP_PMR_SIP_ADDR:192.168.10.11:FFFFFFFF:queue1" >>>>>> -p "ODP_PMR_SIP_ADDR:10.130.69.0:000000FF:queue2" -p >>>>>> "ODP_PMR_SIP_ADDR:10.130.68.0:FFFFFE00:queue3" >>>>>> odp_pool.c:104:odp_pool_init_global(): >>>>>> Pool init global >>>>>> odp_pool.c:105:odp_pool_init_global(): pool_entry_s size 8512 >>>>>> odp_pool.c:106:odp_pool_init_global(): pool_entry_t size 8512 >>>>>> odp_pool.c:107:odp_pool_init_global(): odp_buffer_hdr_t size 216 >>>>>> odp_pool.c:108:odp_pool_init_global(): >>>>>> odp_queue.c:130:odp_queue_init_global():Queue init ... >>>>>> odp_shared_memory.c:296:odp_shm_reserve(): odp_queues: >>>>>> No huge pages, fall back to normal pages, >>>>>> check: /proc/sys/vm/nr_hugepages. >>>>>> >>>>> >>>>> This is an informational message saying that the linux-generic >>>>> implementation was unable to allocate huge pages so it's falling back to >>>>> normal pages. I'm not sure why you're seeing that except that it seems >>>>> that some allocations may have been successful (those in odp_pool.c) while >>>>> those for queue initialization were not. >>>>> >>>>> I'll let others who are more expert in this area chime in with some >>>>> additional thoughts. >>>>> >>>>> >>>>>> >>>>>> >>>>>> >>>>>> Thanks & Regards, >>>>>> >>>>>> *B.Nousilal,* >>>>>> >>>>>> _______________________________________________ >>>>>> lng-odp mailing list >>>>>> [email protected] >>>>>> https://lists.linaro.org/mailman/listinfo/lng-odp >>>>>> >>>>>> >>>>> >>>> >>>> >>>> -- >>>> >>>> >>>> *Thanks & Regards,B.Nousilal,* >>>> >>>> _______________________________________________ >>>> lng-odp mailing list >>>> [email protected] >>>> https://lists.linaro.org/mailman/listinfo/lng-odp >>>> >>>> >>> >> >> >> -- >> >> >> *Thanks & Regards,B.Nousilal,* >> > >
_______________________________________________ lng-odp mailing list [email protected] https://lists.linaro.org/mailman/listinfo/lng-odp
