Hi,

I've added a Flow Stats Request procedure to my RYU app which gets called on a 
timer every 10s. I've also implemented a flow stats reply handler see code 
below...

    def send_flow_stats_request(self, dpid):
        # do we know about this switch
        if not dpid:
            return

        try:
            datapath = self._switches[dpid][0]
        except (KeyError):
            self.logger.info("send_flow_stats_request(), No switch defined for 
%s" %
                (dpid_to_str(dpid),))
            return

        ofp = datapath.ofproto
        ofp_parser = datapath.ofproto_parser

        match = ofp_parser.OFPMatch()
        table_id = 0xff
        out_port = ofp.OFPP_NONE
        req = ofp_parser.OFPFlowStatsRequest(datapath, 0,
                                             match,
                                             table_id,
                                             out_port)
        datapath.send_msg(req)


    @set_ev_cls(ofp_event.EventOFPFlowStatsReply, MAIN_DISPATCHER)
    def flow_stats_reply_handler(self, ev):

        msg = ev.msg
        dp = msg.datapath
        body = ev.msg.body

        if dp.id in self._switches:
            self._switches[dp.id][1].process_flow_stats(body)


I have a flow table on the switch which can grow to greater than 8,000 entries 
and when I send the request above, I would expect to receive 8,000 
OFPFlowStatsReply entries in the reply handler, however I'm only getting 
roughly 600 flow stats reply entries back.

Can anyone tell me if I've hit some sort of ceiling, there's nothing in the 
documentation to suggest there's a maximum number of entries that will be 
returned?

Thanks,
Alan.


------------------------------------------------------------------------------
Check out the vibrant tech community on one of the world's most
engaging tech sites, Slashdot.org! http://sdm.link/slashdot
_______________________________________________
Ryu-devel mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/ryu-devel

Reply via email to