On 1/29/18 9:14 PM, Jamo Luhrsen wrote:
well job [c] (added that link now) had the expected trouble again. No
packet traces there, but I may have it working in this one:

https://jenkins.opendaylight.org/sandbox/job/jamo-netvirt-csit-1node-openstack-ocata-upstream-stateful-carbon/6/

#6 failed to stack :(
(no br-int created on control node)

trying again:
https://jenkins.opendaylight.org/sandbox/job/jamo-netvirt-csit-1node-openstack-ocata-upstream-stateful-carbon/7/

JamO

On 1/29/18 6:15 PM, Jamo Luhrsen wrote:
Arun,

something strange happened.

This job [a] I ran with just TRACE debugging enabled and the CSIT results
were as expected (Bad). Then, I got a patch in place to also enable the
packet capture and ran it on this job [b]. That job was all pass though
and I think it executed a lot faster than usual. I have no explanation
for that yet. Unfortunately, the packet capture logs were not archived.
I need to figure that out next.

hoping there is something in the karaf log between jobs [a] and [b] that
can give some more clues about what's going on.

job [c] is running now just to see if it will show up 100% passing again.

JamO

[a] https://logs.opendaylight.org/sandbox/vex-yul-odl-jenkins-2/jamo-netvirt-csit-1node-openstack-ocata-upstream-stateful-carbon/1/log_full.html.gz [b] https://logs.opendaylight.org/sandbox/vex-yul-odl-jenkins-2/jamo-netvirt-csit-1node-openstack-ocata-upstream-stateful-carbon/4/log_full.html.gz

[c]
https://jenkins.opendaylight.org/sandbox/job/jamo-netvirt-csit-1node-openstack-ocata-upstream-stateful-carbon/5/

On 1/28/18 9:14 PM, D Arunprakash wrote:
Hi Jamo,
I just had an initial look on Thursday and the test cases was failed for different and didn’t get a chance to look over the long weekend.
Today I don’t find the link working anymore, Could you please retrigger with 
packet capture enabled?

Regards,
Arun
-----Original Message-----
From: Jamo Luhrsen [mailto:[email protected]]
Sent: Thursday, January 25, 2018 12:45 PM
To: D Arunprakash <[email protected]>; Jamo Luhrsen <[email protected]>; 
Sam Hague <[email protected]>
Cc: openflowplugin-dev <[email protected]>; Manu B 
<[email protected]>
Subject: Re: [openflowplugin-dev] is dhcp issue fixed on carbon?

my bad. I fixed the problem.

this job should be better:

https://jenkins.opendaylight.org/sandbox/job/netvirt-csit-1node-openstack-ocata-jamo-upstream-stateful-carbon/12/

Thanks,
JamO

On 1/24/18 11:01 PM, D Arunprakash wrote:
Hi Jamo,
All the TCs are failed, please check.

Seems to be basic issue,

===========
Message:    Parent suite setup failed:
No keyword with name 'KarafKeywords.Execute_Controller_Karaf_Command_On_Background log:set TRACE org.opendaylight.openflowplugin.impl' found.
===========

Regards,
Arun

-----Original Message-----
From: Jamo Luhrsen [mailto:[email protected]]
Sent: Thursday, January 25, 2018 11:21 AM
To: D Arunprakash <[email protected]>; Sam Hague
<[email protected]>
Cc: openflowplugin-dev <[email protected]>;
Manu B <[email protected]>; Jamo Luhrsen <[email protected]>
Subject: Re: [openflowplugin-dev] is dhcp issue fixed on carbon?

I didn't add any extra tcpdump yet, but here is a job that will enable TRACE 
for ofp.impl.

it's only running the four connectivity/ folder suites.

https://jenkins.opendaylight.org/sandbox/job/netvirt-csit-1node-openst
ack-ocata-jamo-upstream-stateful-carbon/11/

JamO


On 1/24/18 8:04 PM, D Arunprakash wrote:
Please enable the logs for the below package,

org.opendaylight.openflowplugin.impl

Regards,

Arun

*From:*Sam Hague [mailto:[email protected]]
*Sent:* Thursday, January 25, 2018 9:32 AM
*To:* D Arunprakash <[email protected]>
*Cc:* Anil Vishnoi <[email protected]>; openflowplugin-dev
<[email protected]>; Vishal Thapar
<[email protected]>; Faseela K <[email protected]>;
Josh Hershberg <[email protected]>; Jamo Luhrsen
<[email protected]>; Manu B <[email protected]>
*Subject:* RE: is dhcp issue fixed on carbon?

What logs to enable? Or just the whole openflowplugin, which is very noisy.

Is packet capture on port 6653 good?

On Jan 24, 2018 10:54 PM, "D Arunprakash" <[email protected]
<mailto:[email protected]>> wrote:

      Hi Sam,

      We need to enable Openflowplugin logs and do packet capture for the
      time window when the tunnel port is delete and added back.

      Regards,

      Arun

      *From:*Anil Vishnoi [mailto:[email protected]
      <mailto:[email protected]>]
      *Sent:* Thursday, January 25, 2018 1:17 AM
      *To:* Sam Hague <[email protected] <mailto:[email protected]>>
      *Cc:* D Arunprakash <[email protected]
      <mailto:[email protected]>>; openflowplugin-dev
      <[email protected]
      <mailto:[email protected]>>; Vishal Thapar
      <[email protected] <mailto:[email protected]>>;
      Faseela K <[email protected] <mailto:[email protected]>>;
      Josh Hershberg <[email protected] <mailto:[email protected]>>;
      Jamo Luhrsen <[email protected] <mailto:[email protected]>>;
      Manu B <[email protected] <mailto:[email protected]>>


      *Subject:* Re: is dhcp issue fixed on carbon?

      Hi Sam,

      Looks like Arun is looking at it ?

      Arun, if you are not looking at it currently, please let me know I
      will take a look at it.

      Thanks

      Anil

      On Wed, Jan 24, 2018 at 4:25 AM, Sam Hague <[email protected]
      <mailto:[email protected]>> wrote:

          Adding openflow to thread.

          Anil, could someone take a look at this for carbon? We are
          seeing a connection flapping and end up missing port status
          updates. This leads to stale models and flows.

          This is blocking the carbon sr3.

          On Jan 24, 2018 12:58 AM, "D Arunprakash"
          <[email protected] <mailto:[email protected]>>
          wrote:

              Ignore my previous email.

              The tunnel port got deleted around 18:49:29.373 and added
              back on 18:52:46.26

              2018-01-23T18:49:29.373Z|01979|vconn|DBG|tcp:10.30.170.63:6653
              <http://10.30.170.63:6653>: sent (Success): OFPT_PORT_STATUS
              (OF1.3) (xid=0x0): DEL: 4(tun55fb50d0a2b):
              addr:3e:0c:ed:2e:a9:ba

              2018-01-23T18:52:46.261Z|03083|vconn|DBG|tcp:10.30.170.63:6653
              <http://10.30.170.63:6653>: sent (Success): OFPT_PORT_STATUS
              (OF1.3) (xid=0x0): ADD: 9(tun55fb50d0a2b):
              addr:8a:2f:9f:c6:fe:d9

              Immediately after tunnel delete, I’m seeing so multiple
              switch flaps for quite sometime,

              2018-01-23T18:49:35.155Z|02108|rconn|DBG|br-int<->unix:
              entering ACTIVE

              2018-01-23T18:49:35.155Z|02109|vconn|DBG|unix: sent
              (Success): OFPT_HELLO (OF1.3) (xid=0x75):

              version bitmap: 0x04

              2018-01-23T18:49:35.155Z|02110|vconn|DBG|unix: received:
              OFPT_HELLO (OF1.3) (xid=0x1):

              2018-01-23T18:49:35.307Z|02144|rconn|DBG|br-int<->unix:
              connection closed by peer

              2018-01-23T18:49:35.307Z|02145|rconn|DBG|br-int<->unix:
              entering DISCONNECTED

              2018-01-23T18:49:35.324Z|02146|rconn|DBG|br-int<->unix:
              entering ACTIVE

              Also, I’m seeing error in karaf log

              2018-01-23 18:49:29,378 | WARN  | entLoopGroup-7-3 |
              DeviceContextImpl                | 280 -
              org.opendaylight.openflowplugin.impl - 0.4.3.SNAPSHOT |
              writePortStatusMessage

              2018-01-23 18:49:29,379 | WARN  | entLoopGroup-7-3 |
              DeviceContextImpl                | 280 -
              org.opendaylight.openflowplugin.impl - 0.4.3.SNAPSHOT |
              submit transaction for write port status message

              2018-01-23 18:49:29,379 | WARN  | rd-dispatcher-23 |
              ShardDataTree                    | 184 -
              org.opendaylight.controller.sa
              <http://org.opendaylight.controller.sa>l-distributed-datastore
              - 1.5.3.SNAPSHOT | member-1-shard-inventory-operational:
              Store Tx member-1-datastore-operational-fe-0-chn-8-txn-11-0:
              Data validation failed for path
/(urn:opendaylight:inventory?revision=2013-08-19)nodes/node/node[{(urn:opendaylight:inventory?revision=2013-08-19)id=openflow:246869078989547}]/AugmentationIdentifier{childNames=[(urn:opendaylight:flow:inventory?revision=2013-08-19)port-number,
              (urn:opendaylight:flow:inventory?revision=2013-08-19)stale-group,
              
(urn:opendaylight:flow:inventory?revision=2013-08-19)supported-match-types,
              (urn:opendaylight:flow:inventory?revision=2013-08-19)table,
              (urn:opendaylight:flow:inventory?revision=2013-08-19)group,
              (urn:opendaylight:flow:inventory?revision=2013-08-19)manufacturer,
              (urn:opendaylight:flow:inventory?revision=2013-08-19)software,
              (urn:opendaylight:flow:inventory?revision=2013-08-19)ip-address,
              
(urn:opendaylight:flow:inventory?revision=2013-08-19)serial-number,
              
(urn:opendaylight:flow:inventory?revision=2013-08-19)table-features,
              
(urn:opendaylight:flow:inventory?revision=2013-08-19)supported-actions,
              (urn:opendaylight:flow:inventory?revision=2013-08-19)hardware,
              (urn:opendaylight:flow:inventory?revision=2013-08-19)description,
              
(urn:opendaylight:flow:inventory?revision=2013-08-19)switch-features,
              
(urn:opendaylight:flow:inventory?revision=2013-08-19)supported-instructions,
              (urn:opendaylight:flow:inventory?revision=2013-08-19)stale-meter,
(urn:opendaylight:flow:inventory?revision=2013-08-19)meter]}/(urn:opendaylight:flow:inventory?revision=2013-08-19)table/table[{(urn:opendaylight:flow:inventory?revision=2013-08-19)id=50}]/flow.

              
org.opendaylight.yangtools.yang.data.api.schema.tree.ModifiedNodeDoesNotExistException:
              Node
/(urn:opendaylight:inventory?revision=2013-08-19)nodes/node/node[{(urn:opendaylight:inventory?revision=2013-08-19)id=openflow:246869078989547}]/AugmentationIdentifier{childNames=[(urn:opendaylight:flow:inventory?revision=2013-08-19)port-number,
              (urn:opendaylight:flow:inventory?revision=2013-08-19)stale-group,
              
(urn:opendaylight:flow:inventory?revision=2013-08-19)supported-match-types,
              (urn:opendaylight:flow:inventory?revision=2013-08-19)table,
              (urn:opendaylight:flow:inventory?revision=2013-08-19)group,
              (urn:opendaylight:flow:inventory?revision=2013-08-19)manufacturer,
              (urn:opendaylight:flow:inventory?revision=2013-08-19)software,
              (urn:opendaylight:flow:inventory?revision=2013-08-19)ip-address,
              
(urn:opendaylight:flow:inventory?revision=2013-08-19)serial-number,
              
(urn:opendaylight:flow:inventory?revision=2013-08-19)table-features,
              
(urn:opendaylight:flow:inventory?revision=2013-08-19)supported-actions,
              (urn:opendaylight:flow:inventory?revision=2013-08-19)hardware,
              (urn:opendaylight:flow:inventory?revision=2013-08-19)description,
              
(urn:opendaylight:flow:inventory?revision=2013-08-19)switch-features,
              
(urn:opendaylight:flow:inventory?revision=2013-08-19)supported-instructions,
              (urn:opendaylight:flow:inventory?revision=2013-08-19)stale-meter,
(urn:opendaylight:flow:inventory?revision=2013-08-19)meter]}/(urn:opendaylight:flow:inventory?revision=2013-08-19)table/table[{(urn:opendaylight:flow:inventory?revision=2013-08-19)id=50}]/flow
              does not exist. Cannot apply modification to its children.

              We need to check why there is multiple switch disconnect and
              reconnect and how ofp handles the same.

              Regards,

              Arun

              *From:*Vishal Thapar
              *Sent:* Wednesday, January 24, 2018 9:52 AM
              *To:* Faseela K <[email protected]
              <mailto:[email protected]>>; Sam Hague
              <[email protected] <mailto:[email protected]>>; Josh
              Hershberg <[email protected]
              <mailto:[email protected]>>; D Arunprakash
              <[email protected] <mailto:[email protected]>>
              *Cc:* Jamo Luhrsen <[email protected]
              <mailto:[email protected]>>; Manu B <[email protected]
              <mailto:[email protected]>>
              *Subject:* RE: is dhcp issue fixed on carbon?

              Missed adding most important detail and added Arun.

              Inventory operational is still showing old port and new port
              for some reason. I guess that is what caused problems.

https://logs.opendaylight.org/releng/vex-yul-odl-jenkins-1/netvirt-cs
i
t-1node-openstack-ocata-gate-stateful-carbon/263/log_02_l3.html.gz#s1
-
t25-k4-k2-k1-k2-k56

{"id":"openflow:246869078989547:4","flow-node-inventory:supported":""
,
"flow-node-inventory:peer-features":"","flow-node-inventory:port-numb
e
r":4,"flow-node-inventory:hardware-address":"3e:0c:ed:2e:a9:ba","flow
-
node-inventory:current-feature":"","flow-node-inventory:maximum-speed"
:0,"flow-node-inventory:reason":"add","flow-node-inventory:configurat
i
on":"","flow-node-inventory:advertised-features":"","flow-node-invent
o
ry:current-speed":0,"flow-node-inventory:name":"tun55fb50d0a2b","flow
-
node-inventory:state":{"link-down":false,"blocked":false,"live":false
}
}

{"id":"openflow:246869078989547:9","flow-node-inventory:supported":""
,
"flow-node-inventory:peer-features":"","flow-node-inventory:port-numb
e
r":9,"flow-node-inventory:hardware-address":"8a:2f:9f:c6:fe:d9","flow
-
node-inventory:current-feature":"","flow-node-inventory:maximum-speed"
:0,"flow-node-inventory:reason":"add","flow-node-inventory:configurat
i
on":"","flow-node-inventory:advertised-features":"","flow-node-invent
o
ry:current-speed":0,"flow-node-inventory:name":"tun55fb50d0a2b","flow
-
node-inventory:state":{"link-down":false,"blocked":false,"live":false
}
}

              OVS output from same set of logs:

https://logs.opendaylight.org/releng/vex-yul-odl-jenkins-1/netvirt-cs
i
t-1node-openstack-ocata-gate-stateful-carbon/263/log_02_l3.html.gz#s1
-
t25-k4-k1-k3-k1-k11-k4

              9(tun55fb50d0a2b): addr:8a:2f:9f:c6:fe:d9

                   config: 0

                   state: 0

                   speed: 0 Mbps now, 0 Mbps max

              So for now I’d peg it as OFPlugin issue. It didn’t detect or
              inform us of old port delete and that is why we didn’t
              delete old flows. Though wondering if something else in IFM
              code could’ve handled it, but don’t think we handle OfPort
              number changes, expect a delete+add in such scenarios.
              Faseela can pitch in why we have service binding entry with
              new port number but flow is still using old one.

              Regards,

              Vishal.

              *From:*Vishal Thapar
              *Sent:* 24 January 2018 09:26
              *To:* Faseela K <[email protected]
              <mailto:[email protected]>>; Sam Hague
              <[email protected] <mailto:[email protected]>>; Josh
              Hershberg <[email protected] <mailto:[email protected]>>
              *Cc:* Jamo Luhrsen <[email protected]
              <mailto:[email protected]>>; Manu B <[email protected]
              <mailto:[email protected]>>
              *Subject:* RE: is dhcp issue fixed on carbon?

              Quick analysis:

              Not related to policy stuff. Service binding has entry for
              the new port number but table 220 flow is still using old
              port number.

              { "bound-services": [ { "flow-cookie": 134217735,
              "flow-priority": 9, "instruction": [ { "apply-actions": {
              "action": [ { "order": 0, "output-action": { "max-length":
              0, "output-node-connector": "*9*" } } ] }, "order": 0 } ],
              "service-name": "default.tun55fb50d0a2b",
              "service-priority": 9, "service-type":
              "interface-service-bindings:service-type-flow-based" } ],
              "interface-name": "tun55fb50d0a2b", "service-mode":
              "interface-service-bindings:service-mode-egress" }

{"id":"246869078989547.220.tun55fb50d0a2b.0","priority":9,"table_id":
2
20,"installHw":true,"hard-timeout":0,"match":{"openflowplugin-extensi
o
n-general:extension-list":[{"extension-key":"openflowplugin-extension
-
nicira-match:nxm-nx-reg6-key","extension":{"openflowplugin-extension-
n
icira-match:nxm-nx-reg":{"value":4096,"reg":"nicira-match:nxm-nx-reg6"
}}}]},"cookie":134217735,"flow-name":"default.tun55fb50d0a2b","strict"
:true,"instructions":{"instruction":[{"order":0,"apply-actions":{"act
i
on":[{"order":0,"output-action":{"max-length":0,"output-node-connecto
r ":"*4*"}}]}}]},"barrier":false,"idle-timeout":0}

              cookie=0x8000007, duration=403.965s, table=220, n_packets=0,
              n_bytes=0, priority=9,reg6=0x1000 actions=output:*4*

              In OVS logs you can see this tunnel port getting deleted and
              then coming back in with a different OfPort.

https://logs.opendaylight.org/releng/vex-yul-odl-jenkins-1/netvirt-cs
i
t-1node-openstack-ocata-gate-stateful-carbon/263/compute_2/ovs-vswitc
h
d.log.gz

              It goes from 4 to 9. This happens due to clean up in
              previous suite which doesn’t actually clean up everything
              and leaves entry for that old service binding. Can confirm
              it from interfaces-state entry for same port in first and
              second suites. So we have stale flows and stale service
              bindings for old tunnel port. Could probably check with
              OFPlugin how they handle update of a flow, may probably not
              work.

              We need to check if cleanup has been done completely before
              moving to next suite. This is where the work we been doing
              on tools comes in.

              Regards,

              Vishal.

              *From:*Faseela K
              *Sent:* 24 January 2018 08:10
              *To:* Sam Hague <[email protected]
              <mailto:[email protected]>>; Josh Hershberg
              <[email protected] <mailto:[email protected]>>
              *Cc:* Vishal Thapar <[email protected]
              <mailto:[email protected]>>; Jamo Luhrsen
              <[email protected] <mailto:[email protected]>>; Manu B
              <[email protected] <mailto:[email protected]>>
              *Subject:* RE: is dhcp issue fixed on carbon?

              Looks more or less similar issue, tunnel flow is programmed
              in table 220 with older tunnel’s port number, which was
              deleted in l2 suite. However policy code has not kicked in.
              I will take a detailed look on what is causing this issue now.

              Thanks,

              Faseela

              *From:*Faseela K
              *Sent:* Wednesday, January 24, 2018 7:48 AM
              *To:* 'Sam Hague' <[email protected]
              <mailto:[email protected]>>; Josh Hershberg
              <[email protected] <mailto:[email protected]>>
              *Cc:* Vishal Thapar <[email protected]
              <mailto:[email protected]>>; Jamo Luhrsen
              <[email protected] <mailto:[email protected]>>; Manu B
              <[email protected] <mailto:[email protected]>>
              *Subject:* RE: is dhcp issue fixed on carbon?

              Thanks Sam for initial triaging.

              I will take a look at this.

              *From:*Sam Hague [mailto:[email protected]]
              *Sent:* Wednesday, January 24, 2018 6:54 AM
              *To:* Faseela K <[email protected]
              <mailto:[email protected]>>; Josh Hershberg
              <[email protected] <mailto:[email protected]>>
              *Cc:* Vishal Thapar <[email protected]
              <mailto:[email protected]>>; Jamo Luhrsen
              <[email protected] <mailto:[email protected]>>; Manu B
              <[email protected] <mailto:[email protected]>>
              *Subject:* Re: is dhcp issue fixed on carbon?

              OK, seems pretty consistent that table 220 flows are not
              showing up. Vishal, Faseela, can you see if it is like the
              policymgr one where the bind/unbind was wrong? That seems
              the closest culprit as those were the last patches merged.

              Here is another case where the table 220 flow is missing in
              suite [5] of job [6]. This time the port missing is a tunnel
              port. "9(tun55fb50d0a2b): addr:8a:2f:9f:c6:fe:d9" is missing
              from table 220. And then in suite [7] of the same job this
              port has the same issue where the tunnel port is missing:
              "16(tap28760838-a7): addr:fe:16:3e:26:0a:e3"

              [5]
https://logs.opendaylight.org/releng/vex-yul-odl-jenkins-1/netvirt-cs
i
t-1node-openstack-ocata-gate-stateful-carbon/263/log_02_l3.html.gz#s1
-
t25-k4-k1-k3-k1-k12-k4

              [6]
https://logs.opendaylight.org/releng/vex-yul-odl-jenkins-1/netvirt-cs
i t-1node-openstack-ocata-gate-stateful-carbon/263/log_02_l3.html.gz

              [7]
https://logs.opendaylight.org/releng/vex-yul-odl-jenkins-1/netvirt-cs
i
t-1node-openstack-ocata-gate-stateful-carbon/263/log_04_security_grou
p
.html.gz

              On Tue, Jan 23, 2018 at 3:33 PM, Sam Hague
              <[email protected] <mailto:[email protected]>> wrote:

                  further details for Josh since the original email
                  doesn't have many...

                  - so the "l3.Check Vm Instances Have Ip Address" test
                  fails with the net1 not being able to get all the vm ips
                  for it's three vms.

                  - '[u'None', u'31.0.0.9', u'31.0.0.10']' contains 'None'
                  - this means the first vm of the three did not get a
ip

https://logs.opendaylight.org/releng/vex-yul-odl-jenkins-1/netvirt-cs
i
t-1node-openstack-ocata-upstream-stateful-carbon/298/log_02_l3.html.g
z
#s1-t11-k8

                  - looks at the neutron ports to find which port goes
                  with vm1

https://logs.opendaylight.org/releng/vex-yul-odl-jenkins-1/netvirt-cs
i
t-1node-openstack-ocata-upstream-stateful-carbon/298/log_02_l3.html.g
z
#s1-t11-k9-k1-k4-k1-k2

                  get the missing ip as 31.0.0.6, then look at next log to
                  get the port

                  - look at the 31.0.0.x addresses, we know 31,0.0

https://logs.opendaylight.org/releng/vex-yul-odl-jenkins-1/netvirt-cs
i
t-1node-openstack-ocata-upstream-stateful-carbon/298/log_02_l3.html.g
z
#s1-t11-k9-k1-k8-k2

                  3862fa17-4e7d-4d41-9237-c372fca11c03 | |
                  fa:16:3e:96:06:3f | ip_address='31.0.0.6',
                  subnet_id='697e1b34-1adb-4299-b50f-6527b15260fd' |
ACTIVE |

                  - I know the first vm (and second) are both on the
                  compute_1 so look at the ovs logs on compute_1

https://logs.opendaylight.org/releng/vex-yul-odl-jenkins-1/netvirt-cs
i
t-1node-openstack-ocata-upstream-stateful-carbon/298/log_02_l3.html.g
z
#s1-t11-k9-k2-k1-k2-k1-k11-k4

                  - compute_1, in the ofctl show br-int, we see port 7

                  7(tap3862fa17-4e): addr:fe:16:3e:96:06:3f

                  - then check flows to see if there is a table 220 flow
                  for port 7

https://logs.opendaylight.org/releng/vex-yul-odl-jenkins-1/netvirt-cs
i
t-1node-openstack-ocata-upstream-stateful-carbon/298/log_02_l3.html.g
z
#s1-t11-k9-k2-k1-k2-k1-k12-k4

                  And the table 220 flow for port 7 is not there, so the
                  vm can't get an IP.

                  [3] is the patch vishal pushed to fix a similar issue
                  the first time we saw this. What we found is that the
                  elan tag was being reused, because a port was deleted
                  and then a new one created and the elan tag reused. So
                  you ended up with a tunnel port stomping on a vm port.

                  [3] https://git.opendaylight.org/gerrit/#/c/67009/

                  On Tue, Jan 23, 2018 at 3:07 PM, Sam Hague
                  <[email protected] <mailto:[email protected]>> wrote:

                      Adding Josh to thread.

                      On Tue, Jan 23, 2018 at 2:25 PM, Faseela K
                      <[email protected]
                      <mailto:[email protected]>> wrote:

                          Manu,

                              Could you please take a look at the DHCP
                          failure in the below run?

                               I am caught up with something else, will
                          help you out in initial triaging.

                          Thanks,

                          Faseela

                          *From:*Sam Hague [mailto:[email protected]
                          <mailto:[email protected]>]
                          *Sent:* Monday, January 22, 2018 10:57 PM
                          *To:* Vishal Thapar <[email protected]
                          <mailto:[email protected]>>; Faseela K
                          <[email protected]
                          <mailto:[email protected]>>; Jamo Luhrsen
                          <[email protected] <mailto:[email protected]>>
                          *Subject:* is dhcp issue fixed on carbon?

                          Vishal, Faseela,

                          can you look at this job run to see if the issue
                          you fixed with the policymgr binding is fixed?
                          in this build the whole poligymgr bundle has
                          been removed. This is carbon so I just removed
                          the whole bundle as we would never use it. Could
                          that have uncovered something that the code was
                          doing? If so, then even master and nitrogen
                          should have the issue since there we disabled
                          building policymgr - so should be the same as
                          removing it.

                          Other thing, merged in carbon is the bind/unbind
                          patches for elan and dhcp. Could those have an
                          impact?

                          Thanks, Sam

                          I don't see "7(tap3862fa17-4e):
                          addr:fe:16:3e:96:06:3f" pop up in the table 220
                          flows which was the problem before.

                          3862fa17-4e7d-4d41-9237-c372fca11c03 | |
                          fa:16:3e:96:06:3f | ip_address='31.0.0.6',
                          subnet_id='697e1b34-1adb-4299-b50f-6527b15260fd'
                          | ACTIVE |

                          Thanks, Sam

                          [1]
https://logs.opendaylight.org/releng/vex-yul-odl-jenkins-1/netvirt-cs
i
t-1node-openstack-ocata-upstream-stateful-carbon/298/log_02_l3.html.g
z
#s1-t11-k9-k2-k1-k2-k1-k11-k4



      --

      Thanks

      Anil


_______________________________________________
openflowplugin-dev mailing list
[email protected]
https://lists.opendaylight.org/mailman/listinfo/openflowplugin-dev

_______________________________________________
openflowplugin-dev mailing list
[email protected]
https://lists.opendaylight.org/mailman/listinfo/openflowplugin-dev

Reply via email to