On Sun, Jan 28, 2018 at 6:05 AM, Lionel Orellana <lione...@gmail.com> wrote:

> Thanks Ben.
>
> I can reproduce it with the nodejs and wildfly bjild configs directly off
> the catalog.
>

and you haven't created those buildconfigs previously and then deleted
them, within that project?  This is the first time you're creating the
buildconfig in the project?


>
> My first thought was that the network issue was causing the build pod to
> be terminated and a second one is being created before the first one dies.
> Is that a possibility?
>

I don't think so, once we create a pod for a given build we consider it
done and shouldn't be creating another one, but the logs should shed some
light on that.


> I will get the logs tomorrow.
>
> On Sun, 28 Jan 2018 at 1:51 pm, Ben Parees <bpar...@redhat.com> wrote:
>
>> On Sat, Jan 27, 2018 at 4:06 PM, Lionel Orellana <lione...@gmail.com>
>> wrote:
>>
>>> Hi,
>>>
>>> I'm seeing an random error when running builds. Some builds fail very
>>> quickly with "build pod already exists". This is happening with a number of
>>> build configs and seems to occur when more than one build from different
>>> build configs are running at the same time.
>>>
>>
>> This can happen if you "reset" the build sequence number in your
>> buildconfig (buildconfig.status.latestversion).  Are you recreating
>> buildconfigs that may have previously existed and run builds, or editing
>> the buildconfig in a way that might be reseting the lastversion field to an
>> older value?
>>
>> Are you able to confirm whether or not a pod does exist for that build?
>> The build pod name will be in the form like "buildconfigname-buildnumber-
>> build"
>>
>> If you're able to recreate this consistently(assuming you're sure it's
>> not a case of you having recreated an old buildconfig or reset the
>> buildconfig lastversion sequence value), enabling level 4 logging in your
>> master and reproducing it and then providing us with the logs would be
>> helpful to trace what is happening.
>>
>>
>>
>>>
>>> There is a possibly related error in one of the nodes:
>>>
>>> Jan 28 07:26:39  atomic-openshift-node[10121]: W0128 07:26:39.735522
>>> 10848 docker_sandbox.go:266] NetworkPlugin cni failed on the stat
>>> us hook for pod "nodejs-26-build_bimorl": Unexpected command output
>>> nsenter: cannot open : No such file or directory
>>> Jan 28 07:26:39 atomic-openshift-node[10121]: with error: exit status 1
>>>
>>
>> This shouldn't be related to issues with pod creation (since this error
>> wouldn't occur until after the pod is created and attempting to run on a
>> node), but it's definitely something you'll want to sort out.  I've CCed
>> our networking lead into this thread.
>>
>>
>>
>>>
>>> -bash-4.2$ oc version
>>> oc v3.6.0+c4dd4cf
>>> kubernetes v1.6.1+5115d708d7
>>>
>>> Any ideas?
>>>
>>> Thanks
>>>
>>>
>>> Lionel.
>>>
>>> _______________________________________________
>>> users mailing list
>>> users@lists.openshift.redhat.com
>>> http://lists.openshift.redhat.com/openshiftmm/listinfo/users
>>>
>>>
>>
>>
>> --
>> Ben Parees | OpenShift
>>
>>


-- 
Ben Parees | OpenShift
_______________________________________________
users mailing list
users@lists.openshift.redhat.com
http://lists.openshift.redhat.com/openshiftmm/listinfo/users

Reply via email to