On 11/22/2017 05:55 PM, Nir Soffer wrote:
>
>
> On Wed, Nov 22, 2017 at 7:15 PM Dafna Ron <[email protected]
> <mailto:[email protected]>> wrote:
>
>     Hi,
>
>     we have a failure in 004_basic_sanity.verify_add_vm_template.
>
>     The error seems to be a failure from the api request since
>     although I am seeing errors in the logs I am not sure they are the
>     cause.
>
>
>     **
>
>     *Link and headline ofto suspected patches: *
>
>     *
>
>     core: Cleanup BaseImagesCommand code -
>     https://gerrit.ovirt.org/#/c/83812/*
>
>     *
>     *
>
>     *Link to Job:*
>
>     *http://jenkins.ovirt.org/job/ovirt-master_change-queue-tester/3989*
>
>     *
>     *
>
>     *Link to all logs:*
>
>     
> *http://jenkins.ovirt.org/job/ovirt-master_change-queue-tester/3989/artifact*
>
>
>     *
>     *
>
>     *(Relevant) error snippet from the log: *
>
>     *
>
>     <error>
>
>     **
>
>     lago log:
>
>
>     *
>
>      File "/usr/lib64/python2.7/unittest/case.py", line 369, in run
>         testMethod()
>       File "/usr/lib/python2.7/site-packages/nose/case.py", line 197, in 
> runTest
>         self.test(*self.arg)
>       File "/usr/lib/python2.7/site-packages/ovirtlago/testlib.py", line 129, 
> in wrapped_test
>         test()
>       File "/usr/lib/python2.7/site-packages/ovirtlago/testlib.py", line 59, 
> in wrapper
>         return func(get_test_prefix(), *args, **kwargs)
>       File "/usr/lib/python2.7/site-packages/ovirtlago/testlib.py", line 68, 
> in wrapper
>         return func(prefix.virt_env.engine_vm().get_api(), *args, **kwargs)
>       File 
> "/home/jenkins/workspace/ovirt-master_change-queue-tester/ovirt-system-tests/basic-suite-master/test-scenarios/004_basic_sanity.py",
>  line 500, in verify_add_vm_template
>         lambda: api.vms.get(VM1_NAME).status.state == 'down',
>       File "/usr/lib/python2.7/site-packages/ovirtlago/testlib.py", line 271, 
> in assert_true_within_long
>         assert_equals_within_long(func, True, allowed_exceptions)
>       File "/usr/lib/python2.7/site-packages/ovirtlago/testlib.py", line 258, 
> in assert_equals_within_long
>         func, value, LONG_TIMEOUT, allowed_exceptions=allowed_exceptions
>       File "/usr/lib/python2.7/site-packages/ovirtlago/testlib.py", line 219, 
> in assert_equals_within
>         res = func()
>       File 
> "/home/jenkins/workspace/ovirt-master_change-queue-tester/ovirt-system-tests/basic-suite-master/test-scenarios/004_basic_sanity.py",
>  line 500, in <lambda>
>         lambda: api.vms.get(VM1_NAME).status.state == 'down',
>     '\'NoneType\' object has no attribute \'status\'\n-------------------- >> 
> begin captured logging << --------------------\novirtlago.testlib: ERROR:     
> * Unhandled exception in <function <lambda> at 0x4bafb90>\nTraceback (most 
> recent call last):\n  File 
> "/usr/lib/python2.7/site-packages/ovirtlago/testlib.py", line 219, in 
> assert_equals_within\n    res = func()\n  File 
> "/home/jenkins/workspace/ovirt-master_change-queue-tester/ovirt-system-tests/basic-suite-master/test-scenarios/004_basic_sanity.py",
>  line 500, in <lambda>\n    lambda: api.vms.get(VM1_NAME).status.state == 
> \'down\',\nAttributeError: \'NoneType\' object has no attribute 
> \'status\'\n--------------------- >> end captured logging << 
> ---------------------'
>
>
> This is a test error - code like this is always wrong:
>
>     api.vms.get(VM1_NAME).status.state == something
>
> In this case, api.vms.get() returned None. The test code
> should handle None and fail with meaningfull error, or maybe
> retry the call.
>  
>
>     engine log: **
>     **
>
>     {"jsonrpc": "2.0", "id": "32bfeb29-e70f-48b2-84c7-91cf22ea408c", 
> "result": {"status": "OK", "domain": "2d776fb7-eba8-4ee2-97b6-4350c28ddb39", 
> "voltype": "LEAF", "description": "", "parent": 
> "a1192fcc-4660-41a7-a1a5-a221808f603b", "format": "COW", "generation": 0, 
> "image": "dc033235-ab4b-4ef0-be79-6e8a22bfb0bc", "ctime": "1511365200", 
> "disktype": "DATA", "legality": "LEGAL", "mtime": "0", "apparentsize": 
> "1073741824", "children": [], "pool": "", "capacity": "41126400", "uuid": 
> "6a354b93-86af-41fb-a68b-10897ae3562f", "truesize": "1073741824", "type": 
> "SPARSE", "lease": {"owners": [], "version": null}}}�
>     2017-11-22 10:40:39,454-05 DEBUG 
> [org.ovirt.vdsm.jsonrpc.client.internal.ResponseWorker] (ResponseWorker) [] 
> Message received: {"jsonrpc": "2.0", "id": 
> "32bfeb29-e70f-48b2-84c7-91cf22ea408c", "result": {"status": "OK", "domain": 
> "2d776fb7-eba8-4ee2-97b6-4350c28ddb39", "voltype": "LEAF", "description": "", 
> "parent": "a1192fcc-4660-41a7-a1a5-a221808f603b", "format": "COW", 
> "generation": 0, "image": "dc033235-ab4b-4ef0-be79-6e8a22bfb0bc", "ctime": 
> "1511365200", "disktype": "DATA", "legality": "LEGAL", "mtime": "0", 
> "apparentsize": "1073741824", "children": [], "pool": "", "capacity": 
> "41126400", "uuid": "6a354b93-86af-41fb-a68b-10897ae3562f", "truesize": 
> "1073741824", "type": "SPARSE", "lease": {"owners": [], "version": null}}}
>     2017-11-22 10:40:39,454-05 INFO  
> [org.ovirt.engine.core.vdsbroker.vdsbroker.GetVolumeInfoVDSCommand] 
> (EE-ManagedThreadFactory-engineScheduled-Thread-34) 
> [c90b2f35-7a80-476f-9b23-78beecef7c37] FINISH, GetVolumeInfoVDSCommand, 
> return: 
> org.ovirt.engine.core.common.businessentities.storage.DiskImage@6903fdd6, log 
> id: 85b379d
>     2017-11-22 10:40:39,455-05 INFO  
> [org.ovirt.engine.core.vdsbroker.irsbroker.GetImageInfoVDSCommand] 
> (EE-ManagedThreadFactory-engineScheduled-Thread-34) 
> [c90b2f35-7a80-476f-9b23-78beecef7c37] FINISH, GetImageInfoVDSCommand, 
> return: 
> org.ovirt.engine.core.common.businessentities.storage.DiskImage@6903fdd6, log 
> id: 75d29166
>     2017-11-22 10:40:39,455-05 DEBUG 
> [org.ovirt.engine.core.common.di.interceptor.DebugLoggingInterceptor] 
> (EE-ManagedThreadFactory-engineScheduled-Thread-34) 
> [c90b2f35-7a80-476f-9b23-78beecef7c37] method: runVdsCommand, params: 
> [GetImageInfo, 
> GetImageInfoVDSCommandParameters:{storagePoolId='0762073b-b0ab-4ee2-8af9-dafa930a5fd4',
>  ignoreFailoverLimit='false', 
> storageDomainId='2d776fb7-eba8-4ee2-97b6-4350c28ddb39', 
> imageGroupId='dc033235-ab4b-4ef0-be79-6e8a22bfb0bc', 
> imageId='6a354b93-86af-41fb-a68b-10897ae3562f'}], timeElapsed: 165ms
>     2017-11-22 10:40:39,528-05 DEBUG 
> [org.ovirt.engine.core.common.di.interceptor.DebugLoggingInterceptor] 
> (EE-ManagedThreadFactory-engineScheduled-Thread-34) 
> [c90b2f35-7a80-476f-9b23-78beecef7c37] method: getVmsWithPlugInfo, params: 
> [dc033235-ab4b-4ef0-be79-6e8a22bfb0bc], timeElapsed: 68ms
>     2017-11-22 10:40:39,543-05 INFO  
> [org.ovirt.engine.core.utils.transaction.TransactionSupport] 
> (EE-ManagedThreadFactory-engineScheduled-Thread-34) [] transaction rolled back
>     2017-11-22 10:40:39,543-05 WARN  
> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] 
> (EE-ManagedThreadFactory-engineScheduled-Thread-34) [] Trying to release a 
> shared lock for key: '93d271d5-e0af-4795-b800-55465c7b8826TEMPLATE' , but 
> lock does not exist
>     2017-11-22 10:40:39,543-05 WARN  
> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] 
> (EE-ManagedThreadFactory-engineScheduled-Thread-34) [] Trying to release a 
> shared lock for key: 'fb76119e-9283-4c8b-9185-5d79e2481bc7DISK' , but lock 
> does not exist
>     2017-11-22 10:40:39,543-05 WARN  
> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] 
> (EE-ManagedThreadFactory-engineScheduled-Thread-34) [] Trying to release 
> exclusive lock which does not exist, lock key: 'vm1VM_NAME'
>     2017-11-22 10:40:39,543-05 INFO  [org.ovirt.engine.core.bll.AddVmCommand] 
> (EE-ManagedThreadFactory-engineScheduled-Thread-34) [] Lock freed to object 
> 'EngineLock:{exclusiveLocks='[vm1=VM_NAME]', 
> sharedLocks='[93d271d5-e0af-4795-b800-55465c7b8826=TEMPLATE, 
> fb76119e-9283-4c8b-9185-5d79e2481bc7=DISK]'}'
>     2017-11-22 10:40:39,556-05 ERROR 
> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] 
> (EE-ManagedThreadFactory-engineScheduled-Thread-34) [] EVENT_ID: 
> USER_ADD_VM_FINISHED_FAILURE(60), Failed to complete VM vm1 creation.
>     2017-11-22 10:40:39,556-05 ERROR 
> [org.ovirt.engine.core.bll.tasks.CommandCallbacksPoller] 
> (EE-ManagedThreadFactory-engineScheduled-Thread-34) [] Failed invoking 
> callback end method 'onSucceeded' for command 
> '2760a62f-e4ee-4dd2-b13a-1bb548e3e62d' with exception 'null', the callback is 
> marked for end method retries
>     2017-11-22 10:40:39,935-05 DEBUG 
> [org.ovirt.engine.core.common.di.interceptor.DebugLoggingInterceptor] 
> (EE-ManagedThreadFactory-engineScheduled-Thread-35) [] method: get, params: 
> [d25423f6-34bc-4cde-82fa-c0eedfc85386], timeElapsed: 3ms
>     2017-11-22 10:40:39,949-05 DEBUG 
> [org.ovirt.engine.core.common.di.interceptor.DebugLoggingInterceptor] 
> (EE-ManagedThreadFactory-engineScheduled-Thread-71) [] method: get, params: 
> [7c9ec3eb-2838-4b06-9408-326b67869bdc], timeElapsed: 2ms
>     2017-11-22 10:40:40,606-05 DEBUG 
> [org.ovirt.engine.core.aaa.filters.SsoRestApiAuthFilter] (default task-22) [] 
> Entered SsoRestApiAuthFilter
>
>     vdsm log: 
>     2017-11-22 10:36:11,973-0500 ERROR (jsonrpc/7) [storage.TaskManager.Task] 
> (Task='5b6ab546-b560-41fc-9b56-de465aa55eee') Unexpected error (task:875)
>     Traceback (most recent call last):
>       File "/usr/lib/python2.7/site-packages/vdsm/storage/task.py", line 882, 
> in _run
>         return fn(*args, **kargs)
>       File "<string>", line 2, in getAllTasksInfo
>       File "/usr/lib/python2.7/site-packages/vdsm/common/api.py", line 48, in 
> method
>         ret = func(*args, **kwargs)
>       File "/usr/lib/python2.7/site-packages/vdsm/storage/hsm.py", line 2216, 
> in getAllTasksInfo
>         allTasksInfo = self._pool.getAllTasksInfo()
>       File "/usr/lib/python2.7/site-packages/vdsm/storage/securable.py", line 
> 77, in wrapper
>         raise SecureError("Secured object is not in safe state")
>     SecureError: Secured object is not in safe state
>     2017-11-22 10:36:11,975-0500 INFO  (jsonrpc/7) [storage.TaskManager.Task] 
> (Task='5b6ab546-b560-41fc-9b56-de465aa55eee') aborting: Task is aborted: 
> u'Secured object is not in safe state' - code 100 (task:1181)
>     2017-11-22 10:36:11,976-0500 ERROR (jsonrpc/7) [storage.Dispatcher] 
> FINISH getAllTasksInfo error=Secured object is not in safe state 
> (dispatcher:86)
>     Traceback (most recent call last):
>       File "/usr/lib/python2.7/site-packages/vdsm/storage/dispatcher.py", 
> line 73, in wrapper
>         result = ctask.prepare(func, *args, **kwargs)
>       File "/usr/lib/python2.7/site-packages/vdsm/storage/task.py", line 108, 
> in wrapper
>         return m(self, *a, **kw)
>       File "/usr/lib/python2.7/site-packages/vdsm/storage/task.py", line 
> 1189, in prepare
>         raise self.error
>     SecureError: Secured object is not in safe state
>     2017-11-22 10:36:11,976-0500 INFO  (jsonrpc/7) [jsonrpc.JsonRpcServer] 
> RPC call Host.getAllTasksInfo failed (error 654) in 0.01 seconds 
> (__init__:573)
>     2017-11-22 10:36:11,979-0500 INFO  (jsonrpc/2) [vdsm.api] START 
> getAllTasksStatuses(spUUID=None, options=None) from=::1,52728, 
> task_id=172b5528-cc76-426a-81d3-43143b49a878 (api:46)
>     2017-11-22 10:36:11,979-0500 INFO  (jsonrpc/2) [vdsm.api] FINISH 
> getAllTasksStatuses error=Secured object is not in safe state from=::1,52728, 
> task_id=172b5528-cc76-426a-81d3-43143b49a878 (api:50)
>
>
> This is https://bugzilla.redhat.com/1426429
>
> Not related to the real error.
>
>     2017-11-22 10:36:11,980-0500 ERROR (jsonrpc/2) [storage.TaskManager.Task] 
> (Task='172b5528-cc76-426a-81d3-43143b49a878') Unexpected error (task:875)
>     Traceback (most recent call last):
>       File "/usr/lib/python2.7/site-packages/vdsm/storage/task.py", line 882, 
> in _run
>         return fn(*args, **kargs)
>       File "<string>", line 2, in getAllTasksStatuses
>       File "/usr/lib/python2.7/site-packages/vdsm/common/api.py", line 48, in 
> method
>         ret = func(*args, **kwargs)
>       File "/usr/lib/python2.7/site-packages/vdsm/storage/hsm.py", line 2176, 
> in getAllTasksStatuses
>         allTasksStatus = self._pool.getAllTasksStatuses()
>       File "/usr/lib/python2.7/site-packages/vdsm/storage/securable.py", line 
> 77, in wrapper
>         raise SecureError("Secured object is not in safe state")
>     SecureError: Secured object is not in safe state
>     2017-11-22 10:36:11,980-0500 INFO  (jsonrpc/2) [storage.TaskManager.Task] 
> (Task='172b5528-cc76-426a-81d3-43143b49a878') aborting: Task is aborted: 
> u'Secured object is not in safe state' - code 100 (task:1181)
>     2017-11-22 10:36:11,981-0500 ERROR (jsonrpc/2) [storage.Dispatcher] 
> FINISH getAllTasksStatuses error=Secured object is not in safe state 
> (dispatcher:86)
>     Traceback (most recent call last):
>       File "/usr/lib/python2.7/site-packages/vdsm/storage/dispatcher.py", 
> line 73, in wrapper
>         result = ctask.prepare(func, *args, **kwargs)
>       File "/usr/lib/python2.7/site-packages/vdsm/storage/task.py", line 108, 
> in wrapper
>         return m(self, *a, **kw)
>       File "/usr/lib/python2.7/site-packages/vdsm/storage/task.py", line 
> 1189, in prepare
>         raise self.error
>     SecureError: Secured object is not in safe state
>
>
> The way to find the relevant errors is to find the entry point of this
> flow
> in engine.log, and then follow this in vdsm.log using flow_id=xxx.
>
> Ideally we would send a uuid to engine for every operation, and be
> able to query the operation status using the uuid. This is the way
> storage jobs are managed in vdsm side since 4.1.
>
> If we had this every test would generate a uuids and debugging
> failing test would be much faster.
>
> Nir
>
>     *
>
>     *
>
>     *</error>*
>
>     **
>
>
yes, this was in earlier builds as well (although not always working)
which is why I usually try to find the command in engine and follow the
image uuid in vdsm.

However, in this case, I think the issue is in engine and not in vdsm at
all.

 
>
>     _______________________________________________
>     Devel mailing list
>     [email protected] <mailto:[email protected]>
>     http://lists.ovirt.org/mailman/listinfo/devel
>

_______________________________________________
Devel mailing list
[email protected]
http://lists.ovirt.org/mailman/listinfo/devel

Reply via email to