angela wrote:
> jan damborsky ??:
>   
>> Hi Angela,
>>
>> Please see my response in line.
>>
>> Thank you,
>> Jan
>>
>> angela wrote:
>>     
>>> Hi Jan,
>>>
>>> Please see my comments in line.
>>>
>>> As for the test suite libti, now the sparc support and x86 bug fix 
>>> have been finished.
>>> You can find the baseline and Opensolaris 0811 RC2 test results in 
>>> the following page:
>>> http://opg-qe.central.sun.com/wiki/index.php/Install_libti
>>>       
>> I have go through the test report and it so far it looks good :-)
>> Thanks for testing ths
>>
>>     
>>> BTW, the sparc system I used to test is sun4u :)
>>>       
>> Since we will be supporting sun4v for January release,
>> could I please ask you to give it a try on sun4v as well ?
>> Thanks !
>>     
> I've tested, the result is the same. You may find the results at:
> http://opg-qe.central.sun.com/wiki/index.php/Install_libti
>   

Thanks ! It looks really good.

> And I also filed 2 test driver's CRs for tracking:
> 5794 <http://defect.opensolaris.org/bz/show_bug.cgi?id=5794>  nor     P3 
> OpenSol       qa-installer at defect.opensol...       NEW     
>       TI test driver should provide interface to create multiple zfs volumes
> 5795 <http://defect.opensolaris.org/bz/show_bug.cgi?id=5795>  nor     P3 
> OpenSol       qa-installer at defect.opensol...       NEW     
>       TI test driver should provide interface to create zfs root pool with 
> TI_ATTR_ZFS_RPOOL_PRESERVE setting to TRUE
>   

Thank you very much for capturing those
test driver deficiencies. I will work on them.

Cheers,
Jan

>
> Cheers,
> Angela
>   
>>     
>>> Thank you for your support,
>>> Angela
>>>
>>> jan damborsky ??:
>>>       
>>>> Hi Angela,
>>>>
>>>>
>>>> angela wrote:
>>>>         
>>>>> jan damborsky ??:
>>>>>           
>>>>>> Hi Angela,
>>>>>>
>>>>>> thanks for reporting those issues !
>>>>>>
>>>>>> Please see my comments in line.
>>>>>>
>>>>>> Jan
>>>>>>
>>>>>>
>>>>>> angela wrote:
>>>>>>  
>>>>>>             
>>>>>>> Hi Jan,
>>>>>>>
>>>>>>> When I tried to clean create zfs volume assertion, I found the 
>>>>>>> following
>>>>>>> defects:
>>>>>>>
>>>>>>> 1) it doesn't reflect failures if zfs volume creation pass but 
>>>>>>> enable
>>>>>>> dump on it fail. I filed CR
>>>>>>> <http://defect.opensolaris.org/bz/show_bug.cgi?id=5687>5687
>>>>>>> <http://defect.opensolaris.org/bz/show_bug.cgi?id=5687> to track it.
>>>>>>>
>>>>>>> BTW, by went through zfm_create_volumes(), I found it can create 
>>>>>>> several
>>>>>>> volumes at the same time, right?
>>>>>>>     
>>>>>>>               
>>>>>> Yes, that is correct.
>>>>>>
>>>>>>  
>>>>>>             
>>>>>>> But test_ti doesn't provide this functionality.
>>>>>>> Should we support it or not?
>>>>>>>     
>>>>>>>               
>>>>>> I am not sure, what is your opinion on this ?
>>>>>>   
>>>>>>             
>>>>> What I mean is that, it can only create one volume at the same time 
>>>>> by test_ti.
>>>>> There is no way to creating several volumes at the same time by 
>>>>> calling test_ti.
>>>>> If test_ti doesn't provide interface for this kind of calling, 
>>>>> zfm_create_volumes() can't be tested fully.
>>>>>           
>>>> Understood. I think it is a good idea to enhance test driver
>>>> for supporting this feature, I am just wondering what you
>>>> think about this from QE perspective ?
>>>>         
>>> Maybe we can let test_ti read a file which contains zvols attributes 
>>> per line.
>>> That is:
>>> To creating one zvol, calling as the following:
>>> test_ti -t l -p <zpool> -n <zvol_name> -z <zvol_size> -u <zvol_type>
>>> To creating multi zvols, calling as the following:
>>> test_ti -t l -p <zpool> -f <zvols_attributs_file>
>>> For each line in <zvols_attributs_file>, using white space to 
>>> separate each field, the format for each line is as:
>>> zvol_name zvol_size zvol_type
>>>       
>> That sounds reasonable :-)
>> Could you please file bug for this ?
>>
>>     
>>>>>>  
>>>>>>             
>>>>>>> 2) it failed to release rpool if it has volume with type swap or 
>>>>>>> dump. I
>>>>>>> filed CR5689 
>>>>>>> <http://defect.opensolaris.org/bz/show_bug.cgi?id=5689> to
>>>>>>> track it.
>>>>>>> For information I got, TI should release rpool successfully no 
>>>>>>> matter
>>>>>>> what kind of dataset it has.
>>>>>>>     
>>>>>>>               
>>>>>> Currently, TI assumes that ZFS volume dedicated to swap is called 
>>>>>> 'swap' and
>>>>>> ZFS volume created for dump is called 'dump'. This is compliant with
>>>>>> 'ZFS boot'
>>>>>>   
>>>>>>             
>>>>> So the volume name can only be "swap" and "dump", can not be any 
>>>>> other names, right?
>>>>>           
>>>> There is no technical restriction for this, but right now installer 
>>>> is only creating
>>>> swap and dump with that names. Speaking about dump, I think it is fine
>>>> to assume that it is always named dump, as it make sense to have only
>>>> one created on the system.
>>>> With respect to swap, I think that the approach might be different. 
>>>> People
>>>> are used to create swap devices at will and it might be possible that
>>>> AI installer might support more than one swap created.
>>>>
>>>>         
>>>>> If so,
>>>>> 1) CR5689 will not be a defect
>>>>>           
>>>> I think that we can assume always dump with name 'dump'
>>>> and should take care of any swap device created in the pool.
>>>>
>>>>         
>>>>> 2) test_ti should not accept names other than swap/dump for these 
>>>>> zvols
>>>>>           
>>>> I think that it is out of scope of test driver to do this kind of 
>>>> checks,
>>>> I would like to keep it generic, since the assumptions might change
>>>> in future.
>>>>
>>>>         
>>>>> 3) and I need to modify test assertions accordingly.
>>>>>           
>>>> That is a good suggestion. Right now, I would recommend
>>>> to always create swap with name 'swap' and dump with
>>>> name 'dump' - this will exactly reflect what installer does
>>>> today. Should this change in future, we would accommodate
>>>> test assertions to reflect that.
>>>>         
>>> So I'll modify test assertions.
>>>       
>> Thanks !
>>
>>     
>>>>>> specification (PSARC2006/370):
>>>>>>
>>>>>> ...
>>>>>>
>>>>>> 6.5 Swap/Dump
>>>>>>
>>>>>> On systems with zfs root, swap and dump will be allocated from within
>>>>>> the root pool.  This has several advantages:  it simplifies the 
>>>>>> process of formatting disks for install, and it provides the ability
>>>>>> to re-size the swap and dump areas without having to re-partition 
>>>>>> the disk.
>>>>>> Both swap and dump will be zvols.
>>>>>>
>>>>>> On systems with zfs root, there will be separate swap and dump zvols.
>>>>>> The reasons for this are:
>>>>>>
>>>>>>    1.  The implementation of swap and dump to zvols (not discussed
>>>>>>        here because it isn't a visible interface) does not allow the
>>>>>>        same device to be used for both.
>>>>>>    2.  The total space used for swap and dump is a relatively small
>>>>>>        fraction of a typical disk size so the cost of having separate
>>>>>>        swap and dump devices is small.
>>>>>>    3.  It will be common for systems to need different amounts of 
>>>>>>        swap and dump space.  When they are separate zvols, it's 
>>>>>> easier
>>>>>>        to define the proper size of each.
>>>>>>
>>>>>> Because swap is implemented as a zvol, features such as encryption
>>>>>> and checksumming will automatically work for swap space (it it 
>>>>>> particularly
>>>>>> important that encyrption be supported).
>>>>>>
>>>>>> The swap and dump zvols for a pool will be the datasets:
>>>>>>
>>>>>>     <rootpool>/swap
>>>>>>     <rootpool>/dump
>>>>>>
>>>>>> ...
>>>>>>
>>>>>> For those cases (ZFS volume for swap named <pool>/swap,
>>>>>> ZFS volume for dump named <pool>/dump) releasing of ZFS
>>>>>> pool should succeed. Or is it failing in that case ?
>>>>>>   
>>>>>>             
>>>>> If the volume name is "swap" and "dump", then the releasing can 
>>>>> succeed. See the log I attached.
>>>>> If the volume name is not swap/dump, then it will fail, as CR5689 
>>>>> filed.
>>>>>           
>>>> Thanks for testing this ! Please see my comment above about
>>>> other names used for swap and dump.
>>>>
>>>>         
>>>>>>  
>>>>>>             
>>>>>>> If it is, I think I should add more test assertions on "release 
>>>>>>> rpool"
>>>>>>> feature as well as "create an existent rpool" feature with 
>>>>>>> respect of
>>>>>>> different dataset in it, right?
>>>>>>>     
>>>>>>>               
>>>>>> If ZFS pool to be released contains either regular ZFS filesystem 
>>>>>> or volume,
>>>>>> TI should successfully release that pool. I agree that test 
>>>>>> assertion should
>>>>>> be probably added to cover this scenario.
>>>>>>   
>>>>>>             
>>>>> I'll try to add them.
>>>>> Does the codes of releasing the existent zpool ti_create_target 
>>>>> calling are the same with the codes ti_release_target calling? If 
>>>>> it is been reused, I will only add test assertions only on release 
>>>>> pool.
>>>>>           
>>>> No. Looking at the source code, the code path is different.
>>>> That said, I think it could be considered as a bug.
>>>> But since installer currently uses ti_release_target() for releasing
>>>> the pool, the path for destroying the pool in ti_create_target()
>>>> is not utilized.
>>>> Based on this, it should be sufficient to test releasing the pool
>>>> using ti_release_target().
>>>>
>>>>         
>>>>> BTW, I found when trying to create a zfs root pool, there is an 
>>>>> attribute named TI_ATTR_ZFS_RPOOL_PRESERVE.
>>>>> Is that mean, when trying to create an existent zfs root pool, if 
>>>>> setting this attribute, zfs root pool will be preserved, else TI 
>>>>> will release it firstly and then create a brand new zfs root pool, 
>>>>> right?
>>>>>           
>>>> Yes, that is correct. That said, this code path currently
>>>> can't release the pool if dump or swap are there,
>>>> please see my comment above.
>>>>
>>>>         
>>>>> If so, for preserved case, if the devices used to create zpool are 
>>>>> the same, then nothing happened, right? If the devices used are not 
>>>>> the same, what will happen?
>>>>>           
>>>> If TI_ATTR_ZFS_RPOOL_PRESERVE is set, pool is always left untouched
>>>> regardless of device name provided.
>>>>
>>>>         
>>>>> And also test_ti doesn't have interface to set preserve or not.
>>>>>           
>>>> That is correct - test_ti always sets TI_ATTR_ZFS_RPOOL_PRESERVE
>>>> to FALSE. We might enhance test_ti to provide this.
>>>>
>>>> Thank you,
>>>> Jan
>>>>
>>>>         
>>>>> Thanks,
>>>>> Angela
>>>>>           
>>>>>> Thank you,
>>>>>> Jan
>>>>>>
>>>>>>  
>>>>>>             
>>>>>>> Thanks,
>>>>>>> Angela
>>>>>>>     
>>>>>>>               
>>>>>>   
>>>>>>             
>
> _______________________________________________
> caiman-discuss mailing list
> caiman-discuss at opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/caiman-discuss
>   


Reply via email to