Mike,

I just realized that I forgot to publish my review.  I am offline ATM,
but I will publish it in the next couple of hours.

Do you plan to update your the patch in Review Board?

Sorry for the oversight,
-John




On Jun 12, 2013, at 2:26 AM, Mike Tutkowski
<mike.tutkow...@solidfire.com> wrote:

> Hi Edison, John, and Wei (and whoever else is reading this :) ),
>
> Just an FYI that I believe I have implemented all the areas we wanted
> addressed.
>
> I plan to review the code again tomorrow morning or afternoon, then send in
> another patch.
>
> Thanks for all the work on this everyone!
>
>
> On Tue, Jun 11, 2013 at 12:29 PM, Mike Tutkowski <
> mike.tutkow...@solidfire.com> wrote:
>
>> Sure, that sounds good.
>>
>>
>> On Tue, Jun 11, 2013 at 12:11 PM, Wei ZHOU <ustcweiz...@gmail.com> wrote:
>>
>>> Hi Mike,
>>>
>>> It looks the two feature do not have many conflicts in Java code, except
>>> the cloudstack UI.
>>> If you do not mind, I will merge disk_io_throttling branch into master
>>> this
>>> week, so that you can develop based on it.
>>>
>>> -Wei
>>>
>>>
>>> 2013/6/11 Mike Tutkowski <mike.tutkow...@solidfire.com>
>>>
>>>> Hey John,
>>>>
>>>> The SolidFire patch does not depend on the object_store branch, but - as
>>>> Edison mentioned - it might be easier if we merge the SolidFire branch
>>> into
>>>> the object_store branch before object_store goes into master.
>>>>
>>>> I'm not sure how the disk_io_throttling fits into this merge strategy.
>>>> Perhaps Wei can chime in on that.
>>>>
>>>>
>>>> On Tue, Jun 11, 2013 at 11:07 AM, John Burwell <jburw...@basho.com>
>>> wrote:
>>>>
>>>>> Mike,
>>>>>
>>>>> We have a delicate merge dance to perform.  The disk_io_throttling,
>>>>> solidfire, and object_store appear to have a number of overlapping
>>>>> elements.  I understand the dependencies between the patches to be as
>>>>> follows:
>>>>>
>>>>>        object_store <- solidfire -> disk_io_throttling
>>>>>
>>>>> Am I correct that the device management aspects of SolidFire are
>>> additive
>>>>> to the object_store branch or there are circular dependency between
>>> the
>>>>> branches?  Once we understand the dependency graph, we can determine
>>> the
>>>>> best approach to land the changes in master.
>>>>>
>>>>> Thanks,
>>>>> -John
>>>>>
>>>>>
>>>>> On Jun 10, 2013, at 11:10 PM, Mike Tutkowski <
>>>> mike.tutkow...@solidfire.com>
>>>>> wrote:
>>>>>
>>>>>> Also, if we are good with Edison merging my code into his branch
>>> before
>>>>>> going into master, I am good with that.
>>>>>>
>>>>>> We can remove the StoragePoolType.Dynamic code after his merge and
>>> we
>>>> can
>>>>>> deal with Burst IOPS then, as well.
>>>>>>
>>>>>>
>>>>>> On Mon, Jun 10, 2013 at 9:08 PM, Mike Tutkowski <
>>>>>> mike.tutkow...@solidfire.com> wrote:
>>>>>>
>>>>>>> Let me make sure I follow where we're going here:
>>>>>>>
>>>>>>> 1) There should be NO references to hypervisor code in the storage
>>>>>>> plug-ins code (this includes the default storage plug-in, which
>>>>> currently
>>>>>>> sends several commands to the hypervisor in use (although it does
>>> not
>>>>> know
>>>>>>> which hypervisor (XenServer, ESX(i), etc.) is actually in use))
>>>>>>>
>>>>>>> 2) managed=true or managed=false can be placed in the url field (if
>>>> not
>>>>>>> present, we default to false). This info is stored in the
>>>>>>> storage_pool_details table.
>>>>>>>
>>>>>>> 3) When the "attach" command is sent to the hypervisor in
>>> question, we
>>>>>>> pass the managed property along (this takes the place of the
>>>>>>> StoragePoolType.Dynamic check).
>>>>>>>
>>>>>>> 4) execute(AttachVolumeCommand) in the hypervisor checks for the
>>>> managed
>>>>>>> property. If true for an attach, the necessary hypervisor data
>>>>> structure is
>>>>>>> created and the rest of the attach command executes to attach the
>>>>> volume.
>>>>>>>
>>>>>>> 5) When execute(AttachVolumeCommand) is invoked to detach a volume,
>>>> the
>>>>>>> same check is made. If managed, the hypervisor data structure is
>>>>> removed.
>>>>>>>
>>>>>>> 6) I do not see an clear way to support Burst IOPS in 4.2 unless
>>> it is
>>>>>>> stored in the volumes and disk_offerings table. If we have some
>>> idea,
>>>>>>> that'd be cool.
>>>>>>>
>>>>>>> Thanks!
>>>>>>>
>>>>>>>
>>>>>>> On Mon, Jun 10, 2013 at 8:58 PM, Mike Tutkowski <
>>>>>>> mike.tutkow...@solidfire.com> wrote:
>>>>>>>
>>>>>>>> "+1 -- Burst IOPS can be implemented while avoiding implementation
>>>>>>>> attributes.  I always wondered about the details field.  I think
>>> we
>>>>> should
>>>>>>>> beef up the description in the documentation regarding the
>>> expected
>>>>> format
>>>>>>>> of the field.  In 4.1, I noticed that the details are not
>>> returned on
>>>>> the
>>>>>>>> createStoratePool updateStoragePool, or listStoragePool response.
>>>> Why
>>>>>>>> don't we return it?  It seems like it would be useful for clients
>>> to
>>>> be
>>>>>>>> able to inspect the contents of the details field."
>>>>>>>>
>>>>>>>> Not sure how this would work storing Burst IOPS here.
>>>>>>>>
>>>>>>>> Burst IOPS need to be variable on a Disk Offering-by-Disk Offering
>>>>>>>> basis. For each Disk Offering created, you have to be able to
>>>> associate
>>>>>>>> unique Burst IOPS. There is a disk_offering_details table. Maybe
>>> it
>>>>> could
>>>>>>>> go there?
>>>>>>>>
>>>>>>>> I'm also not sure how you would accept the Burst IOPS in the GUI
>>> if
>>>>> it's
>>>>>>>> not stored like the Min and Max fields are in the DB.
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> *Mike Tutkowski*
>>>>>>> *Senior CloudStack Developer, SolidFire Inc.*
>>>>>>> e: mike.tutkow...@solidfire.com
>>>>>>> o: 303.746.7302
>>>>>>> Advancing the way the world uses the cloud<
>>>>> http://solidfire.com/solution/overview/?video=play>
>>>>>>> *™*
>>>>>>
>>>>>>
>>>>>>
>>>>>> --
>>>>>> *Mike Tutkowski*
>>>>>> *Senior CloudStack Developer, SolidFire Inc.*
>>>>>> e: mike.tutkow...@solidfire.com
>>>>>> o: 303.746.7302
>>>>>> Advancing the way the world uses the
>>>>>> cloud<http://solidfire.com/solution/overview/?video=play>
>>>>>> *™*
>>>>
>>>>
>>>> --
>>>> *Mike Tutkowski*
>>>> *Senior CloudStack Developer, SolidFire Inc.*
>>>> e: mike.tutkow...@solidfire.com
>>>> o: 303.746.7302
>>>> Advancing the way the world uses the
>>>> cloud<http://solidfire.com/solution/overview/?video=play>
>>>> *™*
>>
>>
>>
>> --
>> *Mike Tutkowski*
>> *Senior CloudStack Developer, SolidFire Inc.*
>> e: mike.tutkow...@solidfire.com
>> o: 303.746.7302
>> Advancing the way the world uses the 
>> cloud<http://solidfire.com/solution/overview/?video=play>
>> *™*
>
>
>
> --
> *Mike Tutkowski*
> *Senior CloudStack Developer, SolidFire Inc.*
> e: mike.tutkow...@solidfire.com
> o: 303.746.7302
> Advancing the way the world uses the
> cloud<http://solidfire.com/solution/overview/?video=play>
> *™*

Reply via email to