Re: [Openstack] Grizzly Bear

2013-07-19 Thread Jonathan Proulx
As long as the network the interfaces connect to is properly configured for
the traffic you are trying to put across them there's no reason I know of
that OpenStack should care if they are physical or virtual (or tunnels or
bonded interfaces, or what).

That said I've not tried virtual interfaces.
___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Checkpoint VM and roll back

2013-04-25 Thread Jonathan Proulx
On Thu, Apr 25, 2013 at 2:06 PM, Tim Bell tim.b...@cern.ch wrote:


 I'd like to be able to

 1. checkpoint a running virtual machine
 2. run a test
 3. rollback to the checkpoint from step 1

 Has anyone had experience of doing this using OpenStack (such as with
 snapshots) ?



For slow cycling tests like revising our base operating system images or
deploying large new config management blocks I:

1) boot from a base image
2) deploy and test change
3) on pass snapshot on fail destroy and relaunch last good snapshot
4) recurse as required for multi layered change tests

For faster cycles OpenStack snapshoting may be too cumbersome.  I've used
LVM and ZFS snapshots to provide in operating system roll back and have
read of others doing CoW overlays (unionfs as similar) but that's operating
system specific on openstack.

-Jon
___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] multi-host mode in quantum

2013-04-04 Thread Jonathan Proulx
Hi All,

On Thu, Apr 4, 2013 at 3:23 PM, Henry Gessau ges...@cisco.com wrote:
Unfortunately, I don't think multiple network nodes is the same
multi-host network mode that Xin is asking about.

The following did not make it into grizzly and is now targeted for havana:
https://blueprints.launchpad.net/quantum/+spec/quantum-multihosthttps://blueprints.launchpad.net/quantum/+spec/quantum-multihost

I'm hoping that means there still needs to be a central dhcp server but the
multiple L3 agents mean actual traffic from VMs can be managed directly on
the compute note they are running on and doesn't need to be sent back to a
different network node before being nat'ed or tagged?

Multi-host was the single most important feature for me in the quantum blue
prints, if nat'ed networks need to be piped through a gateway box other
than the physical host the instance is on quantum remains a no go for me.
Having a central (but redundant) dhcp with distributed NAT may actually be
an improvement over having to run dnsmasq literally everywhere.

-Jon
___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Network layout/planning for HA

2013-03-22 Thread Jonathan Proulx
On Fri, Mar 22, 2013 at 03:25:56PM -0400, Samuel Winchenbach wrote:

:It looks like 802.3ad is for making parallel, redundant connections.   I
:only have two NICs in the server; I don't think I would want to bind by
:internal and external NICs together would I?

What you want is up to you :)

You could though if you then use VLANs to segregate the internal and
external traffic over the bonded link.  We're not doing this in our
OpenStack but do have older production virtulization systems where
we're using VLANs over bonded NICs.

-Jon

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Help with VMs

2013-02-19 Thread Jonathan Proulx
On Tue, Feb 19, 2013 at 1:25 PM, Harvey West harvey.w...@btinternet.com wrote:

 This boots ok.  kvm -m 2048 -hda freeBSD.img -boot c
 (note: did not use virtio mods. Assumed these were just optimized NIC/SCSI
 drivers. Which I can live without for the time being)

I ran into the same isssue with FreeBSD images. These are
paravirtualized drivers which are absolutely required to get the VM to
work.  Linux images usually have them already but for FreeBSD (and
Windows) you need to install them by hand. note that when using kvm by
hand you can specify the driver emulation required so this is why it
boots by hand but not via openstack (you could also configure
openstack to use different kvm settings, but the virtio driver is
faster and installing the drivers on FreeBSD isn't too hard. I'm not
much of a *BSD guy but I managed)

-Jon

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


[Openstack] Help, erroneous resource tracker preventing instances from starting

2013-01-07 Thread Jonathan Proulx
Hi All,

I have a growing problem in which compute nodes are puzzlingly over
reporting their resource utilization and thus appearing to be over utilized
when they are in fact empty.  System is Ubuntu 12.04 using cloud archive
Folsom (2012.2-0ubuntu5~cloud0) problem appeared on a single node after
upgrade from Essex some months ago and has now grown to 5 nodes (the lowest
numbered 5 nodes both by IP and lexically by name)

For example on the compute node nova-1:

2013-01-07 10:39:43 INFO nova.compute.manager [-] Updating host status
2013-01-07 10:41:02 AUDIT nova.compute.resource_tracker [-] Free ram (MB):
-397134
2013-01-07 10:41:02 AUDIT nova.compute.resource_tracker [-] Free disk (GB):
-3318
2013-01-07 10:41:02 AUDIT nova.compute.resource_tracker [-] Free VCPUS: -215
2013-01-07 10:41:02 INFO nova.compute.resource_tracker [-] Compute_service
record updated for nova-1

Oddly even though no instances are scheduled teh resource utilization does
vary, for example in the last 5hours:

root@nova-1:~# grep 'Free VCPUS:' /var/log/nova/nova-compute.log|awk
'{print $NF}'|sort -n |uniq -c
156 -218
  3 -216
  5 -215
  2 -214
  2 -212
  1 -211
  1 -210
  5 -209
  5 -208

# but no instances are running
root@nova-1:~# virsh list
 IdName   State


root@nova-1:~#

# nor does OpenStack seem to *think* any instances are running or reserved
by any projects
# as seen by nova-manage service describe_resource nova-1

HOST  PROJECT cpu mem(mb) hdd
nova-1  (total)24   48295 602
nova-1  (used_now)233  4331413740
nova-1  (used_max)  0   0   0
# note lack of a list of tenants here

I can't replicate the issue intetionally but also can't clear appaerent
resource utilization.  Tried direct manipulation of the database but that
gets reset by computenode reports, tried rebooting the nodes.  I can always
fall back to just reinstalling them, but since this is still a
pre-production cluster I'd liek to understand what is happening.

Anyone have an insight into why nova.compute.resource_tracker is so
confused or how I can force it to understand what resources are in use?
Operationally it isn't painful to reinstall, but it does hurt a bit not
knowing what's going on here.

Thanks,
-Jon
___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] How to deploy openstack automatically in your env.

2012-12-03 Thread Jonathan Proulx
On Mon, Dec 03, 2012 at 06:21:54PM +0800, Lei Zhang wrote:
:It is a wired thing that the openstack is a python project. But many tools
:for it are build on ruby?

Puppet (http://puppetlabs.com/) and Chef
(http://www.opscode.com/chef), the main players in configuration
management are both written in Ruby and manage more than just
OpenStack so the code just isn't related

I can't be much help on your origina RHEL question as I run a Debian / 
Ubuntu shop.  But while it is true that deployment tools are better
tested on Ubuntu, certainly any puppe tor chef based solutions
*should* work on RHEL if they don't it would be worth reporting their
failures to the developers of those tools.

There is a Fedora OpenStack wiki at
http://fedoraproject.org/wiki/OpenStack which is linked from
http://www.openstack.org/software/start/ which has other RedHat family
specific links at the bottom of the page

-Jon

:
:
:On Mon, Dec 3, 2012 at 5:44 PM, Joe Breu joseph.b...@rackspace.com wrote:
:
:  Hi Lei,
:
:  We have chef cookbooks to install Openstack located at
: http://github.com/rcbops/chef-cookbooks.
:
:   ---
: Joseph Breu
: Deployment Engineer
: Rackspace Private Cloud
: 210-312-3508
:
:  On Dec 3, 2012, at 9:15 AM, Lei Zhang wrote:
:
:  Hi all,
:
: I search the internet for days and found several automatically tool.
: Including
:
:- devstack
:- puppet+pupet-openstack
:- stackops
:- OneStack
:
: But It seems that all the scripts are well tested on ubuntu not RHEL. How
: could you guys to deploy the openstack automatically, especially on RHEL.
:  --
: Lei Zhang
:
:  Blog: http://jeffrey4l.github.com
: twitter/weibo: @jeffrey4l
:
:  ___
: Mailing list: https://launchpad.net/~openstack
: Post to : openstack@lists.launchpad.net
: Unsubscribe : https://launchpad.net/~openstack
: More help   : https://help.launchpad.net/ListHelp
:
:
:
:
:
:-- 
:Lei Zhang
:
:Blog: http://jeffrey4l.github.com
:twitter/weibo: @jeffrey4l

:___
:Mailing list: https://launchpad.net/~openstack
:Post to : openstack@lists.launchpad.net
:Unsubscribe : https://launchpad.net/~openstack
:More help   : https://help.launchpad.net/ListHelp


___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


[Openstack] [Nova] used_max -vs used_now resources?

2012-11-25 Thread Jonathan Proulx
Hi All,

I'm trying to make sense of 'nova-manage service describe_resource' output
from Folsom running on Ubuntu 12.04 with KVM back end.

I'd thought used_now accounted for running resource use and used_max
included reservations that had not yet been instantiated.  That assumption
seemed to match what I saw under Essex.

now I have situations like this:

HOST  PROJECT cpu mem(mb) hdd
nova-3  (total)24   48295 605
nova-3  (used_now)140  2872323400
nova-3  (used_max)  48192  90
nova-3   956ae20bbb444a8c8f149729198aec63   4
8192  90

where used_now is wildly higher than it seems it should be.  looking on
the compute node itself verifies there are 4 single vCPU instances running
there.

On a different system reporting a high used_now but zero actually running
instances I tried rebooting the compute node hoping to clear state (since
the node pushes these numbers to the database correcting them there doesn't
work).  On reboot it came up claiming over 160 cpus in use.

Is this a bug, a misconfiguration, or a misunderstanging?

Thanks,
-Jon
___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


[Openstack] How to I unconfuse my resource_tracker?

2012-11-06 Thread Jonathan Proulx
My compute nodes are confused about how many resource they have free.  I
suspect this is largely due to RPC timeouts I was experiencing due to a
misconfiguration compounded by high load and a scheduler bug, but not so
much interested in how it got this way as to how to clean it up.

for example on node nova-1 the are actually 7  single CPU instances running
as shown by virsh on the system and in the instances table and 24 available
vCPUs. However it reports -102 VCPUs available:

root@nova-1:~# grep AUDIT /var/log/nova/nova-compute.log|tail -4
2012-11-06 15:49:20 AUDIT nova.compute.resource_tracker [-] Free VCPUS: -102
2012-11-06 15:50:50 AUDIT nova.compute.resource_tracker [-] Free ram (MB):
-205145
2012-11-06 15:50:50 AUDIT nova.compute.resource_tracker [-] Free disk (GB):
-2175
2012-11-06 15:50:50 AUDIT nova.compute.resource_tracker [-] Free VCPUS: -102

nova-manage service describe_resource nova-1
HOST  PROJECT cpu mem(mb) hdd
nova-1  (total)24   48295 605
nova-1  (used_now)126  2534402780
nova-1  (used_max)  7   14336 210
nova-1   3008a142e9524f7295b06ea811908f93   7
14336 210

Both of these match what I see in the compute_nodes table which also tells
tells me there are 88 running_vms (not 7).

Where is the resource_tracker on nova-1 getting this information and how do
I correct it?  (It's clearly pushing this bad info back to the database as
I  first tried to correct it there)

-Jon
___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Scheduler issues in folsom

2012-11-01 Thread Jonathan Proulx
On Wed, Oct 31, 2012 at 10:54 PM, Vishvananda Ishaya
vishvana...@gmail.comwrote:


 My patch here seems to fix the issue in the one scheduler case:


 https://github.com/vishvananda/nova/commit/2eaf796e60bd35319fe6add6dd04359546a21682

 If you could give that a try on your scheduler node and see if it fixes it
 that would be awesome. Also, it would be fery helpful if you can report a
 bug for me to reference in my merge proposal. I will see what I can do to
 write a few tests and have a potential fix for multiple schedulers.


bug is here https://bugs.launchpad.net/nova/+bug/1073956 since you've
reproduced it

If give it a try ==  drop that host_manager.py in place of my (folsom)
file and restart the scheduler I'm still getting the same results when
using a 100 iteration for loop around nova boot  --availability-zone az:host
all 100 end up on nova-1 but I'm suspicious this may avoid the scheduler
entirely not sure how that  availability zone trick for specifying a target
hso tis implemented. My user's case that I'm trying to make work uses
'euca-run-instances -n 500' using that with a value of 200 (and your
host_manager.py) the scheduler immediately puts them all in error state and
doesn't say any more about it, previously it was scheduling them poorly.
It's quite possible I've knocked something loose while banging around (also
possible I need to pull your whole branch, didn''t look as closely as I
should have at what it was based on) so I'm going to recheck my services
and do some more tests but that's what I see at first.

Thanks!
-Jon
___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Scheduler issues in folsom

2012-10-31 Thread Jonathan Proulx
Hi All

While the RetryScheduler may not have been designed specifically to
fix this issue https://bugs.launchpad.net/nova/+bug/1011852 suggests
that it is meant to fix it, well if it is a scheduler race condition
which is my suspicion.

This is my current scheduler config which gives the failure mode I describe:

scheduler_available_filters=nova.scheduler.filters.standard_filters
scheduler_default_filters=AvailabilityZoneFilter,RamFilter,CoreFilter,ComputeFilte
r,RetryFilter
scheduler_max_attempts=30
least_cost_functions=nova.scheduler.least_cost.compute_fill_first_cost_fn
compute_fill_first_cost_fn_weight=1.0
cpu_allocation_ratio=1.0
ram_allocation_ratio=1.0

I'm running the scheduler and api server on a single controller host
and it's pretty consistent about scheduling hundred  instances per
node at first then iteratively rescheduling them elsewhere when
presented with either an single API request to start many instances
(using euca2ools) or a shell loop around nova boot to generate one api
request per server.

the cpu_allocation ratio should limit the scheduler to 24 instances
per compute node regardless how how it's calculating memory, so while
I talked a lot about memory allocation as a motivation it is more
frequent for cpu to actually be the limiting factor in my deployment
and it certainly should.

And yet after attempting to launch 200 m1.tiny instances:

root@nimbus-0:~# nova-manage service describe_resource nova-23
2012-10-31 11:17:56
HOST  PROJECT cpu mem(mb) hdd
nova-23 (total)24   48295 882
nova-23 (used_now)107   56832  30
nova-23 (used_max)107   56320  30
nova-23  98333a1a28e746fa8c629c83a818ad57 106
54272   0
nova-23  3008a142e9524f7295b06ea811908f93   1
2048  30

eventually those bleed off to other systems though not entirely

2012-10-31 11:29:41
HOST  PROJECT cpu mem(mb) hdd
nova-23 (total)24   48295 882
nova-23 (used_now) 43   24064  30
nova-23 (used_max) 43   23552  30
nova-23  98333a1a28e746fa8c629c83a818ad57  42
21504   0
nova-23  3008a142e9524f7295b06ea811908f93   1
2048  30

at this point 12min later out of 200 instances 168 are active 22 are
errored and 10 are still building.  Notably only 23 actual VMs are
running on nova-23:

root@nova-23:~# virsh list|grep instance |wc -l
23

So that's what I see perhaps my assumptions about why I'm seeing it
are incorrect

Thanks,
-Jon

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Scheduler issues in folsom

2012-10-31 Thread Jonathan Proulx
On Wed, Oct 31, 2012 at 1:47 PM, Huang Zhiteng winsto...@gmail.com wrote:
 Hi Jonathan,

 If I understand correctly, that bug is about multiple scheduler

There is only a single process, I was reading it as relating to
include threads within a single process, but they should clearly be
able to serialize this withing a single process, so hoping my reading
is not correct.

What openstack package are you using?

2012.2-0ubuntu5~cloud0  on Ubuntu 12.04 from Ubuntu cloud-archive.

With further testing it appears the allocation ratios are simply being
ignored as the ChanceScheduler also violates my resource allocation
limits I just hadn't been testing at large enough scale to reliably
see it.

I am seeing memory allocations 1.5 default ram_allocation_ratio which
is somewhat curious.  So it's almost as if the RamFilter isn't
filtering rather than falling back to the default allocation ration
(the compute nodes would melt before I got to the 16:1 default cpu
allocation ration so not surprised I'm not getting near that more like
2:1)

again despite:
scheduler_available_filters=nova.scheduler.filters.standard_filters
scheduler_default_filters=AvailabilityZoneFilter,RamFilter,CoreFilter,ComputeFilter,RetryFilter
cpu_allocation_ratio=1.0
ram_allocation_ratio=1.0

Thanks,
-Jon

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


[Openstack] Scheduler issues in folsom

2012-10-30 Thread Jonathan Proulx
Hi All,

I'm having what I consider serious issues with teh scheduler in
Folsom.  It seems to relate to the introdution of threading in the
scheduler.

For a number of local reason we prefer to have instances start on the
compute node with the least amount of free RAM that is still enough to
satisfy the request which is the reverse of the default policy of
scheduling on the system with the most free RAM.  I'm fairly certain
the smae behavior would be seen with that policy as well, and any
other policy that results in a best choice for scheduling the next
instance. 

We have work loads that start hundreds of instances or the same image
and there are plans on scaling this to thousands.  What I'm seeing is
somehting like this:

* user submits API request for 300 instances
* scheduler puts them all on one node
* retry schedule kicks in at some point for the 276 that don't fit
* those 276 are all scheduled on the next best node
* retry cycle repeats with the 252 that don't fit there

I'm not clear exactly where the RetryScheduler in serts itself (I
should probably read it) but the first compute node is very overloaded
handling start up request which results in a fair number of instances
entering ERROR state rather than rescheduling (so not all 276
actually make it to the next round) and the whole process it painfully
slow.  In the end we are lucky to see 50% of the requested instances
actually make it into Active state (and then only becasue we increased
scheduler_max_attempts).

Is that really how it's supposed to work?  With the introduction of
the RetryScheduler as a fix for the scheduling race condition I think
it is, but it is a pretty bad solution for me, unless I'm missing
something, am I?  wouln't be the first time... 

For now I'm working around this by using the ChanceScheduler
(compute_scheduler_driver=nova.scheduler.chance.ChanceScheduler) so
the scheduler threads don't pick a best node.  This is orders of
magnitude faster and consistantly successful in my tests.  It is not
ideal for us as we have a small minority of ciompute nodes with twice
the memory capacity of our standard nodes and would prefer to keep
those available for some of our extra large memory flavors and we'd
also liek to minimize memory fragmentation on the standard sized nodes
for similar reasons.

-Jon

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


[Openstack] Special Zero Size root in Flavor, what exactly does it mean?

2012-10-27 Thread Jonathan Proulx
Hi All,

I know that specifying a zero size root volume in a flavor is
special but what exactly does it mean.  I'm writing up some
documentation on flavors and realized I can't fully explain the
default m1.tiny ...

Thanks,
-Jon

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Using nova-volumes openstack LVM group for other pourposes

2012-10-25 Thread Jonathan Proulx
On Thu, Oct 25, 2012 at 12:48:23PM +0200, Daniel Vázquez wrote:
:If we create a volume out of openstack context (it's is not created on
:nova database) only on system level. Can we after add to openstack?
:reformulate ask, Can we add to openstack an existen volumen (it's
:living on nova-volumes group)? I'm not sure on nova-volume create
:command or horizon ... that behavior it will be with existent lv label
:or some other properties.

It would be possible, it would require manually inserting data in many
places though...this definately isn't a supported option but I do
belive it is technically possible

you might need to follow the naming convention (probably should) of
volume-uuid, this could be doen with lvrename

you will need to make an entry in the volumes table of your nova
database

you will also need an entry in /var/lib/nova/volumes/volume-uuid so
that tgtd knows how to map the iSCSI target back to the correct
device.

that's what comes to mind, it's untested and there may be other
pieces, but there isn't anything special in the logical volume data
or metadata.

If you're trying to plot a migration path for existing VMs it might be
better to use nova to create a new volume and 'dd' to dump the
contents of the old volume into it just to be sure everthing lines up
correctly.

-Jon

:
:2012/10/24 Daniel Vázquez daniel2d2...@gmail.com:
: Yeah!! Jon, I'm agree with you about organization/separation LVM
: groups, this is for very very very special situation.
: Any case if I use nova pattern labeling via logical volumen creation
: or via renaming label, I hope can switch the content of this custom
: logical volument to use with openstack, an attach to a VM in future.
:
:
:
: 2012/10/24 Jonathan Proulx j...@csail.mit.edu:
: On Wed, Oct 24, 2012 at 08:56:26PM +0200, Daniel Vázquez wrote:
: :Hi here!
: :
: :Can we create and use news logical volumes for own/custom use(out of
: :openstack) on nova-volumes openstack LVM group, and use it beside
: :openstack operational?
: :IMO it's LVM and no problem, but it has openstack collateral consequences?
:
: If you are talking about creating random logical volumes for
: non-openstack use in the same volume group nova-volume or cinder is
: using to create volumes (lv are in the same vg but don't otherwise
: interact), yes you can do that without ocnfusing openstack or having
: your volumes trampled.  For example only having one volume group and
: using that for operating system partitions as well as volume-foo
: volumes for cinder
:
: I don't think it's a particularly good idea from an organizational
: standpoint I'd rather have distinct vg's for each purpose so it is
: clear which resources are operating system and which are data, but in
: my environment (a private computing/research cloud with a small admin
: group and 1k users in a few 10's of closely related tenents) it's
: probably more an aesthetic than technical choice.  The larger and more
: diverse your situation the stronger I'd argue for keeping them in
: seperate VGs.
:
: -Jon
:
:___
:Mailing list: https://launchpad.net/~openstack
:Post to : openstack@lists.launchpad.net
:Unsubscribe : https://launchpad.net/~openstack
:More help   : https://help.launchpad.net/ListHelp

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Possible upgrade bug in nova-volume ( cinder)?

2012-10-25 Thread Jonathan Proulx
On Wed, Oct 24, 2012 at 6:32 PM, John Griffith
john.griff...@solidfire.com wrote:

 Hey Jon,

 Cool... Yeah, I had intended for that patch to be a stable/folsom patch but
 shouldn't have submitted it to master :(  The real problem isn't just
 normalizing the lvm names, but also the provider_location information that
 is stored in the DB for when you try to attach to your compute node.

Yup, I had understood that.  We only have about 10 active volumes so
it was easy enough in my case to copy and paste the ids/uuids around
for mysql updates and lvrenames.  So I can now attach both old and new
volumes at my site.

 The cinder version is the same (https://review.openstack.org/#/c/14790/) and
 depending on the feedback it's a candidate for back-port.

 The other possible issues that I've seen people run in to:
 1. volumes directory not being specified correctly
 2. not restarting tgtadm
 3. not having the include statement in /etc/tgt/conf.d

 I think you and I covered most of these in our chat on IRC earlier today...

I tripped over a couple more steps that I didn't know about but
managed to discover on my own as well, like 'cinder-manage db sync' to
set up the tables in the cinder database, which I'm sure are so
obvious as to be nearly invisible to people working on the code and
enough like any other openstack service it wasn't too hard to guess.
I collected up my notes and expanded migration section of the release
notes by a few steps
http://wiki.openstack.org/ReleaseNotes/Folsom#OpenStack_Block_Storage_.28Cinder.29

Thanks again,
-Jon

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


[Openstack] Possible upgrade bug in nova-volume ( cinder)?

2012-10-24 Thread Jonathan Proulx
Hi All,

I'm seeing a bug due to my recent essex to folsom upgrade relating to
LVM back volume storage, I'm not sure where it got introduced most
likely either in nova-volume or in the Ubuntu cloud archive
packaging...I only noticed it after transitioning from
folsom-nova-volume to fosom-cinder but despite thinking I'd tested the
nova-volume service before moving to cinder I'm pretty sure it had to
exist in nova-volume as well (prehaps it was masked becasue I didn't
restart tgtd until cinder)

The symptom is that volumes created under folsom (with nova-volume or
cinder) can be attached.

The reason is that the backing-store devices in both
/var/lib/nova/volumes/* and /var/lib/cinder/volumes/* are all named
/dev/vg/volume-uuid while under essex the volumes were named
/dev/vg/volume-id

To fix this for me I can look up the volumes by ID in the database and
then lvrename the logical volumes (I don't have too many and all on
one volume server right now).

Before I go sifting through postinst scripts and openstack code to see
whre this came from anyone know where I should file this (and has
anyone else run into it)?

-Jon

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Possible upgrade bug in nova-volume ( cinder)?

2012-10-24 Thread Jonathan Proulx
On Wed, Oct 24, 2012 at 2:45 PM, Jonathan Proulx j...@jonproulx.com wrote:

 To fix this for me I can look up the volumes by ID in the database and
 then lvrename the logical volumes (I don't have too many and all on
 one volume server right now).

That maybe the wrong answer as the database (both cinder and the older
nova leavings) has a provider_location that implies the right
logical volume name:

+--++---+
| id   | ec2_id | provider_location

  |
+--++---+
| 25cb6abc-1938-41da-b4a4-7639fa122117 | NULL   | 128.52.x.x:3260,9
iqn.2010-10.org.openstack:volume-001c 1
  |
| 60cd2c0e-6d61-4010-aee2-df738adb3581 | NULL   | 128.52.x.x:3260,4
iqn.2010-10.org.openstack:volume-001a 1
  |
| 67ba5863-9f92-4694-b639-6c9520e0c6f3 | NULL   | 128.52.x.x:3260,2
iqn.2010-10.org.openstack:volume-0016 1
  |
| 7397daa1-f4a7-47d4-b0dc-0b306defdf62 | NULL   | 128.52.x.x:3260,14
iqn.2010-10.org.openstack:volume-0014 1
 |
| 7d8c51bc-9cac-4edf-b1e6-1c37d5a8256f | NULL   | 128.52.x.x:3260,10
iqn.2010-10.org.openstack:volume-7d8c51bc-9cac-4edf-b1e6-1c37d5a8256f
1 |
| 86426e77-e396-489d-9e66-49f0beef46bb | NULL   | 128.52.x.x:3260,16
iqn.2010-10.org.openstack:volume-0019 1
 |
| 98ac28f5-77d8-476b-b3e1-c90a0fd3e880 | NULL   | 128.52.x.x:3260,1
iqn.2010-10.org.openstack:volume-0010 1
  |
| a6e68eae-23a9-483e-bd42-e4b8a7f47dc4 | NULL   | 128.52.x.x:3260,24
iqn.2010-10.org.openstack:volume-a6e68eae-23a9-483e-bd42-e4b8a7f47dc4
1 |
| a89b9891-571c-43be-bc1b-0c346a161d38 | NULL   | 128.52.x.x:3260,9
iqn.2010-10.org.openstack:volume-a89b9891-571c-43be-bc1b-0c346a161d38
1  |
| cbd32221-7794-41d1-abf2-623c49f4ff03 | NULL   | 128.52.x.x:3260,6
iqn.2010-10.org.openstack:volume-001b 1
  |
+--++---+

so I'm also open to suggestions on the right resolution to this.
Should I rename the logical volume sand update the provider_location
or should I make the  /var/lib/cinder/volumes/* files match what is in
the database and LVM (and if I do the latter will something come along
and undo that)?

-Jon

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Using nova-volumes openstack LVM group for other pourposes

2012-10-24 Thread Jonathan Proulx
On Wed, Oct 24, 2012 at 08:56:26PM +0200, Daniel Vázquez wrote:
:Hi here!
:
:Can we create and use news logical volumes for own/custom use(out of
:openstack) on nova-volumes openstack LVM group, and use it beside
:openstack operational?
:IMO it's LVM and no problem, but it has openstack collateral consequences?

If you are talking about creating random logical volumes for
non-openstack use in the same volume group nova-volume or cinder is
using to create volumes (lv are in the same vg but don't otherwise
interact), yes you can do that without ocnfusing openstack or having
your volumes trampled.  For example only having one volume group and
using that for operating system partitions as well as volume-foo
volumes for cinder

I don't think it's a particularly good idea from an organizational
standpoint I'd rather have distinct vg's for each purpose so it is
clear which resources are operating system and which are data, but in
my environment (a private computing/research cloud with a small admin
group and 1k users in a few 10's of closely related tenents) it's
probably more an aesthetic than technical choice.  The larger and more
diverse your situation the stronger I'd argue for keeping them in
seperate VGs.

-Jon

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Possible upgrade bug in nova-volume ( cinder)?

2012-10-24 Thread Jonathan Proulx
On Wed, Oct 24, 2012 at 3:01 PM, John Griffith
john.griff...@solidfire.com wrote:

 Hey Jon,

 Couple of things going on, one is the volume naming (in progress here:
 https://review.openstack.org/#/c/14615/).  I'll take a closer look at some
 of the other issues you pointed out.

Hi John,

On this issue I think the issue you link to above covers my problem
(even if that exact implementation now seems abandoned).  So the
right thing is to normalize all the naming on UUID in my case using
lvrename (or symlinks as the patchset above does) and updating the
cinder db provider_location.  What other issues did you see to look
into?  Seems this covers everything I brought up here and the issues
you helped me with on IRC earlier (thanks) seem like documentation
issues (which I hope to get into the docs once I clean up my notes)

-Jon

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


[Openstack] nova-volume - cinder troubles

2012-10-22 Thread Jonathan Proulx
Hi All,

I'm trying to get from nova-volume to cinder and seem to be tripping
up near the end.

I have ServerA running as the cloud controller (horizon, keystone, glance,
rabbitmq, mysql, nova-api, etc...), ServerB was running nova-volume
and is now running cinder.  I got the DB created (on ServerA) and
pupulated with cinder-manage migrate bits in the release notes.

I disabled osapi_volume on ServerB, restarted nova-api and then
cinder-api as directed.

But api request still seem to be trying (and failing) to use the
nova-volume service, as seen by updates to the nova.volumes table when
trying to create or attach volumes through Horizon or booting from
volume with the python-novaclient.

What piece of plubing am I missing?  Seem like the controller should be
running cinder-api too (but not cinder volume)?

-Jon




___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] nova-volume - cinder troubles

2012-10-22 Thread Jonathan Proulx
On Mon, Oct 22, 2012 at 03:06:15PM -0400, Jonathan Proulx wrote:
:Hi All,
:
:I'm trying to get from nova-volume to cinder and seem to be tripping
:up near the end.
:
:I have ServerA running as the cloud controller (horizon, keystone, glance,
:rabbitmq, mysql, nova-api, etc...), ServerB was running nova-volume

snip

:What piece of plubing am I missing?  Seem like the controller should be
:running cinder-api too (but not cinder volume)?

not out of the woods yet, but this problem is that I needed to start
the cinder-scheduler process on the controller node ServerA above.

-Jon

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Folsom nova-scheduler race condition?

2012-10-10 Thread Jonathan Proulx
On Wed, Oct 10, 2012 at 4:33 AM, Huang Zhiteng winsto...@gmail.com wrote:

 Yes, but it seems Jonathan's filter list doesn't include RetryFilter,
 so it's possible that he ran into a race condition that RetryFilter
 targeted to solve.

Yes, that was it exactly.  Thank you for seeing the obvious think I
was missing...

-Jon

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


[Openstack] Folsom nova-scheduler race condition?

2012-10-09 Thread Jonathan Proulx
Hi All,

Looking for a sanity test before I file a bug.  I very recently
upgraded my install to Folsom (on top of Ubuntu 12.04/kvm).  My
scheduler settings in nova.conf are:

scheduler_available_filters=nova.scheduler.filters.standard_filters
scheduler_default_filters=AvailabilityZoneFilter,RamFilter,CoreFilter,ComputeFilter
least_cost_functions=nova.scheduler.least_cost.compute_fill_first_cost_fn
compute_fill_first_cost_fn_weight=1.0
cpu_allocation_ratio=1.0

This had been working to fill systems based on available RAM and to
not exceed 1:1 allocation ration of CPU resources with Essex.  With
Folsom, if I specify a moderately large number of instances to boot or
spin up single instances in a tight shell loop they will all get
schedule on the same compute node well in excess of the number of
available vCPUs . If I start them one at a time (using --poll in a
shell loop so each instance is started before the next launches) then
I get the expected allocation behaviour.

I see https://bugs.launchpad.net/nova/+bug/1011852 which seems to
attempt to address this issue but as I read it that fix is based on
retrying failures.  Since KVM is capable of over committing both CPU
and Memory I don't seem to get retryable failure, just really bad
performance.

Am I missing something this this fix or perhaps there's a reported bug
I didn't find in my search, or is this really a bug no one has
reported?

Thanks,
-Jon

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Using EC2 through OpenStack

2012-10-02 Thread Jonathan Proulx
On Tue, Oct 02, 2012 at 12:19:45PM +0200, Michaël Van de Borne wrote:
:Hi all,
:
:This might be obvious, but I can't find the answer. Is there a way to
:control EC2 instances using OpenStack?

OpenStack provides the same facility as EC2 but on your own hardware,
so they don't really touch.  Glance (the image store componet of
OpenStack) can use S3 for storage, but I don't think it's a good idea
unless you have a a really fast pipe to Amazon and deep pockets to pay
transfer fees.  Perhaps others have different opinions on that point
(caveat: I haven't actually tried impeneting this that way)

You can use some of the same tools to manage instances on both since
OpenStack provides an EC2 compatible API (I've used euca2ools and  hybridfox
with both).


-Jon


___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] issues with fixed_range blocking new instance creation?

2012-09-07 Thread Jonathan Proulx
On Thu, Sep 6, 2012 at 9:59 PM, Vishvananda Ishaya
vishvana...@gmail.com wrote:
 fixed range is likely fine. I suspect you created your network with a 
 10.0.0.0/24 though. It might be a bit tricky to switch to a larger range now. 
 You may have to create the rest of your fixed ips manually in the db 
 associated with the same network and then change the cidr entry on the 
 network in the db to 10.0.0.0/16. You also might have to manually destroy the 
 bridges and let nova-network recreate them.

That does seem to be the case, if you'd asked me 24hr ago I'd have
sworn I never put /24 anywhere but that's the only explaination that
fits observed reality.  Luckily I haven't quite called the system
production yet so I'm taking a maintenance window now, terminating all
instances and deleting and recreating network

Being able to expand networks more easily would be a nice thing (and
in a dhcp controlled world it's not too disruptive as long as you
don't go pushing the gateway address around), but best I can tell this
is what I've got right now.

Thanks,
-Jon

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


[Openstack] After expanding fixed ip range instances no longer have network

2012-09-07 Thread Jonathan Proulx
Hi All,

Running Essex on Ununtu 12.04 using multi-host  FlatDHCP nova-networking

I ran out of IPs on my fixed_ip range so I shut everything (instances,
nova-network, nova-compute) down deleted the old network and recreated
it with a smaller netmask.  This seems to have almost worked.

I can start more instances than I previously had fixed ip's, the right
ones seem to be being assigned and the mask on the recreated bridge
interfaces is correct, but the (ubuntu-cloudimage) instances can't
seem to see their NIC's any more, or perhaps aren't getting dhcp
properly I'm still trying to force my way in as our instances rather
rely on net access for accounts.

On the compute node things seem OK the bridge is up and the right
things are connected:

root@nova-5:/var/log/nova# brctl show br100
bridge name bridge id   STP enabled interfaces
br100   8000.60eb69d22521   no  eth1
vnet0
vnet1
vnet2


I do notice this in iptables:

Chain nova-network-POSTROUTING (1 references)
target prot opt source   destination
ACCEPT all  --  10.0.0.0/16  nova-5.csail.mit.edu
ACCEPT all  --  10.0.0.0/16  10.128.0.0/24
ACCEPT all  --  10.0.0.0/16  10.0.0.0/16  ! ctstate DNAT


my fixed range is 10.0.0.0/16 not sure where 10.128.0.0/24 comes into
it as I don't use that network, but can't see that as a problem.

Can any one tell me what I've looked?

-Jon

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] After expanding fixed ip range instances no longer have network

2012-09-07 Thread Jonathan Proulx
On Fri, Sep 7, 2012 at 11:33 AM, Jonathan Proulx j...@jonproulx.com wrote:

 Can any one tell me what I've looked?

I assumed stopping and restarting nova-network would restart dnsmasq
since dnsmasq doesn't have it's own init script, but this seems not to
be the case.

dnsmasq is listening on an IP the system nolonger has, I'm sure I'll
find the answer ot this on my own soon enough but how does one
properly restart openstack's dnsmasq (on ubuntu)?  It clearly gets
started with lots of command line args from nova.conf

-Jon

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] After expanding fixed ip range instances no longer have network

2012-09-07 Thread Jonathan Proulx
On Fri, Sep 07, 2012 at 11:59:20AM -0400, Jonathan Proulx wrote:
:On Fri, Sep 7, 2012 at 11:33 AM, Jonathan Proulx j...@jonproulx.com wrote:

:dnsmasq is listening on an IP the system nolonger has, I'm sure I'll
:find the answer ot this on my own soon enough but how does one
:properly restart openstack's dnsmasq (on ubuntu)?  It clearly gets
:started with lots of command line args from nova.conf

For teh record: One way that works is to kill all dnsmasq processes then restart
nova-network. Hopefully there's a cleaner way, but I no longer *need*
to find it.

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] After expanding fixed ip range instances no longer have network

2012-09-07 Thread Jonathan Proulx
Thanks Vish,

Your directions were correct  would have saved me a few hours of
floundering had I just started a little later :)

particularly the `killall dnsmasq`

-Jon

On Fri, Sep 07, 2012 at 09:12:52AM -0700, Vishvananda Ishaya wrote:
:If you are sure this is the issue:
:
:killall dnsmasq
:restart nova-network
:
:Vish
:
:On Sep 7, 2012, at 8:59 AM, Jonathan Proulx j...@jonproulx.com wrote:
:
: On Fri, Sep 7, 2012 at 11:33 AM, Jonathan Proulx j...@jonproulx.com wrote:
: 
: Can any one tell me what I've looked?
: 
: I assumed stopping and restarting nova-network would restart dnsmasq
: since dnsmasq doesn't have it's own init script, but this seems not to
: be the case.
: 
: dnsmasq is listening on an IP the system nolonger has, I'm sure I'll
: find the answer ot this on my own soon enough but how does one
: properly restart openstack's dnsmasq (on ubuntu)?  It clearly gets
: started with lots of command line args from nova.conf
: 
: -Jon
: 
: ___
: Mailing list: https://launchpad.net/~openstack
: Post to : openstack@lists.launchpad.net
: Unsubscribe : https://launchpad.net/~openstack
: More help   : https://help.launchpad.net/ListHelp
:

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


[Openstack] issues with fixed_range blocking new instance creation?

2012-09-06 Thread Jonathan Proulx
Hi All,

my stack stopped launching instances last night  I've been chasing my
tail for a while but I think it think's it's out of fixed IPs.

the system is Essex as packages on Ubuntu 12.04 using FlatDHCP, with
multihost and a fixed range or 10.0.0.0/16

--multi_host=True
--public_interface=eth0
--fixed_range=10.0.0.0/16
--flat_interface=eth1
--flat_network_bridge=br100
--network_manager=nova.network.manager.FlatDHCPManager

the error I'm seeing for all attempts at launching is:

2012-09-06 10:16:43 TRACE nova.compute.manager [instance:
d5aa2d4b-76db-43b7-a655-24edeb7fc495]   File
/usr/lib/python2.7/dist-packages/nova/virt/libvirt/vif.py, line 69,
in _get_configurations
2012-09-06 10:16:43 TRACE nova.compute.manager [instance:
d5aa2d4b-76db-43b7-a655-24edeb7fc495] 'ip_address':
mapping['ips'][0]['ip'],
2012-09-06 10:16:43 TRACE nova.compute.manager [instance:
d5aa2d4b-76db-43b7-a655-24edeb7fc495] IndexError: list index out of
range

and I just noticed that the flat network interfaces are getting /24
netmasks NOT  /16 this combined with 210 running instances and forty
some compute nodes make me think it's out of addresses.

is my syntax for the fixed range wrong, doc give 10.0.0.0/8 as the
example and default so I don't think so.  I'm fine with reverting to
that, don't mind more addresses, but I would like to understand if
this is actually my problem and even if it's not why it's not giving
the mask I asked for.

-Jon

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Instance stuck in deleting state with error

2012-08-02 Thread Jonathan Proulx
On Wed, Aug 1, 2012 at 3:12 PM, Lorin Hochstein
lo...@nimbisservices.com wrote:

 I believe pip gets it from PyPI:
 http://pypi.python.org/pypi/python-novaclient/

Ah, I documented this internally and promptly forgot, this is where my
version of python-novaclient with reset-state came from:
sudo pip install -e
git+https://github.com/openstack/python-novaclient.git#egg=python-novaclient

you may want to verify the version at
http://pypi.python.org/pypi/python-novaclient has it as well.

-Jon

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Instance stuck in deleting state with error

2012-08-01 Thread Jonathan Proulx
FYI

the nova reset-state command exists on my Debian 6.0 workstations
where python-novaclient is installed via pip (presumably the gets
latest version from trunk?) it is not available on my Ubuntu 12.04
systems using standard Ubuntu repository versions (essex).

-Jon


On Wed, Aug 1, 2012 at 2:22 AM, Wolfgang Hennerbichler
wolfgang.hennerbich...@risc-software.at wrote:


 On 07/31/2012 04:17 PM, Rohit Karajgi wrote:

 Hi Wolfgang,

 Have you updated the your python-novaclient?


 no I didn't.


 The 'nova reset-state

 server --active' command is pretty much there. It is an admin
 action in Nova extensions. $ nova help | grep reset reset-state
 Reset the state of an instance


 thanks for the hint.
 Wolfgang



 Regards, Rohit Karajgi | Lead Engineer | NTT Data Global Technology
 Services Private Ltd | w. +91.20.6604.1500 x 378 |  m. +91
 992.242.9639 | rohit.kara...@nttdata.com



 -Original Message- From:
 openstack-bounces+rohit.karajgi=nttdata@lists.launchpad.net
 [mailto:openstack-bounces+rohit.karajgi=nttdata@lists.launchpad.net]
 On Behalf Of Wolfgang Hennerbichler Sent: Tuesday, July 31, 2012
 10:45 AM To: openstack@lists.launchpad.net Subject: Re: [Openstack]
 Instance stuck in deleting state with error

 On 07/30/2012 09:35 PM, Kevin L. Mitchell wrote:

 That said, be aware that there is a reset-state command to
 novaclient, so that you can do Chris's recommended reset without
 having to muck around with the database directly.


 where? nova help | grep reset yields nothing. I think this is one of
 openstack worst weaknesses, that if the status of an instance is in
 error-state and one has to wade through a couple of logfiles
 (scheduler, nova-network, nova-compute) in order to find out what
 really happened. I would be superior if the error itself would be
 reported back to the database.

 Wolfgang

 ___ Mailing list:
 https://launchpad.net/~openstack Post to :
 openstack@lists.launchpad.net Unsubscribe :
 https://launchpad.net/~openstack More help   :
 https://help.launchpad.net/ListHelp

 __


 Disclaimer:This email and any attachments are sent in strictest confidence
 for the sole use of the addressee and may contain legally privileged,
 confidential, and proprietary data.  If you are not the intended recipient,
 please advise the sender by replying promptly to this email and then delete
 and destroy this email and any attachments without any further use, copying
 or forwarding




 --
 DI (FH) Wolfgang Hennerbichler
 Software Development
 Unit Advanced Computing Technologies
 RISC Software GmbH
 A company of the Johannes Kepler University Linz

 IT-Center
 Softwarepark 35
 4232 Hagenberg
 Austria

 Phone: +43 7236 3343 245
 Fax: +43 7236 3343 250
 wolfgang.hennerbich...@risc-software.at
 http://www.risc-software.at


 ___
 Mailing list: https://launchpad.net/~openstack
 Post to : openstack@lists.launchpad.net
 Unsubscribe : https://launchpad.net/~openstack
 More help   : https://help.launchpad.net/ListHelp

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Instance stuck in deleting state with error

2012-08-01 Thread Jonathan Proulx
On Wed, Aug 1, 2012 at 3:40 PM, Lorin Hochstein
lo...@nimbisservices.com wrote:

 What's the use case for resetting an instance to the error state? Is the
 idea to do:

 nova reset-state instance
 nova delete instance

That was my use case, though I ended up doing both in the database
because I was impatient...

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


[Openstack] Instance stuck in deleting state with error

2012-07-30 Thread Jonathan Proulx
Hi All,

I have an instance that has been in this state for a couple days:

| OS-EXT-STS:power_state  | 0|
| OS-EXT-STS:task_state   | deleting|
| OS-EXT-STS:vm_state | error |

with this fault:

| fault | {u'message': u'ProcessExecutionError', u'code': 500,
u'created': u'2012-07-30T16:51:34Z'} |

sending more delete requests seem to have no effect and the node it
thinks it is on doesn't seem to have any state about it.

How to I clean this up, do I just need to muck out the DB and if so
what parts of it?

Thanks,
-Jon

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Instance stuck in deleting state with error

2012-07-30 Thread Jonathan Proulx
Well for posterity or so people can tell me what I did wrong here's
how I seemed to convince the database that  the instance was really
deleted:

UPDATE instances SET
vm_state='deleted',task_state='deleted',deleted=1,deleted_at=now()
WHERE uuid=FOO

And here's what I had to do to the volume:

 UPDATE volumes SET
attach_status='detached',instance_id=NULL,mountpoint=NULL,status='available'
WHERE id=FOO

Any insight into how I might determine what when wrong here (essex on
ubuntu 12.04 rabbitmq mysql) or if I missed any pieces that will bite
me are appreciated.

-Jon

On Mon, Jul 30, 2012 at 1:03 PM, Jonathan Proulx j...@jonproulx.com wrote:
 Hi All,

 I have an instance that has been in this state for a couple days:

 | OS-EXT-STS:power_state  | 0|
 | OS-EXT-STS:task_state   | deleting|
 | OS-EXT-STS:vm_state | error 
 |

 with this fault:

 | fault | {u'message': u'ProcessExecutionError', u'code': 500,
 u'created': u'2012-07-30T16:51:34Z'} |

 sending more delete requests seem to have no effect and the node it
 thinks it is on doesn't seem to have any state about it.

 How to I clean this up, do I just need to muck out the DB and if so
 what parts of it?

 Thanks,
 -Jon

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Instance stuck in deleting state with error

2012-07-30 Thread Jonathan Proulx
Thanks but I'm using KVM so not that bug...figures that's the part of
my stack I left out.

-Jon

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


[Openstack] Ceph performance as volume image store?

2012-07-23 Thread Jonathan Proulx
Hi All,

I've been looking at Ceph as a storage back end.  I'm running a
research cluster and while people need to use it and want it 24x7 I
don't need as many nines as a commercial customer facing service does
so I think I'm OK with the current maturity level as far as that goes,
but I have less of a sense of how far along performance is.

My OpenStack deployment is 768 cores across 64 physical hosts which
I'd like to double in the next 12 months.  What it's used for is
widely varying and hard to classify some uses are hundreds of tiny
nodes others are looking to monopolize the biggest physical system
they can get.  I think most really heavy IO currently goes to our NAS
servers rather than through nova-volumes but that could change.

Anyone using ceph at that scale (or preferably larger)?  Does it keep
up if you keep throwing hardware at it?  My proof of concept ceph
cluster on crappy salvaged hardware has proved the concept to me but
has (unsurprisingly) crappy salvaged performance. Trying to get a
sense of what performance expectations I should have given decent
hardware before I decide if I should buy decent hardware for it...

Thanks,
-Jon

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


[Openstack] [nova] Specifying Schedule Filters Per Request?

2012-07-21 Thread Jonathan Proulx
Hi All,

http://docs.openstack.org/essex/openstack-compute/admin/content/compute-options-reference.html
states that the scheduler_default_filters option in nova.conf controls
which filter class names to use for filtering hosts when not specified
in the request. How can I specify what filters to use in the request
(or can I)?

I've been experimenting with DifferentHostFilter and SameHostFilter
both of which seem very expensive if left in the default list (even if
not passing hints to engage either).  It would be nice if these we
only enabled for the few cases that want them but I don't see any
options to the python-novaclient to pass in a filter list.

Thanks,
-Jon

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


Re: [Openstack] Networking issues in Essex

2012-07-12 Thread Jonathan Proulx

I've only deployed openstack for the first time a couple weeks ago,
but FWIW...

I had similar symptoms on my Essex test deployment (on Ubuntu 12.04)
turned out my problem was taht while the br100 bridge was up and
configured the underlying eth1 physical interface was down so the bits
went nowhere.  'ifconfig eth1 up' fixed all, followed ofcoures by
fixing in /etc/network/interfaces as well so this happens on it's own
in future.

-Jon

On Thu, Jul 12, 2012 at 02:56:57PM +1000, Michael Chapman wrote:
:Hi all, I'm hoping I could get some assistance figuring out my networking
:problems with a small Essex test cluster. I have a small Diablo cluster
:running without any problems but have hit a wall in deploying Essex.
:
:I can launch VMs without issue and access them from the compute host, but
:from there I can't access anything except the host, DNS services, and other
:VMs.
:
:I have separate machines running keystone, glance, postgresql, rabbit-mq
:and nova-api. They're all on the .os domain with 172.22.1.X IPs
:
:I have one machine running nova-compute, nova-network and nova-api, with a
:public address 192.43.239.175 and also an IP on the 172.22.1.X subnet in
:the .os domain. It has the following nova/conf:
:
:--dhcpbridge_flagfile=/etc/nova/nova.conf
:--dhcpbridge=/usr/bin/nova-dhcpbridge
:--logdir=/var/log/nova
:--state_path=/var/lib/nova
:--lock_path=/var/lock/nova
:--force_dhcp_release
:--iscsi_helper=tgtadm
:--libvirt_use_virtio_for_bridges
:--connection_type=libvirt
:--root_helper=sudo nova-rootwrap
:--verbose
:--ec2_private_dns_show_ip
:
:--network_manager=nova.network.manager.FlatDHCPManager
:--rabbit_host=os-amqp.os
:--sql_connection=postgresql://[user]:[password]@os-sql.os/nova
:--image_service=nova.image.glance.GlanceImageService
:--glance_api_servers=os-glance.os:9292
:--auth_strategy=keystone
:--scheduler_driver=nova.scheduler.simple.SimpleScheduler
:--keystone_ec2_url=http://os-key.os:5000/v2.0/ec2tokens
:
:--api_paste_config=/etc/nova/api-paste.ini
:
:--my_ip=192.43.239.175
:--flat_interface=eth0
:--public_interface=eth1
:--multi_host=True
:--routing_source_ip=192.43.239.175
:--network_host=192.43.239.175
:
:--dmz_cidr=$my_ip
:
:--ec2_host=192.43.239.175
:--ec2_dmz_host=192.43.239.175
:
:I believe I'm seeing a natting issue of some sort - my VMs cannot ping
:external IPs, though DNS seems to work.
:ubuntu@monday:~$ ping www.google.com
:PING www.l.google.com (74.125.237.148) 56(84) bytes of data.
:AWKWARD SILENCE
:
:When I do a tcpdump on the compute host things seem fairly normal, even
:though nothing is getting back to the VM
:
:root@ncios1:~# tcpdump icmp -i br100
:tcpdump: verbose output suppressed, use -v or -vv for full protocol decode
:listening on br100, link-type EN10MB (Ethernet), capture size 65535 bytes
:14:35:28.046416 IP 10.0.0.8  syd01s13-in-f20.1e100.net: ICMP echo request,
:id 5002, seq 9, length 64
:14:35:28.051477 IP syd01s13-in-f20.1e100.net  10.0.0.8: ICMP echo reply,
:id 5002, seq 9, length 64
:14:35:29.054505 IP 10.0.0.8  syd01s13-in-f20.1e100.net: ICMP echo request,
:id 5002, seq 10, length 64
:14:35:29.059556 IP syd01s13-in-f20.1e100.net  10.0.0.8: ICMP echo reply,
:id 5002, seq 10, length 64
:
:I've pored over the iptables nat rules and can't see anything amiss apart
:from the masquerades that are automatically added: (I've cut out some empty
:chains for brevity)
:
:root@ncios1:~# iptables -L -t nat -v
:Chain PREROUTING (policy ACCEPT 22 packets, 2153 bytes)
: pkts bytes target prot opt in out source
:destination
:   22  2153 nova-network-PREROUTING  all  --  anyany anywhere
:  anywhere
:   22  2153 nova-compute-PREROUTING  all  --  anyany anywhere
:  anywhere
:   22  2153 nova-api-PREROUTING  all  --  anyany anywhere
:  anywhere
:
:Chain INPUT (policy ACCEPT 12 packets, 1573 bytes)
: pkts bytes target prot opt in out source
:destination
:
:Chain OUTPUT (policy ACCEPT 31 packets, 2021 bytes)
: pkts bytes target prot opt in out source
:destination
:   31  2021 nova-network-OUTPUT  all  --  anyany anywhere
:  anywhere
:   31  2021 nova-compute-OUTPUT  all  --  anyany anywhere
:  anywhere
:   31  2021 nova-api-OUTPUT  all  --  anyany anywhere
:anywhere
:
:Chain POSTROUTING (policy ACCEPT 30 packets, 1961 bytes)
: pkts bytes target prot opt in out source
:destination
:   31  2021 nova-network-POSTROUTING  all  --  anyany anywhere
:anywhere
:   30  1961 nova-compute-POSTROUTING  all  --  anyany anywhere
:anywhere
:   30  1961 nova-api-POSTROUTING  all  --  anyany anywhere
:anywhere
:   30  1961 nova-postrouting-bottom  all  --  anyany anywhere
:  anywhere
:0 0 MASQUERADE  tcp  --  anyany 192.168.122.0/24!
:192.168.122.0/24 masq ports: 1024-65535
:0 0 MASQUERADE  udp  --  anyany 192.168.122.0/24!
:192.168.122.0/24 masq ports: 1024-65535
:0 0 MASQUERADE  all  -- 

Re: [Openstack] [Nova] resource free -vs- allocated utilization?

2012-07-12 Thread Jonathan Proulx
for posterity yes the info isn't hard to find in the database:

mysql select id,vcpus,vcpus_used,memory_mb,memory_mb_used  from compute_nodes;

I'm not terribly keen on SQL as an interface, guess if it bothers me
enough I'll implement a different interface...

On Wed, Jul 11, 2012 at 10:34 PM, Jonathan Proulx j...@jonproulx.com wrote:
 On Wed, Jul 11, 2012 at 8:58 PM, Chris Behrens cbehr...@codestud.com wrote:
 Hi Jon,

 There's actually a review up right now proposing to add an OS API extension
 to be able to give some of this data:

 https://review.openstack.org/#/c/9544/

 that seems to be based on quota limits where as I'm look for just now
 is over all physical limits for all users.

 I don't know how you may be looking to query it, but it's not too difficult
 to get it directly from the instances table in the database, either.

 I guess I'm looking for something like:

 eucalyptus-describe-availibility-zones verbose
 AVAILABILITYZONE |- vm types free / max cpu ram disk
 AVAILABILITYZONE |- m1.small  /  1 128 2
 AVAILABILITYZONE |- c1.medium  /  1 256 5
 AVAILABILITYZONE |- m1.large  /  2 512 10
 AVAILABILITYZONE |- m1.xlarge  /  2 1024 20
 AVAILABILITYZONE |- c1.xlarge  /  4 2048 20

 (well not with the zeros, but it's the first example I could find)

 where with euca2ools and openstack I get essentially the output of
 'nova-manage service list' which is useful but not for the same
 things.  Guess I'll dig into the database shouldn't be too hard to get
 close to what I want, the eucalyptus output also takes into account
 fragmentation, which is nice since if I have 100 free cpu slots but on
 100 different compute nodes I'm in a bit more trouble than if they are
 only spread across 10 since in the later case multicore instances can
 start (for a while anyway).

 Thanks,
 -Jon

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


[Openstack] [Nova] resource free -vs- allocated utilization?

2012-07-11 Thread Jonathan Proulx
is there an easy way to see free -vs- allocated nova resources?

for example that 300 of 1000 cpu cores are in use (and similar with
memory and storage).

Thanks,
-Jon

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


[Openstack] Managing projects roles and quotas (Dashboard -vs- nova-manage)

2012-07-05 Thread Jonathan Proulx
Hi All,

I have a new Essex install on Ubuntu 12.04 and just starting to kick
the tires. I would like to manage resource quotas for different
projects/tenants but having created them in Dashboard nova-manage
doesn't see them, though keystone does and I can join users to tenants
with roles using keystone user-role-add.

The documentation  here
http://docs.openstack.org/developer/nova/runnova/managing.projects.html#setting-quotas
says to manage projects roles and quotas using 'nova-manage'.  Should
I avoid the webUI and keystone and use nova-manage for this or is
there a different way to manage quotas I'm missing?


Thanks,
-Jon

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp


[Openstack] Noob question about api client access

2012-06-24 Thread Jonathan Proulx
Hi All,

I have a shiny new install of openstack and from the controller node
(the one with keystone, glance, horizon and all the nova pieces) I can
run everything as I expect using the nova cli tools upload images
boot instances manage the network, etc.  I can also use the Horizon
wed interface from remote systems just fine (of course all the actual
commands are run locally on the controller)

Where it get weird is when I try to use the nova cli on a remote
client.  Using the same openrc to set up the environment nova
credentials returns the expected result but all the other commands
I've tried (mostly of the list variety) fail like this:

$ nova list
ERROR: n/a (HTTP 400)


The install is Essex on Ubuntu Precise which I deployed using the
Puppet OpenStack module.  The example config puppet deployed for me
was only listening on 127.0.0.1 so I had to change the bind_host in
/etc/keystone/keystone.conf from 127.0.0.1 to 0.0.0.0 to get the 'nova
credentials' to run.  I've been sifting through the OpenStack Compute
Starter Guide and greping for and replacing references to localhost
or 127.0.0.1 in the various configs I alsoe tried adding
--allow_admin_api=true to the nova.conf which was in the openstack
docs but not in my conf,  but I'm clearly still missing something.

Hopefully this is obvious to some one more familiar with all the pieces?

Thanks,
-Jon

___
Mailing list: https://launchpad.net/~openstack
Post to : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp