cowley-tech/content/blog/upgrade-openstack-from-juno-to-kilo/index.md

236 lines
7.4 KiB
Markdown
Raw Normal View History

2024-01-18 20:13:37 +01:00
---
date: 2015-08-11
title: Upgrade Openstack from Juno to Kilo
category: devops
featured_image: http://i.imgur.com/UAyzTqf.gif
---
It's a process that strikes fear into the hearts of Sysadmins
everywhere. This weekend I finally got round to upgrading the Openstack
cluster in my lab to Kilo. As I have no spare machines lying around
(Intel NUC/HP Microserver/similar donations welcome) I did it in place.
Did it go well? Mostly\...
My base was a Juno install from Packstack. If your Juno install was
different, then YMMV, but the idea should transfer. The basic process
was to install the Kilo yum repo, then run an upgrade:
```
sudo yum install http://rdo.fedorapeople.org/openstack-kilo/rdo-release-kilo.rpm
sudo yum upgrade
```
Then a reboot. Finshed\...
No, nothing is ever that simple.
In fact most of the services fail dismally due to DB schema updates.
This is relatively easily fixed though
# Keystone
```
systemctl stop httpd openstack-keystone.service
systemctl disable openstack-keystone.service
sudo -u keystone keystone-manage db_sync
```
The application itself was not updated by the packages, so I collected
the lated code from Github:
```
curl http://git.openstack.org/cgit/openstack/keystone/plain/httpd/keystone.py?h=stable/kilo \
| sudo tee /var/www/cgi-bin/keystone/main /var/www/cgi-bin/keystone/admin
```
Ensure that to Apache config files are as the should be.
`/etc/httpd/conf.d/10-keystone_wsgi_admin.conf`:
```
<VirtualHost *:35357>
ServerName keystone.example.com
WSGIDaemonProcess keystone_admin processes=5 threads=1 user=keystone group=keystone
WSGIProcessGroup keystone_admin
WSGIScriptAlias / /var/www/cgi-bin/keystone/admin
WSGIPassAuthorization On
LogLevel info
ErrorLogFormat "%{cu}t %M"
ErrorLog /var/log/httpd/keystone-error.log
CustomLog /var/log/httpd/keystone-access.log combined
</VirtualHost>
```
and `/etc/httpd/conf.d/10-keystone_wsgi_main.conf`:
```
<VirtualHost *:5000>
ServerName keystone.example.com
WSGIDaemonProcess keystone-public processes=5 threads=1 user=keystone group=keystone display-name=%{GROUP}
WSGIProcessGroup keystone-public
WSGIScriptAlias / /var/www/cgi-bin/keystone/main
WSGIApplicationGroup %{GLOBAL}
WSGIPassAuthorization On
LogLevel info
ErrorLogFormat "%{cu}t %M"
ErrorLog /var/log/httpd/keystone-error.log
CustomLog /var/log/httpd/keystone-access.log combined
</VirtualHost>
```
Now restart Apache:
```
systemctl start httpd.service
```
# Glance
There was nothing surprising here really. Stop the services, update the
database and restart the services.
```
sudo systemctl stop openstack-glance-api.service openstack-glance-registry.service
sudo -u glance glance-manage db_sync
sudo systemctl start openstack-glance-api.service openstack-glance-registry.service
```
# Nova
Again Nova was quite simple. Stop services, update DB and start again.
```
sudo systemctl stop openstack-nova-api.service \
openstack-nova-cert.service openstack-nova-compute.service \
openstack-nova.compute.service openstack-nova-conductor.service \
openstack-nova-consoleauth.service openstack-nova-network.service \
openstack-nova-novncproxy.service openstack-nova-objectstore.service \
openstack-nova-scheduler.service openstack-nova-volume.service
sudo -u nova nova-manage db null_instance_uuid_scan
sudo -u nova "nova-manage db sync
sudo systemctl start openstack-nova-api.service \
openstack-nova-cert.service openstack-nova-compute.service \
openstack-nova.compute.service openstack-nova-conductor.service \
openstack-nova-consoleauth.service openstack-nova-network.service \
openstack-nova-novncproxy.service openstack-nova-objectstore.service \
openstack-nova-scheduler.service openstack-nova-volume.service
```
# Neutron
This need a few tweaks in `/etc/neutron/neutron.conf`.
In the \[DEFAULT\] section, change the value of the rpc\_backend option:
`neutron.openstack.common.rpc.impl_kombu` becomes `rabbit`
In the \[DEFAULT\] section, change the value of the core\_plugin option:
`neutron.plugins.ml2.plugin.Ml2Plugin` becomes `ml2`
In the \[DEFAULT\] section, change the value or values of the
service\_plugins option to use short names. For example:
`neutron.services.l3_router.l3_router_plugin.L3RouterPlugin` becomes
`router`
In the \[DEFAULT\] section, explicitly define a value for the
`nova_region_name` option. For example:
```
[DEFAULT]
...
nova_region_name = regionOne
```
Stop the services and upgrade the DB:
```
sudo systemctl stop neutron-dhcp-agent.service neutron-l3-agent.service \
neutron-metadata-agent.service neutron-openvswitch-agent.service \
neutron-ovs-cleanup.service neutron-server.service
sudo -u neutron neutron-db-manage --config-file /etc/neutron/neutron.conf \
--config-file /etc/neutron/plugins/ml2/ml2_conf.ini upgrade kilo
```
Now you can restart Neutron
```
sudo systemctl start neutron-dhcp-agent.service neutron-l3-agent.service \
neutron-metadata-agent.service neutron-openvswitch-agent.service \
neutron-ovs-cleanup.service neutron-server.service
```
# Horizon
This pretty much worked, but what I did see is that once mu login ticket
expired then I could not login unless I cleared the cookie out.
This is easily fixed by adding `AUTH_USER_MODEL = 'openstack_auth.User'`
to `/etc/openstack-dashboard/local_settings`.
# Cinder
This is what gave me the most problems. Basically, the database for
Cinder itself, the database for Nova volumes and the actual iSCSI target
got out of sync. I ran `nova volume-detach ...` and it got stuck in a
detaching state.
Basically, I had to go through and get it into a know state (volumes
attached to anything) via the back door.
As an admin, force the volume into \"available\" with:
```
nova volume-detach <instance_uuid> <volume_id>
cinder reset-state --state available <volume_id>
```
Using
[targetcli](http://linux-iscsi.org/wiki/Targetcli#Quick_start_guide)
ensure that there are no ACLs on the LUNs. They will be named with the
volume\_id. I\'ll not go into the details of how to use `targetcli`,
just that you remove a *file* from the virtual tree that it creates.
Next up you\'ll need to going to manipulate the Cinder database (hope
you still have your packstack file). Standard disclaimer: You can
royally screw things up here, so tread carefully, use transactions and
take a backup first.
```
update cinder.volumes set attach_status='detached',
status='available' where id ='$volume_id';
```
Now do the same in Nova.
```
delete from block_device_mapping where not deleted
and volume_id='$volume_id'
```
You should now be able re-attach the volume to the instance using the
CLI. However, I had one that persisted in playing silly buggers. I had
to manually update the Cinder DB that is in the *attached* state:
```
update cinder.volumes set attach_status='attached',
status='in-use' where id ='$volume_id';
```
Finally do a full reboot to ensure that everything comes back as you
expect.
I am pretty sure that is everything.
# Conclusion
I think this was the first time I have done an upgrade of Openstack in
place. Considering the fear that this operation puts in people, I think
it went pretty smoothly.
I started the install Friday evening, the upgrade was finished that
night. Most of my lab instances were up and running by Saturday evening
(having spent the day at the beach). All bar one were running Sunday
evening (after another trip to the beach). The last instance (with the
awkward Cinder volume) was running this morning (again, wait for it:
after a trip to the beach yesterday).