Cinder error issues on Liberty

asked 2016-04-05 13:39:44 -0500

fleetdc gravatar image

updated 2016-04-11 11:25:59 -0500

rbowen gravatar image

Hello,

I have installed Openstack Liberty and I am having a problem with Cinder Volumes. I am using RDO running on CentOS 7.1 as I have checked all of my cinder.conf files making sure I followed the Liberty configuration instructions for the controller, compute and storage node. When I run cinder-mange service list I get this

#cinder-manage service list

No handlers could be found for logger "oslo_config.cfg"

2016-04-05 11:57:45.207 13660 DEBUG oslo_db.api [req-532c10c8-f3e8-4ec4-95b2-8e5b8cfcaafb - - - - -] Loading backend 'sqlalchemy' from 'cinder.db.sqlalchemy.api' _load_backend /usr/lib/python2.7/site-packages/oslo_db/api.py:230

/usr/lib/python2.7/site-packages/oslo_db/sqlalchemy/enginefacade.py:241: NotSupportedWarning: Configuration option(s) ['use_tpool'] not supported

  exception.NotSupportedWarning

2016-04-05 11:57:45.234 13660 DEBUG oslo_db.sqlalchemy.engines [req-532c10c8-f3e8-4ec4-95b2-8e5b8cfcaafb - - - - -] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION _check_effective_sql_mode /usr/lib/python2.7/site-packages/oslo_db/sqlalchemy/engines.py:256

Binary           Host                                 Zone             Status     State Updated At

cinder-scheduler controller                           nova             enabled    :-)   2016-04-05 15:57:36

cinder-volume    cloudnode1@lvm                       nova             enabled    XXX   2016-03-28 22:50:41

cinder-volume    blocknode@lvm                        nova             enabled    XXX   None

I have checked my Cinder volume logs on my Storage node along with my compute node with debug set and I see this in my logs:

2016-04-05 12:02:25.184 10926 ERROR cinder.service [-] Manager for service cinder-volume cloudnode1@lvm is reporting problems, not sending heartbeat. Service will appear "down".

2016-04-05 12:02:33.621 10926 DEBUG oslo_service.periodic_task [req-bb5f7cd0-ff99-443b-8211-1679199c413b - - - - -] Running periodic task VolumeManager._publish_service_capabilities run_periodic_tasks /usr/lib/python2.7/site-packages/oslo_service/periodic_task.py:213

2016-04-05 12:02:33.622 10926 DEBUG oslo_service.periodic_task [req-bb5f7cd0-ff99-443b-8211-1679199c413b - - - - -] Running periodic task VolumeManager._report_driver_status run_periodic_tasks /usr/lib/python2.7/site-packages/oslo_service/periodic_task.py:213

2016-04-05 12:02:33.623 10926 WARNING cinder.volume.manager [req-bb5f7cd0-ff99-443b-8211-1679199c413b - - - - -] Update driver status failed: (config name lvm) is uninitialized.

2016-04-05 12:02:35.185 10926 ERROR cinder.service [-] Manager for service cinder-volume cloudnode1@lvm is reporting problems, not sending heartbeat. Service will appear "down".

2016-04-05 12:02:45.191 10926 ERROR cinder.service [-] Manager for service cinder-volume cloudnode1@lvm is reporting problems, not sending heartbeat. Service will appear "down".

2016-04-05 12:02:55.191 10926 ERROR cinder.service [-] Manager for service cinder-volume cloudnode1@lvm is reporting problems, not sending heartbeat. Service will appear "down".

I am confused here is my cinder.conf file from my controller :

[DEFAULT]

verbose = True

debug = True

rpc_backend = rabbit

auth_strategy = keystone

my_ip = 10.0.0.30

[keystone_authtoken]

auth_uri = http://controller:5000

auth_url = http://controller:35357

auth_plugin = password

project_domain_id = default

user_domain_id = default

project_name = service

username = cinder

password = passwd

[database]

connection = mysql://cinder:passwd@controller/cinder

[oslo_messaging_rabbit]

rabbit_host = controller

rabbit_userid = rabbit

rabbit_password = xxxxxx

[oslo_concurrency]


lock_path = /var/lib/cinder/tmp

Here is my cinder.conf file from my compute/ storage node

[DEFAULT]

rpc_backend = rabbit

auth_strategy = keystone

my_ip = 10.0.0.40

glance_host= controller

enabled_backends = lvm

[database]

connection = mysql://cinder:password@10.0.0.30/cinder

[oslo_messaging_rabbit]

rabbit_host = controller

rabbit_userid = rabbit

rabbit_password = passwd

[keystone_authtoken]

auth_uri = http://controller:5000

auth_url = http://controller:35357

auth_plugin = password ...
(more)
edit retag flag offensive close merge delete

2 answers

Sort by ยป oldest newest most voted
0

answered 2016-06-13 01:44:01 -0500

I have encountered the same problem, and i solved it by adding a new disk /dev/sdb, and then run pvcreate /dev/sdb1,vgcreate cinder-volumes /dev/sdb1. Maybe you can try.

edit flag offensive delete link more
0

answered 2017-01-07 04:22:21 -0500

1st, please find the log file. /var/log/cinder/volume.log in cloudnode1 2nd.post the error info out

sometimes, it caused the config file /etc/cinder/cinder.conf, if the volume name (run vgdisplay) mismatch with config file. it also could not send heatbeat correctly.

Jichen

edit flag offensive delete link more

Get to know Ask OpenStack

Resources for moderators

Question Tools

1 follower

Stats

Asked: 2016-04-05 13:39:44 -0500

Seen: 6,598 times

Last updated: Apr 11 '16