Ask Your Question

tekkafedora's profile - activity

2020-06-04 14:13:11 -0500 received badge  Notable Question (source)
2020-05-19 11:11:47 -0500 received badge  Popular Question (source)
2020-05-11 08:26:33 -0500 asked a question horizon connection reset from external network

Hello, I have installed Openstack Queens on CentOS 7 with 3 controllers. At the end of overcloud deploy I got:

 2020-05-04 09:34:19Z [overcloud]: CREATE_COMPLETE  Stack CREATE completed successfully
 Stack overcloud CREATE_COMPLETE 
Host 172.23.0.231 not found in /home/stack/.ssh/known_hosts
Started Mistral Workflow tripleo.deployment.v1.get_horizon_url. Execution ID: 13b02fc2-dadf-41bb-8b68-ca1502e472b4
Overcloud Endpoint: http://172.23.0.231:5000/
Overcloud Horizon Dashboard URL: http://172.23.0.231:80/dashboard
Overcloud rc file: /home/stack/overcloudrc
Overcloud Deployed

Director node has its host ip on 172.19.0/24 network and during undercloud install it set up:

7: br-ctlplane: <broadcast,multicast,up,lower_up> mtu 1500 qdisc noqueue state UNKNOWN group default qlen 1000
    link/ether 56:6f:3d:48:00:14 brd ff:ff:ff:ff:ff:ff
    inet 172.23.0.220/24 brd 172.23.0.255 scope global br-ctlplane
       valid_lft forever preferred_lft forever
    inet 172.23.0.222/32 scope global br-ctlplane
       valid_lft forever preferred_lft forever
    inet 172.23.0.221/32 scope global br-ctlplane
       valid_lft forever preferred_lft forever
    inet6 fe80::546f:3dff:fe48:14/64 scope link 
       valid_lft forever preferred_lft forever
8: docker0: <no-carrier,broadcast,multicast,up> mtu 1500 qdisc noqueue state DOWN group default 
    link/ether 02:42:6a:09:7f:36 brd ff:ff:ff:ff:ff:ff
    inet 172.17.0.1/16 scope global docker0
       valid_lft forever preferred_lft forever

From a desktop on the 172.23.0.0 network I'm able to access the dashboard, while from a desktop on another network (also the network of undercloud director for example, the 172.19.0.0) I receive a "the connection was reset" message in the browser (ERR_CONNECTION_RESET).

Is this the default expected setup usign tripleo? How can I change the configuration to allow connections coming from different networks? If I go on controller node where the cluster resource ip-172.23.0.231 resource is active I can see that:

Going into the haproxy container, the haproxy:

[root@ostack-ctrl1 ~]# docker exec -it haproxy-bundle-docker-1 bash
()[root@ostack-ctrl1 /]#
()[root@ostack-ctrl1 /]# more /etc/haproxy/haproxy.cfg 
....
listen horizon
  bind 172.23.0.231:80 transparent
  mode http
  cookie SERVERID insert indirect nocache
  option forwardfor
  option httpchk
  server ostack-ctrl1.internalapi.localdomain 172.23.0.227:80 check cookie ostack-ctrl1.internalapi.localdomain fall 5 i
nter 2000 rise 2
  server ostack-ctrl2.internalapi.localdomain 172.23.0.235:80 check cookie ostack-ctrl2.internalapi.localdomain fall 5 i
nter 2000 rise 2
  server ostack-ctrl0.internalapi.localdomain 172.23.0.238:80 check cookie ostack-ctrl0.internalapi.localdomain fall 5 i
nter 2000 rise 2
...

As the first in the chain is controller1, I go inside its horizon container and its config is this:

# docker exec -it horizon bash
()[root@ostack-ctrl0 /]# grep ^ALLOW /etc/openstack-dashboard/local_settings
ALLOWED_HOSTS = ['*', ]
()[root@ostack-ctrl0 /]# 

And the if I "tail -f horizon.log", when I try to open the horizon dashboard page coming from an ip on 172.19.0.0 network I see this inside it:

2020-05-11 12:18:38,961 61 ...
(more)
2020-05-10 07:39:34 -0500 received badge  Famous Question (source)
2020-05-08 02:06:10 -0500 received badge  Enthusiast
2020-05-06 17:49:15 -0500 asked a question Train on CentOS 8 methods?

Hello, I have completed a Queens Openstack installation using Tripleo and CentOS 7 and configuring 2 compute nodes, 3 controller nodes and 3 ceph OSD nodes, with containers and Pacemaker, to be similar enough with OSP 13.

Now I would like to try Openstack Train using CentOS 8, to get a configuration similar to the OSP 16 one.

At this page: https://docs.openstack.org/train/deploy/index.html (link text) I see:

  • Deploying OpenStack using Ansible in Docker Containers (kolla) Guide => both CentOS 7 and 8 referenced

  • OpenStack-Ansible Deployment Guide => only CentOS 7 referenced

  • TripleO Deployment Guide => only CentOS 7 referenced

Are they consistent so that to use CentOS 8 I can only go with the first option and not with TripleO? Is TripleO going to be deprecated, perhaps?

Thanks in advance, Gianluca

2020-05-06 14:00:05 -0500 received badge  Notable Question (source)
2020-05-04 18:55:14 -0500 received badge  Popular Question (source)
2020-05-04 07:55:46 -0500 answered a question Error during ceph nodes setup in copy ceph keys

Problem solved.

The reason was that for overcloud deployment I used a custom roles_data.yaml file (called with "-r") containing CephStorage role, expecting the workflow to merge and parse also the default ones.

Instead it processed ONLY the custom role defined inside roles_data.yaml; I added Controller and Compute roles in my custom file and all went ok.

Gianluca

2020-05-04 01:26:03 -0500 asked a question Error during ceph nodes setup in copy ceph keys

I'm setting up a small test lab for HA Openstack environment based on Queens and CentOS 7 (to be the best similar with OSP 13, as my target). Director is CentOS 7.8 with ansible ansible-2.6.19-1.el7.ans.noarch and undercloud deployed.

Idea is to have 2 compute, 3 controllers, 3 ceph storage nodes

As the first step of overcloud deploy the correct 3 ceph nodes candidates for OSD role are powered on and begin to be configured, but at a certain point I have with this error

2020-05-03 22:31:07Z [AllNodesDeploySteps.CephStorageDeployment_Step1.0]: CREATE_COMPLETE  state changed
2020-05-03 22:31:07Z [AllNodesDeploySteps.CephStorageDeployment_Step1]: CREATE_COMPLETE  Stack CREATE completed successfully
2020-05-03 22:31:08Z [AllNodesDeploySteps.CephStorageDeployment_Step1]: CREATE_COMPLETE  state changed
2020-05-03 22:31:08Z [AllNodesDeploySteps.WorkflowTasks_Step2]: CREATE_IN_PROGRESS  state changed
2020-05-03 22:31:09Z [AllNodesDeploySteps.WorkflowTasks_Step2]: CREATE_COMPLETE  state changed
2020-05-03 22:31:09Z [AllNodesDeploySteps.WorkflowTasks_Step2_Execution]: CREATE_IN_PROGRESS  state changed
2020-05-03 22:33:49Z [AllNodesDeploySteps.WorkflowTasks_Step2_Execution]: CREATE_FAILED  resources.WorkflowTasks_Step2_Execution: Failure caused by error in tasks: ceph_base_ansible_workflow

  ceph_base_ansible_workflow [task_ex_id=b9eee634-5f66-47f9-bef2-57e46cd6d80f] -> Failure caused by error in tasks: ceph_install

  ceph_install [task_e
2020-05-03 22:33:49Z [AllNodesDeploySteps]: CREATE_FAILED  Resource CREATE failed: resources.WorkflowTasks_Step2_Execution: Failure caused by error in tasks: ceph_base_ansible_workflow
...
overcloud.AllNodesDeploySteps.WorkflowTasks_Step2_Execution:
  resource_type: OS::TripleO::WorkflowSteps
  physical_resource_id: 71e920fe-b781-4d41-b082-cf27c7bbdb4c
  status: CREATE_FAILED
  status_reason: |
    resources.WorkflowTasks_Step2_Execution: Failure caused by error in tasks: ceph_base_ansible_workflow

      ceph_base_ansible_workflow [task_ex_id=b9eee634-5f66-47f9-bef2-57e46cd6d80f] -> Failure caused by error in tasks: ceph_install

      ceph_install [task_ex_id=567a8737-4abe-4b24-89e9-6116d4e8eff2] -> One or more actions had failed.
...
     Unexpected error while running command.
    Command: ansible-playbook /usr/share/ceph-ansible/site-docker.yml.sample --user tripleo-admin --become --become-user root --extra-vars {"ireallymeanit": "yes", "osd_pool_default_pgp_num": 16, "osd_pool_default_pg_num": 16} --inventory-file /tmp/ansible-mistral-actionV7vbpj/inventory.yaml --private-key /tmp/ansible-mistral-actionV7vbpj/ssh_private_key --skip*** package-install,with_pkg
    Exit code: 2

And this in /var/log/mistral/ceph-install-workflow.log:

2020-05-04 00:33:45,311 p=17037 u=mistral |  TASK [ceph-osd : include_tasks common.yml] *************************************
2020-05-04 00:33:45,311 p=17037 u=mistral |  Monday 04 May 2020  00:33:45 +0200 (0:00:00.145)       0:01:48.161 ************ 
2020-05-04 00:33:45,517 p=17037 u=mistral |  included: /usr/share/ceph-ansible/roles/ceph-osd/tasks/common.yml for 172.23.0.239, 172.23.0.229, 
172.23.0.234
2020-05-04 00:33:45,562 p=17037 u=mistral |  TASK [ceph-osd : create bootstrap-osd and osd directories] *********************
2020-05-04 00:33:45,562 p=17037 u=mistral |  Monday 04 May 2020  00:33:45 +0200 (0:00:00.251)       0:01:48.413 ************ 
2020-05-04 00:33:45,793 p=17037 u=mistral |  ok: [172.23.0.239] => (item=/var/lib/ceph/bootstrap-osd/)
2020-05-04 00:33:45,845 p=17037 u=mistral |  ok: [172.23.0.229] => (item=/var/lib/ceph/bootstrap-osd/)
2020-05-04 00:33:45,891 p=17037 u=mistral |  ok: [172.23.0.234] => (item=/var/lib/ceph/bootstrap-osd/)
2020-05-04 00:33:45,995 p=17037 u=mistral |  ok: [172.23.0.239] => (item=/var/lib/ceph/osd/)
2020-05-04 00:33:46,044 p=17037 u=mistral |  ok: [172.23.0.229] => (item=/var/lib/ceph/osd/)
2020-05-04 00:33:46,086 p=17037 u=mistral |  ok: [172.23 ...
(more)
2020-05-01 09:56:19 -0500 received badge  Famous Question (source)
2020-04-29 05:29:35 -0500 answered a question Small HA installation: ceph nodes not considered?

Thanks for answering.

For me there is no problem to have mon and mgr services to run on osd nodes, the problem is how to compile instackenv.json and instruct deploy command (through env.yaml files) to do so....

I was basing on some links like:

https://access.redhat.com/documentation/en-us/red_hat_openstack_platform/13/html/director_installation_and_usage/chap-requirements#sect-Environment_Requirements (https://access.redhat.com/documentati...)

https://docs.openstack.org/project-deploy-guide/tripleo-docs/latest/provisioning/profile_matching.html (https://docs.openstack.org/project-de...)

But it is not clear to me how to create the map to say the installer to use the 3 dedicated ceph nodes I have chosen for that.... can you give a tip about it?

You say "without OSD you can't store data", but I would expect the deploy command to setup OSD nodess... while my 3 nodes elected to be the ceph ones are "correctly" (in the sense that they are powered on and then off) introspected, but then they seem to have been excluded from the whole workflow operation... Can you confirm that at the stage where compute and controller nodes are powered on, also ceph nodes should have been powered on at the same time? Or are they expected to be powered on only in a second moment?

Are my settings below in jstackenv.json correct for the candidate ceph nodes? Or what do I have to change?

"capabilities": "profile:ceph-storage,node:ceph-2,boot_option:local"
2020-04-29 02:26:00 -0500 received badge  Notable Question (source)
2020-04-29 02:26:00 -0500 received badge  Popular Question (source)
2020-04-28 16:42:27 -0500 asked a question Small HA installation: ceph nodes not considered?

Hello, sorry for long post, I try to give more details possible below. I'm setting up a small test lab for HA Openstack environment based on Queens and CentOS 7 (to be the best similar with OSP 13, as my target) and the Openstack nodes will be oVirt VMs. Also the director is a VM in oVirt.

Idea is to have 2 compute, 3 controllers, 3 ceph storage nodes (for image, block, object and manila). The nodes have 1 60Gb root disk; the ceph nodes have 2 more disks (100Gb for journal and 150Gb for OSD).

I have installed undercloud and I have made up some combinations of instackenv.json file for introspection and all nodes are correctly introspected with VMs powered on and off. I have 4 questions:

  • which value to use for this small storage cluster and ovveride default ceph parameters (pgnum, mon_max_pg_per_osd, ecc.) without getting errors during deploy?

  • what is the correct parameter to set in instackenv.json or through "openstack baremetal node set --property .." command to have a map for ceph OSD for the 3 designated hosts?

  • at which stage of the overcloud deploy are ceph nodes expected to be powered on and installed?

  • is it correct that in this architecture layout mon, mgr and mds are deployed on controller nodes as docker containers while only OSD on the dedicated storage nodes?

Thanks, Gianluca

Details: For ceph OSD nodes I have tried to give these capabilities in instackenv.json file:

"name": "ostack-ceph2",
"capabilities": "profile:ceph-storage,node:ceph-2,boot_option:local"

with then a scheduler_hints_env.yaml file of this type:

parameter_defaults:
  ControllerSchedulerHints:
    'capabilities:node': 'controller-%index%'
  ComputeSchedulerHints:
    'capabilities:node': 'compute-%index%'
  CephStorageSchedulerHints:
    'capabilities:node': 'ceph-%index%'
  HostnameMap:
    overcloud-controller-0: ostack-ctrl0
    overcloud-controller-1: ostack-ctrl1
    overcloud-controller-2: ostack-ctrl2
    overcloud-novacompute-0: ostack-compute0
    overcloud-novacompute-1: ostack-compute1
    overcloud-ceph-storage-0: ostack-ceph0
    overcloud-ceph-storage-1: ostack-ceph1
    overcloud-ceph-storage-2: ostack-ceph2

But while compute and controllers are deployed ok and their hostnames are also correctly mapped, ceph nodes remain untouched, not even powered on; I don't know if it depends on expected workflow and they need to be set up only at a final stage that doesn't arrive.

To accomplish this, for ceph I'm giving to overcloud deploy these environment files:

-e /usr/share/openstack-tripleo-heat-templates/environments/ceph-ansible/ceph-ansible.yaml \
-e /usr/share/openstack-tripleo-heat-templates/environments/ceph-ansible/ceph-mds.yaml \
-e /usr/share/openstack-tripleo-heat-templates/environments/manila-cephfsnative-config.yaml \

I'm also using this env file:

parameter_defaults:
  ControllerCount: 3
  ComputeCount: 2
  CephCount: 3

BTW: I receive at very beginning that CephCount parameter is ignored (?) Initially I received errors during Ceph setup due to low PGs defaut numbers:

"stderr": "Error ERANGE:  pg_num 128 size 3 would mean 768 total pgs, which exceeds max 750 (mon_max_pg_per_osd 250 * num_in_osds 3)"

So I'm trying to change with this env file:

parameter_defaults:
  CephPoolDefaultSize: 3
  CephPoolDefaultPgNum: 64
  CephConfigOverrides:
    mon_max_pg_per_osd: 400

Right now the deploy seems stuck after step

2020-04-28 14:04:27Z [overcloud.AllNodesDeploySteps.ControllerDeployment_Step5.2]: CREATE_COMPLETE  state changed

and on controller nodes I have:

[root@ostack-ctrl0 ~]# ceph -s
  cluster:
    id:     5d194678-8950-11ea-b8c5-566f3d480013
    health: HEALTH_WARN
            1 MDSs report slow metadata IOs
            Reduced data ...
(more)