部署ocp报错ERROR 4669 (HY000): cluster not exist

【 使用环境 】 测试环境
【 OB or 其他组件 】ocp
【 使用版本 】ocp-4.0.3-ce
【问题描述】黑屏方式部署ocp报错集群找不到
【复现路径】问题出现前后相关操作
【问题现象及影响】
报错信息如下:
1
Traceback (most recent call last):
File “pipeline/install_ocp.py”, line 47, in
run(context)
File “pipeline/install_ocp.py”, line 37, in run
install_ocp_pipeline.run()
File “/root/installer/core/pipeline.py”, line 10, in run
task.run()
File “/root/installer/task/metadb_init.py”, line 37, in run
raise Exception(“init metadb failed, {} {} {}”.format(init_metadb_result.code, init_metadb_result.stdout, init_metadb_result.stderr))
Exception: init metadb failed, 1 ERROR 4669 (HY000): cluster not exist

配置文件内容如下:

OCP deploy config

Note:

Do not use 127.0.0.1 or hostname as server address

When a server has both public ip and private ip, if private ip is connectable, use private ip for faster connection

If a vip is configured, it should be already created and bonded to the right server and port, the installation script won’t do any work with vip maintainance, just use it to connect to the service

Ignore precheck errors

It’s recommanded to not ignore precheck errors

precheck_ignore: true

Create an obcluster as OCP’s metadb

create_metadb_cluster: false

Clean OCP’s metadb cluster when uninstall

clean_metadb_cluster: false

Metadb cluster deploy config

ob_cluster:

name: obcluster

home_path: /data/apps/oceanbase

root_password: ‘rootpass’

The directory for data storage, it’s recommanded to use an independent path

data_path: /data/ocp

The directory for clog, ilog, and slog, it’s recommanded to use an independent path.

redo_path: /data/logs/ocp

sql_port: 2881

rpc_port: 2882

zones:

- name: zone1

  servers:

    - 10.20.19.225

- name: zone2

  servers:

    - 10.20.19.81

- name: zone3

  servers:

    - 10.20.19.97

custom obd config for obcluster

custom_config:

- key: devname

value: eth0

- key: cpu_count

value: 0

- key: memory_limit

value: 0G

- key: system_memory

value: 30G

- key: __min_full_resource_pool_memory

value: 5368709120

- key: datafile_maxsize

value: 0

- key: datafile_next

value: 0

Meta user info

meta:

tenant: meta_tenant

user: meta_user

password: q4SD77ha0PPTN8aZ

database: meta_database

cpu: 2

# Memory configs in GB, 4 means 4GB

memory: 4

Monitor user info

monitor:

tenant: monitor_tenant

user: root

password: q4SD77ha0PPTN8aZ

database: monitor_database

cpu: 4

# Memory configs in GB, 8 means 8GB

memory: 8

Obproxy to connect metadb cluster

obproxy:

home_path: /data/apps/obproxy

port: 2883

servers:

- 10.20.30.125

custom config for obproxy

custom_config:

- key: clustername

value: obcluster

Vip is optional, if vip is not configured, one of obproxy servers’s address will be used

vip:

address: 1.1.1.1

port: 2883

Ssh auth config

ssh:

port: 22

user: root

auth method, support password and pubkey

auth_method: password

timeout: 10

password: wXR@mG%Vjj3EkBAa

OCP config

ocp:

ocp container’s name

name: ‘ocp’

OCP process listen port and log dir on host

process:

port: 8080

log_dir: /data/logs/ocp

servers:

- 10.20.30.125

OCP container’s resource

resource:

cpu: 2

# Memory configs in GB, 8 means 8GB

memory: 8

Vip is optional, if vip is not configured, one of ocp servers’s address will be used

vip:

address: 1.1.1.1

port: 8080

OCP basic auth config, used when upgrade ocp

auth:

user: root

password: ***********

options:

# enable front-end tracking, non-business sensitive data such as system information, user behavior, and front-end anomalies will be collected to improve the product experience and stability

enable_web_tracking: true

OCP metadb config, for ocp installation, if “create_metadb_cluster” is configured true, this part will be replaced with the configuration of metadb cluster and obproxy

metadb:

host: 10.20.30.125

port: 2883

meta_user: meta_user@meta_tenant#obcluster

meta_password: q4SD77ha0PPTN8aZ

meta_database: meta_database

monitor_user: monitor_user@monitor_tenant#obcluster

monitor_password: q4SD77ha0PPTN8aZ

monitor_database: monitor_database

【附件】

集群名称似乎不对,看下实际集群名和安装配置里的是否一致

你好,总结问题是 由于mysql -h10.20.30.125 -P2883 -umeta_user@meta_tenant#cluster_name -p这个命令连接的时候就报错集群不存在。当删除集群名字后就没有问题了。想问下是哪里操作的不对,还是因为版本的问题啊

1 个赞

这个集群名应该用你实际的集群名称

如果带#cluster_name这个的话 需要连接proxy对应的ip和端口而不是observer的ip端口

目前问题解决了。确实是集群名称没有写对。 部署名称和集群名称是两个概念,切勿搞混走弯路。
show parameters like ‘%cluster%’; 查看下实际的集群名称后再部署。 谢谢大家

1 个赞

下次优先 仍个 单号。 我要 赚积分。 哈哈哈