unknown module 'rsync_sst-undo_dir' error when syncing node - mysql

I have a galera cluster with 3 nodes in QA environement.. There was a abrupt shut down of the nodes and I restarted and bootstrapped the first nodewhich is working fine. The second node when I try to start gives the error , "rsyncd[30958]: unknown module 'rsync_sst-undo_dir' tried from XXX" .
In mysql log :
[Warning] WSREP: 0.0 : State transfer to 1.0 (x) failed: -255 (Unknown error 255) [ERROR] WSREP: /home/buildbot/buildbot/build/gcs/src/gcs_group.cpp:gcs_group_handle_join_msg():1207: Will never receive state. Need to abort.
Not sure what causes this error,
rsyncd[30958]: unknown module 'rsync_sst-undo_dir' tried from XXX systemd[1]: mariadb.service: main process exited, code=killed, status=6/ABRT unknown module 'rsync_sst-undo_dir' "
Any pointers would be very helpful .
Thanks
Any pointers would be very helpful .

Related

MySQL Ubuntu 19.10 Installation Error - System Schema directory does not exist

I have attempted to conduct a clean install of MySQL on my ubuntu19.10 install
However during installs I continue to get the following errors while attempting to install MySql on Ubuntu 191.10.
within /var/log/MySql/error.log:
2020-03-31T02:04:00.133031Z 1 [ERROR] [MY-013178] [Server] Execution of server-side SQL statement 'CREATE SCHEMA mysql DEFAULT COLLATE utf8mb4_0900_ai_ci' failed with error code = 1049, error message = 'System schema directory does not exist.'.
2020-03-31T02:04:00.133738Z 0 [ERROR] [MY-010020] [Server] Data Dictionary initialization failed.
2020-03-31T02:04:00.134159Z 0 [ERROR] [MY-010119] [Server] Aborting
2020-03-31T02:04:00.662756Z 0 [System] [MY-010910] [Server] /usr/sbin/mysqld: Shutdown complete (mysqld 8.0.19-0ubuntu0.19.10.3) (Ubuntu).
Installing gives the following errors:
$ sudo apt-get install mysql-server
Reading package lists... Done
Building dependency tree
Reading state information... Done
mysql-server is already the newest version (8.0.19-0ubuntu0.19.10.3).
0 upgraded, 0 newly installed, 0 to remove and 0 not upgraded.
2 not fully installed or removed.
After this operation, 0 B of additional disk space will be used.
Do you want to continue? [Y/n] y
Setting up mysql-server-8.0 (8.0.19-0ubuntu0.19.10.3) ...
Renaming removed key_buffer and myisam-recover options (if present)
mysqld will log errors to /var/log/mysql/error.log
2020-03-31T02:04:00.134301Z 0 [ERROR] [MY-010946] [Server] Failed to start mysqld daemon. Check mysqld error log.
Warning: Unable to start the server.
AppArmor parser error for /etc/apparmor.d/usr.sbin.mysqld in /etc/apparmor.d/usr.sbin.mysqld at line 9: Could not open 'abstractions/mysql'
Job for mysql.service failed because the control process exited with error code.
See "systemctl status mysql.service" and "journalctl -xe" for details.
invoke-rc.d: initscript mysql, action "start" failed.
● mysql.service - MySQL Community Server
Loaded: loaded (/lib/systemd/system/mysql.service; enabled; vendor preset: enabled)
Active: activating (auto-restart) (Result: exit-code) since Mon 2020-03-30 21:04:03 CDT; 9ms ago
Process: 17035 ExecStartPre=/usr/share/mysql/mysql-systemd-start pre (code=exited, status=1/FAILURE)
Mar 30 21:04:03 ubuntu-vm systemd[1]: mysql.service: Control process exited, code=exited, status=1/FAILURE
Mar 30 21:04:03 ubuntu-vm systemd[1]: mysql.service: Failed with result 'exit-code'.
Mar 30 21:04:03 ubuntu-vm systemd[1]: Failed to start MySQL Community Server.
dpkg: error processing package mysql-server-8.0 (--configure):
installed mysql-server-8.0 package post-installation script subprocess returned error exit status 1
No apport report written because the error message indicates its a followup error from a previous failure.
dpkg: dependency problems prevent configuration of mysql-server:
mysql-server depends on mysql-server-8.0; however:
Package mysql-server-8.0 is not configured yet.
dpkg: error processing package mysql-server (--configure):
dependency problems - leaving unconfigured
Errors were encountered while processing:
mysql-server-8.0
mysql-server
E: Sub-process /usr/bin/dpkg returned an error code (1)
Try to install mariadb first sudo apt install mariadb-server
After this you can normally use mysql but if you have any error try to reinstall mysql with this command : sudo apt install mysql-server
Mariadb is the same as MySQL it’s just open source project after the buy of MySQL by oracle so an employee have quit MySQL to create a new SGBD with an high compatibility of MySQL.

mariadb 10.1.13 galera cluster : error

mariadb 10.1.x galera cluster setting.
first node 192.168.159.132
/etc/mysql/my.cnf
[galera]
# Mandatory settings
wsrep_on=ON
wsrep_provider=/usr/lib/galera/libgalera_smm.so
wsrep_cluster_address=gcomm://192.168.159.132,192.168.159.135,192.168.159.136
binlog_format=row
default_storage_engine=InnoDB
innodb_autoinc_lock_mode=2
first node 192.168.159.132
$ sudo service mysql bootstrap
* Bootstrapping the cluster mysqld
Starting mysql (via systemctl): mysql.service
Job for mariadb.service failed because the control process exited with error code. See "systemctl status mariadb.service" and "journalctl -xe" for details.
$ systemctl status mariadb.service
[ERROR] WSREP: gcs/src/gcs_core.cpp:gcs_core_open():208: Failed to open backend connection: -110 (Connection timed out)
[ERROR] WSREP: gcs/src/gcs.cpp:gcs_open():1379: Failed to open channel 'my_wsrep_cluster' at 'gcomm://192.168.32.251,192.168.32.252,192.168.32...ection timed out)
[ERROR] WSREP: gcs connect failed: Connection timed out
[ERROR] WSREP: wsrep::connect(gcomm://192.168.32.251,192.168.32.252,192.168.32.253) failed: 7
[ERROR] Aborting
systemd[1]: mariadb.service: Main process exited, code=exited, status=1/FAILURE
systemd[1]: Failed to start MariaDB database server.
systemd[1]: mariadb.service: Unit entered failed state.
systemd[1]: mariadb.service: Failed with result 'exit-code'.
Why "Galera Cluster" not Start?
How To checking "Connection Time Out"?
There is a Error in your config. After gcom you only use one : like:
wsrep_cluster_address=gcomm://192.168.159.132,192.168.159.135,192.168.159.136
their is an error:
use format as below:
wsrep_cluster_address="gcomm://first_ip,second_ip,third_ip"
The galera configuration looks fine. But you have to provide an extra argument while bootstrapping the first node. Try this:
mysqld --wsrep-new-cluster
For the first node wsrep_cluster_address can be empty. For next nodes you have to provide at least one ip address that is already a part of cluster. But it's a good practice to provide the list of all the nodes that you are going to use. It wont give you any error for mentioning a node in cluster-address which is not present in the cluster yet. So you can use, wsrep_cluster_address=gcomm://192.168.159.132,192.168.159.135,192.168.159.136 for every node.
You are getting "connection timed out" because the node is trying to connect to the other nodes listed in wsrep_cluster_address, which aren't online as this is the very first node in the cluster. So, during bootstrap you should start the first node with empty wsrep_cluster_address :.wsrep_cluster_address='gcomm://'.

Mysql server can not start as root

start mysql server as root failed.
1.service mysqld start
Starting mysqld: [FAILED]
2.mysqld
error log:
2016-02-23T01:19:16.961857Z 0 [Warning] TIMESTAMP with implicit DEFAULT value is deprecated. Please use --explicit_defaults_for_timestamp server option (see documentation for more details).
2016-02-23T01:19:16.963350Z 0 [Note] mysqld (mysqld 5.7.10) starting as process 20924 ...
2016-02-23T01:19:16.965669Z 0 [ERROR] Fatal error: Please read "Security" section of the manual to find out how to run mysqld as root!
2016-02-23T01:19:16.965715Z 0 [ERROR] Aborting
2016-02-23T01:19:16.965737Z 0 [Note] Binlog end
2016-02-23T01:19:16.965818Z 0 [Note] mysqld: Shutdown complete
but can start mysql service as mysql user, or use mysqld_safe as root can start too. what's wrong?
/etc/init.d/msyqld start
MySQL Daemon failed to start.
Starting mysqld: [FAILED]
if ! /bin/kill -0 $safe_pid 2>/dev/null; then
echo "MySQL Daemon failed to start."
ret=1
break
fi
... ...
if [ $ret -eq 0 ]; then
action $"Starting $prog: " /bin/true
touch $lockfile
else
action $"Starting $prog: " /bin/false
fi
I found this Code Snippets in the script /etc/init.d/mysqld. but don't know why?
Add this line to /etc/my.cnf in [mysqld] section
explicit_defaults_for_timestamp = 1
Check this link
Since you can start as the mysql user, it is likely you have it currently configured to run as root.
Mysql's reference manual says:
On Linux, for installations performed using a MySQL repository, RPM packages, or Debian packages, the MySQL server mysqld should be started by the local mysql operating system user. Starting by another operating system user is not supported by the init scripts that are included as part of the installation.
/etc/selinux/config
#SELINUX=enforcing
SELINUX=disabled

logstash-forwarder json invalid character

At some point my logstash forwarder used to work properly but at some point fail to start with this error message
Failed unmarshalling json: invalid character ']' looking for beginning of value logstash
This message appears only if i run forwarder with this command:
$ /opt/logstash-forwarder/bin/logstash-forwarder -config="/etc/logstash-forwarder"
If i run it as a service it fails without any error message, if i restart the service getting this
Restarting log shipper logstash-forwarder
start-stop-daemon: warning: failed to kill PID: No such process
...done.
The first time stopped forwarder working i found in syslog this message:
connection from UDP: [54.194.184.218]:63707->[10.38.143.181] Nov 21
12:41:38 logstash-forwarder[PID]: 2014/11/21 12:41:38.692517 Read
error looking for ack: EOF Nov 21 12:41:38 logstash-forwarder[PID]:
2014/11/21 12:41:38.692592 Setting trusted CA from file:
/etc/pki/tls/certs/logstash-forwarder.crt Nov 21 12:41:38
logstash-forwarder[PID]: 2014/11/21 12:41:38.692827 Connecting to
X.X.X.X:5000 (X.X.X.X) Nov 21 12:41:38 logstash-forwarder[PID]:
2014/11/21 12:41:38.694128 Failure connecting to X.X.X.X: dial tcp
X.X.X.X:5000: connection refused
At server side logstash and ES cluster seems to work fine..
I found many similar problems with mine but not any solution.

Percona Xtradbcluster: Error while getting data from donor node

I've a problem in joining new node to percona xtradbcluster.
here is my log :
I tested the test_sstuser working on honor without any problem .
but in new nodes I get these errors :
2014-05-31 04:01:40 2391 [Warning] WSREP: Failed to prepare for incremental state transfer: Local state UUID (00000000-0000-0000-0000-000000000000) does not match group state UUID (02dc9b38-e825-11e3-a67b-4bfddfde0c98): 1 (Operation not permitted)
at galera/src/replicator_str.cpp:prepare_for_IST():447. IST will be unavailable.
WSREP_SST: [ERROR] Error while getting data from donor node: exit codes: 1 0 (20140531 04:01:40.934)
WSREP_SST: [ERROR] Cleanup after exit with status:32 (20140531 04:01:40.936)
2014-05-31 04:01:40 2391 [ERROR] WSREP: Process completed with error: wsrep_sst_xtrabackup-v2 --role 'joiner' --address '192.168.0.2' --auth 'test_sstuser:123456' --datadir '/var/lib/mysql/' --defaults-file '/etc/mysql/my.cnf' --parent '2391' '' : 32 (Broken pipe)
2014-05-31 04:01:40 2391 [ERROR] WSREP: Failed to read uuid:seqno from joiner script.
2014-05-31 04:01:40 2391 [ERROR] WSREP: SST failed: 32 (Broken pipe)
2014-05-31 04:01:40 2391 [ERROR] Aborting
2014-05-31 04:01:41 2391 [Warning] WSREP: 1.0 (test.host.com): State transfer to 0.0 (master-2.test.net) failed: -32 (Broken pipe)
2014-05-31 04:01:41 2391 [ERROR] WSREP: gcs/src/gcs_group.c:gcs_group_handle_join_msg():723: Will never receive state. Need to abort.
how can I solve this problem ?
Best regards
Ali
I had the same issue, with another more message:
WSREP_SST: [ERROR] xtrabackup process ended without creating '/path/to/mysql/xtrabackup_galera_info' (20150108 13:09:23.012)
The issue was that I was using xtrabackup and not xtrabackup-v2 in my /etc/mysql/my.cnf:
wsrep_sst_method = xtrabackup
Setting it to:
wsrep_sst_method = xtrabackup-v2
fixed the issue.
Some recommend not to set this variable.
I had the same issue. The fix is simple, just add the following to my.cnf in the [mysqld] section innodb_log_file_size=48M
(Use whatever size log file you need, 48M is OK for small Dbs)
Then follow the instructions to restart mysql with a new log file size. The basics are (but READ the Manual!) stop mysql remove the ib_logfile* files from /var/lib/mysql restart mysql
Be sure not to be using innodb fast shutdown (read the manual!)
We just experienced this error on 1 node out of a 3-node cluster. Manually rsyncing the /var/lib/mysql from a healthy node (we stopped that node during the rsync to not change files while syncing) to the broken node was what we needed to make it go again.