Version End of Life. 31 July 2020
This is a bugfix release.
tpm now outputs a note and recommendation for performing backups of your cluster when installation has been completed.
tpm would fail during installation if the current directory was not writable by the current user.
When performing a tpm update in a cluster with an active witness, the host with the witness will not be restarted correctly resulting in the witness being down on that host.
Issues: CT-673, CT-691
During a relay failover within a Composite Active/Active or Multi-Site/Active-Active deployment, if the communications had also failed between sites when the failover occured the manager would be unable to determine the correct Primary of the remote site.
In a deployment, single cluster or Composite Active/Active where there is either the potential for high-latency across sites, or high latency within a site due to high loads on the connectors, the manager could mis-identify this high latency as a failure. This would trigger a quorum validation. These would be reported as network hangs, even though the result of the quorum check would be valid.
To address this, the processing of router notifications processed by the connector and all other operations have been separated. This reduces the change of a heartbeat gap between hosts and therefore the connectors are available to the managers even under high loads or latency.
Release 5.3.3 is a bug fix release.
The output from thl list now includes the name of the file for the correspnding THL event. For example:SEQ# = 0 / FRAG# = 0 (last frag) - FILE = thl.data.0000000001 - TIME = 2018-08-29 12:40:57.0 - EPOCH# = 0 - EVENTID = mysql-bin.000050:0000000000000508;-1 - SOURCEID = demo-c11 - METADATA = [mysql_server_id=5;dbms_type=mysql;tz_aware=true;is_metadata=true;service=alpha;shard=tungsten_alpha;heartbeat=MASTER_ONLINE] - TYPE = com.continuent.tungsten.replicator.event.ReplDBMSEvent - OPTIONS = [foreign_key_checks = 1, unique_checks = 1, time_zone = '+00:00', ##charset = US-ASCII]