Showing entries 1 to 10 of 82
10 Older Entries »
Displaying posts with tag: Clustering (reset)
Geo-Scale MySQL in AWS Webinar

Learn how to build a global, multi-region MySQL cloud back-end capable of serving hundreds of millions of online multiplayer game accounts. Find out how Riot Games serves a globally distributed audience with low-latency, fast response times for read traffic, rapid-failover automated high availability, simple administration, system visibility, and stability.

 

Watch the Geo-Scale MySQL in AWS Webinar.

Watch the AWS Re:Invent Talk, here

Read the Riot Games blog post, here

ProxySQL Native Support for Percona XtraDB Cluster (PXC)

ProxySQL in its versions up to 1.x did not natively support Percona XtraDB Cluster (PXC). Instead, it relied on the flexibility offered by the scheduler. This approach allowed users to implement their own preferred way to manage the ProxySQL behaviour in relation to the Galera events.

From version 2.0 we can use native ProxySQL support for PXC.. The mechanism to activate native support is very similar to the one already in place for group replication.

In brief it is based on the table [runtime_]mysql_galera_hostgroups and the information needed is mostly the same:

  • writer_hostgroup: the hostgroup ID that refers to the WRITER
[Read more]
Tungsten Clustering 5.3.6 and Tungsten Replicator 5.3.6 Released

Continuent is pleased to announce that Tungsten Clustering 5.3.6 and Tungsten Replicator 5.3.6 are now available!

Our 5.3.6 release fixes a number of bugs and has been released to improve stability.

Highlights common to both products:

  • Now, instead of searching for a master with appropriate role (i.e. matching the slave preferred role) until timeout is reached, the replicator will loop twice before accepting connection to any host no matter what its role is. (CT-712)
  • Changing the state machine so that RESTORING is not a substate of OFFLINE:NORMAL, but instead of OFFLINE so that a transition from OFFLINE:NORMAL:RESTORING to ONLINE is not possible any longer. Now it will not be possible to transition from
    OFFLINE:RESTORING to ONLINE (CT-797)
  • When an heartbeat is inserted, it will now use the JVM timezone instead of hardcoded UTC. (CT-803)
  • Now skipping files that are not …
[Read more]
Tungsten Clustering 6.0.4 and Tungsten Replicator 6.0.4 Released

Continuent is pleased to announce that Tungsten Clustering 6.0.4 and Tungsten Replicator 6.0.4 are now available!

Our v6.0.4 release fixes a number of bugs and introduces some new features, with improvements across the board in a variety of different components.

Some of the key improvements include:

  • When installing from an RPM, the installation would automatically restart the Connector during the installation. This behavior can now be controlled by setting the parameter --no-connectors within the configuration to prevent tpm from restarting the Connectors during the automated update processing.
  • Cross-site Replicators within a composite multimaster deployment can now be configured to point to slaves by default, and to prefer slaves over masters during operation using the new --policy-relay-from-slave=true option
  • You can now enable an option so …
[Read more]
We’re hiring: Database Clustering Technical Sales & Support Engineer

We are growing!

Consequently, Continuent needs additional talented staff.

We are currently looking for a person who would be a good fit for our “Database Clustering Technical Sales & Support Engineer” position, in the US Pacific timezone.

If you know someone who would be a good fit for this role and our company’s culture, please send them our way.

Continuent Clustering 5.3.3/5.3.4 and Tungsten Replicator 5.3.3/5.3.4 Released

Continuent is pleased to announce that Continuent Clustering 5.3.4 and Tungsten Replicator 5.3.4 are now available!

Release 5.3.4 was released shortly after 5.3.3 due to a specific bug in our reporting tool tpm diag. All of the major changes except this one fix are in the 5.3.3 release.

Our 5.3.3/4 release fixes a number of bugs and has been released to improve stability in certain parts of the products.

Highlights common to both products:

  • Fixed an issue with LOAD DATA INFILE
  • Replicator now outputs the filename of the file when using thl to show events
  • tpm diag has been improved the way we extract MySQL and support Net:SSH options

Highlights in the clustering product:

  • Tungsten Manager stability has been improved by identifying some memory leaks.
  • Tungsten Connector has fixed a number of bugs relating to bridge mode, …
[Read more]
Watch the webinar replay: How Bluefin ensures 24/7 operation and application availability with Continuent Clustering

Watch the relay of this webinar and learn how Bluefin Payment Systems provides 24/7/365 operation and application availability for their PayConex payment gateway and Decryptx decryption-as-a-service, essential to point-of-sale (POS) solutions in retail, mobile, call centers and kiosks.

We discuss why Bluefin uses Continuent Clustering, and how Bluefin runs two co-located data centers with multimaster replication between each cluster in each data center, with full failover within the cluster and between clusters, handling 350 million records each month.

Watch this webinar replay at https://youtu.be/crlgsflH7Gw

Multi-Cloud SaaS Applications: Speed + Availability = Success!

In this blog post, we talk about how to run applications across multiple clouds (i.e. AWS, Google Cloud, Microsoft Azure) using Continuent Clustering. You want your business-critical applications to withstand node, datacenter, availability-zone or regional failures. For SaaS apps, you also want to bring data close to your application users for faster response times and a better user experience. With cross-cloud capability, Continuent also helps avoid lock-in to any particular cloud provider.

The key to success for the database layer is to be available and respond rapidly.

From both a business and operational perspective, spreading the application across cloud environments from different vendors provides significant protection against vendor-specific outages and vendor lock-in. Running on multiple platforms provides greater …

[Read more]
Global Read-Scaling using Continuent Clustering

Did you know that Continuent Clustering supports having clusters at multiple sites world-wide with either active-active or active-passive replication meshing them together?

Not only that, but we support a flexible hybrid model that allows for a blended architecture using any combination of node types. So mix-and-match your highly available database layer on bare metal, Amazon Web Services (AWS), Azure, Google Cloud, VMware, etc.

In this article we will discuss using the Active/Passive model to scale reads worldwide.

The model is simple: select one site as the Primary where all writes will happen. The rest of the sites will pull events as quickly as possible over the WAN and make the data available to all local clients. This means your application gets the best of both worlds:

  • Simple deployment with no application changes needed. All writes …
[Read more]
Mastering Continuent Clustering Series: Converting a standalone cluster to a Composite Primary/DR topology using INI configuration

In this blog post, we demonstrate how to convert a single standalone cluster into a Composite Primary/DR topology running in two data centers.

Our example starting cluster has 5 nodes (1 master and 4 slaves) and uses service name alpha. Our target cluster will have 6 nodes (3 per cluster) in 2 member clusters alpha_east and alpha_west in composite service alpha.

This means that we will reuse the existing service name alpha as the name of the new composite service, and create two new service names, one for each cluster (alpha_east and alpha_west).

Below is an INI file extract example for our starting standalone cluster with 5 nodes:

[defaults]
...

[alpha]
connectors=db1,db2,db3,db4,db5
master=db1
members=db1,db2,db3,db4,db5
topology=clustered

To convert the above configuration to a Composite Primary/DR:

  1. First you must stop all services on all existing nodes:
    shell> stopall …
[Read more]
Showing entries 1 to 10 of 82
10 Older Entries »