Showing entries 12601 to 12610 of 44118
« 10 Newer Entries | 10 Older Entries »
Monitoring MySQL in the 21st Century

In todays computing environments, the use of cloud based technologies--both public and private--along with various popular devops automation tools, are essential to effective and efficient IT operations.

MySQL Enterprise Monitor (MEM) 3 integrates very well into these modern operations environments. To demonstrate that point, I've created two short video demos that walk you through some helpful examples.

The first one walks through how you can monitor MySQL instances in the public cloud, using Amazon RDS as an example provider:


The second one walks through some examples of how you can integrate MEM 3 with your private cloud and devops automation tools:


You can find additional video demos, and more information about MEM 3  …

[Read more]
MySQL Connect 2014, Save The Date!

We're very pleased to announce that MySQL Connect will next year take place from Monday September 29 to Thursday October 2, 2014, in San Francisco. And there will be a welcome reception on Sunday September 28 evening.

During the past 2 years we’ve run MySQL Connect during the weekend prior to Oracle OpenWorld, and we repeatedly got feedback that avoiding the weekend would be better. Well, this is now done! MySQL Connect will be fully aligned with Oracle OpenWorld. This will also enable us to have more room to schedule sessions across the 4 days and to be better integrated within Oracle OpenWorld overall, which was also part of the feedback we’ve got.

We will provide more details down the road but wanted to make sure you could save the dates and make your plans accordingly.

The Call for Papers will likely open in March, so start thinking about your topics now in order to submit them swiftly when it …

[Read more]
Seconds_Behind_Master fluctuating wildly? Check for events caught in a loop

Recently I was working with a customer where we noticed that Seconds_Behind_Master fluctuating from an expected value of 0 seconds behind to a fairly high six figure value.  The servers were configured in a master-master relationship and used 5 figure server_id values, and we had just migrated this cluster from one data centre to another by re-pointing masters.  Seeing large fluctuations in Seconds_Behind_Master can often be explained by long running queries being processed by the SQL_THREAD, however SHOW PROCESSLIST indicated that there were no long running replication events, and we had no other indication that the server was lagging  due to resource constraints — CPU, disk, and memory were under-utilized.

We then moved our investigation to manual review of the binary log where events appeared normal (5 digit server_id values) until every once in a while we would see a rash of server_id 21 events.. Wait, what?  I …

[Read more]
How to get the MySQL Community Team at Your Local Event

December is half way through the budget year for the MySQL Community Team and we carefully meter out our small budget to make the most effective use from it. But December is also the time when existing conferences start the tasks of preparing for next year and brave souls who have never been part of a conference start out from scratch to create a new show. The MySQL Community Team would love to help with your conference.

First we need to know about your conference. We do scour the announcement sites like callingallpapers.com and lanyrd.com but your show may have been inadvertently skipped. Tell us the dates, the location, how many people you are expecting, your target audience (PHP Developers, DBAs, left handed actuaries from the Pacific Rim), and what makes your show unique (at least for your area). We spread the word about shows to …

[Read more]
Quick and dirty concurrent operations from the shell

Let’s say that you want to measure something in your database, and for that you need several operations to happen in parallel. If you have a capable programming language at your disposal (Perl, Python, Ruby, PHP, or Java would fit the bill) you can code a test that sends several transactions in parallel.

But if all you have is the shell and the mysql client, things can be trickier. Today I needed such a parallel result, and I only had mysql and bash to accomplish the task.

In the shell, it’s easy to run a loop:

for N in $(seq 1 10)
do
mysql -h host1 -e "insert into sometable values($N)"
done

But this does run queries sequentially, and each session will open and close before the next one starts. Therefore there is no concurrency at all.
Then I thought that the method for parallel execution in the shell is to run things in the background, and then collect the results. …

[Read more]
One more InnoDB gap lock to avoid

While troubleshooting deadlocks for a customer, I came around an interesting situation involving InnoDB gap locks. For a non-INSERT write operation where the WHERE clause does not match any row, I expected there should’ve been no locks to be held by the transaction, but I was wrong. Let’s take a look at this table and and example UPDATE.

mysql> SHOW CREATE TABLE preferences \G
*************************** 1. row ***************************
       Table: preferences
Create Table: CREATE TABLE `preferences` (
  `numericId` int(10) unsigned NOT NULL,
  `receiveNotifications` tinyint(1) DEFAULT NULL,
  PRIMARY KEY (`numericId`)
) ENGINE=InnoDB DEFAULT CHARSET=latin1
1 row in set (0.00 sec)
mysql> BEGIN;
Query OK, 0 rows affected (0.00 sec)
mysql> SELECT COUNT(*) FROM preferences;
+----------+
| COUNT(*) |
+----------+
|        0 |
+----------+
1 row in set …
[Read more]
MySQL 5.7.3: Deep dive into 1mil QPS with InnoDB Memcached

As you probably already know, in MySQL 5.7.3 release, InnoDB Memcached reached a record of over 1 million QPS on read only workload. The overview of the benchmark and testing results can be seen in an earlier blog by Dimitri. In this blog, I will spend sometime on the detail changes we have made to achieve this record number.

First thanks to Facebook’s Yoshinori with his bug#70172 that brought our attention to this single commit read only load test. We have been focussing on operation with large batch size. This bug prompted us to do a series of optimization on single commit read only queries and these optimizations eliminated almost all major bottlenecks from the InnoDB Memcached plugin itself.

If you are just …

[Read more]
MySQL 5.7.3: Deep dive into 1mil QPS with InnoDB & Memcached

As you probably already know, in MySQL 5.7.3 release, InnoDB Memcached reached a record of over 1 million QPS on a read only load. The overview of the benchmark and testing results can be seen in an earlier blog by Dimitri. In this blog, I will spend sometime on the detail changes we have made to achieve this number.

First thanks to Facebook's Yoshinori with his bug#70172 that brought our attention to this single commit read only load test. We have been focussing on operation with large batch size. This bug prompted us to do a series of optimization on single commit read only queries and these optimizations eliminate almost all major bottlenecks from the InnoDB Memcached plugin itself.


If you are just …

[Read more]
Percona and Continuent present: Multi-Data Center MySQL with Continuent Tungsten

Many users are challenged with how to setup multi-master, multi-site MySQL clusters. How to do you go from a single database server to a scalable cluster? Or move from a brittle MySQL replication system to a transparent, manageable database cluster? In this joint webinar with Continuent, Percona examines how Continuent Tungsten addresses these problems. 

First, we discuss the read-write

MariaDB world record price per row 0.0000005$ on a single DELL R710

Don't look at an industry benchmark here, it's a real client story.

200 Billion records in a month and it should be transactional but not durable.

For regular workload we use LOAD DATA INFILE into partitioned InnoDB, but here we have estimated 15TB of RAID storage. This is a lot of disks and it can't no more stay inside a single server internal storage.

MariaDB 5.5 come with TokuDB storage engine for compression, but is it possible in the time frame impose by the workload?

We start benchmarking 380G of raw input data files,  6 Billion rows.

First let's check the compression with the dataset.


Great job my TokuDB 1/5, without tuning a single parameter other than durability! …

[Read more]
Showing entries 12601 to 12610 of 44118
« 10 Newer Entries | 10 Older Entries »