Showing entries 1 to 10 of 233
10 Older Entries »
Displaying posts with tag: monitoring (reset)
Monitoring NDBCluster Copying Alter Progress


MySQL NDB Cluster has great support for online (inplace) schema changes, but it is still sometimes necessary to perform an offline (copying) ALTER TABLE. These are relatively expensive to make as the entire table is copied into a new table which eventually replace the old table.

One example where a copying ALTER TABLE is required is when upgrading from MySQL NDB Cluster 7.2 or earlier to MySQL NDB Cluster 7.3 or later. The format used for temporal columns changed between these version (corresponding to MySQL Server 5.5 to 5.6). In order to take advantage of the new temporal format, a table rebuild is required.

Note: Support for the old temporal format has been removed in MySQL 8.0. So, you must …

[Read more]
What Does I/O Latencies and Bytes Mean in the Performance and sys Schemas?


The Performance Schema and sys schema are great for investigating what is going on in MySQL including investigating performance issues. In my work in MySQL Support, I have a several times heard questions whether a peak in the InnoDB Data File I/O – Latency graph in MySQL Enterprise Monitor (MEM) or some values from the corresponding tables and view in the Performance Schema and sys schema are cause for concern. This blog will discuss what these observations means and how to use them.

The Tables and Views Involved

[Read more]
How NOT to Monitor Your Database

Do you have experience putting out backend database fires? What were some things you wished you had done differently? Proactive database monitoring is more cost efficient, manageable, and sanity-saving than reactive monitoring. We reviewed some of the most common mistakes - too many log messages, metric “melting pots,” retroactive changes, incomplete visibility, undefined KPIs - and put together an action plan on how to prevent them. From our experience, we've listed out the top 5 biggest (and preventable!) database monitoring pitfalls.

Log Levels

There never seem to be enough logging levels to capture the desired granularity and relevance of a log message accurately. Is it INFO, TRACE, or DEBUG? What if it’s DEBUG but it’s for a condition we should WARN about? Is there really a linear hierarchy here? If you’re like most people, you’ve seen …

[Read more]
Monitor Critical Databases Confidently with the Sensitive Data Vault

Building extremely deep monitoring as a SaaS product has a drawback: we capture too much data for some customers’ compliance requirements. As a result, some companies have been unable to deploy us, or have had to redact data before sending it to our cloud platform. To address this, we built the Sensitive Data Vault, a highly secure, completely on-premises storage module for the most critically private data that must never leave the customer’s firewall.


What is it?

The VividCortex Sensitive Data Vault is a new component of the overall VividCortex solution that you deploy inside your firewall. It ensures that the data never leaves your servers and never enters the VividCortex cloud environment. It consists of:

  • a Go service that the VividCortex collector agent communicates with
  • a customer-maintained MySQL or PostgreSQL database that the Go application uses
[Read more]
How SendGrid Ships Better Code Faster with VividCortex

VividCortex CEO Baron Schwartz and SendGrid DBA Silvia Botros teamed up to discuss how performance monitoring leads to better, faster code deployment. This 30-minute webinar covers: 

  • How developers can deploy code faster and more safely.
  • A close-up view of a health- and monitoring-focused work environment.
  • How database monitoring fits into a culture of DevOps and lean, agile development. 

"Now, with VividCortex, whenever we have an issue that's impacting the mail processing throughput, we can very quickly go and answer, "What was running at that time? What was the most expensive query? What was taking up all the load?" Within an hour, we can typically figure out exactly which JIRA WAR to go to."  —Silvia Botros, Lead MySQL DBA

Take a look:



Reading Amazon RDS MySQL/Aurora log file from terminal.


At Mydbops we support a good number of clients on AWS cloud (Aurora and RDS).

Amazon Relational Database Service (RDS) is providing the cloud based database service. It is the cost-efficient, resize able & ease to manage. As in any other DBaaS, If you need to analyse the log files (Error log / Slow log), you need to login the console and manually download the files.

Logging into the console seems simple, But this is a bit complex operation when it comes to incorporate that in a day to day operation and automation. In this blog i would like to share my experience in making this into a straightforward process for downloading the log files directly from command line without console GUI.


Following tools are to be installed for this operation.

[Read more]
Presentation: Ansible is our Wishbone

This presentation was made at LSPE event in Bangalore (India) held at Walmart labs on 10-03-2018. This presentation focuses how we have harnessed the power of Ansible at Mydbops.



Just Released: New Features!

I’m pleased to announce that VividCortex has just released a number of new features since our last product update. If you have any questions regarding any of the updates below, contact us!

Short URLs

You can now share links with your coworkers and collaborate in VividCortex using URLs short enough to be copied and pasted into emails and chat apps, letting you get work done faster with less hassle. Just click the Link icon in the top right of the screen and copy the URL:

When anyone with appropriate permissions clicks the link, they’ll be taken directly to the URL you were looking at when you grabbed the link. Click “Freeze Time Range” to let your coworkers see the time frame you were viewing, no matter when they click the link. For more details, take a look at our …

[Read more]
Announcement: p99 Percentile Metrics

I’m pleased to announce that VividCortex now offers 99th percentile metrics to help understand latency outliers in a query workload. These metrics provide visibility beyond the average latency, help identify the worst outliers, and improve focus on the customer experience. They are offered for all of the databases we currently support when using On-Host monitoring.

What You'll See

Latency percentile metrics are one of our most popular feature requests, so we know this will make a lot of you very happy! We actually started collecting these metrics some time ago; you’ll have p99 latency metrics for the last couple of months if you look back at your historical metrics. These metrics are captured globally for an environment, per-query, per-database, per-user, per-verb, per-caller and per-custom-query-tag.

Why Did We Choose to Implement This? 

It is extremely useful for a lot of reasons. Averages can be …

[Read more]
Percona Monitoring and Management 1.8.0 Is Now Available

The latest release of Percona Monitoring and Management (PMM) 1.8.0 is available as of February 27, 2018.

Percona Monitoring and Management is a free and open-source platform for managing and monitoring MySQL and MongoDB performance. You can run PMM in your own environment for maximum security and reliability. It provides thorough time-based analysis for MySQL and MongoDB servers to ensure that your data works as efficiently as possible.

This release introduces many improvements in the user interface and optimizations of performance.

New landing page

The home page has been removed in favor of the …

[Read more]
Showing entries 1 to 10 of 233
10 Older Entries »