Most modern websites and applications ground on data collection, storage, and analysis. Databases take an active part in building the entire Web environment. That’s why it is crucial to ensure correct data retrieval from databases and appropriate ways of data manipulation. To modify your data properly, you will need to execute SQL queries. The current […]
In the second article of the series uncovering how to get started with MySQL, we talk about the ways of connecting to MySQL Server. You can connect to MySQL Server using MySQL Client, dbForge Studio for MySQL, and MySQL Workbench. In this article, we consider each method in detail. Connecting to MySQL Using the MySQL Command-Line ClientConnecting to […]
We are glad to inform our MySQL users, that a new version of our Visual Studio plugin for MySQL database development, dbForge Fusion for MySQL, v.6.6, has been just rolled out. Connectivity The new version features a massive update of connectivity options implemented to ensure that you can easily connect and work with the latest […]
The post New Connectivity Options in dbForge Fusion for MySQL, v6.6 appeared first on Devart Blog.
Learn how to safely rename a column in a MySQL database using a universal multi-featured tool – dbForge Studio for MySQL. To paraphrase Heraclitus, the only constant is change. That concerns databases too. With time, business requirements change forcing the database development process to adapt. Thus, database developers increasingly face the necessity to perform structural database […]
Logical backup’s are of great use in data migration across cloud environments and table level recoveries. The new Mysql shell 8.0.22 ,has introduced a couple of new utilities named util.dumpTable() and util.exportTable() to export individual tables from a MySQL. Prior to 8.0.22 it is not possible to make a backup of single table using MySQL Shell.
MySQL Shell’s new table dump utility util.dumpTables() from this we can take the dump of the specific tables of the schema using this utility.it works in the same way as the instance dump utility util.dumpInstance() …
In this article, you’ll learn how to connect to Azure Database for MySQL and discover three scenarios of database migration from MySQL to Azure using dbForge Studio for MySQL. The article demonstrates how to connect to Azure Database for MySQL Server via dbForge Studio for MySQL. It also explores three common approaches of using the […]
The post Azure Database for MySQL: How to Connect and Migrate Databases with dbForge Studio for MySQL appeared first on Devart Blog.
As a Database Engineer in Mydbops. We tend to solve multiple complex problems for our esteemed customers. To control the System resources and scale up /down based on needed we are evaluating Dockers and Kubernetes.
Docker is a set of platform as a service products that uses OS-level virtualization to deliver software in packages called Containers. Containers are isolated from one another and bundle their own software, libraries and configuration files; they can communicate with each other through well-defined channels.It’s more lightweight than standard Containers and boots up in seconds.
Docker also is easy to use when you need a simple, single instance. What is great about Docker though is that it allows configuring multiple versions of MySQL.
Docker can be installed with yum repository or apt-get repository based on your linux …[Read more]
In my previous Blog i have explained how we can integrate MySQL tools ( Percona Tool kit ) with systemd service, At Mydbops works with multiple clients ,we get various requirements from customers a few may be a complex ones ,one of the requirement was to setup a binlog server and copy the Production binlogs instantly .
To satisfy customer request i have made a simple shell script and integrated in systemd service which does a similar job as of Maxscale Binlog Server ( Pervious blog )
In this blog i will explain how to setup a simple binlog server.
Pre-requites for setting up the Binlog server.
- A standalone remote server with required disk.
- The mysqlclient to be …
In my day to day tasks as a DB Engineer at Mydbops we have been using multiple MySQL tools for multiple use cases to ensure an optimal performance and availability for servers managed by our Remote DBA Team.
Some of the problems we face when we demonise the process or use screen for running processes.
- The daemon process gets killed when the server reboot happens. …
One problem I have experienced during logical exporting and importing databases is the tools don’t output any progress indicator (or anything at all really) and using the -v (verbose) switch outputs every single MySQL command being run from the import file. As, importing and exporting of databases is one of the most frequent activity being performed , I Just felt the need of some tool that could help me check the approximate progress of import and export of databases going on.
So, I searched and fortunately could find a tool named Pipeviewer (PV) which allows a user to see the progress of data through a pipeline, by giving information such as time elapsed, percentage completed (with progress bar), current throughput rate, total data transferred, and ETA .
On Centos and RHEL :
[root@vm1 vagrant]# yum install pv
[root@vm1 …[Read more]