Showing entries 81 to 90 of 312
« 10 Newer Entries | 10 Older Entries »
Displaying posts with tag: Backup (reset)
MySQL Enterprise Backup (MEB) and Oracle Storage Cloud

MEB 3.12.0 and above support cloud backup and restore using OpenStack-compatible object stores ("Swift"). This allows MySQL database users with Oracle Storage Cloud account to take backups and store them directly in the cloud and restore them from there.

The following steps illustrate how to set up and use MEB with Oracle Storage Cloud :

1) Create Oracle Storage Cloud account at https://cloud.oracle.com/storage . Once service gets activated, make a note of the following credentials that will be required in further steps :

  • Username

  • Password

  • Identity domain name

  • Service Instance Name : Customer-specified name of the service instance

[Read more]
How to install and configure AutoMySQLBackup

In this blog article we will show you how to install AutoMySQLBackup on a Linux VPS. AutoMySQLBackup is very useful utility for creating daily, weekly or monthly backups of one or more MySQL databases from one of more MySQL servers. It dumps the databases and compress them in to archives.it comes with many features such as:   Email notification of backups   Backup Compression and Encryption   Configurable backup rotation   Incremental database backups As usual, log in to your server as user root ssh root@IP and execute the following command to make sure that all services ar up to […]

MySQL-Docker operations. - Part 2: Customizing MySQL in Docker


Previous Episodes:


After seeing the basics of deploying a MySQL server in Docker, in this article we will lay the foundations to customising a node and eventually using more than one server, so that we can cover replication in the next one.
Enabling GTID: the dangerous approach.To enable GTID, you need to set five variables in the database server:

  • master-info-repository=table
  • relay-log-info-repository=table
  • enforce-gtid-consistency
  • gtid_mode=ON
  • log-bin=mysql-bin

For MySQL 5.6, you also need to set log-slave-updates, but we won't deal with such ancient versions here.
Using the method …

[Read more]
Binlog Servers for Simplifying Point in Time Recovery

A common way to implement point in time recovery capability is:

to regularly do a full backup of a database, and to save the binary logs of that database (or from its master if doing backups on a slave).

When point in time recovery is required you need to:

restore a backup, and apply the binary logs up to the point of recovery.

(Step # 2 and # b above are the ones that will be simplified

TwinDB Really Loves Backups

A week or two ago one of my former colleagues (at Percona) Jevin Real gave a talk titled Evolving Backups Strategy, Deploying pyxbackup at Percona Live 2015 in Amsterdam. I think Jervin raised some very good points about where MySQL backup solutions in general fall short. There are definitely a lot of tools and scripts out there that claim to do MySQL backups correctly, but don’t actually do it correctly. What I am more interested though is in measuring TwinDB against the points that Jervin highlighted to see if TwinDB falls short too.

Dependencies

We distribute TwinDB agent as a package that can be installed using the standard OS package management system. For example, using YUM on CentOS, RHEL and Amazon Linux, or using APT …

[Read more]
MySQL 5.7 : Playing with mysqlpump

MySQL 5.7 comes with a new backup tool, named mysqlpump, which is almost the same as mysqldump with the ability of extracting data in parallel threads.

I tried a little experiment. Using a server containing 11 databases, with a total of 300 tables and about 20 million rows (roughly ≈ 10GB,) I used both mysqldump and mysqlpump to get a backup.

mysqldump --all-databases  > dump.sql
mysqlpump --all-databases \
--add-drop-database --add-drop-table --skip-watch-progress \
--default-parallelism=10 \
--parallel-schemas=db,db1,db2 \
--parallel-schemas=db3,db4,db5 \
--parallel-schemas=db6,db7,db8 \
--parallel-schemas=db9,db10 > pump.sql

The backup with mysqldump took 3 minutes and 33 seconds. The one with mysqlpump took 2 minutes and 55 …

[Read more]
Introducing mysqlpump

Starting with MySQL 5.7.8, we are shipping a new client utility called mysqlpump that performs logical backups, producing a set of SQL statements that can be run to reproduce the original schema objects and table data. The goal of mysqlpump is to have a modern utility that is extendable and has native support for parallelization. …

mysqlpump — A Database Backup Program

The MySQL 5.7 Release Notes  for version 5.7.8 are out. Besides the new JSON data type, there is also a new tool, called mysqlpump, which offers the following features:

Parallel processing of databases, and of objects within databases, to speed up the dump process Better control over which databases and database objects (tables, views, stored programs, user accounts) to dump Dumping of user

Backing up and restoring tables named with special characters

Introduction

The names of databases and tables within MySQL are known as identifiers. In the simplest case these identifiers are just strings of certain ASCII characters (the basic Latin letters, the digits 0-9, the dollar sign and the underscore). However, if an identifier is placed in quotes, it can contain any character of the full Unicode Basic Multilingual Plane (except U+0000). We say that a character is a special character if it is permitted in a quoted identifier but not in an unquoted identifier.

MySQL Enterprise Backup (MEB) 3.12.1 introduces support for proper handling of table and database names with special characters. In MEB versions prior to 3.12.1 database and table names were represented as ASCII strings and the same string was used on the command line, internally within MEB and in filenames.  This caused MEB to fail some …

[Read more]
Deploy TwinDB agent using Chef on your MySQL fleet

We are big fans of automation which is also one of the reasons why we started TwinDB. We want to remove the pain around manual redundant jobs and automate them so that you can focus on what is important for your business.

Why Chef?

Awesome Chef

Chef is a great example of how automation can be successfully applied to infrastructure such that your infrastructure becomes easily deployable, manageable, testable and well documented. Whenever, I talk about the benefits of using Chef, I greatly highlight how writing Chef cookbooks essentially documents your infrastructure. Isn’t it great to have a well documented infrastructure, that anyone can reference to? How many times have I not heard someone tell me how they have no idea how a particular service was deployed, or what was deployed on a particular machine? Many many times, …

[Read more]
Showing entries 81 to 90 of 312
« 10 Newer Entries | 10 Older Entries »