Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 7 years ago.
Improve this question
I'm quite new to mysql and in database management in general....
I've to solve this scenario:
In the development stage the web site has the database in the local machine and some tables are dedicated to contain information data used by the application ,during the development the records of that tables grow and when we move to production we want to update the production server with the new data...
Can someone advise the best practice to automate the update process from the local to the production database.
Thanks in advance
The road to doing this successfully is to have each database know how far it has migrated.
You should absolute use something like Liquibase or Flyway to do it. If you have a simple database environment these two will work. Both of these will track changes in version files that the database keep track of.
If you need more complexity, like in a sharded environment, you probably need to roll your own tool for this.
You should mention different .sql files for each environment like,
development.sql,
staging.sql,
production.sql
And you need to write shellscript to execute this script while deployment process.
Also, you need to maintain one constant to get current environment.
Related
Closed. This question needs details or clarity. It is not currently accepting answers.
Want to improve this question? Add details and clarify the problem by editing this post.
Closed 7 years ago.
Improve this question
Is it possible to make an mySQL database on the web so that any person from any network can access it? For example if somebody in Europe downloads my program, it will still be able to connect to the mySQL database without problems.
Currently, I am using Connector J with a locally hosted mySQL database.
Edit: Basically I want to create an online mySQL database (not locally hosted) where anybody can download my program and the program from any internet connection can read and write the data from the server online.
Also are there any free online mySQL hosting services?
Edit:
For example, java games access data online from a database, but they are not on the same database. How can I achieve this? Are mySQL databases still the way to go?
It's difficult to understand your purpose but what I perceived from your question is maybe you are looking for some sort of cloud app service like Heroku. Take a look at their ClearDB add-on and see if that fulfills your need.
[This was supposed to be a comment but due to the lack of my Reputation points I have to submit it as an answer]
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 8 years ago.
Improve this question
I'm a newbie and never created a database. I want create a MySQL Database which should run on a Server (other people want use it too)
What are the essential requirements to create this in an easy way?
Are there any recommended free programs which help me to reach my goal ?
Thank you
1) Get an environment
A great VPS provider is Digital Ocean
Or go for shared hosting
2) Install MySQL service
Depending on your environment and OS
Shared hosting will probably have MySQL already installed and a nice GUI for you to use to set up your databases
If you choose Digital Ocean, look at their documentation: https://www.digitalocean.com/community/tutorials/how-to-install-linux-apache-mysql-php-lamp-stack-on-centos-6
3) Set up a MySQL database and users
Plenty of materials on this
This should be it's own question once you've reached this stage, should you need help
I hope that will help you get further to your goal.
Referral links are included
Closed. This question is opinion-based. It is not currently accepting answers.
Want to improve this question? Update the question so it can be answered with facts and citations by editing this post.
Closed 9 years ago.
Improve this question
I have almost a TB of data to be loaded to MYSQL database regularly once in a week.
The server used is of lower configuration and takes a long time for every upload.
Can anyone please suggest me a tool or efficient technology to handle this.
I personally found that LOAD DATA INFILE works best for me. Check it out: http://dev.mysql.com/doc/refman/5.1/en/load-data.html.
But as #duffymo said, if your server simply can't handle this, it doesn't matter how you upload data, it might not be physically possible to go faster (disks can write only this much data per second).
It's not a matter of efficiency. No software will fix this. Your problem is server and network.
1TB per week? In a single instance of MySQL on an under-powered server? With no sharding or replication? I sincerely doubt that.
But if you must continue, maybe you should look into Hadoop. Keep your data in the Hadoop file system. You won't have to move it anywhere. Use Hive for SQL and let map-reduce help with the processing.
Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 10 years ago.
Improve this question
I am administrating mediawiki for my organisation. We use it as our Intranet site. It has accumulated a huge organisational knowledge base. I have make sure that mediawiki is always up and running. Knowledge base always backed up.
Is there a way to take continuous back of mediawiki files and databases? My mediawiki is hosted on LAMPP server with Debian OS.
I am trying to find a way to automate backup process.
It depends on what you mean by "continuous". If you want a copy of the database running that is always the same as the main database, you will need to set up "replication" - see http://dev.mysql.com/doc/refman/5.1/en/replication.html for how to do that.
If you want a database backup that is relatively current, then running mysqldump every hour or so is a pretty good solution.
You'll need to backup the files separately, because they are in your file system not the database. Look at running rsync every hour or so.
Why do you want a "continuous" backup and how would you use it? Do either of these approaches answer your question?
Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 10 years ago.
Improve this question
I have a MySQL DB which should be versioned with SVN. I dont want the full DB, only the structure and selected tables. I searched the net and found some information, but nothing seems to really work in a reliable way. Any experience or hints?
Thanks :)
Check out this script which automates the process, allowing for specific selection of databases and exclusion of tables [disclaimer I am the author] - http://mysql-svn-backup.redant.com.au/
Use mysqldump to export the data you want into a file and put this into SVN. Using cron, you can automize this to run in specific timeslots
The question is: what exactly do you want to version, and why?
I propose you version an SQL file that you can import to create your database. Any tool can be used to create this SQL file (basic tool: mysqldump), which you can then save into your SVN repository. You will be able to track new tables being created by comparing revisions of SQL files.
You can automate this process by adding a CRON job to automatically dump and commit the file every 2 hours.