For my current environment, I am running 2 separate Linux boxes each running MySQL. Each server should have the exact same db (one of the servers should host the db and the other one points to it). I know this could be possible using "federated tables" but the application that the database supports has hundreds of tables. It would be very recursive to point each table to the other db.
Does anyone have any suggestions how this could be possible? I have looked all over and have not found anything quite like this scenario.
Related
Good morning!
I have 2 tables that are stored on different databases - different mySQL hostnames.
How to I join them?
Is there a way to open a connection to a database, being connected to another, using commands at the MYSQL Workbench GUI interface (not command prompt) and then query it?
Many thanks,
I am new to Database programming and there is some general questions that I would like to ask. I created the schema in my localhost using mySQL and linked to eclipse. There are some problems that I do not know how to approach.
One of my friends would like to help to develop at his personal machine, but he could not link to my database server. So one way is to copy the schema to his mySQL and change the connection string, are there any better ways?
If I would to release the project and run it on different machines, will it affect the databases operation since the schema resides in my local server.
Is there ways to just like attach the database inside the project since it is a local database and I am not accessing it from any other programs.
Sorry if my questions sound very stupid. I am really new.
Your approach of running local test MySQL instance for every developer sounds fine.
However, if your application never needs any data shared (essentially database is always local as you stated), then you should ask yourself if MySQL really an appropriate solution for your application. I think SQLite is a better choice in that situation - it is likely to work faster than MySQL for local access, and it has essentially zero setup - no database daemon to run or anything like that which greatly simplifies application install.
Granted, SQLite has its limits, but I often use SQLite for my projects, and only if they grow large, I might migrate them to MySQL or PostgreSQL if task requires it.
Typical signs that SQLite may not cut it:
Many clients (10+) need to access database for writing
Total database size is very large - more than 10GB total
This is a challenging question, but will still ask it. I know nothing is impossible. Impossible simply means not discovered yet.
What I would like to do is connect all my FTPs and SQL DBs from my 10 servers and pool them all together to combine the resources and create a share resource environment. Therefore creating a ftp and sql cluster. I do not want to purchase any software. Just wondering if there is some method maybe in PHP. I dont want to clone my sites db to the 10 servers, but wondering is there a way to bridge 10 sql dbs together over 10 servers to create a super computer like environment?
MySQL Cluster is an Open-Source solution.
Distributed FTP Daemon seems to be a promising project.
I have two servers. One is located in our office, and its MySQL base contains our offers, our clients etc.
The second server is located at our hosting provider's datacenter. It uses the same database structure and the same offers, and I use it for our website.
I was synchronizing these two servers manually, by sending json from one server to another each few hours, but now I need a real-time synchronization.
Which way should I use?
Master-slave replication from company server to website server. The problem is, that our slave website database has its own changeable tables too. For example, orders, user sessions, viewcounts and so on. And I need to send somehow that tables to master server at office.
To use only one database for both servers. Problem is, that there could be up to 100 queries each pageview, and I think that running each query through internet could be quiet slow.
We cannot use only one server for all tasks because we are unable to provide a stable low-latency internet connection in our office. So when internet is down, our site or our CRM system would be down to.
May be there is a third and best way to do this?
You can try Data Comparison tool in dbForge Studio for MySQL. It will connect to two different MySQL servers (using simple connection, SSL, SSH or HTTP tunnel), compare them and show differences; then it will offer to run synchronization script or view/save it.
There is also stand-alone dbForge Data Compare tool.
I'm developing an app which will have a central database users can add entries to. The database will have to be on a server somewhere but I want the users to be able to add entries offline. The app will sync to the main db when connection is available. So, I supose I need 2 databases - the main one sitting on a server (preferably linux) and a small one on each client machine to use as a buffer when offline. The app will be coded in c# for windows. I'm having trouble deciding what databases to use and whether I can leverage any replication technology to make this easier. Also, I don't want to pay for anything ;) So I guess my questions are...
Will I have any trouble writing code in ADO.NET to move data from something like SQL Compact Edition to MySQL?
Are there any replication solutions which will move stuff from local to main database for me
I've recently discovered IBM's db2 expressC but I'm not sure if it's serverless as well as server installed. Does anyone know?
Firebird can be server or serverless. Can I replicate between them. Is the server mode capable of heavy use?
Firebird can be server or serverless.
Can I replicate between them.
Yes.
Is the server mode capable of heavy
use?
Define 'heavy use'. I've had production systems with 200 simultaneous users pumping 20 transactions/minute each on databases in the 10-20GB range. I'm sure there are many larger deployments out there.
Also, what you describe seem like the 'briefcase model'. You should look into it if you haven't already done so. Maybe the solution is not replication at the database level, but rather a smarter fat client.
Just answering two of your questions; I don't know about DB2 or Firebird.
Will I have any trouble writing code in ADO.NET to move data from something like SQL Compact Edition to MySQL?
That should be very trivial; install MySQL Connector/NET and you're good to go.
Are there any replication solutions which will move stuff from local to main database for me
SQL Server replication is made for this, but I don't suppose it would work with MySQL.