Closed. This question is opinion-based. It is not currently accepting answers.
Want to improve this question? Update the question so it can be answered with facts and citations by editing this post.
Closed 7 years ago.
Improve this question
Please explain why redisdb is used in some cases. which are theese cases which redis db is used? Is redis good for managing big database or something?
Redis is an in memory data store. One popular use for Redis is as a state server, saving session information. Redis is supposed to be fast because it works with memory. So, if you used redis to hold session information, you will only hit the
machine's memory. If you used a database like mysql for session state, there is a high likelihood that disk I/O will be required.
Redis doesn't have a strong persistence mechanism and as such should not be used as an application's database.
Related
Closed. This question is opinion-based. It is not currently accepting answers.
Want to improve this question? Update the question so it can be answered with facts and citations by editing this post.
Closed 1 year ago.
Improve this question
both innodb_buffer_pool and redis they are all suppert memery. and they are all support LRU. so why should i use both redis and mysql at same time?, what's the point of using redis in my case.
my application is coding use golang(game server). It provides http api. but applications are becoming more complex. it has 100 sql query per request (20% insert/update, 80% select).because the game hasn't been released yet so
the data set is small, about 400000 rows, stress testing in progress,i want to improving it. i increased innodb_buffer_pool_size,the performance improved but it didn't meet my expectations. so i want using redis.
thanks for you answer
If the data is changing, the buffer_pool can give you up-to-the-second data; no external caching mechanism can.
It is usually folly to have a cache in front of another cache.
How big is the dataset? What do the queries look like? (That is, please provide more info so we can come closer to answering your question.)
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 4 years ago.
Improve this question
i work on a project with node.js and a mysql database.
i have a connection between them with the npm modul mysql.
Now my question:
is it better to send a SQL command and sort the data in node.js or multiple detailed sql commands?
what is faster / more performant?
Thanks.
Without knowing the exact SQL queries, I would say that database operations are faster compared to your own implementation. Many smart people have worked to assure performance, accuracy, atomicity, concurrency etc. of the Mysql engine.
Even if you can gain marginal improvements in some aspect with your own code, it is unlikely that you will be able to justify the investment.
Closed. This question is opinion-based. It is not currently accepting answers.
Want to improve this question? Update the question so it can be answered with facts and citations by editing this post.
Closed 6 years ago.
Improve this question
Could we use Orion internal MongoDB data base to store other information?
For example using Cygnus to persist data in it?
Or should it be dedicated to Orion?
Thanks and best regards.
Yes, you can.
From a functional point of view, as long as you don't use the same DB that Orion uses (by default orion) or (in the case of using -multiservice) the DB used by individual tenants (by default, the ones starting with orion- prefix) there shouldn't be any conflict.
From a non functional point of view, you should check that the MongoDB instance (or cluster) is sized properly to support the aggregated load combining all the applications using it (Orion and others).
Closed. This question is opinion-based. It is not currently accepting answers.
Want to improve this question? Update the question so it can be answered with facts and citations by editing this post.
Closed 9 years ago.
Improve this question
I want to make a simple script that periodically performs SNMP GET and stores the output to MySQL database. It will be running on Windows.
What would be the least painful choice of programming language to write it in? Powershell, Java, …?
Python has SNMP (i.e. NET-SNMP) bindings. It will handle MySQL database as well. In that way you may use those APIs to create your own scanning application with SNMP communication channel and MySQL database storage of collected data. Of course there are tons of other solutions/combinations possible... however I recommend this one, because in such a simple case you don't have to engage heavy JVM.
Python has SNMP(Net-SNMP) binding and it's better that you can use RRD database to store output as it consists time interval itself.Python has RRD api also and you can handle easily
Closed. This question is opinion-based. It is not currently accepting answers.
Want to improve this question? Update the question so it can be answered with facts and citations by editing this post.
Closed 6 years ago.
Improve this question
I'm thinking of using DB Connection Pool in J2EE.
Database: MySQL
Servlets: http://www.webdevelopersjournal.com/columns/connection_pool.html
Listeners: http://onjava.com/onjava/2006/04/19/database-connection-pooling-with-tomcat.html
now which is a better option & why? Also, any working source code will help me save time.
I really wouldn't go for a homegrown connection pool as in your 1st link. There are a lot of factors you need to take account with and those are surely not covered by the example given in that article. Forget about it, don't homegrow a connection pool, it's a way too important core piece of your webapplication.
Go for a container managed connection pool with an existing connection pooling implementation like mentioned in your 2nd link. You have nowadays the choice of under each DBCP and C3P0. DBCP is already inbuilt in Tomcat, but it's singlethreaded and thus not really performant when talking about a busy-visited website. Alternatively (and Tomcat-specific), there's also a tomcat-jdbc which removes this DBCP limitation, but as it's fairly new, I'd rather go for C3P0 as it's already been thoroughly developed and maintained for ages and is nowadays been very robust and performant.