Many user space factors can affect recording performance. Such as the:
- Dozens of settings in each database server configuration.
- Table structure and settings.
- Connection Settings.
- Request Settings.
Do you swallow warnings or exceptions? It is assumed that a MySQL sample at face value will cause a duplicate key error. This can fail without doing anything at all. What Cassandra could have done in the same case is not familiar to me.
My limited experience with Cassandra tells me one thing about inserts, while the performance of everything else decreases as the data grows, the inserts seem to support the same speed. How quickly it compares with MySQL, I have not tested.
This may not be so fast, because the inserts are fast, but rather not slow. If you want a more meaningful test, you need to enable parallelism and more scripting options, such as large datasets, packages of different sizes, etc. More sophisticated tests can check the latency for data availability after insertion and read speed over time.
It wonโt surprise me if Cassandra is the first port of call to insert data into the queue or just add it. This is customizable if you look at the level of consistency. MySQL also allows you to balance performance and reliability / availability, although each will have its own options for what they allow and do not allow.
In addition, if you do not enter the internal organs, it can be difficult to say why one works better than the other.
I did some tests of the use case that I had for Cassandra some time ago. For the test, he would first insert tens of thousands of lines. I had to disconnect the script for a few seconds, because otherwise the queries executed after the fact would not see the data, and the results would be incompatible between the implementations that I tested.
If you really need quick inserts, add the file to ramdisk.
jgmjgm
source share