I am working on a web2 project that would like to have thousands of lines per day for users. for processing the size of the data that I created such a database: one .mdf and .ldf file as Minor DataBase and 1 Major DB for saving and querying the addresses of user accounts and database files.
I worked for several months for this plan, and now I can easily deal with it. I want to know if it is worth processing the sheer size of independent data? in your opinion, better performance? opening a connection of many small .mdf files or just a huge database.
after that I will parse the mdf repository on several computers.
they are all handled by C # and linq (.net4)
// Later descriptions
I built this plan and it works great. for example: opening each small mdf file takes 1 second and requests it in 0.0sec. it makes a static time for each connection, but in one database for the 50rows system it should find them, for example, 200,000 lines and takes about 4-5 seconds on my system with a simple query of choice using the main key.
for another instance, I want to get a line between 500,000 lines for linking the contents of the page and select 50 comments from the 2milmions line, and also count the number of votes for each comment, the number of views per day, week, month and total. the number of likes, the response of comments and getting more data from 2-3 other tables, these queries are heavy and take longer than a small database of subordinates.
I think that good design and processes should work easily for the system.
the only problem is that small subordinate databases with sql server files take up more than the physical size of about 3 MB per database.
source share