Large-scale dataset processing

From online discussion groups and blogs, I have seen that many of the interview questions are related to processing a large-scale array. I wonder if there is a systematic approach to analyzing this type of question? Or, more specifically, is there any data structure or algorithms that can be used to solve this problem? Any suggestions really appreciated.

+5
source share
4 answers

Large-scale datasets belong to several categories that I have seen, each of which presents various problems for you.

  • , . :
    • ,
    • , , ( , !)
    • .
  • , , - . - , , .
  • , . , , , :-) .
    • , . . / , CouchDB Hadoop, .
  • , . ( ) ( , ). , , , . :
    • ( )
    • ( , )

, , , :

  • , . , .
  • , . , ?
+8

"" - , "" ( , SMP - ).

+1

. , , . , , , B + Trees.

+1

, , . , .

- . , , , .

- - , .

- , .

The DBMS can greatly simplify data access, but it adds some system overhead.

0
source

All Articles