I am surprised that no one mentioned Excel. As Brian Ripley said (see Slide 7):
Let's not feel: the most widely used piece of statistics software is Excel.
Indeed, Excel is a great tool for adding columns of numbers. Having said that, if the analysis you are doing is more complex, you should definitely use the right programming language.
Of the three obvious data management languages ββ(R, MATLAB, and Python), R has the best data management tools. See this other SO question for a more detailed comparison.
EDIT: While re-reading this, I am voicing a pretty pro-excel. I would like to expand my answer in order to maintain my reputation.
Excel causes me more problems than benefits. Its widespread use in my organization is mostly harmful. It is very difficult to keep track of where the data comes from and how your calculations work. Debugging Excel models is almost impossible. It encourages local data warehouses instead of central databases. It does not work with diff tools, and this makes reproducibility of your science difficult. From a semantic point of view, it does not separate data and what-is-done-into-data. The idea that all your variables need a location is distracting from understanding. The plotting possibilities are ridiculously terrible.
All that said, Excel is useful for a few specific things:
As a CSV viewer. Of course, R has a View function, but it's not so pretty.
Really simple data mining. Sorting, filtering, grouping columns. I find that this can be done a little faster using the point and click interface than with the code. Of course, you will have to write code later for reproducibility, but in the initial stages of Excel, itβs not bad for this.
Charts are distinctive and easily distinguishable. If you see someone giving a presentation with a graph drawn in Excel, you cannot trust the results.
What is it. Anything else is a mess.
Richie cotton
source share