While crawling through the wikileaks online volumes of information I realized how difficult it is to actually make any meaning out of all those millions of cables, items.
Can anyone here point me to some tools/software that can help me tag, categorize, group, search, translate these articles easily ?
Which softwares do you use anyway ?
If you are needing to group, search, and translate datasets from the Wikileaks database then I would suggest you take a look at the DataDrivenJournalism.net website, which offers a ton of different articles and tools that may be of use to you. In terms of searching through all of the cables that Wikileaks has to offer I recently came across a project called Cable Search, which is still in the beginning phases, but that you might find useful. It is a search interface that is based on 251287 cables, including ones that have been retracted from the original source. When searching, you can prepend the modifier "-" to signal words you don't want to appear in a document. Or use quotation marks for more than one word or intitle: for searching in titles. You can find more information here.
I hope that helps you out a little bit. Good luck in your search :)
That's really wonderful and relevant piece of info Hanna. Thank you for the same. I hope few other people will be able to contribute more info on this.
This info might be old, but a few years ago I looked at a tool called OpenCalias, at http://www.opencalais.com/ for semantic analysis. I'm not sure how easy it is to use, in theory it sounded pretty neat and would help get meaning out of a bunch of text.