I have a composition of several 100 thousandth legal documents (mainly from the European Union) - laws, comments, court documents, etc. I am trying algorithmically to give them some idea.
I modeled a known relationship (temporary, this-change-this, etc.). But at the level of one document, I would like for me to have the best tools to quickly understand. I am open to ideas, but here is a more specific question:
For example: are there NLP methods for defining relevant / conflicting parts of documents as opposed to a template? Recently leaked TTIP documents are thousands of pages of data tables, but one sentence somewhere out there could destroy the industry.
I have played with google new Parsey McParface and other NLP solutions in the past, but as long as they work impressively, I'm not sure how good they are in isolation.
source share