Natural Language Processing is a pursuit perched between computer science and linguistics. Its primary focus is using computers to analyze natural language. Basically, computers are used to process text to determine what people meant when they said (or wrote) something.
We have architected and deployed systems (or software) that are used to discover entities, sentiment, authors, or veracity (truth). If you have needs or interest in this area, we are happy to discuss these topics at length.
Publicly available data sources we have used include Wikipedia, USENET news postings, NIST's Reuters articles and Project Gutenberg texts.
Generally, we leverage the use of the Resource Description Framework to represent metadata about the text. Then the metadata is queried to extract relationships, both explicit and implicit, between textual attributes in a meaningful way. If you have an interest in representing or querying textual metadata we would be happy to help.