aaron's picture

scientio has got something - using 'concept structures'

| | | | | |

Edmonds. 2007. Using concept structures for efficient document comparison and location. Conference Proceeding

I was so pleased to finally (and serendipitously, I might add) find a computer science article that describes what I was trying to do with my masters work, from outside the discipline.

This is a quick read. Spells out terms very clearly for non-adepts, so I found it to be quite accessible.

aaron's picture

googling 'textual analytics'

| | | | |


InfoProfiler beta - Features

Information sources:

InfoProfiler can deal with all content sources where text is available in a machine-readable format e.g., World Wide Web, SEC Filings, Proprietary databases, and company Intranets.

Extracts Text and Information pieces:

InfoProfiler has inbuilt text extraction capabilities and can extract information from a website as well as from a given set of websites. It understands the structure of a webpage and decides whether it is a news, forum, review, bulletin board or a blog. InfoProfiler can also remove unwanted text portions (such as, advertisements) from a given information source.

aaron's picture

Tools for non-linear access to multi-disciplinary domain knowledge

| | | | | | | | | | | |

Every existential assertion creates a cascade of existential assumptions.

Take, for example the brief definition: Dimensions are measures of magnitude or extent. In order to understand this defintion, the concepts mentioned on the surface--measures, magnitudes and extents--must each exist and I must be able to know as much. The existential requirements of this relatively simple assertion are thus complicated by knowledge requirements.

So we know that knowledge is involved. But what do we know about knowledge? We know that the nature of knowledge, its very possibilty, and the range of what can be known relies upon the nature of existence, its possiblities, and the range of what can exist.

aaron's picture

tex.tuals process

| | | | | | | | |

As I read through a text (I've remarked on the paper-to-digital conversion process elsewhere) I want to be able to highlight and capture whole swaths of text.

It is critically important to repurposing my captured snippets (bits, fits, blobs, fragments, portions) that I am able to recontextualize them easily.

How can this recontextualization happen?

First; the bibliographic information must be embedded in each bit.

Also, its relative location in the linear flow of the text must be recorded, so as to be able to quickly pull up various degrees of context around the bit.

Syndicate content   bookmark tuals 0.1 on

Bg Bottom