On Wed, Oct 14, 2009 at 1:03 PM, Anca Luca <ancapaula.luca(a)xwiki.com> wrote:
* the backing annotation storage is based on
components, currently with an
XWiki
objects implementation but a different service can be easily created (there
is
one implemented for Scribo annotations stored in RDF)
Where is the source for the xwiki-based annotation storage? Is this it?
http://hg.nuxeo.org/sandbox/scribo/file/a9c5293ae8b3/scribo-xwiki-documents…
some other part of
http://hg.nuxeo.org/sandbox/scribo/summary ?
Also, I request a simple feature: The addition of an extra, separate,
string-property alongside every annotation. This would allow an arbitrary
JSON structure (text) to be entered, containing additional information.
As my interest is in semantic video and audio annotations, this kind of
"back-link" would allow easy addition of structured information alongside an
annotation. One of the issues with "annotation" in general is that there can
be a variety of information one wants to supply. For example, in
a structured document, one may want to annotate a named element, an
unnamed-element positionally defined within a larger named structure, as
well as an "extent" , in case the annotation doesn't correspond to an exact
DOM structure.
With streaming media, you will want to annotate with a start and a stop
time, but also there may be the need for much additional information to be
cross-referenced: for example, if using automatic indexing, structure or
dialog detection in a video or audio stream, one will typically have a
variety of other information one may need attached to any particular
annotation. So in my need for annotation, I may want to include a JSON
structure like this alongside every audio annotation:
{"title": "Michael Jackson Special", "file": "
http://www.aspenpublicradio.com/podcasts/sp/soulpatrol_mjspecialjul09.mp3&q…quot;,
"type": "sound", "link":
"http://www.aspenpublicradio.org/support.php",
"tags": "Music", "image":
"/Misc_icon_Media_Clip_1813.jpg", "author": "KAJX,
Soul Patrol Podcast", "description": "Clip: segment32 start:
00:15:12.600
end: 00:19:01.400 length: 228.80000000000007s.
analyzer=VampQmSegmenter:segment", "date":
"2009-07-05T08:00:00Z", "start":
912.6, "duration": 228.80000000000007, "captions.file":
"/xwiki/bin/view/Episodes/www_aspenpublicradio_com_podcasts_sp_soulpatrol_mjspecialjul09_mp3?xpage=plain&analyzer=VampQmSegmenter%3Asegment",
"trainspodder.analysis": "VampQmSegmenter:segment"}
This is from a real-world example:
http://nielsmayer.com/trainspodder-prototype-2009-10-16.jpg
... This discussion is timely, as I'm now implementing the functionality
behind the "add comment" button in the timeline popup bubble. Thus my
interest in finding the source code and it's status for your annotations
back-end for use outside your project....
Regarding Scribo,
http://trueg.wordpress.com/2009/05/14/scribo-getting-natural-language-into-…
that the main focus is on natural language processing for
documents, which is a laudable goal, as I spent some time in college writing
in prolog for just this purpose and then spent 8 years in research working
alongside NLP people.
See before the web existed, we did "web" over email:
http://nielsmayer.com/p93-shepherd.pdf and worried about semantic tagging of
"intent", and automating conversation structures based on tracking
conversational moves....
Today, however, a vast unexplored chunk of content and "intelligence" (and
it's dialectic, stupidity) out there is in audio and video form. For that,
semantic-audio technologies on which I base my work come in very handy,
e.g.
http://www.omras2.org/ . Is scribo intending to focus any effort on
semantic tagging audio and video content? Given the shared platform, Xwiki,
perhaps there's room for collaboration?
Also, please note that there is a good deal of potential with the "idiot
savant" technology such as that employed by
http://www.google.com/voice .
Excited to see things from the labs hit the mainstream, I installed my new
free google voice "Call Me" button on my website... the performance of the
transcription capabilities are less than stellar (it's own "annotation"
animated beautifully in synchrony with the recorded message, dynamically
underlining each spoken word in the transcript, many of which were
completely wrong):
*Niels Mayer* () .....11:10 PMHey this is a groovy tester google voice on my
homepage. I wanna see how prizes, as I can release my number. What they're
gonna do. I have keep number private sector so. Ohh no it's not facts of
course it's not gonna do it. Anyways, it's all very excited and an call on
the web and see what it actually and they call.
-- Niels
http://nielsmayer.com