Getting ino the semantic web and RDF using N3 (original) (raw)
W3C | Semantic Web| SWAP
Examples - Getting into RDF & Semantic Web using N3
This is a collection of examples to accompany the Primer. (The files have URIs starting http://www.w3.org/2000/10/swap/test/ if you are reading this on paper.) These are designed as examples, to show how something can be done, not as working products.
Remember, to convert an RDF document from n3 syntax into xml syntax,cwm xxx.n3 -rdf
should do it among other things. I haven't exhaustively tested all combinations of n3 and RDF in all cases.
Merging graphs
When explaining why the semantic web needs to be a web not a tree, I used slides showing information in blue from one source (blue links) with that from another sources (red and green links).
- The first diagram shows some information maybe from someone's personal information that they will attend an event which has a given home page. It also has the person's email address and given name.
- The second diagramshows how information about the meeting has been read from another source, and also privacy policy link has been picked up from the meeting page indicating what the privacy policy is.
The key point, of course, is that URIs are used to identify the overlap concepts such as the person, meeting and its home page. This allows the graphs to be merged. It is a simple RDF example, let's see what it looks like:
- The blue arcs (in N3, in XML made by
cwm -rdf blue.n3
) - The red arcs (in N3, in XML)
- The green arcs (in N3, in XML)
- The merged graph (in N3 made by
cwm red.n3 blue.n3 green.n3 > white.n3
, in XML)
Ontologies in DAML
Family relationships - GEDCOM
Here is an ontology for geneology, (where from?). Thanks to Jos De Roo for putting this into N3.
- Some rules relating relationships. One's father's sister is ones' aunt, and so on. (gedcom/gedcom-relations.n3)
- Some data of some imaginary families (gedcom/gedcom-facts.n3);
- The same data after processing by the rules, with each person's extended family relationships calculated (gedcom/lots.n3)
- gedcom/Makefileshows (for unix people) the commands to buld them
TAMBIS
The owners of the next two ontologies retain all rights, so you should not copy them for the purposes of making your own ontology.
These examples do not contain rules, just ontology information in DAML. DAML allows you to describe the properties and class you create and the relationships between them. Here are two fairly large examples from biochemistry. They seem to have been generated using the OILED software. They are just here as examples.
The tambis ontology was provided as an example by Ian Horrocks. Files intest/tambis
- tambis/tambis-full.daml was the original file;
- tambis/tambis-no-ents.daml has the XML entities expanded, but with universally quanitified variables still unidentified;
- tambis/tambis-forSome.n3 has existential variables declared as such (by a hack)
- tambis/tambis.n3 is what the ontology looks like in N3
BioPathWays
Eric Neumann kindly provided a sketch ontology from the biopathways consortium.
- pathway/Makefile has the recipes, very similar to above for Tambis
- pathway/pathway.daml original file in RDF
- pathway/pathway.noents with XML entities expanded
- pathway/pathway.n3 result in N3
- pathway/pathway.rdf result regenerated in RDF
Schema validation I
In the semantic web, there are many different forms of document validity. One useful form is to check that the document doesn't violate any rules associated with the vocabularies it uses, and that for each namespace it uses, every term it uses is actually mentioned in the schema for that namespace.
- daml-ex.n3 An example schema using DAML
- invalid-ex.n3 A document which declares terms in an invalid way
- schema-rules.n3 A set of rules about the DAML vocabulary
- schema-filter.n3 A filter to remove all the worlkings and preserve just the result
The command line I used with the closed world machine was
cwm daml-ex.n3 invalid-ex.n3 schema-rules.n3 -think -filter=schema-filter.n3
This combines the schema, the document and the rules, and draws all possible conclusions, then filters out any conclusions of invalidity.
The rules are by no means complete, and this is by no means the only way to do validation. A special-purpose program would be more efficient of course. However, this demonstrates how easy it is when data is a semantic form to process it using a general purpose engine.
(update Mar 2002: see validate.n3 discussedbelow)
Rules making rules
The a transitive property, like ancestor, is one such that if Fred is Joe's ancenstor, and Joe is Bill's ancenstor, then Fred is Bill's ancenstor. That (for all Fred, Joe and Bill) is a rule for "ancenstor", and the fact that that is true for any transitive property in place of "ancenstor" is the definition of transitivity.
- rules12.n3 has the "implies" rules to draw the above conclusion in two stages, first concluding a specific rule for ancenstor from the general rule for transitivity, and secondly using it in the specific case of Fred and company.
- rules13.n3 has hte same, but uses "log:means" (a bidirectional implication) to indicate that transitivity is that and only that. "Means" is not a primitive for cwm, so it is defined in terms of log:implies. This doesn't then acheive anything spectacularly new otherwise.
KIF to RDF: Converting axioms
There were some axioms for RDF and RDFS and DAML prodcued by Deborah McGuiness and Fikes for the DAML work, in KIF. KIF list syntax is pretty like N3 list syntax so a sed file converts the KIF to N3 list syntax. Then, a bunch of rules decorate the lists with their equivalent in N3. The result is quite a mess, but all the intermediate results are labelled as "log:Chaff". This magic class has the feature in cwm that -purge removes any statements from the store which mention (as subject, verb or object) anything which is in class Chaff.
The process is very recursive, and so uses the -think option on cwm, which iterates over all rules. It takes (2001/02) a long time, partly beccause it is not directed in its simple-minded approach, and partly because it is not optimised. The rules are by no means complete, so the axoim set is not translated in its entirity. This example is unfinished but still may be instructive.
- dpo/Makefile contains the recipes - files are all in the dpo subdirectory
- dpo/axioms.txt was cut and pasted from the axiomatic semantics document
- dpo/axioms.n3 is the result ofmaking this into N3 format for the KIF lists
- dpo/axioms.sed was the stream editor file used to do that;
- dpo/kludge.n3 labels everything in the file as an axiom
- dpo/consts.n3 is a list of all the constants which are deemed the same in KIF and N3
- dpo/convert.n3 has the actual rules for deriving some N3 expressions from some KIF expressions;
- dpo/result.n3 is the result of conversion
- dpo/purged,n3 is the result of converting and purging the log:Chaff;
- dpo/purged2.n3 is the result of purging it still futher (too far!) removing all lists as well
- dpo/noKIF.n3 is the file usedto do that, to label all lists as log:Chaff.
Integrating applications: Roadmaps, Organizations and diagrams
Many see the semantic web's impact on the industry as a tool for Enterprise Application Integration. That is, just as the web integrates human-oriented information systems, so the Semantic Web integrates applications in which data has well-defined meaning. Here are some files which represent data which deal with technical architecture and the organizational structure around W3C. Imagine that these had been produced from specific applications for technical design and organizational management. Now it happens that there is a need bit of free softrware called GRphviz which is a tool for drawing diagrams. Putting all three together we can draw digrams of relationships between technical architecure and consortium organization. We can also make rule files to extract subsets of this data when focussing on a specific problem or area.
- How the data is processed - diagram
- Technical architecture; diagram, data
- Organizational diagram, data
- Subsets: technical architecture planned for: web services, semantic web, both
Following links: Schema validation II
The schema validation above looked for inconsistencies with a schema. The example worked by the file and the schemas being given to cwm on the command line. However, it would be more useful to find the schemas automatically. By using notation3's formulae, it is possible to extract all the predicates used in an RDF document. If the namespace's URI scheme is dereferencable (such as http:), or if the user has a local schema catalog, then one should be able to look up the schema in real time. The is done with cwm's builtinlog:semantics
which gives the N3 formula corresponding to the parsing of the semantic content of a resource.
- The rule file which checks schema validity - validate.n3
Note how, when validate.n3 follows a URI to a schema, it doesn't just believe it - it looks objectively at what it says. People early on with the Semantic Web were always afraid that inconsistencies could be sown which would trip up browsing agents. It is fundamental that you don't beleive everying on the Web. The log:semantics and log:includes built-ins allow rules to look objectively at other documents to see what they say.
As well as checking that the prediccates used are indeed declared as Properties in the schema, this file also checks consistency as the example above. This involves, for each schema, finding the set of facts which the schema implies. In other words, you don't have to state explicitly in the schema that something is (say) a class, if you have already said that it is a subclass of something. The validate.n3 rules takes each schema, merges it with the schema rules, and finds all the conclusions. This gives an augmented schema -- the schema plus the things implied by it according to the defined meanings of the terms (like subclass) used. This uses the built-ins log:merge
(for merging formulae) andlog:conclusion
(for doing, in effect, a cwm --think
internally on a specfic formula).
Secure links - toward trust on the web
Try using the crypto module (thanks to Sean Palmer) for making rules which only accept information from resources on the web if they have the right cryptographic hash, or the right digital signature.
Secure applications: Delegated authentication
Adding cryptographic functions to an semantic web engine makes it into a trust engine. This is important for the
Suppose we want to give access to the W3C "member lounge" web site to the emplyees of member organizations. We trust any organization's appointed representative to say who is an employee. This can be done very simply using public key cryprography. The bare bones of the system are mapped out in part of the cwm cryptography test suite.
Appendix - Test files
There is a regresion test for cwm which uses many files.
Parser test files
These are simple semantic web examples in N3 and RDF syntaxes made typically in order to test the software, but you can use them to test your understanding.
- s1.n3 some simple schema information
- animal.rdf a little DAML ontology about an animal - from a DAML discussion
- double-conclusion.n3 Testing that a reasoner wouldn't draw duplicate conclusions
References
Tim BL, with his director hat off Id:Examples.html,v1.202002/05/1519:49:25timblExpId: Examples.html,v 1.20 2002/05/15 19:49:25 timbl Exp Id:Examples.html,v1.202002/05/1519:49:25timblExp