General Actions:
Log-in
Register
cameleon
Main
Search
Page Actions:
Export
▼
:
Export as PDF
Export as RTF
Export as HTML
More actions
▼
:
Print preview
View Source
Collaborative and Automatic Methods for the Multilingualisation of Lexica and Ontologies
»
Panels
»
Quick Links Panel
»
First CAMELEON SHARED TASK
Wiki source code of
First CAMELEON SHARED TASK
Hide line numbers
1: == Multilingualization of ontologies == 2: 3: Ontologies are language-independent abstract representations of a portion of human knowledge. However, in practice, the labels and descriptions of the concepts and relations represented in an ontology are often expressed using natural language, to allow humans to more easily interpret the information encoded. Having such ontologies enables the pursuit of the ultimate goal of the semantic web, which is to use the information represented in these ontologies to allow automatic reasoning on large amounts of data. 4: 5: Therefore, towards the multilingual semantic web, there is a need for efficient automatic techniques that allow the translation of ontologies from one language to another. On the one hand, general-purpose machine translation systems, like Google translate, present lexical choice and ambiguity resolution problems, yielding bad-quality translations due to the limited available context and the specialized domain modelled by the ontology. On the other hand, the use of specific hand-crafted lexicons is onerous and such resources are only rarely available for a given language pair and domain. Furthermore, most of the available resources are targeted to English or use English as a pivot language. 6: 7: == Task description == 8: 9: The goal of this first edition of the CAMELEON shared task is to explore the problem of automatic ontology translation. The teams will be free to use whatever resources, tools and techniques they consider relevant to solve the problem. We will provide a set of ontologies in English, French and Portuguese as test set. The goal of the systems is to generate as output the labels for an ontology in a language L2 having as input the ontology in language L1. 10: 11: Therefore, each registered team will have the possibility to implement a system that translates between the three language pairs in each direction. The systems will be evaluated in the following translation directions: 12: 13: * //Input//: English ontology, //Output//: French ontology 14: * //Input//: English ontology, //Output//: Portuguese ontology 15: * //Input//: French ontology, //Output//: English ontology 16: * //Input//: French ontology, //Output//: Portuguese ontology 17: * //Input//: Portuguese ontology, //Output//: English ontology 18: * //Input//: Portuguese ontology, //Output//: French ontology 19: 20: If a team chooses which translation directions (language pair + direction) they want to target, from one to all. Each translation direction will be evaluated through a separate set of evaluation scores. To measure the performance of a system, we will compare each automatic translation obtained with the reference manual translation, using standard ontology matching measures like exact match and edit string distance. 21: 22: We would like to remember that this shared task is not at all a competition. It is an opportunity to focus on a given task which is of common interest within CAMELEON. 23: 24: Furthermore, we invite the teams to provide manual translations to the labels of each provided ontology. We will then use the manual translations to enrich the ontologies and to apply metrics to evaluate the precision of multiple translations. 25: 26: == Team registration == 27: 28: Each team willing to participate in the shared task should send, before December 21, 2012, an email to both task organizers carlosramisch at gmail.com and cassia.ts at gmail.com, containing: 29: 30: * Name of the team members and a name identifying the team; 31: * The translation directions targeted (please use the numbers 1 to 6 above, for instance, directions 3 and 5) 32: * A brief description (max. 1 paragraph) of the intended system. 33: 34: In this first edition, each participant team should include at least one member of the CAMELEON project. 35: 36: == The data set == 37: 38: The [[MultiFarm dataset>>http://web.informatik.uni-mannheim.de/multifarm/]] is composed of a set of ontologies describing the domain of organising conferences (the domain being well understandable for different researchers). The ontologies were developed within the [[OntoFarm project>>http://nb.vse.cz/svatek/ontofarm.html]]. These OntoFarm ontologies were manually translated into eight different languages (Chinese, Czech, Dutch, French, German, Portuguese, Russian, and Spanish) and the corresponding alignments between these ontologies were also manually set. For this shared task, we will use the MultiFarm translations in English, French and Portuguese. 39: 40: %The details about the input and output formats will be posted on the shared task website soon. 41: 42: === Open test set === 43: 44: **Attention**: A new version of the dataset was made available on May 01, 2013. Please use this last version as the previous one contained a spurious "null" class. 45: 46: We are making available a set of three ontologies for testing. This data is made available using the standard format [[OWL>>http://en.wikipedia.org/wiki/Web_Ontology_Language]]. We expect the participant systems to receive as an input one ontology in OWL, and the output will be another OWL ontology. The output ontology will be identical to the input one, except for the rdfs:label elements. For example, if the input ontology has labels in French, the output ontology will have the same labels translated into Portuguese, like in the example below: 47: 48: [[image:examplesharedTaskCAMELEON.png||width="50%"]][[image:examplesharedTaskCAMELEONpt.png||width="50%"]] 49: 50: * [[Download the open test set>>attach:CAMELEONsharedTask2013testData-v2.zip]] 51: 52: The evaluation will be performed both on the open test set and on a blind test set. The participants should send the output of the systems on the open test set before --April 30-- May 31, 2013. The blind test set will be made available soon after receiving the outputs of the systems on the open test set, and participants will be requested to return the outputs in 24h. The blind test set consists of a set of different ontologies in the same domain of conference organization. The evaluation will be performed using standard measures of ontology alignment calculated with the [[OAEI API>>http://oaei.ontologymatching.org/2011/align.html]]. 53: 54: == Evaluation and baseline [[image:new.gif]] == 55: 56: We are making available a package containing a script for calculating the evaluation metrics and a baseline translation for each direction. The tool can be downloaded here: 57: 58: * [[Download the evaluation and baseline package>>attach:eval-baseline.zip]] 59: 60: === Evaluation script === 61: 62: Once you have downloaded and unzipped the //eval-baseline.zip// package, you must place your results in the corresponding folder. In order to evaluate the systems in a systematic way, you have to put the file containing your translated ontology in the folder //translations/direction/ontology// and name it //mysystem.owl//. A //direction// should be named //language1-language2//, where //language1// is the source language and //language2// is a different target language. An //ontology// is one of the three ontologies in the open test set: //cmt//, //confOf// or //conference//. 63: 64: You MUST use the same filename across the different folders (for example, //mysystem.owl//). Please modify the variables in the script //run-eval.sh// according to the ontologies, directions and system you want to evaluate. 65: 66: === Evaluation metrics === 67: 68: The script //run-eval.sh// evaluates all the systems, directions and ontologies at once. You should edit the script variable //systems// to provide as many system names as translated ontologies you generated. If all the files are in their right place in the folders, the evaluation results will be a series of //direction-system.html// files in the //results// folder. For example, the baseline system //bing// for the direction //en-fr// generated the following results: 69: 70: |= algo |= |=**BLEU** |= |= |=|= |=**bing-equal** |= |=|= |=**bing-edit** |= 71: | test | Average | 1-grams | 2-grams | 3-grams | | Prec. | Rec. | FMeas. | | Prec. | Rec. | FMeas. 72: | cmt | 0.2503 | 0.4400 | 0.2380 | 0.1900 | | 1.00 | 0.25 | 0.40 | | 0.56 | 0.56 | 0.56 73: | confOf | 0.2533 | 0.5410 | 0.2300 | 0.1650 | | 1.00 | 0.24 | 0.39 | | 0.72 | 0.72 | 0.72 74: | conference | 0.2697 | 0.5690 | 0.2810 | 0.1720 | | 0.96 | 0.21 | 0.35 | | 0.63 | 0.63 | 0.63 75: |**H-mean** |**0.2574** |** 0.5103** |**0.2477** |**0.1750** | |**0.99** |**0.23** |**0.38** | |**0.63** |**0.63** |**0.63** 76: 77: All these automatic measures compare somehow the translation obtained by your system with the reference translation, generated by a human translator. In the future, we would like to generate multiple reference translations in order to yield a fairer evaluation. 78: 79: * **BLEU**: This is a standard measure in the evaluation of machine translation. See [[this paper>>http://acl.ldc.upenn.edu/P/P02/P02-1040.pdf]] for more details. It computes the proportion of common n-grams between the generated translation and the reference translation. In the next columns, values are broken down according to 1-, 2- and 3-grams. It is expected that there are fewer 2- and 3-grams in common than the number of 1-grams. Since the translations are short, we calculate BLEU only for up to 3-grams (instead of the typical value of up to 4-grams). When including 4-grams, some of the values for the baseline system were 0.0 because the system didn't share any 4-gram with the reference translation. 80: 81: * **bing-equal** and **bing-edit**: These two measures use an ontology alignment approach to evaluate the systems. Each correspondence between the system and reference labels are considered as a reference alignment. Therefore, we run two simple ontology alignment algorithms: equal and edit. The first one only aligns labels when they are 100% equal. The second aligns similar labels using a Levenstein string distance measure. Then, these alignments are compared with the reference alignments that should have been found if the translations were perfect. This gives an idea of the proportion of 100% correct translations and the proportion of translations where the system translation is reasonably similar to the expected reference translation. 82: 83: === Baseline system === 84: 85: The baseline system uses the [[Bing>>http://www.bing.com/translator]] standard translation engine, which is freely available, for translating the labels. No domain adaptation or tweaking was performed. The ontologies translated by Bing are placed in the folder translations/direction/ontology/bing.owl. 86: 87: The baseline evaluation measures are available in the results folder. The baseline translations were generated by Roger Granada. 88: 89: == Presentation of the systems == 90: 91: The strategy and results of each team will be presented in a short paper (max. 2 pages) in English. The format of the paper will be announced later. This paper will not be published publicly, but only internally on the CAMELEON wiki. Each team will have the opportunity to present their results during the Fourth CAMELEON workshop, around July 2013 in Porto Alegre (either in person or by videoconference). 92: 93: In the future, we would like to transform the resulting set of papers into a larger paper to be submitted to a conference or journal, and the short papers will be used as a starting point. Everybody willing to contribute to this submission will be welcome as a co-author. 94: 95: == The CAMELEON reading group == 96: 97: From the moment when the teams are registered, we intend to create a reading group on multilingualization of ontologies. This reading group will read one article per week, suggested by one of the participants, and then we will have a weekly skype meeting to discuss the paper. The skype meeting will be coordinated by the person who suggested the paper. The goal of this reading group is to familiarize the participants with the domains of ontology multilingualization, ontology matching, machine translation and multilingual lexical resources. 98: 99: The participation in the reading group is non-mandatory for the registered teams and open to other participants. 100: 101: == Important dates == 102: 103: * Registration open - November 30, 2012 104: * Test data available - --December 14, 2012-- December 17, 2012 105: * Team registration deadline - --December 21, 2012-- February 28, 2013 106: * System outputs due - --April 30, 2013-- --May 31, 2013-- July 26, 2013 107: * Description paper due (max. 2 pages) - --May 31, 2013 -- --June 30, 2013-- August 30, 2013 108: * CAMELEON workshop - to be announced (December 2013). 109: 110: == Reading suggestions == 111: 112: * [[A Machine Learning Approach to Multilingual and Cross-lingual Ontology Matching>>http://disi.unitn.it/~~p2p/RelatedWork/Matching/70310656.pdf]] 113: 114: * [[Cross-lingual entity matching and infobox alignment in Wikipedia>>http://www.sciencedirect.com/science/article/pii/S0306437912001299]] 115: 116: * [[Using Domain-specific and Collaborative Resources for Term Translation>>http://www.aclweb.org/anthology-new/W/W12/W12-4210.pdf]] 117: 118: * [[Simple methods for dealing with term variation and term alignment>>http://www.limsi.fr/pz/tia2011/Proceedings/pdf/TIA15.pdf]] 119: 120: * [[Babouk: Focused Web Crawling for Corpus Compilation and Automatic Terminology Extraction>>http://ieeexplore.ieee.org/xpl/articleDetails.jsp?arnumber=6040719]]
Navigation
About
Members
Publications
Resources
Workshops
CAMELEON 2011
CAMELEON 2012
CAMELEON 2012b
CAMELEON 2013
Tutorials
Shared task
CAMELEON 2014
Tutorials
CAMELEON 2015
Intranet
Wiki Dashboard
Document Index
Sandbox
My Recent Modifications
XWiki Scheduler Job Class
|
XWikiAllGroup
|
ceramisch
|
Admin
|
NewPage