The database used for the SWS 2013 evaluation has been collected thanks to a joint effort from several participating institutions that provided search utterances and queries on multiple languages and acoustic conditions (see Table 1). The database is available to the community for research purposes. Fell free to evaluate your query-by-example approaches for keyword spotting (spoken term detection). The database contains 20 hours of utterance audio (the data you search in), 500 development and 500 evaluation audio queries (the data you search for), scoring scripts and references.
Here ( -ws.org/Vol-1043/mediaeval2013_submission_92.pdf) you can find the MediaEval SWS2013 task description, and here ( -ws.org/Vol-1043/) are the system descriptions of particular teams which participated in MediaEval SWS2013 evaluations. An overview paper discussing achieved results was published at SLTU 2014 and is available here. If you publish any results based on this SWS2013 database, please refer the paper (bibtex).
2013) Multilingual
This Social Policy Report from the Society for Research in Child Development (SRCD)*, examines how to best support the development and learning of children who are multilingual, and offers recommendations for policy and practice.
On the Create Variation Label page, in Site Template Language, select the language to be used in the multilingual user interface (MUI) of the source site. The choice is based on which language packs are available with your Microsoft 365 subscription. Language packs are needed only if you want to use MUI for the administrative pages of the site. Language packs are not required for variations.
The Rome Workshop was the sixth in a series of W3C workshops that survey and share information about currently available best practices and standards that can help content creators and localizers address the needs of the multilingual Web, including the Semantic Web. These workshops provide an important opportunity to identify gaps that need to be addressed and this workshop shifts the focus to emphasize practical methods of making multilingualism a reality on the Web. The workshop was also designed as an opportunity for participants to network and share information between and across the various different communities involved in enabling the multilingual Web.
As an example, I installed German language pack on SharePoint 2010 English and enabled it as alternate language for a site in its settings. Please, note, that absolutely the same steps are working for SharePoint 2013 on-premise as well:
If you use SharePoint 2013 you will not find language menu in SharePoint UI, you have to change it in the browser settings: -language-in-sharepoint-2013/. In Internet Explorer open Internet Options and click Languages button. Here you can set the default language for your browser.
This last meeting is the possibility to get to know the members of the MOLTO Consortium, to attend talks and demos, to discuss the work, and to plan future collaboration. The program features presentations on the core technologies of MOLTO (tools for translating and for building translation systems), on research aspects (how to scale up the scope of MOLTO methods) and on use cases of these ideas (mathematics, museum object descriptions, business applications). The intended audience of the meeting includes computational linguists, computer scientists, developers of multilingual websites, museum curators, public sector officers, and members of the translation industry.
The final project meeting of MOLTO will be held in Barcelona, on 23 May 2013, hosted and organized by UPC in the beautiful Rectorat building:UPC, CAMPUS NORD - Edif. R, C. JORDI GIRONA, 31, 08034 BARCELONA, SPAIN.
In this paper we present three term weightingapproaches for multi-lingual documentsummarization and give results onthe DUC 2002 data as well as on the2013 Multilingual Wikipedia feature articlesdata set. We introduce a new intervalboundednonnegative matrix factorization.We use this new method, latent semanticanalysis (LSA), and latent Dirichlet allocation(LDA) to give three term-weightingmethods for multi-document multi-lingualsummarization. Results on DUC and TACdata, as well as on the MultiLing 2013data, demonstrate that these methods arevery promising, since they achieve oraclecoverage scores in the range of humansfor 6 of the 10 test languages. Finally,we present three term weighting approachesfor the MultiLing13 single documentsummarization task on the Wikipediafeatured articles. Our submissions signifi-cantly outperformed the baseline in 19 outof 41 languages.
EconPapers FAQ Archive maintainers FAQ Cookies at EconPapers Format for printing The RePEc blog The RePEc plagiarism page Multi-source, multilingual information extraction and summarization edited by Thierry Poibeau, Horacio Saggion, Jakub Piskorski, Roman Yangarber (eds.) springer, 2013. 323 pp. $129.00 (isbn: 978-3-642-28568-4 [print] 978-3-642-28569-1 [online])José L. Vicedo and David TomásJournal of the Association for Information Science & Technology, 2013, vol. 64, issue 7, 1519-1521Date: 2013References: Add references at CitEc Citations: Track citations by RSS feedDownloads: (external link) (text/html)Access to full text is restricted to subscribers.Related works:This item may be available elsewhere in EconPapers: Search for items with the same title.Export reference: BibTeX RIS (EndNote, ProCite, RefMan) HTML/TextPersistent link: :bla:jinfst:v:64:y:2013:i:7:p:1519-1521Ordering information: This journal article can be ordered from ... bs.asp?ref=2330-1635Access Statistics for this articleMore articles in Journal of the Association for Information Science & Technology from Association for Information Science & TechnologyBibliographic data for series maintained by Wiley Content Delivery (Obfuscate( 'wiley.com', 'contentdelivery' )). var addthis_config = "data_track_clickback":true; var addthis_share = url:" :bla:jinfst:v:64:y:2013:i:7:p:1519-1521"Share This site is part of RePEc and all the data displayed here is part of the RePEc data set. Is your work missing from RePEc? Here is how to contribute. Questions or problems? Check the EconPapers FAQ or send mail to Obfuscate( 'oru.se', 'econpapers' ). EconPapers is hosted by the Örebro University School of Business.
I use Google translate all the time on various language wikipedia articles in languages I can't read, and it would be great to have this as a way to supplement that activity. Sometimes I can read the google translation, but usually I just need to guess my way through the translated text and more often than not, in those cases my assumptions turn out to be dead wrong. This type of solid translation would be great as an extra support to get the gist of those automatic translations. Jane023 (talk) 07:17, 7 August 2013 (UTC)Reply[reply]
In the example given, there's a call to a template to express the statement that Berlin is the capital of Germany. But doesn't Wikidata already store (or at least, have the capacity for storing) facts like these? It seems redundant to re-express this fact outside of Wikidata. In a sense, as futuristic as this proposal sounds, it seems not ambitious enough. If the capacity exists to generate text in any language, why not just have automatically-generated articles? Yaron K. (talk) 13:16, 7 August 2013 (UTC)Reply[reply]
As I have interpreted the idea, it should be possible to use a version independant template that will generate for any language: an infobox, article text and categories for an article using data from wikidata but text betwen the dataobjects from elsewhere (wiktionary, in the template, from a very copact machinetranslation tool). Seing this from a static viewpoitn this is corresponding to a botgeneration from Wikidata, but there is several stronger benefits. This general template can reuse translations from one type of obejcts (say towns in Malaysia to town from Mali etc). Even stronger would be the possibility to extend the template without the need to modify the different articles. Say for a town it could start with X is a town i y community with 777 inhabitants. it could then later extend to include a table of number of inhabitants at different times and some more data of the town, like if a new adm level of a District will be introduced by the authotirtes. And then both the text part, data part and infobox will be extended at the same time in all versions. And to continue the template could strat to introduce when the city was founded important builings ,if it is close to a lake etc, and thne the textpart in the articles will be very much extended. So wikidata as it is, but then also a new enitity that include these type of version independant templates (or these seen as a special type of data in Wikidata).Anders Wennersten (talk) 18:36, 7 August 2013 (UTC)Reply[reply]
Very interesting proposal. We have been working on a similar approach demonstrated by our prototype system called AceWiki-GF. It builds upon the controlled natural language ACE and the grammar framework GF. See this paper for the details: -conferences.org/sites/default/files/papers2013/kaljurand.pdf Tokuhn (talk) 19:09, 7 August 2013 (UTC)Reply[reply]
Most excellent idea. I wonder if a side effect would be to make smaller massively multilingual wikis/projects feasible? Few entities that run wikis in the world have the capacity to maintain hundreds of independent wikis, or even more than one) and if they deal with multiple languages at all, have a smattering of pages translated, and no language-specific search.
I wonder if "N5: Caching. Since the content of a page will depend on the language settings of the user, an appropriate caching mechanism needs to be designed and deployed" is really necessary? Content negotiation could just direct user from multilingualwiki to xx.multilingulawiki, users could then navigate to other languages in ways they expect, it'd be external crawl/search friendly, and no special caching mechanism would be needed. But this is a triviality compared to the rest of the proposal. Mike Linksvayer (talk) 02:30, 9 August 2013 (UTC)Reply[reply] 2ff7e9595c
Comments