On Monday, December 2 at 1.30 pm in Room P-702 (Paulinum), Mohamed Morsey will give a final rehearsal for his PhD defense ?E*fficient Extraction and Query Benchmarking of Wikipedia Data?. Guests are encouraged to both provide feedback about improvements to the talk and ask preparatory questions.
As always, Bachelor and Master students are able to get points for attendance and there is complimentary coffee and cake after the session.
E*fficient Extraction and Query Benchmarking of Wikipedia Data
The thesis consists of two major parts:
- Semantic Data Extraction: the objective of that part is to extract data from semi-structured source, i.e. Wikipedia, and transform it into a networked knowledge base, i.e. DBpedia. Furthermore, maintaining the up-to-dateness of that knowledge base to be always in synchronization with Wikipedia.
- Triplestore Performance Evaluation: normally the semantic data is stored on a triplestore, e.g. Virtuoso, in order to enable the efficient querying of that data. In that part we have developed a new benchmark for evaluating and contrasting the performance of various triplestores.
On Monday, November 18 at 3 ? 4 pm (not 1.30 ? 2.30 pm!) in Room P-702 (Paulinum), Andreas Nareike and Natanael Arndt will present their current project ?Electronic Resource Management in context of Libraries?, current research and give an prospect of there PhD topics. As always, Bachelor and Master students are able to get points for attendance and there is complimentary coffee and cake after the session.
Electronic Resource Management in libraries
Within the results of this project we want to apply semantic web technologies, to manage for instance licenses for electronic journals at the Leipzig University Library. We focus on building a scalable and reusable, intelligent data management platform.
The platform should accumulate and homogenize the heterogeneous data in in RDF and any other representation formats from different provenances.
Our current research
When reusing ontologies or vocabularies in an application context there occurs a gap between the domain ontology and the program logic. In parallel with the software product line method from software engineering we introduce the application schema as substantiation of the domain ontology. This application schema defines a subset of terms from the domain vocabularies resp. ontologies and enriches the resulting schema by application specific restrictions.
While keeping the terminological knowledge in domain ontologies, this approach encourages application engineers to a greater reuse of existing domain vocabularies in combination with software components.
With the aid of the Semantic Web Application Framework OntoWiki we implement a use-case within the context of library information systems as application on the Web of Data.
Thereby we utilize the application to describe forms, component
interfaces and restricted views for rolls in workflows.
On Monday, November 11 at 1.30 ? 2.15 pm in Room P-702 (Paulinum), Muhammad Saleem will present the ISWC Big Data Track Challenge winning paper ?Fostering Serendipity through Big Linked Data? and ?DAW: Duplicate-AWare Federated Query Processing over the Web of Data?. As always, Bachelor and Master students are able to get points for attendance and there will be complimentary coffee and Berliners after the session.
?Fostering Serendipity through Big Linked Data? and ?DAW? by Muhammad Saleem (30 minutes + question time)
Muhammad Saleem completed his bachelor in Computer Software Engineering from N-W.F.P University of Engineering and Technology and Master in Computer Science and Engineering from Hanyang University, South Korea. Currently, he is working as a PhD student at Agile Knowledge Engineering and Semantic Web (AKSW) University of Leipzig, Germany. His research interests includes federated SPARQL query processing over Linked Data, knowledge extraction and database management.
He will give a brief talk about two papers 1) DAW: Duplicate-AWare Federated Query Processing over the Web of Data 2) Fostering Serendipity through Big Linked Data presented at ISWC2013. DAW is a duplicate-aware approach to SPARQL federated query processing to achieve the same recall while querying fewer number of sources. It can be used in combination with any federated SPARQL query engine to optimize the number of sources it selects, thus reducing the overall network traffic as well as the query execution time of existing engines. While the second paper aim to foster serendipity using Big Data triplification, its continuous integration, and visualization. As a proof of concept, the integration, visualization of the constant flow of bio-medical publications with the 7.36 billion large Linked Cancer Genome Atlas (TCGA) dataset is shown in the paper.
Welcome to the Website of the Institute for Applied Informatics (InfAI) e.V.
Founded in 2006, the InfAI – Institute for Applied Informatics e.V. – pursues the promotion of science and research in the areas of computer science and business computing for the public good. The InfAI is a recognized adjunct institute of the University of Leipzig.
We look forward to meet you!
JPEG 9 with lossless JPEG compression immediately available
The Independent JPEG Group (IJG) at the Leipzig Institute for Applied Informatics has released the new version of the software library „libjpeg“. This software library is the basis for the processing of images in the ubiquitous JPEG format and is used worldwide in a countless number of applications of software vendors and in the photographic industry.
It is available under a fee-free license with open source code. After 15 years of continuous development, the upgrade to the new version 9 enables for the first time an effective lossless image compression mode and thus offers end users and industrial adopters a high added value. The Code is ready for download at the IJG Development Site.
The Big Clean
The Big Clean is a free, 1-day conference that addresses the challenges associated with reuse of public sector data. The event rises to the occasion of teaching the members of public how to recycle public data and reuse it in the pursuit of their own goals, ranging from the use of data to power applications, drive societal changes and provide grounds for a data-driven journalism. The Big Clean is focused on three principal topics:
- Screen-scraping: the skill of distilling data out of web pages and other poorly structured sources
- Data refining: the techniques of transforming raw data into usable data
- Data-driven journalism: the craft of telling stories with data
- Date and time: Saturday 3rd November, 2012, from 9 AM to 5 PM (GMT+1)
- Location: National Technical Library, Prague, Czech Republic (map)
- Web site: http://bigclean.techlib.cz/
- Facebook event: ((http://www.facebook.com/events/413790828682484/
- Costs: the admission is free
- How to register: Please use this form
INFORMATICS 2010 – The 40th Annual Conference of the Society for Computer Science in Leipzig
The 40th Annual Conference of the Society for Computer Science (GI) takes place from 27th September until 2nd October 2010 in Leipzig. Workshops, panels and other events will be held from Tuesday, 28th September 2010, until Thursday, 30th September 2010. Tutorials as well as the graduate programme will complement this conference on Monday and Friday.
Experts from science and industry will give a thorough overview on the most important new trends around the conference's topic of Service Science – New Perspectives for Computer Science. The focus will be on the „E-Humanities“ as a link between applied informatics and the humanities. The highlight on Wednesday, 29th September 2010, will be the Computer Science Day with invited talks on the conference's topic. Likewise, the 100th birthday of Konrad Zuse, the co-inventor of the first modern computer, will be honoured through talks and an exhibition.
Workshop proposals should be handed in by 10th January 2010.
Further information can be found at the Society's website for the annual conference.
Conference SABRE 2009 – Software, Agents, and Services for Business, Research and E-Sciences
March 23 – 25, 2009, Mediencampus Villa Ida, Leipzig, Germany
Next year's SABRE 2009 – Software, Agents and Services for Business, Research, and E-Sciences will feature the following Conferences and Events:
- BPSC – Business Process and Services Computing
- InnoLogIST – Intern. Conference on Innovative Logistics Information Systems and Technologies
- ISSS 2009 – International Symposium on Services Science
- MDD&PL – Model-Driven Development and Product Lines: Synergies and Experience
- TMS – Text Mining Services
Last Modification: 2013-01-14 17:57:04 by Roland Fassauer