Planet RDF

It's triples all the way down

January 27

Libby Miller: Tangible enough

Thinking about the purpose of prototypes:

Make new and upcoming technologies and standards tangible enough to help people think through the consequences of them.

Technology is moving fast, but it is also unevenly distributed, and the consequences – good and bad – of emerging technologies may only become apparent as they move into the mainstream. By making these consequences tangible early we can choose between possible futures.


Posted at 20:14

Libby Miller: What is Radiodan for?

This is my view only, and there’s a certain amount of thinking out loud / lack of checking / potentially high bullshit level.

Yesterday I was asked to comment on a Radiodan doc and this popped out:

Posted at 11:59

January 25

Libby Miller: A quick Radiodan: Exclusively Archers

I made one of these a few months ago – they’re super simple – but Chris Lynas asked me about it, so I thought I should write it up quickly.

It’s an internet radio that turns itself on for

Posted at 14:29

January 23

Cambridge Semantics: Big Data Industry News Watch

A round up of recent industry news on the topics of Big Data and Enterprise Data Management

Posted at 15:00

Libby Miller: A quick analysis of wifi cards for using a Raspberry Pi as an access point

When Radiodan can’t access the web, it throws up an access point (AP) created by the Pi: you connect directly to that and it displays the available wifi points in a webpage as a captive portal, and asks you to add the password for the one you want. It’s not easy to get credentials for wifi to objects with no user interface, and this is the best one we’ve found so far (

Posted at 09:37

January 20

Cambridge Semantics: Putting the Smarts in Data Integration

Driving business value from your data often requires integration across many sources. These integration projects can be time consuming, expensive and difficult to manage. Any short cuts can compromise on quality and reuse. In many industries, non-compliance with data governance rules can put you firm’s reputation at risk and expose you to large fines.

Traditional data integration methods require point to point mapping of source and target systems. This effort typically requires a team of both business SMEs and technology professionals. These mappings are time consuming to create and code and errors in the ETL (Extract, Transform, and Load) process require iterative cycles through the process.

Posted at 20:47

AKSW Group - University of Leipzig: Two AKSW Papers at #WWW2015 in Florence, Italy

Hello Community!
We are very pleased to announce that two of our papers were accepted for presentation at WWW 2015.  The papers cover novel approaches for Key Discovery while Linking Ontologies and a benchmark framework for entity annotation systems. In more detail, we will present the following papers:
Visit us from the 18th to the 22nd May in Florence, Italy and enjoy the talks. More information on these publications at
Ricardo on behalf of AKSW

Posted at 15:09

Bob DuCharme: R (and SPARQL), part 2

Retrieve data from a SPARQL endpoint, graph it and more, then automate it.

Posted at 13:32

January 15

Cambridge Semantics: 2014 - What a year!

Happy New Year !!

Posted at 19:03

Redlink: Redlink, The Data Linking API

Today Sergio Fernández and John Pereira have attended the Vienna Semantic Web Meetup to present Redlink, The Data Linking API to the very active Viennese community.


At  the meetup page you can find more photos.


Posted at 14:50

January 14

Ebiquity research group UMBC: 2015 Ontology Summit: Internet of Things: Toward Smart Networked Systems and Societies

The Internet of Things (IoT) is the interconnection of uniquely identifiable embedded computing devices within the existing Internet infrastructure.

The theme of the 2015 Ontology Summit is Internet of Things: Toward Smart Networked Systems and Societies. The Ontology Summit is an annual series of events (first started by Ontolog and NIST in 2006) that involve the ontology community and communities related to each year’s theme.

The 2015 Summit will hold a virtual discourse over the next three months via mailing lists and online panel sessions augmented conference calls. The Summit will culminate in a two-day face-to-face workshop on 13-14 April 2015 in Arlington, VA. The Summit’s goal is to explore how ontologies can play a significant role in the realization of smart networked systems and societies in the Internet of Things.

The Summit’s initial launch session will take place from 12:30pm to 2:00pm EDT on Thursday, January 15th and will include overview presentations from each of the four technical tracks. See the 2015 Ontology Summit for more information, the schedule and details on how to participate in these free an open events.

Posted at 18:17

January 13

Bob DuCharme: R (and SPARQL), part 1

Or, R for RDF people.

Posted at 13:26

January 12

Dublin Core Metadata Initiative: DC-2015 website and Call for Participation open

2015-01-12, DCMI and the host of DC-2015, Universidade Estadual Paulista--Sáo Paulo State University (UNESP), are pleased to announce the publication of the Call for Participation at and the opening of the DC-2015 website at DC-2015 will take place in Sáo Paulo, Brazil on 1-5 September 2015. Just as DCMI celebrates it's 20th anniversary this year, it also celebrates the first time it's Annual Meeting and International Conference have been located in South America. Watch the conference website at for updates as DCMI and UNESP develop an exciting program around the conference theme of "Metadata and Ubiquitous Access to Culture, Science and Digital Humanities".

Posted at 23:59

Dublin Core Metadata Initiative: Shanghai Library joins DCMI as an Institutional Member

2015-01-12, DCMI is very pleased to announce that the Shanghai Library ( has joined DCMI as an Institutional Member. The Shanghai Library is the second largest library in China--second only to the National Library. The Shanghai Library was founded in 1952. In October 1995, the Shanghai Library and the Institute of Scientific and Technical Information of Shanghai merged to become a comprehensive research public library and center for industrial information. It is also the branch of the National Cultural Information Resource Sharing Project in Shanghai, the main library of the Shanghai Central Library System, Shanghai Ancient Books Protection Center and the "Pioneer Technology Development Research Center" of the Shanghai soft science research base. The DCMI Institutional Member Program is open to all public sector organizations interested in supporting DCMI while participating actively in DCMI governance. Please see the DCMI membership page at for more details about DCMI's membership programs.

Posted at 23:59

Frederick Giasson: Open Semantic Framework 3.2 Released

Structured Dynamics is happy to announce the immediate availability of the Open Semantic Framework version 3.2. This is the second important OSF release in a month and a half. triple_120

This new major release of OSF changes the way the web services communicate with the triple store. Originally, OSF web services were using a ODBC channel to communicate with the triple store (Virtuoso). This new release uses the SPARQL HTTP endpoints of the triple store to send queries to it. This is the only changes that occurs in this new version, but as you will see bellow, this is a major one.

Why switching to HTTP?

The problem with using ODBC as the primary communication channel between the OSF web services and the triple store is that it was adding a lot of complexity into OSF. Because the UnixODBC drivers that are shipped with Ubuntu had issues with Virtuoso, we had to use the iODBC drivers to make sure that everything was working properly. This situation forced us to recompile PHP5 such that it uses iODBC instead of UnixODBC as the ODBC drivers for PHP5.

This was greatly complexifying the deployment of OSF since we couldn’t use the default PHP5 packages that shipped with Ubuntu, but had to maintain our own ones that were working with iODBC.

The side effect of this is that system administrators couldn’t upgrade their Ubuntu instances normally since PHP5 needed to be upgraded using particular packages created for that purpose.

Now that OSF doesn’t use ODBC to communicate with the triple store, all this complexity goes away since no special handling is now required. All of the default Ubuntu packages can be used like system administrators normally do.

With this new version, the installation and deployment of a OSF instance has been greatly simplified.

Supports New Triple Stores

Another problem with using ODBC is that it was limiting the number of different triple stores that could be used for operating OSF. In fact, people could only use Virtuoso with their OSF instance.

This new release opens new opportunities. OSF still ships with Virtuoso Open Source as its default triple store, however any triple store that has the following characteristics could replace Virtuoso in OSF:

  1. It has a SPARQL HTTP endpoint
  2. It supports SPARQL 1.1 and SPARQL Update 1.1
  3. It supports SPARQL Update queries that can be sent to the SPARQL HTTP endpoint
  4. It supports the SPARQL 1.1 Query Results JSON Format
  5. It supports the SPARQL 1.1 Graph Store HTTP Protocol via a HTTP endpoint (optional, only required by the Datasets Management Tool)

Deploying a new OSF 3.2 Server

Using the OSF Installer

OSF 3.2 can easily be deployed on a Ubuntu 14.04 LTS server using the osf-installer application. It can easily be done by executing the following commands in your terminal:

mkdir -p /usr/share/osf-installer/

cd /usr/share/osf-installer/


chmod 755


./osf-installer --install-osf -v

Using a Amazon AMI

If you are an Amazon AWS user, you also have access to a free AMI that you can use to create your own OSF instance. The full documentation for using the OSF AMI is available here.

Upgrading Existing Installations

It is not possible to automatically upgrade previous versions of OSF to OSF 3.2. It is possible to upgrade a older instance of OSF to OSF version 3.2, but only manually. If you have this requirement, just let me know and I will write about the upgrade steps that are required to upgrade these instances to OSF version 3.2.


Now that the triple store’s SPARQL HTTP endpoint requires it to be enabled with SPARQL Update rights, it is more important than ever to make sure that the SPARQL HTTP endpoint of the triple store is only available to the OSF web services.

This can be done by properly configuring your firewall or proxy such that only local traffic, or traffic coming from the OSF web service processes, can reach the endpoint.

The SPARQL endpoint that should be exposed to the outside World is OSF’s SPARQL endpoint, which adds an authentication layer above the triple store’s endpoint, and restricts potentially armful SPARQL queries.


This new version of the Open Semantic Framework greatly simplifies its deployment and its maintenance. It also enables other triple stores that exist on the market to be used for OSF instead of Virtuoso Open Source.

Posted at 18:15

January 11

Tetherless World Constellation group RPI: Another AGU and we all get wet from the rain in San Fran…

The 2014 Meeting of the American Geophysical Union in the wet city of San Francisco has not yet faded from memory. Unfortunately, it may be remembered for the “year of the RFID mess” over the great science progress. However, let’s start with the positive. Rensselaer’s Tetherless World was well represented – see what we did at = Patrick, Stephan, Marshall, Evan and Paulo (representing others including Linyun and Han) in talks, posters covering both research and project progress, and the academic booth (go RPI!). This year, we presented in Informatics (IN) and Education (ED) sessions with talks and many posters. Just on a logistics note, I was very pleased to have the exhibit hall adjoined to one of the poster halls this year. This made the task of moving between them and not missing one or the other, much easier. Hope that continues. It was another excellent year for Informatics; I’ve misplaced the stats but suffice to say increasing numbers of abstracts, great student contributions and a sea of new faces. A continuing treat is the Leptoukh Lecture (honouring Greg L, whom I still miss very much). This year, Dr. Bryan Lawrence (working in the UK, but actually a Kiwi) gave a tour de force lecture on computation and data aspects of climate science. The attendance was excellent, clearly pulling in a wide cross-section of attendees from well beyond the IN folks. Thanks Bryan. This year was the change over for Informatics leadership with Kerstin Lehnert taking over from Michael Piasecki as President – thanks Michael for your leadership and efforts over the last two years. Ruth Duerr (NSIDC) came in as President-Elect and Anne Wilson (CU/LASP) as secretary. Diversity rules in Informatics!!!

In regard to IN poster sessions, we saw an increase in the flash mob approach. What is that you ask? It is where, at an appointed time during the poster session, the session convener arranges for all poster presenters to be present. After having also advertised by twitter, email and general coercion, they gather poster attendees around each poster (in order, down the row). The presenter has 5 minutes to present their poster and then the mob moves on. It has shown to be a very effective way of engaging attendees and the presenters. If the session organiser has pre-planned it, the sequencing can also be very effective. After each has been presented, may attendees stay to quiz specific posters they were interested in. The one aspect that makes this style hard is the general noise level in the poster hall. Poster presenters need to “speak up” and project their voice: not all are prepared for that but it is very good practice!

I am author / co-author on quite a few presentations each year. This year I had two posters (both invited) as lead. You can see them via the link above. Sixth generation of data and information architectures, and Anatomy and Physiology of Data Science drew quite a lot of interest. But I must say, I did enjoy getting to stand with Mark Parsons at our poster “Why Data Citation Misses the Point” (I will add that to the website) and elaborate on our premise. Interestingly, we had a lot of agreement with the work — we’d hope to provoke arguments (!! as usual !!). Now to find time to write that up.

I want to acknowledge the excellent presentation of other works I was co-author on. The TWCers noted above are indeed skilled and knowledgeable researchers and practitioners. I know that but it is always excellent to have peers approach me to tell me that and how impressed they are with both the work and the people!

And the RFID issue – just go here and see for yourselves:

See all of you next December.


Posted at 00:04

January 08

W3C Data Activity: CSV on the Web: new drafts including JSON and RDF conversion

The CSV on the Web Working Group has published four drafts. Alongside updates to the existing Model for Tabular Data and Metadata and Metadata Vocabulary for Tabular Data documents, are two new documents. These describe mechanisms for generating JSON and … Continue reading

Posted at 17:57

January 07

Orri Erling: DBpedia Usage Report

We've just published the latest DBpedia Usage Report, covering v3.3 (released July, 2009) to v3.9 (released September, 2013); v3.10 (sometimes called "DBpedia 2014"; released September, 2014) will be included in the next report.

We think you'll find some interesting details in the statistics. There are also some important notes about Virtuoso configuration options and other sneaky technical issues that can surprise you (as they did us!) when exposing an ad-hoc query server to the world.

Posted at 20:12

January 06

W3C Data Activity: Spatial Data on the Web WG launched

It was 10 months ago today, 6th March 2014, that the Linking Geospatial Data workshop in London came to an end with Bart De Lathouwer of the OGC and I standing side by side announcing that our two organizations would … Continue reading

Posted at 15:01

January 03

Libby Miller: Raspberry Pi podcast-player-in-a-box – step by step


Podcast-player-in-a-box is a way to associate a physical object (a plastic card) with a possibly-changing list of audio files. When you put the card in the box it plays the audio.

It’s inspired by

Posted at 16:54

January 01

W3C Read Write Web Community Group: Read Write Web — Q4 Summary — 2014


The web ponders moving further towards SSL, with the W3C TAG publishing a draft finding on how this could be more easily achieved.  There was a great review by the EFF on progress, as well as some interesting suggestions by timbl.

Linked data continues it’s inexorable march towards the mainstream with steady progress throughout the quarter and whole year.  Some good reviews are available here, here, and here.  With a look forward to what we may see in 2015.  A cool ontology viewer called VOWL also caught the eye.

There were some more discussion regarding the HTTP PATCH verb and how it applies to data, with specs and implementations reaching readiness.  A comprehensive wishlist covering much of the future of LDP and RWW was posted by Sandro, as well as a new authentication system, called SPOT (Simple Page-Owner Token).

Communications and Outreach

Henry Story delivered an outstanding presentation at Scala eXchange conference in London, where he outlined the current state of play of the read write web and decentralized social web.  An overview of the project is available on github, as well as source code.

Some conversations took place in the identity credentials group and open badges  which aims to allow writing of achievements, via badges, on servers, in images and data structures, and using digital signatures.


Community Group

The LDP Patch specification is now reaching readiness and I believe the integration into GOLD is going to happen as we speak. GOLD has also now integrated JSON LD.

Community group has added a slack instance, which allows slightly more realtime chat, an API and many other features.

A stub wiki page has been added on the concept of “nanotations“, linking to Kingley’s blog explanation, feel free to add your own examples!


Some initial work has started on intelligent personal assistants.  Juergen has written a sioc bot which is able to take real time conversations and convert them to linked data.  Leveraging adapters in hubot the code is available on git and was up and running in just a couple of days.

I’ve also been working on a linked data robot that allows simple transfer of credit (aka marking) from one URI to another.  The hope is to build out a linked data based transfer and reputation system.  A slightly related side project I’ve started is virtual wallet, which will allow holding of web currencies and transfer between WebIDs, much more standards work to be done here…


Last but not Least…

An interesting system called webhose has been launched.  “The API – Easily integrate data from hundreds of thousands of global online sources: message boards & forums, blogs, comments, reviews, news and more”.  Seems like a neat bridge to pull into your apps news from many web2.0 data sources!

Posted at 10:47

December 30

Ebiquity research group UMBC: PhD defense: Varish Mulwad — Inferring the Semantics of Tables


Dissertation Defense

TABEL — A Domain Independent and Extensible Framework
for Inferring the Semantics of Tables

Varish Vyankatesh Mulwad

8:00am Thursday, 8 January 2015, ITE325b

Tables are an integral part of documents, reports and Web pages in many scientific and technical domains, compactly encoding important information that can be difficult to express in text. Table-like structures outside documents, such as spreadsheets, CSV files, log files and databases, are widely used to represent and share information. However, tables remain beyond the scope of regular text processing systems which often treat them like free text.

This dissertation presents TABEL — a domain independent and extensible framework to infer the semantics of tables and represent them as RDF Linked Data. TABEL captures the intended meaning of a table by mapping header cells to classes, data cell values to existing entities and pair of columns to relations from an given ontology and knowledge base. The core of the framework consists of a module that represents a table as a graphical model to jointly infer the semantics of headers, data cells and relation between headers. We also introduce a novel Semantic Message Passing scheme, which incorporates semantics into message passing, to perform joint inference over the probabilistic graphical model. We also develop and explore a “human-in-the-loop” paradigm, presenting plausible models of user interaction with our framework and its impact on the quality of inferred semantics.

We present techniques that are both extensible and domain agnostic. Our framework supports the addition of preprocessing modules without affecting existing ones, making TABEL extensible. It also allows background knowledge bases to be adapted and changed based on the domains of the tables, thus making it domain independent. We demonstrate the extensibility and domain independence of our techniques by developing an application of TABEL in the healthcare domain. We develop a proof of concept for an application to generate meta-analysis reports automatically, which is built on top of the semantics inferred from tables found in medical literature.

A thorough evaluation with experiments over dataset of tables from the Web and medical research reports presents promising results.

Committee: Drs. Tim Finin (chair), Tim Oates, Anupam Joshi, Yun Peng, Indrajit Bhattacharya (IBM Research) and L. V. Subramaniam (IBM Research)

Posted at 00:07

December 23

Tetherless World Constellation group RPI: American Geophysical Union Informatics

A held post from last year – just releasing it….

I’ve  lost count of how many AGU meetings I’ve been to, except for knowing that this was my 11th consecutive year to the Fall Meeting in recent times. I am thinking about this since I also received my 25th year AGU pin this year. Ouch. To say there was a lot going on at AGU, is like saying it gets busy around the shops during the holidays. So, it was an average year for me in terms of length of day and tiredness, etc. Each year, I have at least one stand out memory. This year it began with the number of colleagues from solar and space physics that I bumped into (and remembered and they remembered me) and had very interesting and relevant conversations with (about software, and data, and science). Next in line was Simon Cox’s Leptoukh lecture – an excellent tour de force to demonstrate what taking a few steps back and conceiving a core observations and measurement model can do to impact a significant number of application fields. Well done mate. The RPI Tetherless World contributions were (again) very strong – to see what I mean – take a look at: My appreciation goes to Patrick, Marshall, Yu, Linyun, Evan and Massimo for your efforts – the booth, the posters, the talks – (and Jin, Han, John, and others left back on the ranch) – all provided an excellent showcase of our (RPI/TWC) collective work. Now, on to the science – informatics to be specific in all its discipline-specific forms – the Special Focus Group ESSI ( is thriving with increasingly diverse participants and new faces appearing each year. As far as topics, I’ll spare you all a word cloud but “Data” was the word. The other word was, well, “Big” – in the sense that a number of sessions succumbed to Big Data (or at least the Era of Big Data, one phrase I prefer)… and more than just informatics; Union, Education, Global Environmental Change, … Thus, I’m okay with that. With a meeting that big science highlights are hard to capture in a shortish blog.   Having been around long enough, it is normal (or even required) for me to be critical of certain aspects of the meetings logistics/ organization as they affect people and the efficacy of the scientific exchange itself. I have shared those concerns as well as the positive aspects to the appropriate people/ committees. If any of you wish to pass your observations (positive and otherwise) to me, I will pass them on. One thing I cannot let pass is the new AGU data policy that was approved and pre-released during the meeting. I am sure that there will be some noise about this in the days to come.

Posted at 00:31

December 22

Cambridge Semantics: Adding Clarity to Big Data Analytics

We recently announced the integration of our Anzo Smart Data Platform (Anzo SDP) with the KeyLines network visualization tool. The advancement will enable business analysts and IT professionals in Global 2000 companies to gain new “big picture” business insights from their big data queries on diverse data.

The combination of Anzo SDP and Cambridge Intelligence’s KeyLines solution will especially benefit use cases in pharma and financial services including: 

  • Drug discovery – The ability to visualize the connection between research and results in this heavily data-driven process can help avoid effort duplication, identify gaps in understanding, and ensure discovery is more cost efficient.
  • Clinical trials – Ensuring a conclusive trial is key to any drug’s success. Network visualization techniques can help find potential participants with the required profiles, analyze trial results and ensure overall vigilance.
  • Compliance Surveillance – Link and visualize activities, web logs, email and phone archives, IM communications, and other sources to uncover potential violations of regulatory requirements as well as internal policies and procedures violations.

Posted at 16:46

Redlink: Join Our Team!

Redlink – Our vision is to build outstanding software based on linked data, smart document indexes and content analysis.
We are hiring.  Join the team!

(1) Java Developer

We are seeking innovative, organized and proactive individuals with a positive, energetic attitude with Bachelor’s degree in Computer Science or related discipline. Skills should include:

  • Object-oriented analysis and design skills
  • Development of software components and products using Java/J2EE technologies
  • Server side Java application development experience
  • Willing to learn and adapt to new programming languages or paradigms
  • Knowledge of Spring framework, CDI or related IoC frameworks
  • Familiar with Maven build tools
  • Experience with designing and implementing RESTful web services (JAX-RS)
  • Knowledge of relational databases (PostgreSQL) and Object-Relational tools (Hibernate)
  • Experience writing high-quality, testable, maintainable, and well-documented code
  • Ability to work independently as well as with a team
  • Self-disciplined with strong attention to detail

These things are not required but are certainly pluses:

  • Semantic Web technologies (RDF, SPARQL)
  • Schema-less databases
  • Experience with Continuous Integration tools
  • Basic Frontend Web development
  • Experience working in Agile/Scrum

 Please send your application to

(2) Frontend developer

We are looking for developers for our dashboard and various client projects (mobile and web applications). You are strong in understanding customer requirements, designing prototypes and software solutions with are based on standards and  UI frameworks. Your  technical skills include:

  • Fluent knowledge of Javascript, html and CSS
  • Good knowledge of AngularJS and tools such as Bower, npm Grunt, Git, SASS
  • Good knowledge of REST /SOAP services

You want to work closely with a multidisciplinary, agile team.

 Please send your application to

(3) UI Designer

We are seeking for a creative, but user experience expert. New mobile and interactive experience within a mobile and web UI for humans and machines is what you will create for us:

  • Knowledge and tools for UI and UX-Design
  • Good understanding of front-end tools
  • Creativity and ability to lead the communication with team and customers

You want to work closely with a multidisciplinary, agile team.

 Please send your application to

Posted at 12:45

Copyright of the postings is owned by the original blog authors. Contact us.