Upload
3-round-stones
View
556
Download
0
Embed Size (px)
DESCRIPTION
Overview of US EPA's Linked Data Service to launch in early 2013. Open data published using the Linked Data model increases search engines' ability to find and display high value data sets. Linked Data enables policy makers, analysts and developers to more readily access and re-use data.
Citation preview
Linked Data at US EPA
1-Feb-2013Linked Data Workshop with EPA OSWER
By Bernadette Hyland, David Wood & Luke Ruth
1These slides will walk us through a common workflow use case comparing the Linked Data Service to Envirofacts.
Agenda• Intros ...
• Trends in data management
• Government data publication
• Update on EPA Linked Data Service
• EPA OSWER moving towards Linked Data?
• Review Next steps ...
2
Trends in government data management
3
4
Headlines and agency memos about government transparency with open data and various government Web sites.... innovation challenges based on open government data
... High energy datapalooza’s are emerging with awards ranging from a couple thousand to $100k+. These challenges open the doors to innovation for better healthcare solutions and more efficient use of energy, to name but a few. They all require access to and re-use of HIGH QUALITY DATA.
In 2012, we read many headlines about big data and world’s search engines and social media sites.
5Photo credit: http://www.flickr.com/photos/glennharper/4452247708/5
However, while there is lots of gold to be mined from public data, it is an uncomfortable time for Government IT and business managers who are tasked with data management programs.
Most people are having a difficult time keeping up. If you feel like you are hanging on while the world changes too fast, you are not alone.
Photo credit: http://www.flickr.com/photos/glennharper/4452247708/
6
Who is sharing their data as Linked Data? Small and large commercial and government organizations, NGOs, Non-profits ... plus many universities. Governments in the last few years have been responding to Open Government initiatives that mandate publishing open government data. Some are careful, slow-moving entities who simply needed to find real solutions to real problems.
GovernmentsGoals: Governmental transparency and/or improved
internal efficiencies (data warehouses)
7
Big DataSimple dataComplex dataLegacy data
8
KEY POINT: Search, discovery and data access approaches have evolved over the last decade and techniques are beginning to come together. GoPubMed was launched in 2002 as the first semantic search portal. Later, Microsoft’s Bing, Google’s Knowledge Graph are two of the other well known search engines employing semantic techniques.
Semantic search systems generally considers the context of search, location, intent, variation of words, synonyms and concepts. Semantic search has roots in linguistic research and NLP.
Big data research has grown to include the MapReduce algorithm for handling really large data sets, often measured in terabytes or greater. This is the kind of data that people at the Large Hadron Collider at CERN are working on to provide insights into how the universe works, including the recent discovery of the Higgs Boson, the particle that gives mass to matter.
Under the big top tent of semantic search we’re dealing with different types of content, big, public, complex and legacy data. Simple, complex and legacy data comes in small, medium and large sizes.
Many government agencies by contrast have lots of small to medium data sets in structured databases, like Oracle. These databases (and the systems that depend upon them) are not going away however fewer new data warehouse projects are likely to be started. Data warehouses are widely recognized to be costly to create and maintain, and change SLOWLY.
The biggest win for governments worldwide who adopt a Web architecture for data publishing is combining data sets to discover new or previously uncontemplated relationships.
“Big Data Is Important, but Open Data Is More Valuable”As change agents, enterprise architects can help
their organizations become richer through strategies such as open data.
David Newman, VP Research, Gartner
9
Open data refers to the idea that certain data should be freely available to everyone to use and republish as they wish, without restrictions from copyright, patents or other forms of control.
The term “open data” has gained popularity with open data initiatives including data.gov.uk, data.gov and other government data catalog sites.
Enterprise architects are playing an important role in fostering information-sharing practices. Access to, and use of, open data will be particularly critical for a business that operate using the Web; organizations should focus on using open data to enhance business practices that generate growth and innovation.
Open data + open standards + open platforms
Highly scalable computing &
hosting via the CloudInternational Data Exchange
Standards
5 Star Data (Linked Data)
Open Source tools
10
A Web-oriented approach to information sharing has impacted how scientists, researchers, regulators and the public interacts with government.
Linked data lowers the barriers to re-use and interoperability among multiple, distributed and heterogeneous data sources.
Access to high-quality Linked Open Data via the Web means millions of researchers and developers will be able to shorten the time-consuming research process involving data cleansing and modeling.
11
How do we get a loose coupling of shared data over Web architectures? By using the structured data model for the Web: RDF.
There is a project to create freely available data on the Web in this way, which is known as the Linked Open Data project.
W3C sees Linked Data as the set of best practices and technologies to support worldwide data access, integration and creative re-use of authoritative data.
12
The mission of the Government Linked Data (GLD) Working Group is to provide standards and other information which help governments around the world publish their data as effective and usable Linked Data using Semantic Web technologies.
13
We are 16 months into the Government Linked Data Working group’s two year charter.
14
A sound government information management strategy requires providing CONTEXT and CONFIDENCE to those accessing and potentially re-using your data.
Giving people have timely access to information, for disaster preparedness, scientific research, policy and research, the network effect of people helping people is our greatest hope.
On the heels of the recent East Coast hurricane that devastated parts of New York and New Jersey, government executives suggested that fear of cyber-doom scenarios may be taking too much of our thinking & planning. According to Secretary Panetta, it may be driving us to unrealistic and potentially dangerous responses to threats that don’t exist.
The reality is that when disaster strike, people come together and help one another. We don’t see paralysis, panic and social collapse.
During today’s session, I’ll describe how several agencies and private sector organizations are using Web technologies and semantics to improve information access and discovery. Simply put, semantic technologies provide CONTEXT.
Open Government Data
15
“We’re moving from managing documents to managing discrete pieces of open data and content which can be tagged, shared, secured, mashed up and presented in the way that is most useful for the consumer of that information.”
-- Report on Digital Government: Building a 21st Century Platform to Better Serve the American People
Growing chorus ...
16
The Digital Government Strategy sets out to accomplish three things: Access to high quality digital information & services; procure and manage devices, applications, and data in smart, secure and affordable ways; and unlock the power of government data to spur innovation.
Governments around the world are defining detailed digital services plans based on open data, open APIs and open source data platforms. They are defining how governments are publishing data with an eye towards improving access and re-use. Administrators and program managers are committing to delivery of digital services using semantic technologies broadly, and Linked Data specifically.
Big data
• Simple data
• Complex data
• Legacy data
Integrating ...
17
We need to find ways to fit things together that wasn’t originally intended to fit together.
NB: This is the Musée du Louvre which has evolved from a late 12th Century fortress under Phillip II, extended over centuries to incorporate the landmark Inverted Pyramid architected by I.M. Pei that was completed in 1993.
A recent competition to house its new galleries for Islamic art opened this year, 2012. It continues to accommodate new works for art & galleries in new & previously unanticipated ways.
Today, we need to understand the context of big data + complex data + public data and legacy data into one consistent whole.
18
September 2011: 295 datasets that meet the LOD Cloud criteria, consisting of over 31 billion RDF triples and are interlinked by around 504 million links.
THERE IS A PROCESS
PublishConvertDescribeNameModelIdentify
Maintain
19
Take comfort in the fact that there is a familiar process. It is similar to the process & roles of traditional data modeling.
Creating Linked Data requires that we identify the data, model exemplar records -- what you are going to carry forward & what you are going to leave behind.
Name all of the NOUNs. Turn the records into URIs.
Next, describe RESOURCES with vocabularies.
Write a script or process to convert from canonical form to RDF. Then publish. Maintain over time.
3 Round Stones produces the leading platform for the publication of reusable data on the Web. Our commercially supported Open Source platform is used by the Fortune 2000 and US Government agencies to collect, publish and reuse data, both on the public Internet and behind institutional firewalls.
20
Our goal is to produce the leading platform for the publication of reusable data on the Web.
Callimachushttp://callimachusproject.orghttp://3roundstones.com
21
Callimachus is that platform. It is available via 3roundstones.com or its Open Source site callimachusproject.org.
CONTENTMANAGEMENT
SYSTEM
LINKED DATAMANAGEMENT
SYSTEM
Callimachus
UNSTRUCTURED
TEXT
TEXT
STRUCTURED
DATA
DATA
22
Callimachus may be compared to a distributed CMS. CMS’s manage mostly unstructured information. Callimachus, by contrast to a CMS, manages primarily structured Linked Data. We call this a Linked Data management system.
23
Callimachus started in 2009 as a simple online RDF editor. Users could fill out HTML forms, which would create RDF behind the scenes. The resulting RDF could be viewed as HTML and, of course, shared and combined with other RDF data. Since then HTML5 has allowed us to hack the browser environment much less and extend its capabilities.
24
Clinical Trials + enterprise linked
data
US Legislation + enterprise data
DBpedia + enterprise datasets
Data driven Web apps using Callimachus
24
Callimachus integrates (very) well with other enterprise systems as well as Web content. It can form an entire application or part of one.NB: Mention Documentum, Oracle via HTTP
25
• US HHS committed to making a vast array of open data more readily available to improve health care delivery & reduce costs in 2013 and beyond.
• In 2012, Sentara created a Web application that integrates authoritative data from 5 different sources including content from NLM, NOAA, EPA and DBpedia
• This application utilizes open data, open standards and an open source data platform
User
NOAA US EPA AirNow
DBpediaNational Library of Medicine
US EPA SunWise
26
US EPA Linked Data
• Cloud-based Linked Data provision of 3 core programs:
• 2.9M Facilities
• 100K substances• 25 years of toxic pollution reports• FISMA compliant
• 16 Callimachus templates
• Official launch Feb 2013
27
28
Envirofacts, EPA’s older system.
29
EPA’s new Linked Data system. Cooperation without coordination. Data reuse breaks the back of API gridlock. Clay Shirky stole that from me :)
30
This data is exactly the same data used to create the interface. Unlike traditional database-driven applications, the data is immediately accessible for reuse by third parties. This prevents data duplication, allows for tracking of provenance and avoids reinventing the wheel.
We’ve Seen This Before
31
Like HTML and RDF, credit cards have a human-readable side and a machine-readable side.
Contractor (3 Round Stones, Inc.)
Public
Application, Script or automated client
Web Browser
SPARQL endpointREST APIResource URIs
Linked Data management systemlocated at a Tier 1 Cloud Provider
(FISMA compliant)
RDF Database
Registered developer
32
Introduce Callimachus, an open source, open data platform based on open standards. 3 Round Stones provides commercial support for Callimachus and is a major contributor to the OS project.
Users of Callimachus see a generated Web interface, but can also directly access the data via REST or SPARQL.
SPARQL Named Queries (like stored procedures) allow for automated conversion to different formats for reuse in non-RDF environments.
From WikipediaFrom EPA
Open Street Map
33
Data may be easily combined from several sources.
HOW IT IS DONE TODAY ...
34
Audience for EPA Data
• Middle school student doing a science project
• Concerned citizen worried about local pollution
• Environmental Science PhD from EPA
• Doctor from NIH writing a research paper
35To try and understand the advantages and disadvantages of both systems, we need to know the audience for the system. That presents a problem though. It’s nearly impossible to know your audience at any given moment. Even if it were possible, the audience is so varying that it would be unwise to cater to a single group at the expense of another.
The audience could be a middle school student, a concerned local citizen, a PhD collecting information for a report, or a doctor writing a research paper. We just don’t know.
For example, if the system was designed to accommodate a 6th grader, the system would be over simplified and thin. If it was designed with a PhD or a doctor in mind, the average citizen could be overwhelmed and find the system complicated and verbose.
That’s why the goal should be to make the simple things easy and the complicated things possible.
How much mercury did Hanson Permanente Cement
release in 2004?
36With that in mind, let’s walk through our example trying to keep all our audience members in mind. Let’s pretend our audience members live in Cupertino, California and want to know about the local Cement plant. The question is - How much mercury did Hanson Permanente Cement release in 2004?
37The process starts out much the same. The user enters their zip code into the search field, which in this case is 95014, and are presented with the results. From here though, the workflow is quite different.
Envirofacts
38
39Rather than immediately returning a list of the facilities in that zip code, Envirofacts gives users the option to verify their location, drag, drop, and resize a map to match their request. While this provides a high level of granularity it is making a simple thing harder than it needs to be for a lot of users. While some people may know the physical boundary of their zip code, the average user would most likely trust the application to take care of that.
40Envirofacts then returns a page that looks like this (with some rows obviously cut from the table). It’s great that right up front, the user is able to both see the results and have access to the data. There is a link they can paste into their browser and get the raw data as well as a button to click that will download the results as CSV. The big problem here though is that the data comes back with formatting that renders it nearly unusable and content that adds very little value above what is present on the screen.
Copying and pasting the link at the top of the screen gives us the following data:
41It comes back in what *appears* to be CSV. However, there is no actual indication of that. There are no column headers, no descriptions of what fields represent, the text is difficult to partition and understand, and there are no links to documentation on the use of the API or how the query is structured. All we know is that it is data from “this report”.
The other link - the CSV Table link - returns the following data:
42Having the table as CSV is theoretically useful but it too comes back without the necessary structure to understand and use the data. This time there are column headers but it is unclear what exactly they represent. Many of the fields are just the URL links found in the HTML table from the website. There doesn’t appear to be raw data - just links to where the raw data could potentially be found.
Moving on from accessing the data, we will try to find the facility we are looking for - Hanson Permanente.
Finding Hanson Permanente
43Finding Hanson Permanente can only be done in this table by scrolling down and finding "Hanson" in the alphabetically sorted list of facilities. There are no options to sort on the contents of another column or search within the table. The unfortunate side effect of this is that by the time you scroll down you can no longer see the column headers. Simply looking at this screen, you can see that there are 8 separate reports that can be viewed, but it is unclear how they are differentiated and what each contains.
The key here is that the data reflects internal EPA systems - which are unknown to the majority of users. By doing this, Envirofacts is implicitly asking users to become expert on internal EPA systems which they either are not capable of, or do not have the time for.
Finding Mercury Released in 2004
44Because most users do not have this knowledge, the first report they'll most likely click is the Summary Report. The “Summary Report” brings us to a long page where after quite a bit of scrolling we can see the Toxic Releases for 2011. However, unlike the previous search results, this data is not available for download or retrieval by any means other than screen-scraping or re-keying. It is also a limited dataset and does not have the data for 2004.
Compliance Report
45The Summary, Facility, AFS, BR, RCRA, TRI, TSCA Reports at their top level do not have the data about Mercury either. It is actually contained in the Compliance Report. However, like the other tables, there is no way to download this data and repurpose it for other applications. The other source of confusion is that this data can be found in multiple places depending on its originating report, and it can be unclear whether the data is in fact the same.
For example, this data can also be found by drilling down in the TRI Report by clicking:“View Report”->”P2 Report (Report)”->”P2 Report”-> and then manipulating the view based on the year and view you want. These graphs and charts ultimately contain very interesting and relevant data but they are so obscured and inaccessible that it becomes extremely difficult to create anything new.
Potential Audience
• Middle school student doing a science project
• Concerned citizen worried about local pollution
• Environmental Science PhD from EPA
• Doctor from NIH writing a research paper
✔
XX
X
46Who did we cater to? The middle school student? Probably not. The concerned citizen? Unless that citizen happens to have specific knowledge of the EPA system and a great deal of experience navigating technology, most likely not. What about the Environmental Impact PhD and the Doctor from NIH? They may have the knowledge to understand column names, chemical compounds, and reporting a bit better but only the Environmental Science PhD with a working knowledge of EPA’s system can determine enough information to make use of it. The doctor on the other hand still is still working against the system itself to find the data behind it.
Linked Data
47
Now let’s look at the same workflow in the Linked Data Service.
Finding Hanson Permanente
48
By keeping the application simple - and letting the results be viewed either as a table or a map - the user can adjust their search as they see fit without extra navigation. Also, by having the data in a table that can searched or sorted however the user sees fit, finding a specific facility is as easy as typing the name in or sorting on relevant criteria. This is made possible by exposing the data, rather than containing it in a standard HTML table.
I fully recognize that Envirofacts could offer identical functionality by tweaking their application, but the key underlying point is that this application was created very cheaply and quickly *because* the data is modeled as Linked Data. When the developing environment is a Web Browser, and the data is described and Linked, an application can be a simple XHMTL page with JavaScript, instead of a heavy-weight dedicated application.
Finding Mercury Released in 20041
2
49
There are two very important things to note on this page. 1 is that on any facility’s page, there is always an option to download the data. This data is available in two formats (RDF/XML and Turtle). With the click of a button a user can have all of the data that was used to drive the creation of the current page, which means he or she can repurpose that data into any new application. Note here that this download is not an extract, summary, or recreation of the data - it is literally the *same* data that was used to drive that page.
2 is that because this page is “data-driven”, navigation relies on exploring the data, not the system that contains it. On the same page where we get information like it’s latitude and longitude, we can also find a link to a report detailing exactly how much mercury was released in 2004. We could easily do an in-page search for 2004 or Mercury to identify the releases associated with those terms.
TRI Report
50
Rather than aggregating the data for presentation, the actual report is presented with the raw data continuously available in the top right of the page.
A subtle difference to be pointed out here is the difference in the name of the facility. Previously it was identified as Hanson Permanente, but now it is known as Lehigh Southwest Cement Co. During the modeling phase, the Linked Data was created to implicitly include this relationship (which is known via the mapping of EPA FRS identifiers). On the other hand, pulling down the CSV files would not give the user any obvious way of understanding this relationship.
Data Reuse
51
Lastly, giving users the ability to grab the data off any page, at any time during navigation, strongly facilitates the reuse of data. These graphs are not natively embedded in the webpage of a given facility. Rather, by downloading the data the user can quickly and easily make new and different visualizations for a report or presentation.
For example, this history of air stack pollution reports was made with a single parameterized SPARQL query and a single JavaScript pattern. This could very easily be applied to any number of facilities, changed to a bar graph, or altered in any number of other ways with very little effort thanks to the fact it was modeled using Linked Data.
Potential Audience
• Middle school student doing a science project
• Concerned citizen worried about local pollution
• Environmental Science PhD from EPA
• Doctor from NIH writing a research paper
✔
✔
✔
✔
52
Linked Data allowed us to reach all the members of our potential audience by giving the user options, aggregating based on relevance rather than data source, and by exposing the data that drives the service for reuse.
The middle school student or concerned citizen that want to know the location of a facility, the amount of a particular chemical it released, and the year it was released in never have to click any of the options in the Linked Data box. They can simply use the interface, explore the data, and find what they need in a read-only experience.
The Environmental Science PhD is still able to find what he is looking for with Linked Data but can do so in a much more intuitive way. The doctor from NIH is now able to find the data they’re interested in and if they choose to take the next step, download the actual data behind the page. By quickly and easily obtaining the raw data, anyone from scientists to journalists can generate their own applications without any knowledge of the Linked Data Service itself.
What Callimachus is
53
Subject
Predicate Object
54
The heart of Callimachus is a template engine used to navigate, visualize and build applications upon Linked Data. Here we see some typical RDF data, with a subject, a predicate and an object.
Subject
Object(Predicate is defined in a template)
55
Callimachus can use that data to build complex Web pages.
Subject Predicate
(Object gets filled in when template is evaluated)
56
It does this with a template language that is simply XHTML with RDFa markup. There are some extensions for syntactic convenience.
Templates
• Written in XHTML+RDFa (declarative pattern);
• Parsed to create SPARQL queries;
• Query results are filled into the same template.
57
HTTP GETrequest
Class
Viewable
XMLtemplate apply.xsl
HTML
SPAR
QL
quer
y
RD
F re
spon
se
HTTPresponse
Web serverRDF Store
Resource
Controller
XHTML+RDFatemplate
TemplateEngine
58
Callimachus is implemented as a Web MVC architecture with an underlying RDF DB. The process shown demonstrates how a view is generated from a Web request.
Create/edit templates are HTML forms
59
Callimachus templates may also be used to create or edit RDF data by generating HTML forms.
60
Callimachus provides a pseudo file system that is used to store and represent content, including RDF/OWL data, named SPARQL queries, schemata, templates, etc. The pseudo file system provides a common view of content that is abstracted from its actual storage location; RDF data is stored in an RDF store whereas file-oriented content is stored in a BLOB store.
61
Documents, including data and ontologies, can be uploaded via drag-and-drop when using an HTML5-compliant browser. File upload via a separate interface is available for older browsers.
Contractor (3 Round Stones, Inc.)
Public
Application, Script or automated client
Web Browser
SPARQL endpointREST APIResource URIs
Linked Data management systemlocated at a Tier 1 Cloud Provider
(FISMA compliant)
RDF Database
Registered developer
62
Users of Callimachus see a generated Web interface, but can also directly access the data via REST or SPARQL. SPARQL Named Queries (like stored procedures) allow for automated conversion to different formats for reuse in non-RDF environments.
63
Callimachus can associate SPARQL queries with URLs, so that they are executed when their URL is resolved. We call these “named queries” and they are analogous to stored procedures in a relational database. Named queries can accept parameters, which allows them to be a very flexible way to manage routine access to queries that can drive visualizations.
64
The view of a named query displays its results. Results, like template results, are naturally cached to increase performance.
65
The results of named queries may be formatted in a variety of ways, or arbitrarily transformed via XProc pipelines and XSLT. This screenshot shows the results of a named query being used to drive a Google Chart widget. Callimachus also has stock transforms for d3 visualizations.
Leather tags holding metadata
Papyrus rolls
66
Credits
David NewmanGartner: “Innovation Insight: Linked Data Drives Innovation Through Information-Sharing Network Effects” Published: 15 December 2011
David Wood, ed. Linking Government Data, Springer (2011) http://3roundstones.com/linking-government-data/
US Executive Branch
Digital Government Strategy: Building a 21st Century Platform to Better Serve the American People, http://www.whitehouse.gov/sites/default/files/omb/egov/digital-government/digital-government.html
W3C Linked Data Cookbook http://www.w3.org/2011/gld/wiki/Linked_Data_Cookbook
All other photos and images © 2010-2012 3 Round Stones, Inc. and released under a CC-by-sa licenseAll other photos and images © 2010-2012 3 Round Stones, Inc. and released under a CC-by-sa license
67
This work is Copyright © 2011-2012 3 Round Stones Inc.It is licensed under the Creative Commons Attribution 3.0 Unported LicenseFull details at: http://creativecommons.org/licenses/by/3.0/
You are free:
to Share — to copy, distribute and transmit the work
to Remix — to adapt the work
Under the following conditions:Attribution. You must attribute the work in the manner specified by the author or licensor (but not in any way that suggests that they endorse you or your use of the work).
Share Alike. If you alter, transform, or build upon this work, you may distribute the resulting work only under the same or similar license to this one.
68
This presentation is licensed under a Creative Commons BY-SA license, allowing you to share and remix its contents as long as you give us attribution and share alike.