Open information may mean more errors get published

Making huge troves of raw government data freely available fast may mean accepting that much of that data are only as good as the people and processes behind them, a panel of government officials said Monday.

For example, in response to massive public demand for information during the 2010 BP oil spill, Energy Department and Environmental Protection Agency workers were struggling to push out data on ocean samples within 24 hours of when they were gathered.

That not only put a massive strain on the agency, diverting money and resources from other priorities, but also, in some cases, led to different types of data being improperly mixed in a way that was more confusing than helpful, said Tim Crawford, a senior adviser at EPA.

The best the agency could do, Crawford said, was to label the data sets to make clear what had gone through a full agency review and what hadn't.

"Being able to classify that information as basically a beta version, saying, 'This is at your own risk; we don't stand behind the values you see here,' that's very important," Crawford said. "Then you go down the list and say, 'Well, it's been looked at a little bit; it's in the stream; it's certified at various levels.' "

EPA is working on a standard protocol for describing how thoroughly data have been reviewed before they're released, Crawford said.

Crawford spoke at a panel discussion titled Ensuring Federal Data's Accuracy at the Excellence in Government conference hosted by Government Executive Media Group.

Data from EPA and other agencies are released most often these days on Data.gov, an Obama administration open government initiative that marked its second anniversary Sunday.

The site now holds roughly 3,000 government-generated data sets, but has been criticized by open government groups, which say it's less a tool for government transparency than a data dump for information like "the population count of wild horses and burros."

In some instances, Data.gov has become a single repository for information that was already routinely publicly released, such as the FBI's uniform crime reports , but never before pulled into a single location.

In those cases, the Data.gov platform, which includes comment and response sections, can be a check on data errors, allowing the FBI and other agencies to improve their data through crowd sourcing, according to Sanjeev Bhagowalia, a deputy associate administrator who works on Data.gov for the General Services Administration.

"You can now say, 'Hey, wait a second. You say in this particular place there are only five crimes, but I went to the police precinct and I was able to see 25 crimes,' " Bhagowalia said.

Bhagowalia said GSA is considering melding data.gov with the often arduous Freedom of Information Act process so that noncontroversial requested information can simply be posted to the website.

Stay up-to-date with federal news alerts and analysis — Sign up for GovExec's email newsletters.
FROM OUR SPONSORS
JOIN THE DISCUSSION
Close [ x ] More from GovExec
 
 

Thank you for subscribing to newsletters from GovExec.com.
We think these reports might interest you:

  • Sponsored by G Suite

    Cross-Agency Teamwork, Anytime and Anywhere

    Dan McCrae, director of IT service delivery division, National Oceanic and Atmospheric Administration (NOAA)

    Download
  • Data-Centric Security vs. Database-Level Security

    Database-level encryption had its origins in the 1990s and early 2000s in response to very basic risks which largely revolved around the theft of servers, backup tapes and other physical-layer assets. As noted in Verizon’s 2014, Data Breach Investigations Report (DBIR)1, threats today are far more advanced and dangerous.

    Download
  • Federal IT Applications: Assessing Government's Core Drivers

    In order to better understand the current state of external and internal-facing agency workplace applications, Government Business Council (GBC) and Riverbed undertook an in-depth research study of federal employees. Overall, survey findings indicate that federal IT applications still face a gamut of challenges with regard to quality, reliability, and performance management.

    Download
  • PIV- I And Multifactor Authentication: The Best Defense for Federal Government Contractors

    This white paper explores NIST SP 800-171 and why compliance is critical to federal government contractors, especially those that work with the Department of Defense, as well as how leveraging PIV-I credentialing with multifactor authentication can be used as a defense against cyberattacks

    Download
  • Toward A More Innovative Government

    This research study aims to understand how state and local leaders regard their agency’s innovation efforts and what they are doing to overcome the challenges they face in successfully implementing these efforts.

    Download
  • From Volume to Value: UK’s NHS Digital Provides U.S. Healthcare Agencies A Roadmap For Value-Based Payment Models

    The U.S. healthcare industry is rapidly moving away from traditional fee-for-service models and towards value-based purchasing that reimburses physicians for quality of care in place of frequency of care.

    Download
  • GBC Flash Poll: Is Your Agency Safe?

    Federal leaders weigh in on the state of information security

    Download

When you download a report, your information may be shared with the underwriters of that document.