The open government directive has specific requirements like the following: Within 45 days, each agency shall identify and publish online in an open format at least three high-value data sets ... and register those data sets via data.gov... These must be data sets not previously available online or in a downloadable format. Tracking those specific requirements on data.gov (separate from the open government dashboard ...more »
This entry is a consensus recommendation of seven organizations that work on government transparency of which OpenTheGovernment.org is one. Require agencies to indicate whether a high value dataset was previously unavailable, available only with a FOIA request, available only for purchase, or available, but in a less user-friendly format. Going forward, this will make it easier to track how agencies are complying with ...more »
On the main data.gov website, there is the Agency Participation tab. It does what you expect. It shows agencies and related statistics. What it doesn't do is simplify search. If you present this information, the statistics entice you to click on the participant to see what and why. If you can generate the numbers, you can populate a hyperlinked search to the datasets, tools, and more. Having to go back and then filter ...more »
The current wikipedia Data.gov page is a great, open , neutral platform for the entire comunity to post the links to the latest global to local data.gov sites and the apps that have been generated by developers. It can become an authoritative reference for the growing open data community. Here is the link : http://en.wikipedia.org/wiki/Data.gov
It would be great if users could see all the Data Sets that have been suggested. This could cut down on duplicate data set suggestions.
In light of some of the recent issues in the environmental community: It may be good to have a list of datasets that the government has/uses that can't be released and the reasons why. Ie. we can't release dataset x because it contains dataset y which we purchased from company z (and due to contractual/legal reasons we can't release dataset y). Another example would be data that is provided from other countries ...more »
Lets say you publish the number of cases cleared each month in 2009 but later learn that the numbers for some months were understated due to a server problem. If someone has already downloaded and is using the defective dataset, how are they notified that they need to download the dataset once more in order to have accurate data?
This entry is a consensus recommendation of seven organizations that work on government transparency of which OpenTheGovernment.org is one.
Consider making some of the datasets available as feeds that are constantly up to date, rather than as static datasets that are pulled down and then reposted on an occasional basis.
Government institutions may try to briefly surface their data on Data.gov to earn points on the "open data dashboard", only to later take them down. I believe this is disingenuous, and clearly goes against the spirit of the Open Government Directive. The case I'm specifically referring to is the data tracking Broadband Stimulus funding via the BIP/BTOP programs. The data available at the following link has not been ...more »
The federal government is an information-intensive organization and it is imperative that data within and across the federal government be well managed. A central repository of data about data—also referred to as meta data—can be an effective data-management tool. However, a repository, as proposed here, is not to be confused with a data dictionary that merely gives definitions of data. A repository can be used to manage ...more »
All government agencies that provide data to the public should offer the data in standardized forms. This would include both the format (CSV, XML, SQL Server, Access, Oracle, etc.) and structure. For example, the State Energy Data System at EIA provides its annual data update in CSV, with columns for the variable name, location, date, and value. This would likely need to be modified for various applications, but if ...more »
Make sure the site is 508 accessible since its managed by GSA who are responsible for 508!
Many developers would like to mine the data at a large scale which may include a majority or all of the available data. It is technically possible to make a web scraper or mirror the site but this would be a long, slow process that would consume a great deal of bandwidth both for the client and for the servers housing the data. What would be an optimal way to pull down all the data would be to distribute it via Bit Torrent. ...more »
When searching for interesting datasets it can be very frustrating to see identical datasets (except by year) cluttering the search results. To me, this seems like just an attempt to inflate the counts by treating separate years or geography as separate datasets. Here is my proposed simple rule: If the dataset metadata fields (i.e. like "Coverage Date") are the same, it is the SAME dataset. Just because the instance ...more »
Mahsup URL: http://salary-range-career-industry.findthebest.com/ We used a database from the U.S Bureau of Labor Statistics. We wanted to make a searchable salaries database. We formatted the database along the top of the screen, jobs are listed by categories including Job Title, Industry Title, Total Employment, Hourly Mean Pay and Annual Mean Salary. For example, if you want to sort jobs in order of average salaries, ...more »
Actively market data.gov to school systems as a tool for education. You need a certain level of sophistication and familiarity to use these datasets effectively. Increasing familiarity with students through curriculum integration is a high impact way to increase the pool of people using these data sets now and in the future.
As additional datasets become exposed through RDF, and described through RDFS/OWL-based ontological definitions, I would like to see the ability to mashup DATA.GOV datasets with the rest of the Linked Open Data cloud, using SPARQL.
Government/Public Sector Practice
As currently written, the CONOP only addresses internal activities (means) and doesn't identify the outcomes (ends) that would result from successful implementation of Data.gov. In paragraph 1 the CONOP states "Data.gov is a flagship Administration initiative intended to allow the public to easily find, access, understand, and use data that are generated by the Federal government.", yet there is no discussion about "what ...more »
I belive you need to have a page that shows users how to use the site and how to perform searches. I consider myself pretty tech savvy but I could not easily determine what to do to get the data.
I would like to recommend that the Data.gov dashboard be developed (aligned) with the Federal Enterprise Architecture (FEA) Business Reference Model (BRM) and Service Component Reference Model (SRM). The BRM provides a view of the federal government’s core lines of business (the products and services it delivers to its citizens, the private sector, and other government agencies), and the SRM classifies the internal ...more »
i would love to see data.gov partner with state and local entities to produce up to date crisis management data for state and local emergency management responders... one goal of data.gov seems to be to enrich/better the lives of citizens and enable them to make more informed/educated decisions within their lives...since most issues within citizens lives lie with the heart of state local entities, it seems partnering ...more »
I suppose this isn't a terribly common problem, but not all of us grew up with this technology, or have been able to afford or access it too long between the place of residence not having acess to the web or the costs of equipment being too high for limited budgets. I can't speak for all, but sometimes the maze can be overwhelmingly confusing, trying to find what it is you're looking for, and then understand what it ...more »
The cost of poor information is HUGE; over 50% of the Data warehouse projects fail due to poor information. There is no way any agency will provide quality information without producing quality database schemas. Very few agencies pay any attention to the quality of the schema and as a result, (a) development and maintenance costs skyrocket, (b) applications are delivered very late, (c) performance is not ideal and (d) ...more »
Just need a lot of basic features for ease of finding info, such as filter, sort, search, for many data sources, including within/without sites and pages etc...
Currently users can rate the data usefulness etc but the rating isn't that meaningful because it isn't related to what the user is looking for. Adding an ability for users to tag datasets with keywords will make it easier for people to run more specific searches. And, if you allow users to associate ratings to keywords the rating system will deliver much richer information.