Pulse


Pulse

By GCN Staff


DARPA to mine 'big code' to improve software reliability

During the past decade information technologies have driven productivity gains that are essential to U.S. economic competitiveness, and computing systems now control a significant portion of the critical infrastructure.

As a result, tremendous public and commercial resources are devoted to ensuring that programs are correct, especially at scale. Yet, in spite of sizeable efforts by developers, software defects remain at the root of most system errors and security vulnerabilities.

To address the predicament, the Defense Advanced Research Projects Agency wants to advance the way software is built, debugged, verified, maintained and understood by combining principles of big data analytics with software analysis.

DARPA said its Mining and Understanding Software Enclaves (MUSE) program would facilitate new ways to dramatically improve software correctness and help develop radically different approaches for automatically constructing and repairing complex software, according to its announcement.

“Our goal is to apply the principles of big data analytics to identify and understand deep commonalities among the constantly evolving corpus of software drawn from the hundreds of billions of lines of open source code available today,” said Suresh Jagannathan, DARPA program manager in the announcement.

“We’re aiming to treat programs—more precisely, facts about programs—as data, discovering new relationships (enclaves) among this ‘big code’ to build better, more robust software.”

Central to MUSE’s approach is the creation of a community infrastructure that would incorporate a continuously operating specification-mining engine, the agency said. This engine would use “deep program analyses and big data analytics to create a public database containing … inferences about salient properties, behaviors and vulnerabilities of software drawn from the hundreds of billions of lines of open source code available today.”

“The collective knowledge gleaned from this effort would facilitate new mechanisms for dramatically improving software reliability and help develop radically different approaches for automatically constructing and repairing complex software,” DARPA said in describing the project.

Posted on Mar 24, 2014 at 9:20 AM0 comments


What's next for predictive analytics?

Predicative analytics, a statistical or data mining approach that determines outcomes using a series of algorithms and techniques used on both structured and unstructured data, has become a technology high in demand.  Although not a new method, its ease of use, inexpensive computing power and organizations increasing amounts of data have driven its adoption.

The technology is being used for retention analysis, fraud detection, medical diagnosis and risk assessment, to name a few. Fern Halper, director of research for advanced analytics at TDWI, highlighted four trends about predicative analytics in a blog post on TDWI website.

Techniques: Although the top three predicative analytics techniques are linear regression, decision trees and clustering, others have become more widely used. These include time series data analysis, which can be applied to weather observations, stock market prices, and machine-generated data; machine learning, which can uncover previously unknown patterns in data; and ensemble modeling, in which predictions from a group of models are used to generate more accurate results.

Open source: Open source solutions enable a wide community to engage in innovation. The R language, a free software environment for data manipulation, statistics and graphics has become one of the most popular open source solutions addressing predicative analytics in the enterprise.

In-memory analytics: In-memory analytics processes data in random-access memory rather than on disk, which lets users to analyze large data sets more effectively.

Predictive analytics in the cloud: The use of the public cloud for analytics appears to be increasing. Organizations are starting to investigate the cloud for business intelligence, and users are putting their big data in the cloud where so can process real-time data as well as running predictive models on extremely large multisource data sets.

Posted on Mar 17, 2014 at 10:41 AM2 comments


USGS to merge National Atlas and National Map programs

The National Atlas of the United States and the National Map will be combined into a single source for geospatial and cartographic information,  the U.S. Geological Survey announced.

The purpose of the merger is to streamline access to information from the USGS National Geospatial Program, which will help set priorities for its civilian mapping role and “consolidate core investments,” the USGS said.

The National Atlas provides a map-like view of the geospatial and geostatistical collected for the United States. It was designed to “enhance and extend our geographic knowledge and understanding and to foster national self-awareness,” according to its website. It will be removed from service on Sept. 30, 2014, as part of the conversion.  Some of the products and services from the National Atlas will continue to be available from The National Map, while others will not, the agency said.  A National Atlas transition page is online and will be updated with the latest news on the continuation or disposition of National Atlas products and services.

The National Map is designed to improve and deliver topographic information for the United States. It has been used for scientific analysis and emergency response, according to the USGS website.

In an effort to make the transition an easy one, the agency said it would post updates to the National Map and National Atlas websites during the conversion, including the “availability of the products and services currently delivered by nationalatlas.gov.”

“We recognize how important it is for citizens to have access to the cartographic and geographic information of our nation, said National Geospatial Program Director Mark DeMulder. “We are committed to providing that access through nationalmap.gov.”

Posted on Mar 14, 2014 at 8:56 AM0 comments


Disorder in the court? Check your beacon

Apple introduced its iBeacon at a developers conference last June, an “indoor positioning system” it described as "a new class of low-powered, low-cost transmitters that can notify nearby iOS 7 devices of their presence."

Like other beacons, Apple’s iBeacon works in the Bluetooth Low Energy frequency range to send  notifications to other devices within a network of “location-aware” beacons mounted on nearby surfaces.

While the technology is new, initial interest has focused on applications in the retail, events or energy savings, but the technology has also spawned ideas for applications in the public-sector community. In a blog on the National Center for State Courts website, court technology consultant Jim McMillan proposed several, including apps for courthouse navigation and public safety.

“It would be great to be able to provide an automatic guide to assist the public in what is called ‘wayfaring’ though a courthouse facility,” McMillan wrote. Or a courtroom beacon system could be set up to locate over-scheduled attorneys trying to keep up with the typical delays and sudden shifts in the docket.

Finally, beacons could help tighten courthouse security, wrote McMillan, “by automatically locating and calling the closest bailiffs or deputies to the courtroom if there is a problem that they could address.”

For other location-aware applications of beacon technology, “our imagination is the only limit,” he said. 

Posted on Mar 13, 2014 at 9:01 AM1 comments


Survey: Organizations seek better data governance tools

Survey: 2013 Data Governance Survey Results, by Rand Secure Data

Key Points: In spite of the fact that organizations expect overall data growth in the next year between 26 percent and 50 percent, with the median amount of data stored between 20TB to 50TB, respondents reported that their data management tools (archiving, backup and ediscovery) don’t meet expectations.

Only 8 percent of organizations with more than 500 TB of data rated their data archiving solutions as excellent, and over 32 percent of respondents overall ranked their data and email archiving solutions as average or poor. The more data an organization has, the lower the solutions were ranked.

Of the 98 percent of respondents who said they have backup solutions in place, 25 percent of them said they were extremely satisfied with their current solution, while just under 40 percent said they are or may be looking to implement or upgrade their backup solution within the next year. Nearly 15 percent more of survey respondents however indicated archiving is more important than backup to their organizations.

Nearly 20 percent of respondents reported they do not have an ediscovery solution in their organization, while only 10 percent with a solution rated it as excellent. Over one quarter rated their solution as poor, and more than half (51 percent) are or may be looking to upgrade their solution in the next year.

Bottom line: The dissatisfaction over data and email archiving systems combined with their importance within organizations signify a need in the market for new and innovative data and email archiving solutions.

Posted on Mar 12, 2014 at 12:19 PM0 comments