What is Data Loss Prevention Shelfware?

What is Data Loss Prevention (DLP) Shelfware?

With the huge volume of sensitive data companies must deal with today, it’s no wonder data loss protection (DLP) still tops the list of administrator concerns.

Despite the attention, however, it seems that getting DLP right remains easier said than done.

Even major government agencies, with ample resources with which to secure their information, are still struggling. The unique challenge of protecting against digital data loss was recently brought to the fore by an internal audit of the IRS, one of the biggest holders of Personally Identifying Information (PII) in the world.

According to reports, the IRS has been grappling to get their DLP online for over nine years.

Due to delays and various hurdles, for nearly a decade, the tax authority has paid a third-party contractor for data protection services for both “Data at rest” and “data in use” capabilities it is unable to use.   “The inspector general report estimated the total cost of licensing for the unused capabilities to be $1.5 million over four years, $1.2 million of which was paid out by IRS.”[i] What’s more unbelievable is the audit revealed that the IRS will still be unable to meet a revised June 2020 deadline to get these modules in place!

Government Data Leaking

But beyond the wastefulness of their network security efforts, what was interesting to note were the specific data loss problems the IRS regularly has to deal with. Both IRS and its parent organization the Treasury Department have experienced several incidents over the recent period where employees have been caught leaking taxpayer data. These and other data compromises have drawn the attention of lawmakers who want to know what the agency is doing to stop employees from abusing their access and protect citizens’ private information.

The IRS has been scrambling to find solutions to its data loss debacle. Like many organizations, the Service has considered turning to artificially intelligent tools, hoping it’ll be the magic bullet for their DLP deficiencies.

Unfortunately, the businesses that go the AI route often find they merely traded one problem for a different, more complicated one.

The Cons of Machine Learning

Many cybersecurity solutions and programs with machine learning functions rely on regular expression patterns to function. These algorithms predetermine what “sensitive data” is and decide what controls and safety measures are activated in any given scenario.  This leads to serious issues in identification accuracy.

Markers that are meant to be specific–but in the context of high volumes of data are pretty generic–end up producing mountains of false positives, rendering the DLP coverage essentially useless. Failing to see data in context also leads to false negatives, allowing important files to slip through the cracks.

DLP Powered by Science

Rather than rely on set models, GTB programs regularly analyze data with smart algorithms.   This approach virtually eliminates false positives by sharpening in on relevant data and only real exfiltration threats.  False negatives are also prevented with these methods.  GTB tracks and controls sensitive data even when elements of a file or stream are changed.

GTB’s Security Manager gives networks all of the advantages and protections of smart system tools without any of the inefficiencies of rigid algorithms and unadaptive models.

[i] https://fcw.com/articles/2019/08/27/irs-unused-data-security-software.aspx

Visibility: Accurately, discover sensitive data; detect and address broken business process, or insider threats including sensitive data breach attempts.

Protection: Automate data protection, breach prevention and incident response both on and off the network; for example, find and quarantine sensitive data within files exposed on user workstations, FileShares and cloud storage.

Notification: Alert and educate users on violations to raise awareness and educate the end user about cybersecurity and corporate policies.

Education: Start target cyber-security training; e.g., identify end-users violating policies and train them.

  • Employees and organizations have knowledge and control of the information leaving the organization, where it is being sent, and where it is being preserved.
  • Ability to allow user classification to give them influence in how the data they produce is controlled, which increases protection and end-user adoption.
  • Control your data across your entire domain in one Central Management Dashboard with Universal policies.
  • Many levels of control together with the ability to warn end-users of possible non-compliant – risky activities, protecting from malicious insiders and human error.
  • Full data discovery collection detects sensitive data anywhere it is stored, and provides strong classification, watermarking, and other controls.
  • Delivers full technical controls on who can copy what data, to what devices, what can be printed, and/or watermarked.
  • Integrate with GRC workflows.
  • Reduce the risk of fines and non-compliance.
  • Protect intellectual property and corporate assets.
  • Ensure compliance within industry, regulatory, and corporate policy.
  • Ability to enforce boundaries and control what types of sensitive information can flow where.
  • Control data flow to third parties and between business units.