Self-paced

Explore our extensive collection of courses designed to help you master various subjects and skills. Whether you're a beginner or an advanced learner, there's something here for everyone.

Bootcamp

Learn live

Join us for our free workshops, webinars, and other events to learn more about our programs and get started on your journey to becoming a developer.

Upcoming live events

Learning library

For all the self-taught geeks out there, here is our content library with most of the learning materials we have produced throughout the years.

It makes sense to start learning by reading and watching videos about fundamentals and how things work.

Search from all Lessons


LoginGet Started
← Back to Lessons
Edit on Github

DLP Strategies and Technologies

What is Data at Rest?

Data Loss Prevention (DLP) is a crucial aspect of modern cybersecurity. This lesson covers key DLP strategies and technologies, focusing on protecting 3 main strategies:

  • Data at rest.
  • Data in motion.
  • Data in use.

What is Data at Rest?

Stored data not currently being processed or transmitted. For example: Files on hard drives, databases, cloud storage

How to protect data at rest?

There are 3 main strategies to protect data at rest:

Encryption

A process of converting data into a coded form to prevent unauthorized access. If anyone tries to access the data, they will see a coded version, they won't be able to understand it.

Tools for encrypting data at rest

When comparing encryption tools for data at rest, consider the following criteria:

CriteriaBitLocker (Windows)FileVault (macOS)VeraCrypt (cross-platform)LUKS (Linux)AWS Key Management Service
Encryption StrengthAES-256AES-256AES, Twofish, SerpentAES, Twofish, SerpentAES-256
Key ManagementTPM-based, AD integrationIntegrated with Apple IDUser-managedUser-managedFully managed by AWS
Performance ImpactLowLowModerateLowMinimal (cloud-based)
Platform CompatibilityWindows onlymacOS onlyWindows, macOS, LinuxLinux onlyCloud-agnostic
Ease of UseHighHighModerateModerateHigh
Integration CapabilitiesStrong with Windows ecosystemStrong with Apple ecosystemLimitedStrong with Linux systemsExtensive AWS service integration
ComplianceFIPS 140-2FIPS 140-2Not certifiedFIPS 140-2Multiple (FIPS 140-2, HIPAA, etc.)
Recovery OptionsRecovery key, AD recoveryRecovery key, iCloud recoveryRescue diskBackup of LUKS headerKey rotation, multi-region keys
Audit and ReportingLimitedLimitedLimitedLimitedComprehensive
CostIncluded with WindowsIncluded with macOSFree (open-source)Free (open-source)Pay-per-use model

Access Controls

A system of restricting access to data based on user roles and permissions. This ensures that only authorized individuals can view or modify sensitive information.

Tools for implementing access controls for data at rest

When comparing access control tools for data at rest, consider the following criteria:

CriteriaActive Directory (Windows)OpenLDAP (cross-platform)IAM (AWS)Azure AD (Microsoft Cloud)Okta (Cloud-based)
Authentication MethodsKerberos, NTLMLDAP, SASLMulti-factorMulti-factor, SAMLMulti-factor, SAML
Authorization GranularityHighModerateHighHighHigh
Platform CompatibilityWindows-centricCross-platformAWS servicesMicrosoft ecosystemCloud-agnostic
ScalabilityGoodModerateExcellentExcellentExcellent
Single Sign-On (SSO)YesLimitedYesYesYes
ComplianceHIPAA, SOC 2Depends on implementationMultiple (HIPAA, PCI DSS, etc.)Multiple (GDPR, HIPAA, etc.)SOC 2, ISO 27001
Audit and ReportingComprehensiveLimitedComprehensiveComprehensiveComprehensive
Integration CapabilitiesStrong with WindowsModerateExtensive AWS integrationStrong with Microsoft servicesExtensive third-party integrations
User ExperienceGoodBasicGoodGoodExcellent
CostIncluded with Windows ServerFree (open-source)Pay-per-useSubscription-basedSubscription-based

Data Classification

A process of categorizing data based on its sensitivity and importance to the organization. This helps in applying appropriate security measures and access controls to different types of data.

Tools for implementing data classification

When comparing data classification tools, consider the following criteria:

CriteriaMicrosoft Information ProtectionTitus Classification SuiteBoldon James ClassifierVaronis Data Classification EngineBigID Data Intelligence
Classification MethodsContent-based, context-basedRule-based, machine learningUser-driven, automatedContent-based, behavior-basedAI/ML-based, pattern recognition
Supported File TypesOffice, PDF, imagesWide range of file typesOffice, CAD, PDFStructured and unstructured dataStructured and unstructured data
Integration CapabilitiesStrong with Microsoft ecosystemExtensive third-party integrationsGood with various DLP solutionsStrong with Windows file systemsBroad data source support
Automation LevelHighHighModerate to HighHighVery High
User InterfaceIntuitiveUser-friendlyCustomizableAdmin-focusedModern and intuitive
ScalabilityExcellentGoodGoodExcellentExcellent
Reporting and AnalyticsComprehensiveDetailedGoodAdvancedIn-depth analytics
Compliance SupportGDPR, CCPA, HIPAA, etc.Multiple regulationsCustomizable for various standardsGDPR, CCPA, HIPAA, etc.Comprehensive privacy regulations
Cloud SupportNative cloud integrationCloud and on-premisesCloud and on-premisesHybrid environmentsCloud-native, multi-cloud
CostSubscription-basedLicense-basedLicense-basedSubscription-basedSubscription-based

Data in Motion

Data in motion refers to information that is being transmitted over a network, such as through email, file transfers, or web traffic.

Strategies for securing data in motion

When comparing strategies for securing data in motion, consider the following criteria:

CriteriaSSL/TLSIPsec VPNSFTPHTTPSNetwork Segmentation
Encryption StrengthStrong (AES)Strong (AES, 3DES)Strong (AES)Strong (AES)N/A (complementary)
Protocol LevelTransportNetworkApplicationApplicationNetwork
AuthenticationCertificate-basedPre-shared key, certificatesPassword, key-basedCertificate-basedN/A
IntegrityYesYesYesYesN/A
Performance ImpactLowModerateLowLowLow
Ease of ImplementationModerateComplexEasyEasyModerate
CompatibilityWide supportMost devicesMost systemsWeb-basedNetwork infrastructure
Use CasesWeb, email, appsSite-to-site, remote accessFile transfersWeb applicationsInternal network protection
CompliancePCI DSS, HIPAAHIPAA, GDPRPCI DSS, HIPAAPCI DSS, HIPAAPart of defense-in-depth
CostLow to moderateModerate to highLowLowModerate to high

Tools for implementing data in motion protection

When comparing tools for protecting data in motion, consider the following criteria:

CriteriaCisco AnyConnectOpenVPNWinSCPLet's EncryptCisco ASA
TypeVPN ClientVPN SolutionSFTP ClientSSL/TLS Certificate AuthorityFirewall/VPN
Encryption MethodsAES-256OpenSSL (various)AES, 3DESRSA, ECDSAAES, 3DES
PlatformsWindows, macOS, Linux, iOS, AndroidCross-platformWindowsAny web serverNetwork appliance
Ease of UseUser-friendlyModerateUser-friendlyEasy to implementComplex
ScalabilityHighModerateN/A (client-side)HighHigh
IntegrationCisco ecosystemFlexibleStandaloneWeb serversCisco ecosystem
ComplianceFIPS 140-2ConfigurableSupports complianceWidely acceptedFIPS 140-2
CostCommercialFree/CommercialFreeFreeCommercial
Additional FeaturesPosture assessmentCustom scriptsFile synchronizationAutomatic renewalsIntrusion prevention
SupportEnterprise-gradeCommunity/CommercialCommunityCommunityEnterprise-grade

Data in Use

Data in use refers to information that is actively being processed, accessed, or manipulated by applications or users. This includes data that is currently in a computer's RAM, CPU cache, or being displayed on a screen. For example, when a user opens a sensitive financial spreadsheet and is actively editing the cells, that data is considered "in use." Similarly, when an application is processing credit card information during a transaction, that data is also in use. Protecting data in use is crucial because it's often in its most vulnerable and exposed state during active processing.

Notable Case for Data in Use: Meltdown and Spectre Vulnerabilities

A famous case where data in use was potentially breached involves the Meltdown and Spectre vulnerabilities, discovered in 2018. These vulnerabilities affected nearly all modern processors, including those from Intel, AMD, and ARM.

  • Meltdown: Allowed malicious programs to access higher-privileged parts of a computer's memory.
  • Spectre: Tricked applications into accessing arbitrary locations in memory.

Impact:

  • These vulnerabilities could potentially allow attackers to read sensitive data in use, such as passwords, encryption keys, and other confidential information directly from the processor's memory.
  • The vulnerabilities affected data that was actively being processed by the CPU, making it a clear example of a threat to data in use.

Significance:

  • These vulnerabilities were particularly severe because they were hardware-level issues, affecting billions of devices worldwide.
  • They highlighted the importance of securing data not just at rest or in motion, but also while it's being actively processed by the CPU.

Mitigation:

  • Addressing these vulnerabilities required a combination of hardware updates, operating system patches, and in some cases, changes to application code.
  • The incident led to increased focus on hardware-level security and the development of new processor designs with enhanced security features.

This case underscores the critical importance of protecting data in use and the complex challenges involved in securing information at the hardware level.

Strategies for securing data in use

When comparing strategies for securing data in use, consider the following criteria:

CriteriaScreen Capture PreventionCopy-Paste RestrictionsWatermarkingApplication SandboxingMemory Encryption
Protection LevelModerateModerateLow to ModerateHighVery High
User Experience ImpactModerateHighLowLow to ModerateLow
Implementation ComplexityModerateLowLowHighVery High
EffectivenessGood for visual dataGood for text dataGood for tracingExcellent for isolationExcellent for sensitive data
CompatibilityOS-dependentApplication-specificWide compatibilityOS and app dependentHardware-dependent
Bypass DifficultyModerateLow to ModerateLowHighVery High
Performance ImpactLowLowLowModerateModerate to High
Compliance SupportHIPAA, GDPRPCI DSS, HIPAACopyright protectionHIPAA, GDPR, PCI DSSFIPS 140-2, HIPAA
CostModerateLowLowModerate to HighHigh
ScalabilityGoodExcellentExcellentModerateModerate

Tools for implementing data in use protection

When comparing tools for protecting data in use, consider the following criteria:

CriteriaSymantec DLPMcAfee DLP EndpointDigital GuardianCrowdStrike FalconForcepoint DLP
Screen Capture PreventionYesYesYesLimitedYes
Copy-Paste ControlYesYesYesNoYes
WatermarkingYesLimitedYesNoYes
Application ControlYesYesYesYesYes
Memory ScanningYesLimitedYesYesLimited
OS CompatibilityWindows, macOSWindows, macOS, LinuxWindows, macOS, LinuxWindows, macOS, LinuxWindows, macOS
Cloud IntegrationYesYesYesNativeYes
User Behavior AnalyticsYesLimitedYesYesYes
Incident ResponseBuilt-inBuilt-inAdvancedAdvancedBuilt-in
Deployment OptionsOn-prem, CloudOn-prem, CloudOn-prem, Cloud, HybridCloud-nativeOn-prem, Cloud

Integrating DLP into Existing Infrastructure

Network Integration

Network integration for DLP involves two key aspects.

  • Network Gateways: First, it requires implementing DLP solutions at network gateways, which are the entry and exit points of network traffic. This allows for monitoring and controlling data as it moves in and out of the organization's network.

  • Network Security Tools: Second, it involves integrating DLP with existing network security tools such as firewalls and proxy servers. This integration enables a more comprehensive approach to data protection by combining the traffic filtering capabilities of firewalls and proxies with the data-centric controls of DLP systems. Together, these measures create a robust defense against unauthorized data exfiltration and help ensure that sensitive information remains within the organization's secure network boundaries.

Endpoint Integration

Similar to how Endpoint Detection and Response (EDR) works, endpoint integration in Data Loss Prevention (DLP) involves the installation of specialized software agents on individual user devices such as computers, laptops, and mobile devices. These agents serve as local sentinels, continuously monitoring activities on the device and enforcing predetermined security policies. They track actions like file transfers, copy-paste operations, and application usage, ensuring that sensitive data is not mishandled or leaked.

When a user attempts an action that violates the established DLP policies, the agent can intervene in real-time, either by blocking the action outright or by issuing a warning to the user. This approach allows organizations to extend their data protection measures beyond the network perimeter, safeguarding sensitive information even when devices are used off-site or disconnected from the corporate network.

Cloud Integration

Cloud integration in Data Loss Prevention (DLP) involves extending data protection measures to various cloud. It typically involves using API-based connections to link DLP systems with cloud platforms, allowing organizations to apply consistent data protection policies across their on-premises and cloud environments.

Application Integration

Application integration in the context of Data Loss Prevention (DLP) involves incorporating DLP controls directly into custom-built applications and integrating them with commonly used productivity suites. For custom applications, DLP controls can be embedded during the development process, allowing for tailored protection mechanisms that align with the specific data handling requirements of the application.

When it comes to widely used productivity suites like Microsoft 365 or Google Workspace, DLP integration typically involves leveraging built-in DLP features or using third-party solutions that seamlessly connect with these platforms.

Best Practices for DLP Implementation

  1. Start with a data discovery and classification exercise
  2. Develop clear, enforceable policies
  3. Implement DLP in phases, starting with critical data
  4. Regularly review and update DLP rules and policies
  5. Provide user training and awareness programs
  6. Monitor DLP performance and adjust as needed

By understanding and implementing these DLP strategies and technologies, organizations can significantly reduce the risk of data breaches and unauthorized data exposure.