Skip to content

MSP Breaches: From Isolated Incidents to Supply-Chain Crises

An ISO 27001‑aligned incident response framework helps your MSP treat incidents as portfolio‑level risk, not isolated tickets. By designing once for your own platforms and multi‑tenant services, you can understand blast radius, coordinate response across customers and evidence your actions for auditors and regulators. This information is general and does not constitute legal or regulatory advice, and you should obtain professional advice for specific legal or regulatory questions.

Preparation feels invisible until the one day it becomes the only thing that matters.

Why MSP incidents behave like supply‑chain failures

MSP incidents behave like supply‑chain failures because compromise of one shared tool can cascade across many customers simultaneously. When attackers abuse remote management, identity or backup platforms, they gain leverage over dozens of tenants at once. A robust ISO 27001‑aligned framework forces you to analyse that blast radius up front and plan how you will detect, contain and recover from platform‑level events, rather than treating each alert as an isolated problem.

For a traditional single organisation, a compromised server or phishing incident usually affects one environment and one management chain. As an MSP, your reality is different. A single weakness in remote monitoring software, backup infrastructure or identity tooling can expose dozens or hundreds of customers at once.

A majority of organisations in the 2025 ISMS.online survey reported being impacted by at least one third‑party or vendor‑related security incident in the past year.

Real‑world examples include widely reported cases such as the Kaseya VSA ransomware attack, where attackers compromised a remote management platform and pushed malicious code to many MSP tenants in a single action, or abusing a shared identity service to create privileged accounts across customer estates.

When attackers target MSPs, they often aim for the tools you use to reach client systems. If a remote administration platform or central identity service is compromised, the attacker can deploy malware or create backdoor accounts at scale. This is why you need to think in terms of blast radius: which services, customers and data could be affected if a shared component fails, and how quickly you can identify and contain that spread.

An ISO 27001‑aligned framework pushes you to formalise this thinking. Preparation work includes mapping which services and tools are in scope, who owns them, what would constitute a major incident in each and how incidents in those tools might appear across tenants. A structured ISMS platform such as ISMS.online can help you document this shared tooling, define responsibilities and keep these maps current as your service catalogue evolves.

It also encourages you to log and classify events across all customer environments in a consistent way, so you can spot patterns that indicate a systemic problem rather than treating every alert as a separate issue. Over time, this becomes the difference between catching a platform‑level compromise early and discovering it only after many customers report symptoms independently.

Where visibility gaps undermine “due care”

Visibility gaps undermine “due care” because they leave you unable to reconstruct timelines, prove control operation or show reasonable effort when a multi‑tenant incident occurs. If logs are inconsistent, incomplete or poorly correlated across customers and shared tools, both your technical response and your audit storey suffer, and it becomes harder to demonstrate that you acted responsibly.

Your ability to manage incidents across many tenants depends on the visibility you have into their environments and your own platforms. Limited log retention, inconsistent onboarding and siloed monitoring tools all create blind spots. From an ISO 27001 perspective, those blind spots make it hard to prove that your controls are working or that you have exercised reasonable care when something goes wrong. Logging and monitoring controls in the ISO 27001 annex are designed to reduce this uncertainty by setting expectations for what you capture and retain as part of an information security management system, including specific Annex A controls on event logging, monitoring and incident management in ISO/IEC 27001.

You might, for example, have rich telemetry from some high‑value customers but only basic logs from smaller ones. Or you might collect logs centrally but store them in ways that make it difficult to tie specific events to specific tenants or services. When an incident arises, you then struggle to answer simple but critical questions: when did this start, which systems are affected and how far has it spread?

A sound incident response framework forces you to decide what “enough visibility” means for each service tier and to document it. That includes defining standard log sources, retention periods and correlation rules, and ensuring time synchronisation so timelines remain trustworthy. It also means making conscious choices about where you accept residual risk and clearly recording those decisions, rather than allowing gaps to emerge by accident and only discovering them when the stakes are highest.

The economic case for treating incidents as shared risk

Treating incidents as shared risk makes economic sense because one unmanaged multi‑customer breach can severely damage profitability and may wipe out the gains from many years of margin on affected services. Designing a reusable framework, with standard playbooks and evidence paths, is usually far cheaper than absorbing the cost of a single large‑scale failure, and it supports the kind of governance ISO 27001 expects you to show during reviews and audits. Industry analyses of major cyber incidents, including consulting research such as Gartner’s work on incident impact economics, consistently highlight how recovery, legal and reputational costs from a single large event can far exceed the apparent savings from under‑investing in preparation.

Many MSPs initially feel that large‑scale supply‑chain scenarios are theoretical. Day to day, you may see more password resets, phishing tickets and minor outages than platform‑level compromises. The temptation is to treat those as purely operational nuisances and to tackle improvements in a piecemeal way. However, the economics change when you consider the impact of a single, unmanaged multi‑customer incident that hits shared tools at the core of your services.

A serious event that affects many tenants at once can trigger contractual penalties, prolonged downtime, staff overtime and, in the worst cases, customer loss. It also consumes leadership attention and can draw scrutiny from regulators or insurers. When you compare that to the investment required to design a reusable, ISO 27001‑aligned framework-standardised playbooks, clear roles, centralised evidence capture and regular management review-the business case often becomes clearer and easier to defend to decision‑makers.

By reframing incident response as protection for your entire customer portfolio, not just individual tickets, you build support for systematic improvements. That might include prioritising threat detection coverage on shared platforms, strengthening access control around your own tooling, rehearsing platform‑level response scenarios and reporting on incident trends at a portfolio level so leadership can see the return on that investment.

Learning from recurring multi‑tenant patterns

Recurring multi‑tenant incident patterns are one of your best sources of practical improvement ideas. When you capture root causes and themes across customers, you can harden shared controls, adjust service baselines and refine your incident catalogue in ways that reduce both risk and rework, while also giving you evidence of continual improvement to share with auditors.

Even without headline‑grabbing breaches, your historical incidents hold valuable signals. Recurring misconfigurations, weak backup practices, unpatched remote access or inconsistent onboarding steps can show up across customers. Each of these patterns is both a security risk and a commercial risk: the same underlying issue can produce similar incidents again and again, eroding margins and trust.

In an ISO 27001 context, this is where structured post‑incident reviews and risk treatment come in. Rather than closing incidents once systems are restored, you capture root causes, control failures and lessons learned in a disciplined way. Those findings then feed into your risk register, improvement plans and, ultimately, your services catalogue. For example, you might introduce a minimum hardening baseline for new customers, a standard backup verification service tier or additional monitoring requirements on your own platforms.

MSPs that excel here treat multi‑tenant incidents as signals to strengthen shared controls, not just as isolated problems to fix. Over time, this mindset reduces incident volume and impact, while also giving you credible stories to share with customers about how you have improved your protections based on real‑world experience. It also gives you concrete examples to reference in ISO 27001 management reviews and internal audits, demonstrating that you are learning and adapting rather than standing still.

Moving from firefighting to a framework

Moving from firefighting to a framework means turning improvised heroics into a small set of standard patterns your engineers can apply consistently. When you codify the incident types that matter most and define how they are logged, led and reviewed, you make large‑scale events more survivable and easier to explain to auditors and customers, without losing the ability to use professional judgement.

When every incident is handled as a one‑off emergency, engineers improvise with whatever tools and knowledge they have. That can work in the short term but does not scale. Different analysts take different steps, evidence quality varies and the organisation struggles to show auditors or customers a consistent, governed approach. This is where ISO 27001s emphasis on standardisation, documented procedures and continual improvement becomes a strength rather than a paperwork burden.

A framework approach means defining a small set of standard playbooks for the incident types that matter most to your services-ransomware, business email compromise, cloud account abuse, platform compromise-and making them easy to follow. It also means deciding how incidents will be logged, who will lead, what approvals are needed for major actions and how you will record the outcome in a way that feeds directly into risk and improvement processes.

If you adopt a platform such as ISMS.online to hold your policies, risk records, incident logs and improvements, you gain a single source of truth that supports both operations and audits. Instead of searching through scattered documents and tickets after a major incident, you can point to a coherent management system showing how you prepared, responded and learned, and you can show that this system aligns with the ISO 27001 controls and clauses your certification depends on.

Book a demo


Why Internal‑Only Incident Response Fails in an MSP World

Internal‑only incident response fails in an MSP world because it assumes one network, one hierarchy and one set of obligations. Your reality involves many customers, shared tools and overlapping regulations, so your process must be designed for multi‑tenant, shared‑responsibility incidents rather than single‑organisation outages. An ISO 27001‑aligned approach helps you surface those assumptions, adjust them and then prove how they work in practice.

One‑organisation assumptions vs multi‑tenant reality

Internal‑only plans fail because they assume you own every asset, control every user and can convene decision‑makers inside one organisation. As an MSP, you coordinate activity across many customers, tools and time zones, and incidents often straddle your platforms, customer networks and upstream cloud services. Your incident design has to reflect that complexity, not hide it behind a single‑company playbook or informal habits.

Most legacy incident plans were written for in‑house IT teams. They assume you own all assets, control all users and can convene the right stakeholders quickly. They also tend to rely on a single ticketing system and informal communication-conference calls, chat threads, email chains-that may work when there is only one business to involve and a narrow set of decision‑makers to satisfy.

As an MSP, you rarely have that luxury. You may be supporting tens or hundreds of customers, each with their own policies, contacts and expectations. Your teams work across time zones and tools, from professional services automation platforms to remote monitoring and management suites and multiple security products. Incidents can start in your environment, in a customer network or inside a third‑party cloud service, and often require coordinated action and clear hand‑offs between organisations.

An ISO 27001‑aligned process acknowledges this complexity. It encourages you to define scope clearly (what is covered, what is out of scope), to document interfaces with external parties and to map out how incidents move through your organisation and your customers’ organisations. That structure makes it easier to scale, train new staff and demonstrate control, while also providing a foundation for more explicit shared‑responsibility models later on.

Coordination failures as a design problem

Coordination failures in MSP incidents are usually design problems, not individual mistakes. If you do not define who leads triage, who declares major incidents or who talks to customers and regulators, you guarantee confusion when a serious event hits several tenants at once, even if your people are skilled and well intentioned.

If you think back to recent complex incidents, you may recognise patterns: duplicated investigations between your team and a customer SOC, mixed messages going to business stakeholders, delays in communication with cloud providers or confusion over who should notify regulators. These are not just execution issues; they are symptoms of a process that was not designed for shared responsibility or tested against realistic multi‑party scenarios.

ISO 27001 expects you to define roles and responsibilities clearly, including for outsourced services, through requirements on organisational roles, responsibilities and authorities and supplier relationships in the main clauses and Annex A of ISO/IEC 27001. For an MSP, this translates into explicit agreements about who leads incident triage, who has authority to declare a major incident, who handles external communication and how handovers occur. Simple responsibility matrices and escalation paths are not bureaucracy for its own sake-they are a way to reduce chaos when time and trust are under pressure.

By addressing these coordination gaps in your framework, and revisiting them after major incidents or exercises, you can reduce mean time to respond, avoid duplicated work and limit the risk of inconsistent statements. That makes life easier for your engineers, more reassuring for customers and more defensible in audits that probe how multi‑party incidents are actually handled.

Why ticketing workflows are not a full incident framework

Ticketing workflows are not a full incident framework because they track work items but rarely express detection logic, decision thresholds or learning. ISO 27001 expects you to define how incidents are identified, classified, escalated and reviewed, and most ticket queues simply cannot show that bigger picture on their own, even when you configure fields and priorities carefully.

It is tempting to assume that because you have ticket queues, priorities and SLAs, you already have an incident response framework. In reality, ticketing tools are only one part of the storey. They tell you that something is being worked on, but they rarely capture the full context of detection, decision‑making, communication and learning that ISO 27001 cares about when it assesses the maturity of your ISMS.

A robust framework specifies how incidents are identified and classified, which thresholds trigger escalation, what information must be captured and which actions must be taken before closure. It also describes how related incidents across customers will be correlated, how evidence will be stored and how post‑incident reviews will feed back into your risk and control environment. These elements sit above any individual tool and give auditors confidence that you are not relying purely on ad hoc effort.

You can certainly implement much of this inside your existing tools. For example, you can add specific fields, workflows and approval steps to your professional services automation platform and integrate it with security tools. However, you still need an overarching design that ties those tool‑level configurations back to documented policies and ISO 27001 objectives. Without that, auditors and customers may see only a patchwork of tickets rather than a governed process that you can explain, test and improve.

The human cost of improvised response

Improvised response takes a human toll because it forces engineers to rebuild process, documentation and communication from memory during every incident. Over time that increases error rates and burnout, and makes it much harder to prove to auditors that you follow a consistent approach that respects the limits of human attention and workload.

When your process assumes that analysts can juggle many incidents, manually gather evidence and remember varied customer requirements on the fly, you increase both error rates and fatigue. Engineers end up re‑inventing workflows for each client, searching through old tickets for templates and trying to keep track of different severity scales and reporting obligations in their heads or personal notes.

Over time, this wears people down and makes it harder to keep response quality high. From a management‑system perspective, it also undermines your ability to monitor performance: if every analyst follows a slightly different path, your metrics will be noisy and your improvement efforts unfocused. It becomes difficult to show whether changes in tools or training have actually improved outcomes, because your baseline is inconsistent.

Aligning with ISO 27001 encourages you to respect the limits of human attention. You design workflows that minimise unnecessary variation, automate repetitive steps where possible and provide clear guidance so staff are not forced to improvise during every incident. That makes the work more sustainable, reduces the likelihood of critical details being missed and gives you a stronger platform for training, succession planning and performance review.

Customer communication as a first‑class concern

Customer communication has to be a first‑class concern because even technically competent response can damage trust if tenants feel uninformed or misled. Standardising notifications, updates and reports across customers lets you meet contractual and regulatory expectations while giving account managers clear, consistent messages to share, especially when multiple tenants are affected at once.

Internal‑only plans often treat external communication as an afterthought. In an MSP context, that can be a serious mistake. A technically competent response that leaves customers confused or uninformed can still damage relationships and trigger complaints. When different account managers share conflicting updates, trust erodes quickly and customers may escalate beyond your normal channels.

A multi‑tenant framework should therefore include standard communication patterns: initial notifications, regular status updates, incident summaries and post‑incident reports. It should also account for regulatory deadlines-for example, when customers have obligations to notify authorities about personal data breaches and need timely information from you to do so. These expectations can be reflected in both your internal runbooks and your external service agreements.

Designing these communication flows up front, and connecting them to your internal incident states, helps ensure that customers feel supported and that you meet contractual and regulatory commitments. It also gives your teams clear scripts and expectations when the pressure is on, reducing improvisation and conflict between technical and customer‑facing staff.




ISMS.online gives you an 81% Headstart from the moment you log on

ISO 27001 made easy

We’ve done the hard work for you, giving you an 81% Headstart from the moment you log on. All you have to do is fill in the blanks.




What ISO 27001 Really Demands from Incident Response (for an MSP, Not a Single Enterprise)

For an MSP, ISO 27001 demands that incident response sit inside a governed information security management system, rather than operating as a loose collection of technical runbooks. You are expected to plan, operate, monitor and improve incident processes that cover both your own platforms and the services you provide to customers, and to treat incidents as evidence about how well your controls truly work.

Almost all organisations in the State of Information Security 2025 report list achieving or maintaining security certifications such as ISO 27001 or SOC 2 as a priority.

Incident response as part of the ISMS lifecycle

Incident response belongs inside your ISO 27001 lifecycle because incidents are one of the clearest tests of whether your controls work. You identify risks, implement and operate controls, observe how incidents actually unfold and then adjust design, training and technology based on what you learn, rather than assuming your original design was perfect.

At its core, ISO 27001 requires you to establish, implement, maintain and continually improve an information security management system, as set out in the standard’s main requirements clause for the ISMS in ISO/IEC 27001. Incident management sits inside this cycle. You identify risks that could lead to incidents, select and implement controls, monitor how well they work and improve them based on results and events. Controls on logging, event handling and communication in the standard’s annex are all part of this picture. These include Annex A controls on event logging, monitoring and information security incident management, which together underpin your incident capability in ISO/IEC 27001.

In practical terms for your MSP, this means that when you design your incident response process, you do so with the same discipline as any other control. You define its purpose, scope, interfaces and ownership. You plan how it will be resourced and measured, and how often it will be reviewed in management meetings. You also ensure that incident outcomes feed back into your risk assessments and treatment plans in a traceable, repeatable way.

Because your incidents often span multiple tenants and shared platforms, integration with the ISMS lifecycle is especially important. Platform‑level events may reveal weaknesses in your own tool configuration or access model, while tenant‑level incidents may point to patterns you should address in shared baselines. Treating these signals as formal inputs to your management system helps you strengthen your overall posture rather than merely fixing isolated symptoms.

Translating standards into concrete expectations means turning ISO language about events, incidents and responsibilities into visible policies, procedures and records. Auditors expect to see not only that you understand the principles, but that you have put them into practice in a way that fits your multi‑tenant services and can be explained to staff and customers.

ISO 27001’s annex and related guidance standards, including the ISO/IEC 27035 incident management series and cyber‑incident management resources from bodies such as ENISA’s cyber incident management guidance, set expectations for incident reporting, response and learning. They talk about defining events and incidents, establishing responsibilities, ensuring prompt reporting, documenting actions and reviewing lessons. Controls on logging, event handling and communications all contribute to a coherent incident capability, and related standards in the same family describe typical phases of incident management: preparation, identification, assessment, response and learning.

To make these expectations meaningful for your MSP, you translate them into tangible artefacts and behaviours such as:

  • An incident management policy that defines terms, scope and principles staff recognise.
  • Procedural documents that describe how to handle incident types mapped to your actual services.
  • Role descriptions and responsibility matrices for internal teams, customers and key suppliers.
  • Logging and monitoring requirements, including retention periods and time synchronisation.
  • Templates for incident records and reviews that capture information you will later need.
  • Training that explains when and how staff should report incidents and use your tools.

By mapping each of these back to specific ISO 27001 controls and clauses in supporting documentation, you can show auditors and customers that your implementation is grounded in recognised practice, not just in‑house habits. This mapping also helps you keep your framework aligned as the standard evolves and as you add new services or regulatory obligations.

Scoping decisions and their consequences

Scoping decisions shape what you must evidence because they determine whether incidents in customer environments sit inside or outside your formal management system. If you are not explicit about where the boundary lies, regulators and customers may assume you control more than you have planned for, and you may find yourself unable to provide the level of evidence they expect.

A crucial decision for MSPs is how to define the scope of the ISMS in relation to customer environments. Some choose to include only their own infrastructure and platforms; others extend the scope to cover specific managed services or even entire customer estates. Each approach has implications for incident response, evidence and audit.

If you exclude customer environments from scope, you still need to show how incidents affecting those environments are handled in relation to your services, but you may have more limited evidence obligations. If you include them, you commit to demonstrating a higher degree of control and documentation, which can strengthen customer trust but may demand more effort, more integration work and more careful documentation of shared responsibilities.

Whichever path you choose, it is important to be explicit and consistent. Your incident process, risk treatments and audit narratives should align with the defined scope and be reflected in your statement of applicability. Ambiguity here can lead to uncomfortable questions later, especially if a major incident prompts deeper scrutiny from customers, regulators or certification bodies.

Continual improvement and meaningful metrics

Continual improvement in incident response depends on metrics that actually inform decisions, not vanity numbers. When you track detection, containment and learning in ways that align with your risks and objectives, management reviews become opportunities to strengthen your framework rather than tick‑box exercises, and your incident data becomes an asset rather than a burden.

ISO 27001’s emphasis on continual improvement means you should not treat incident response as “finished” once you have a documented process. Instead, you monitor how it performs, review incidents and near misses and adjust your controls, playbooks and training accordingly. For an MSP, this often means analysing both tenant‑level and platform‑level incidents to see where shared improvements will have the greatest effect.

Rather than tracking only basic numbers, you can define indicators that relate to your objectives and risks-for example, average time to detect incidents across tenants, the proportion of incidents detected by your own monitoring versus reported by customers or the percentage of high‑impact incidents that result in completed post‑incident reviews with documented actions. You might also track notification timeliness against contractual and regulatory commitments and the rate at which agreed improvements are implemented.

These metrics inform your management reviews and can also be used in discussions with customers and auditors to demonstrate maturity. The key is to choose measures that reflect reality and support decisions, rather than statistics that look impressive but do not drive improvement. Once you understand which metrics matter, the next question is who does what in a multi‑party incident and how you coordinate those responsibilities.




From Single‑Org IR Plan to MSP Shared‑Responsibility Model

Moving from a single‑organisation incident response plan to an MSP shared‑responsibility model means making “who does what, when” explicit across your own teams, customers and critical suppliers. An ISO 27001‑aligned framework provides the structure to document these roles, decision points and handovers before a crisis hits, so that you are not negotiating responsibilities in the middle of an outage.

Defining who leads and who supports

Defining who leads and who supports is essential, because multi‑party incidents involving your services, your customers and vendors can stall if everyone waits for someone else to act. A shared‑responsibility model gives your teams and customers a common map of leadership, support and escalation paths that they can follow under pressure.

In many incidents, especially those affecting customer systems, multiple parties must act. You may provide monitoring, triage and technical response; the customer retains responsibility for certain changes or for regulatory notifications; and upstream providers manage parts of the underlying infrastructure. Without a shared map of responsibilities, confusion can slow down response and create disputes about who should have done what, and when.

A practical approach is to build a responsibility matrix that covers common incident scenarios. For each, you outline who detects and declares the incident, who leads technical containment and recovery, who approves high‑risk actions and who communicates with different audiences. You also note dependencies on third parties and how to engage them, including any special escalation routes or response commitments.

This matrix becomes a reference for your internal teams and a communication tool with customers and suppliers. It can be incorporated into policies, runbooks and customer agreements, and revisited after major incidents to see whether it still reflects reality. Over time, it turns abstract “shared responsibility” language into something you can train, audit and refine.

Aligning your shared‑responsibility model with regulatory expectations ensures that customers can meet their notification duties and you can defend your part in the process. Many regimes assume cooperation between controllers, processors and service providers, so your framework should reflect how you support customers’ legal obligations without taking on responsibilities you cannot realistically fulfil.

Data protection laws and sector regulations often assume that controllers, processors and service providers will cooperate on incident handling and notifications. Under frameworks such as the EU’s General Data Protection Regulation, breach notification provisions expect controllers and processors to work together so controllers can meet their duties to notify supervisory authorities and affected individuals within required timeframes, as set out in GDPR Article 33.

By aligning your shared‑responsibility model with these expectations, you reduce the risk of surprises if a regulator asks how a multi‑party incident was handled. You can, for example, specify that you will provide initial technical findings within a defined time window, support root cause analysis and assist with evidence for notifications, while making clear that final legal decisions sit with the customer.

It is worth involving legal and privacy specialists in designing and reviewing this model, so that it accurately reflects contractual and regulatory duties across jurisdictions. Clear design up front reduces friction when real incidents occur and makes it easier to defend your actions if they are later scrutinised in audits, regulatory reviews or insurance assessments.

Extending the model to cloud and SaaS providers

Extending your model to cloud and SaaS providers recognises that many incidents originate in layers you do not fully control. By defining escalation paths, expectations and information flows with these vendors, you avoid improvising critical relationships while customers are waiting for answers and regulators are watching the clock.

Your services probably depend on various cloud and software‑as‑a‑service platforms-identity providers, backup services, security tools, collaboration suites. When incidents originate in those layers, response can be complex: you may need to work with both the customer and the vendor to investigate, contain and remediate. Each party has different levers and obligations, and misalignment can cause delay.

A robust shared‑responsibility model therefore includes escalation paths and expectations for these providers. That might involve knowing how to raise high‑priority tickets, what information to supply, how they will communicate about incidents and what support they will provide with forensics or recovery. You then weave those expectations into your own playbooks so analysts know when and how to involve upstream partners and what to expect from them.

Documenting these relationships helps you demonstrate to auditors and customers that you have not overlooked critical dependencies. It also highlights gaps where you might want to renegotiate terms, seek alternative providers or add compensating controls in your own environment so that you are not wholly dependent on a vendor’s response.

Testing whether the model works in practice

Testing your shared‑responsibility model in practice shows whether the diagrams and matrices actually help people during an incident. Exercises that involve customers and suppliers reveal gaps in contacts, expectations and decision rights before a live incident exposes them, and help you refine both your model and your runbooks.

Even a well‑designed shared‑responsibility model can fail if it remains theoretical. To build confidence, you should test it through exercises that involve all key parties. Tabletop simulations, where you walk through realistic scenarios with customers and suppliers, are particularly useful because they reveal both technical and human issues without the risk of production impact.

In these sessions, you can check whether contact details are current, whether people understand their roles and whether there are any unanticipated bottlenecks. You can also identify differences in expectations-for example, how quickly customers expect to be updated, or how much information suppliers are willing to share. Those insights often prompt small but important changes to contracts, runbooks or escalation paths.

The results of these exercises feed back into your documentation and agreements. Over time, you build a model that has been validated by practice, not just by design, and you gain evidence you can present in ISO 27001 management reviews and internal audits to show that you test and improve your shared‑responsibility arrangements deliberately.




climbing

Embed, expand and scale your compliance, without the mess. IO gives you the resilience and confidence to grow securely.




A Dual‑Domain ISO 27001 Incident Response Framework for MSPs

A dual‑domain ISO 27001 incident response framework treats your MSP’s own platforms and your customers’ environments as distinct domains, governed by one lifecycle and one set of principles. This lets you design once, then reuse and adapt across tenants without confusing who leads in which situations or blurring the line between your responsibilities and those of your customers.

Defining MSP platform incidents and customer incidents

Defining MSP platform and customer incidents separately helps you prioritise the most dangerous scenarios without losing sight of tenant‑specific events. Platform incidents threaten many customers at once and demand top‑tier governance, while customer incidents can highlight patterns that point back to shared weaknesses in your own services and tooling.

In the platform domain, incidents centre on tools and services you operate: remote management platforms, monitoring infrastructure, shared authentication, hosting platforms and internal networks. A compromise here-such as an attacker taking over your remote management platform and pushing malicious agents-can have wide impact, so you treat these incidents as top‑tier events with stronger controls, more senior oversight and closer linkage to risk and business continuity planning.

In the customer domain, incidents occur in networks, systems and applications that you manage on behalf of clients. Some may be contained to one tenant-for example, a single‑tenant ransomware outbreak or misconfigured firewall-while others may reveal weaknesses that also exist elsewhere. For each domain, you define how incidents are detected, who is on point and what thresholds trigger involvement from the other domain. A customer ransomware event might start in the customer domain but become a platform incident if evidence suggests your shared tooling was the entry point.

The lifecycle-prepare, detect, assess, respond, recover, learn-remains the same in both domains. What differs is the scope, stakeholders and specific actions. By expressing these differences explicitly in policies, playbooks and training, you avoid confusion about who leads in which situations and make it easier for auditors and customers to understand how you handle platform‑level versus tenant‑level risk.

Standardising triage and severity across tenants

Standardising triage and severity across tenants allows your analysts to work consistently while still honouring customer‑specific sensitivities. A common classification model underpins portfolio‑wide reporting, service design and regulatory response, and makes it easier to explain your approach to auditors who want to see how you prioritise and escalate incidents.

Analysts working on your SOC or service desk should not have to learn a new classification scheme for every customer. At the same time, customers may have different regulatory obligations and risk appetites. The solution is to design a standard severity and classification model that applies everywhere, then allow controlled per‑customer extensions that are clearly documented.

For example, you might define a small set of incident categories-such as data breach, denial of service, malware infection, account compromise and service disruption-and a severity scale based on impact and urgency. You then agree with each customer how these map to their own internal scales and what additional triggers they may have, such as regulatory thresholds or sector‑specific reporting rules.

This shared model enables cross‑tenant reporting and analytics, because incidents can be compared and aggregated. It also supports consistent service commitments and escalation paths, and it aligns neatly with ISO 27001’s expectation that you define clear criteria and responsibilities for incident handling. When an auditor asks how you distinguish events from incidents or low‑impact from major cases, you can show them a simple model that applies across your portfolio.

Balancing structure and flexibility

Balancing structure and flexibility means giving engineers clear guardrails without scripting every technical move. Your framework should demand certain checks, approvals and records, while leaving space for professional judgement about how to investigate and contain a specific threat in a specific customer context.

A common concern is that a formal framework will be too rigid for real‑world incidents. To avoid this, you design guardrails rather than scripts. Guardrails specify the minimum steps that must happen-such as logging, initial assessment, classification, approvals for disruptive actions and documentation of outcomes-but leave room for engineers to choose technical tactics appropriate to the situation and the tools available.

For example, a playbook may say that when a potential account compromise is detected, you must verify the alert, identify affected systems, decide whether to reset credentials or block access, preserve relevant logs and inform the customer. It does not need to dictate exactly which commands or tools to use to perform those checks, as long as those methods are consistent with your control environment and evidence needs.

This balance respects professional judgement while still delivering the consistency and evidence that ISO 27001 and customers expect. It also helps you adapt as tools and threats change, because you update guardrails and examples rather than rewriting deeply scripted procedures every time you switch a product.

Making the framework visible and usable

Making the framework visible and usable ensures it does not live only in policy documents. When you present the dual‑domain lifecycle through diagrams, training and embedded runbooks, analysts and customers can see how incidents flow and where they fit in, and your incident processes move from theory to daily practice.

A dual‑domain framework only adds value if people can understand and apply it. Visual representations such as swimlane diagrams, state transitions or high‑level flow charts can help. They show, at a glance, how incidents move between MSP and customer lanes, when key decisions are made and where communication occurs, so that staff are not left guessing during a crisis.

These visuals can be included in training material, shared with customers as part of onboarding and referenced in audits. They also help new staff grasp how the pieces fit together quickly, which is especially valuable in high‑turnover environments. Combined with clear documentation and embedded runbooks in your tools, they turn the framework from a static document into something that is actually used and refined.




Making It Real: Workflows, Runbooks, and Evidence for Audits

Making your incident framework real means wiring it into the workflows, runbooks and records your teams use every day. When incident handling, learning and evidence capture all follow the same patterns, you can respond faster, reduce mistakes and give auditors the artefacts they expect from an ISO 27001‑aligned ISMS, instead of scrambling to reconstruct events after the fact.

Choosing high‑value scenarios for playbooks keeps your framework focused on incidents that can hurt many customers or your core services. By standardising a handful of realistic, high‑impact cases, you avoid both dangerous gaps and overwhelming teams with low‑value detail that they cannot remember or maintain.

You do not need a unique playbook for every conceivable event. Instead, you identify scenarios that are both likely and high impact for your customers and services. Common examples include ransomware or other destructive malware, business email compromise, cloud account abuse and compromise of a remote management platform. These align naturally with the threats ISO 27001 expects you to consider in your risk assessments.

For each scenario, you define a runbook that follows your standard lifecycle. It sets out who owns initial triage, what checks to perform, which containment options exist, how to involve the customer and what to document along the way. The language should be tool‑agnostic enough that it remains valid if you change suppliers, while still practical enough for analysts to follow during a stressful incident.

Over time, you can refine these playbooks based on real events. Post‑incident reviews highlight steps that were missed or unnecessary, communication that caused confusion or controls that did not work as expected. You then update the runbooks and share changes with relevant staff, turning hard‑won experience into institutional memory instead of relying on individuals to remember what worked last time.

Automating evidence capture and making audit‑readiness normal

Automating evidence capture makes audit‑readiness normal because incident records are created as a by‑product of doing the work, not as a separate, painful task. When tickets, logs and post‑incident reviews line up, you can show auditors a coherent storey without last‑minute reconstruction or guesswork about what really happened.

A frequent pain point in audits is assembling incident evidence. If you rely on manual note‑taking and ad hoc document storage, you may find yourself piecing together timelines from emails, chat logs and screenshots. That is stressful, time‑consuming and prone to gaps, particularly when staff have changed roles since the incident occurred.

To avoid this, you build evidence capture into your workflows. For example, you can ensure that every significant incident has a dedicated record in your case management tool, with fields for detection source, classification, affected services, decisions, approvals and communication summaries. You can integrate these records with logs from monitoring tools so that the technical evidence and the narrative stay linked and can be retrieved together.

An ISMS platform such as ISMS.online can then act as the repository for policies, risk registers, incident logs, corrective actions and reviews. When auditors or customers ask how you handle incidents, you can show them a coherent set of records that align with your ISO 27001 scope and controls, rather than improvising. This also helps internal management reviews, because decision‑makers can see patterns, track progress and prioritise improvements based on real data.

Embedding runbooks into tools analysts already use

Embedding runbooks into tools analysts already use makes it easier to follow the framework under pressure. When guidance sits one click away inside tickets, chat or automation platforms, your ISO‑aligned process becomes the default, not an optional extra, and analysts are more likely to follow it consistently.

Runbooks are most useful when they are close at hand. If they live only in a document library that no one opens, analysts will revert to memory and improvisation. To counter this, you integrate guidance into the tools people already use to manage incidents, so they encounter the right prompts at the right time.

That might mean adding quick‑link fields in tickets that open relevant playbooks, using checklists within your professional services automation platform, embedding decision trees in your chat or collaboration platform, or wiring your security automation tool to present recommended actions for certain alert types. The goal is to make the ISO‑aligned path the path of least resistance, so that the easiest way to work is also the most compliant and evidence‑friendly.

When your tools reinforce your framework, adoption improves and you get more consistent data. This in turn strengthens your ability to analyse incidents, refine playbooks and demonstrate control. It also reduces the cognitive load on engineers, who no longer have to remember every step unaided in the middle of a complex investigation.

Testing that your evidence model survives real incidents

Testing that your evidence model survives real incidents ensures that the records you rely on for audits and insurance claims actually get created. Exercises should check not only technical response but also whether timelines, decisions and approvals are captured in ways a third party could understand and trust months or years later.

Planned exercises and simulations are invaluable here. You can run tabletop sessions where teams walk through a scenario step by step, or more technical exercises where red‑team activities generate real alerts. In either case, you explicitly include evidence capture as part of the objectives, not just technical containment.

During these exercises, you not only watch how quickly and effectively teams respond, but also review the resulting records. Did the right tickets get created? Were fields filled consistently? Is there enough information to reconstruct decisions and actions? Would an external party, such as an auditor, insurer or regulator, understand what happened and why you chose particular actions?

By treating these questions as part of your exercise objectives, you improve both operational readiness and audit‑readiness. The lessons you learn feed back into your runbooks, workflows and training programmes, and they give you concrete examples to reference in ISO 27001 internal audits and management reviews when you discuss the effectiveness of your incident controls.




ISMS.online supports over 100 standards and regulations, giving you a single platform for all your compliance needs.

ISMS.online supports over 100 standards and regulations, giving you a single platform for all your compliance needs.




Multi‑Tenant SLAs, Contracts, and Regulatory Alignment

In a multi‑tenant world, your incident response framework has to align with SLAs, contracts and regulatory duties, or you risk over‑promising to Sales while Legal, Privacy and Security try to enforce a different reality. ISO 27001 gives you a structured way to make these expectations explicit, evidence‑backed and testable through internal audit and management review.

Managing third‑party risk and tracking supplier compliance was cited as a top challenge by 41% of organisations in the 2025 ISMS.online survey.

Encoding the framework into SLAs and agreements

Encoding your framework into SLAs and agreements turns internal intent into external promises that Sales and Legal can stand behind. Clear definitions of incidents, severities, response times and cooperation make it easier to defend your position when customers or insurers scrutinise a major event and ask how your commitments are grounded in governance.

Shared‑responsibility models and incident processes are only as strong as the agreements that support them. When contracts are vague about response times, notification triggers and cooperation, misunderstandings are likely when incidents occur. To avoid this, you translate key elements of your framework into customer‑facing documents that use clear, non‑technical language and align with your operational capabilities.

The 2025 ISMS.online survey highlights that common supplier requirements now include ISO 27001, ISO 27701, GDPR, Cyber Essentials, SOC 2 and emerging AI‑governance standards.

That includes defining what counts as a security incident, how severities are determined, what response targets apply, how and when you will notify customers and what you expect from them in return. It also covers how evidence will be shared, how joint investigations will be conducted and how disputes will be escalated. These commitments should reflect your ISO 27001 controls and be informed by data from past incidents and exercises.

By basing this language on your ISO‑aligned process, and reviewing it regularly through management review and internal audit, you ensure that sales promises, legal obligations and operational capabilities are in sync. This alignment reduces the risk of over‑commitment and gives you a clearer storey to tell in tenders and due‑diligence exercises, where customers increasingly compare providers on how well their SLAs reflect reality.

Reflecting regulatory and insurance requirements

Reflecting regulatory and insurance requirements in your framework ensures that customers’ legal teams and privacy officers can use your incident support to meet their obligations. When you explain who will provide which information, and how quickly, you reduce the risk of missed deadlines or policy disputes and demonstrate that you understand your role in wider compliance chains.

About two‑thirds of organisations in the 2025 ISMS.online survey say the speed and volume of regulatory change are making security and privacy compliance harder to sustain.

Many of your customers operate under regulations that specify how quickly certain incidents must be reported to authorities or affected individuals. For example, under GDPR Article 33, controllers are expected to notify the relevant supervisory authority of certain personal data breaches without undue delay and, where feasible, within 72 hours of becoming aware of them. Cyber insurance policies may also impose conditions on incident response, such as maintaining a tested plan or engaging specific types of specialists when particular thresholds are reached. Regulators and insurers increasingly ask how third‑party and supply‑chain incidents are handled, not just how internal processes operate, as reflected in industry analyses such as Aon’s cyber insurance trends reporting.

Your framework should account for these external drivers. In contracts and playbooks, you can clarify which notifications you support directly, what information you will provide and how timelines are coordinated. You can also document how you will cooperate with customers’ legal and compliance teams when they make regulatory decisions, and how you will support evidence for insurance claims if a significant loss occurs.

This clarity benefits everyone. Customers gain confidence that their obligations can be met; you reduce the risk of being blamed for missed deadlines; and insurers and auditors see that you have thought through your role in the broader compliance ecosystem, consistent with ISO 27001’s emphasis on understanding interested parties and external requirements.

Designing service tiers without breaking the framework

Designing service tiers without breaking the framework lets you offer commercial choice while maintaining one coherent incident lifecycle. The core process stays consistent; higher tiers add depth of monitoring, investigation and reporting rather than completely different ways of working, so that metrics and lessons remain comparable.

A practical solution is to keep one core framework for all services, with common definitions, lifecycles and evidence requirements. Service tiers then affect the depth of monitoring, the scope of response, the level of reporting and the involvement of specialists, not the existence of the process itself. For example, all customers might benefit from standard classification and communication, while higher tiers receive more proactive containment and richer reporting.

A simple way to think about this is:

Element Core tier (all customers) Higher tier (selected customers)
Classification & scope Standard categories and severity model Same model plus customer‑specific triggers
Monitoring & triage Baseline alerting on agreed services Enhanced telemetry and analyst review
Reporting & learning Standard incident and review summaries Extended reporting, metrics and joint workshops

This approach supports both commercial flexibility and governance. You can still compare metrics across tiers and maintain a single set of management reviews, while offering customers choices that fit their risk appetite and budget. It also makes it easier to show auditors that your service differentiation does not undermine the core controls required by ISO 27001.

Handling cross‑border and multi‑regime incidents

Handling cross‑border and multi‑regime incidents means acknowledging that one event can trigger several legal and regulatory regimes at once. Your framework should make room for customer‑specific legal decisions while committing you to provide timely, accurate technical information across jurisdictions, so customers can meet their obligations without unrealistic expectations about your role.

When you serve customers in multiple jurisdictions, a single incident can trigger overlapping regulatory regimes. A breach affecting a European subsidiary of a global customer might involve both local data protection laws and sector‑specific rules from their home country. Your framework needs to be able to accommodate such complexity and avoid assuming that one set of rules applies everywhere. Financial supervisors such as the UK’s Financial Conduct Authority, in guidance on outsourcing and cloud/ICT arrangements like FG18/5, highlight how issues in cross‑border services can engage multiple regulatory frameworks at once.

You do not have to become a global legal expert, but you should at least ensure that your shared‑responsibility model and playbooks leave room for customer‑specific regulatory decisions. You can, for example, agree that the customer will lead on interpreting laws and drafting notifications, while you commit to providing timely technical details and support in agreed formats and timeframes, regardless of jurisdiction.

By recognising these nuances in advance and documenting them in agreements and runbooks, you avoid making assumptions that could later be judged as careless. You also reassure customers’ legal and privacy teams that you understand the role of third‑party providers in a multi‑regime environment and that your ISO 27001 framework is flexible enough to support their obligations.

Demonstrating that your SLAs are grounded in reality

Demonstrating that your SLAs are grounded in reality reassures customers, auditors and insurers that headline promises are backed by tested processes and real performance data. It is far easier to negotiate favourable terms when you can point to measured outcomes and internal review cycles rather than only policy wording.

Customers, auditors and insurers increasingly ask not just for SLAs and policies but for evidence that they are realistic and tested. Sharing aggregated metrics about incident response performance, summaries of past exercises and examples of improvements made after incidents can go a long way toward building confidence. These materials also demonstrate that you take internal audit and management review seriously.

Because ISO 27001 already expects you to measure and review your controls, you can reuse those mechanisms to support this external assurance. For instance, you might track how often you meet or exceed response targets, how many significant incidents result in completed reviews and how quickly agreed improvements are implemented. You can present these results as part of customer governance meetings, tender responses or due‑diligence processes.

When you can back up your contractual promises with data and documented learning, you strengthen your position in negotiations and build trust. You also give yourself early warning if SLAs are drifting away from what your teams can realistically deliver, so you can adjust either the commitments or the resourcing before problems become public.




Book a Demo With ISMS.online Today

ISMS.online helps you operationalise an ISO 27001‑aligned incident response framework by bringing policies, risks, incidents and improvements together in one management system your MSP teams can use every day. Instead of chasing scattered documents and ticket trails, you can build a repeatable, auditable way to protect customers, model shared responsibility and prove professionalism across your portfolio.

Understanding where you are today

Understanding where you are today gives you a realistic starting point for improving incident response. By mapping your current tools, habits and pain points against a structured ISMS, you can see which strengths to build on, which gaps to close first and how your existing work maps onto ISO 27001 expectations without throwing everything away.

A useful first step is to assess where your current incident approach sits on the spectrum from ad hoc reaction to governed framework. You might already have strong elements-experienced engineers, robust tooling, informal playbooks-but lack consistent documentation, metrics or clear links to risk management. A structured conversation or demo can help you see how those pieces could be organised inside an ISMS and how a dual‑domain or shared‑responsibility model would look in practice.

During that discussion, you can explore how ISMS.online represents incident processes, risks, controls and actions. You can also test your assumptions about scope, responsibilities and evidence. Even if you decide to move forward gradually, having a clearer picture of your starting point makes planning easier and helps you prioritise high‑value changes that your teams and customers will feel quickly.

Piloting a repeatable approach with a focused scope

Piloting a repeatable approach with a focused scope lets you prove value quickly without overwhelming teams. Starting with a handful of high‑impact scenarios and services, you can demonstrate better consistency, evidence and customer communication before scaling up, and you can show that the framework works in the real world rather than just on paper.

Moving to an ISO 27001‑aligned framework does not have to mean overhauling everything at once. Many MSPs find it effective to start with a limited set of services or customers and a handful of high‑impact incident scenarios. They design and implement playbooks, workflows and records for that subset, then expand as confidence grows and results become visible in metrics and audits.

ISMS.online can support this incremental approach. You can build an initial incident management policy, define roles and responsibilities, and create incident records and review templates for your chosen scenarios. As you run real incidents or exercises, you capture outcomes in the platform and adjust your design. The lessons from this pilot then inform how you roll out to the rest of your business, across both platform and customer domains.

Protecting teams from overload while you improve

Protecting teams from overload while you improve is critical if you want long‑term adoption. Clear expectations, practical templates and integrated tooling help engineers spend less time on administration and more time on meaningful incident work, so they see the framework as support rather than extra bureaucracy.

A common concern is that formalising incident response will overwhelm engineers or compliance staff. The opposite can be true if you design carefully. By clarifying expectations, simplifying documentation and providing ready‑made templates and workflows, you can reduce the cognitive load on individuals. Instead of inventing processes on the fly, they follow a known path that fits their tools and daily habits.

Working with ISMS.online, you can see how to align configuration with your existing tools so that people are not forced to duplicate effort. For example, incident records in the ISMS can be linked to tickets or cases in your operational systems, and corrective actions can be tracked alongside other improvements. This reduces friction and helps everyone see how their work fits into the bigger picture of your ISO 27001‑aligned incident framework.

Involving the right stakeholders from the start

Involving the right stakeholders from the start ensures that your incident framework supports Security, Service Delivery, Legal and Privacy teams, rather than surprising them later. Shared workshops anchored in a live ISMS view help each group see how their needs are reflected and how shared‑responsibility and dual‑domain concepts translate into day‑to‑day decisions.

Incident response touches many functions: security leadership, service delivery, legal and compliance, sales and account management, and finance. If you design your framework in isolation, you may later discover conflicts with contracts, pricing or regulatory positions. Bringing the right people into early discussions helps avoid that and accelerates agreement on changes.

An initial workshop or demo that includes these stakeholders can surface priorities, constraints and opportunities. You can explore questions such as which services should be in scope first, how SLAs and security schedules might need to change and what metrics matter to each audience. ISMS.online can act as the shared canvas for those conversations, showing how different needs can be reflected in one management system and how responsibilities are documented and tested.

Building a business case grounded in experience

Building a business case grounded in experience makes it easier to justify investment to executives, boards and owners. Examples from similar MSPs show how better incident response can cut audit effort, reduce losses and strengthen your sales storey, turning abstract risk language into outcomes that business stakeholders recognise.

As you consider investing time and resources in enhancing your incident framework and ISMS, it helps to base your case on concrete examples. Learning from MSPs that have already made this journey-seeing how they reduced audit preparation time, improved response consistency or strengthened their sales storey-can make your own plans more persuasive and less theoretical.

By engaging with ISMS.online, you can access such examples and understand what worked in organisations similar to yours. You can then use that insight to shape your own targets, timelines and success measures. Instead of arguing from theory, you present a path backed by real‑world outcomes and aligned with ISO 27001s expectations for continual improvement and management review.

If you want to move from scattered incident handling to a coherent, ISO 27001‑aligned framework that supports your growth ambitions, a conversation with the ISMS.online team is a practical place to start. You bring your knowledge of your customers and services; they bring experience in building and operating information security management systems. Together, you can design an approach that fits your business, supports your dual‑domain and shared‑responsibility models and stands up to scrutiny when it matters most.

Book a demo



Frequently Asked Questions

How does an ISO 27001‑aligned incident response framework actually work for an MSP?

An ISO 27001‑aligned incident response framework gives your MSP one consistent way to handle incidents across your own platforms and every managed customer.

How does this framework sit inside an ISMS for a multi‑tenant MSP?

In ISO 27001, incident response is a core part of your information security management system (ISMS), not a bolt‑on runbook. For an MSP that means your ISMS must cover:

  • Your shared platforms and tooling (RMM, backup, identity, PSA, SOC stack).
  • Every customer environment that depends on those platforms.
  • The ways one weakness in a shared component can cascade across tenants.

A practical ISO 27001‑aligned framework will normally:

  • Follow a simple lifecycle such as Prepare → Detect → Assess → Respond → Recover → Learn.
  • Tie policies, severity definitions, roles and communication rules to that lifecycle.
  • Require structured records and post‑incident reviews that feed into risk management and management review.

An information security management system such as ISMS.online becomes the place that holds:

  • Your incident policy and severity model.
  • The playbooks you expect engineers to follow.
  • The incident records, risks and improvements that prove the framework is alive.

That single view is what lets you show auditors and customers that you are running incident response systematically at MSP scale, rather than reacting ad‑hoc in tickets and chat tools.

How does this framework shape day‑to‑day incident handling?

In daily operations, the framework gives your teams the same starting pattern for every significant security issue, regardless of where it begins:

  • Capture the alert source, affected tenant and suspected scope.
  • Classify impact and urgency using your shared severity scale.
  • Assign an incident commander and customer‑facing lead.
  • Track actions, approvals and communication in a consistent structure.
  • Close with a short review and feed any new risks or improvements into your ISMS.

Because the lifecycle is repeatable, you stop reinventing the process every time an alert comes in. Over time that consistency is what turns incident response from late‑night fire‑fighting into a managed service you can explain quickly to prospects, auditors and insurers, backed by real examples from your ISMS.online environment.


How is an MSP‑ready incident response framework different from a standard internal IR plan?

An MSP‑ready incident response framework is designed for many customers and shared platforms, while a typical internal plan assumes one organisation with one leadership chain and one risk appetite.

What really changes when the same weakness can affect dozens of tenants?

In a single organisation, most incidents:

  • Are limited to one network or application stack.
  • Are handled by one leadership and legal team.
  • Sit inside one set of contracts and regulations.

In an MSP, the same vulnerability or misconfiguration can show up everywhere:

  • A backup platform issue can undermine restore capability across a whole customer portfolio.
  • A mistake in identity or SSO configuration can expose multiple tenants at once.
  • An RMM agent abused by an attacker can push tools or ransomware into many environments in minutes.

To cope with that reality, MSP‑ready incident response usually introduces:

  • A two‑lane view – every incident is quickly classified as “customer‑specific” or “MSP platform/tooling”, with a clear rule for reclassifying when you discover a shared cause.
  • A shared severity model – high, medium and low mean the same thing to the SOC, service desk, account managers and leadership across all tenants.
  • Contract‑aware handling: – who must be informed, through which channel, and within what time for each type of customer or regulator.
  • Portfolio‑level visibility: – records that show how you handled a single issue across many customers, not just one ticket per tenant.

Many MSPs find it helpful to sketch incidents as two swimlanes – “MSP” and “Customer” – passing through the same phases from preparation to lessons learned, with arrows showing when a local issue reveals a shared‑platform root cause.

How does this change your incident management maturity over time?

Once you adopt an MSP‑specific view, your SOC and engineering teams:

  • Use the same playbooks for both single‑tenant and platform‑wide incidents.
  • Escalate a “customer‑only” issue to an “MSP platform incident” using defined triggers.
  • Produce consistent reports for customers and leadership, regardless of which tenant was initially affected.

That consistency makes it easier to answer tough questions from larger customers and ISO 27001 auditors about shared platforms and supply‑chain risk. When you can show portfolio‑wide incident data and reviews in ISMS.online instead of isolated tickets, you are demonstrating that your operation is designed for multi‑tenant risk, not just for a single internal IT environment.


How should an MSP define who does what with customers and suppliers during incidents?

You protect relationships by defining who does what, and when, across your MSP, your customers and key cloud or SaaS providers before any major incident happens.

What belongs in a shared‑responsibility model that works under pressure?

A shared‑responsibility model is easier to use when it is built around a few realistic scenarios, such as:

  • Ransomware in a single tenant after a phishing attack.
  • A compromise in shared tooling such as RMM, backup or identity.
  • Abuse of a cloud or SaaS account hosted with a major provider.

For each scenario, your model should clarify:

  • Which groups are involved (MSP SOC, customer IT or security, legal/privacy, cloud or SaaS provider).
  • Who is expected to spot an issue first and who can formally declare an incident.
  • Who leads the incident: technical lead, incident commander and customer‑facing lead.
  • Who talks to regulators, affected end‑customers, law enforcement, insurers and the press.
  • When something that starts as “customer‑only” must be treated as an “MSP platform incident” and handled differently.
  • Typical time windows for first triage, customer updates, regulatory notifications and closure.

A simple RACI‑style table with rows like “Detect”, “Declare”, “Contain”, “Notify regulators/customers” and “Post‑incident review”, and columns for MSP, customer and provider roles, is often enough to make expectations clear.

Keeping this shared‑responsibility model in your ISMS alongside statements of work, SLAs and incident playbooks makes it much easier to:

  • Align contracts to how incidents will actually run.
  • Train staff and partners to the same expectations.
  • Show auditors and procurement teams that you have thought through shared responsibility.

When you build the model once in ISMS.online and link it to customer‑specific arrangements, it becomes a reusable asset you can reference at onboarding, during security reviews and whenever a major incident touches several parties.


How can MSPs design incident playbooks that engineers actually follow?

Engineers are far more likely to follow incident playbooks when they feel like light guardrails rather than heavy scripts, and when they are wired directly into the tools your teams already use.

How do you embed playbooks into everyday work without adding friction?

Usable playbooks focus on the essentials: decisions, approvals and evidence. You can weave them into day‑to‑day engineering work by:

  • Linking specific incident runbooks directly from your PSA or service desk ticket types, such as “Security incident – ransomware suspected” or “Security incident – privileged account misuse”.
  • Embedding short checklists in tickets that cover key steps and approvals, for example: “Severity confirmed”, “Customer informed”, “Backups verified”, “Forensic copy captured”.
  • Triggering additional tasks or approval steps automatically when certain conditions are met, such as a particular severity level or involvement of regulated data.
  • Referencing a formal incident record in your ISMS from each operational ticket so all evidence and decisions trace back to one structured entry.

Visually, a typical ticket might include:

  • A selected “Security incident” category.
  • A four‑to‑six‑item checklist aligned with your ISO 27001 process.
  • A link to the relevant MSP playbook stored in ISMS.online.
  • A field that holds the ID of the formal incident record for that event.

When your ISMS and your operational tools reinforce each other this way, engineers spend more time analysing incidents and less time searching for documentation. At the same time, you end up with incident records that satisfy ISO 27001 requirements and customer security teams, rather than a patchwork of screenshots, chats and ad‑hoc notes.

How does this design improve engineering behaviour and learning?

Because the playbooks are close to the work and deliberately lightweight:

  • Engineers stop seeing them as bureaucracy and start treating them as standard safety steps.
  • Handover between shifts and teams gets smoother because everyone works from the same structure.
  • Post‑incident reviews have better data, so the improvements you make in ISMS.online reflect what really happens in your MSP.

Over time, you can refine your playbooks based on actual incidents, retire steps that nobody needs, and add checks that repeatedly prove useful. That keeps the framework credible, avoids documentation bloat and helps you show auditors and customers that your incident response is improving based on real evidence.


What incident evidence should an ISO 27001 auditor expect to see from an MSP?

An ISO 27001 auditor usually wants to see structured, repeatable records for significant incidents that show how you detected, assessed, handled and learned from them across both MSP platforms and affected customers.

What does an audit‑ready incident record look like for a multi‑tenant MSP?

For each serious incident, an audit‑ready record will normally include:

  • When and how you first became aware of the problem, and which detection source or monitoring tool raised it.
  • How you assessed impact and urgency, including the severity level and a short justification.
  • Which services, systems and customers were affected, and whether the issue was limited to one tenant or related to a shared MSP platform.
  • The containment, eradication and recovery actions you took, with a clear link to who carried them out and when.
  • Who you informed, including customers, regulators, partners or insurers, with times and channels used.
  • When you declared the incident closed, and any residual risks or follow‑up items.
  • What you learned and what changed as a result, such as updated risks, strengthened controls, new training or refined policies.

For MSPs, auditors also look for portfolio‑level discipline, for example:

  • A clear distinction between incidents that stay inside one customer environment and those that stem from MSP platforms or shared tooling.
  • Evidence that serious platform or multi‑tenant incidents are reviewed at a portfolio level, not only inside individual tickets.
  • A visible link between important incidents and your risk assessment, risk treatment plans and management review outputs.

You can capture all of this through a simple structure with headings like “Overview”, “Timeline”, “Impact”, “Actions”, “Communications” and “Lessons learned”, implemented as fields or forms in your ISMS. When those records live in ISMS.online and are completed as part of normal work, you can answer auditor and customer questions quickly using a small set of well‑organised examples rather than assembling partial evidence from multiple systems.

How can you turn those records into a commercial advantage?

Well‑structured incident records do more than satisfy auditors. When appropriate, you can:

  • Share anonymised examples during security reviews with prospects to show how you behave during real incidents.
  • Demonstrate that your ISO 27001 framework is lived in operations, not just written in a policy.
  • Differentiate yourself from MSPs who talk about best practice but cannot show concrete evidence.

Because ISMS.online keeps incident, risk and improvement data in one place, the same evidence that underpins certification also becomes a powerful way to reassure enterprise customers, procurement teams and cyber insurers that your MSP is prepared for difficult days, not only quiet ones.


How can an MSP adopt ISO 27001‑aligned incident response without overwhelming the team?

You make ISO 27001‑aligned incident response sustainable by starting small, focusing on real risks and expanding once the first scope is working in production.

What does a realistic first 90‑day plan look like for an MSP?

A 90‑day approach that many MSPs can handle alongside existing work might look like this:

  1. Define scope: Decide which shared tools and customer segments you will cover first, for example RMM, backup and identity platforms across your most important customers.
  2. Set simple rules: Write a short incident policy and a clear severity model so everyone understands what counts as an incident, who can declare one and how you describe impact.
  3. Create focused runbooks: Draught two short playbooks in engineer‑friendly language, such as one for suspected ransomware and one for account compromise.
  4. Design records and reviews: Configure a straightforward incident record template and a post‑incident review form in your ISMS with only the fields you genuinely need.
  5. Exercise the process: Run at least one tabletop walk‑through with internal stakeholders and, where possible, a cooperative customer using a likely scenario.
  6. Refine and plan expansion: Capture what helped and what slowed you down, then refine your playbooks, templates and severity model before planning how to extend coverage.

You can treat this as three phases: scope and design in the first month, pilots and exercises in the second, and refinement plus planning in the third. That pace is usually fast enough to show value to leadership without burning out engineers.

How do you grow the framework without losing control?

After the first 90 days, the goal is to keep moving in small, predictable steps:

  • Extend the same framework to additional services or customer tiers one at a time.
  • Add playbooks for new patterns you actually see in your tickets rather than trying to cover every theoretical risk.
  • Feed serious incidents into your risk and management review meetings so investment and process changes stay visible.
  • Regularly prune and adjust templates and checklists in ISMS.online so they reflect how your MSP currently operates.

If you want to accelerate that journey, using ISMS.online gives you a structured starting point for ISO 27001‑aligned incident response, including components suited to MSPs. That lets your team focus on tailoring scope, roles and playbooks to your business instead of designing everything from scratch, while still ending up with an approach that engineers respect, auditors understand and customers trust.



Mark Sharron

Mark Sharron leads Search & Generative AI Strategy at ISMS.online. His focus is communicating how ISO 27001, ISO 42001 and SOC 2 work in practice - tying risk to controls, policies and evidence with audit-ready traceability. Mark partners with product and customer teams so this logic is embedded in workflows and web content - helping organisations understand, prove security, privacy and AI governance with confidence.

Take a virtual tour

Start your free 2-minute interactive demo now and see
ISMS.online in action!

platform dashboard full on mint

We’re a Leader in our Field

4/5 Stars
Users Love Us
Leader - Spring 2026
High Performer - Spring 2026 Small Business UK
Regional Leader - Spring 2026 EU
Regional Leader - Spring 2026 EMEA
Regional Leader - Spring 2026 UK
High Performer - Spring 2026 Mid-Market EMEA

"ISMS.Online, Outstanding tool for Regulatory Compliance"

— Jim M.

"Makes external audits a breeze and links all aspects of your ISMS together seamlessly"

— Karen C.

"Innovative solution to managing ISO and other accreditations"

— Ben H.