DORA Incident Reporting (2026 Guide)

You find out about an ICT disruption late on a Friday afternoon. Operations wants technical facts first. Legal wants to know if a regulator must be notified. Compliance asks whether the event is major under DORA. Management wants a timeline, and your incident record is still incomplete. If that sounds familiar, you are exactly the audience for this guide.
DORA incident reporting is one of the areas where institutions often realize that having an incident process is not the same as having a reporting-ready incident process. Under the Digital Operational Resilience Act, firms need more than internal ticketing and post-event summaries. They need a structured way to identify, classify, document, escalate, and report serious ICT-related incidents within strict timelines.
DORApp was built to simplify DORA compliance for EU financial institutions through a modular approach, turning complex regulatory requirements into structured, manageable workflows with report-ready processes. In this article, you will get a practical view of what DORA incident reporting requires, how it fits into the wider regulation, and what your team should be doing now to stay ready in 2026.
What DORA incident reporting actually covers
DORA incident reporting sits inside the broader operational resilience framework established by Regulation (EU) 2022/2554. If you need a wider primer first, it helps to start with what is dora and then come back to the reporting detail.
At a high level, DORA requires financial entities to manage ICT-related incidents in a disciplined, traceable way. That includes identifying incidents, assessing their impact, deciding whether they meet major incident criteria, and reporting them to the relevant competent authority where required. It is not only about cyberattacks. Service outages, third-party failures, data integrity problems, and other ICT disruptions may all become relevant depending on impact and context.
From a regulatory standpoint, incident reporting connects directly to several of DORA’s five pillars. It relies on sound ICT risk management, clear governance, testing and resilience capability, and strong third-party oversight. If your institution still treats these as separate workstreams, incident reporting often exposes the gaps.
For a broader legal and structural overview, you may also want to read dora regulation explained and digital operational resilience act dora.
When an incident becomes a reporting issue
Here is where many teams get stuck. Not every operational problem becomes a DORA incident report. You first need a defensible classification process that looks at impact, critical services, customers affected, duration, geographic spread, and other relevant criteria under the applicable standards and guidance.
In practice, this means your first challenge is not writing the report. It is deciding whether the event is reportable at all. That is why classification quality matters so much. If the classification step is weak, every later reporting step becomes harder to defend.
A useful next read here is dora incident classification. It helps explain why institutions should avoid rushing into a “major” or “non-major” label before the key impact fields are complete enough to support that decision.
Classification is not the same as technical severity
Your IT team may call something “critical” because a production system is down. That may be operationally correct, but DORA reporting looks at regulatory significance, not only technical urgency. A short outage in a non-critical internal tool may be severe internally but not reportable. A third-party issue affecting a critical service across entities could be reportable even if the underlying technical failure seems limited at first glance.
Think of it this way, DORA asks how the incident affects the resilience of regulated services and the institution’s obligations, not only how difficult it is for engineers to fix.
ICT risk context matters
Reporting decisions also sit inside your wider risk framework. If teams do not share a common understanding of dependencies, service criticality, or provider concentration, incident classification may become inconsistent. That is one reason cross-functional coordination matters so much. For background, see what is ict risk.
Major incident criteria and thresholds: what typically makes an incident “major”
Here is the thing, many teams understand that “impact matters,” but still struggle to explain impact in a way that holds up under review. For DORA purposes, a major ICT-related incident is typically not defined by a single dramatic symptom. It is defined by a combination of indicators that show meaningful disruption to critical services, significant customer impact, or material operational risk.
In most cases, authorities and technical standards focus on criteria in a few broad buckets. The exact thresholds and how they are interpreted can vary by competent authority, entity type, and national practice, so you should treat this section as an operational overview and align your internal rule set with your legal and compliance teams where needed.
The criteria regulators typically look for
While the detailed classification logic belongs in your formal process, a practical working view usually includes:
This does not replace your formal incident classification method. It is a compact reminder of what usually needs to be visible in the record before you can defend a “major” conclusion.
Decision inputs to capture in the first hours
Many reporting delays happen because teams did not capture the right fields early. Even if values are provisional, it helps to collect decision inputs immediately so you can justify why you classified the incident as major or non-major at the time.
In practice, that usually means capturing:
Think of it this way, you are not only classifying the incident, you are preserving the reasoning behind the classification. That reasoning is often what protects you later when timelines are tight and facts evolve.

Reporting stages and timelines you need to manage
Under DORA, incident reporting is typically a staged process rather than a single one-time submission. Based on the technical standards currently in force, institutions may need to prepare an initial report, follow with an intermediate update, and then submit a final report once the incident has been fully assessed and closed.
The exact workflow can vary depending on your competent authority setup and the specifics of the event, but the discipline is the same. You need enough structured data early, enough governance around approvals, and enough process control to keep deadlines from slipping while the incident is still unfolding.
The deadlines are short for a reason
For major ICT-related incidents, timing matters. Institutions generally need to move quickly after awareness and classification. As reflected in DORApp’s incident management documentation, common configured reporting logic follows the staged timeline used in the technical standards: an initial report within 4 hours from major classification and no later than 24 hours from awareness, an intermediate report within 72 hours from initial submission, and a final report within one month from the intermediate stage or latest accepted update, based on the configured process.
If you want the full reporting stage breakdown, dora major incident reporting is the best companion article.
Estimated data is sometimes unavoidable
The reality is that you rarely know everything at the start. Early incident records may contain estimates for customer impact, transaction disruption, financial effect, or root cause status. That is usually acceptable where the reporting framework allows it, provided your institution updates later stages with better-confirmed information and keeps a clear audit trail of what changed and why.
What many people overlook is that regulators generally understand uncertainty in the first hours of a live incident. What creates problems is not uncertainty itself, but unstructured reporting, missing rationale, or poor traceability.
RTS, ITS, and reporting templates: what authorities expect in practice
Now, when it comes to day-to-day execution, DORA incident reporting is not only “send a report fast.” It is “send the right information in the right structure.” This is where Regulatory Technical Standards (RTS) and Implementing Technical Standards (ITS) matter operationally, because they typically shape what information must be captured, how it is categorized, and how it is submitted.
You do not need to be a legal drafter to feel this. If your incident record does not line up with the reporting structure, teams usually end up rewriting, reformatting, and re-arguing decisions under time pressure. A template-driven reality changes the workflow.
Why RTS and ITS matter beyond legal interpretation
In practical terms, RTS and ITS often drive:
For most small business owners and entrepreneurs, templates can feel like a paperwork issue. For regulated institutions, templates are an operational requirement. They are also a strong hint about what regulators expect to be traceable in your internal evidence base.
What template-driven reporting changes in practice
When reporting is template-driven, you typically need to work with structured inputs rather than relying on narrative alone. That often means:
The difference often comes down to whether the incident record is built as a “report draft” from the start, or as a technical ticket that someone tries to convert into regulatory language later.
How to operationalize template readiness
Consider this, your first submission should not be your first test of the reporting structure. Teams that report well typically do a few practical things in advance:
If you want to reduce stress on the day, build your reporting process so it behaves like a repeatable workflow, not an emergency writing exercise.
What good reporting looks like in practice
A strong DORA incident report usually reflects good preparation long before the incident happened. The institutions that handle reporting well tend to have already mapped critical services, linked providers to those services, defined ownership, and clarified who signs off what.
From a practical standpoint, good reporting usually includes five things:
Your incident record should act as the source of truth
If the reporting team has to rebuild the story from email chains, spreadsheets, vendor calls, and scattered technical tickets, you will lose time and confidence. A cleaner approach is to keep one canonical incident record and attach evidence, decisions, actions, and reporting drafts to that same record.
Platforms like DORApp streamline the incident and Register of Information process through structured imports, intuitive record management, validation checks, and compliant report generation workflows. That does not replace internal accountability, but it can reduce the manual burden that often causes reporting delays.
Third-party involvement often changes the picture
Many major incidents involve external providers in some way, whether through direct service failure, subcontracting breakdowns, or delayed evidence from a vendor. In 2026, that matters even more because regulators are paying closer attention to third-party oversight, CTPP designation effects, and subcontracting risk under Delegated Regulation (EU) 2025/532.
Consider this, if your team cannot quickly identify which provider supports the affected service, what contract applies, and whether critical functions are involved, your reporting may slow down at exactly the wrong moment.
You can also browse related topic collections under Incident Reporting and DORA Fundamentals.

DORA incident reporting examples by scenario
Example-driven thinking helps because many “major incident” debates happen in gray areas. The goal is not to memorize scenarios. It is to recognize what should be recorded early, what typically triggers escalation, and where teams get blocked when they try to produce a defensible report.
Scenario 1: Third-party outage affecting a critical customer-facing service
A key external provider has an outage that degrades your customer-facing service. Your own infrastructure is stable, but customers cannot complete essential actions for hours.
What to record early:
What likely triggers escalation:
Common blockers and how to handle them:
Scenario 2: Data integrity issue with limited downtime but high operational impact
A processing job runs successfully, but produces incorrect outputs. Systems stay up, yet records may be unreliable. Customers may not notice immediately, but your institution faces a high integrity risk.
What to record early:
What likely triggers escalation:
Common blockers and how to handle them:
Scenario 3: Cyber incident where early scope is uncertain
Security detects suspicious activity. You take containment actions quickly, but you do not yet know whether data was accessed, whether lateral movement occurred, or whether multiple services were affected.
What to record early:
What likely triggers escalation:
Common blockers and how to handle them:
Scenario 4: Repeated instability that does not look dramatic, until it does
A service experiences repeated short disruptions, intermittent latency spikes, and partial functionality loss. Each event seems minor, but the cumulative impact becomes significant.
What to record early:
What likely triggers escalation:
Common blockers and how to handle them:
Across all scenarios, the practical theme is the same. IT, security, vendor management, compliance, legal, and management typically need clear handoffs and sign-offs. If you only bring compliance in at the end, you often lose the traceability you needed at the start.
Common reporting mistakes institutions still make
Most reporting failures are not caused by a lack of effort. They usually come from process design problems that only become obvious during a live incident.
Starting with narrative instead of structured facts
Teams often begin with long descriptions and too few standardized fields. That makes reporting harder because deadlines usually depend on timestamps, impact values, service references, and classification logic. Narrative matters, but structured data comes first.
Separating incident response from regulatory reporting
Operational responders and compliance teams sometimes work in parallel with little synchronization. One group contains the issue, the other tries to interpret it for reporting. In practice, this can create duplicate records, missing evidence, and inconsistent timelines.
Waiting for perfect information
This is one of the most common traps. Teams delay because they want certainty on root cause or full impact before moving forward. DORA reporting does not reward avoidable delay. It generally expects timely, defensible reporting that can be updated as facts improve.
Treating the incident as finished after submission
The report is not the finish line. Post-incident review, corrective actions, and feedback into risk management are all part of a mature resilience program. This is one reason articles like DORA Pillars Explained: Complete Breakdown (2026) are useful, because they show how incident reporting connects with the rest of the DORA framework.
How tools can support the process without replacing judgment
No platform can make regulatory judgment calls for you. Your institution still needs governance, accountable decision-makers, and legal and compliance review where required. What tools can do is make the process more structured, more visible, and less dependent on manual patchwork.
With features such as automated workflows, configurable review gates, audit trail, structured evidence handling, and a data model that auto-converts to XBRL for DORA reporting, DORApp allows teams to work with incomplete but improving data instead of waiting for perfect inputs. Based on the available documentation, it also supports incident lifecycle stages, reporting workspace preparation, and integration with related compliance records.
Support should reduce friction, not create a second problem
If a reporting solution is too rigid, teams may avoid using it during real incidents. If it is too loose, it may not support defensible reporting. The right balance usually means guided structure with room for staged updates, estimates where allowed, and clear approval routing.
That practical balance reflects Dorapp’s broader approach to clarity and efficiency, shaped in part by founder Matevž Rostaher’s background across FinTech, InsurTech, and RegTech environments where operational pressure and regulatory expectations often meet.

Why 2026 feels different from initial compliance
The first phase of DORA was about getting frameworks, policies, and registers in place by the January 2025 application date. The mood in 2026 is different. Regulators increasingly expect proof that your processes actually work under pressure.
Under DORA, this means incident reporting is no longer just a documented policy area. It is part of your institution’s evidence of operational resilience. Authorities may look not only at whether you submitted reports, but whether your classification logic, data quality, governance trail, and third-party coordination were credible.
It also helps to remember where this requirement came from and how it evolved. For that context, see DORA European Commission Timeline and History (2026).
Explore how DORApp can support your DORA compliance journey with a 14-day free trial. If you prefer to talk through your incident reporting workflow first, you can also book a demo and see how the platform approaches reporting, validation, and operational governance in practice.
Disclaimer: The information in this article is intended for general informational and educational purposes only. It does not constitute professional technical, legal, financial, or regulatory advice. Website performance outcomes, platform capabilities, and business results will vary depending on your specific circumstances, goals, and implementation. Always evaluate tools and platforms based on your own needs and, where relevant, seek professional guidance.
Regulated industry note: This article is for informational purposes only and does not constitute financial, legal, or regulatory advice. DORA compliance requirements may vary based on your institution type, size, and national regulatory framework. Content referencing regulated industries is provided for general context only and should not be interpreted as legal, regulatory, compliance, or financial advice. If you operate in a regulated sector, always consult qualified financial, legal, and compliance professionals for guidance specific to your situation.
Frequently Asked Questions
What is DORA incident reporting in simple terms?
DORA incident reporting is the process EU financial entities use to assess and report serious ICT-related incidents to their competent authority under the Digital Operational Resilience Act. In simple terms, it means you need a clear way to identify important incidents, determine whether they are major, document the impact, and submit reports within the required timelines. It is not limited to cyberattacks. System outages, provider failures, and data integrity issues may also become relevant if they significantly affect critical services or operations.
Does every ICT incident need to be reported under DORA?
No. DORA does not require every technical issue or operational disruption to be reported externally. The key question is whether the incident meets the criteria for reportability, especially whether it qualifies as a major ICT-related incident under the applicable rules and guidance. Your institution should have a documented classification process to assess impact, duration, affected services, customers, and other relevant factors. Internal recording is still important for non-major incidents, because they may reveal patterns, recurring weaknesses, or cumulative risk over time.
What are the main reporting stages under DORA?
The reporting flow generally follows staged submissions for major incidents. Based on current technical standards and implementation practice, this usually means an initial report, an intermediate report, and a final report. The initial stage focuses on early facts and immediate impact. The intermediate stage updates the regulator on developments, containment, and revised impact. The final stage closes the loop with root cause, complete impact assessment, and remediation context. Exact handling may vary by authority and institution setup, so you should validate the process against your legal and supervisory obligations.
How fast do you need to report a major incident?
The timelines are short. As reflected in DORApp’s documented reporting workflow, common DORA-aligned handling uses an initial report within 4 hours from major classification and no later than 24 hours from awareness, followed by an intermediate report within 72 hours from initial submission and a final report within one month from the intermediate stage or latest accepted update. Because timing can be interpreted through configured workflows and national practice, your institution should confirm the exact operational rule set with legal and compliance advisors.
What information usually blocks a DORA incident report from being submitted?
The fields that most often cause delay are usually not the most technical ones. Common blockers include the awareness timestamp, classification rationale, affected service mapping, legal entity context, impact quantification, and third-party involvement details. If your institution has not linked services, providers, and criticality in advance, these gaps may only become visible during a live incident. That is why mature teams work on data readiness before the next event happens, not during the event itself.
How does incident reporting relate to the Register of Information?
The Register of Information is not the same thing as incident reporting, but the two are closely connected. If a reportable incident involves an ICT third-party service provider, your reporting process becomes much easier when the affected provider, service arrangement, and dependencies are already mapped in the Register of Information. In practice, institutions with poor register quality often struggle to identify which contracts, services, and critical functions are affected. Good register governance can therefore improve both reporting speed and reporting quality.
Can a tool automate DORA incident reporting?
A tool can automate parts of the workflow, but it should not replace human regulatory judgment. Good platforms may help collect structured data, track deadlines, validate mandatory fields, route approvals, and generate report-ready outputs. They may also support audit trail and evidence handling. What they cannot do on their own is decide your institution’s legal position or guarantee that a classification decision is correct in every context. Final accountability still sits with your institution’s responsible functions and decision-makers.
Why is 2026 such an important year for DORA incident reporting?
2026 is the point where many institutions are moving from framework setup to operational proof. Regulators are increasingly focused on whether processes work in practice, not only whether policies exist on paper. That means institutions may face more scrutiny on data quality, auditability, cross-functional coordination, and third-party oversight. With broader supervisory maturity, CTPP developments, and more operational testing of resilience capabilities, incident reporting is becoming a practical evidence point for whether your DORA program is actually functioning.
Who inside the institution should own DORA incident reporting?
There is rarely a single team that can handle it alone. Operational IT or security teams usually manage the technical response, but compliance, legal, risk, vendor management, and senior management often need to be involved depending on the event. A strong operating model defines who owns incident creation, who validates classification, who approves submissions, and who maintains the evidence trail. The best setups reduce confusion before an incident occurs by documenting roles, escalation points, and review gates in advance.
What should you do if some incident information is still unknown?
You usually should not freeze the process while waiting for perfect information. Where the reporting framework allows estimates or provisional values, use them carefully and document that they are provisional. Then update the later reporting stages as facts become clearer. What matters most is that your institution can explain what was known at the time, why the judgment was reasonable, and how the information evolved. Strong traceability and version control are often more useful than false certainty in the first hours of a major event.
What are the requirements for DORA reporting?
DORA reporting requirements typically center on having a defensible way to identify and classify ICT-related incidents, decide whether they qualify as major, and submit staged reports to your competent authority within defined timelines. In practice, this usually means you need structured incident records that capture awareness time, impact indicators, affected services, third-party involvement, and a clear rationale for your classification decisions, with an audit trail showing updates and approvals over time. The exact expectations can vary based on authority and entity type, so it is common to validate your internal workflow with legal and compliance teams.
What is the required incident reporting timeframe under DORA?
DORA major incident reporting is generally time-sensitive and staged. As reflected in DORApp’s documented workflow, common DORA-aligned handling uses an initial report within 4 hours from major classification and no later than 24 hours from awareness, followed by an intermediate report within 72 hours from initial submission and a final report within one month from the intermediate stage or latest accepted update. Because supervisory practice and configured workflows can affect how timing is applied, your institution should confirm the operational rule set with its legal and compliance advisors.
What are the 8 types of reportable incidents?
DORA does not usually work as a simple “pick one of eight incident types and report it” model. Reportability is typically driven by impact and classification indicators, not only by the incident label. That said, institutions often categorize incidents internally into common groups to support consistent triage and reporting, such as service availability outages, severe performance degradation, data integrity issues, data confidentiality events, cyber incidents, third-party service disruptions, change or deployment failures, and infrastructure or network failures. Your competent authority may expect standardized categories and structured fields aligned with the applicable technical standards, so it helps to keep internal categorization consistent with the reporting taxonomy you use.
What is a DORA incident reporting template?
A DORA incident reporting template is the structured format, often aligned to technical standards and supervisory expectations, that defines which fields and sections must be completed for each reporting stage. In practice, templates push teams toward consistent taxonomies, clearer mandatory fields, and better traceability across initial, intermediate, and final submissions. If your incident record is already organized around template-style fields, reporting often becomes faster because you are populating a defined structure rather than rewriting the incident from scratch under deadline pressure.
Key Takeaways
Conclusion
DORA incident reporting is not just a compliance formality. It is a practical test of whether your institution can turn operational disruption into clear, timely, and defensible regulatory action. The teams that do this well usually have one thing in common: they prepare the structure before the next incident arrives. They define critical services, connect provider data, clarify approval routes, and make sure incident records can support reporting without a last-minute scramble.
If your current process still depends on disconnected spreadsheets, inbox threads, and manual reconstruction, that is usually a sign that reporting maturity needs attention. The good news is that this can be improved with the right operating model and better supporting tools.
If you want a broader foundation, explore the Dorapp blog’s DORA resources first. If you are actively reviewing how to operationalize reporting workflows, validations, and structured evidence handling, DORApp is worth exploring at dorapp.eu. A short demo or trial can help you see whether its approach fits your institution’s process.
About the Author
Matevž Rostaher is Co-Founder and Product Owner of DORApp. He brings deep experience in building secure and compliant ICT solutions for the financial sector and is positioned by DORApp as an expert trusted by financial institutions on complex regulatory and operational challenges. DORApp’s own webinar materials list him as CEO and Co-Founder of Skupina Novum d.o.o. and CEO and Co-Founder of FJA OdaTeam d.o.o. His articles should carry the voice of someone who understands not just compliance requirements, but the systems and delivery realities behind them.