{"id":828,"date":"2026-04-22T13:25:57","date_gmt":"2026-04-22T13:25:57","guid":{"rendered":"https:\/\/satoriocs.com\/blogs\/?p=828"},"modified":"2026-04-22T13:37:59","modified_gmt":"2026-04-22T13:37:59","slug":"ai-ml-based-samd-in-2026-what-indian-medical-device-manufacturers-must-know-about-fda-guidance-the-eu-ai-act-and-total-product-lifecycle-compliance","status":"publish","type":"post","link":"https:\/\/satoriocs.com\/blogs\/ai-ml-based-samd-in-2026-what-indian-medical-device-manufacturers-must-know-about-fda-guidance-the-eu-ai-act-and-total-product-lifecycle-compliance\/","title":{"rendered":"AI\/ML-Based SaMD in 2026: What Indian Medical Device Manufacturers Must Know About FDA Guidance, the EU AI Act, and Total Product Lifecycle Compliance"},"content":{"rendered":"\n<p><strong>Author:<\/strong> Regulatory Affairs Team, Satori One Click Solutions LLP | <strong>Reading time:<\/strong> ~14 minutes<\/p>\n\n\n\n<hr class=\"wp-block-separator has-alpha-channel-opacity\"\/>\n\n\n\n<p>If you are an Indian medical device manufacturer with an AI-powered software product \u2014 a diagnostic imaging tool, a clinical decision support system, a remote patient monitoring application \u2014 2026 is not a year to wait and watch.<\/p>\n\n\n\n<p>Two of the world&#8217;s most consequential regulatory changes are happening right now, simultaneously. The US FDA has fundamentally restructured how it expects manufacturers to develop, validate, and maintain AI-enabled medical devices across their entire lifecycle. And the EU AI Act \u2014 the world&#8217;s first comprehensive law governing artificial intelligence \u2014 has begun its full enforcement phase for high-risk AI systems, which includes virtually every AI product intended for medical use.<\/p>\n\n\n\n<p>At the same time, India&#8217;s own CDSCO released its first-ever draft guidance on Medical Device Software in October 2025, including explicit provisions for AI and ML-based SaMD. The regulatory world your AI product must navigate in 2026 looks nothing like it did even two years ago.<\/p>\n\n\n\n<p>This guide explains what has changed, what it means for Indian manufacturers targeting the US and EU markets, and the specific compliance steps you need to take today.<\/p>\n\n\n\n<hr class=\"wp-block-separator has-alpha-channel-opacity\"\/>\n\n\n\n<h2 class=\"wp-block-heading\">What Is SaMD \u2014 and Why Does It Matter for Indian Manufacturers?<\/h2>\n\n\n\n<p>Software as a Medical Device (SaMD) is standalone software that performs a medical purpose independently \u2014 without being embedded in a physical hardware device. Examples include:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>AI-powered radiology tools that detect abnormalities in CT or MRI scans<\/li>\n\n\n\n<li>Clinical decision support software that recommends treatment protocols<\/li>\n\n\n\n<li>Electrocardiogram (ECG) analysis applications<\/li>\n\n\n\n<li>Wearable-connected patient monitoring systems<\/li>\n\n\n\n<li>Ophthalmology screening tools using computer vision<\/li>\n\n\n\n<li>Predictive algorithms for early sepsis or deterioration detection<\/li>\n<\/ul>\n\n\n\n<p>India has a rapidly growing SaMD sector. The government&#8217;s Production Linked Incentive (PLI) Scheme for Medical Devices, the National Medical Devices Policy, and active investment in MedTech innovation at IITs have made India a credible originator \u2014 not just an importer \u2014 of medical software. Companies from Bangalore, Hyderabad, Pune, and Mumbai are building AI diagnostic tools that they want to sell in the US, Europe, Canada, and Australia.<\/p>\n\n\n\n<p>But here is the challenge most Indian SaMD manufacturers face when they try to go global: the regulatory requirements for AI-based medical devices in the US and EU are more demanding, more nuanced, and more rapidly evolving than for any other product category. Getting them wrong delays your market entry by 12 to 24 months. Getting them right \u2014 early \u2014 gives you a significant competitive advantage over rivals who are still figuring it out.<\/p>\n\n\n\n<hr class=\"wp-block-separator has-alpha-channel-opacity\"\/>\n\n\n\n<h2 class=\"wp-block-heading\">Part 1: FDA&#8217;s New Approach to AI\/ML SaMD \u2014 The Total Product Lifecycle (TPLC) Framework<\/h2>\n\n\n\n<h3 class=\"wp-block-heading\">The Core Problem FDA Was Solving<\/h3>\n\n\n\n<p>Traditional medical device regulation was built on a straightforward model: design a device, test it, submit it for clearance, sell it. The product you submitted is the product on the market. Changes require new submissions.<\/p>\n\n\n\n<p>AI breaks this model completely. Machine learning models can update. Algorithms can drift. A model trained on one patient population may perform differently on another. An AI that is performing well today may be performing poorly six months from now as the real-world data it processes diverges from its training data. The FDA recognised by 2019 that its &#8220;traditional paradigm was not designed for adaptive artificial intelligence and machine learning technologies.&#8221;<\/p>\n\n\n\n<p>The answer FDA developed over several years is the Total Product Lifecycle (TPLC) approach \u2014 a framework that governs not just the moment a device is cleared, but every stage of its existence from design through deployment through ongoing real-world performance.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">The Key FDA Guidance Documents (2021\u20132026)<\/h3>\n\n\n\n<p>The FDA has built its AI\/ML regulatory framework through a series of guidance documents. Indian manufacturers targeting the US market need to understand all of them:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>October 2021 \u2014 Good Machine Learning Practice (GMLP) for Medical Device Development: Guiding Principles<\/strong> This joint guidance from FDA, Health Canada, and the UK&#8217;s MHRA established 10 core principles for how AI\/ML medical devices should be designed, developed, and validated. It covers multi-disciplinary teams, appropriate data management, clinical relevance of reference datasets, model performance testing, and the importance of human-AI workflow design.<\/li>\n<\/ul>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>April 2023 \u2014 Predetermined Change Control Plans (PCCPs): Guiding Principles<\/strong> PCCPs are one of the most important concepts for AI SaMD manufacturers to understand. A PCCP is a plan that you submit to FDA before market authorisation, describing the kinds of changes you expect to make to your AI model after launch \u2014 and the protocol you will follow to implement those changes safely \u2014 without requiring a brand-new submission for each update. This is critical for adaptive AI systems.<\/li>\n<\/ul>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>June 2024 \u2014 Transparency for Machine Learning-Enabled Medical Devices<\/strong> This guidance addresses how manufacturers should communicate to users, healthcare providers, and patients about how their AI works, what its known limitations are, and how its performance has been validated.<\/li>\n<\/ul>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>January 2025 \u2014 Artificial Intelligence-Enabled Device Software Functions: Lifecycle Management and Marketing Submission Recommendations (Draft)<\/strong> This is the most comprehensive FDA guidance on AI SaMD to date. Published in January 2025, it applies the TPLC framework and specifies exactly what manufacturers must include in their marketing submissions. Key requirements include:<\/li>\n<\/ul>\n\n\n\n<ol class=\"wp-block-list\">\n<li>A detailed model description (architecture, training methodology, intended function)<\/li>\n\n\n\n<li>Data lineage and train\/test\/validation splits \u2014 where your data came from and how it was used<\/li>\n\n\n\n<li>Performance tied to specific clinical claims \u2014 not just technical accuracy but clinical relevance<\/li>\n\n\n\n<li>Bias analysis and mitigation \u2014 evidence that the model performs equitably across relevant patient subgroups (gender, age, ethnicity, comorbidities)<\/li>\n\n\n\n<li>Human-AI workflow documentation \u2014 how the device fits into clinical practice, and what the human oversight mechanisms are<\/li>\n\n\n\n<li>Post-market monitoring plan \u2014 how you will track real-world performance and detect algorithmic drift<\/li>\n\n\n\n<li>Predetermined Change Control Plan (if you intend to update the model post-approval)<\/li>\n<\/ol>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>August 2025 \u2014 Final PCCP Guidance<\/strong> FDA finalised its PCCP guidance in August 2025, formalising the mechanism for pre-authorised algorithm modifications. This is now the pathway Indian manufacturers should design their AI development roadmaps around from the start.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">What This Means in Practice for Indian SaMD Manufacturers<\/h3>\n\n\n\n<p>As of early 2026, FDA has authorised over 1,350 AI-enabled medical devices \u2014 roughly double the number from 2022. The market is growing fast. But the compliance bar has risen just as rapidly.<\/p>\n\n\n\n<p>For an Indian company seeking FDA 510(k) clearance or De Novo authorisation for an AI SaMD product, the January 2025 draft guidance effectively defines the new minimum standard. Your submission must demonstrate TPLC thinking \u2014 not just that your device performs well today, but that you have a disciplined system for monitoring it, updating it responsibly, and maintaining performance over time.<\/p>\n\n\n\n<p>Three specific gaps trip up Indian SaMD submissions most frequently:<\/p>\n\n\n\n<p><strong>1. Training data diversity.<\/strong> FDA expects bias analysis across patient subgroups. If your model was trained entirely on Indian patient data (or, conversely, entirely on Western patient data), you must document this clearly and address the implications for performance in the intended US patient population. The regulatory scrutiny on algorithmic bias has increased significantly.<\/p>\n\n\n\n<p><strong>2. Algorithm Change Protocol (ACP).<\/strong> Many Indian manufacturers still treat their AI model as a static product. FDA now expects a documented plan for how changes will be managed. Companies without an ACP or PCCP in place will find post-approval model updates require completely new submissions \u2014 causing costly delays.<\/p>\n\n\n\n<p><strong>3. Post-market surveillance design.<\/strong> FDA expects you to specify your monitoring metrics before market authorisation. What is your baseline false-positive rate? How will you detect distribution shift? What will trigger a safety review? These must be pre-specified in your submission, not figured out after your product is on the market.<\/p>\n\n\n\n<hr class=\"wp-block-separator has-alpha-channel-opacity\"\/>\n\n\n\n<h2 class=\"wp-block-heading\">Part 2: The EU AI Act \u2014 What Indian Exporters Must Comply With From August 2026<\/h2>\n\n\n\n<h3 class=\"wp-block-heading\">The World&#8217;s First Comprehensive AI Law<\/h3>\n\n\n\n<p>The EU Artificial Intelligence Act (Regulation (EU) 2024\/1689) entered into force on 1 August 2024 and is being phased in progressively. For medical device manufacturers, the most important deadline is <strong>August 2026<\/strong>, when the high-risk AI obligations become fully applicable to new AI-powered SaMD placed on the EU market.<\/p>\n\n\n\n<p>This is not a distant horizon. If you are planning to launch an AI medical device in Europe, or if you already have CE-marked medical software in the EU, you need to be assessing and building compliance now.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">How the EU AI Act Overlaps with MDR and IVDR<\/h3>\n\n\n\n<p>Before the EU AI Act, Indian manufacturers entering the European market for medical software needed to comply with the EU Medical Device Regulation (MDR 2017\/745) or the In Vitro Diagnostic Regulation (IVDR 2017\/746). Those requirements remain fully in force.<\/p>\n\n\n\n<p>The EU AI Act does not replace MDR or IVDR. It adds a second, overlapping regulatory layer on top of them. Any AI system that is part of a CE-marked medical device, or that functions as a standalone SaMD, is classified as <strong>high-risk AI<\/strong> under the AI Act&#8217;s Article 6(1). This applies automatically and cannot be avoided by product design.<\/p>\n\n\n\n<p>The result is <strong>dual compliance<\/strong>: your AI SaMD must satisfy both MDR\/IVDR technical documentation requirements and the new AI Act Annex IV requirements. In practical terms, this means your Notified Body will audit your technical file twice \u2014 once against MDR standards, and once against AI Act standards.<\/p>\n\n\n\n<p>The Medical Device Coordination Group (MDCG) released MDCG 2025-6 in June 2025 \u2014 its first FAQ formally clarifying how the AI Act overlaps with MDR and IVDR. If you have a Notified Body relationship for EU market access, this document is essential reading.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">The Core EU AI Act Requirements for High-Risk SaMD<\/h3>\n\n\n\n<p>Under the EU AI Act, high-risk AI systems \u2014 including all medical AI \u2014 must meet the following requirements:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Data governance and data quality.<\/strong> Your training, validation, and testing datasets must be subject to appropriate data governance practices. You must demonstrate that datasets are relevant, representative, free of errors, and complete. For Indian manufacturers training on Indian patient data, you must address how this data represents the intended EU patient population.<\/li>\n\n\n\n<li><strong>Transparency and explainability.<\/strong> Users and healthcare providers must be able to understand how your AI reaches its outputs. &#8220;Black box&#8221; AI that cannot provide interpretable reasoning is significantly harder to approve in the EU than in other markets.<\/li>\n\n\n\n<li><strong>Human oversight.<\/strong> The AI Act requires that high-risk AI systems be designed to allow effective oversight by humans. Your device must include mechanisms for clinical staff to override, correct, or reject AI outputs. This must be built into the product design and documented in your technical file.<\/li>\n\n\n\n<li><strong>Robustness, accuracy, and cybersecurity.<\/strong> Your AI must be resilient to errors, faults, and adversarial inputs, and must maintain consistent performance throughout its lifecycle.<\/li>\n\n\n\n<li><strong>Quality Management System aligned with AI Act requirements.<\/strong> Your QMS (typically aligned with ISO 13485 for medical devices) must now also demonstrate data governance, bias mitigation, transparency, and cybersecurity controls \u2014 requirements that go beyond what ISO 13485 alone covers.<\/li>\n\n\n\n<li><strong>Post-market monitoring for AI-specific drift.<\/strong> Similar to FDA&#8217;s approach, the EU AI Act requires real-world performance monitoring and bias-drift assessment as part of your ongoing PMS obligations. Annual PDF reports are not sufficient \u2014 you need live monitoring processes.<\/li>\n\n\n\n<li><strong>Registration in the EU AI database.<\/strong> High-risk AI systems must be registered in a new EU AI database (currently being built). This will become a standard part of market access alongside EUDAMED registration for medical devices.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Timeline for Indian Manufacturers<\/h3>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>August 2024:<\/strong> EU AI Act entered into force<\/li>\n\n\n\n<li><strong>August 2025:<\/strong> GPAI (General Purpose AI) model obligations applied \u2014 relevant if your SaMD is built on a foundation model<\/li>\n\n\n\n<li><strong>August 2026:<\/strong> High-risk AI obligations fully applicable to NEW AI-powered SaMD entering the EU market<\/li>\n\n\n\n<li><strong>August 2027:<\/strong> Full compliance required for existing high-risk AI medical devices already on the EU market<\/li>\n<\/ul>\n\n\n\n<p>If you are launching a new AI SaMD in the EU after August 2026, you must comply from day one. If you already have an AI product on the EU market, you have until August 2027 to achieve full compliance \u2014 but given the documentation work involved, that timeline is already tight.<\/p>\n\n\n\n<hr class=\"wp-block-separator has-alpha-channel-opacity\"\/>\n\n\n\n<h2 class=\"wp-block-heading\">Part 3: India&#8217;s CDSCO \u2014 The New Draft Guidance on Medical Device Software<\/h2>\n\n\n\n<p>While the focus for exporters is on FDA and EU compliance, Indian manufacturers must also understand what is changing at home. In October 2025, CDSCO released its first comprehensive Draft Guidance on Medical Device Software \u2014 a 76-page document that explicitly covers AI\/ML-based SaMD for the first time.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">What the CDSCO Draft Guidance Changes<\/h3>\n\n\n\n<p>Prior to this guidance, India&#8217;s Medical Device Rules 2017 applied to SaMD in principle, but lacked detailed software-specific direction. Developers struggled with threshold questions \u2014 does my software qualify as a medical device? How is it classified? What documentation is required? The new draft guidance answers these questions directly.<\/p>\n\n\n\n<p><strong>Key provisions relevant to AI\/ML SaMD:<\/strong><\/p>\n\n\n\n<p>The guidance distinguishes clearly between Software in a Medical Device (SiMD) \u2014 software embedded in physical hardware like insulin pumps \u2014 and Software as a Medical Device (SaMD), which operates independently. AI-powered diagnostic tools, ECG analysis apps, and clinical decision support systems fall under SaMD.<\/p>\n\n\n\n<p>A risk-based classification framework now applies: Class A (low risk) through Class D (highest risk). The classification is determined by two factors \u2014 the significance of the information the software provides and the severity of the healthcare condition it addresses. AI tools that directly inform diagnosis or treatment in critical clinical scenarios will typically be classified as Class C or D.<\/p>\n\n\n\n<p>For AI\/ML-based SaMD, submission documentation must include an Algorithm Change Protocol (ACP) \u2014 a document describing how changes to the AI algorithm will be managed throughout the product lifecycle. This mirrors FDA&#8217;s approach and reflects the CDSCO&#8217;s intent to align with global digital health regulatory standards.<\/p>\n\n\n\n<p>Class A and B SaMD are licensed by state authorities; Class C and D fall under CDSCO&#8217;s Central Licensing Authority with more rigorous review requirements.<\/p>\n\n\n\n<p><strong>What this means for exporters:<\/strong> The CDSCO guidance, once finalised, creates a domestic regulatory foundation that is explicitly aligned with FDA and EU approaches. Indian manufacturers who build their AI SaMD compliance programmes to meet the higher bar of FDA and EU requirements will find that CDSCO compliance follows naturally \u2014 not the other way around. Design for the most demanding market first; domestic compliance becomes significantly easier as a result.<\/p>\n\n\n\n<hr class=\"wp-block-separator has-alpha-channel-opacity\"\/>\n\n\n\n<h2 class=\"wp-block-heading\">Part 4: The Five Compliance Actions Indian SaMD Manufacturers Must Take Now<\/h2>\n\n\n\n<h3 class=\"wp-block-heading\">1. Classify Your AI System Under All Three Frameworks Simultaneously<\/h3>\n\n\n\n<p>Before investing further in development or market entry planning, classify your AI product under all relevant frameworks:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>US FDA:<\/strong> Determine device class (I, II, III), identify the appropriate submission pathway (510(k), De Novo, PMA), and identify predicate devices<\/li>\n\n\n\n<li><strong>EU AI Act + MDR\/IVDR:<\/strong> Confirm your device is high-risk AI (almost certainly yes), determine your MDR risk class, identify your Notified Body requirements<\/li>\n\n\n\n<li><strong>CDSCO:<\/strong> Classify under the draft SaMD framework (Class A\u2013D) and identify your licensing authority<\/li>\n<\/ul>\n\n\n\n<p>Classification drives everything downstream \u2014 documentation requirements, timelines, costs, and strategy. Many Indian manufacturers waste 6\u201312 months because they begin documentation before completing a thorough regulatory classification analysis.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">2. Build Your Data Governance Framework Before Training Your Model<\/h3>\n\n\n\n<p>Both FDA and the EU AI Act place extraordinary emphasis on data quality, data lineage, and bias analysis. This is an area where Indian manufacturers consistently underinvest at the beginning of a project, then pay the price during regulatory review.<\/p>\n\n\n\n<p>Specifically, you need to document:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Where your training, validation, and test data came from<\/li>\n\n\n\n<li>How the data was labelled, curated, and quality-controlled<\/li>\n\n\n\n<li>Whether the data represents the patient population you intend to serve (demographic diversity, comorbidity range, equipment variability)<\/li>\n\n\n\n<li>Your bias analysis methodology \u2014 how you evaluated and addressed performance disparities across patient subgroups<\/li>\n<\/ul>\n\n\n\n<p>If your device will be sold in the US or EU, bias analysis across gender, age, and ethnicity is non-negotiable. Build this into your data collection and annotation plan from the start \u2014 retrofitting it after model training is extremely costly.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">3. Design Your Algorithm Change Protocol (ACP) Before Launch<\/h3>\n\n\n\n<p>One of the most consequential decisions you will make about your AI product is whether to design it as a locked model (no post-market learning) or an adaptive model (capable of updating). The regulatory implications are significant.<\/p>\n\n\n\n<p>Locked models follow a more predictable approval pathway but require a new submission for any meaningful algorithm change. Adaptive models \u2014 or locked models with a PCCP in place \u2014 allow pre-authorised updates but require significantly more upfront documentation.<\/p>\n\n\n\n<p>As of 2025, fewer than 10% of FDA-cleared AI devices had an authorised PCCP. This means most manufacturers are paying the price of new submissions every time they want to improve their model. Indian manufacturers entering the market now have the opportunity to design PCCP-readiness in from the start \u2014 giving them a long-term competitive advantage in speed and cost of product iteration.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">4. Build Human Oversight Into Product Design \u2014 Not as an Afterthought<\/h3>\n\n\n\n<p>Both FDA and the EU AI Act require that high-risk AI systems allow effective human oversight. For clinical AI, this means your product design must include:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Clear display of AI confidence scores or uncertainty estimates<\/li>\n\n\n\n<li>Mechanisms for clinicians to override, accept, or reject AI outputs<\/li>\n\n\n\n<li>Audit trails of AI-assisted decisions<\/li>\n\n\n\n<li>User interface design that supports rather than supplants clinical judgment<\/li>\n<\/ul>\n\n\n\n<p>Products that present AI outputs as definitive answers, without uncertainty quantification or override capability, will face significant friction in both FDA submissions and EU Notified Body assessments. Design for human-AI collaboration from your first prototype.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">5. Align Your QMS for Dual Compliance From Day One<\/h3>\n\n\n\n<p>Your Quality Management System is the backbone of regulatory approval in every market. For AI SaMD manufacturers targeting the US and EU, a QMS that meets only ISO 13485 is no longer sufficient.<\/p>\n\n\n\n<p>Your QMS must also demonstrate:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Data governance procedures (for FDA GMLP and EU AI Act requirements)<\/li>\n\n\n\n<li>Software lifecycle management aligned with IEC 62304<\/li>\n\n\n\n<li>Risk management aligned with ISO 14971<\/li>\n\n\n\n<li>Cybersecurity controls (increasingly scrutinised in both FDA and EU reviews)<\/li>\n\n\n\n<li>Bias monitoring and algorithmic drift detection procedures<\/li>\n\n\n\n<li>Post-market surveillance processes that include AI-specific performance metrics<\/li>\n<\/ul>\n\n\n\n<p>Building a QMS that satisfies all these requirements simultaneously is complex \u2014 but it is far less costly than building separate systems for each market, or rebuilding your QMS after a regulatory rejection.<\/p>\n\n\n\n<hr class=\"wp-block-separator has-alpha-channel-opacity\"\/>\n\n\n\n<h2 class=\"wp-block-heading\">How Satori Helps Indian SaMD Manufacturers Navigate Global Compliance<\/h2>\n\n\n\n<p>At Satori One Click Solutions LLP, we work with Indian medical device manufacturers at every stage of their global market access journey. For AI\/ML-based SaMD companies, our regulatory support spans:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Pre-submission strategy:<\/strong> We help you classify your device correctly under FDA, EU MDR\/AI Act, and CDSCO frameworks simultaneously \u2014 before you commit to a documentation or development approach. Getting this right at the start saves 6\u201318 months.<\/li>\n\n\n\n<li><strong>FDA submission preparation:<\/strong> We prepare your 510(k) or De Novo submissions with full TPLC documentation \u2014 model descriptions, data lineage, bias analysis, PCCP development, and post-market monitoring plans aligned with FDA&#8217;s January 2025 draft guidance.<\/li>\n\n\n\n<li><strong>EU MDR + AI Act dual compliance:<\/strong> We support your CE marking pathway under MDR\/IVDR and build the AI Act Annex IV documentation layer your Notified Body will require from August 2026 onwards.<\/li>\n\n\n\n<li><strong>CDSCO SaMD licensing:<\/strong> We guide your domestic licensing under the new CDSCO Medical Device Software framework, including Algorithm Change Protocol preparation and QMS alignment.<\/li>\n\n\n\n<li><strong>US Agent and Indian Agent services:<\/strong> Satori provides FDA US Agent services for Indian manufacturers required to register with the FDA, and Indian Agent services for foreign companies registering with CDSCO.<\/li>\n\n\n\n<li><strong>Ongoing compliance support:<\/strong> Regulatory requirements for AI SaMD are evolving faster than in any other product category. We provide ongoing monitoring of FDA, EU, CDSCO, Health Canada, and TGA regulatory changes affecting AI medical devices, so your compliance posture stays current without requiring constant internal resource investment.<\/li>\n<\/ul>\n\n\n\n<hr class=\"wp-block-separator has-alpha-channel-opacity\"\/>\n\n\n\n<h2 class=\"wp-block-heading\">Frequently Asked Questions<\/h2>\n\n\n\n<p><strong>Does my AI diagnostic app qualify as a SaMD?<\/strong> If your software is intended to perform a medical purpose \u2014 diagnosis, treatment, monitoring, prevention, or prediction \u2014 independently of a physical hardware device, it is almost certainly classified as SaMD in India, the US, and the EU. The key factor is intended use, not how the software is technically built. Consult a regulatory specialist before assuming your product falls outside SaMD scope.<\/p>\n\n\n\n<p><strong>Do I need FDA clearance before I can sell my AI medical device in the US?<\/strong> Yes, with very limited exceptions. Most AI SaMD products require 510(k) clearance, De Novo classification, or PMA approval before US market entry. The appropriate pathway depends on your device classification and whether a substantially equivalent predicate device exists. Satori can assess your pathway in a pre-submission consultation.<\/p>\n\n\n\n<p><strong>When does EU AI Act compliance become mandatory for my product?<\/strong> For new AI-powered SaMD entering the EU market, full high-risk AI obligations apply from August 2026. For existing products already on the EU market, the deadline is August 2027. If you have a CE-marked AI medical device today, you have less than 18 months to achieve full AI Act compliance alongside your existing MDR\/IVDR obligations.<\/p>\n\n\n\n<p><strong>Our AI model was trained on Indian patient data. Can we use this for US or EU submissions?<\/strong> You can, but you must address the representativeness and bias implications thoroughly. FDA and the EU AI Act both require bias analysis across patient subgroups in the intended market population. Training data from an Indian population may have demographic, environmental, and equipment characteristics that differ meaningfully from US or EU patients. A well-documented data governance and bias analysis plan can address this \u2014 but it must be proactive, not retrofitted.<\/p>\n\n\n\n<p><strong>What is a Predetermined Change Control Plan (PCCP) and do we need one?<\/strong> A PCCP is a document submitted to FDA as part of your market authorisation that describes categories of changes you plan to make to your AI model post-approval, and the processes you will follow to implement them safely without requiring a new submission for each update. PCCPs are optional but strongly recommended for adaptive AI systems. Manufacturers without a PCCP must submit a new 510(k) for any significant algorithm change \u2014 which can take 6\u201312 months per update. Given the rapid development cycles of most AI products, designing a PCCP from the start is almost always the right strategic decision.<\/p>\n\n\n\n<hr class=\"wp-block-separator has-alpha-channel-opacity\"\/>\n\n\n\n<h2 class=\"wp-block-heading\">Summary: The 2026 Regulatory Landscape for AI SaMD in Brief<\/h2>\n\n\n\n<figure class=\"wp-block-table\"><table><thead><tr><th><\/th><th><\/th><th><\/th><\/tr><\/thead><tbody><tr><td>India<\/td><td>CDSCO MDR 2017 + Draft MDS Guidance (Oct 2025)<\/td><td>Risk classification (A\u2013D), Algorithm Change Protocol, QMS per ISO 13485<\/td><\/tr><tr><td>United States<\/td><td>FDA TPLC Framework + January 2025 Draft Guidance + PCCP Final Guidance<\/td><td>TPLC documentation, bias analysis, PCCP, post-market monitoring plan<\/td><\/tr><tr><td>European Union<\/td><td>EU MDR\/IVDR + EU AI Act (Regulation (EU) 2024\/1689)<\/td><td>Dual compliance: MDR technical file + AI Act Annex IV; mandatory from August 2026 for new products<\/td><\/tr><tr><td>Canada<\/td><td>Health Canada SaMD guidance (AI-specific rules under development)<\/td><td>GMLP alignment; Health Canada is developing AI-specific guidance harmonised with FDA<\/td><\/tr><tr><td>Australia<\/td><td>TGA SaMD framework<\/td><td>IMDRF SaMD alignment; TGA references FDA and EU guidance closely<\/td><\/tr><\/tbody><\/table><\/figure>\n\n\n\n<hr class=\"wp-block-separator has-alpha-channel-opacity\"\/>\n\n\n\n<h2 class=\"wp-block-heading\">Ready to Build Your Global AI SaMD Regulatory Strategy?<\/h2>\n\n\n\n<p>The regulatory landscape for AI medical devices is more complex and more consequential than for any other product category \u2014 but it is navigable with the right partner and the right approach.<\/p>\n\n\n\n<p>At <a href=\"http:\/\/satoriocs.com\" data-type=\"link\" data-id=\"satoriocs.com\">Satori One Click Solutions LLP<\/a>, we combine regulatory expertise across CDSCO, FDA, Health Canada, EMA\/MDR, TGA, and MHRA with deep practical experience in medical device compliance. We work with Indian manufacturers building AI diagnostic tools, clinical decision support systems, and connected health platforms who want to reach global markets efficiently and sustainably.<\/p>\n\n\n\n<p>Contact our team for a complimentary regulatory assessment for your AI SaMD product. We will help you understand your classification, your submission pathway, your documentation requirements, and your timeline \u2014 so you can make confident strategic decisions about your global market entry.<\/p>\n\n\n\n<p><strong>Email:<\/strong> satoriocs@gmail.com <\/p>\n\n\n\n<p><strong>Contact:<\/strong> +91 98290 98077  <\/p>\n\n\n\n<p><strong>Website:<\/strong> www.satoriocs.com<\/p>\n\n\n\n<hr class=\"wp-block-separator has-alpha-channel-opacity\"\/>\n\n\n\n<p><em>Satori One Click Solutions LLP is a global pharma, medical device, and healthcare product regulatory consultancy operating from offices in India, Canada, and the United States. We provide end-to-end regulatory affairs, quality management, legal compliance, and market access services across USFDA, Health Canada, CDSCO, EMA, MHRA, and TGA jurisdictions.<\/em><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Author: Regulatory Affairs Team, Satori One Click Solutions LLP | Reading time: ~14 minutes If you are an Indian medical device manufacturer with an AI-powered software product \u2014 a diagnostic imaging tool, a clinical decision support system, a remote patient monitoring application \u2014 2026 is not a year to wait and watch. Two of the [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":836,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"pagelayer_contact_templates":[],"_pagelayer_content":"","footnotes":""},"categories":[1,4],"tags":[159,156,157,152,149,150,153,139,154,158,155,35,151],"class_list":["post-828","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-home","category-medical-device-regulations-global-compliance","tag-ai-compliance-2026","tag-ai-medical-devices","tag-cdsco-medical-device-software","tag-eu-ai-act","tag-fda-ai-guidance","tag-iec-62304","tag-indian-medical-device-manufacturers","tag-iso-13485","tag-medical-device-registration","tag-pccp","tag-regulatory-affairs-india","tag-samd","tag-tplc"],"_links":{"self":[{"href":"https:\/\/satoriocs.com\/blogs\/wp-json\/wp\/v2\/posts\/828","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/satoriocs.com\/blogs\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/satoriocs.com\/blogs\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/satoriocs.com\/blogs\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/satoriocs.com\/blogs\/wp-json\/wp\/v2\/comments?post=828"}],"version-history":[{"count":6,"href":"https:\/\/satoriocs.com\/blogs\/wp-json\/wp\/v2\/posts\/828\/revisions"}],"predecessor-version":[{"id":844,"href":"https:\/\/satoriocs.com\/blogs\/wp-json\/wp\/v2\/posts\/828\/revisions\/844"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/satoriocs.com\/blogs\/wp-json\/wp\/v2\/media\/836"}],"wp:attachment":[{"href":"https:\/\/satoriocs.com\/blogs\/wp-json\/wp\/v2\/media?parent=828"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/satoriocs.com\/blogs\/wp-json\/wp\/v2\/categories?post=828"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/satoriocs.com\/blogs\/wp-json\/wp\/v2\/tags?post=828"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}