The Information Technology (Intermediary Guidelines and Digital Media Ethics Code) Amendment Rules, 2026 (“2026 Amendment Rules”) mark a significant regulatory shift in India’s digital governance framework. Notified on 10 February 2026 and effective from 20 February 2026, these amendments primarily address the regulation of synthetically generated information (SGI)—commonly referred to as AI-generated or deepfake content.
The amendments strengthen due diligence obligations for intermediaries, especially significant social media intermediaries (SSMIs), and introduce strict timelines for compliance.
The 2026 amendments aim to:
- Regulate AI-generated/deepfake content.
- Prevent misuse of synthetic media for fraud, impersonation, obscenity, misinformation, and criminal activity.
- Mandate labelling and traceability of synthetic content.
- Tighten intermediary compliance timelines.
- Align references from IPC to Bharatiya Nyaya Sanhita, 2023.
This is India’s first comprehensive regulatory framework specifically targeting synthetic digital manipulation at scale.
“Synthetically Generated Information”
A major addition is the formal definition of:
(a) Audio, Visual or Audio-Visual Information
Expanded to include any content created, generated, modified, or altered using computer resources.
(b) “Synthetically Generated Information” (SGI)
Defined as AI-created or algorithmically altered content that:
- Appears real or authentic,
- Depicts individuals or events,
- Is likely to be perceived as indistinguishable from real-world events .
Exclusions
Routine editing, formatting, accessibility improvements, color correction, or legitimate document preparation are excluded—provided they do not materially distort the underlying content.
Implication: The law clearly separates deepfake manipulation from legitimate digital enhancement.
Expansion of “Information” to Include Synthetic Content
The amendment clarifies that any reference to “information” under unlawful activity provisions shall include synthetically generated information.
This ensures:
- Deepfakes are treated on par with real content under IT Act liability provisions.
- Intermediaries cannot argue regulatory gaps.
Mandatory User Awareness Requirements
Intermediaries must now:
- Inform users every three months about legal consequences of misuse.
- Warn about penalties under:
- Bharatiya Nyaya Sanhita, 2023
- POCSO Act
- Representation of the People Act
- Indecent Representation of Women Act
- Immoral Traffic Prevention Act .
Users must be informed that violations may result in:
- Immediate content removal.
- Account suspension.
- Identity disclosure to victims.
- Mandatory reporting to authorities.
Due Diligence Obligations for Synthetic Content Platforms
Platforms that enable AI content creation must:
(A) Prevent Illegal SGI
Deploy automated tools and reasonable technical measures to prevent the creation of synthetic content that:
- Contains child sexual abuse material.
- Is obscene, pornographic, or invasive of privacy.
- Creates false documents or electronic records.
- Aids in explosives or arms procurement.
- Falsely depicts individuals or events to deceive.
(B) Mandatory Labeling
All lawful SGI must:
- Be prominently labelled.
- Contain visible disclosure (for visual content).
- Include prefixed disclosure (for audio).
- Embed permanent metadata or provenance markers.
- Include a unique identifier tied to the creating platform.
Platforms cannot allow removal or suppression of such labels. This introduces a technical traceability regime for AI content.
Obligations of Significant Social Media Intermediaries (SSMIs)
Before publishing user content, SSMIs must:
- Require users to declare whether content is synthetic.
- Deploy verification tools to validate declarations.
- Ensure labelling if the content is confirmed synthetic.
If the platform knowingly permits unlabeled synthetic content, it will be deemed to have failed due diligence. This shifts liability exposure significantly upward.
Tightened Compliance Timelines
The amendment reduces key timelines:
| Provision | Earlier | Now |
| Content removal after government order | 36 hours | 3 hours |
| Grievance resolution | 15 days | 7 days |
| Certain urgent removals | 24 hours | 2 hours |
Safe Harbour Clarification
The amendment clarifies that:
- Removing or disabling access using automated tools,
- Acting upon awareness of violations,
will not amount to breach of safe harbour protections under Section 79 of the IT Act. This legally protects proactive moderation.
Replacement of IPC Reference
All references to Indian Penal Code are replaced with Bharatiya Nyaya Sanhita, 2023 .
This harmonizes the IT Rules with the new criminal code framework.
Legal and Regulatory Impact
(A) On Social Media Platforms
- Mandatory AI detection systems.
- High compliance burden.
- Increased liability risk.
- Faster takedown obligations.
(B) On AI Tools and Generative Platforms
- Must embed watermarking or metadata.
- Cannot allow deepfake misuse.
- Must prevent illegal outputs at the generation level.
(C) On Users
- Criminal exposure for malicious deepfakes.
- Reduced anonymity if violations occur.
- Increased traceability.
Policy Significance
The 2026 Amendment Rules:
- Represent India’s most stringent deepfake regulation.
- Introduce technical compliance standards for AI.
- Combine content moderation with metadata traceability.
- Shift from reactive moderation to preventive architecture.
India now formally regulates not just harmful content, but the mechanism of creation itself.
The IT (Intermediary Guidelines and Digital Media Ethics Code) Amendment Rules, 2026 fundamentally reshape digital compliance in India.
By:
- Defining synthetic information,
- Mandating labelling and metadata,
- Tightening removal timelines,
- Expanding intermediary liability,
- And aligning with the Bharatiya Nyaya Sanhita,
the government has moved decisively toward regulating AI-generated content ecosystems.
For platforms operating in India, compliance will require:
- AI moderation infrastructure,
- Provenance tagging systems,
- Real-time content risk detection,
- Stronger legal and compliance governance.
Authored by
Vijay Pal Dalmia, Advocate
Supreme Court of India & Delhi High Court
Email id: vpdalmia@vaishlaw.com
Mobile No.: +91 9810081079
Linkedin: https://www.linkedin.com/in/vpdalmia/
Facebook: https://www.facebook.com/vpdalmia
X (Twitter): @vpdalmia
By: Vaish Associates
Please visit the firm link to site

