
The Indian Computer Emergency Response Team (CERT-In) has warned against vulnerabilities in AI design, training and interaction mechanism
The cybersecurity watchdog noted that all the AI apps are not safe and therefore advised users signing up for them to consider using an anonymous account not linked to their personal or professional identity
As per CERT-In’s latest advisory, the “vulnerabilities” include technical issues such as data poisoning, adversarial attacks, model inversion, prompt injection and hallucination exploitation
The Indian Computer Emergency Response Team (CERT-In) has warned against multiple “vulnerabilities” in AI design, training and interaction mechanism.
The cybersecurity watchdog noted that all the AI apps are not safe and therefore advised users signing up for them to consider using an anonymous account not linked to their personal or professional identity.
As per the latest advisory, the “vulnerabilities” include technical issues such as data poisoning, adversarial attacks, model inversion, prompt injection and hallucination exploitation.
The advisory further said that AI has accelerated automating of routine tasks, fostering creativity and supporting business functions such as customer services, logistics, medical diagnosis and cybersecurity.
“Artificial Intelligence has become a hallmark of innovation, revolutionising industries ranging from healthcare to communications. AI is increasingly used to handle activities traditionally undertaken by humans,” CERT-In said.
The note highlights six major types of attacks that pose significant threats to AI applications’ security, reliability and trustworthiness:
Data poisoning: Manipulating training data to produce inaccurate or malicious outputs
Adversarial attacks: Changing inputs to trick AI models into giving wrong predictions
Model inversion: Extracting sensitive information about training data through analysis
Model stealing: Copying AI models by repeatedly querying them
Prompt injection: Introducing malicious instructions to bypass AI safeguards
Hallucination exploitation: Taking advantage of AI’s tendency to generate fabricated outputs
Threat actors can take advantage of the rising demand for AI apps to create fake apps designed to trick users into downloading them, the advisory said.
If someone downloads these fake AI apps on their devices, it maximises the opportunity to install malware designed to steal all their data, the advisory says, asking users to practice due diligence before clicking the ‘download’ button in order to minimise AI cybersecurity risks.
This also comes days after the Ministry of Electronics and Information Technology (MeitY) recommended setting up a dedicated AI governance board to review and authorise AI applications in the country.
The framework called for empowering the proposed board with powers to ensure AI initiatives align with legal instruments and address ethical considerations. In addition to the governance board, MeitY has also advocated for an “AI Ethics Committee” to design and integrate standard AI practices into all project stages, according to the competency framework.
The regulatory push also parallels rapid developments in India’s AI ecosystem, with both public and private sectors making significant investments. Earlier this month, MeitY launched the IndiaAI Compute Portal, a unified datasets platform AIKosha, and an accelerator programme for homegrown AI startups as part of the larger INR 10,300 Cr IndiaAI Mission.
In the private sector, Jio Platforms recently announced plans for a cloud-based PC to help users deploy compute-intensive AI applications, alongside developing “JioBrain,” a machine learning-as-a-service offering for enterprises.
According to a BCG and Nasscom report, India’s AI market could reach $17 billion by 2027 with a CAGR of 25-35%. IDC forecasts AI spending in India will hit $6 billion by 2027, growing at 33.7% annually from 2022.