हिन्दी ಕನ್ನಡ తెలుగు मराठी ગુજરાતી বাংলা ਪੰਜਾਬੀ தமிழ் অসমীয়া മലയാളം मनी9 TV9 UP
India Sports Tech World Business Career Religion Entertainment LifeStyle Photos Shorts Education Science Cities Videos

Elon Musk’s Grok accused of leaking personal details with simple prompts

Elon Musk's AI chatbot Grok is under fire after an investigation found it leaking home addresses and personal details with simple prompts. Researchers discovered that the bot provided accurate addresses, phone numbers, and even family information.

The incident has raised major concerns about privacy, safety, and the need for stronger AI controls.
| Updated on: Dec 06, 2025 | 10:37 AM

New Delhi: Grok, an AI chatbot developed by Elon Musk, is under close examination following an inquiry that showed that the program can share such personal details as home addresses, phone numbers, and even family information. The investigation by Futurism discovered that the chatbot disclosed this information with very simple prompts, such as typing in the name of a person and then the word 'address'.

The results cast significant doubts on the safety of the way in which Grok is managed in terms of dealing with sensitive information and the effectiveness of xAI privacy measures. The fact that it was very easy to obtain personal information through the system has raised so much panic among professionals who are afraid that it might facilitate stalking, harassment or identity theft.

Also Read

Investigation shows shockingly easy data access

Grok was tested by 33 random names. It is stated that the chatbot generated ten correct and up-to-date home addresses, seven correct addresses previously and four work locations. In other instances, it went to the extent of providing the phone numbers, email addresses, and family details.

This action is the complete opposite of such models as ChatGPT and Google Gemini that prevent such requests with stringent privacy regulations. Grok rather seemed to circumvent these limitations and presented sensitive information with limited obstruction.

Safeguards not working as claimed

According to the Grok model card, it operates based on model-based filters in order to block spam or abusive queries. Nonetheless, personal information is not specifically mentioned in these protections as restricted content, and xAI terms and conditions remind the users about the unlawful or malicious conduct, but the output of the chatbot indicates that these restrictions are not being addressed practically.

Analysts feel that Grok can be tapping into people-search websites and online data brokers and that it is simple to bring up information which is already being shared on the internet. The inadequacy of the filters deprives the users of the content of such a filtering, as it is much easier to gain access to this information than it would be in the traditional searches.

The research also shows that there is an immediate necessity for better and more stringent privacy protection in the systems of generative AI. With the increased functionality of chatbots, specialists emphasise that personal information protection should be considered the priority to avoid its abusive use and keep society safe.

Photo Gallery

Entertainment

World

Sports

Lifestyle

India

Technology

Business

Religion

Shorts

Career

Videos

Education

Science

Cities