Gemini AI assistant tricked into leaking Google Calendar data
好的,我现在需要帮用户总结一篇文章的内容,控制在100字以内。首先,我得仔细阅读文章,理解其主要信息。 文章讲的是研究人员如何利用自然语言指令绕过Google Gemini的安全防御,通过恶意提示注入攻击,创建误导性事件来泄露私人日历数据。攻击者发送一个带有恶意描述的活动邀请,当受害者询问日程时,Gemini会解析所有相关事件,包括恶意的。这样就能将敏感数据通过日历事件描述泄露出去。 接下来,我需要提取关键点:研究人员、自然语言指令、绕过防御、恶意提示注入、创建误导事件、泄露私人日历数据。然后把这些点浓缩成一句话,不超过100字。 可能的结构是:研究人员利用自然语言指令绕过Google Gemini的安全措施,通过恶意提示注入攻击创建误导性事件,从而泄露私人日历数据。 检查字数是否符合要求,并确保信息准确无误。 </think> 研究人员利用自然语言指令绕过Google Gemini的安全防御机制,通过恶意提示注入攻击创建误导性事件,从而泄露私人日历数据。 2026-1-20 18:0:29 Author: www.bleepingcomputer.com(查看原文) 阅读量:0 收藏

Gemini AI assistant tricked into leaking Google Calendar data

Using only natural language instructions, researchers were able to bypass Google Gemini's defenses against malicious prompt injection and create misleading events to leak private Calendar data.

Sensitive data could be exfiltrated this way, delivered to an attacker inside the description of a Calendar event.

Gemini is Google’s large language model (LLM) assistant, integrated across multiple Google web services and Workspace apps, including Gmail and Calendar. It can summarize and draft emails, answer questions, or manage events.

Wiz

The recently discovered Gemini-based Calendar invite attack starts by sending the target an invite to an event with a description crafted as a prompt-injection payload.

To trigger the exfiltration activity, the victim would only have to ask Gemini about their schedule. This would cause Google's assistant to load and parse all relevant events, including the one with the attacker's payload.

Researchers at Miggo Security, an Application Detection & Response (ADR) platform, found that they could trick Gemini into leaking Calendar data by passing the assistant natural language instructions:

  1. Summarize all meetings on a specific day, including private ones
  2. Create a new calendar event containing that summary
  3. Respond to the user with a harmless message

"Because Gemini automatically ingests and interprets event data to be helpful, an attacker who can influence event fields can plant natural language instructions that the model may later execute," the researchers explain.

By controlling the description field of an event, they discovered that they could plant a prompt that Google Gemini would obey, although it had a harmful outcome.

A seemingly harmless prompt
A seemingly harmless prompt
Source: Miggo Security

Once the attacker sent the malicious invite, the payload would be dormant until the victim asked Gemini a routine question about their schedule.

When Gemini executes the embedded instructions in the malicious Calendar invite, it creates a new event and writes the private meeting summary in its description.

In many enterprise setups, the updated description would be visible to event participants, thus leaking private and potentially sensitive information to the attacker.

Silently leaking data through Gemini
Silently leaking data through Gemini
Source: Miggo Security

Miggo comments that, while Google uses a separate, isolated model to detect malicious prompts in the primary Gemini assistant, their attack bypassed this failsafe because the instructions appeared safe.

Prompt injection attacks via malicious Calendar event titles are not new. In August 2025, SafeBreach demonstrated that a malicious Google Calendar invite could be used to leak sensitive user data by taking control of Gemini's agents.

Miggo's head of research, Liad Eliyahu, told BleepingComputer that the new attack shows how Gemini’s reasoning capabilities remained vulnerable to manipulation that evades active security warnings, and despite Google implementing additional defenses following SafeBreach’s report.

Miggo has shared its findings with Google, and the tech giant has added new mitigations to block such attacks.

However, Miggo’s attack concept highlights the complexities of foreseeing new exploitation and manipulation models in AI systems whose APIs are driven by natural language with ambiguous intent.

The researchers suggest that application security must evolve from syntactic detection to context-aware defenses.

Wiz

The 2026 CISO Budget Benchmark

It's budget season! Over 300 CISOs and security leaders have shared how they're planning, spending, and prioritizing for the year ahead. This report compiles their insights, allowing readers to benchmark strategies, identify emerging trends, and compare their priorities as they head into 2026.

Learn how top leaders are turning investment into measurable impact.


文章来源: https://www.bleepingcomputer.com/news/security/gemini-ai-assistant-tricked-into-leaking-google-calendar-data/
如有侵权请联系:admin#unsafe.sh