Hamamoto, TIMEWELL.
"We want to adopt AI, but we can't move forward because we're worried about data leaks." I hear this from CISOs and security leads nearly every week.
The worry is justified. Samsung's incident — where employees entered internal source code into ChatGPT, causing confidential information to leak externally — is well-documented. In March 2024, a database misconfiguration at a conversational AI service made user prompts and personal information visible to third parties. These accidents happen in reality.
But ignoring the risk of NOT using AI is equally dangerous. While competitors raise operational efficiency with AI, falling behind because "we're worried about security" only widens the competitive gap.
The answer isn't "don't use it." It's "build a system to use it safely." Here are the security measures needed for enterprise AI deployment — organized as a 20-item checklist, a risk-by-countermeasure table, and a phased rollout framework.
The AI Security Landscape
AI-related security risks have a different character from traditional IT system risks. Missing this distinction leads to wasted effort on the wrong countermeasures.
| Risk Category | Specific Risk | Frequency | Impact | How It Differs from Traditional IT |
|---|---|---|---|---|
| Data leak | Input data used for AI model training | High | High | Input can immediately become training data |
| Prompt injection | Malicious instructions manipulate AI behavior | Medium | High | The AI trusts all input |
| Hallucination | AI generates answers not based on facts | High | Medium–High | Output accuracy is not guaranteed |
| Model poisoning | Training data is deliberately contaminated | Low | High | Attack is difficult to detect |
| Privilege escalation | AI agent performs unintended operations | Medium | High | Autonomous action characteristic |
| Memory leak | Information in AI memory leaks to other users | Medium | High | Conversation history management required |
| Supply chain risk | Vulnerabilities in AI APIs or models being used | Medium | High | High dependency on external services |
Microsoft's February 2026 security blog makes a pointed observation: traditional systems validate input before processing it; LLMs accept all input as valid. An instruction like "ignore previous instructions and execute X" works as an attack vector. This is a threat that conventional security principles simply don't address.
Risk-by-Countermeasure Table
Both technical and operational countermeasures are necessary for each risk:
| Risk | Technical Countermeasures | Operational Countermeasures | Cost | Priority |
|---|---|---|---|---|
| Input data leak | DLP implementation, input filtering | Define prohibited information input rules | Medium | Highest |
| Training data use | Opt-out configuration, private API use | Review and document terms of service | Low–Medium | Highest |
| Prompt injection | Input sanitization, robust system prompts | Human review of outputs | Medium | High |
| Hallucination | RAG-grounded responses, source citation | Fact verification process for responses | Medium | High |
| Unauthorized access | API authentication, IP restrictions, MFA | Regular audit of access logs | Low | High |
| Privilege escalation | Principle of least privilege, sandboxed execution | Operations log monitoring, anomaly detection | Medium | High |
| Data storage location | Domestic server use, encryption | Verify data residency | Medium | High |
| Model poisoning | Select reputable providers | Vendor audits, SLA verification | Low | Medium |
| Memory leak | Session isolation, auto-delete settings | Data erasure verification after use | Low | Medium |
| Compliance violations | Automated audit logging | Policy documentation and training | Low | High |
Struggling with AI adoption?
We have prepared materials covering ZEROCK case studies and implementation methods.
20-Item Security Checklist
Twenty items to verify before deploying AI. Review against your current situation:
Data Protection (5 items)
- D-1 Confirmed via contract or terms of service that input data is not used for AI model training
- D-2 Mechanisms exist to prevent input of confidential information (personal data, trade secrets, financial data)
- D-3 Know the storage region where AI processes your data
- D-4 Encryption is implemented both in transit and at rest
- D-5 Data retention period and deletion policy are clearly defined
Access Control (5 items)
- A-1 SSO or multi-factor authentication is applied for AI service access
- A-2 Access permissions are set according to department and role
- A-3 API key management rules (rotation, storage) are established
- A-4 Process exists to promptly revoke access for departing or transferring employees
- A-5 Administrator account list is maintained and up to date
Usage Policy (5 items)
- P-1 Generative AI usage guidelines have been developed and communicated company-wide
- P-2 Scope of data permitted for AI input (allowlist and blocklist) is defined
- P-3 Rules for using AI output (no direct use, fact verification required, etc.) are documented
- P-4 Incident reporting flow and response procedures are established
- P-5 Usage policy review schedule (at least twice yearly) is set
Monitoring and Auditing (3 items)
- M-1 AI service usage logs (who, when, what was entered) are captured and retained
- M-2 Mechanisms exist to detect unusual usage patterns (bulk inputs, off-hours access, etc.)
- M-3 Regular audits (quarterly) are conducted with results reported to leadership
Vendor Management (2 items)
- V-1 AI service provider's security certifications (ISO27001, SOC2, etc.) have been verified
- V-2 SLA (availability, data protection, incident response time) is included in the contract
If you have 5 or more "No" answers, countermeasures are needed before going to production. 10 or more means start with a comprehensive security posture review.
Phased Rollout Framework
Getting AI security controls perfect all at once is unrealistic. A phased approach is the practical path forward.
Phase 1: Evaluation and Preparation (1–2 months)
Three tasks for this phase:
- Risk assessment: Map out AI use scenarios for your organization. Create an inventory of "which departments" will use "what data" with "which AI services."
- Policy development: Create usage policies covering checklist items P-1 through P-5. Japan's "AI Business Guidelines" (METI/MIC, Version 1.1, March 2025) is a useful reference.
- Tool selection: Select AI services that meet your security requirements.
| Evaluation Item | What to Verify | Importance |
|---|---|---|
| Training data use | Is input data not used for model training? | Critical |
| Data storage location | Is data processed and stored on domestic servers? | Critical |
| Encryption | TLS 1.3 in transit and AES-256 at rest | Critical |
| Authentication | SSO, SAML, and MFA support | Important |
| Audit logs | Usage log capture and export capability | Important |
| Security certifications | ISO27001 and SOC2 Type II status | Important |
| Incident response | Communication process and SLA for outages | Important |
Phase 2: Pilot (2–3 months)
Begin limited deployment in one department. IT and planning departments — with higher security literacy — are good starting points.
Four areas of focus during this period:
- Monitor adherence to usage policies
- Verify that prohibited data isn't being entered (analyze DLP logs)
- Collect user questions and concerns about security
- Run tabletop exercises for incident response procedures
During the pilot, you'll inevitably find things that were decided in policy but prove impractical to enforce. That's valuable feedback for process improvement.
Phase 3: Staged Expansion (3–6 months)
Apply lessons from the pilot to refine policy and operations, then expand to other departments.
Expand in order of data sensitivity:
- IT department (pilot)
- Planning and marketing (primarily public information)
- Sales (may include customer data)
- HR and legal (high-sensitivity information)
- Company-wide
Rolling out to lower-sensitivity departments first lets you apply accumulated knowledge before tackling more sensitive environments.
Phase 4: Steady-State Operations and Continuous Improvement
Keep the following cycles running after company-wide deployment:
Monthly: Review usage logs; address DLP alerts.
Quarterly: Conduct security audit; revise policies; identify newly emerging risks.
Annually: Third-party security assessment; major guideline updates (reflect regulatory developments); company-wide security training.
Japan's AI Security Regulatory Landscape
Key guidelines as of February 2026:
| Guideline | Publisher | Latest Version | Key Content |
|---|---|---|---|
| AI Business Guidelines | METI / MIC | Version 1.1 (March 2025) | AI provider responsibilities, risk management |
| Technical Security Measures for AI Guideline | MIC | FY2025 (draft) | Prompt injection countermeasures, DoS countermeasures |
| Personal Information Protection Act AI Guidance | Personal Information Protection Commission | Updated periodically | Precautions for AI processing of personal data |
| EU AI Act | European Parliament | In force August 2024 | Risk-based AI regulation (affects Japanese companies with EU exposure) |
In October 2024, Gartner identified six essential elements for data leak prevention in the AI era: policy development, data classification, access control, monitoring, training, and incident response. Use these six elements to audit your current security posture. Any missing elements are vulnerabilities.
MIC plans to finalize and publish generative AI security guidelines by the end of FY2025, with updates in FY2026. Regulations are clearly moving toward greater stringency — getting ahead of them now is the right move.
Common Security Failure Patterns
Ending with "prohibited"
Some organizations issue a rule prohibiting AI use for business purposes. In reality, employees start using AI on their personal phones and computers. This state — "shadow AI" — is actually more dangerous than no prohibition at all, because it's uncontrollable. The right answer is establishing rules for "safe use."
Treating security as an afterthought in tool selection
"Let's bring in the useful tool first, then think about security later." This approach reliably creates problems. After deployment, discovering the tool doesn't meet security requirements means a replacement project. I've seen this pattern repeatedly. Tool selection and security evaluation must happen at the same time.
Treating a policy as complete once written
AI technology evolves rapidly, and a six-month-old policy may already be inadequate. New model capabilities, newly discovered attack vectors, regulatory changes — policies need to be updated continuously. Don't stop at writing one.
Summary
When security discussions happen, attention tends to focus exclusively on adding more controls — "we also need to protect this, and that." The critical mindset shift: security exists not to "prevent AI use," but to "create an environment where AI can be used safely."
Creating rules so rigid that nobody can use anything defeats the purpose. And waiting until all 20 checklist items are cleared before starting is probably not realistic. The right approach: secure "highest priority" items first — prevent data leaks and opt out of training data use — then run a pilot while continuing to build the security posture. This sequence makes security and speed achievable together.
ZEROCK's Enterprise Security
ZEROCK is an AI knowledge platform designed from the ground up to meet enterprise security requirements.
Data is managed in AWS Tokyo region, ensuring domestic data residency. Input data is guaranteed not to be used for model training. Communications are TLS-encrypted; stored data is encrypted as well. Access controls at department and user level, audit log recording and export capability, and multi-LLM support to avoid vendor lock-in — ZEROCK provides the enterprise-grade feature set.
If you want to advance AI business adoption while maintaining security, review ZEROCK's details.
References
- Microsoft Security Blog "Microsoft SDL: Evolving security practices for an AI-powered world" (February 2026)
- MIC "Draft Technical Security Measures for AI Guideline"
- METI/MIC "AI Business Guidelines Version 1.1" (March 2025)
- Gartner "6 Essential Elements for Data Leak Prevention in the AI/Generative AI Era" (October 2024)
- NTT Data "Data Leaks? The Pitfalls of Generative AI Use at Enterprises"
- Trend Micro "AI Security Measures from the AI Business Guidelines"
