Upgrade to Pro
— share decks privately, control downloads, hide ads and more …
Speaker Deck
Features
Speaker Deck
PRO
Sign in
Sign up for free
Search
Search
[ACL 2026 Demo] Fast-MIA: Efficient and Scalabl...
Search
Sponsored
·
Your Podcast. Everywhere. Effortlessly.
Share. Educate. Inspire. Entertain. You do you. We'll handle the rest.
→
Shotaro Ishihara
May 12, 2026
Research
1
0
Share
[ACL 2026 Demo] Fast-MIA: Efficient and Scalable Membership Inference for LLMs
https://arxiv.org/abs/2510.23074
https://github.com/Nikkei/fast-mia
Shotaro Ishihara
May 12, 2026
More Decks by Shotaro Ishihara
See All by Shotaro Ishihara
Fast-MIA: Efficient and Scalable Membership Inference for LLMs
upura
0
1
JAPAN AI CUP Prediction Tutorial
upura
2
1k
情報技術の社会実装に向けた応用と課題:ニュースメディアの事例から / appmech-jsce 2025
upura
0
360
日本語新聞記事を用いた大規模言語モデルの暗記定量化 / LLMC2025
upura
0
630
Quantifying Memorization in Continual Pre-training with Japanese General or Industry-Specific Corpora
upura
1
100
JOAI2025講評 / joai2025-review
upura
0
1.6k
AI エージェントを活用した研究再現性の自動定量評価 / scisci2025
upura
1
240
JSAI2025 企画セッション「人工知能とコンペティション」/ jsai2025-competition
upura
0
120
生成的推薦の人気バイアスの分析:暗記の観点から / JSAI2025
upura
0
390
Other Decks in Research
See All in Research
COFFEE-Japan PROJECT Impact Report(海ノ向こうコーヒー)
ontheslope
0
1.5k
英語教育 “研究” のあり方:学術知とアウトリーチの緊張関係
terasawat
1
930
[BlackHatAsia2026] Hidden Telemetry: Uncovering TraceLogging ETW Providers You're Not Using (Yet)
asuna_jp
1
380
東京大学工学部計数工学科、計数工学特別講義の説明資料
kikuzo
0
340
Using our influence and power for patient safety
helenbevan
0
340
【NICOGRAPH2025】Photographic Conviviality: ボディペイント・ワークショップによる 同時的かつ共生的な写真体験
toremolo72
0
230
2026年度 生成AI を活用した論文執筆ガイド/ワークショップ / 2026 Academic Year Guide to Writing Papers Using Generative AI - Workshop
ks91
PRO
0
120
「行ける・行けない表」による地域公共交通の性能評価
bansousha
0
140
存立危機事態の再検討
jimboken
0
280
Collective Predictive Coding and World Models in LLMs: A System 0/1/2/3 Perspective on Hierarchical Physical AI (IEEE SII 2026 Plenary Talk)
tanichu
1
380
Self-Hosted WebAssembly Runtime for Runtime-Neutral Checkpoint/Restore in Edge–Cloud Continuum
chikuwait
0
510
はじまりの クエスチョンブック —余暇と豊かさにあふれた社会とは?
culturaltransition
PRO
0
400
Featured
See All Featured
Raft: Consensus for Rubyists
vanstee
141
7.4k
State of Search Keynote: SEO is Dead Long Live SEO
ryanjones
0
190
Building a Scalable Design System with Sketch
lauravandoore
463
34k
Amusing Abliteration
ianozsvald
1
160
Scaling GitHub
holman
464
140k
Code Reviewing Like a Champion
maltzj
528
40k
The browser strikes back
jonoalderson
0
1k
コードの90%をAIが書く世界で何が待っているのか / What awaits us in a world where 90% of the code is written by AI
rkaga
61
43k
DevOps and Value Stream Thinking: Enabling flow, efficiency and business value
helenjbeal
1
180
What’s in a name? Adding method to the madness
productmarketing
PRO
24
4k
Navigating Weather and Climate Data
rabernat
0
180
Stop Working from a Prison Cell
hatefulcrawdad
274
21k
Transcript
Hiromu Takahashi and Shotaro Ishihara ACL 2026 System Demonstrations Fast-MIA:
Efficient and Scalable Membership Inference for LLMs
uv run --with vllm python main.py \ --config config/sample.yaml 1.
High-throughput batch inference using vLLM (about 5 times faster individually) 2. Cross-method caching architecture (Reduce the total processing time for benchmarking multiple methods) https://github.com/Nikkei/fast-mia Fast-MIA: Efficient and Scalable 2 LLM LOSS vLLM backend batch inference Shared Cache Reuse across methods PPL/zlib Min-K% Prob DC-PDD Lowercase PAC ReCaLL Con-ReCall SaMIA ……
Membership Inference Attack (MIA) on LLMs 3 LLM Is this
text included? Text Pre-training Data • Calculate the log-likelihood, etc. • Various methods have been proposed.
Challenges in MIA on LLMs 4 LLM Is this text
included? Text Pre-training Data • Calculate the log-likelihood, etc. • Various methods have been proposed. 1. Growing computational demands for individual MIA methods. 2. Redundant computation across methods for benchmarking.
We introduce Fast-MIA 5 1. Growing computational demands for individual
MIA methods. 2. Redundant computation across methods for benchmarking. LLM LOSS vLLM backend batch inference Shared Cache Reuse across methods PPL/zlib Min-K% Prob DC-PDD Lowercase PAC ReCaLL Con-ReCall SaMIA …… 1. High-throughput batch inference using vLLM. 2. Cross-method caching architecture.
uv run --with vllm python main.py \ --config config/sample.yaml How
to Use: https://github.com/Nikkei/fast-mia 6 model: model_id: "huggyllama/llama-30b" data: data_path: "swj0419/WikiMIA" format: "huggingface" text_length: 32 methods: - type: "loss"
AUC Reproducibility and Speed 7 Left: Fast-MIA Right: Transformers-based implementations
Inference time (the number of inferences) The cache is working
8
uv run --with vllm python main.py \ --config config/sample.yaml 1.
High-throughput batch inference using vLLM (about 5 times faster individually) 2. Cross-method caching architecture (Reduce the total processing time for benchmarking multiple methods) https://github.com/Nikkei/fast-mia Contributions Welcome 9 LLM LOSS vLLM backend batch inference Shared Cache Reuse across methods PPL/zlib Min-K% Prob DC-PDD Lowercase PAC ReCaLL Con-ReCall SaMIA ……