Show HN: Context Gateway – Compress agent context before it hits the LLM
TL;DR Highlight
A proxy that sits between AI agents (Claude Code, Cursor, etc.) and LLM APIs to intercept, log, and modify requests in real time.
Who Should Read
Developers building or debugging AI agent systems who want observability and control over what gets sent to LLM APIs, and security engineers monitoring agent behavior.
Core Mechanics
- The proxy intercepts all API calls between agents (Claude Code, Cursor, Cline, etc.) and the underlying LLM APIs (Anthropic, OpenAI, etc.).
- It provides real-time logging of requests and responses, including the full prompt/completion, token counts, and latency.
- Beyond logging, it can modify requests on the fly — injecting additional context, filtering out sensitive data, or redirecting to a different model.
- Use cases: debugging unexpected agent behavior, monitoring for prompt injection attempts, enforcing data privacy policies, and cost attribution by agent.
- The proxy is positioned as 'insert once, observe everything' — configure your agent's API base URL to point to the proxy and all traffic flows through it.
Evidence
- The tool was demonstrated with real Claude Code sessions, showing the full context window content that agents send to the API.
- HN commenters found the full-context logging surprisingly revealing — agents often include much more context than users expect.
- Security-focused commenters liked the prompt injection detection angle — being able to see exactly what's being sent helps catch adversarial content in tool outputs.
- Some noted that any proxy adds latency and becomes a potential single point of failure — important to design for fallback.
How to Apply
- Point your agent's API base URL to the proxy during development/debugging — understanding exactly what gets sent to the API is essential for iterating on prompts and tools.
- Use the proxy to enforce data privacy policies: configure it to filter or redact PII before requests leave your infrastructure.
- For cost attribution: use the proxy to tag requests by agent/project and track token usage per workload.
- In production, deploy the proxy with redundancy and a fallback to direct API calls — don't let the observability layer become a reliability bottleneck.
Code Example
# 1. Install
curl -fsSL https://compresr.ai/api/install | sh
# 2. Run TUI wizard (select agent, configure model/API key, set threshold)
context-gateway
# 3. Check compression logs
cat logs/history_compaction.jsonlTerminology
Related Papers
Training an LLM in Swift, Part 1: Taking matrix mult from Gflop/s to Tflop/s
Apple Silicon에서 Swift로 직접 행렬 곱셈 커널을 구현하며 CPU, SIMD, AMX, GPU(Metal)를 단계별로 최적화해 Gflop/s에서 Tflop/s 수준까지 성능을 높이는 과정을 상세히 설명한 글이다. 프레임워크 없이 LLM 학습의 핵심 연산을 밑바닥부터 구현하고 싶은 개발자에게 Apple Silicon의 성능 한계를 체감할 수 있는 드문 자료다.
Removing fsync from our local storage engine
FractalBits가 fsync 없이 SSD 전용 KV 스토리지 엔진을 구현해 동일 조건 대비 약 65% 높은 쓰기 성능을 달성한 설계 방법을 공유했다. fsync의 메타데이터 오버헤드를 피하기 위해 사전 할당, O_DIRECT, SSD 원자 쓰기 단위 정렬 저널을 조합한 구조가 핵심이다.
Google Chrome silently installs a 4 GB AI model on your device without consent
Google Chrome이 사용자 동의 없이 Gemini Nano 4GB 모델 파일을 자동 다운로드하고, 삭제해도 재다운로드되는 문제가 발견됐다. GDPR 위반 가능성과 수십억 대 기기에 적용될 때의 환경 비용 문제가 제기되고 있다.
How OpenAI delivers low-latency voice AI at scale
OpenAI redesigned its WebRTC stack to serve real-time voice AI to over 900 million users, detailing the design decisions and trade-offs of a relay + transceiver split architecture.
Efficient Test-Time Inference via Deterministic Exploration of Truncated Decoding Trees
Deterministic Leaf Enumeration (DLE) cuts self-consistency’s redundant sampling by deterministically exploring a tree of possible sequences, simultaneously improving math/code reasoning performance and speed.