CoreTechX On-Prem
Infrastructure Planner
Rough per-H100 sizing: users, workloads, and supporting CPU, memory, and storage.
Planning unit
1x NVIDIA H100
H100 GPU count
1
1
8
OCR pipeline
Page image processing on local GPU and CPU
Semantic indexing
Embedding edited page content locally
AI assistant
Chat and assistant requests served by local LLM
Planning assumptions
Local LLM size
Small
Medium
Large
Extra large
Embedding size
Small
Medium
Large
Peak load factor
i
Accounts for busy-hour concentration. Higher values reduce estimated users because the same daily work arrives in shorter peak windows.
Low
Standard
High
Storage retention
90 days
180 days
365 days
2 years