We introduce SciReplicate-Bench, the first benchmark targeting the reliable reproduction of algorithmic results described in scientific papers with agentic LLM pipelines. The benchmark covers memory management, tool grounding, and execution tracking, enabling systematic evaluation of agent behaviors. Project resources are available on the website.