I read from Asymptotic Statistics (Van der Vaart, 1998) Theorem 5.7 that, to let the M-estimator $\hat{\theta}_n$ converge in probability to $\theta_0$, two conditions need to be satisfied, one of which is for every $\epsilon>0$, $$\sup_{\theta:d(\theta,\theta_0)\geq\epsilon}M(\theta)<M(\theta_0).$$
The inequality looks complexed, involving the metric $d$, which puzzles me. Why can't we remove the metric, expressing it as $$\forall\theta\neq\theta_0,M(\theta)<M(\theta_0)$$ Aren't the two equivalent?