https://github.com/AI-Hypercomputer/JetStream/blob/main/docs/observability-prometheus-metrics-in-jetstream-server.md only mentions the prometheus_port if the jetstream runs with maxengine_server.
However, such option doesn't exist under https://github.com/AI-Hypercomputer/jetstream-pytorch. I wonder if jetstream-pytorch also expose prometheus metrics related to the inference server?