From eb854c3e0332cfcf66e65d2d991ac1cc5633ec9b Mon Sep 17 00:00:00 2001
From: Xuan Gu <xuan.gu@liu.se>
Date: Fri, 24 Feb 2023 12:51:50 +0000
Subject: [PATCH] Update README.md

---
 README.md | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)

diff --git a/README.md b/README.md
index 7a11d32..980a4d1 100644
--- a/README.md
+++ b/README.md
@@ -115,4 +115,4 @@ Ref: https://pytorch-lightning.readthedocs.io/en/stable/common/trainer.html
 - Use as large batch_size as possible for a more stable benchmark result. For single node, use 256; for multi-node, use 128.  
 - Benchmarking with dim = 2, nodes = 1, gpus = 8, batch_size = 128, 256 takes ~2mins.  
 - Specify the paths for enroot cache and data, see this [page](https://gitlab.liu.se/xuagu37/run-pytorch-and-tensorflow-containers-with-nvidia-enroot#set-path-to-user-container-storage).
-- (20220222) For single node case, command ```srun singularity/enroot``` gives slightly worse throughput performance compared with running ```singularity/enroot``` directly. Command ```srun enroot``` stopped working for multi-node case.
+- (20220222) ```srun enroot ...```stopped working for multi-node case. Use pyxis instead. See the script ```benchmark_multi_node.sbatch```.
-- 
GitLab