Browse Source

Fixed logging level in distributed mode (#4284)

Co-authored-by: fkwong <huangfuqiang@transai.cn>
modifyDataloader
imyhxy GitHub 3 years ago
parent
commit
771ac6c53d
No known key found for this signature in database GPG Key ID: 4AEE18F83AFDEB23
1 changed files with 1 additions and 1 deletions
  1. +1
    -1
      utils/torch_utils.py

+ 1
- 1
utils/torch_utils.py View File

@@ -23,7 +23,6 @@ try:
except ImportError:
thop = None

logging.basicConfig(format="%(message)s", level=logging.INFO)
LOGGER = logging.getLogger(__name__)


@@ -108,6 +107,7 @@ def profile(input, ops, n=10, device=None):
# profile(input, [m1, m2], n=100) # profile over 100 iterations

results = []
logging.basicConfig(format="%(message)s", level=logging.INFO)
device = device or select_device()
print(f"{'Params':>12s}{'GFLOPs':>12s}{'GPU_mem (GB)':>14s}{'forward (ms)':>14s}{'backward (ms)':>14s}"
f"{'input':>24s}{'output':>24s}")

Loading…
Cancel
Save