diff --git a/data-loader/cli/src/main/java/com/scalar/db/dataloader/cli/command/dataimport/ImportCommandOptions.java b/data-loader/cli/src/main/java/com/scalar/db/dataloader/cli/command/dataimport/ImportCommandOptions.java index aa8ecd5a88..d2e4b96e3d 100755 --- a/data-loader/cli/src/main/java/com/scalar/db/dataloader/cli/command/dataimport/ImportCommandOptions.java +++ b/data-loader/cli/src/main/java/com/scalar/db/dataloader/cli/command/dataimport/ImportCommandOptions.java @@ -145,7 +145,8 @@ public class ImportCommandOptions { @CommandLine.Option( names = {"--data-chunk-size", "-dcs"}, paramLabel = "", - description = "Maximum number of records to be included in a single data chunk", + description = + "Maximum number of records to be included in a single data chunk (default: 500)", defaultValue = "500") protected int dataChunkSize; @@ -153,21 +154,22 @@ public class ImportCommandOptions { names = {"--transaction-size", "-ts"}, paramLabel = "", description = - "Maximum number of put operations that are grouped together into one ScalarDB distributed transaction, only supported in ScalarDB transaction mode", + "Maximum number of put operations that are grouped together into one ScalarDB distributed transaction, only supported in ScalarDB transaction mode (default: 100)", defaultValue = "100") protected int transactionSize; @CommandLine.Option( names = {"--split-log-mode", "-slm"}, paramLabel = "", - description = "Split log file into multiple files based on data chunks", + description = "Split log file into multiple files based on data chunks (default: false)", defaultValue = "false") protected boolean splitLogMode; @CommandLine.Option( names = {"--data-chunk-queue-size", "-qs"}, paramLabel = "", - description = "Maximum number of data chunks that can be kept at a time for processing", + description = + "Maximum number of data chunks that can be kept at a time for processing (default: 256)", defaultValue = "256") protected int dataChunkQueueSize;