# HG changeset patch # User Mikael Berthe # Date 1539530857 -7200 # Node ID 3fa13770e97066f45294ec6982c9e0903bd035c7 # Parent bce80b708ddbff0c9c08b08638ae4ec44b571bfb Rename JSON size fields Make the field names more explicit... diff -r bce80b708ddb -r 3fa13770e970 goduf.go --- a/goduf.go Sun Oct 14 15:45:30 2018 +0200 +++ b/goduf.go Sun Oct 14 17:27:37 2018 +0200 @@ -61,20 +61,20 @@ // Results contains the results of the duplicates search type Results struct { - Groups []ResultSet `json:"groups"` // List of duplicate sets - Duplicates uint `json:"duplicates"` // Number of duplicates - NumberOfSets uint `json:"number_of_sets"` // Number of duplicate sets - RedundantDataSize uint64 `json:"redundant_data_size"` // Redundant data size - RedundantDataSizeH string `json:"redundant_data_size_h"` // Same, human-readable - TotalFileCount uint `json:"total_file_count"` // Total number of checked files - TotalSize uint64 `json:"total_size"` // Total size for checked files - TotalSizeH string `json:"total_size_h"` // Same, human-readable + Groups []ResultSet `json:"groups"` // List of duplicate sets + Duplicates uint `json:"duplicates"` // Number of duplicates + NumberOfSets uint `json:"number_of_sets"` // Number of duplicate sets + RedundantDataSizeBytes uint64 `json:"redundant_data_size_bytes"` // Redundant data size + RedundantDataSizeHuman string `json:"redundant_data_size_human"` // Same, human-readable + TotalFileCount uint `json:"total_file_count"` // Total number of checked files + TotalSizeBytes uint64 `json:"total_size_bytes"` // Total size for checked files + TotalSizeHuman string `json:"total_size_human"` // Same, human-readable } // ResultSet contains a group of identical duplicate files type ResultSet struct { - Size uint64 `json:"size"` // Size of each item - Paths []string `json:"paths"` // List of file paths + FileSize uint64 `json:"file_size"` // Size of each item + Paths []string `json:"paths"` // List of file paths } type fileObj struct { @@ -523,8 +523,8 @@ size := uint64(l[0].Size()) // We do not count the size of the 1st item // so we get only duplicate size. - results.RedundantDataSize += size * uint64(len(l)-1) - newSet := ResultSet{Size: size} + results.RedundantDataSizeBytes += size * uint64(len(l)-1) + newSet := ResultSet{FileSize: size} for _, f := range l { newSet.Paths = append(newSet.Paths, f.FilePath) results.Duplicates++ @@ -532,10 +532,10 @@ results.Groups = append(results.Groups, newSet) } results.NumberOfSets = uint(len(results.Groups)) - results.RedundantDataSizeH = formatSize(results.RedundantDataSize, true) + results.RedundantDataSizeHuman = formatSize(results.RedundantDataSizeBytes, true) results.TotalFileCount = data.cmpt - results.TotalSize = data.totalSize - results.TotalSizeH = formatSize(data.totalSize, true) + results.TotalSizeBytes = data.totalSize + results.TotalSizeHuman = formatSize(data.totalSize, true) return results, nil } diff -r bce80b708ddb -r 3fa13770e970 output.go --- a/output.go Sun Oct 14 15:45:30 2018 +0200 +++ b/output.go Sun Oct 14 17:27:37 2018 +0200 @@ -62,7 +62,7 @@ if !summaryOnly { for i, g := range results.Groups { fmt.Printf("\nGroup #%d (%d files * %v):\n", i+1, - len(g.Paths), formatSize(g.Size, true)) + len(g.Paths), formatSize(g.FileSize, true)) for _, f := range g.Paths { fmt.Println(f) } @@ -81,7 +81,7 @@ myLog.Println(0, "Final count:", results.Duplicates, "duplicate files in", len(results.Groups), "sets") myLog.Println(0, "Redundant data size:", - formatSize(results.RedundantDataSize, false)) + formatSize(results.RedundantDataSizeBytes, false)) } func displayResultsJSON(results Results) {