File: fdupes.1

package info (click to toggle)
fdupes 1:1.6.1-2
  • links: PTS, VCS
  • area: main
  • in suites: bullseye, buster, sid
  • size: 200 kB
  • sloc: ansic: 1,319; makefile: 43
file content (143 lines) | stat: -rw-r--r-- 3,575 bytes parent folder | download | duplicates (2)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
.TH FDUPES 1
.\" NAME should be all caps, SECTION should be 1-8, maybe w/ subsection
.\" other parms are allowed: see man(7), man(1)
.SH NAME
fdupes \- finds duplicate files in a given set of directories
.SH SYNOPSIS
.B fdupes
[
.I options
]
.I DIRECTORY
\|.\|.\|.

.SH "DESCRIPTION"
Searches the given path for duplicate files. Such files are found by
comparing file sizes and MD5 signatures, followed by a 
byte-by-byte comparison.

.SH OPTIONS
.TP
.B -r --recurse
for every directory given follow subdirectories encountered within
.TP
.B -R --recurse:
for each directory given after this option follow subdirectories
encountered within (note the ':' at the end of option; see the
Examples section below for further explanation)
.TP
.B -s --symlinks
follow symlinked directories
.TP
.B -H --hardlinks
normally, when two or more files point to the same disk area they are
treated as non-duplicates; this option will change this behavior
.TP
.B -n --noempty
exclude zero-length files from consideration
.TP
.B -f --omitfirst
omit the first file in each set of matches
.TP
.B -A --nohidden
exclude hidden files from consideration
.TP
.B -1 --sameline
list each set of matches on a single line
.TP
.B -S --size
show size of duplicate files
.TP
.B -m --summarize
summarize duplicate files information
.TP
.B -q --quiet
hide progress indicator
.TP
.B -d --delete
prompt user for files to preserve, deleting all others (see
.B CAVEATS
below)
.TP
.B -N --noprompt
when used together with \-\-delete, preserve the first file in each
set of duplicates and delete the others without prompting the user 
.TP
.B -I --immediate
delete duplicates as they are encountered, without
grouping into sets; implies --noprompt
.TP
.B -p --permissions
don't consider files with different owner/group or permission bits as duplicates
.TP
.B -o --order\fR=\fIWORD\fR
order files according to WORD:
time - sort by mtime, name - sort by filename
.TP
.B -i --reverse
reverse order while sorting
.TP
.B -v --version
display fdupes version
.TP
.B -h --help
displays help
.SH "SEE ALSO"
.\" Always quote multiple words for .SH
.BR md5sum (1)
.SH NOTES
Unless
.B -1
or
.B --sameline
is specified, duplicate files are listed
together in groups, each file displayed on a separate line. The
groups are then separated from each other by blank lines.

When
.B -1
or
.B --sameline
is specified, spaces and backslash characters  (\fB\e\fP) appearing
in a filename are preceded by a backslash character.

.SH EXAMPLES
.TP
.B fdupes a --recurse: b
will follow subdirectories under b, but not those under a.
.TP
.B fdupes a --recurse b
will follow subdirectories under both a and b.

.SH CAVEATS
If fdupes returns with an error message such as
.B fdupes: error invoking md5sum
it means the program has been compiled to use an external
program to calculate MD5 signatures (otherwise, fdupes uses
internal routines for this purpose), and an error has occurred
while attempting to execute it. If this is the case, the 
specified program should be properly installed prior 
to running fdupes.

When using
.B \-d
or
.BR \-\-delete ,
care should be taken to insure against
accidental data loss.

When used together with options
.B \-s
or
.BR \-\-symlink ,
a user could accidentally
preserve a symlink while deleting the file it points to.

Furthermore, when specifying a particular directory more than
once, all files within that directory will be listed as their
own duplicates, leading to data loss should a user preserve a
file without its "duplicate" (the file itself!).

.SH AUTHOR
Adrian Lopez <adrian2@caribe.net>