Duplicate File Finder User Manual

Previous: Known Deficiencies    Contents    Next: Java 1.4.x Deficiencies

Part 7:
Future Plans for Duplicate File Finder

I have significant future plans for Duplicate File Finder. These include in no particular order:

  1. Inclusion of a Bad Messages window. This is for if Duplicate File Finder is unable to find or read a file for example.

  2. Include an option to generate file checksums only if file sizes are identical.

  3. Include an option limiting file searching to files that fall within a user specified file size range (minimum and maximum file sizes).

  4. Exclusion of directories or files from a search.

  5. Filter to examine or exclude only files with certain extensions (very windows like though).

  6. Have a switchable default to ignore directories named "cache".

  7. Detection of if full directories match each other.

  8. Removal of files from lists without deletion.

  9. Moving found duplicate files to another location.

  10. Replace a duplicated file with a link to another identical, existing file.

  11. Move a duplicated file to the users trash directory.

  12. Ability to mark files in multiple groups for a later single step processing (delete, move, ...)

  13. Recognition of problem file names.

  14. Recognition of broken links. This is already kind of implemented, but you have to run Duplicate File Finder from the command line to get a hint that there might be a broken link.

  15. An option to view a file in its normal application before processing it.

  16. With image files, create a list of files, either automatically or a user defined list, and view all those files in the configured external viewer.

  17. Addition of an auto delete/move function based on certain customisable rules. These rules would be for duplicate files in the same directory, and to be able to specify that files in certain specified directories have duplicates in other specified directories, the files in the other directories should be automatically deleted/moved, and deleting the oldest or newest files.

  18. Be able to specify directories in which duplicate files should not be deleted.

  19. Defining if the auto delete function should really delete or just move files to the users trash directory.

  20. A protection mechanism to ensure you don't accidentially delete all files in a duplicated group leaving you with no copy of a file which had duplicates.

  21. A protection mechanism to ensure files in system directories (/bin, /etc, ...) are not deleted. This mechanism is primarily for those who insist on running applications as root.

  22. Option to save the current list of found duplicates so processing can continue at a later date.

  23. Option to export the found list of duplicates to a text file (.txt), or a comma delimited file (.csv) for import into a spreadsheet program.

  24. Making use of a real database for the storage of file information (DB2?).

  25. If information about files on a removable media is to be stored in a database, the inclusion of a name of the media, for example the name of the CD the files are stored on, in the database.

  26. Recognition of images that are different in their binary make up, but similar to the eye. I have no experience in this area, so any help is welcome.

  27. Recognition of audio files, particularly mp3 and ogg vorbis, that are different in their binary make up but similar to the ear. As with images, I have no experience in this area so any help is welcome.

  28. Possible recognition of files that would be the same but where one file is truncated.

  29. Generation of a final report of what was found and what was done. This would include a count of the total number of files and directories examined, the total number of duplicates found, the total number and names of files deleted, moved, or linked, files that could not be deleted or otherwise processed, the total number and names of files and directories with problem names. All of this will be sumarised in a list comprising of each summary, and clicking on that list item will expand it to show all the information inside. For example, clicking on the list item with the count of problem file names will expand that list to show the problem file names.

  30. Estimation of how long a search will take. This will be based on what types of searches will be done, how much there is to search in terms of Megabytes, and the CPU bogomips count. If it is estimated a search will take a long than a user configurable warning time limit (default 30 mins), a window advising of this will be displayed.

  31. An interface redesign. Although functional at the moment, it's not the best. Suggestions are welcome.

  32. An option to save the current settings as either a search profile or as a default for when Duplicate File Finder is started. All saved values would be stored in ~/.dff.

  33. Ability to set a scheduled job so that Duplicate File Finder starts at a set time.

  34. Future versions will be written using C++. Java was originally chosen because of the ease of writing in it.

  35. The ability to run Duplicate File Finder as a daemon to check files that are downloaded from the net and show a popup indicating that you have a file that matches your criteria and where that file is. But wait a minute - this is going to extremes!!! This feature, if it is ever implemented, will definately be the last that gets implemented.

Other suggestions and requests are welcome.

Previous: Known Deficiencies    Contents    Next: Java 1.4.x Deficiencies