new duplicate processing thoughts
This commit is contained in:
13
TODO
13
TODO
@@ -19,10 +19,15 @@
|
|||||||
scan storage/import dir:
|
scan storage/import dir:
|
||||||
ignore *thumb*
|
ignore *thumb*
|
||||||
|
|
||||||
scan storage_dir
|
deal with duplicates differently, and process in this order:
|
||||||
* need to find / remove duplicate files from inside storage_dir and import_dir
|
1) any duplicate from inside import dir -> storage dir => delete all import dir duplicate files
|
||||||
-- in fact not sure what will happen if I try this right now, I think it might sort of work, only the dup display per file won't be able to
|
2) any duplicate from inside storage dir -> storage dir =>
|
||||||
use jex.path for all sets of files, only those dups in the original source of the scan
|
if regex match yyyy/yyyymmdd in a set of files and no other regex match -> keep regex files & delete others
|
||||||
|
if regex match yyyy/yyyymmdd in a set of files and other regex matches -> present to user to choose
|
||||||
|
if no regex match yyyy/yyyymmdd in a set of files -> present to user to choose
|
||||||
|
if regex match yyyy/yyyymmdd in a dir of files and no other regex match -> keep regex dir files & delete others
|
||||||
|
all other dir sets -> present to user to choose
|
||||||
|
3) any duplicate from inside import dir -> import dir == Do as we do now
|
||||||
|
|
||||||
-- started on some basic optimisations (commit logs every 100 logs, not each log)
|
-- started on some basic optimisations (commit logs every 100 logs, not each log)
|
||||||
- with debugs: import = 04:11, getfiledetails== 0:35:35
|
- with debugs: import = 04:11, getfiledetails== 0:35:35
|
||||||
|
|||||||
Reference in New Issue
Block a user