mirror of
https://github.com/arsenetar/dupeguru.git
synced 2026-01-25 16:11:39 +00:00
Compare commits
5 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
552e6b7836 | ||
|
|
28f70b281b | ||
|
|
32d9b573c0 | ||
|
|
fc76a843d5 | ||
|
|
06607aabb2 |
@@ -15,7 +15,8 @@ cocoa/*/build
|
|||||||
cocoa/*/dg_cocoa.plugin
|
cocoa/*/dg_cocoa.plugin
|
||||||
qt/base/*_rc.py
|
qt/base/*_rc.py
|
||||||
qt/base/*_ui.py
|
qt/base/*_ui.py
|
||||||
qt/se/*_ui.py
|
qt/*/*_ui.py
|
||||||
|
qt/pe/modules/block/block.c
|
||||||
help_se/dupeguru_help
|
help_se/dupeguru_help
|
||||||
help_me/dupeguru_me_help
|
help_me/dupeguru_me_help
|
||||||
help_pe/dupeguru_pe_help
|
help_pe/dupeguru_pe_help
|
||||||
1
.hgtags
1
.hgtags
@@ -5,3 +5,4 @@ a8f232f880b6f9ada565d472996a627ebf69b6e9 before-tiger-drop
|
|||||||
321d15e818cf9a3f1fc037543090bb2fca2cccd7 me5.7.0
|
321d15e818cf9a3f1fc037543090bb2fca2cccd7 me5.7.0
|
||||||
adc73ccd14b1386cb04dee773c53a2d126800e31 se2.9.0
|
adc73ccd14b1386cb04dee773c53a2d126800e31 se2.9.0
|
||||||
cbcf9c80fee4c908ef2efbf1c143c9e47676c9b2 pe1.8.0
|
cbcf9c80fee4c908ef2efbf1c143c9e47676c9b2 pe1.8.0
|
||||||
|
61c4101851bdea3cb37dfb76f0d404c78c7c594c se2.9.1
|
||||||
|
|||||||
@@ -23,7 +23,7 @@
|
|||||||
<key>CFBundleSignature</key>
|
<key>CFBundleSignature</key>
|
||||||
<string>hsft</string>
|
<string>hsft</string>
|
||||||
<key>CFBundleVersion</key>
|
<key>CFBundleVersion</key>
|
||||||
<string>1.8.0</string>
|
<string>1.8.1</string>
|
||||||
<key>NSMainNibFile</key>
|
<key>NSMainNibFile</key>
|
||||||
<string>MainMenu</string>
|
<string>MainMenu</string>
|
||||||
<key>NSPrincipalClass</key>
|
<key>NSPrincipalClass</key>
|
||||||
|
|||||||
@@ -38,7 +38,7 @@ class PyDupeGuru(PyApp):
|
|||||||
self.app.scanner.ignore_list.Clear()
|
self.app.scanner.ignore_list.Clear()
|
||||||
|
|
||||||
def clearPictureCache(self):
|
def clearPictureCache(self):
|
||||||
self.app.scanner.cached_blocks.clear()
|
self.app.scanner.clear_picture_cache()
|
||||||
|
|
||||||
def doScan(self):
|
def doScan(self):
|
||||||
return self.app.start_scanning()
|
return self.app.start_scanning()
|
||||||
|
|||||||
@@ -11,8 +11,7 @@ import logging
|
|||||||
import plistlib
|
import plistlib
|
||||||
import re
|
import re
|
||||||
|
|
||||||
from Foundation import *
|
from AppKit import NSBundle, NSUserDefaults, NSURL
|
||||||
from AppKit import *
|
|
||||||
from appscript import app, k, CommandError
|
from appscript import app, k, CommandError
|
||||||
|
|
||||||
from hsutil import io
|
from hsutil import io
|
||||||
@@ -23,7 +22,7 @@ from hsutil.cocoa import as_fetch
|
|||||||
from core import fs
|
from core import fs
|
||||||
from core import app_cocoa, directories
|
from core import app_cocoa, directories
|
||||||
from . import data
|
from . import data
|
||||||
from .cache import string_to_colors, Cache
|
from .cache import string_to_colors
|
||||||
from .scanner import ScannerPE
|
from .scanner import ScannerPE
|
||||||
|
|
||||||
mainBundle = NSBundle.mainBundle()
|
mainBundle = NSBundle.mainBundle()
|
||||||
@@ -74,7 +73,7 @@ def get_iphoto_database_path():
|
|||||||
|
|
||||||
def get_iphoto_pictures(plistpath):
|
def get_iphoto_pictures(plistpath):
|
||||||
if not io.exists(plistpath):
|
if not io.exists(plistpath):
|
||||||
raise InvalidPath(self)
|
return []
|
||||||
s = io.open(plistpath).read()
|
s = io.open(plistpath).read()
|
||||||
# There was a case where a guy had 0x10 chars in his plist, causing expat errors on loading
|
# There was a case where a guy had 0x10 chars in his plist, causing expat errors on loading
|
||||||
s = s.replace('\x10', '')
|
s = s.replace('\x10', '')
|
||||||
@@ -124,9 +123,8 @@ class Directories(directories.Directories):
|
|||||||
|
|
||||||
def add_path(self, path):
|
def add_path(self, path):
|
||||||
if path == Path('iPhoto Library'):
|
if path == Path('iPhoto Library'):
|
||||||
if path in self:
|
if path not in self:
|
||||||
raise AlreadyThereError()
|
self._dirs.append(path)
|
||||||
self._dirs.append(path)
|
|
||||||
else:
|
else:
|
||||||
directories.Directories.add_path(self, path)
|
directories.Directories.add_path(self, path)
|
||||||
|
|
||||||
@@ -136,8 +134,7 @@ class DupeGuruPE(app_cocoa.DupeGuru):
|
|||||||
app_cocoa.DupeGuru.__init__(self, data, 'dupeGuru Picture Edition', appid=5)
|
app_cocoa.DupeGuru.__init__(self, data, 'dupeGuru Picture Edition', appid=5)
|
||||||
self.scanner = ScannerPE()
|
self.scanner = ScannerPE()
|
||||||
self.directories = Directories()
|
self.directories = Directories()
|
||||||
p = op.join(self.appdata, 'cached_pictures.db')
|
self.scanner.cache_path = op.join(self.appdata, 'cached_pictures.db')
|
||||||
self.scanner.cached_blocks = Cache(p)
|
|
||||||
|
|
||||||
def _do_delete(self, j):
|
def _do_delete(self, j):
|
||||||
def op(dupe):
|
def op(dupe):
|
||||||
|
|||||||
@@ -10,8 +10,6 @@ import os
|
|||||||
import logging
|
import logging
|
||||||
import sqlite3 as sqlite
|
import sqlite3 as sqlite
|
||||||
|
|
||||||
import hsutil.sqlite
|
|
||||||
|
|
||||||
from _cache import string_to_colors
|
from _cache import string_to_colors
|
||||||
|
|
||||||
def colors_to_string(colors):
|
def colors_to_string(colors):
|
||||||
@@ -35,31 +33,10 @@ def colors_to_string(colors):
|
|||||||
class Cache(object):
|
class Cache(object):
|
||||||
"""A class to cache picture blocks.
|
"""A class to cache picture blocks.
|
||||||
"""
|
"""
|
||||||
def __init__(self, db=':memory:', threaded=True):
|
def __init__(self, db=':memory:'):
|
||||||
def create_tables():
|
|
||||||
sql = "create table pictures(path TEXT, blocks TEXT)"
|
|
||||||
self.con.execute(sql);
|
|
||||||
sql = "create index idx_path on pictures (path)"
|
|
||||||
self.con.execute(sql)
|
|
||||||
|
|
||||||
self.dbname = db
|
self.dbname = db
|
||||||
if threaded:
|
self.con = None
|
||||||
self.con = hsutil.sqlite.ThreadedConn(db, True)
|
self._create_con()
|
||||||
else:
|
|
||||||
self.con = sqlite.connect(db, isolation_level=None)
|
|
||||||
try:
|
|
||||||
self.con.execute("select * from pictures where 1=2")
|
|
||||||
except sqlite.OperationalError: # new db
|
|
||||||
create_tables()
|
|
||||||
except sqlite.DatabaseError, e: # corrupted db
|
|
||||||
logging.warning('Could not create picture cache because of an error: %s', str(e))
|
|
||||||
self.con.close()
|
|
||||||
os.remove(db)
|
|
||||||
if threaded:
|
|
||||||
self.con = hsutil.sqlite.ThreadedConn(db, True)
|
|
||||||
else:
|
|
||||||
self.con = sqlite.connect(db, isolation_level=None)
|
|
||||||
create_tables()
|
|
||||||
|
|
||||||
def __contains__(self, key):
|
def __contains__(self, key):
|
||||||
sql = "select count(*) from pictures where path = ?"
|
sql = "select count(*) from pictures where path = ?"
|
||||||
@@ -108,9 +85,36 @@ class Cache(object):
|
|||||||
except sqlite.DatabaseError, e:
|
except sqlite.DatabaseError, e:
|
||||||
logging.warning('DatabaseError while setting %r for key %r: %s', value, key, str(e))
|
logging.warning('DatabaseError while setting %r for key %r: %s', value, key, str(e))
|
||||||
|
|
||||||
|
def _create_con(self, second_try=False):
|
||||||
|
def create_tables():
|
||||||
|
sql = "create table pictures(path TEXT, blocks TEXT)"
|
||||||
|
self.con.execute(sql);
|
||||||
|
sql = "create index idx_path on pictures (path)"
|
||||||
|
self.con.execute(sql)
|
||||||
|
|
||||||
|
self.con = sqlite.connect(self.dbname, isolation_level=None)
|
||||||
|
try:
|
||||||
|
self.con.execute("select * from pictures where 1=2")
|
||||||
|
except sqlite.OperationalError: # new db
|
||||||
|
create_tables()
|
||||||
|
except sqlite.DatabaseError, e: # corrupted db
|
||||||
|
if second_try:
|
||||||
|
raise # Something really strange is happening
|
||||||
|
logging.warning('Could not create picture cache because of an error: %s', str(e))
|
||||||
|
self.con.close()
|
||||||
|
os.remove(self.dbname)
|
||||||
|
self._create_con(second_try=True)
|
||||||
|
|
||||||
def clear(self):
|
def clear(self):
|
||||||
sql = "delete from pictures"
|
self.close()
|
||||||
self.con.execute(sql)
|
if self.dbname != ':memory:':
|
||||||
|
os.remove(self.dbname)
|
||||||
|
self._create_con()
|
||||||
|
|
||||||
|
def close(self):
|
||||||
|
if self.con is not None:
|
||||||
|
self.con.close()
|
||||||
|
self.con = None
|
||||||
|
|
||||||
def filter(self, func):
|
def filter(self, func):
|
||||||
to_delete = [key for key in self if not func(key)]
|
to_delete = [key for key in self if not func(key)]
|
||||||
|
|||||||
@@ -26,20 +26,21 @@ BLOCK_COUNT_PER_SIDE = 15
|
|||||||
# collection made by the main process.
|
# collection made by the main process.
|
||||||
RESULTS_QUEUE_LIMIT = multiprocessing.cpu_count() * 2
|
RESULTS_QUEUE_LIMIT = multiprocessing.cpu_count() * 2
|
||||||
|
|
||||||
def prepare_pictures(pictures, cached_blocks, j=job.nulljob):
|
def prepare_pictures(pictures, cache_path, j=job.nulljob):
|
||||||
# The MemoryError handlers in there use logging without first caring about whether or not
|
# The MemoryError handlers in there use logging without first caring about whether or not
|
||||||
# there is enough memory left to carry on the operation because it is assumed that the
|
# there is enough memory left to carry on the operation because it is assumed that the
|
||||||
# MemoryError happens when trying to read an image file, which is freed from memory by the
|
# MemoryError happens when trying to read an image file, which is freed from memory by the
|
||||||
# time that MemoryError is raised.
|
# time that MemoryError is raised.
|
||||||
|
cache = Cache(cache_path)
|
||||||
prepared = [] # only pictures for which there was no error getting blocks
|
prepared = [] # only pictures for which there was no error getting blocks
|
||||||
try:
|
try:
|
||||||
for picture in j.iter_with_progress(pictures, 'Analyzed %d/%d pictures'):
|
for picture in j.iter_with_progress(pictures, 'Analyzed %d/%d pictures'):
|
||||||
picture.dimensions
|
picture.dimensions
|
||||||
picture.unicode_path = unicode(picture.path)
|
picture.unicode_path = unicode(picture.path)
|
||||||
try:
|
try:
|
||||||
if picture.unicode_path not in cached_blocks:
|
if picture.unicode_path not in cache:
|
||||||
blocks = picture.get_blocks(BLOCK_COUNT_PER_SIDE)
|
blocks = picture.get_blocks(BLOCK_COUNT_PER_SIDE)
|
||||||
cached_blocks[picture.unicode_path] = blocks
|
cache[picture.unicode_path] = blocks
|
||||||
prepared.append(picture)
|
prepared.append(picture)
|
||||||
except IOError as e:
|
except IOError as e:
|
||||||
logging.warning(unicode(e))
|
logging.warning(unicode(e))
|
||||||
@@ -49,6 +50,7 @@ def prepare_pictures(pictures, cached_blocks, j=job.nulljob):
|
|||||||
raise
|
raise
|
||||||
except MemoryError:
|
except MemoryError:
|
||||||
logging.warning('Ran out of memory while preparing pictures')
|
logging.warning('Ran out of memory while preparing pictures')
|
||||||
|
cache.close()
|
||||||
return prepared
|
return prepared
|
||||||
|
|
||||||
def get_match(first, second, percentage):
|
def get_match(first, second, percentage):
|
||||||
@@ -57,7 +59,7 @@ def get_match(first, second, percentage):
|
|||||||
return Match(first, second, percentage)
|
return Match(first, second, percentage)
|
||||||
|
|
||||||
def async_compare(ref_id, other_ids, dbname, threshold):
|
def async_compare(ref_id, other_ids, dbname, threshold):
|
||||||
cache = Cache(dbname, threaded=False)
|
cache = Cache(dbname)
|
||||||
limit = 100 - threshold
|
limit = 100 - threshold
|
||||||
ref_blocks = cache[ref_id]
|
ref_blocks = cache[ref_id]
|
||||||
pairs = cache.get_multiple(other_ids)
|
pairs = cache.get_multiple(other_ids)
|
||||||
@@ -70,10 +72,10 @@ def async_compare(ref_id, other_ids, dbname, threshold):
|
|||||||
percentage = 0
|
percentage = 0
|
||||||
if percentage >= threshold:
|
if percentage >= threshold:
|
||||||
results.append((ref_id, other_id, percentage))
|
results.append((ref_id, other_id, percentage))
|
||||||
cache.con.close()
|
cache.close()
|
||||||
return results
|
return results
|
||||||
|
|
||||||
def getmatches(pictures, cached_blocks, threshold=75, match_scaled=False, j=job.nulljob):
|
def getmatches(pictures, cache_path, threshold=75, match_scaled=False, j=job.nulljob):
|
||||||
def empty_out_queue(queue, into):
|
def empty_out_queue(queue, into):
|
||||||
try:
|
try:
|
||||||
while True:
|
while True:
|
||||||
@@ -82,9 +84,9 @@ def getmatches(pictures, cached_blocks, threshold=75, match_scaled=False, j=job.
|
|||||||
pass
|
pass
|
||||||
|
|
||||||
j = j.start_subjob([3, 7])
|
j = j.start_subjob([3, 7])
|
||||||
pictures = prepare_pictures(pictures, cached_blocks, j)
|
pictures = prepare_pictures(pictures, cache_path, j)
|
||||||
j = j.start_subjob([9, 1], 'Preparing for matching')
|
j = j.start_subjob([9, 1], 'Preparing for matching')
|
||||||
cache = cached_blocks
|
cache = Cache(cache_path)
|
||||||
id2picture = {}
|
id2picture = {}
|
||||||
dimensions2pictures = defaultdict(set)
|
dimensions2pictures = defaultdict(set)
|
||||||
for picture in pictures:
|
for picture in pictures:
|
||||||
@@ -95,6 +97,7 @@ def getmatches(pictures, cached_blocks, threshold=75, match_scaled=False, j=job.
|
|||||||
dimensions2pictures[picture.dimensions].add(picture)
|
dimensions2pictures[picture.dimensions].add(picture)
|
||||||
except ValueError:
|
except ValueError:
|
||||||
pass
|
pass
|
||||||
|
cache.close()
|
||||||
pictures = [p for p in pictures if hasattr(p, 'cache_id')]
|
pictures = [p for p in pictures if hasattr(p, 'cache_id')]
|
||||||
pool = multiprocessing.Pool()
|
pool = multiprocessing.Pool()
|
||||||
async_results = []
|
async_results = []
|
||||||
@@ -108,7 +111,7 @@ def getmatches(pictures, cached_blocks, threshold=75, match_scaled=False, j=job.
|
|||||||
others = [pic for pic in others if not pic.is_ref]
|
others = [pic for pic in others if not pic.is_ref]
|
||||||
if others:
|
if others:
|
||||||
cache_ids = [f.cache_id for f in others]
|
cache_ids = [f.cache_id for f in others]
|
||||||
args = (ref.cache_id, cache_ids, cached_blocks.dbname, threshold)
|
args = (ref.cache_id, cache_ids, cache_path, threshold)
|
||||||
async_results.append(pool.apply_async(async_compare, args))
|
async_results.append(pool.apply_async(async_compare, args))
|
||||||
if len(async_results) > RESULTS_QUEUE_LIMIT:
|
if len(async_results) > RESULTS_QUEUE_LIMIT:
|
||||||
result = async_results.pop(0)
|
result = async_results.pop(0)
|
||||||
|
|||||||
@@ -10,12 +10,18 @@
|
|||||||
from core.scanner import Scanner
|
from core.scanner import Scanner
|
||||||
|
|
||||||
from . import matchbase
|
from . import matchbase
|
||||||
|
from .cache import Cache
|
||||||
|
|
||||||
class ScannerPE(Scanner):
|
class ScannerPE(Scanner):
|
||||||
cached_blocks = None
|
cache_path = None
|
||||||
match_scaled = False
|
match_scaled = False
|
||||||
threshold = 75
|
threshold = 75
|
||||||
|
|
||||||
def _getmatches(self, files, j):
|
def _getmatches(self, files, j):
|
||||||
return matchbase.getmatches(files, self.cached_blocks, self.threshold, self.match_scaled, j)
|
return matchbase.getmatches(files, self.cache_path, self.threshold, self.match_scaled, j)
|
||||||
|
|
||||||
|
def clear_picture_cache(self):
|
||||||
|
cache = Cache(self.cache_path)
|
||||||
|
cache.clear()
|
||||||
|
cache.close()
|
||||||
|
|
||||||
|
|||||||
@@ -137,18 +137,3 @@ class TCCacheSQLEscape(TestCase):
|
|||||||
except KeyError:
|
except KeyError:
|
||||||
self.fail()
|
self.fail()
|
||||||
|
|
||||||
|
|
||||||
class TCCacheThreaded(TestCase):
|
|
||||||
def test_access_cache(self):
|
|
||||||
def thread_run():
|
|
||||||
try:
|
|
||||||
c['foo'] = [(1,2,3)]
|
|
||||||
except sqlite.ProgrammingError:
|
|
||||||
self.fail()
|
|
||||||
|
|
||||||
c = Cache()
|
|
||||||
t = threading.Thread(target=thread_run)
|
|
||||||
t.start()
|
|
||||||
t.join()
|
|
||||||
self.assertEqual([(1,2,3)], c['foo'])
|
|
||||||
|
|
||||||
|
|||||||
@@ -1,3 +1,9 @@
|
|||||||
|
- date: 2010-01-15
|
||||||
|
version: 1.8.1
|
||||||
|
description: |
|
||||||
|
* Improved scanning speed when ref directories are involved. (#77)
|
||||||
|
* Show a message dialog at the end of the scan if no duplicates are found. (#81)
|
||||||
|
* Fixed a crash when adding the iPhoto library twice. [Mac OS X] (#80)
|
||||||
- date: 2009-12-16
|
- date: 2009-12-16
|
||||||
version: 1.8.0
|
version: 1.8.0
|
||||||
description: |
|
description: |
|
||||||
|
|||||||
@@ -56,7 +56,7 @@ class File(fs.File):
|
|||||||
class DupeGuru(DupeGuruBase):
|
class DupeGuru(DupeGuruBase):
|
||||||
LOGO_NAME = 'logo_pe'
|
LOGO_NAME = 'logo_pe'
|
||||||
NAME = 'dupeGuru Picture Edition'
|
NAME = 'dupeGuru Picture Edition'
|
||||||
VERSION = '1.8.0'
|
VERSION = '1.8.1'
|
||||||
DELTA_COLUMNS = frozenset([2, 5, 6])
|
DELTA_COLUMNS = frozenset([2, 5, 6])
|
||||||
|
|
||||||
def __init__(self):
|
def __init__(self):
|
||||||
@@ -65,7 +65,7 @@ class DupeGuru(DupeGuruBase):
|
|||||||
def _setup(self):
|
def _setup(self):
|
||||||
self.scanner = ScannerPE()
|
self.scanner = ScannerPE()
|
||||||
self.directories.fileclasses = [File]
|
self.directories.fileclasses = [File]
|
||||||
self.scanner.cached_blocks = Cache(op.join(self.appdata, 'cached_pictures.db'))
|
self.scanner.cache_path = op.join(self.appdata, 'cached_pictures.db')
|
||||||
DupeGuruBase._setup(self)
|
DupeGuruBase._setup(self)
|
||||||
|
|
||||||
def _update_options(self):
|
def _update_options(self):
|
||||||
|
|||||||
@@ -22,6 +22,6 @@ class MainWindow(MainWindowBase):
|
|||||||
title = "Clear Picture Cache"
|
title = "Clear Picture Cache"
|
||||||
msg = "Do you really want to remove all your cached picture analysis?"
|
msg = "Do you really want to remove all your cached picture analysis?"
|
||||||
if self._confirm(title, msg, QMessageBox.No):
|
if self._confirm(title, msg, QMessageBox.No):
|
||||||
self.app.scanner.cached_blocks.clear()
|
self.app.scanner.clear_picture_cache()
|
||||||
QMessageBox.information(self, title, "Picture cache cleared.")
|
QMessageBox.information(self, title, "Picture cache cleared.")
|
||||||
|
|
||||||
Reference in New Issue
Block a user