Compare commits
49 Commits
da54b3a454
...
main
Author | SHA1 | Date | |
---|---|---|---|
22d59aa221 | |||
ed0b5ba441 | |||
d240bd953b | |||
bb64d8449c | |||
b0a4c2939c | |||
3e364b1858 | |||
0da76f7fb5 | |||
d7c42f5c1d | |||
374c46bc48 | |||
aca658e32d | |||
acd71df302 | |||
0fd431f6f7 | |||
9322f754bf | |||
486cf074d0 | |||
bf23bb5a4c | |||
d7946c2aaf | |||
f54b4b1d9d | |||
a2765b0582 | |||
1f37684862 | |||
5307b5d8df | |||
130b7dec4a | |||
6452f2e50d | |||
f52219cb31 | |||
5a93dacdad | |||
798ae49c8b | |||
8d6db630d4 | |||
29f58e7fe7 | |||
fe3f045c6e | |||
8ce1ca3354 | |||
75d60339ee | |||
8a9aec4884 | |||
033c68593b | |||
260a13688a | |||
e04469938d | |||
8bdfb282b9 | |||
9baa4dbc17 | |||
5cf25089da | |||
cf2c61b92e | |||
cc4e973bf9 | |||
7ede0dee72 | |||
840e97cff0 | |||
50fcfb9513 | |||
776ca68e3e | |||
95fa44fb97 | |||
d0c7ed792a | |||
97802d9111 | |||
89c9b4ebce | |||
a65cadf106 | |||
4922ceb678 |
@ -1,6 +1,6 @@
|
|||||||
repos:
|
repos:
|
||||||
- repo: https://github.com/pre-commit/pre-commit-hooks
|
- repo: https://github.com/pre-commit/pre-commit-hooks
|
||||||
rev: v4.6.0
|
rev: v5.0.0
|
||||||
hooks:
|
hooks:
|
||||||
- id: trailing-whitespace
|
- id: trailing-whitespace
|
||||||
args: [--markdown-linebreak-ext=.gitignore]
|
args: [--markdown-linebreak-ext=.gitignore]
|
||||||
@ -13,16 +13,16 @@ repos:
|
|||||||
- id: go-imports
|
- id: go-imports
|
||||||
args: [-w]
|
args: [-w]
|
||||||
- repo: https://github.com/golangci/golangci-lint
|
- repo: https://github.com/golangci/golangci-lint
|
||||||
rev: v1.60.3
|
rev: v1.64.5
|
||||||
hooks:
|
hooks:
|
||||||
- id: golangci-lint
|
- id: golangci-lint
|
||||||
- repo: https://github.com/asottile/setup-cfg-fmt
|
- repo: https://github.com/asottile/setup-cfg-fmt
|
||||||
rev: v2.5.0
|
rev: v2.7.0
|
||||||
hooks:
|
hooks:
|
||||||
- id: setup-cfg-fmt
|
- id: setup-cfg-fmt
|
||||||
|
|
||||||
- repo: https://github.com/asottile/reorder-python-imports
|
- repo: https://github.com/asottile/reorder-python-imports
|
||||||
rev: v3.13.0
|
rev: v3.14.0
|
||||||
hooks:
|
hooks:
|
||||||
- id: reorder-python-imports
|
- id: reorder-python-imports
|
||||||
args: [--py38-plus, --add-import, 'from __future__ import annotations']
|
args: [--py38-plus, --add-import, 'from __future__ import annotations']
|
||||||
@ -31,12 +31,12 @@ repos:
|
|||||||
hooks:
|
hooks:
|
||||||
- id: add-trailing-comma
|
- id: add-trailing-comma
|
||||||
- repo: https://github.com/asottile/pyupgrade
|
- repo: https://github.com/asottile/pyupgrade
|
||||||
rev: v3.17.0
|
rev: v3.19.1
|
||||||
hooks:
|
hooks:
|
||||||
- id: pyupgrade
|
- id: pyupgrade
|
||||||
args: [--py38-plus]
|
args: [--py38-plus]
|
||||||
exclude: tests
|
exclude: tests
|
||||||
- repo: https://github.com/hhatto/autopep8
|
- repo: https://github.com/hhatto/autopep8
|
||||||
rev: v2.3.1
|
rev: v2.3.2
|
||||||
hooks:
|
hooks:
|
||||||
- id: autopep8
|
- id: autopep8
|
||||||
|
189
BasicMap.go
189
BasicMap.go
@ -1,189 +0,0 @@
|
|||||||
package ch
|
|
||||||
|
|
||||||
import (
|
|
||||||
"cmp"
|
|
||||||
"errors"
|
|
||||||
"fmt"
|
|
||||||
"math/bits"
|
|
||||||
"slices"
|
|
||||||
"sync"
|
|
||||||
|
|
||||||
"gitea.narnian.us/lordwelch/goimagehash"
|
|
||||||
)
|
|
||||||
|
|
||||||
type basicMapStorage struct {
|
|
||||||
hashMutex sync.RWMutex
|
|
||||||
|
|
||||||
ids map[ID]*[]ID
|
|
||||||
hashes [3][]structHash
|
|
||||||
}
|
|
||||||
|
|
||||||
type structHash struct {
|
|
||||||
hash uint64
|
|
||||||
ids *[]ID
|
|
||||||
}
|
|
||||||
|
|
||||||
func (b *basicMapStorage) Atleast(hashKind goimagehash.Kind, maxDistance int, searchHash uint64) []Result {
|
|
||||||
hashType := int(hashKind) - 1
|
|
||||||
matchingHashes := make([]Result, 0, 100) // hope that we don't need all of them
|
|
||||||
for _, storedHash := range b.hashes[hashType] {
|
|
||||||
distance := bits.OnesCount64(searchHash ^ storedHash.hash)
|
|
||||||
if distance <= maxDistance {
|
|
||||||
matchingHashes = append(matchingHashes, Result{ToIDList(*storedHash.ids), distance, Hash{storedHash.hash, hashKind}})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return matchingHashes
|
|
||||||
}
|
|
||||||
func (b *basicMapStorage) GetMatches(hashes []Hash, max int, exactOnly bool) ([]Result, error) {
|
|
||||||
var foundMatches []Result
|
|
||||||
b.hashMutex.RLock()
|
|
||||||
defer b.hashMutex.RUnlock()
|
|
||||||
resetTime()
|
|
||||||
defer logTime(fmt.Sprintf("Search Complete: max: %v ExactOnly: %v", max, exactOnly))
|
|
||||||
|
|
||||||
if exactOnly { // exact matches are also found by partial matches. Don't bother with exact matches so we don't have to de-duplicate
|
|
||||||
for _, hash := range hashes {
|
|
||||||
hashType := int(hash.Kind) - 1
|
|
||||||
index, hashFound := b.findHash(hashType, hash.Hash)
|
|
||||||
if hashFound {
|
|
||||||
foundMatches = append(foundMatches, Result{
|
|
||||||
Distance: 0,
|
|
||||||
Hash: hash,
|
|
||||||
IDs: ToIDList(*b.hashes[hashType][index].ids),
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
logTime("Search Exact")
|
|
||||||
// If we have exact matches don't bother with other matches
|
|
||||||
if len(foundMatches) > 0 && exactOnly {
|
|
||||||
return foundMatches, nil
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
foundHashes := make(map[uint64]struct{})
|
|
||||||
totalPartialHashes := 0
|
|
||||||
for _, hash := range hashes {
|
|
||||||
for _, match := range b.Atleast(hash.Kind, max, hash.Hash) {
|
|
||||||
_, alreadyMatched := foundHashes[match.Hash.Hash]
|
|
||||||
if alreadyMatched {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
foundHashes[match.Hash.Hash] = struct{}{}
|
|
||||||
foundMatches = append(foundMatches, match)
|
|
||||||
}
|
|
||||||
|
|
||||||
}
|
|
||||||
fmt.Println("Total partial hashes tested:", totalPartialHashes, len(foundHashes))
|
|
||||||
go b.printSizes()
|
|
||||||
return foundMatches, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (b *basicMapStorage) findHash(hashType int, hash uint64) (int, bool) {
|
|
||||||
return slices.BinarySearchFunc(b.hashes[hashType], hash, func(e structHash, t uint64) int {
|
|
||||||
return cmp.Compare(e.hash, t)
|
|
||||||
})
|
|
||||||
}
|
|
||||||
func (b *basicMapStorage) InsertHash(hashType int, hash uint64, ids *[]ID) {
|
|
||||||
index, hashFound := b.findHash(hashType, hash)
|
|
||||||
if hashFound {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
b.hashes[hashType] = slices.Insert(b.hashes[hashType], index, structHash{hash, ids})
|
|
||||||
}
|
|
||||||
|
|
||||||
func (b *basicMapStorage) MapHashes(hash ImageHash) {
|
|
||||||
for _, ih := range hash.Hashes {
|
|
||||||
var (
|
|
||||||
hashType = int(ih.Kind) - 1
|
|
||||||
)
|
|
||||||
ids, ok := b.ids[hash.ID]
|
|
||||||
if !ok {
|
|
||||||
ids = &[]ID{hash.ID}
|
|
||||||
b.ids[hash.ID] = ids
|
|
||||||
}
|
|
||||||
|
|
||||||
b.InsertHash(hashType, ih.Hash, ids)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func (b *basicMapStorage) DecodeHashes(hashes SavedHashes) error {
|
|
||||||
for hashType, sourceHashes := range hashes.Hashes {
|
|
||||||
b.hashes[hashType] = make([]structHash, len(sourceHashes))
|
|
||||||
for savedHash, idlistLocation := range sourceHashes {
|
|
||||||
b.hashes[hashType] = append(b.hashes[hashType], structHash{savedHash, &hashes.IDs[idlistLocation]})
|
|
||||||
for _, id := range hashes.IDs[idlistLocation] {
|
|
||||||
b.ids[id] = &hashes.IDs[idlistLocation]
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
for hashType := range b.hashes {
|
|
||||||
slices.SortFunc(b.hashes[hashType], func(a, b structHash) int {
|
|
||||||
return cmp.Compare(a.hash, b.hash)
|
|
||||||
})
|
|
||||||
}
|
|
||||||
b.printSizes()
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (b *basicMapStorage) printSizes() {
|
|
||||||
// fmt.Println("Size of", "hashes:", size.Of(b.hashes)/1024/1024, "MB")
|
|
||||||
// fmt.Println("Size of", "ids:", size.Of(b.ids)/1024/1024, "MB")
|
|
||||||
// fmt.Println("Size of", "basicMapStorage:", size.Of(b)/1024/1024, "MB")
|
|
||||||
|
|
||||||
}
|
|
||||||
|
|
||||||
func (b *basicMapStorage) EncodeHashes() (SavedHashes, error) {
|
|
||||||
hashes := SavedHashes{
|
|
||||||
Hashes: [3]map[uint64]int{
|
|
||||||
make(map[uint64]int),
|
|
||||||
make(map[uint64]int),
|
|
||||||
make(map[uint64]int),
|
|
||||||
},
|
|
||||||
}
|
|
||||||
idmap := map[*[]ID]int{}
|
|
||||||
|
|
||||||
for _, ids := range b.ids {
|
|
||||||
if _, ok := idmap[ids]; ok {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
idmap[ids] = len(hashes.IDs)
|
|
||||||
hashes.IDs = append(hashes.IDs, *ids)
|
|
||||||
}
|
|
||||||
|
|
||||||
for hashType, hashToID := range b.hashes {
|
|
||||||
for _, hash := range hashToID {
|
|
||||||
hashes.Hashes[hashType][hash.hash] = idmap[hash.ids]
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return hashes, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (b *basicMapStorage) AssociateIDs(newids []NewIDs) error {
|
|
||||||
for _, newid := range newids {
|
|
||||||
ids, found := b.ids[newid.OldID]
|
|
||||||
if !found {
|
|
||||||
msg := "No IDs belonging to " + string(newid.OldID.Domain) + " exist on this server"
|
|
||||||
return errors.New(msg)
|
|
||||||
}
|
|
||||||
*ids = InsertID(*ids, newid.NewID)
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (b *basicMapStorage) GetIDs(id ID) IDList {
|
|
||||||
ids, found := b.ids[id]
|
|
||||||
if !found {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
return ToIDList(*ids)
|
|
||||||
}
|
|
||||||
|
|
||||||
func NewBasicMapStorage() (HashStorage, error) {
|
|
||||||
storage := &basicMapStorage{
|
|
||||||
hashMutex: sync.RWMutex{},
|
|
||||||
ids: make(map[ID]*[]ID),
|
|
||||||
hashes: [3][]structHash{},
|
|
||||||
}
|
|
||||||
return storage, nil
|
|
||||||
}
|
|
11
CHDB.go
Normal file
11
CHDB.go
Normal file
@ -0,0 +1,11 @@
|
|||||||
|
package ch
|
||||||
|
|
||||||
|
type CHDB interface {
|
||||||
|
// OpenCHDB(path string, comicvinePath string, deleteExisting bool) (CHDB, error)
|
||||||
|
PathHashed(path string) bool
|
||||||
|
PathDownloaded(path string) bool
|
||||||
|
AddPath(path string)
|
||||||
|
CheckURL(url string) bool
|
||||||
|
AddURL(url string)
|
||||||
|
Close() error
|
||||||
|
}
|
177
CHDB_bolt.go
Normal file
177
CHDB_bolt.go
Normal file
@ -0,0 +1,177 @@
|
|||||||
|
package ch
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"log"
|
||||||
|
"os"
|
||||||
|
"path/filepath"
|
||||||
|
"slices"
|
||||||
|
|
||||||
|
bolt "go.etcd.io/bbolt"
|
||||||
|
)
|
||||||
|
|
||||||
|
type CHDBBolt struct {
|
||||||
|
comicvinePath string
|
||||||
|
db *bolt.DB
|
||||||
|
deleteExisting bool
|
||||||
|
}
|
||||||
|
|
||||||
|
func OpenCHDBBolt(path string, comicvinePath string, deleteExisting bool) (CHDBBolt, error) {
|
||||||
|
path, _ = filepath.Abs(path)
|
||||||
|
err := os.MkdirAll(filepath.Dir(path), 0o755)
|
||||||
|
if err != nil {
|
||||||
|
panic("Unable to create directory " + filepath.Dir(path))
|
||||||
|
}
|
||||||
|
db, err := bolt.Open(path, 0o644, nil)
|
||||||
|
if err != nil {
|
||||||
|
return CHDBBolt{comicvinePath, db, deleteExisting}, fmt.Errorf("failed to open database: %w", err)
|
||||||
|
}
|
||||||
|
err = db.Update(func(tx *bolt.Tx) error {
|
||||||
|
|
||||||
|
_, err = tx.CreateBucketIfNotExists([]byte("paths"))
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("failed to create bucket %v: %w", "paths", err)
|
||||||
|
}
|
||||||
|
_, err = tx.CreateBucketIfNotExists([]byte("bad_urls"))
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("failed to create bucket %v: %w", "paths", err)
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
})
|
||||||
|
if err != nil {
|
||||||
|
db.Close()
|
||||||
|
return CHDBBolt{comicvinePath, db, deleteExisting}, fmt.Errorf("failed to init database: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
return CHDBBolt{comicvinePath, db, deleteExisting}, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c CHDBBolt) Import(paths []string, bad_urls []string) {
|
||||||
|
slices.Sort(paths)
|
||||||
|
slices.Sort(bad_urls)
|
||||||
|
c.db.Update(func(tx *bolt.Tx) error {
|
||||||
|
p := tx.Bucket([]byte("paths"))
|
||||||
|
b := tx.Bucket([]byte("bad_urls"))
|
||||||
|
|
||||||
|
for _, path := range paths {
|
||||||
|
p.Put([]byte(path), []byte{})
|
||||||
|
}
|
||||||
|
for _, url := range bad_urls {
|
||||||
|
b.Put([]byte(url), []byte{})
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c CHDBBolt) Dump() (paths []string, bad_urls []string) {
|
||||||
|
|
||||||
|
c.db.View(func(tx *bolt.Tx) error {
|
||||||
|
p := tx.Bucket([]byte("paths"))
|
||||||
|
b := tx.Bucket([]byte("bad_urls"))
|
||||||
|
paths = make([]string, 0, p.Inspect().KeyN)
|
||||||
|
bad_urls = make([]string, 0, b.Inspect().KeyN)
|
||||||
|
b.ForEach(func(k, v []byte) error {
|
||||||
|
bad_urls = append(bad_urls, string(k)+"")
|
||||||
|
return nil
|
||||||
|
})
|
||||||
|
p.ForEach(func(k, v []byte) error {
|
||||||
|
paths = append(paths, string(k)+"")
|
||||||
|
return nil
|
||||||
|
})
|
||||||
|
return nil
|
||||||
|
})
|
||||||
|
return paths, bad_urls
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c CHDBBolt) PathHashed(path string) bool {
|
||||||
|
path, _ = filepath.Rel(c.comicvinePath, path)
|
||||||
|
|
||||||
|
tx, err := c.db.Begin(false)
|
||||||
|
if err != nil {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
defer tx.Rollback()
|
||||||
|
b := tx.Bucket([]byte("paths"))
|
||||||
|
dbRes := b.Get([]byte(path))
|
||||||
|
if dbRes != nil {
|
||||||
|
if c.deleteExisting {
|
||||||
|
os.Remove(filepath.Join(c.comicvinePath, path))
|
||||||
|
}
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c CHDBBolt) PathDownloaded(path string) bool {
|
||||||
|
relPath, _ := filepath.Rel(c.comicvinePath, path)
|
||||||
|
|
||||||
|
tx, err := c.db.Begin(false)
|
||||||
|
if err != nil {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
defer tx.Rollback()
|
||||||
|
b := tx.Bucket([]byte("paths"))
|
||||||
|
dbRes := b.Get([]byte(relPath))
|
||||||
|
if dbRes == nil {
|
||||||
|
|
||||||
|
f, err := os.Open(path)
|
||||||
|
if err == nil {
|
||||||
|
defer f.Close()
|
||||||
|
}
|
||||||
|
return !os.IsNotExist(err)
|
||||||
|
}
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c CHDBBolt) AddPath(path string) {
|
||||||
|
relPath, _ := filepath.Rel(c.comicvinePath, path)
|
||||||
|
|
||||||
|
tx, err := c.db.Begin(true)
|
||||||
|
if err != nil {
|
||||||
|
c.db.Logger().Errorf("Failed to open transaction: %v", err)
|
||||||
|
}
|
||||||
|
defer tx.Rollback()
|
||||||
|
b := tx.Bucket([]byte("paths"))
|
||||||
|
|
||||||
|
err = b.Put([]byte(relPath), []byte{})
|
||||||
|
if err != nil {
|
||||||
|
log.Println(fmt.Errorf("Failed to insert %v (%v) into paths: %w", path, relPath, err))
|
||||||
|
}
|
||||||
|
tx.Commit()
|
||||||
|
if c.deleteExisting {
|
||||||
|
_ = os.Remove(path)
|
||||||
|
_ = RmdirP(filepath.Dir(path))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c CHDBBolt) CheckURL(url string) bool {
|
||||||
|
|
||||||
|
tx, err := c.db.Begin(true)
|
||||||
|
if err != nil {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
defer tx.Rollback()
|
||||||
|
b := tx.Bucket([]byte("bad_urls"))
|
||||||
|
return b.Get([]byte(url)) != nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c CHDBBolt) AddURL(url string) {
|
||||||
|
|
||||||
|
tx, err := c.db.Begin(true)
|
||||||
|
if err != nil {
|
||||||
|
c.db.Logger().Errorf("Failed to open transaction: %v", err)
|
||||||
|
}
|
||||||
|
defer tx.Rollback()
|
||||||
|
b := tx.Bucket([]byte("bad_urls"))
|
||||||
|
|
||||||
|
err = b.Put([]byte(url), []byte{})
|
||||||
|
if err != nil {
|
||||||
|
log.Println(fmt.Errorf("Failed to insert %v into bad_urls: %w", url, err))
|
||||||
|
}
|
||||||
|
tx.Commit()
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c CHDBBolt) Close() error {
|
||||||
|
return c.db.Close()
|
||||||
|
}
|
142
CHDB_sqlite.go
Normal file
142
CHDB_sqlite.go
Normal file
@ -0,0 +1,142 @@
|
|||||||
|
package ch
|
||||||
|
|
||||||
|
import (
|
||||||
|
"database/sql"
|
||||||
|
"fmt"
|
||||||
|
"log"
|
||||||
|
"os"
|
||||||
|
"path/filepath"
|
||||||
|
|
||||||
|
_ "modernc.org/sqlite"
|
||||||
|
)
|
||||||
|
|
||||||
|
type CHDBSqlite struct {
|
||||||
|
comicvinePath string
|
||||||
|
sql *sql.DB
|
||||||
|
deleteExisting bool
|
||||||
|
}
|
||||||
|
|
||||||
|
func OpenCHDBSqlite(path string, comicvinePath string, deleteExisting bool) (CHDBSqlite, error) {
|
||||||
|
path, _ = filepath.Abs(path)
|
||||||
|
err := os.MkdirAll(filepath.Dir(path), 0o755)
|
||||||
|
if err != nil {
|
||||||
|
panic("Unable to create directory " + filepath.Dir(path))
|
||||||
|
}
|
||||||
|
println(fmt.Sprintf("file://%s?&_pragma=busy_timeout(500)&_pragma=journal_mode(wal)", path))
|
||||||
|
sql, err := sql.Open("sqlite", fmt.Sprintf("file://%s?&_pragma=busy_timeout(500)&_pragma=journal_mode(wal)", path))
|
||||||
|
if err != nil {
|
||||||
|
return CHDBSqlite{comicvinePath, sql, deleteExisting}, fmt.Errorf("Failed to open database: %w", err)
|
||||||
|
}
|
||||||
|
err = sql.Ping()
|
||||||
|
if err != nil {
|
||||||
|
return CHDBSqlite{comicvinePath, sql, deleteExisting}, fmt.Errorf("Failed to open database: %w", err)
|
||||||
|
}
|
||||||
|
_, err = sql.Exec(`
|
||||||
|
CREATE TABLE IF NOT EXISTS paths(
|
||||||
|
path STRING PRIMARY KEY
|
||||||
|
);
|
||||||
|
CREATE TABLE IF NOT EXISTS bad_urls(
|
||||||
|
url STRING PRIMARY KEY
|
||||||
|
);
|
||||||
|
`)
|
||||||
|
if err != nil {
|
||||||
|
err = fmt.Errorf("Failed to create table: %w", err)
|
||||||
|
}
|
||||||
|
return CHDBSqlite{comicvinePath, sql, deleteExisting}, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s CHDBSqlite) Dump() (paths []string, bad_urls []string) {
|
||||||
|
|
||||||
|
rows, err := s.sql.Query("SELECT path from paths")
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
for rows.Next() {
|
||||||
|
var value string
|
||||||
|
err = rows.Scan(&value)
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
paths = append(paths, value)
|
||||||
|
}
|
||||||
|
rows.Close()
|
||||||
|
|
||||||
|
rows, err = s.sql.Query("SELECT url from bad_urls")
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
for rows.Next() {
|
||||||
|
var value string
|
||||||
|
err = rows.Scan(&value)
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
bad_urls = append(bad_urls, value)
|
||||||
|
}
|
||||||
|
rows.Close()
|
||||||
|
return paths, bad_urls
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s CHDBSqlite) PathHashed(path string) bool {
|
||||||
|
path, _ = filepath.Rel(s.comicvinePath, path)
|
||||||
|
dbPath := ""
|
||||||
|
|
||||||
|
if s.deleteExisting {
|
||||||
|
_ = s.sql.QueryRow("SELECT path FROM paths where path=?", path).Scan(&dbPath)
|
||||||
|
|
||||||
|
if dbPath == path {
|
||||||
|
os.Remove(filepath.Join(s.comicvinePath, path))
|
||||||
|
}
|
||||||
|
return dbPath == path
|
||||||
|
}
|
||||||
|
count := 0
|
||||||
|
_ = s.sql.QueryRow("SELECT count(path) FROM paths where path=?", path).Scan(&count)
|
||||||
|
return count > 0
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s CHDBSqlite) PathDownloaded(path string) bool {
|
||||||
|
relPath, _ := filepath.Rel(s.comicvinePath, path)
|
||||||
|
|
||||||
|
count := 0
|
||||||
|
_ = s.sql.QueryRow("SELECT count(path) FROM paths where path=?", relPath).Scan(&count)
|
||||||
|
if count != 1 {
|
||||||
|
f, err := os.Open(path)
|
||||||
|
if err == nil {
|
||||||
|
defer f.Close()
|
||||||
|
}
|
||||||
|
return !os.IsNotExist(err)
|
||||||
|
}
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s CHDBSqlite) AddPath(path string) {
|
||||||
|
relPath, _ := filepath.Rel(s.comicvinePath, path)
|
||||||
|
_, err := s.sql.Exec("INSERT INTO paths VALUES(?) ON CONFLICT DO NOTHING", relPath)
|
||||||
|
if err != nil {
|
||||||
|
log.Println(fmt.Errorf("Failed to insert %v into paths: %w", relPath, err))
|
||||||
|
}
|
||||||
|
|
||||||
|
if s.deleteExisting {
|
||||||
|
_ = os.Remove(path)
|
||||||
|
_ = RmdirP(filepath.Dir(path))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s CHDBSqlite) CheckURL(url string) bool {
|
||||||
|
count := 0
|
||||||
|
_ = s.sql.QueryRow("SELECT count(url) FROM bad_urls where url=?", url).Scan(&count)
|
||||||
|
return count > 0
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s CHDBSqlite) AddURL(url string) {
|
||||||
|
_, err := s.sql.Exec("INSERT INTO bad_urls VALUES(?) ON CONFLICT DO NOTHING", url)
|
||||||
|
if err != nil {
|
||||||
|
log.Println(fmt.Errorf("Failed to insert %v into bad_urls: %w", url, err))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s CHDBSqlite) Close() error {
|
||||||
|
return s.sql.Close()
|
||||||
|
}
|
31
cmd/bolt-migrate/main.go
Normal file
31
cmd/bolt-migrate/main.go
Normal file
@ -0,0 +1,31 @@
|
|||||||
|
package main
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"os"
|
||||||
|
|
||||||
|
ch "gitea.narnian.us/lordwelch/comic-hasher"
|
||||||
|
)
|
||||||
|
|
||||||
|
func main() {
|
||||||
|
fmt.Printf("cv path: %s Sqlite path: %s Bolt path: %s\n", os.Args[1], os.Args[2], os.Args[3])
|
||||||
|
sql, err := ch.OpenCHDBSqlite(os.Args[2], os.Args[1], false)
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
db, err := ch.OpenCHDBBolt(os.Args[3], os.Args[1], false)
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
paths, bad_urls := sql.Dump()
|
||||||
|
fmt.Printf("Dumped %d %d", len(paths), len(bad_urls))
|
||||||
|
db.Import(paths, bad_urls)
|
||||||
|
// for _, path := range paths {
|
||||||
|
// db.AddPath(filepath.Join(os.Args[1], path))
|
||||||
|
// }
|
||||||
|
// for _, url := range bad_urls {
|
||||||
|
// db.AddURL(url)
|
||||||
|
// }
|
||||||
|
sql.Close()
|
||||||
|
db.Close()
|
||||||
|
}
|
File diff suppressed because it is too large
Load Diff
@ -9,5 +9,5 @@ import (
|
|||||||
)
|
)
|
||||||
|
|
||||||
func Notify(sig chan os.Signal) {
|
func Notify(sig chan os.Signal) {
|
||||||
signal.Notify(sig, os.Interrupt, syscall.SIGABRT, syscall.SIGQUIT, syscall.SIGTERM)
|
signal.Notify(sig, os.Interrupt, syscall.SIGQUIT, syscall.SIGTERM)
|
||||||
}
|
}
|
||||||
|
366
cmd/comic-hasher/server.go
Normal file
366
cmd/comic-hasher/server.go
Normal file
@ -0,0 +1,366 @@
|
|||||||
|
package main
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bufio"
|
||||||
|
"cmp"
|
||||||
|
"context"
|
||||||
|
"encoding/json"
|
||||||
|
"fmt"
|
||||||
|
"image"
|
||||||
|
"io/fs"
|
||||||
|
"log"
|
||||||
|
"net/http"
|
||||||
|
"net/url"
|
||||||
|
"os"
|
||||||
|
"path/filepath"
|
||||||
|
"strconv"
|
||||||
|
"strings"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
ch "gitea.narnian.us/lordwelch/comic-hasher"
|
||||||
|
"gitea.narnian.us/lordwelch/goimagehash"
|
||||||
|
"golang.org/x/exp/slices"
|
||||||
|
)
|
||||||
|
|
||||||
|
type Server struct {
|
||||||
|
httpServer *http.Server
|
||||||
|
mux *CHMux
|
||||||
|
BaseURL *url.URL
|
||||||
|
hashes ch.HashStorage
|
||||||
|
Context context.Context
|
||||||
|
cancel func()
|
||||||
|
signalQueue chan os.Signal
|
||||||
|
readerQueue chan string
|
||||||
|
hashingQueue chan ch.Im
|
||||||
|
mappingQueue chan ch.ImageHash
|
||||||
|
onlyHashNewIDs bool
|
||||||
|
version string
|
||||||
|
}
|
||||||
|
|
||||||
|
type CHMux struct {
|
||||||
|
version string
|
||||||
|
*http.ServeMux
|
||||||
|
}
|
||||||
|
|
||||||
|
func (CHM *CHMux) ServeHTTP(w http.ResponseWriter, r *http.Request) {
|
||||||
|
w.Header().Set("Server", "Comic-Hasher "+CHM.version)
|
||||||
|
CHM.ServeMux.ServeHTTP(w, r)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *Server) authenticated(w http.ResponseWriter, r *http.Request) (string, bool) {
|
||||||
|
return strings.TrimSpace("lordwelch"), true
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *Server) setupAppHandlers() {
|
||||||
|
s.mux.HandleFunc("/add_cover", s.addCover)
|
||||||
|
s.mux.HandleFunc("/match_cover_hash", s.matchCoverHash)
|
||||||
|
s.mux.HandleFunc("/associate_ids", s.associateIDs)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *Server) associateIDs(w http.ResponseWriter, r *http.Request) {
|
||||||
|
user, authed := s.authenticated(w, r)
|
||||||
|
if !authed || user == "" {
|
||||||
|
http.Error(w, "Invalid Auth", http.StatusForbidden)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
var (
|
||||||
|
values = r.URL.Query()
|
||||||
|
domain = ch.Source(strings.ToLower(strings.TrimSpace(values.Get("domain"))))
|
||||||
|
ID = strings.ToLower(strings.TrimSpace(values.Get("id")))
|
||||||
|
newDomain = ch.Source(strings.ToLower(strings.TrimSpace(values.Get("newDomain"))))
|
||||||
|
newID = strings.ToLower(strings.TrimSpace(values.Get("newID")))
|
||||||
|
)
|
||||||
|
if ID == "" {
|
||||||
|
msg := "No ID Provided"
|
||||||
|
log.Println(msg)
|
||||||
|
writeJson(w, http.StatusBadRequest, result{Msg: msg})
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if domain == "" {
|
||||||
|
msg := "No domain Provided"
|
||||||
|
log.Println(msg)
|
||||||
|
writeJson(w, http.StatusBadRequest, result{Msg: msg})
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if newID == "" {
|
||||||
|
msg := "No newID Provided"
|
||||||
|
log.Println(msg)
|
||||||
|
writeJson(w, http.StatusBadRequest, result{Msg: msg})
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if newDomain == "" {
|
||||||
|
msg := "No newDomain Provided"
|
||||||
|
log.Println(msg)
|
||||||
|
writeJson(w, http.StatusBadRequest, result{Msg: msg})
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if newDomain == domain {
|
||||||
|
msg := "newDomain cannot be the same as the existing domain"
|
||||||
|
log.Println(msg)
|
||||||
|
writeJson(w, http.StatusBadRequest, result{Msg: msg})
|
||||||
|
return
|
||||||
|
}
|
||||||
|
log.Printf("Attempting to associate %s:%s to %s:%s", domain, ID, newDomain, newID)
|
||||||
|
err := s.hashes.AssociateIDs([]ch.NewIDs{{
|
||||||
|
OldID: ch.ID{
|
||||||
|
Domain: &domain,
|
||||||
|
ID: ID,
|
||||||
|
},
|
||||||
|
NewID: ch.ID{
|
||||||
|
Domain: &newDomain,
|
||||||
|
ID: newID,
|
||||||
|
},
|
||||||
|
}})
|
||||||
|
|
||||||
|
if err == nil {
|
||||||
|
writeJson(w, http.StatusOK, result{Msg: "New ID added"})
|
||||||
|
} else {
|
||||||
|
writeJson(w, http.StatusOK, result{Msg: err.Error()})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
type result struct {
|
||||||
|
Results []ch.Result `json:"results,omitempty"`
|
||||||
|
Msg string `json:"msg,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
func writeJson(w http.ResponseWriter, status int, res result) {
|
||||||
|
w.Header().Set("Content-Type", "application/json; charset=utf-8")
|
||||||
|
w.Header().Set("X-Content-Type-Options", "nosniff")
|
||||||
|
var (
|
||||||
|
bytes []byte
|
||||||
|
err error
|
||||||
|
)
|
||||||
|
if bytes, err = json.Marshal(res); err != nil {
|
||||||
|
bytes, _ = json.Marshal(result{Msg: fmt.Sprintf("Failed to create json: %s", err)})
|
||||||
|
}
|
||||||
|
w.WriteHeader(status)
|
||||||
|
_, _ = w.Write(bytes)
|
||||||
|
_, _ = w.Write([]byte("\n"))
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *Server) matchCoverHash(w http.ResponseWriter, r *http.Request) {
|
||||||
|
user, authed := s.authenticated(w, r)
|
||||||
|
if !authed || user == "" {
|
||||||
|
http.Error(w, "Invalid Auth", http.StatusForbidden)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
var (
|
||||||
|
values = r.URL.Query()
|
||||||
|
ahashStr = strings.TrimSpace(values.Get("ahash"))
|
||||||
|
dhashStr = strings.TrimSpace(values.Get("dhash"))
|
||||||
|
phashStr = strings.TrimSpace(values.Get("phash"))
|
||||||
|
maxStr = strings.TrimSpace(values.Get("max"))
|
||||||
|
exactOnly = strings.ToLower(strings.TrimSpace(values.Get("exactOnly"))) != "false"
|
||||||
|
simple = strings.ToLower(strings.TrimSpace(values.Get("simple"))) == "true"
|
||||||
|
ahash uint64
|
||||||
|
dhash uint64
|
||||||
|
phash uint64
|
||||||
|
max int = 8
|
||||||
|
max_tmp int
|
||||||
|
err error
|
||||||
|
hashes []ch.Hash
|
||||||
|
)
|
||||||
|
|
||||||
|
if simple {
|
||||||
|
writeJson(w, http.StatusBadRequest, result{Msg: "Simple results are no longer Supported"})
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
if ahash, err = strconv.ParseUint(ahashStr, 16, 64); err != nil && ahashStr != "" {
|
||||||
|
log.Printf("could not parse ahash: %s", ahashStr)
|
||||||
|
writeJson(w, http.StatusBadRequest, result{Msg: "hash parse failed"})
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if ahash > 0 {
|
||||||
|
hashes = append(hashes, ch.Hash{Hash: ahash, Kind: goimagehash.AHash})
|
||||||
|
}
|
||||||
|
if dhash, err = strconv.ParseUint(dhashStr, 16, 64); err != nil && dhashStr != "" {
|
||||||
|
log.Printf("could not parse dhash: %s", dhashStr)
|
||||||
|
writeJson(w, http.StatusBadRequest, result{Msg: "hash parse failed"})
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if dhash > 0 {
|
||||||
|
hashes = append(hashes, ch.Hash{Hash: dhash, Kind: goimagehash.DHash})
|
||||||
|
}
|
||||||
|
if phash, err = strconv.ParseUint(phashStr, 16, 64); err != nil && phashStr != "" {
|
||||||
|
log.Printf("could not parse phash: %s", phashStr)
|
||||||
|
writeJson(w, http.StatusBadRequest, result{Msg: "hash parse failed"})
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if phash > 0 {
|
||||||
|
hashes = append(hashes, ch.Hash{Hash: phash, Kind: goimagehash.PHash})
|
||||||
|
}
|
||||||
|
if max_tmp, err = strconv.Atoi(maxStr); err != nil && maxStr != "" {
|
||||||
|
log.Printf("Invalid Max: %s", maxStr)
|
||||||
|
writeJson(w, http.StatusBadRequest, result{Msg: fmt.Sprintf("Invalid Max: %s", maxStr)})
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if maxStr != "" {
|
||||||
|
max = max_tmp
|
||||||
|
}
|
||||||
|
|
||||||
|
if max > 8 {
|
||||||
|
log.Printf("Max must be less than 9: %d", max)
|
||||||
|
writeJson(w, http.StatusBadRequest, result{Msg: fmt.Sprintf("Max must be less than 9: %d", max)})
|
||||||
|
return
|
||||||
|
}
|
||||||
|
matches, err := s.hashes.GetMatches(hashes, max, exactOnly)
|
||||||
|
slices.SortFunc(matches, func(a ch.Result, b ch.Result) int {
|
||||||
|
return cmp.Compare(a.Distance, b.Distance)
|
||||||
|
})
|
||||||
|
log.Println(err)
|
||||||
|
if len(matches) > 0 {
|
||||||
|
var msg string = ""
|
||||||
|
if err != nil {
|
||||||
|
msg = err.Error()
|
||||||
|
}
|
||||||
|
|
||||||
|
writeJson(w, http.StatusOK, result{
|
||||||
|
Results: matches,
|
||||||
|
Msg: msg,
|
||||||
|
})
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
writeJson(w, http.StatusNotFound, result{Msg: "No hashes found"})
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *Server) addCover(w http.ResponseWriter, r *http.Request) {
|
||||||
|
user, authed := s.authenticated(w, r)
|
||||||
|
if !authed || user == "" {
|
||||||
|
http.Error(w, "Invalid Auth", http.StatusForbidden)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if true {
|
||||||
|
w.WriteHeader(http.StatusNotImplemented)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
var (
|
||||||
|
values = r.URL.Query()
|
||||||
|
domain = strings.TrimSpace(values.Get("domain"))
|
||||||
|
ID = strings.TrimSpace(values.Get("id"))
|
||||||
|
)
|
||||||
|
|
||||||
|
if ID == "" {
|
||||||
|
log.Println("No ID Provided")
|
||||||
|
writeJson(w, http.StatusBadRequest, result{Msg: "No ID Provided"})
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if domain == "" {
|
||||||
|
log.Println("No domain Provided")
|
||||||
|
writeJson(w, http.StatusBadRequest, result{Msg: "No Domain Provided"})
|
||||||
|
return
|
||||||
|
}
|
||||||
|
i, format, err := image.Decode(r.Body)
|
||||||
|
if err != nil {
|
||||||
|
msg := fmt.Sprintf("Failed to decode Image: %s", err)
|
||||||
|
log.Println(msg)
|
||||||
|
writeJson(w, http.StatusBadRequest, result{Msg: msg})
|
||||||
|
return
|
||||||
|
}
|
||||||
|
log.Printf("Decoded %s image from %s", format, user)
|
||||||
|
select {
|
||||||
|
case <-s.Context.Done():
|
||||||
|
log.Println("Recieved quit")
|
||||||
|
return
|
||||||
|
default:
|
||||||
|
}
|
||||||
|
s.hashingQueue <- ch.Im{Im: i, Format: format, ID: ch.ID{Domain: ch.NewSource(domain), ID: ID}}
|
||||||
|
writeJson(w, http.StatusOK, result{Msg: "Success"})
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *Server) mapper(done func()) {
|
||||||
|
defer done()
|
||||||
|
for hash := range s.mappingQueue {
|
||||||
|
s.hashes.MapHashes(hash)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *Server) hasher(workerID int, done func(int)) {
|
||||||
|
defer done(workerID)
|
||||||
|
for image := range s.hashingQueue {
|
||||||
|
start := time.Now()
|
||||||
|
if image.NewOnly && len(s.hashes.GetIDs(image.ID)) > 0 {
|
||||||
|
log.Printf("Skipping existing hash with ID: %s found", image.ID)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
hash := ch.HashImage(image)
|
||||||
|
if *hash.ID.Domain == "" || hash.ID.ID == "" {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
select {
|
||||||
|
// TODO: Check channel pipelines
|
||||||
|
case s.mappingQueue <- hash:
|
||||||
|
default:
|
||||||
|
}
|
||||||
|
|
||||||
|
elapsed := time.Since(start)
|
||||||
|
log.Printf("Hashing took %v: worker: %v. %s: %064b id: %s\n", elapsed, workerID, hash.Hashes[0].Kind, hash.Hashes[0].Hash, hash.ID)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *Server) reader(workerID int, done func(i int)) {
|
||||||
|
defer done(workerID)
|
||||||
|
for path := range s.readerQueue {
|
||||||
|
|
||||||
|
id := ch.ID{
|
||||||
|
Domain: ch.NewSource(filepath.Base(filepath.Dir(filepath.Dir(path)))),
|
||||||
|
ID: filepath.Base(filepath.Dir(path)),
|
||||||
|
}
|
||||||
|
if len(s.hashes.GetIDs(id)) > 0 {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
file, err := os.Open(path)
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
i, format, err := image.Decode(bufio.NewReader(file))
|
||||||
|
file.Close()
|
||||||
|
if err != nil {
|
||||||
|
continue // skip this image
|
||||||
|
}
|
||||||
|
|
||||||
|
im := ch.Im{
|
||||||
|
Im: i,
|
||||||
|
Format: format,
|
||||||
|
ID: id,
|
||||||
|
NewOnly: s.onlyHashNewIDs,
|
||||||
|
}
|
||||||
|
select {
|
||||||
|
case s.hashingQueue <- im:
|
||||||
|
default:
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *Server) HashLocalImages(opts Opts) {
|
||||||
|
if opts.coverPath == "" {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
go func() {
|
||||||
|
log.Println("Hashing covers at ", opts.coverPath)
|
||||||
|
start := time.Now()
|
||||||
|
err := filepath.WalkDir(opts.coverPath, func(path string, d fs.DirEntry, err error) error {
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
select {
|
||||||
|
case <-s.Context.Done():
|
||||||
|
log.Println("Recieved quit")
|
||||||
|
err = s.httpServer.Shutdown(context.TODO())
|
||||||
|
return fmt.Errorf("Recieved quit: %w", err)
|
||||||
|
default:
|
||||||
|
}
|
||||||
|
if d.IsDir() {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
s.readerQueue <- path
|
||||||
|
return nil
|
||||||
|
})
|
||||||
|
elapsed := time.Since(start)
|
||||||
|
log.Println("Err:", err, "local hashing took", elapsed)
|
||||||
|
}()
|
||||||
|
}
|
@ -106,7 +106,7 @@ func main() {
|
|||||||
debugImage(debugim, 8, 8)
|
debugImage(debugim, 8, 8)
|
||||||
}
|
}
|
||||||
|
|
||||||
hash := ch.HashImage(ch.Im{Im: im, Format: format, ID: ch.ID{Domain: ch.Source(ch.ComicVine), ID: "nothing"}})
|
hash := ch.HashImage(ch.Im{Im: im, Format: format, ID: ch.ID{Domain: ch.NewSource(ch.ComicVine), ID: "nothing"}})
|
||||||
|
|
||||||
fmt.Println("ahash: ", goimagehash.NewImageHash(hash.Hashes[0].Hash, hash.Hashes[0].Kind).BinString())
|
fmt.Println("ahash: ", goimagehash.NewImageHash(hash.Hashes[0].Hash, hash.Hashes[0].Kind).BinString())
|
||||||
fmt.Println("dhash: ", goimagehash.NewImageHash(hash.Hashes[1].Hash, hash.Hashes[1].Kind).BinString())
|
fmt.Println("dhash: ", goimagehash.NewImageHash(hash.Hashes[1].Hash, hash.Hashes[1].Kind).BinString())
|
||||||
|
701
cv/cv.go
Normal file
701
cv/cv.go
Normal file
@ -0,0 +1,701 @@
|
|||||||
|
package cv
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bufio"
|
||||||
|
"bytes"
|
||||||
|
"cmp"
|
||||||
|
"context"
|
||||||
|
"encoding/json"
|
||||||
|
"errors"
|
||||||
|
"fmt"
|
||||||
|
"io"
|
||||||
|
"io/fs"
|
||||||
|
"log"
|
||||||
|
"net/http"
|
||||||
|
"net/url"
|
||||||
|
"os"
|
||||||
|
"path"
|
||||||
|
"path/filepath"
|
||||||
|
"strconv"
|
||||||
|
"strings"
|
||||||
|
"sync"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"slices"
|
||||||
|
|
||||||
|
ch "gitea.narnian.us/lordwelch/comic-hasher"
|
||||||
|
)
|
||||||
|
|
||||||
|
type Download struct {
|
||||||
|
URL string
|
||||||
|
Dest string
|
||||||
|
IssueID string
|
||||||
|
Image *bytes.Buffer
|
||||||
|
}
|
||||||
|
|
||||||
|
type Issue struct {
|
||||||
|
ID int `json:"id"`
|
||||||
|
IssueNumber string `json:"issue_number"`
|
||||||
|
Image struct {
|
||||||
|
IconURL string `json:"icon_url,omitempty"`
|
||||||
|
MediumURL string `json:"medium_url,omitempty"`
|
||||||
|
ScreenURL string `json:"screen_url,omitempty"`
|
||||||
|
ScreenLargeURL string `json:"screen_large_url,omitempty"`
|
||||||
|
SmallURL string `json:"small_url,omitempty"`
|
||||||
|
SuperURL string `json:"super_url,omitempty"`
|
||||||
|
ThumbURL string `json:"thumb_url"`
|
||||||
|
TinyURL string `json:"tiny_url,omitempty"`
|
||||||
|
OriginalURL string `json:"original_url"`
|
||||||
|
ImageTags string `json:"image_tags"`
|
||||||
|
} `json:"image"`
|
||||||
|
Volume struct {
|
||||||
|
ID int `json:"id"`
|
||||||
|
} `json:"volume"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type CVResult struct {
|
||||||
|
// Error string `json:"error"`
|
||||||
|
// Limit int `json:"limit"`
|
||||||
|
Offset int `json:"offset"`
|
||||||
|
NumberOfPageResults int `json:"number_of_page_results"`
|
||||||
|
NumberOfTotalResults int `json:"number_of_total_results"`
|
||||||
|
StatusCode int `json:"status_code"`
|
||||||
|
Results []Issue `json:"results"`
|
||||||
|
// Version string `json:"version"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type CVDownloader struct {
|
||||||
|
APIKey string
|
||||||
|
JSONPath string
|
||||||
|
ImagePath string
|
||||||
|
ImageTypes []string
|
||||||
|
SendExistingImages bool
|
||||||
|
KeepDownloadedImages bool
|
||||||
|
Context context.Context
|
||||||
|
FinishedDownloadQueue chan Download
|
||||||
|
|
||||||
|
fileList []string
|
||||||
|
totalResults int
|
||||||
|
imageWG sync.WaitGroup
|
||||||
|
downloadQueue chan *CVResult
|
||||||
|
imageDownloads chan download
|
||||||
|
notFound chan download
|
||||||
|
chdb ch.CHDB
|
||||||
|
bufPool *sync.Pool
|
||||||
|
get_id func(id ch.ID) ch.IDList
|
||||||
|
only_hash_new_ids bool
|
||||||
|
}
|
||||||
|
|
||||||
|
var (
|
||||||
|
ErrQuit = errors.New("quit")
|
||||||
|
ErrInvalidPage = errors.New("invalid ComicVine page")
|
||||||
|
ErrInvalidIndex = errors.New("invalid page index")
|
||||||
|
ErrDownloadFail = errors.New("download failed")
|
||||||
|
ErrMissingPage = errors.New("page missing")
|
||||||
|
ErrUpdateNeeded = errors.New("update needed")
|
||||||
|
)
|
||||||
|
|
||||||
|
func (c *CVDownloader) readJson() ([]*CVResult, error) {
|
||||||
|
var issues []*CVResult
|
||||||
|
for _, filename := range c.fileList {
|
||||||
|
if c.hasQuit() {
|
||||||
|
return nil, ErrQuit
|
||||||
|
}
|
||||||
|
result, err := c.loadIssues(filename)
|
||||||
|
if err != nil {
|
||||||
|
if err == ErrInvalidPage {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
return issues, err
|
||||||
|
}
|
||||||
|
|
||||||
|
c.totalResults = max(result.NumberOfTotalResults, c.totalResults)
|
||||||
|
issues = append(issues, result)
|
||||||
|
}
|
||||||
|
return issues, nil
|
||||||
|
}
|
||||||
|
func (c *CVDownloader) loadIssues(filename string) (*CVResult, error) {
|
||||||
|
tmp := &CVResult{Results: make([]Issue, 0, 100)}
|
||||||
|
file, err := os.Open(filepath.Join(c.JSONPath, filename))
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
bytes, err := io.ReadAll(file)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
err = json.Unmarshal(bytes, tmp)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
if getOffset(filename) != tmp.Offset {
|
||||||
|
return nil, ErrInvalidPage
|
||||||
|
}
|
||||||
|
return tmp, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func Get(url string) (*http.Response, error, func()) {
|
||||||
|
ctx, cancel := context.WithTimeout(context.Background(), time.Second*20)
|
||||||
|
req, err := http.NewRequestWithContext(ctx, "GET", url, nil)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err, cancel
|
||||||
|
}
|
||||||
|
resp, err := http.DefaultClient.Do(req)
|
||||||
|
return resp, err, cancel
|
||||||
|
}
|
||||||
|
|
||||||
|
func getOffset(name string) int {
|
||||||
|
i, _ := strconv.Atoi(name[3 : len(name)-1-4])
|
||||||
|
return i
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *CVDownloader) findDownloadedPage(offset int) int {
|
||||||
|
index := offset / 100
|
||||||
|
if index < len(c.fileList) && getOffset(c.fileList[index]) == offset { // If it's in order and it's not missing it should be here
|
||||||
|
return index
|
||||||
|
}
|
||||||
|
index, found := slices.BinarySearchFunc(c.fileList, offset, func(a string, b int) int {
|
||||||
|
return cmp.Compare(getOffset(a), b)
|
||||||
|
})
|
||||||
|
if found {
|
||||||
|
return index
|
||||||
|
}
|
||||||
|
return -1
|
||||||
|
}
|
||||||
|
func (c *CVDownloader) getDownloadedIssues(offset int, update bool) (*CVResult, error) {
|
||||||
|
index := c.findDownloadedPage(offset)
|
||||||
|
if index < 0 {
|
||||||
|
return nil, ErrMissingPage
|
||||||
|
}
|
||||||
|
issue, err := c.loadIssues(c.fileList[index])
|
||||||
|
if err != nil || issue == nil {
|
||||||
|
err = fmt.Errorf("Failed to read page at offset %d: %w", offset, err)
|
||||||
|
os.Remove(filepath.Join(c.JSONPath, c.fileList[index]))
|
||||||
|
c.fileList = slices.Delete(c.fileList, index, index+1)
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
c.totalResults = max(c.totalResults, issue.NumberOfTotalResults)
|
||||||
|
|
||||||
|
if update && (len(issue.Results) == 0 || issue.Results[0].IssueNumber == "") {
|
||||||
|
err = fmt.Errorf("Deleting page %d to update records from cv", offset)
|
||||||
|
os.Remove(filepath.Join(c.JSONPath, c.fileList[index]))
|
||||||
|
c.fileList = slices.Delete(c.fileList, index, index+1)
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
if c.totalResults == issue.Offset+issue.NumberOfPageResults {
|
||||||
|
if index != len(c.fileList)-1 {
|
||||||
|
err = fmt.Errorf("Wrong index: expected %d got %d", len(c.fileList), index)
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
log.Println("Deleting the last page to detect new comics")
|
||||||
|
os.Remove(filepath.Join(c.JSONPath, c.fileList[index]))
|
||||||
|
c.fileList = slices.Delete(c.fileList, index, index+1)
|
||||||
|
}
|
||||||
|
|
||||||
|
return issue, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// updateIssues c.downloadQueue must not be closed before this function has returned
|
||||||
|
func (c *CVDownloader) updateIssues() (int, error) {
|
||||||
|
base_url, err := url.Parse("https://comicvine.gamespot.com/api/issues/?sort=date_added,id:asc&format=json&field_list=id,issue_number,image,volume")
|
||||||
|
if err != nil {
|
||||||
|
log.Fatal(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
query := base_url.Query()
|
||||||
|
query.Add("api_key", c.APIKey)
|
||||||
|
base_url.RawQuery = query.Encode()
|
||||||
|
c.totalResults = max(c.totalResults, 1)
|
||||||
|
failCount := 0
|
||||||
|
prev := -1
|
||||||
|
offset := 0
|
||||||
|
retry := func(url string, err error) bool {
|
||||||
|
if errors.Is(err, context.Canceled) {
|
||||||
|
log.Println("Server closed")
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
log.Printf("Failed to download %#v at offset %v: %v Attempt #%d", url, offset, err, failCount+1)
|
||||||
|
if prev == offset {
|
||||||
|
sleepTime := time.Second * 36
|
||||||
|
if failCount > 2 {
|
||||||
|
sleepTime = time.Minute * 10
|
||||||
|
}
|
||||||
|
log.Println("This page failed to download, lets wait for", sleepTime, "and hope it works")
|
||||||
|
select {
|
||||||
|
case <-c.Context.Done(): // allows us to return immediately even during a timeout
|
||||||
|
return false
|
||||||
|
case <-time.After(sleepTime):
|
||||||
|
}
|
||||||
|
}
|
||||||
|
prev = offset
|
||||||
|
failCount += 1
|
||||||
|
offset -= 100
|
||||||
|
return failCount < 15
|
||||||
|
}
|
||||||
|
updated := 0
|
||||||
|
for offset = 0; offset <= c.totalResults; offset += 100 {
|
||||||
|
if c.hasQuit() {
|
||||||
|
return offset - 100, ErrQuit
|
||||||
|
}
|
||||||
|
|
||||||
|
issue, err := c.getDownloadedIssues(offset, updated < 9)
|
||||||
|
if err == nil && issue != nil {
|
||||||
|
|
||||||
|
prev = -1
|
||||||
|
failCount = 0
|
||||||
|
select {
|
||||||
|
case <-c.Context.Done(): // allows us to return immediately even during a timeout
|
||||||
|
return offset - 100, ErrQuit
|
||||||
|
case c.downloadQueue <- issue:
|
||||||
|
}
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if errors.Is(err, ErrInvalidIndex) {
|
||||||
|
return offset - 100, err
|
||||||
|
}
|
||||||
|
if err != nil && !errors.Is(err, ErrMissingPage) {
|
||||||
|
log.Println(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
log.Println("Starting download at offset", offset)
|
||||||
|
issue = &CVResult{}
|
||||||
|
URI := (*base_url)
|
||||||
|
query = base_url.Query()
|
||||||
|
query.Add("offset", strconv.Itoa(offset))
|
||||||
|
URI.RawQuery = query.Encode()
|
||||||
|
|
||||||
|
select {
|
||||||
|
case <-c.Context.Done(): // Allows us to return immediately even during a timeout
|
||||||
|
return offset - 100, ErrQuit
|
||||||
|
case <-time.After(10 * time.Second): // Enforces a minimum 10s wait between API hits
|
||||||
|
}
|
||||||
|
|
||||||
|
resp, err, cancelDownloadCTX := Get(URI.String())
|
||||||
|
if err != nil {
|
||||||
|
cancelDownloadCTX()
|
||||||
|
if retry(URI.String(), err) {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
// Fail and let comic-hasher try the whole thing again later
|
||||||
|
return offset - 100, fmt.Errorf("%w: %w", ErrDownloadFail, err)
|
||||||
|
}
|
||||||
|
if resp.StatusCode != 200 {
|
||||||
|
cancelDownloadCTX()
|
||||||
|
if retry(URI.String(), nil) {
|
||||||
|
_ = resp.Body.Close()
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
msg, _ := io.ReadAll(resp.Body)
|
||||||
|
_ = resp.Body.Close()
|
||||||
|
return offset - 100, fmt.Errorf("%w: response code: %d Message: %s", ErrDownloadFail, resp.StatusCode, string(msg))
|
||||||
|
}
|
||||||
|
|
||||||
|
file, err := os.Create(filepath.Join(c.JSONPath, "cv-"+strconv.Itoa(offset)+".json"))
|
||||||
|
if err != nil {
|
||||||
|
log.Fatal(err)
|
||||||
|
}
|
||||||
|
body := io.TeeReader(resp.Body, file)
|
||||||
|
err = json.NewDecoder(bufio.NewReader(body)).Decode(issue)
|
||||||
|
_ = resp.Body.Close()
|
||||||
|
_ = file.Close()
|
||||||
|
if err != nil || issue.Offset != offset {
|
||||||
|
os.Remove(filepath.Join(c.JSONPath, "cv-"+strconv.Itoa(offset)+".json"))
|
||||||
|
cancelDownloadCTX()
|
||||||
|
if retry(URI.String(), err) {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
return offset - 100, fmt.Errorf("%w: %w", ErrDownloadFail, err)
|
||||||
|
}
|
||||||
|
cancelDownloadCTX()
|
||||||
|
if issue.NumberOfTotalResults > c.totalResults {
|
||||||
|
c.totalResults = issue.NumberOfTotalResults
|
||||||
|
}
|
||||||
|
prev = -1
|
||||||
|
failCount = 0
|
||||||
|
updated += 1
|
||||||
|
select {
|
||||||
|
case c.downloadQueue <- issue:
|
||||||
|
}
|
||||||
|
c.insertIssuePage(offset)
|
||||||
|
log.Printf("Downloaded %s/cv-%v.json", c.JSONPath, offset)
|
||||||
|
}
|
||||||
|
return offset, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
type download struct {
|
||||||
|
url string
|
||||||
|
dest string
|
||||||
|
offset int
|
||||||
|
volumeID int
|
||||||
|
issueID int
|
||||||
|
finished bool
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *CVDownloader) start_downloader() {
|
||||||
|
for i := range 5 {
|
||||||
|
go func() {
|
||||||
|
log.Println("starting downloader", i)
|
||||||
|
for dl := range c.imageDownloads {
|
||||||
|
if dl.finished {
|
||||||
|
|
||||||
|
select {
|
||||||
|
case c.FinishedDownloadQueue <- Download{
|
||||||
|
URL: dl.url,
|
||||||
|
Dest: dl.dest,
|
||||||
|
IssueID: strconv.Itoa(dl.issueID),
|
||||||
|
}:
|
||||||
|
c.imageWG.Done()
|
||||||
|
}
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
dir := filepath.Dir(dl.dest)
|
||||||
|
resp, err, cancelDownload := Get(dl.url)
|
||||||
|
if err != nil {
|
||||||
|
cancelDownload()
|
||||||
|
log.Println("Failed to download", dl.volumeID, "/", dl.issueID, dl.url, err)
|
||||||
|
c.imageWG.Done()
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
cleanup := func() {
|
||||||
|
resp.Body.Close()
|
||||||
|
cancelDownload()
|
||||||
|
c.imageWG.Done()
|
||||||
|
}
|
||||||
|
if resp.StatusCode == 404 {
|
||||||
|
|
||||||
|
c.notFound <- dl
|
||||||
|
cleanup()
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if resp.StatusCode != 200 {
|
||||||
|
log.Println("Failed to download", dl.url, resp.StatusCode)
|
||||||
|
cleanup()
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
if c.KeepDownloadedImages {
|
||||||
|
_ = os.MkdirAll(dir, 0o755)
|
||||||
|
image, err := os.Create(dl.dest)
|
||||||
|
if err != nil {
|
||||||
|
log.Println("Unable to create image file", dl.dest, err)
|
||||||
|
os.Remove(dl.dest)
|
||||||
|
image.Close()
|
||||||
|
cleanup()
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
log.Println("downloading", dl.dest)
|
||||||
|
_, err = io.Copy(image, resp.Body)
|
||||||
|
image.Close()
|
||||||
|
if err != nil {
|
||||||
|
log.Println("Failed when downloading image", err)
|
||||||
|
os.Remove(dl.dest)
|
||||||
|
cleanup()
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
c.FinishedDownloadQueue <- Download{
|
||||||
|
URL: dl.url,
|
||||||
|
Dest: dl.dest,
|
||||||
|
IssueID: strconv.Itoa(dl.issueID),
|
||||||
|
}
|
||||||
|
|
||||||
|
} else {
|
||||||
|
image := c.bufPool.Get().(*bytes.Buffer)
|
||||||
|
image.Reset()
|
||||||
|
log.Println("downloading", dl.dest)
|
||||||
|
_, err = io.Copy(image, resp.Body)
|
||||||
|
if err != nil {
|
||||||
|
log.Println("Failed when downloading image", err)
|
||||||
|
cleanup()
|
||||||
|
os.Remove(dl.dest)
|
||||||
|
// Something failed let this buffer GC instead of saving it
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
c.FinishedDownloadQueue <- Download{
|
||||||
|
URL: dl.url,
|
||||||
|
Dest: dl.dest,
|
||||||
|
IssueID: strconv.Itoa(dl.issueID),
|
||||||
|
Image: image,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
cleanup()
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *CVDownloader) handleNotFound() {
|
||||||
|
for failedDownload := range c.notFound {
|
||||||
|
c.chdb.AddURL(failedDownload.url)
|
||||||
|
log.Printf("Not found: volumeID: %d issueID: %d Offset: %d URL: %s\n", failedDownload.volumeID, failedDownload.issueID, failedDownload.offset, failedDownload.url)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *CVDownloader) downloadImages() {
|
||||||
|
defer func() {
|
||||||
|
log.Println("Waiting for final images to complete download")
|
||||||
|
c.imageWG.Wait()
|
||||||
|
}()
|
||||||
|
go c.start_downloader()
|
||||||
|
|
||||||
|
go c.handleNotFound()
|
||||||
|
added := 0
|
||||||
|
for list := range c.downloadQueue {
|
||||||
|
log.Printf("Checking downloads at offset %v\r", list.Offset)
|
||||||
|
for _, issue := range list.Results {
|
||||||
|
type image struct {
|
||||||
|
url string
|
||||||
|
name string
|
||||||
|
}
|
||||||
|
imageURLs := []image{{issue.Image.IconURL, "icon_url"}, {issue.Image.MediumURL, "medium_url"}, {issue.Image.ScreenURL, "screen_url"}, {issue.Image.ScreenLargeURL, "screen_large_url"}, {issue.Image.SmallURL, "small_url"}, {issue.Image.SuperURL, "super_url"}, {issue.Image.ThumbURL, "thumb_url"}, {issue.Image.TinyURL, "tiny_url"}, {issue.Image.OriginalURL, "original_url"}}
|
||||||
|
for _, image := range imageURLs {
|
||||||
|
if len(c.ImageTypes) > 0 && !slices.Contains(c.ImageTypes, image.name) {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if c.chdb.CheckURL(image.url) {
|
||||||
|
log.Printf("Skipping known bad url %s", image.url)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if strings.HasSuffix(image.url, "6373148-blank.png") {
|
||||||
|
c.notFound <- download{
|
||||||
|
url: image.url,
|
||||||
|
offset: list.Offset,
|
||||||
|
volumeID: issue.Volume.ID,
|
||||||
|
issueID: issue.ID,
|
||||||
|
}
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
uri, err := url.ParseRequestURI(image.url)
|
||||||
|
if err != nil {
|
||||||
|
c.notFound <- download{
|
||||||
|
url: image.url,
|
||||||
|
offset: list.Offset,
|
||||||
|
volumeID: issue.Volume.ID,
|
||||||
|
issueID: issue.ID,
|
||||||
|
finished: true,
|
||||||
|
}
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
ext := strings.TrimSuffix(strings.ToLower(path.Ext(uri.Path)), "~original")
|
||||||
|
if ext == "" || (len(ext) > 4 && !slices.Contains([]string{".avif", ".webp", ".tiff", ".heif"}, ext)) {
|
||||||
|
ext = ".jpg"
|
||||||
|
}
|
||||||
|
dir := filepath.Join(c.ImagePath, strconv.Itoa(issue.Volume.ID), strconv.Itoa(issue.ID))
|
||||||
|
path := filepath.Join(dir, image.name+ext)
|
||||||
|
|
||||||
|
ids := c.get_id(ch.ID{
|
||||||
|
Domain: ch.NewSource(ch.ComicVine),
|
||||||
|
ID: strconv.Itoa(issue.ID),
|
||||||
|
})
|
||||||
|
if c.chdb.PathDownloaded(path) || c.only_hash_new_ids && len(ids) > 0 {
|
||||||
|
if _, err = os.Stat(path); c.SendExistingImages && err == nil {
|
||||||
|
// We don't add to the count of added as these should be processed immediately
|
||||||
|
log.Printf("Sending Existing image %v/%v %v", issue.Volume.ID, issue.ID, path)
|
||||||
|
c.imageWG.Add(1)
|
||||||
|
c.imageDownloads <- download{
|
||||||
|
url: image.url,
|
||||||
|
dest: path,
|
||||||
|
offset: list.Offset,
|
||||||
|
volumeID: issue.Volume.ID,
|
||||||
|
issueID: issue.ID,
|
||||||
|
finished: true,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
continue // If it exists assume it is fine, adding some basic verification might be a good idea later
|
||||||
|
}
|
||||||
|
added++
|
||||||
|
|
||||||
|
c.imageWG.Add(1)
|
||||||
|
c.imageDownloads <- download{
|
||||||
|
url: image.url,
|
||||||
|
dest: path,
|
||||||
|
offset: list.Offset,
|
||||||
|
volumeID: issue.Volume.ID,
|
||||||
|
issueID: issue.ID,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if added > 200 {
|
||||||
|
// On a clean single image type run each page would have 100 downloads of a single cover type but stuff happens so we only wait once we have sent 200 to the queue
|
||||||
|
log.Println("waiting for", added, "downloads at offset", list.Offset)
|
||||||
|
beforeWait := time.Now()
|
||||||
|
c.imageWG.Wait()
|
||||||
|
waited := time.Since(beforeWait)
|
||||||
|
added = 0
|
||||||
|
// If we had to wait for the arbitrarily picked time of 7.4 seconds it means we had a backed up queue (slow hashing can also cause it to wait longer), lets wait to give the CV servers a break
|
||||||
|
if waited > time.Duration(7.4*float64(time.Second)) {
|
||||||
|
t := 10 * time.Second
|
||||||
|
log.Println("Waiting for", t, "at offset", list.Offset, "had to wait for", waited)
|
||||||
|
select {
|
||||||
|
case <-time.After(t):
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
// Things are too fast we can't depend CV being slow to manage our download speed
|
||||||
|
// We sleep for 3 seconds so we don't overload CV
|
||||||
|
time.Sleep(3 * time.Second)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *CVDownloader) cleanBadURLs() error {
|
||||||
|
|
||||||
|
var indexesToRemove []int
|
||||||
|
list:
|
||||||
|
for i, jsonFile := range c.fileList {
|
||||||
|
list, err := c.loadIssues(jsonFile)
|
||||||
|
if err != nil {
|
||||||
|
indexesToRemove = append(indexesToRemove, i)
|
||||||
|
os.Remove(filepath.Join(c.JSONPath, jsonFile))
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
for _, issue := range list.Results {
|
||||||
|
for _, url := range []string{issue.Image.IconURL, issue.Image.MediumURL, issue.Image.ScreenURL, issue.Image.ScreenLargeURL, issue.Image.SmallURL, issue.Image.SuperURL, issue.Image.ThumbURL, issue.Image.TinyURL, issue.Image.OriginalURL} {
|
||||||
|
if c.chdb.CheckURL(url) {
|
||||||
|
indexesToRemove = append(indexesToRemove, i)
|
||||||
|
if err := os.Remove(filepath.Join(c.JSONPath, jsonFile)); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
// We've removed the entire page, lets see if the new url works
|
||||||
|
continue list
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
slices.Reverse(indexesToRemove)
|
||||||
|
for _, i := range indexesToRemove {
|
||||||
|
c.fileList = slices.Delete(c.fileList, i, min(i+1, len(c.fileList)-1))
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *CVDownloader) hasQuit() bool {
|
||||||
|
select {
|
||||||
|
case <-c.Context.Done():
|
||||||
|
return true
|
||||||
|
default:
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *CVDownloader) cleanDirs() {
|
||||||
|
_ = filepath.WalkDir(c.ImagePath, func(path string, d fs.DirEntry, err error) error {
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
if d.IsDir() {
|
||||||
|
path, _ = filepath.Abs(path)
|
||||||
|
err := ch.RmdirP(path)
|
||||||
|
// The error is only for the first path value. EG ch.RmdirP("/test/t") will only return the error for os.Remove("/test/t") not os.Remove("test")
|
||||||
|
if err == nil {
|
||||||
|
return filepath.SkipDir
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
})
|
||||||
|
}
|
||||||
|
func (c *CVDownloader) insertIssuePage(offset int) {
|
||||||
|
index, found := slices.BinarySearchFunc(c.fileList, offset, func(a string, b int) int {
|
||||||
|
return cmp.Compare(getOffset(a), b)
|
||||||
|
})
|
||||||
|
if found {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
c.fileList = slices.Insert(c.fileList, index, fmt.Sprintf("cv-%v.json", offset))
|
||||||
|
}
|
||||||
|
|
||||||
|
func NewCVDownloader(ctx context.Context, bufPool *sync.Pool, only_hash_new_ids bool, get_id func(id ch.ID) ch.IDList, chdb ch.CHDB, workPath, APIKey string, imageTypes []string, keepDownloadedImages, sendExistingImages bool, finishedDownloadQueue chan Download) *CVDownloader {
|
||||||
|
return &CVDownloader{
|
||||||
|
Context: ctx,
|
||||||
|
JSONPath: filepath.Join(workPath, "_json"),
|
||||||
|
ImagePath: filepath.Join(workPath, "_image"),
|
||||||
|
APIKey: APIKey,
|
||||||
|
bufPool: bufPool, // Only used if keepDownloadedImages is false to save memory on byte buffers. The buffers get sent back via finishedDownloadQueue
|
||||||
|
FinishedDownloadQueue: finishedDownloadQueue,
|
||||||
|
SendExistingImages: sendExistingImages,
|
||||||
|
KeepDownloadedImages: keepDownloadedImages,
|
||||||
|
ImageTypes: imageTypes,
|
||||||
|
chdb: chdb,
|
||||||
|
get_id: get_id,
|
||||||
|
only_hash_new_ids: only_hash_new_ids,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func DownloadCovers(c *CVDownloader) {
|
||||||
|
var (
|
||||||
|
err error
|
||||||
|
)
|
||||||
|
c.downloadQueue = make(chan *CVResult) // This is just json it shouldn't take up much more than 122 MB
|
||||||
|
c.imageDownloads = make(chan download, 1) // These are just URLs should only take a few MB
|
||||||
|
c.notFound = make(chan download, 1) // Same here
|
||||||
|
os.MkdirAll(c.JSONPath, 0o777)
|
||||||
|
f, _ := os.Create(filepath.Join(c.ImagePath, ".keep"))
|
||||||
|
f.Close()
|
||||||
|
if !c.KeepDownloadedImages {
|
||||||
|
log.Println("Cleaning directories")
|
||||||
|
c.cleanDirs()
|
||||||
|
}
|
||||||
|
log.Println("Reading json")
|
||||||
|
var d *os.File
|
||||||
|
d, err = os.Open(c.JSONPath)
|
||||||
|
c.fileList, err = d.Readdirnames(-1)
|
||||||
|
for i := len(c.fileList) - 1; i >= 0; i-- {
|
||||||
|
if !strings.Contains(c.fileList[i], "json") {
|
||||||
|
c.fileList = slices.Delete(c.fileList, i, i+1)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if err != nil {
|
||||||
|
panic(fmt.Errorf("Unable to open path for json files: %w", err))
|
||||||
|
}
|
||||||
|
|
||||||
|
slices.SortFunc(c.fileList, func(x, y string) int {
|
||||||
|
return cmp.Compare(getOffset(x), getOffset(y))
|
||||||
|
})
|
||||||
|
if len(c.fileList) > 0 {
|
||||||
|
c.totalResults = getOffset(c.fileList[len(c.fileList)-1])
|
||||||
|
}
|
||||||
|
log.Println("Number of pages", len(c.fileList), "Expected Pages:", c.totalResults/100)
|
||||||
|
log.Println("Updating issues now")
|
||||||
|
|
||||||
|
dwg := sync.WaitGroup{}
|
||||||
|
dwg.Add(1)
|
||||||
|
go func() {
|
||||||
|
c.downloadImages()
|
||||||
|
dwg.Done()
|
||||||
|
}()
|
||||||
|
|
||||||
|
offset, err := c.updateIssues()
|
||||||
|
if err != nil {
|
||||||
|
log.Printf("Failed to download CV Covers: %s", err)
|
||||||
|
}
|
||||||
|
issueCount := len(c.fileList) * 100
|
||||||
|
|
||||||
|
log.Println("Number of issues", issueCount, " expected:", c.totalResults)
|
||||||
|
|
||||||
|
close(c.downloadQueue) // sends only happen in c.updateIssues which has already been called
|
||||||
|
// We don't drain here as we want to process them
|
||||||
|
|
||||||
|
log.Println("Waiting for downloaders")
|
||||||
|
dwg.Wait()
|
||||||
|
close(c.imageDownloads)
|
||||||
|
for dw := range c.imageDownloads {
|
||||||
|
fmt.Println("Skipping cv download", dw.issueID)
|
||||||
|
}
|
||||||
|
close(c.notFound)
|
||||||
|
for dw := range c.notFound {
|
||||||
|
fmt.Println("Skipping not found", dw.issueID)
|
||||||
|
}
|
||||||
|
|
||||||
|
// We drain this at the end because we need to wait for the images to download
|
||||||
|
for dw := range c.downloadQueue {
|
||||||
|
fmt.Println("Skipping page download", dw.Offset)
|
||||||
|
}
|
||||||
|
|
||||||
|
log.Println("Completed downloading images")
|
||||||
|
log.Println("Last offset", offset)
|
||||||
|
}
|
90
go.mod
90
go.mod
@ -2,60 +2,70 @@ module gitea.narnian.us/lordwelch/comic-hasher
|
|||||||
|
|
||||||
go 1.23.0
|
go 1.23.0
|
||||||
|
|
||||||
|
toolchain go1.24.0
|
||||||
|
|
||||||
|
// Main comic-hasher
|
||||||
|
require (
|
||||||
|
gitea.narnian.us/lordwelch/goimagehash v0.0.0-20250130004139-e91c39c79e0d
|
||||||
|
github.com/disintegration/imaging v1.6.3-0.20201218193011-d40f48ce0f09
|
||||||
|
github.com/json-iterator/go v1.1.12
|
||||||
|
github.com/kr/pretty v0.2.1
|
||||||
|
github.com/vmihailenco/msgpack v4.0.4+incompatible
|
||||||
|
go.etcd.io/bbolt v1.4.0
|
||||||
|
golang.org/x/exp v0.0.0-20250218142911-aa4b98e5adaa
|
||||||
|
golang.org/x/image v0.25.0
|
||||||
|
)
|
||||||
|
|
||||||
|
// Storage types
|
||||||
|
require (
|
||||||
|
github.com/mattn/go-sqlite3 v1.14.24
|
||||||
|
github.com/ncruces/go-sqlite3 v0.26.0
|
||||||
|
gonum.org/v1/gonum v0.16.0
|
||||||
|
modernc.org/sqlite v1.35.0
|
||||||
|
)
|
||||||
|
|
||||||
|
// other commands
|
||||||
require (
|
require (
|
||||||
gitea.narnian.us/lordwelch/goimagehash v0.0.0-20240812025715-33ff96e45f00
|
|
||||||
github.com/fmartingr/go-comicinfo/v2 v2.0.2
|
github.com/fmartingr/go-comicinfo/v2 v2.0.2
|
||||||
github.com/kr/pretty v0.1.0
|
github.com/mholt/archiver/v4 v4.0.0-alpha.9
|
||||||
github.com/mattn/go-sqlite3 v1.14.22
|
golang.org/x/text v0.25.0
|
||||||
github.com/mholt/archiver/v4 v4.0.0-alpha.8
|
|
||||||
github.com/ncruces/go-sqlite3 v0.18.1
|
|
||||||
golang.org/x/image v0.19.0
|
|
||||||
golang.org/x/text v0.17.0
|
|
||||||
gonum.org/v1/gonum v0.15.1
|
|
||||||
modernc.org/sqlite v1.32.0
|
|
||||||
)
|
)
|
||||||
|
|
||||||
require (
|
require (
|
||||||
github.com/vmihailenco/msgpack/v5 v5.4.1
|
github.com/STARRY-S/zip v0.1.0 // indirect
|
||||||
github.com/vmihailenco/tagparser/v2 v2.0.0 // indirect
|
github.com/andybalholm/brotli v1.1.1 // indirect
|
||||||
)
|
github.com/bodgit/plumbing v1.3.0 // indirect
|
||||||
|
github.com/bodgit/sevenzip v1.5.2 // indirect
|
||||||
require (
|
github.com/bodgit/windows v1.0.1 // indirect
|
||||||
github.com/andybalholm/brotli v1.0.4 // indirect
|
github.com/dsnet/compress v0.0.2-0.20230904184137-39efe44ab707 // indirect
|
||||||
github.com/bodgit/plumbing v1.2.0 // indirect
|
|
||||||
github.com/bodgit/sevenzip v1.3.0 // indirect
|
|
||||||
github.com/bodgit/windows v1.0.0 // indirect
|
|
||||||
github.com/connesc/cipherio v0.2.1 // indirect
|
|
||||||
github.com/disintegration/imaging v1.6.3-0.20201218193011-d40f48ce0f09 // indirect
|
|
||||||
github.com/dsnet/compress v0.0.1 // indirect
|
|
||||||
github.com/dustin/go-humanize v1.0.1 // indirect
|
github.com/dustin/go-humanize v1.0.1 // indirect
|
||||||
github.com/golang/mock v1.6.0 // indirect
|
github.com/golang/protobuf v1.5.2 // indirect
|
||||||
github.com/golang/snappy v0.0.4 // indirect
|
github.com/golang/snappy v0.0.4 // indirect
|
||||||
github.com/google/uuid v1.6.0 // indirect
|
github.com/google/uuid v1.6.0 // indirect
|
||||||
github.com/hashicorp/errwrap v1.0.0 // indirect
|
github.com/hashicorp/errwrap v1.1.0 // indirect
|
||||||
github.com/hashicorp/go-multierror v1.1.1 // indirect
|
github.com/hashicorp/go-multierror v1.1.1 // indirect
|
||||||
github.com/hashicorp/golang-lru/v2 v2.0.7 // indirect
|
github.com/hashicorp/golang-lru/v2 v2.0.7 // indirect
|
||||||
github.com/klauspost/compress v1.15.9 // indirect
|
github.com/klauspost/compress v1.17.11 // indirect
|
||||||
github.com/klauspost/pgzip v1.2.5 // indirect
|
github.com/klauspost/pgzip v1.2.6 // indirect
|
||||||
github.com/kr/text v0.1.0 // indirect
|
github.com/kr/text v0.1.0 // indirect
|
||||||
github.com/mattn/go-isatty v0.0.20 // indirect
|
github.com/mattn/go-isatty v0.0.20 // indirect
|
||||||
|
github.com/modern-go/concurrent v0.0.0-20180228061459-e0a39a4cb421 // indirect
|
||||||
|
github.com/modern-go/reflect2 v1.0.2 // indirect
|
||||||
github.com/ncruces/go-strftime v0.1.9 // indirect
|
github.com/ncruces/go-strftime v0.1.9 // indirect
|
||||||
github.com/ncruces/julianday v1.0.0 // indirect
|
github.com/ncruces/julianday v1.0.0 // indirect
|
||||||
github.com/nwaples/rardecode/v2 v2.0.0-beta.2 // indirect
|
github.com/nwaples/rardecode/v2 v2.0.0-beta.4 // indirect
|
||||||
github.com/pierrec/lz4/v4 v4.1.15 // indirect
|
github.com/pierrec/lz4/v4 v4.1.21 // indirect
|
||||||
github.com/remyoudompheng/bigfft v0.0.0-20230129092748-24d4a6f8daec // indirect
|
github.com/remyoudompheng/bigfft v0.0.0-20230129092748-24d4a6f8daec // indirect
|
||||||
github.com/tetratelabs/wazero v1.8.0 // indirect
|
github.com/sorairolake/lzip-go v0.3.5 // indirect
|
||||||
|
github.com/tetratelabs/wazero v1.9.0 // indirect
|
||||||
github.com/therootcompany/xz v1.0.1 // indirect
|
github.com/therootcompany/xz v1.0.1 // indirect
|
||||||
github.com/ulikunitz/xz v0.5.10 // indirect
|
github.com/ulikunitz/xz v0.5.12 // indirect
|
||||||
go4.org v0.0.0-20200411211856-f5505b9728dd // indirect
|
go4.org v0.0.0-20230225012048-214862532bf5 // indirect
|
||||||
golang.org/x/exp v0.0.0-20240808152545-0cdaa3abc0fa // indirect
|
golang.org/x/sys v0.33.0 // indirect
|
||||||
golang.org/x/sys v0.24.0 // indirect
|
google.golang.org/appengine v1.6.8 // indirect
|
||||||
modernc.org/gc/v3 v3.0.0-20240107210532-573471604cb6 // indirect
|
google.golang.org/protobuf v1.26.0 // indirect
|
||||||
modernc.org/libc v1.55.3 // indirect
|
gopkg.in/check.v1 v1.0.0-20201130134442-10cb98267c6c // indirect
|
||||||
modernc.org/mathutil v1.6.0 // indirect
|
modernc.org/libc v1.61.13 // indirect
|
||||||
modernc.org/memory v1.8.0 // indirect
|
modernc.org/mathutil v1.7.1 // indirect
|
||||||
modernc.org/strutil v1.2.0 // indirect
|
modernc.org/memory v1.8.2 // indirect
|
||||||
modernc.org/token v1.1.0 // indirect
|
|
||||||
)
|
)
|
||||||
|
|
||||||
replace golang.org/x/text v0.17.0 => github.com/lordwelch/text v0.0.0-20240505231825-4893f344170f
|
|
||||||
|
221
go.sum
221
go.sum
@ -15,32 +15,32 @@ cloud.google.com/go/pubsub v1.1.0/go.mod h1:EwwdRX2sKPjnvnqCa270oGRyludottCI76h+
|
|||||||
cloud.google.com/go/storage v1.0.0/go.mod h1:IhtSnM/ZTZV8YYJWCY8RULGVqBDmpoyjwiyrjsg+URw=
|
cloud.google.com/go/storage v1.0.0/go.mod h1:IhtSnM/ZTZV8YYJWCY8RULGVqBDmpoyjwiyrjsg+URw=
|
||||||
cloud.google.com/go/storage v1.5.0/go.mod h1:tpKbwo567HUNpVclU5sGELwQWBDZ8gh0ZeosJ0Rtdos=
|
cloud.google.com/go/storage v1.5.0/go.mod h1:tpKbwo567HUNpVclU5sGELwQWBDZ8gh0ZeosJ0Rtdos=
|
||||||
dmitri.shuralyov.com/gpu/mtl v0.0.0-20190408044501-666a987793e9/go.mod h1:H6x//7gZCb22OMCxBHrMx7a5I7Hp++hsVxbQ4BYO7hU=
|
dmitri.shuralyov.com/gpu/mtl v0.0.0-20190408044501-666a987793e9/go.mod h1:H6x//7gZCb22OMCxBHrMx7a5I7Hp++hsVxbQ4BYO7hU=
|
||||||
gitea.narnian.us/lordwelch/goimagehash v0.0.0-20240812025715-33ff96e45f00 h1:RNqy72W8N/mlnZGxvPoC9ch+zI3GlAGVYbBGpXOHmuY=
|
gitea.narnian.us/lordwelch/goimagehash v0.0.0-20250130004139-e91c39c79e0d h1:mFnVC/tEHk6woq6FBulwzGcuNdYn+zNhXNBILuetQJs=
|
||||||
gitea.narnian.us/lordwelch/goimagehash v0.0.0-20240812025715-33ff96e45f00/go.mod h1:kLCabSskchnLGV41s6YVXZdnLYwAxKwdXPlEuyFhC9E=
|
gitea.narnian.us/lordwelch/goimagehash v0.0.0-20250130004139-e91c39c79e0d/go.mod h1:UDwa7njhbB5nzxIjHbT9Mjlve9GYn3wzxAcQax1XRvE=
|
||||||
github.com/BurntSushi/toml v0.3.1/go.mod h1:xHWCNGjB5oqiDr8zfno3MHue2Ht5sIBksp03qcyfWMU=
|
github.com/BurntSushi/toml v0.3.1/go.mod h1:xHWCNGjB5oqiDr8zfno3MHue2Ht5sIBksp03qcyfWMU=
|
||||||
github.com/BurntSushi/xgb v0.0.0-20160522181843-27f122750802/go.mod h1:IVnqGOEym/WlBOVXweHU+Q+/VP0lqqI8lqeDx9IjBqo=
|
github.com/BurntSushi/xgb v0.0.0-20160522181843-27f122750802/go.mod h1:IVnqGOEym/WlBOVXweHU+Q+/VP0lqqI8lqeDx9IjBqo=
|
||||||
github.com/andybalholm/brotli v1.0.4 h1:V7DdXeJtZscaqfNuAdSRuRFzuiKlHSC/Zh3zl9qY3JY=
|
github.com/STARRY-S/zip v0.1.0 h1:eUER3jKmHKXjv+iy3BekLa+QnNSo1Lqz4eTzYBcGDqo=
|
||||||
github.com/andybalholm/brotli v1.0.4/go.mod h1:fO7iG3H7G2nSZ7m0zPUDn85XEX2GTukHGRSepvi9Eig=
|
github.com/STARRY-S/zip v0.1.0/go.mod h1:qj/mTZkvb3AvfGQ2e775/3AODRvB4peSw8KNMvrM8/I=
|
||||||
github.com/bodgit/plumbing v1.2.0 h1:gg4haxoKphLjml+tgnecR4yLBV5zo4HAZGCtAh3xCzM=
|
github.com/andybalholm/brotli v1.1.1 h1:PR2pgnyFznKEugtsUo0xLdDop5SKXd5Qf5ysW+7XdTA=
|
||||||
github.com/bodgit/plumbing v1.2.0/go.mod h1:b9TeRi7Hvc6Y05rjm8VML3+47n4XTZPtQ/5ghqic2n8=
|
github.com/andybalholm/brotli v1.1.1/go.mod h1:05ib4cKhjx3OQYUY22hTVd34Bc8upXjOLL2rKwwZBoA=
|
||||||
github.com/bodgit/sevenzip v1.3.0 h1:1ljgELgtHqvgIp8W8kgeEGHIWP4ch3xGI8uOBZgLVKY=
|
github.com/bodgit/plumbing v1.3.0 h1:pf9Itz1JOQgn7vEOE7v7nlEfBykYqvUYioC61TwWCFU=
|
||||||
github.com/bodgit/sevenzip v1.3.0/go.mod h1:omwNcgZTEooWM8gA/IJ2Nk/+ZQ94+GsytRzOJJ8FBlM=
|
github.com/bodgit/plumbing v1.3.0/go.mod h1:JOTb4XiRu5xfnmdnDJo6GmSbSbtSyufrsyZFByMtKEs=
|
||||||
github.com/bodgit/windows v1.0.0 h1:rLQ/XjsleZvx4fR1tB/UxQrK+SJ2OFHzfPjLWWOhDIA=
|
github.com/bodgit/sevenzip v1.5.2 h1:acMIYRaqoHAdeu9LhEGGjL9UzBD4RNf9z7+kWDNignI=
|
||||||
github.com/bodgit/windows v1.0.0/go.mod h1:a6JLwrB4KrTR5hBpp8FI9/9W9jJfeQ2h4XDXU74ZCdM=
|
github.com/bodgit/sevenzip v1.5.2/go.mod h1:gTGzXA67Yko6/HLSD0iK4kWaWzPlPmLfDO73jTjSRqc=
|
||||||
|
github.com/bodgit/windows v1.0.1 h1:tF7K6KOluPYygXa3Z2594zxlkbKPAOvqr97etrGNIz4=
|
||||||
|
github.com/bodgit/windows v1.0.1/go.mod h1:a6JLwrB4KrTR5hBpp8FI9/9W9jJfeQ2h4XDXU74ZCdM=
|
||||||
github.com/census-instrumentation/opencensus-proto v0.2.1/go.mod h1:f6KPmirojxKA12rnyqOA5BBL4O983OfeGPqjHWSTneU=
|
github.com/census-instrumentation/opencensus-proto v0.2.1/go.mod h1:f6KPmirojxKA12rnyqOA5BBL4O983OfeGPqjHWSTneU=
|
||||||
github.com/chzyer/logex v1.1.10/go.mod h1:+Ywpsq7O8HXn0nuIou7OrIPyXbp3wmkHB+jjWRnGsAI=
|
github.com/chzyer/logex v1.1.10/go.mod h1:+Ywpsq7O8HXn0nuIou7OrIPyXbp3wmkHB+jjWRnGsAI=
|
||||||
github.com/chzyer/readline v0.0.0-20180603132655-2972be24d48e/go.mod h1:nSuG5e5PlCu98SY8svDHJxuZscDgtXS6KTTbou5AhLI=
|
github.com/chzyer/readline v0.0.0-20180603132655-2972be24d48e/go.mod h1:nSuG5e5PlCu98SY8svDHJxuZscDgtXS6KTTbou5AhLI=
|
||||||
github.com/chzyer/test v0.0.0-20180213035817-a1ea475d72b1/go.mod h1:Q3SI9o4m/ZMnBNeIyt5eFwwo7qiLfzFZmjNmxjkiQlU=
|
github.com/chzyer/test v0.0.0-20180213035817-a1ea475d72b1/go.mod h1:Q3SI9o4m/ZMnBNeIyt5eFwwo7qiLfzFZmjNmxjkiQlU=
|
||||||
github.com/client9/misspell v0.3.4/go.mod h1:qj6jICC3Q7zFZvVWo7KLAzC3yx5G7kyvSDkc90ppPyw=
|
github.com/client9/misspell v0.3.4/go.mod h1:qj6jICC3Q7zFZvVWo7KLAzC3yx5G7kyvSDkc90ppPyw=
|
||||||
github.com/connesc/cipherio v0.2.1 h1:FGtpTPMbKNNWByNrr9aEBtaJtXjqOzkIXNYJp6OEycw=
|
|
||||||
github.com/connesc/cipherio v0.2.1/go.mod h1:ukY0MWJDFnJEbXMQtOcn2VmTpRfzcTz4OoVrWGGJZcA=
|
|
||||||
github.com/davecgh/go-spew v1.1.0/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38=
|
github.com/davecgh/go-spew v1.1.0/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38=
|
||||||
github.com/davecgh/go-spew v1.1.1 h1:vj9j/u1bqnvCEfJOwUhtlOARqs3+rkHYY13jYWTU97c=
|
github.com/davecgh/go-spew v1.1.1 h1:vj9j/u1bqnvCEfJOwUhtlOARqs3+rkHYY13jYWTU97c=
|
||||||
github.com/davecgh/go-spew v1.1.1/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38=
|
github.com/davecgh/go-spew v1.1.1/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38=
|
||||||
github.com/disintegration/imaging v1.6.3-0.20201218193011-d40f48ce0f09 h1:MJFqtdxTq94XqUgg7DcGCaOIXrDTJE/tPHK66Jshguc=
|
github.com/disintegration/imaging v1.6.3-0.20201218193011-d40f48ce0f09 h1:MJFqtdxTq94XqUgg7DcGCaOIXrDTJE/tPHK66Jshguc=
|
||||||
github.com/disintegration/imaging v1.6.3-0.20201218193011-d40f48ce0f09/go.mod h1:44/5580QXChDfwIclfc/PCwrr44amcmDAg8hxG0Ewe4=
|
github.com/disintegration/imaging v1.6.3-0.20201218193011-d40f48ce0f09/go.mod h1:44/5580QXChDfwIclfc/PCwrr44amcmDAg8hxG0Ewe4=
|
||||||
github.com/dsnet/compress v0.0.1 h1:PlZu0n3Tuv04TzpfPbrnI0HW/YwodEXDS+oPKahKF0Q=
|
github.com/dsnet/compress v0.0.2-0.20230904184137-39efe44ab707 h1:2tV76y6Q9BB+NEBasnqvs7e49aEBFI8ejC89PSnWH+4=
|
||||||
github.com/dsnet/compress v0.0.1/go.mod h1:Aw8dCMJ7RioblQeTqt88akK31OvO8Dhf5JflhBbQEHo=
|
github.com/dsnet/compress v0.0.2-0.20230904184137-39efe44ab707/go.mod h1:qssHWj60/X5sZFNxpG4HBPDHVqxNm4DfnCKgrbZOT+s=
|
||||||
github.com/dsnet/golib v0.0.0-20171103203638-1ea166775780/go.mod h1:Lj+Z9rebOhdfkVLjJ8T6VcRQv3SXugXy999NBtR9aFY=
|
github.com/dsnet/golib v0.0.0-20171103203638-1ea166775780/go.mod h1:Lj+Z9rebOhdfkVLjJ8T6VcRQv3SXugXy999NBtR9aFY=
|
||||||
github.com/dustin/go-humanize v1.0.1 h1:GzkhY7T5VNhEkwH0PVJgjz+fX1rhBrR7pRT3mDkpeCY=
|
github.com/dustin/go-humanize v1.0.1 h1:GzkhY7T5VNhEkwH0PVJgjz+fX1rhBrR7pRT3mDkpeCY=
|
||||||
github.com/dustin/go-humanize v1.0.1/go.mod h1:Mu1zIs6XwVuF/gI1OepvI0qD18qycQx+mFykh5fBlto=
|
github.com/dustin/go-humanize v1.0.1/go.mod h1:Mu1zIs6XwVuF/gI1OepvI0qD18qycQx+mFykh5fBlto=
|
||||||
@ -58,13 +58,13 @@ github.com/golang/mock v1.1.1/go.mod h1:oTYuIxOrZwtPieC+H1uAHpcLFnEyAGVDL/k47Jfb
|
|||||||
github.com/golang/mock v1.2.0/go.mod h1:oTYuIxOrZwtPieC+H1uAHpcLFnEyAGVDL/k47Jfbm0A=
|
github.com/golang/mock v1.2.0/go.mod h1:oTYuIxOrZwtPieC+H1uAHpcLFnEyAGVDL/k47Jfbm0A=
|
||||||
github.com/golang/mock v1.3.1/go.mod h1:sBzyDLLjw3U8JLTeZvSv8jJB+tU5PVekmnlKIyFUx0Y=
|
github.com/golang/mock v1.3.1/go.mod h1:sBzyDLLjw3U8JLTeZvSv8jJB+tU5PVekmnlKIyFUx0Y=
|
||||||
github.com/golang/mock v1.4.0/go.mod h1:UOMv5ysSaYNkG+OFQykRIcU/QvvxJf3p21QfJ2Bt3cw=
|
github.com/golang/mock v1.4.0/go.mod h1:UOMv5ysSaYNkG+OFQykRIcU/QvvxJf3p21QfJ2Bt3cw=
|
||||||
github.com/golang/mock v1.4.3/go.mod h1:UOMv5ysSaYNkG+OFQykRIcU/QvvxJf3p21QfJ2Bt3cw=
|
|
||||||
github.com/golang/mock v1.6.0 h1:ErTB+efbowRARo13NNdxyJji2egdxLGQhRaY+DUumQc=
|
|
||||||
github.com/golang/mock v1.6.0/go.mod h1:p6yTPP+5HYm5mzsMV8JkE6ZKdX+/wYM6Hr+LicevLPs=
|
|
||||||
github.com/golang/protobuf v1.2.0/go.mod h1:6lQm79b+lXiMfvg/cZm0SGofjICqVBUtrP5yJMmIC1U=
|
github.com/golang/protobuf v1.2.0/go.mod h1:6lQm79b+lXiMfvg/cZm0SGofjICqVBUtrP5yJMmIC1U=
|
||||||
github.com/golang/protobuf v1.3.1/go.mod h1:6lQm79b+lXiMfvg/cZm0SGofjICqVBUtrP5yJMmIC1U=
|
github.com/golang/protobuf v1.3.1/go.mod h1:6lQm79b+lXiMfvg/cZm0SGofjICqVBUtrP5yJMmIC1U=
|
||||||
github.com/golang/protobuf v1.3.2/go.mod h1:6lQm79b+lXiMfvg/cZm0SGofjICqVBUtrP5yJMmIC1U=
|
github.com/golang/protobuf v1.3.2/go.mod h1:6lQm79b+lXiMfvg/cZm0SGofjICqVBUtrP5yJMmIC1U=
|
||||||
github.com/golang/protobuf v1.3.3/go.mod h1:vzj43D7+SQXF/4pzW/hwtAqwc6iTitCiVSaWz5lYuqw=
|
github.com/golang/protobuf v1.3.3/go.mod h1:vzj43D7+SQXF/4pzW/hwtAqwc6iTitCiVSaWz5lYuqw=
|
||||||
|
github.com/golang/protobuf v1.5.0/go.mod h1:FsONVRAS9T7sI+LIUmWTfcYkHO4aIWwzhcaSAoJOfIk=
|
||||||
|
github.com/golang/protobuf v1.5.2 h1:ROPKBNFfQgOUMifHyP+KYbvpjbdoFNs+aK7DXlji0Tw=
|
||||||
|
github.com/golang/protobuf v1.5.2/go.mod h1:XVQd3VNwM+JqD3oG2Ue2ip4fOMUkwXdXDdiuN0vRsmY=
|
||||||
github.com/golang/snappy v0.0.4 h1:yAGX7huGHXlcLOEtBnF4w7FQwA26wojNCwOYAEhLjQM=
|
github.com/golang/snappy v0.0.4 h1:yAGX7huGHXlcLOEtBnF4w7FQwA26wojNCwOYAEhLjQM=
|
||||||
github.com/golang/snappy v0.0.4/go.mod h1:/XxbfmMg8lxefKM7IXC3fBNl/7bRcc72aCRzEWrmP2Q=
|
github.com/golang/snappy v0.0.4/go.mod h1:/XxbfmMg8lxefKM7IXC3fBNl/7bRcc72aCRzEWrmP2Q=
|
||||||
github.com/google/btree v0.0.0-20180813153112-4030bb1f1f0c/go.mod h1:lNA+9X1NB3Zf8V7Ke586lFgjr2dZNuvo3lPJSGZ5JPQ=
|
github.com/google/btree v0.0.0-20180813153112-4030bb1f1f0c/go.mod h1:lNA+9X1NB3Zf8V7Ke586lFgjr2dZNuvo3lPJSGZ5JPQ=
|
||||||
@ -73,6 +73,10 @@ github.com/google/go-cmp v0.2.0/go.mod h1:oXzfMopK8JAjlY9xF4vHSVASa0yLyX7SntLO5a
|
|||||||
github.com/google/go-cmp v0.3.0/go.mod h1:8QqcDgzrUqlUb/G2PQTWiueGozuR1884gddMywk6iLU=
|
github.com/google/go-cmp v0.3.0/go.mod h1:8QqcDgzrUqlUb/G2PQTWiueGozuR1884gddMywk6iLU=
|
||||||
github.com/google/go-cmp v0.3.1/go.mod h1:8QqcDgzrUqlUb/G2PQTWiueGozuR1884gddMywk6iLU=
|
github.com/google/go-cmp v0.3.1/go.mod h1:8QqcDgzrUqlUb/G2PQTWiueGozuR1884gddMywk6iLU=
|
||||||
github.com/google/go-cmp v0.4.0/go.mod h1:v8dTdLbMG2kIc/vJvl+f65V22dbkXbowE6jgT/gNBxE=
|
github.com/google/go-cmp v0.4.0/go.mod h1:v8dTdLbMG2kIc/vJvl+f65V22dbkXbowE6jgT/gNBxE=
|
||||||
|
github.com/google/go-cmp v0.5.5/go.mod h1:v8dTdLbMG2kIc/vJvl+f65V22dbkXbowE6jgT/gNBxE=
|
||||||
|
github.com/google/go-cmp v0.6.0 h1:ofyhxvXcZhMsU5ulbFiLKl/XBFqE1GSq7atu8tAmTRI=
|
||||||
|
github.com/google/go-cmp v0.6.0/go.mod h1:17dUlkBOakJ0+DkrSSNjCkIjxS6bF9zb3elmeNGIjoY=
|
||||||
|
github.com/google/gofuzz v1.0.0/go.mod h1:dBl0BpW6vV/+mYPU4Po3pmUjxk6FQPldtuIdl/M65Eg=
|
||||||
github.com/google/martian v2.1.0+incompatible/go.mod h1:9I4somxYTbIHy5NJKHRl3wXiIaQGbYVAs8BPL6v8lEs=
|
github.com/google/martian v2.1.0+incompatible/go.mod h1:9I4somxYTbIHy5NJKHRl3wXiIaQGbYVAs8BPL6v8lEs=
|
||||||
github.com/google/pprof v0.0.0-20181206194817-3ea8567a2e57/go.mod h1:zfwlbNMJ+OItoe0UupaVj+oy1omPYYDuagoSzA8v9mc=
|
github.com/google/pprof v0.0.0-20181206194817-3ea8567a2e57/go.mod h1:zfwlbNMJ+OItoe0UupaVj+oy1omPYYDuagoSzA8v9mc=
|
||||||
github.com/google/pprof v0.0.0-20190515194954-54271f7e092f/go.mod h1:zfwlbNMJ+OItoe0UupaVj+oy1omPYYDuagoSzA8v9mc=
|
github.com/google/pprof v0.0.0-20190515194954-54271f7e092f/go.mod h1:zfwlbNMJ+OItoe0UupaVj+oy1omPYYDuagoSzA8v9mc=
|
||||||
@ -84,8 +88,9 @@ github.com/google/uuid v1.6.0 h1:NIvaJDMOsjHA8n1jAhLSgzrAzy1Hgr+hNrb57e+94F0=
|
|||||||
github.com/google/uuid v1.6.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo=
|
github.com/google/uuid v1.6.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo=
|
||||||
github.com/googleapis/gax-go/v2 v2.0.4/go.mod h1:0Wqv26UfaUD9n4G6kQubkQ+KchISgw+vpHVxEJEs9eg=
|
github.com/googleapis/gax-go/v2 v2.0.4/go.mod h1:0Wqv26UfaUD9n4G6kQubkQ+KchISgw+vpHVxEJEs9eg=
|
||||||
github.com/googleapis/gax-go/v2 v2.0.5/go.mod h1:DWXyrwAJ9X0FpwwEdw+IPEYBICEFu5mhpdKc/us6bOk=
|
github.com/googleapis/gax-go/v2 v2.0.5/go.mod h1:DWXyrwAJ9X0FpwwEdw+IPEYBICEFu5mhpdKc/us6bOk=
|
||||||
github.com/hashicorp/errwrap v1.0.0 h1:hLrqtEDnRye3+sgx6z4qVLNuviH3MR5aQ0ykNJa/UYA=
|
|
||||||
github.com/hashicorp/errwrap v1.0.0/go.mod h1:YH+1FKiLXxHSkmPseP+kNlulaMuP3n2brvKWEqk/Jc4=
|
github.com/hashicorp/errwrap v1.0.0/go.mod h1:YH+1FKiLXxHSkmPseP+kNlulaMuP3n2brvKWEqk/Jc4=
|
||||||
|
github.com/hashicorp/errwrap v1.1.0 h1:OxrOeh75EUXMY8TBjag2fzXGZ40LB6IKw45YeGUDY2I=
|
||||||
|
github.com/hashicorp/errwrap v1.1.0/go.mod h1:YH+1FKiLXxHSkmPseP+kNlulaMuP3n2brvKWEqk/Jc4=
|
||||||
github.com/hashicorp/go-multierror v1.1.1 h1:H5DkEtf6CXdFp0N0Em5UCwQpXMWke8IA0+lD48awMYo=
|
github.com/hashicorp/go-multierror v1.1.1 h1:H5DkEtf6CXdFp0N0Em5UCwQpXMWke8IA0+lD48awMYo=
|
||||||
github.com/hashicorp/go-multierror v1.1.1/go.mod h1:iw975J/qwKPdAO1clOe2L8331t/9/fmwbPZ6JB6eMoM=
|
github.com/hashicorp/go-multierror v1.1.1/go.mod h1:iw975J/qwKPdAO1clOe2L8331t/9/fmwbPZ6JB6eMoM=
|
||||||
github.com/hashicorp/golang-lru v0.5.0/go.mod h1:/m3WP610KZHVQ1SGc6re/UDhFvYD7pJ4Ao+sR/qLZy8=
|
github.com/hashicorp/golang-lru v0.5.0/go.mod h1:/m3WP610KZHVQ1SGc6re/UDhFvYD7pJ4Ao+sR/qLZy8=
|
||||||
@ -93,38 +98,43 @@ github.com/hashicorp/golang-lru v0.5.1/go.mod h1:/m3WP610KZHVQ1SGc6re/UDhFvYD7pJ
|
|||||||
github.com/hashicorp/golang-lru/v2 v2.0.7 h1:a+bsQ5rvGLjzHuww6tVxozPZFVghXaHOwFs4luLUK2k=
|
github.com/hashicorp/golang-lru/v2 v2.0.7 h1:a+bsQ5rvGLjzHuww6tVxozPZFVghXaHOwFs4luLUK2k=
|
||||||
github.com/hashicorp/golang-lru/v2 v2.0.7/go.mod h1:QeFd9opnmA6QUJc5vARoKUSoFhyfM2/ZepoAG6RGpeM=
|
github.com/hashicorp/golang-lru/v2 v2.0.7/go.mod h1:QeFd9opnmA6QUJc5vARoKUSoFhyfM2/ZepoAG6RGpeM=
|
||||||
github.com/ianlancetaylor/demangle v0.0.0-20181102032728-5e5cf60278f6/go.mod h1:aSSvb/t6k1mPoxDqO4vJh6VOCGPwU4O0C2/Eqndh1Sc=
|
github.com/ianlancetaylor/demangle v0.0.0-20181102032728-5e5cf60278f6/go.mod h1:aSSvb/t6k1mPoxDqO4vJh6VOCGPwU4O0C2/Eqndh1Sc=
|
||||||
|
github.com/json-iterator/go v1.1.12 h1:PV8peI4a0ysnczrg+LtxykD8LfKY9ML6u2jnxaEnrnM=
|
||||||
|
github.com/json-iterator/go v1.1.12/go.mod h1:e30LSqwooZae/UwlEbR2852Gd8hjQvJoHmT4TnhNGBo=
|
||||||
github.com/jstemmer/go-junit-report v0.0.0-20190106144839-af01ea7f8024/go.mod h1:6v2b51hI/fHJwM22ozAgKL4VKDeJcHhJFhtBdhmNjmU=
|
github.com/jstemmer/go-junit-report v0.0.0-20190106144839-af01ea7f8024/go.mod h1:6v2b51hI/fHJwM22ozAgKL4VKDeJcHhJFhtBdhmNjmU=
|
||||||
github.com/jstemmer/go-junit-report v0.9.1/go.mod h1:Brl9GWCQeLvo8nXZwPNNblvFj/XSXhF0NWZEnDohbsk=
|
github.com/jstemmer/go-junit-report v0.9.1/go.mod h1:Brl9GWCQeLvo8nXZwPNNblvFj/XSXhF0NWZEnDohbsk=
|
||||||
github.com/kisielk/gotool v1.0.0/go.mod h1:XhKaO+MFFWcvkIS/tQcRk01m1F5IRFswLeQ+oQHNcck=
|
github.com/kisielk/gotool v1.0.0/go.mod h1:XhKaO+MFFWcvkIS/tQcRk01m1F5IRFswLeQ+oQHNcck=
|
||||||
github.com/klauspost/compress v1.4.1/go.mod h1:RyIbtBH6LamlWaDj8nUwkbUhJ87Yi3uG0guNDohfE1A=
|
github.com/klauspost/compress v1.4.1/go.mod h1:RyIbtBH6LamlWaDj8nUwkbUhJ87Yi3uG0guNDohfE1A=
|
||||||
github.com/klauspost/compress v1.15.9 h1:wKRjX6JRtDdrE9qwa4b/Cip7ACOshUI4smpCQanqjSY=
|
github.com/klauspost/compress v1.17.11 h1:In6xLpyWOi1+C7tXUUWv2ot1QvBjxevKAaI6IXrJmUc=
|
||||||
github.com/klauspost/compress v1.15.9/go.mod h1:PhcZ0MbTNciWF3rruxRgKxI5NkcHHrHUDtV4Yw2GlzU=
|
github.com/klauspost/compress v1.17.11/go.mod h1:pMDklpSncoRMuLFrf1W9Ss9KT+0rH90U12bZKk7uwG0=
|
||||||
github.com/klauspost/cpuid v1.2.0/go.mod h1:Pj4uuM528wm8OyEC2QMXAi2YiTZ96dNQPGgoMS4s3ek=
|
github.com/klauspost/cpuid v1.2.0/go.mod h1:Pj4uuM528wm8OyEC2QMXAi2YiTZ96dNQPGgoMS4s3ek=
|
||||||
github.com/klauspost/pgzip v1.2.5 h1:qnWYvvKqedOF2ulHpMG72XQol4ILEJ8k2wwRl/Km8oE=
|
github.com/klauspost/pgzip v1.2.6 h1:8RXeL5crjEUFnR2/Sn6GJNWtSQ3Dk8pq4CL3jvdDyjU=
|
||||||
github.com/klauspost/pgzip v1.2.5/go.mod h1:Ch1tH69qFZu15pkjo5kYi6mth2Zzwzt50oCQKQE9RUs=
|
github.com/klauspost/pgzip v1.2.6/go.mod h1:Ch1tH69qFZu15pkjo5kYi6mth2Zzwzt50oCQKQE9RUs=
|
||||||
github.com/kr/pretty v0.1.0 h1:L/CwN0zerZDmRFUapSPitk6f+Q3+0za1rQkzVuMiMFI=
|
|
||||||
github.com/kr/pretty v0.1.0/go.mod h1:dAy3ld7l9f0ibDNOQOHHMYYIIbhfbHSm3C4ZsoJORNo=
|
github.com/kr/pretty v0.1.0/go.mod h1:dAy3ld7l9f0ibDNOQOHHMYYIIbhfbHSm3C4ZsoJORNo=
|
||||||
|
github.com/kr/pretty v0.2.1 h1:Fmg33tUaq4/8ym9TJN1x7sLJnHVwhP33CNkpYV/7rwI=
|
||||||
|
github.com/kr/pretty v0.2.1/go.mod h1:ipq/a2n7PKx3OHsz4KJII5eveXtPO4qwEXGdVfWzfnI=
|
||||||
github.com/kr/pty v1.1.1/go.mod h1:pFQYn66WHrOpPYNljwOMqo10TkYh1fy3cYio2l3bCsQ=
|
github.com/kr/pty v1.1.1/go.mod h1:pFQYn66WHrOpPYNljwOMqo10TkYh1fy3cYio2l3bCsQ=
|
||||||
github.com/kr/text v0.1.0 h1:45sCR5RtlFHMR4UwH9sdQ5TC8v0qDQCHnXt+kaKSTVE=
|
github.com/kr/text v0.1.0 h1:45sCR5RtlFHMR4UwH9sdQ5TC8v0qDQCHnXt+kaKSTVE=
|
||||||
github.com/kr/text v0.1.0/go.mod h1:4Jbv+DJW3UT/LiOwJeYQe1efqtUx/iVham/4vfdArNI=
|
github.com/kr/text v0.1.0/go.mod h1:4Jbv+DJW3UT/LiOwJeYQe1efqtUx/iVham/4vfdArNI=
|
||||||
github.com/lordwelch/text v0.0.0-20240505231825-4893f344170f h1:RMKTfrT4gjJfmB/aWuvCcFxUSvWAJfOAc5khGL6ASjk=
|
|
||||||
github.com/lordwelch/text v0.0.0-20240505231825-4893f344170f/go.mod h1:18ZOQIKpY8NJVqYksKHtTdi31H5itFRjB5/qKTNYzSU=
|
|
||||||
github.com/mattn/go-isatty v0.0.20 h1:xfD0iDuEKnDkl03q4limB+vH+GxLEtL/jb4xVJSWWEY=
|
github.com/mattn/go-isatty v0.0.20 h1:xfD0iDuEKnDkl03q4limB+vH+GxLEtL/jb4xVJSWWEY=
|
||||||
github.com/mattn/go-isatty v0.0.20/go.mod h1:W+V8PltTTMOvKvAeJH7IuucS94S2C6jfK/D7dTCTo3Y=
|
github.com/mattn/go-isatty v0.0.20/go.mod h1:W+V8PltTTMOvKvAeJH7IuucS94S2C6jfK/D7dTCTo3Y=
|
||||||
github.com/mattn/go-sqlite3 v1.14.22 h1:2gZY6PC6kBnID23Tichd1K+Z0oS6nE/XwU+Vz/5o4kU=
|
github.com/mattn/go-sqlite3 v1.14.24 h1:tpSp2G2KyMnnQu99ngJ47EIkWVmliIizyZBfPrBWDRM=
|
||||||
github.com/mattn/go-sqlite3 v1.14.22/go.mod h1:Uh1q+B4BYcTPb+yiD3kU8Ct7aC0hY9fxUwlHK0RXw+Y=
|
github.com/mattn/go-sqlite3 v1.14.24/go.mod h1:Uh1q+B4BYcTPb+yiD3kU8Ct7aC0hY9fxUwlHK0RXw+Y=
|
||||||
github.com/mholt/archiver/v4 v4.0.0-alpha.8 h1:tRGQuDVPh66WCOelqe6LIGh0gwmfwxUrSSDunscGsRM=
|
github.com/mholt/archiver/v4 v4.0.0-alpha.9 h1:EZgAsW6DsuawxDgTtIdjCUBa2TQ6AOe9pnCidofSRtE=
|
||||||
github.com/mholt/archiver/v4 v4.0.0-alpha.8/go.mod h1:5f7FUYGXdJWUjESffJaYR4R60VhnHxb2X3T1teMyv5A=
|
github.com/mholt/archiver/v4 v4.0.0-alpha.9/go.mod h1:5D3uct315OMkMRXKwEuMB+wQi/2m5NQngKDmApqwVlo=
|
||||||
github.com/ncruces/go-sqlite3 v0.18.1 h1:iN8IMZV5EMxpH88NUac9vId23eTKNFUhP7jgY0EBbNc=
|
github.com/modern-go/concurrent v0.0.0-20180228061459-e0a39a4cb421 h1:ZqeYNhU3OHLH3mGKHDcjJRFFRrJa6eAM5H+CtDdOsPc=
|
||||||
github.com/ncruces/go-sqlite3 v0.18.1/go.mod h1:eEOyZnW1dGTJ+zDpMuzfYamEUBtdFz5zeYhqLBtHxvM=
|
github.com/modern-go/concurrent v0.0.0-20180228061459-e0a39a4cb421/go.mod h1:6dJC0mAP4ikYIbvyc7fijjWJddQyLn8Ig3JB5CqoB9Q=
|
||||||
|
github.com/modern-go/reflect2 v1.0.2 h1:xBagoLtFs94CBntxluKeaWgTMpvLxC4ur3nMaC9Gz0M=
|
||||||
|
github.com/modern-go/reflect2 v1.0.2/go.mod h1:yWuevngMOJpCy52FWWMvUC8ws7m/LJsjYzDa0/r8luk=
|
||||||
|
github.com/ncruces/go-sqlite3 v0.26.0 h1:dY6ASfuhSEbtSge6kJwjyJVC7bXCpgEVOycmdboKJek=
|
||||||
|
github.com/ncruces/go-sqlite3 v0.26.0/go.mod h1:46HIzeCQQ+aNleAxCli+vpA2tfh7ttSnw24kQahBc1o=
|
||||||
github.com/ncruces/go-strftime v0.1.9 h1:bY0MQC28UADQmHmaF5dgpLmImcShSi2kHU9XLdhx/f4=
|
github.com/ncruces/go-strftime v0.1.9 h1:bY0MQC28UADQmHmaF5dgpLmImcShSi2kHU9XLdhx/f4=
|
||||||
github.com/ncruces/go-strftime v0.1.9/go.mod h1:Fwc5htZGVVkseilnfgOVb9mKy6w1naJmn9CehxcKcls=
|
github.com/ncruces/go-strftime v0.1.9/go.mod h1:Fwc5htZGVVkseilnfgOVb9mKy6w1naJmn9CehxcKcls=
|
||||||
github.com/ncruces/julianday v1.0.0 h1:fH0OKwa7NWvniGQtxdJRxAgkBMolni2BjDHaWTxqt7M=
|
github.com/ncruces/julianday v1.0.0 h1:fH0OKwa7NWvniGQtxdJRxAgkBMolni2BjDHaWTxqt7M=
|
||||||
github.com/ncruces/julianday v1.0.0/go.mod h1:Dusn2KvZrrovOMJuOt0TNXL6tB7U2E8kvza5fFc9G7g=
|
github.com/ncruces/julianday v1.0.0/go.mod h1:Dusn2KvZrrovOMJuOt0TNXL6tB7U2E8kvza5fFc9G7g=
|
||||||
github.com/nwaples/rardecode/v2 v2.0.0-beta.2 h1:e3mzJFJs4k83GXBEiTaQ5HgSc/kOK8q0rDaRO0MPaOk=
|
github.com/nwaples/rardecode/v2 v2.0.0-beta.4 h1:sdiJxQdPjECn2lh9nLFFhgLCf+0ulDU5rODbtERTlUY=
|
||||||
github.com/nwaples/rardecode/v2 v2.0.0-beta.2/go.mod h1:yntwv/HfMc/Hbvtq9I19D1n58te3h6KsqCf3GxyfBGY=
|
github.com/nwaples/rardecode/v2 v2.0.0-beta.4/go.mod h1:yntwv/HfMc/Hbvtq9I19D1n58te3h6KsqCf3GxyfBGY=
|
||||||
github.com/pierrec/lz4/v4 v4.1.15 h1:MO0/ucJhngq7299dKLwIMtgTfbkoSPF6AoMYDd8Q4q0=
|
github.com/pierrec/lz4/v4 v4.1.21 h1:yOVMLb6qSIDP67pl/5F7RepeKYu/VmTyEXvuMI5d9mQ=
|
||||||
github.com/pierrec/lz4/v4 v4.1.15/go.mod h1:gZWDp/Ze/IJXGXf23ltt2EXimqmTUXEy0GFuRQyBid4=
|
github.com/pierrec/lz4/v4 v4.1.21/go.mod h1:gZWDp/Ze/IJXGXf23ltt2EXimqmTUXEy0GFuRQyBid4=
|
||||||
github.com/pmezard/go-difflib v1.0.0 h1:4DBwDE0NGyQoBHbLQYPwSUPoCMWR5BEzIk/f1lZbAQM=
|
github.com/pmezard/go-difflib v1.0.0 h1:4DBwDE0NGyQoBHbLQYPwSUPoCMWR5BEzIk/f1lZbAQM=
|
||||||
github.com/pmezard/go-difflib v1.0.0/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4=
|
github.com/pmezard/go-difflib v1.0.0/go.mod h1:iKH77koFhYxTK1pcRnkKkqfTogsbg7gZNVY4sRDYZ/4=
|
||||||
github.com/prometheus/client_model v0.0.0-20190812154241-14fe0d1b01d4/go.mod h1:xMI15A0UPsDsEKsMN9yxemIoYk6Tm2C1GtYGdfGttqA=
|
github.com/prometheus/client_model v0.0.0-20190812154241-14fe0d1b01d4/go.mod h1:xMI15A0UPsDsEKsMN9yxemIoYk6Tm2C1GtYGdfGttqA=
|
||||||
@ -132,33 +142,43 @@ github.com/remyoudompheng/bigfft v0.0.0-20230129092748-24d4a6f8daec h1:W09IVJc94
|
|||||||
github.com/remyoudompheng/bigfft v0.0.0-20230129092748-24d4a6f8daec/go.mod h1:qqbHyh8v60DhA7CoWK5oRCqLrMHRGoxYCSS9EjAz6Eo=
|
github.com/remyoudompheng/bigfft v0.0.0-20230129092748-24d4a6f8daec/go.mod h1:qqbHyh8v60DhA7CoWK5oRCqLrMHRGoxYCSS9EjAz6Eo=
|
||||||
github.com/rogpeppe/go-internal v1.3.0/go.mod h1:M8bDsm7K2OlrFYOpmOWEs/qY81heoFRclV5y23lUDJ4=
|
github.com/rogpeppe/go-internal v1.3.0/go.mod h1:M8bDsm7K2OlrFYOpmOWEs/qY81heoFRclV5y23lUDJ4=
|
||||||
github.com/rwcarlsen/goexif v0.0.0-20190401172101-9e8deecbddbd/go.mod h1:hPqNNc0+uJM6H+SuU8sEs5K5IQeKccPqeSjfgcKGgPk=
|
github.com/rwcarlsen/goexif v0.0.0-20190401172101-9e8deecbddbd/go.mod h1:hPqNNc0+uJM6H+SuU8sEs5K5IQeKccPqeSjfgcKGgPk=
|
||||||
|
github.com/sorairolake/lzip-go v0.3.5 h1:ms5Xri9o1JBIWvOFAorYtUNik6HI3HgBTkISiqu0Cwg=
|
||||||
|
github.com/sorairolake/lzip-go v0.3.5/go.mod h1:N0KYq5iWrMXI0ZEXKXaS9hCyOjZUQdBDEIbXfoUwbdk=
|
||||||
github.com/stretchr/objx v0.1.0/go.mod h1:HFkY916IF+rwdDfMAkV7OtwuqBVzrE8GR6GFx+wExME=
|
github.com/stretchr/objx v0.1.0/go.mod h1:HFkY916IF+rwdDfMAkV7OtwuqBVzrE8GR6GFx+wExME=
|
||||||
|
github.com/stretchr/objx v0.4.0/go.mod h1:YvHI0jy2hoMjB+UWwv71VJQ9isScKT/TqJzVSSt89Yw=
|
||||||
|
github.com/stretchr/objx v0.5.0/go.mod h1:Yh+to48EsGEfYuaHDzXPcE3xhTkx73EhmCGUpEOglKo=
|
||||||
|
github.com/stretchr/testify v1.3.0/go.mod h1:M5WIy9Dh21IEIfnGCwXGc5bZfKNJtfHm1UVUgZn+9EI=
|
||||||
github.com/stretchr/testify v1.4.0/go.mod h1:j7eGeouHqKxXV5pUuKE4zz7dFj8WfuZ+81PSLYec5m4=
|
github.com/stretchr/testify v1.4.0/go.mod h1:j7eGeouHqKxXV5pUuKE4zz7dFj8WfuZ+81PSLYec5m4=
|
||||||
github.com/stretchr/testify v1.7.1/go.mod h1:6Fq8oRcR53rry900zMqJjRRixrwX3KX962/h/Wwjteg=
|
github.com/stretchr/testify v1.7.1/go.mod h1:6Fq8oRcR53rry900zMqJjRRixrwX3KX962/h/Wwjteg=
|
||||||
github.com/stretchr/testify v1.8.4 h1:CcVxjf3Q8PM0mHUKJCdn+eZZtm5yQwehR5yeSVQQcUk=
|
github.com/stretchr/testify v1.8.0/go.mod h1:yNjHg4UonilssWZ8iaSj1OCr/vHnekPRkoO+kdMU+MU=
|
||||||
github.com/stretchr/testify v1.8.4/go.mod h1:sz/lmYIOXD/1dqDmKjjqLyZ2RngseejIcXlSw2iwfAo=
|
github.com/stretchr/testify v1.8.1/go.mod h1:w2LPCIKwWwSfY2zedu0+kehJoqGctiVI29o6fzry7u4=
|
||||||
github.com/tetratelabs/wazero v1.8.0 h1:iEKu0d4c2Pd+QSRieYbnQC9yiFlMS9D+Jr0LsRmcF4g=
|
github.com/stretchr/testify v1.10.0 h1:Xv5erBjTwe/5IxqUQTdXv5kgmIvbHo3QQyRwhJsOfJA=
|
||||||
github.com/tetratelabs/wazero v1.8.0/go.mod h1:yAI0XTsMBhREkM/YDAK/zNou3GoiAce1P6+rp/wQhjs=
|
github.com/stretchr/testify v1.10.0/go.mod h1:r2ic/lqez/lEtzL7wO/rwa5dbSLXVDPFyf8C91i36aY=
|
||||||
|
github.com/tetratelabs/wazero v1.9.0 h1:IcZ56OuxrtaEz8UYNRHBrUa9bYeX9oVY93KspZZBf/I=
|
||||||
|
github.com/tetratelabs/wazero v1.9.0/go.mod h1:TSbcXCfFP0L2FGkRPxHphadXPjo1T6W+CseNNY7EkjM=
|
||||||
github.com/therootcompany/xz v1.0.1 h1:CmOtsn1CbtmyYiusbfmhmkpAAETj0wBIH6kCYaX+xzw=
|
github.com/therootcompany/xz v1.0.1 h1:CmOtsn1CbtmyYiusbfmhmkpAAETj0wBIH6kCYaX+xzw=
|
||||||
github.com/therootcompany/xz v1.0.1/go.mod h1:3K3UH1yCKgBneZYhuQUvJ9HPD19UEXEI0BWbMn8qNMY=
|
github.com/therootcompany/xz v1.0.1/go.mod h1:3K3UH1yCKgBneZYhuQUvJ9HPD19UEXEI0BWbMn8qNMY=
|
||||||
github.com/ulikunitz/xz v0.5.6/go.mod h1:2bypXElzHzzJZwzH67Y6wb67pO62Rzfn7BSiF4ABRW8=
|
github.com/ulikunitz/xz v0.5.8/go.mod h1:nbz6k7qbPmH4IRqmfOplQw/tblSgqTqBwxkY0oWt/14=
|
||||||
github.com/ulikunitz/xz v0.5.10 h1:t92gobL9l3HE202wg3rlk19F6X+JOxl9BBrCCMYEYd8=
|
github.com/ulikunitz/xz v0.5.12 h1:37Nm15o69RwBkXM0J6A5OlE67RZTfzUxTj8fB3dfcsc=
|
||||||
github.com/ulikunitz/xz v0.5.10/go.mod h1:nbz6k7qbPmH4IRqmfOplQw/tblSgqTqBwxkY0oWt/14=
|
github.com/ulikunitz/xz v0.5.12/go.mod h1:nbz6k7qbPmH4IRqmfOplQw/tblSgqTqBwxkY0oWt/14=
|
||||||
github.com/vmihailenco/msgpack/v5 v5.4.1 h1:cQriyiUvjTwOHg8QZaPihLWeRAAVoCpE00IUPn0Bjt8=
|
github.com/vmihailenco/msgpack v4.0.4+incompatible h1:dSLoQfGFAo3F6OoNhwUmLwVgaUXK79GlxNBwueZn0xI=
|
||||||
github.com/vmihailenco/msgpack/v5 v5.4.1/go.mod h1:GaZTsDaehaPpQVyxrf5mtQlH+pc21PIudVV/E3rRQok=
|
github.com/vmihailenco/msgpack v4.0.4+incompatible/go.mod h1:fy3FlTQTDXWkZ7Bh6AcGMlsjHatGryHQYUTf1ShIgkk=
|
||||||
github.com/vmihailenco/tagparser/v2 v2.0.0 h1:y09buUbR+b5aycVFQs/g70pqKVZNBmxwAhO7/IwNM9g=
|
github.com/xyproto/randomstring v1.0.5 h1:YtlWPoRdgMu3NZtP45drfy1GKoojuR7hmRcnhZqKjWU=
|
||||||
github.com/vmihailenco/tagparser/v2 v2.0.0/go.mod h1:Wri+At7QHww0WTrCBeu4J6bNtoV6mEfg5OIWRZA9qds=
|
github.com/xyproto/randomstring v1.0.5/go.mod h1:rgmS5DeNXLivK7YprL0pY+lTuhNQW3iGxZ18UQApw/E=
|
||||||
github.com/yuin/goldmark v1.3.5/go.mod h1:mwnBkeHKe2W/ZEtQ+71ViKU8L12m81fl3OWwC1Zlc8k=
|
github.com/yuin/goldmark v1.4.13/go.mod h1:6yULJ656Px+3vBD8DxQVa3kxgyrAnzto9xy5taEt/CY=
|
||||||
|
go.etcd.io/bbolt v1.4.0 h1:TU77id3TnN/zKr7CO/uk+fBCwF2jGcMuw2B/FMAzYIk=
|
||||||
|
go.etcd.io/bbolt v1.4.0/go.mod h1:AsD+OCi/qPN1giOX1aiLAha3o1U8rAz65bvN4j0sRuk=
|
||||||
go.opencensus.io v0.21.0/go.mod h1:mSImk1erAIZhrmZN+AvHh14ztQfjbGwt4TtuofqLduU=
|
go.opencensus.io v0.21.0/go.mod h1:mSImk1erAIZhrmZN+AvHh14ztQfjbGwt4TtuofqLduU=
|
||||||
go.opencensus.io v0.22.0/go.mod h1:+kGneAE2xo2IficOXnaByMWTGM9T73dGwxeWcUqIpI8=
|
go.opencensus.io v0.22.0/go.mod h1:+kGneAE2xo2IficOXnaByMWTGM9T73dGwxeWcUqIpI8=
|
||||||
go.opencensus.io v0.22.2/go.mod h1:yxeiOL68Rb0Xd1ddK5vPZ/oVn4vY4Ynel7k9FzqtOIw=
|
go.opencensus.io v0.22.2/go.mod h1:yxeiOL68Rb0Xd1ddK5vPZ/oVn4vY4Ynel7k9FzqtOIw=
|
||||||
go.opencensus.io v0.22.3/go.mod h1:yxeiOL68Rb0Xd1ddK5vPZ/oVn4vY4Ynel7k9FzqtOIw=
|
go.opencensus.io v0.22.3/go.mod h1:yxeiOL68Rb0Xd1ddK5vPZ/oVn4vY4Ynel7k9FzqtOIw=
|
||||||
go4.org v0.0.0-20200411211856-f5505b9728dd h1:BNJlw5kRTzdmyfh5U8F93HA2OwkP7ZGwA51eJ/0wKOU=
|
go4.org v0.0.0-20230225012048-214862532bf5 h1:nifaUDeh+rPaBCMPMQHZmvJf+QdpLFnuQPwx+LxVmtc=
|
||||||
go4.org v0.0.0-20200411211856-f5505b9728dd/go.mod h1:CIiUVy99QCPfoE13bO4EZaz5GZMZXMSBGhxRdsvzbkg=
|
go4.org v0.0.0-20230225012048-214862532bf5/go.mod h1:F57wTi5Lrj6WLyswp5EYV1ncrEbFGHD4hhz6S1ZYeaU=
|
||||||
golang.org/x/crypto v0.0.0-20190308221718-c2843e01d9a2/go.mod h1:djNgcEr1/C05ACkg1iLfiJU5Ep61QUkGW8qpdssI0+w=
|
golang.org/x/crypto v0.0.0-20190308221718-c2843e01d9a2/go.mod h1:djNgcEr1/C05ACkg1iLfiJU5Ep61QUkGW8qpdssI0+w=
|
||||||
golang.org/x/crypto v0.0.0-20190510104115-cbcb75029529/go.mod h1:yigFU9vqHzYiE8UmvKecakEJjdnWj3jj499lnFckfCI=
|
golang.org/x/crypto v0.0.0-20190510104115-cbcb75029529/go.mod h1:yigFU9vqHzYiE8UmvKecakEJjdnWj3jj499lnFckfCI=
|
||||||
golang.org/x/crypto v0.0.0-20190605123033-f99c8df09eb5/go.mod h1:yigFU9vqHzYiE8UmvKecakEJjdnWj3jj499lnFckfCI=
|
golang.org/x/crypto v0.0.0-20190605123033-f99c8df09eb5/go.mod h1:yigFU9vqHzYiE8UmvKecakEJjdnWj3jj499lnFckfCI=
|
||||||
golang.org/x/crypto v0.0.0-20191011191535-87dc89f01550/go.mod h1:yigFU9vqHzYiE8UmvKecakEJjdnWj3jj499lnFckfCI=
|
golang.org/x/crypto v0.0.0-20191011191535-87dc89f01550/go.mod h1:yigFU9vqHzYiE8UmvKecakEJjdnWj3jj499lnFckfCI=
|
||||||
|
golang.org/x/crypto v0.0.0-20210921155107-089bfa567519/go.mod h1:GvvjBRRGRdwPK5ydBHafDWAxML/pGHZbMvKqRZ5+Abc=
|
||||||
golang.org/x/exp v0.0.0-20190121172915-509febef88a4/go.mod h1:CJ0aWSM057203Lf6IL+f9T1iT9GByDxfZKAQTCR3kQA=
|
golang.org/x/exp v0.0.0-20190121172915-509febef88a4/go.mod h1:CJ0aWSM057203Lf6IL+f9T1iT9GByDxfZKAQTCR3kQA=
|
||||||
golang.org/x/exp v0.0.0-20190306152737-a1d7652674e8/go.mod h1:CJ0aWSM057203Lf6IL+f9T1iT9GByDxfZKAQTCR3kQA=
|
golang.org/x/exp v0.0.0-20190306152737-a1d7652674e8/go.mod h1:CJ0aWSM057203Lf6IL+f9T1iT9GByDxfZKAQTCR3kQA=
|
||||||
golang.org/x/exp v0.0.0-20190510132918-efd6b22b2522/go.mod h1:ZjyILWgesfNpC6sMxTJOJm9Kp84zZh5NQWvqDGG3Qr8=
|
golang.org/x/exp v0.0.0-20190510132918-efd6b22b2522/go.mod h1:ZjyILWgesfNpC6sMxTJOJm9Kp84zZh5NQWvqDGG3Qr8=
|
||||||
@ -167,13 +187,13 @@ golang.org/x/exp v0.0.0-20191030013958-a1ab85dbe136/go.mod h1:JXzH8nQsPlswgeRAPE
|
|||||||
golang.org/x/exp v0.0.0-20191129062945-2f5052295587/go.mod h1:2RIsYlXP63K8oxa1u096TMicItID8zy7Y6sNkU49FU4=
|
golang.org/x/exp v0.0.0-20191129062945-2f5052295587/go.mod h1:2RIsYlXP63K8oxa1u096TMicItID8zy7Y6sNkU49FU4=
|
||||||
golang.org/x/exp v0.0.0-20191227195350-da58074b4299/go.mod h1:2RIsYlXP63K8oxa1u096TMicItID8zy7Y6sNkU49FU4=
|
golang.org/x/exp v0.0.0-20191227195350-da58074b4299/go.mod h1:2RIsYlXP63K8oxa1u096TMicItID8zy7Y6sNkU49FU4=
|
||||||
golang.org/x/exp v0.0.0-20200207192155-f17229e696bd/go.mod h1:J/WKrq2StrnmMY6+EHIKF9dgMWnmCNThgcyBT1FY9mM=
|
golang.org/x/exp v0.0.0-20200207192155-f17229e696bd/go.mod h1:J/WKrq2StrnmMY6+EHIKF9dgMWnmCNThgcyBT1FY9mM=
|
||||||
golang.org/x/exp v0.0.0-20240808152545-0cdaa3abc0fa h1:ELnwvuAXPNtPk1TJRuGkI9fDTwym6AYBu0qzT8AcHdI=
|
golang.org/x/exp v0.0.0-20250218142911-aa4b98e5adaa h1:t2QcU6V556bFjYgu4L6C+6VrCPyJZ+eyRsABUPs1mz4=
|
||||||
golang.org/x/exp v0.0.0-20240808152545-0cdaa3abc0fa/go.mod h1:akd2r19cwCdwSwWeIdzYQGa/EZZyqcOdwWiwj5L5eKQ=
|
golang.org/x/exp v0.0.0-20250218142911-aa4b98e5adaa/go.mod h1:BHOTPb3L19zxehTsLoJXVaTktb06DFgmdW6Wb9s8jqk=
|
||||||
golang.org/x/image v0.0.0-20190227222117-0694c2d4d067/go.mod h1:kZ7UVZpmo3dzQBMxlp+ypCbDeSB+sBbTgSJuh5dn5js=
|
golang.org/x/image v0.0.0-20190227222117-0694c2d4d067/go.mod h1:kZ7UVZpmo3dzQBMxlp+ypCbDeSB+sBbTgSJuh5dn5js=
|
||||||
golang.org/x/image v0.0.0-20190802002840-cff245a6509b/go.mod h1:FeLwcggjj3mMvU+oOTbSwawSJRM1uh48EjtB4UJZlP0=
|
golang.org/x/image v0.0.0-20190802002840-cff245a6509b/go.mod h1:FeLwcggjj3mMvU+oOTbSwawSJRM1uh48EjtB4UJZlP0=
|
||||||
golang.org/x/image v0.0.0-20191009234506-e7c1f5e7dbb8/go.mod h1:FeLwcggjj3mMvU+oOTbSwawSJRM1uh48EjtB4UJZlP0=
|
golang.org/x/image v0.0.0-20191009234506-e7c1f5e7dbb8/go.mod h1:FeLwcggjj3mMvU+oOTbSwawSJRM1uh48EjtB4UJZlP0=
|
||||||
golang.org/x/image v0.19.0 h1:D9FX4QWkLfkeqaC62SonffIIuYdOk/UE2XKUBgRIBIQ=
|
golang.org/x/image v0.25.0 h1:Y6uW6rH1y5y/LK1J8BPWZtr6yZ7hrsy6hFrXjgsc2fQ=
|
||||||
golang.org/x/image v0.19.0/go.mod h1:y0zrRqlQRWQ5PXaYCOMLTW2fpsxZ8Qh9I/ohnInJEys=
|
golang.org/x/image v0.25.0/go.mod h1:tCAmOEGthTtkalusGp1g3xa2gke8J6c2N565dTyl9Rs=
|
||||||
golang.org/x/lint v0.0.0-20181026193005-c67002cb31c3/go.mod h1:UVdnD1Gm6xHRNCYTkRU2/jEulfH38KcIWyp/GAMgvoE=
|
golang.org/x/lint v0.0.0-20181026193005-c67002cb31c3/go.mod h1:UVdnD1Gm6xHRNCYTkRU2/jEulfH38KcIWyp/GAMgvoE=
|
||||||
golang.org/x/lint v0.0.0-20190227174305-5b3e6a55c961/go.mod h1:wehouNa3lNwaWXcvxsM5YxQ5yQlVC4a0KAMCusXpPoU=
|
golang.org/x/lint v0.0.0-20190227174305-5b3e6a55c961/go.mod h1:wehouNa3lNwaWXcvxsM5YxQ5yQlVC4a0KAMCusXpPoU=
|
||||||
golang.org/x/lint v0.0.0-20190301231843-5614ed5bae6f/go.mod h1:UVdnD1Gm6xHRNCYTkRU2/jEulfH38KcIWyp/GAMgvoE=
|
golang.org/x/lint v0.0.0-20190301231843-5614ed5bae6f/go.mod h1:UVdnD1Gm6xHRNCYTkRU2/jEulfH38KcIWyp/GAMgvoE=
|
||||||
@ -189,9 +209,9 @@ golang.org/x/mod v0.0.0-20190513183733-4bf6d317e70e/go.mod h1:mXi4GBBbnImb6dmsKG
|
|||||||
golang.org/x/mod v0.1.0/go.mod h1:0QHyrYULN0/3qlju5TqG8bIK38QM8yzMo5ekMj3DlcY=
|
golang.org/x/mod v0.1.0/go.mod h1:0QHyrYULN0/3qlju5TqG8bIK38QM8yzMo5ekMj3DlcY=
|
||||||
golang.org/x/mod v0.1.1-0.20191105210325-c90efee705ee/go.mod h1:QqPTAvyqsEbceGzBzNggFXnrqF1CaUcvgkdR5Ot7KZg=
|
golang.org/x/mod v0.1.1-0.20191105210325-c90efee705ee/go.mod h1:QqPTAvyqsEbceGzBzNggFXnrqF1CaUcvgkdR5Ot7KZg=
|
||||||
golang.org/x/mod v0.2.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA=
|
golang.org/x/mod v0.2.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA=
|
||||||
golang.org/x/mod v0.4.2/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA=
|
golang.org/x/mod v0.6.0-dev.0.20220419223038-86c51ed26bb4/go.mod h1:jJ57K6gSWd91VN4djpZkiMVwK6gcyfeH4XE8wZrZaV4=
|
||||||
golang.org/x/mod v0.20.0 h1:utOm6MM3R3dnawAiJgn0y+xvuYRsm1RKM/4giyfDgV0=
|
golang.org/x/mod v0.23.0 h1:Zb7khfcRGKk+kqfxFaP5tZqCnDZMjC5VtUBs87Hr6QM=
|
||||||
golang.org/x/mod v0.20.0/go.mod h1:hTbmBsO62+eylJbnUtE2MGJUyE7QWk4xUqPFrRgJ+7c=
|
golang.org/x/mod v0.23.0/go.mod h1:6SkKJ3Xj0I0BrPOZoBy3bdMptDDU9oJrpohJ3eWZ1fY=
|
||||||
golang.org/x/net v0.0.0-20180724234803-3673e40ba225/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4=
|
golang.org/x/net v0.0.0-20180724234803-3673e40ba225/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4=
|
||||||
golang.org/x/net v0.0.0-20180826012351-8a410e7b638d/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4=
|
golang.org/x/net v0.0.0-20180826012351-8a410e7b638d/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4=
|
||||||
golang.org/x/net v0.0.0-20190108225652-1e06a53dbb7e/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4=
|
golang.org/x/net v0.0.0-20190108225652-1e06a53dbb7e/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4=
|
||||||
@ -205,8 +225,9 @@ golang.org/x/net v0.0.0-20190620200207-3b0461eec859/go.mod h1:z5CRVTTTmAJ677TzLL
|
|||||||
golang.org/x/net v0.0.0-20190724013045-ca1201d0de80/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
|
golang.org/x/net v0.0.0-20190724013045-ca1201d0de80/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
|
||||||
golang.org/x/net v0.0.0-20191209160850-c0dbc17a3553/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
|
golang.org/x/net v0.0.0-20191209160850-c0dbc17a3553/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
|
||||||
golang.org/x/net v0.0.0-20200202094626-16171245cfb2/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
|
golang.org/x/net v0.0.0-20200202094626-16171245cfb2/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
|
||||||
golang.org/x/net v0.0.0-20200222125558-5a598a2470a0/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
|
golang.org/x/net v0.0.0-20210226172049-e18ecbb05110/go.mod h1:m0MpNAwzfU5UDzcl9v0D8zg8gWTRqZa9RBIspLL5mdg=
|
||||||
golang.org/x/net v0.0.0-20210405180319-a5a99cb37ef4/go.mod h1:p54w0d4576C0XHj96bSt6lcn1PtDYWL6XObtHCRCNQM=
|
golang.org/x/net v0.0.0-20220722155237-a158d28d115b/go.mod h1:XRhObCWvk6IyKnWLug+ECip1KBveYUHfp+8e9klMJ9c=
|
||||||
|
golang.org/x/net v0.7.0/go.mod h1:2Tu9+aMcznHK/AK1HMvgo6xiTLG5rD5rZLDS+rp2Bjs=
|
||||||
golang.org/x/oauth2 v0.0.0-20180821212333-d2e6202438be/go.mod h1:N/0e6XlmueqKjAGxoOufVs8QHGRruUQn6yWY3a++T0U=
|
golang.org/x/oauth2 v0.0.0-20180821212333-d2e6202438be/go.mod h1:N/0e6XlmueqKjAGxoOufVs8QHGRruUQn6yWY3a++T0U=
|
||||||
golang.org/x/oauth2 v0.0.0-20190226205417-e64efc72b421/go.mod h1:gOpvHmFTYa4IltrdGE7lF6nIHvwfUNPOp7c8zoXwtLw=
|
golang.org/x/oauth2 v0.0.0-20190226205417-e64efc72b421/go.mod h1:gOpvHmFTYa4IltrdGE7lF6nIHvwfUNPOp7c8zoXwtLw=
|
||||||
golang.org/x/oauth2 v0.0.0-20190604053449-0f29369cfe45/go.mod h1:gOpvHmFTYa4IltrdGE7lF6nIHvwfUNPOp7c8zoXwtLw=
|
golang.org/x/oauth2 v0.0.0-20190604053449-0f29369cfe45/go.mod h1:gOpvHmFTYa4IltrdGE7lF6nIHvwfUNPOp7c8zoXwtLw=
|
||||||
@ -218,9 +239,9 @@ golang.org/x/sync v0.0.0-20181221193216-37e7f081c4d4/go.mod h1:RxMgew5VJxzue5/jJ
|
|||||||
golang.org/x/sync v0.0.0-20190227155943-e225da77a7e6/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM=
|
golang.org/x/sync v0.0.0-20190227155943-e225da77a7e6/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM=
|
||||||
golang.org/x/sync v0.0.0-20190423024810-112230192c58/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM=
|
golang.org/x/sync v0.0.0-20190423024810-112230192c58/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM=
|
||||||
golang.org/x/sync v0.0.0-20190911185100-cd5d95a43a6e/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM=
|
golang.org/x/sync v0.0.0-20190911185100-cd5d95a43a6e/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM=
|
||||||
golang.org/x/sync v0.0.0-20210220032951-036812b2e83c/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM=
|
golang.org/x/sync v0.0.0-20220722155255-886fb9371eb4/go.mod h1:RxMgew5VJxzue5/jJTE5uejpjVlOe/izrB70Jof72aM=
|
||||||
golang.org/x/sync v0.8.0 h1:3NFvSEYkUoMifnESzZl15y791HH1qU2xm6eCJU5ZPXQ=
|
golang.org/x/sync v0.14.0 h1:woo0S4Yywslg6hp4eUFjTVOyKt0RookbpAHG4c1HmhQ=
|
||||||
golang.org/x/sync v0.8.0/go.mod h1:Czt+wKu1gCyEFDUtn0jG5QVvpJ6rzVqr5aXyt9drQfk=
|
golang.org/x/sync v0.14.0/go.mod h1:1dzgHSNfp02xaA81J2MS99Qcpr2w7fw1gpm99rleRqA=
|
||||||
golang.org/x/sys v0.0.0-20180830151530-49385e6e1522/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY=
|
golang.org/x/sys v0.0.0-20180830151530-49385e6e1522/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY=
|
||||||
golang.org/x/sys v0.0.0-20190215142949-d0b11bdaac8a/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY=
|
golang.org/x/sys v0.0.0-20190215142949-d0b11bdaac8a/go.mod h1:STP8DvDyc/dI5b8T5hshtkjS+E42TnysNCUPdjciGhY=
|
||||||
golang.org/x/sys v0.0.0-20190312061237-fead79001313/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs=
|
golang.org/x/sys v0.0.0-20190312061237-fead79001313/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs=
|
||||||
@ -233,19 +254,27 @@ golang.org/x/sys v0.0.0-20190726091711-fc99dfbffb4e/go.mod h1:h1NjWce9XRLGQEsW7w
|
|||||||
golang.org/x/sys v0.0.0-20191204072324-ce4227a45e2e/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs=
|
golang.org/x/sys v0.0.0-20191204072324-ce4227a45e2e/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs=
|
||||||
golang.org/x/sys v0.0.0-20191228213918-04cbcbbfeed8/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs=
|
golang.org/x/sys v0.0.0-20191228213918-04cbcbbfeed8/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs=
|
||||||
golang.org/x/sys v0.0.0-20200212091648-12a6c2dcc1e4/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs=
|
golang.org/x/sys v0.0.0-20200212091648-12a6c2dcc1e4/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs=
|
||||||
golang.org/x/sys v0.0.0-20200223170610-d5e6a3e2c0ae/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs=
|
|
||||||
golang.org/x/sys v0.0.0-20201119102817-f84b799fce68/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs=
|
golang.org/x/sys v0.0.0-20201119102817-f84b799fce68/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs=
|
||||||
golang.org/x/sys v0.0.0-20210330210617-4fbd30eecc44/go.mod h1:h1NjWce9XRLGQEsW7wpKNCjG9DtNlClVuFLEZdDNbEs=
|
golang.org/x/sys v0.0.0-20210615035016-665e8c7367d1/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg=
|
||||||
golang.org/x/sys v0.0.0-20210510120138-977fb7262007/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg=
|
golang.org/x/sys v0.0.0-20220520151302-bc2c85ada10a/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg=
|
||||||
|
golang.org/x/sys v0.0.0-20220722155257-8c9f86f7a55f/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg=
|
||||||
|
golang.org/x/sys v0.5.0/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg=
|
||||||
golang.org/x/sys v0.6.0/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg=
|
golang.org/x/sys v0.6.0/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg=
|
||||||
golang.org/x/sys v0.24.0 h1:Twjiwq9dn6R1fQcyiK+wQyHWfaz/BJB+YIpzU/Cv3Xg=
|
golang.org/x/sys v0.33.0 h1:q3i8TbbEz+JRD9ywIRlyRAQbM0qF7hu24q3teo2hbuw=
|
||||||
golang.org/x/sys v0.24.0/go.mod h1:/VUhepiaJMQUp4+oa/7Zr1D23ma6VTLIYjOOTFZPUcA=
|
golang.org/x/sys v0.33.0/go.mod h1:BJP2sWEmIv4KK5OTEluFJCKSidICx8ciO85XgH3Ak8k=
|
||||||
golang.org/x/term v0.0.0-20201126162022-7de9c90e9dd1/go.mod h1:bj7SfCRtBDWHUb9snDiAeCFNEtKQo2Wmx5Cou7ajbmo=
|
golang.org/x/term v0.0.0-20201126162022-7de9c90e9dd1/go.mod h1:bj7SfCRtBDWHUb9snDiAeCFNEtKQo2Wmx5Cou7ajbmo=
|
||||||
|
golang.org/x/term v0.0.0-20210927222741-03fcf44c2211/go.mod h1:jbD1KX2456YbFQfuXm/mYQcufACuNUgVhRMnK/tPxf8=
|
||||||
|
golang.org/x/term v0.5.0/go.mod h1:jMB1sMXY+tzblOD4FWmEbocvup2/aLOaQEp7JmGp78k=
|
||||||
golang.org/x/text v0.0.0-20170915032832-14c0d48ead0c/go.mod h1:NqM8EUOU14njkJ3fqMW+pc6Ldnwhi/IjpwHt7yyuwOQ=
|
golang.org/x/text v0.0.0-20170915032832-14c0d48ead0c/go.mod h1:NqM8EUOU14njkJ3fqMW+pc6Ldnwhi/IjpwHt7yyuwOQ=
|
||||||
golang.org/x/text v0.3.0/go.mod h1:NqM8EUOU14njkJ3fqMW+pc6Ldnwhi/IjpwHt7yyuwOQ=
|
golang.org/x/text v0.3.0/go.mod h1:NqM8EUOU14njkJ3fqMW+pc6Ldnwhi/IjpwHt7yyuwOQ=
|
||||||
golang.org/x/text v0.3.1-0.20180807135948-17ff2d5776d2/go.mod h1:NqM8EUOU14njkJ3fqMW+pc6Ldnwhi/IjpwHt7yyuwOQ=
|
golang.org/x/text v0.3.1-0.20180807135948-17ff2d5776d2/go.mod h1:NqM8EUOU14njkJ3fqMW+pc6Ldnwhi/IjpwHt7yyuwOQ=
|
||||||
golang.org/x/text v0.3.2/go.mod h1:bEr9sfX3Q8Zfm5fL9x+3itogRgK3+ptLWKqgva+5dAk=
|
golang.org/x/text v0.3.2/go.mod h1:bEr9sfX3Q8Zfm5fL9x+3itogRgK3+ptLWKqgva+5dAk=
|
||||||
golang.org/x/text v0.3.3/go.mod h1:5Zoc/QRtKVWzQhOtBMvqHzDpF6irO9z98xDceosuGiQ=
|
golang.org/x/text v0.3.3/go.mod h1:5Zoc/QRtKVWzQhOtBMvqHzDpF6irO9z98xDceosuGiQ=
|
||||||
|
golang.org/x/text v0.3.7/go.mod h1:u+2+/6zg+i71rQMx5EYifcz6MCKuco9NR6JIITiCfzQ=
|
||||||
|
golang.org/x/text v0.3.8/go.mod h1:E6s5w1FMmriuDzIBO73fBruAKo1PCIq6d2Q6DHfQ8WQ=
|
||||||
|
golang.org/x/text v0.7.0/go.mod h1:mrYo+phRRbMaCq/xk9113O4dZlRixOauAjOtrjsXDZ8=
|
||||||
|
golang.org/x/text v0.25.0 h1:qVyWApTSYLk/drJRO5mDlNYskwQznZmkpV2c8q9zls4=
|
||||||
|
golang.org/x/text v0.25.0/go.mod h1:WEdwpYrmk1qmdHvhkSTNPm3app7v4rsT8F2UD6+VHIA=
|
||||||
golang.org/x/time v0.0.0-20181108054448-85acf8d2951c/go.mod h1:tRJNPiyCQ0inRvYxbN9jk5I+vvW/OXSQhTDSoE431IQ=
|
golang.org/x/time v0.0.0-20181108054448-85acf8d2951c/go.mod h1:tRJNPiyCQ0inRvYxbN9jk5I+vvW/OXSQhTDSoE431IQ=
|
||||||
golang.org/x/time v0.0.0-20190308202827-9d24e82272b4/go.mod h1:tRJNPiyCQ0inRvYxbN9jk5I+vvW/OXSQhTDSoE431IQ=
|
golang.org/x/time v0.0.0-20190308202827-9d24e82272b4/go.mod h1:tRJNPiyCQ0inRvYxbN9jk5I+vvW/OXSQhTDSoE431IQ=
|
||||||
golang.org/x/tools v0.0.0-20180917221912-90fa682c2a6e/go.mod h1:n7NCudcB/nEzxVGmLbDWY5pfWTLqBcC2KZ6jyYvM4mQ=
|
golang.org/x/tools v0.0.0-20180917221912-90fa682c2a6e/go.mod h1:n7NCudcB/nEzxVGmLbDWY5pfWTLqBcC2KZ6jyYvM4mQ=
|
||||||
@ -272,15 +301,14 @@ golang.org/x/tools v0.0.0-20191227053925-7b8e75db28f4/go.mod h1:TB2adYChydJhpapK
|
|||||||
golang.org/x/tools v0.0.0-20200130002326-2f3ba24bd6e7/go.mod h1:TB2adYChydJhpapKDTa4BR/hXlZSLoq2Wpct/0txZ28=
|
golang.org/x/tools v0.0.0-20200130002326-2f3ba24bd6e7/go.mod h1:TB2adYChydJhpapKDTa4BR/hXlZSLoq2Wpct/0txZ28=
|
||||||
golang.org/x/tools v0.0.0-20200207183749-b753a1ba74fa/go.mod h1:TB2adYChydJhpapKDTa4BR/hXlZSLoq2Wpct/0txZ28=
|
golang.org/x/tools v0.0.0-20200207183749-b753a1ba74fa/go.mod h1:TB2adYChydJhpapKDTa4BR/hXlZSLoq2Wpct/0txZ28=
|
||||||
golang.org/x/tools v0.0.0-20200212150539-ea181f53ac56/go.mod h1:TB2adYChydJhpapKDTa4BR/hXlZSLoq2Wpct/0txZ28=
|
golang.org/x/tools v0.0.0-20200212150539-ea181f53ac56/go.mod h1:TB2adYChydJhpapKDTa4BR/hXlZSLoq2Wpct/0txZ28=
|
||||||
golang.org/x/tools v0.1.1/go.mod h1:o0xws9oXOQQZyjljx8fwUC0k7L1pTE6eaCbjGeHmOkk=
|
golang.org/x/tools v0.1.12/go.mod h1:hNGJHUnrk76NpqgfD5Aqm5Crs+Hm0VOH/i9J2+nxYbc=
|
||||||
golang.org/x/tools v0.24.0 h1:J1shsA93PJUEVaUSaay7UXAyE8aimq3GW0pjlolpa24=
|
golang.org/x/tools v0.30.0 h1:BgcpHewrV5AUp2G9MebG4XPFI1E2W41zU1SaqVA9vJY=
|
||||||
golang.org/x/tools v0.24.0/go.mod h1:YhNqVBIfWHdzvTLs0d8LCuMhkKUgSUKldakyV7W/WDQ=
|
golang.org/x/tools v0.30.0/go.mod h1:c347cR/OJfw5TI+GfX7RUPNMdDRRbjvYTS0jPyvsVtY=
|
||||||
golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0=
|
golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0=
|
||||||
golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0=
|
golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0=
|
||||||
golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0=
|
golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0=
|
||||||
golang.org/x/xerrors v0.0.0-20200804184101-5ec99f83aff1/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0=
|
gonum.org/v1/gonum v0.16.0 h1:5+ul4Swaf3ESvrOnidPp4GZbzf0mxVQpDCYUQE7OJfk=
|
||||||
gonum.org/v1/gonum v0.15.1 h1:FNy7N6OUZVUaWG9pTiD+jlhdQ3lMP+/LcTpJ6+a8sQ0=
|
gonum.org/v1/gonum v0.16.0/go.mod h1:fef3am4MQ93R2HHpKnLk4/Tbh/s0+wqD5nfa6Pnwy4E=
|
||||||
gonum.org/v1/gonum v0.15.1/go.mod h1:eZTZuRFrzu5pcyjN5wJhcIhnUdNijYxX1T2IcrOGY0o=
|
|
||||||
google.golang.org/api v0.4.0/go.mod h1:8k5glujaEP+g9n7WNsDg8QP6cUVNI86fCNMcbazEtwE=
|
google.golang.org/api v0.4.0/go.mod h1:8k5glujaEP+g9n7WNsDg8QP6cUVNI86fCNMcbazEtwE=
|
||||||
google.golang.org/api v0.7.0/go.mod h1:WtwebWUNSVBH/HAw79HIFXZNqEvBhG+Ra+ax0hx3E3M=
|
google.golang.org/api v0.7.0/go.mod h1:WtwebWUNSVBH/HAw79HIFXZNqEvBhG+Ra+ax0hx3E3M=
|
||||||
google.golang.org/api v0.8.0/go.mod h1:o4eAsZoiT+ibD93RtjEohWalFOjRDx6CVaqeizhEnKg=
|
google.golang.org/api v0.8.0/go.mod h1:o4eAsZoiT+ibD93RtjEohWalFOjRDx6CVaqeizhEnKg=
|
||||||
@ -294,6 +322,8 @@ google.golang.org/appengine v1.4.0/go.mod h1:xpcJRLb0r/rnEns0DIKYYv+WjYCduHsrkT7
|
|||||||
google.golang.org/appengine v1.5.0/go.mod h1:xpcJRLb0r/rnEns0DIKYYv+WjYCduHsrkT7/EB5XEv4=
|
google.golang.org/appengine v1.5.0/go.mod h1:xpcJRLb0r/rnEns0DIKYYv+WjYCduHsrkT7/EB5XEv4=
|
||||||
google.golang.org/appengine v1.6.1/go.mod h1:i06prIuMbXzDqacNJfV5OdTW448YApPu5ww/cMBSeb0=
|
google.golang.org/appengine v1.6.1/go.mod h1:i06prIuMbXzDqacNJfV5OdTW448YApPu5ww/cMBSeb0=
|
||||||
google.golang.org/appengine v1.6.5/go.mod h1:8WjMMxjGQR8xUklV/ARdw2HLXBOI7O7uCIDZVag1xfc=
|
google.golang.org/appengine v1.6.5/go.mod h1:8WjMMxjGQR8xUklV/ARdw2HLXBOI7O7uCIDZVag1xfc=
|
||||||
|
google.golang.org/appengine v1.6.8 h1:IhEN5q69dyKagZPYMSdIjS2HqprW324FRQZJcGqPAsM=
|
||||||
|
google.golang.org/appengine v1.6.8/go.mod h1:1jJ3jBArFh5pcgW8gCtRJnepW8FzD1V44FJffLiz/Ds=
|
||||||
google.golang.org/genproto v0.0.0-20180817151627-c66870c02cf8/go.mod h1:JiN7NxoALGmiZfu7CAH4rXhgtRTLTxftemlI0sWmxmc=
|
google.golang.org/genproto v0.0.0-20180817151627-c66870c02cf8/go.mod h1:JiN7NxoALGmiZfu7CAH4rXhgtRTLTxftemlI0sWmxmc=
|
||||||
google.golang.org/genproto v0.0.0-20190307195333-5fe7a883aa19/go.mod h1:VzzqZJRnGkLBvHegQrXjBqPurQTc5/KpmUdxsrq26oE=
|
google.golang.org/genproto v0.0.0-20190307195333-5fe7a883aa19/go.mod h1:VzzqZJRnGkLBvHegQrXjBqPurQTc5/KpmUdxsrq26oE=
|
||||||
google.golang.org/genproto v0.0.0-20190418145605-e7d98fc518a7/go.mod h1:VzzqZJRnGkLBvHegQrXjBqPurQTc5/KpmUdxsrq26oE=
|
google.golang.org/genproto v0.0.0-20190418145605-e7d98fc518a7/go.mod h1:VzzqZJRnGkLBvHegQrXjBqPurQTc5/KpmUdxsrq26oE=
|
||||||
@ -314,8 +344,13 @@ google.golang.org/grpc v1.23.0/go.mod h1:Y5yQAOtifL1yxbo5wqy6BxZv8vAUGQwXBOALyac
|
|||||||
google.golang.org/grpc v1.26.0/go.mod h1:qbnxyOmOxrQa7FizSgH+ReBfzJrCY1pSN7KXBS8abTk=
|
google.golang.org/grpc v1.26.0/go.mod h1:qbnxyOmOxrQa7FizSgH+ReBfzJrCY1pSN7KXBS8abTk=
|
||||||
google.golang.org/grpc v1.27.0/go.mod h1:qbnxyOmOxrQa7FizSgH+ReBfzJrCY1pSN7KXBS8abTk=
|
google.golang.org/grpc v1.27.0/go.mod h1:qbnxyOmOxrQa7FizSgH+ReBfzJrCY1pSN7KXBS8abTk=
|
||||||
google.golang.org/grpc v1.27.1/go.mod h1:qbnxyOmOxrQa7FizSgH+ReBfzJrCY1pSN7KXBS8abTk=
|
google.golang.org/grpc v1.27.1/go.mod h1:qbnxyOmOxrQa7FizSgH+ReBfzJrCY1pSN7KXBS8abTk=
|
||||||
|
google.golang.org/protobuf v1.26.0-rc.1/go.mod h1:jlhhOSvTdKEhbULTjvd4ARK9grFBp09yW+WbY/TyQbw=
|
||||||
|
google.golang.org/protobuf v1.26.0 h1:bxAC2xTBsZGibn2RTntX0oH50xLsqy1OxA9tTL3p/lk=
|
||||||
|
google.golang.org/protobuf v1.26.0/go.mod h1:9q0QmTI4eRPtz6boOQmLYwt+qCgq0jsYwAQnmE0givc=
|
||||||
gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0=
|
gopkg.in/check.v1 v0.0.0-20161208181325-20d25e280405/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0=
|
||||||
gopkg.in/check.v1 v1.0.0-20180628173108-788fd7840127/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0=
|
gopkg.in/check.v1 v1.0.0-20180628173108-788fd7840127/go.mod h1:Co6ibVJAznAaIkqp8huTwlJQCZ016jof/cbN4VW5Yz0=
|
||||||
|
gopkg.in/check.v1 v1.0.0-20201130134442-10cb98267c6c h1:Hei/4ADfdWqJk1ZMxUNpqntNwaWcugrBjAiHlqqRiVk=
|
||||||
|
gopkg.in/check.v1 v1.0.0-20201130134442-10cb98267c6c/go.mod h1:JHkPIbrfpd72SG/EVd6muEfDQjcINNoR0C8j2r3qZ4Q=
|
||||||
gopkg.in/errgo.v2 v2.1.0/go.mod h1:hNsd1EY+bozCKY1Ytp96fpM3vjJbqLJn88ws8XvfDNI=
|
gopkg.in/errgo.v2 v2.1.0/go.mod h1:hNsd1EY+bozCKY1Ytp96fpM3vjJbqLJn88ws8XvfDNI=
|
||||||
gopkg.in/yaml.v2 v2.2.2/go.mod h1:hI93XBmqTisBFMUTm0b8Fm+jr3Dg1NNxqwp+5A1VGuI=
|
gopkg.in/yaml.v2 v2.2.2/go.mod h1:hI93XBmqTisBFMUTm0b8Fm+jr3Dg1NNxqwp+5A1VGuI=
|
||||||
gopkg.in/yaml.v3 v3.0.0-20200313102051-9f266ea9e77c/go.mod h1:K4uyk7z7BCEPqu6E+C64Yfv1cQ7kz7rIZviUmN+EgEM=
|
gopkg.in/yaml.v3 v3.0.0-20200313102051-9f266ea9e77c/go.mod h1:K4uyk7z7BCEPqu6E+C64Yfv1cQ7kz7rIZviUmN+EgEM=
|
||||||
@ -326,30 +361,28 @@ honnef.co/go/tools v0.0.0-20190106161140-3f1c8253044a/go.mod h1:rf3lG4BRIbNafJWh
|
|||||||
honnef.co/go/tools v0.0.0-20190418001031-e561f6794a2a/go.mod h1:rf3lG4BRIbNafJWhAfAdb/ePZxsR/4RtNHQocxwk9r4=
|
honnef.co/go/tools v0.0.0-20190418001031-e561f6794a2a/go.mod h1:rf3lG4BRIbNafJWhAfAdb/ePZxsR/4RtNHQocxwk9r4=
|
||||||
honnef.co/go/tools v0.0.0-20190523083050-ea95bdfd59fc/go.mod h1:rf3lG4BRIbNafJWhAfAdb/ePZxsR/4RtNHQocxwk9r4=
|
honnef.co/go/tools v0.0.0-20190523083050-ea95bdfd59fc/go.mod h1:rf3lG4BRIbNafJWhAfAdb/ePZxsR/4RtNHQocxwk9r4=
|
||||||
honnef.co/go/tools v0.0.1-2019.2.3/go.mod h1:a3bituU0lyd329TUQxRnasdCoJDkEUEAqEt0JzvZhAg=
|
honnef.co/go/tools v0.0.1-2019.2.3/go.mod h1:a3bituU0lyd329TUQxRnasdCoJDkEUEAqEt0JzvZhAg=
|
||||||
modernc.org/cc/v4 v4.21.4 h1:3Be/Rdo1fpr8GrQ7IVw9OHtplU4gWbb+wNgeoBMmGLQ=
|
modernc.org/cc/v4 v4.24.4 h1:TFkx1s6dCkQpd6dKurBNmpo+G8Zl4Sq/ztJ+2+DEsh0=
|
||||||
modernc.org/cc/v4 v4.21.4/go.mod h1:HM7VJTZbUCR3rV8EYBi9wxnJ0ZBRiGE5OeGXNA0IsLQ=
|
modernc.org/cc/v4 v4.24.4/go.mod h1:uVtb5OGqUKpoLWhqwNQo/8LwvoiEBLvZXIQ/SmO6mL0=
|
||||||
modernc.org/ccgo/v4 v4.19.2 h1:lwQZgvboKD0jBwdaeVCTouxhxAyN6iawF3STraAal8Y=
|
modernc.org/ccgo/v4 v4.23.16 h1:Z2N+kk38b7SfySC1ZkpGLN2vthNJP1+ZzGZIlH7uBxo=
|
||||||
modernc.org/ccgo/v4 v4.19.2/go.mod h1:ysS3mxiMV38XGRTTcgo0DQTeTmAO4oCmJl1nX9VFI3s=
|
modernc.org/ccgo/v4 v4.23.16/go.mod h1:nNma8goMTY7aQZQNTyN9AIoJfxav4nvTnvKThAeMDdo=
|
||||||
modernc.org/fileutil v1.3.0 h1:gQ5SIzK3H9kdfai/5x41oQiKValumqNTDXMvKo62HvE=
|
modernc.org/fileutil v1.3.0 h1:gQ5SIzK3H9kdfai/5x41oQiKValumqNTDXMvKo62HvE=
|
||||||
modernc.org/fileutil v1.3.0/go.mod h1:XatxS8fZi3pS8/hKG2GH/ArUogfxjpEKs3Ku3aK4JyQ=
|
modernc.org/fileutil v1.3.0/go.mod h1:XatxS8fZi3pS8/hKG2GH/ArUogfxjpEKs3Ku3aK4JyQ=
|
||||||
modernc.org/gc/v2 v2.4.1 h1:9cNzOqPyMJBvrUipmynX0ZohMhcxPtMccYgGOJdOiBw=
|
modernc.org/gc/v2 v2.6.3 h1:aJVhcqAte49LF+mGveZ5KPlsp4tdGdAOT4sipJXADjw=
|
||||||
modernc.org/gc/v2 v2.4.1/go.mod h1:wzN5dK1AzVGoH6XOzc3YZ+ey/jPgYHLuVckd62P0GYU=
|
modernc.org/gc/v2 v2.6.3/go.mod h1:YgIahr1ypgfe7chRuJi2gD7DBQiKSLMPgBQe9oIiito=
|
||||||
modernc.org/gc/v3 v3.0.0-20240107210532-573471604cb6 h1:5D53IMaUuA5InSeMu9eJtlQXS2NxAhyWQvkKEgXZhHI=
|
modernc.org/libc v1.61.13 h1:3LRd6ZO1ezsFiX1y+bHd1ipyEHIJKvuprv0sLTBwLW8=
|
||||||
modernc.org/gc/v3 v3.0.0-20240107210532-573471604cb6/go.mod h1:Qz0X07sNOR1jWYCrJMEnbW/X55x206Q7Vt4mz6/wHp4=
|
modernc.org/libc v1.61.13/go.mod h1:8F/uJWL/3nNil0Lgt1Dpz+GgkApWh04N3el3hxJcA6E=
|
||||||
modernc.org/libc v1.55.3 h1:AzcW1mhlPNrRtjS5sS+eW2ISCgSOLLNyFzRh/V3Qj/U=
|
modernc.org/mathutil v1.7.1 h1:GCZVGXdaN8gTqB1Mf/usp1Y/hSqgI2vAGGP4jZMCxOU=
|
||||||
modernc.org/libc v1.55.3/go.mod h1:qFXepLhz+JjFThQ4kzwzOjA/y/artDeg+pcYnY+Q83w=
|
modernc.org/mathutil v1.7.1/go.mod h1:4p5IwJITfppl0G4sUEDtCr4DthTaT47/N3aT6MhfgJg=
|
||||||
modernc.org/mathutil v1.6.0 h1:fRe9+AmYlaej+64JsEEhoWuAYBkOtQiMEU7n/XgfYi4=
|
modernc.org/memory v1.8.2 h1:cL9L4bcoAObu4NkxOlKWBWtNHIsnnACGF/TbqQ6sbcI=
|
||||||
modernc.org/mathutil v1.6.0/go.mod h1:Ui5Q9q1TR2gFm0AQRqQUaBWFLAhQpCwNcuhBOSedWPo=
|
modernc.org/memory v1.8.2/go.mod h1:ZbjSvMO5NQ1A2i3bWeDiVMxIorXwdClKE/0SZ+BMotU=
|
||||||
modernc.org/memory v1.8.0 h1:IqGTL6eFMaDZZhEWwcREgeMXYwmW83LYW8cROZYkg+E=
|
modernc.org/opt v0.1.4 h1:2kNGMRiUjrp4LcaPuLY2PzUfqM/w9N23quVwhKt5Qm8=
|
||||||
modernc.org/memory v1.8.0/go.mod h1:XPZ936zp5OMKGWPqbD3JShgd/ZoQ7899TUuQqxY+peU=
|
modernc.org/opt v0.1.4/go.mod h1:03fq9lsNfvkYSfxrfUhZCWPk1lm4cq4N+Bh//bEtgns=
|
||||||
modernc.org/opt v0.1.3 h1:3XOZf2yznlhC+ibLltsDGzABUGVx8J6pnFMS3E4dcq4=
|
modernc.org/sortutil v1.2.1 h1:+xyoGf15mM3NMlPDnFqrteY07klSFxLElE2PVuWIJ7w=
|
||||||
modernc.org/opt v0.1.3/go.mod h1:WdSiB5evDcignE70guQKxYUl14mgWtbClRi5wmkkTX0=
|
modernc.org/sortutil v1.2.1/go.mod h1:7ZI3a3REbai7gzCLcotuw9AC4VZVpYMjDzETGsSMqJE=
|
||||||
modernc.org/sortutil v1.2.0 h1:jQiD3PfS2REGJNzNCMMaLSp/wdMNieTbKX920Cqdgqc=
|
modernc.org/sqlite v1.35.0 h1:yQps4fegMnZFdphtzlfQTCNBWtS0CZv48pRpW3RFHRw=
|
||||||
modernc.org/sortutil v1.2.0/go.mod h1:TKU2s7kJMf1AE84OoiGppNHJwvB753OYfNl2WRb++Ss=
|
modernc.org/sqlite v1.35.0/go.mod h1:9cr2sicr7jIaWTBKQmAxQLfBv9LL0su4ZTEV+utt3ic=
|
||||||
modernc.org/sqlite v1.32.0 h1:6BM4uGza7bWypsw4fdLRsLxut6bHe4c58VeqjRgST8s=
|
modernc.org/strutil v1.2.1 h1:UneZBkQA+DX2Rp35KcM69cSsNES9ly8mQWD71HKlOA0=
|
||||||
modernc.org/sqlite v1.32.0/go.mod h1:UqoylwmTb9F+IqXERT8bW9zzOWN8qwAIcLdzeBZs4hA=
|
modernc.org/strutil v1.2.1/go.mod h1:EHkiggD70koQxjVdSBM3JKM7k6L0FbGE5eymy9i3B9A=
|
||||||
modernc.org/strutil v1.2.0 h1:agBi9dp1I+eOnxXeiZawM8F4LawKv4NzGWSaLfyeNZA=
|
|
||||||
modernc.org/strutil v1.2.0/go.mod h1:/mdcBmfOibveCTBxUl5B5l6W+TTH1FXPLHZE6bTosX0=
|
|
||||||
modernc.org/token v1.1.0 h1:Xl7Ap9dKaEs5kLoOQeQmPWevfnk/DM5qcLcYlA8ys6Y=
|
modernc.org/token v1.1.0 h1:Xl7Ap9dKaEs5kLoOQeQmPWevfnk/DM5qcLcYlA8ys6Y=
|
||||||
modernc.org/token v1.1.0/go.mod h1:UGzOrNV1mAFSEB63lOFHIpNRUVMvYTc6yu1SMY/XTDM=
|
modernc.org/token v1.1.0/go.mod h1:UGzOrNV1mAFSEB63lOFHIpNRUVMvYTc6yu1SMY/XTDM=
|
||||||
rsc.io/binaryregexp v0.2.0/go.mod h1:qTv7/COck+e2FymRvadv62gMdZztPaShugOCi3I+8D8=
|
rsc.io/binaryregexp v0.2.0/go.mod h1:qTv7/COck+e2FymRvadv62gMdZztPaShugOCi3I+8D8=
|
||||||
|
178
hashing.go
178
hashing.go
@ -7,10 +7,14 @@ import (
|
|||||||
"image"
|
"image"
|
||||||
"log"
|
"log"
|
||||||
"math/bits"
|
"math/bits"
|
||||||
"runtime"
|
|
||||||
"slices"
|
"slices"
|
||||||
|
"strings"
|
||||||
|
"sync"
|
||||||
|
"unsafe"
|
||||||
|
|
||||||
"gitea.narnian.us/lordwelch/goimagehash"
|
"gitea.narnian.us/lordwelch/goimagehash"
|
||||||
|
json "github.com/json-iterator/go"
|
||||||
|
"github.com/vmihailenco/msgpack"
|
||||||
)
|
)
|
||||||
|
|
||||||
//go:embed hashes.gz
|
//go:embed hashes.gz
|
||||||
@ -39,9 +43,12 @@ const (
|
|||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
ComicVine Source = "comicvine.gamespot.com"
|
ComicVine Source = "comicvine.gamespot.com"
|
||||||
|
SavedHashVersion int = 2
|
||||||
)
|
)
|
||||||
|
|
||||||
|
var sources *sync.Map = newSourceMap()
|
||||||
|
|
||||||
type Source string
|
type Source string
|
||||||
|
|
||||||
type Match struct {
|
type Match struct {
|
||||||
@ -50,20 +57,21 @@ type Match struct {
|
|||||||
}
|
}
|
||||||
|
|
||||||
type ID struct {
|
type ID struct {
|
||||||
Domain Source
|
Domain *Source
|
||||||
ID string
|
ID string
|
||||||
}
|
}
|
||||||
|
|
||||||
type Result struct {
|
type Result struct {
|
||||||
IDs IDList
|
Hash Hash
|
||||||
Distance int
|
ID ID
|
||||||
Hash Hash
|
Distance int
|
||||||
|
EquivalentIDs []ID
|
||||||
}
|
}
|
||||||
|
|
||||||
type Im struct {
|
type Im struct {
|
||||||
Im image.Image
|
Im image.Image
|
||||||
Format string
|
Format string
|
||||||
ID ID
|
ID ID
|
||||||
|
NewOnly bool
|
||||||
}
|
}
|
||||||
|
|
||||||
type ImageHash struct {
|
type ImageHash struct {
|
||||||
@ -76,29 +84,85 @@ type Hash struct {
|
|||||||
Kind goimagehash.Kind
|
Kind goimagehash.Kind
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (id *ID) Compare(target ID) int {
|
||||||
|
return cmp.Or(
|
||||||
|
strings.Compare(string(*id.Domain), string(*target.Domain)),
|
||||||
|
strings.Compare(id.ID, target.ID),
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
func newSourceMap() *sync.Map {
|
||||||
|
m := &sync.Map{}
|
||||||
|
for s := range []Source{ComicVine} {
|
||||||
|
m.Store(s, &s)
|
||||||
|
}
|
||||||
|
return m
|
||||||
|
}
|
||||||
|
|
||||||
|
func NewSource[E string | Source](s E) *Source {
|
||||||
|
s2 := Source(strings.ToLower(Clone(string(s))))
|
||||||
|
sp, _ := sources.LoadOrStore(s2, &s2)
|
||||||
|
return sp.(*Source)
|
||||||
|
}
|
||||||
|
|
||||||
// IDList is a map of domain to ID eg IDs["comicvine.gamespot.com"] = []string{"1235"}
|
// IDList is a map of domain to ID eg IDs["comicvine.gamespot.com"] = []string{"1235"}
|
||||||
// Maps are extremely expensive in go for small maps this should only be used to return info to a user no internal code should use this
|
// Maps are extremely expensive in go for small maps this should only be used to return info to a user or as a map containing all IDs for a source
|
||||||
type IDList map[Source][]string
|
type IDList map[Source][]string
|
||||||
|
|
||||||
type OldSavedHashes map[Source]map[string][3]uint64
|
//go:noinline pragma
|
||||||
|
func (a *ID) DecodeMsgpack(dec *msgpack.Decoder) error {
|
||||||
|
var s struct {
|
||||||
|
Domain, ID string
|
||||||
|
}
|
||||||
|
err := dec.Decode(&s)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
type SavedHashes struct {
|
a.ID = Clone(s.ID)
|
||||||
IDs [][]ID
|
a.Domain = NewSource(s.Domain)
|
||||||
Hashes [3]map[uint64]int
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
//go:noinline pragma
|
||||||
|
func Clone(s string) string {
|
||||||
|
if len(s) == 0 {
|
||||||
|
return ""
|
||||||
|
}
|
||||||
|
b := make([]byte, len(s))
|
||||||
|
copy(b, s)
|
||||||
|
return unsafe.String(&b[0], len(b))
|
||||||
|
}
|
||||||
|
|
||||||
|
//go:noinline pragma
|
||||||
|
func (a *ID) UnmarshalJSON(b []byte) error {
|
||||||
|
var s struct {
|
||||||
|
Domain, ID string
|
||||||
|
}
|
||||||
|
if err := json.Unmarshal(b, &s); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
a.ID = Clone(s.ID)
|
||||||
|
domain := Clone(s.Domain)
|
||||||
|
a.Domain = NewSource(domain)
|
||||||
|
|
||||||
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func ToIDList(ids []ID) IDList {
|
func ToIDList(ids []ID) IDList {
|
||||||
idlist := IDList{}
|
idlist := IDList{}
|
||||||
for _, id := range ids {
|
for _, id := range ids {
|
||||||
idlist[id.Domain] = Insert(idlist[id.Domain], id.ID)
|
idlist[*id.Domain] = Insert(idlist[*id.Domain], id.ID)
|
||||||
}
|
}
|
||||||
return idlist
|
return idlist
|
||||||
}
|
}
|
||||||
func InsertID(ids []ID, id ID) []ID {
|
func InsertIDp(ids []*ID, id *ID) []*ID {
|
||||||
index, itemFound := slices.BinarySearchFunc(ids, id, func(e ID, t ID) int {
|
index, itemFound := slices.BinarySearchFunc(ids, id, func(existing, target *ID) int {
|
||||||
return cmp.Or(
|
return cmp.Or(
|
||||||
cmp.Compare(e.Domain, t.Domain),
|
cmp.Compare(*existing.Domain, *target.Domain),
|
||||||
cmp.Compare(e.ID, t.ID),
|
cmp.Compare(existing.ID, target.ID),
|
||||||
)
|
)
|
||||||
})
|
})
|
||||||
if itemFound {
|
if itemFound {
|
||||||
@ -106,51 +170,18 @@ func InsertID(ids []ID, id ID) []ID {
|
|||||||
}
|
}
|
||||||
return slices.Insert(ids, index, id)
|
return slices.Insert(ids, index, id)
|
||||||
}
|
}
|
||||||
func (s *SavedHashes) InsertHash(hash Hash, id ID) {
|
|
||||||
for i, h := range s.Hashes {
|
|
||||||
if h == nil {
|
|
||||||
s.Hashes[i] = make(map[uint64]int)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
hashType := int(hash.Kind) - 1
|
func InsertID(ids []ID, id ID) []ID {
|
||||||
idx, hashFound := s.Hashes[hashType][hash.Hash]
|
index, itemFound := slices.BinarySearchFunc(ids, id, func(existing, target ID) int {
|
||||||
if !hashFound {
|
return cmp.Or(
|
||||||
idx = len(s.IDs)
|
cmp.Compare(*existing.Domain, *target.Domain),
|
||||||
s.IDs = append(s.IDs, make([]ID, 0, 3))
|
cmp.Compare(existing.ID, target.ID),
|
||||||
|
)
|
||||||
|
})
|
||||||
|
if itemFound {
|
||||||
|
return ids
|
||||||
}
|
}
|
||||||
s.IDs[idx] = InsertID(s.IDs[idx], id)
|
return slices.Insert(ids, index, id)
|
||||||
s.Hashes[hashType][hash.Hash] = idx
|
|
||||||
}
|
|
||||||
|
|
||||||
func ConvertSavedHashes(oldHashes OldSavedHashes) SavedHashes {
|
|
||||||
t := SavedHashes{}
|
|
||||||
idcount := 0
|
|
||||||
for _, ids := range oldHashes {
|
|
||||||
idcount += len(ids)
|
|
||||||
}
|
|
||||||
t.IDs = make([][]ID, 0, idcount)
|
|
||||||
t.Hashes[0] = make(map[uint64]int, idcount)
|
|
||||||
t.Hashes[1] = make(map[uint64]int, idcount)
|
|
||||||
t.Hashes[2] = make(map[uint64]int, idcount)
|
|
||||||
for domain, sourceHashes := range oldHashes {
|
|
||||||
for id, hashes := range sourceHashes {
|
|
||||||
idx := len(t.IDs)
|
|
||||||
t.IDs = append(t.IDs, []ID{{domain, id}})
|
|
||||||
for hashType, hash := range hashes {
|
|
||||||
t.Hashes[hashType][hash] = idx
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
fmt.Println("Expected number of IDs", idcount)
|
|
||||||
idcount = 0
|
|
||||||
for _, ids := range t.IDs {
|
|
||||||
idcount += len(ids)
|
|
||||||
}
|
|
||||||
fmt.Println("length of hashes", len(t.Hashes[0])+len(t.Hashes[1])+len(t.Hashes[2]))
|
|
||||||
fmt.Println("Length of ID lists", len(t.IDs))
|
|
||||||
fmt.Println("Total number of IDs", idcount)
|
|
||||||
return t
|
|
||||||
}
|
}
|
||||||
|
|
||||||
type NewIDs struct {
|
type NewIDs struct {
|
||||||
@ -161,14 +192,14 @@ type NewIDs struct {
|
|||||||
type HashStorage interface {
|
type HashStorage interface {
|
||||||
GetMatches(hashes []Hash, max int, exactOnly bool) ([]Result, error)
|
GetMatches(hashes []Hash, max int, exactOnly bool) ([]Result, error)
|
||||||
MapHashes(ImageHash)
|
MapHashes(ImageHash)
|
||||||
DecodeHashes(hashes SavedHashes) error
|
DecodeHashes(hashes *SavedHashes) error
|
||||||
EncodeHashes() (SavedHashes, error)
|
EncodeHashes() (*SavedHashes, error)
|
||||||
AssociateIDs(newIDs []NewIDs) error
|
AssociateIDs(newIDs []NewIDs) error
|
||||||
GetIDs(id ID) IDList
|
GetIDs(id ID) IDList
|
||||||
}
|
}
|
||||||
|
|
||||||
func Atleast(maxDistance int, searchHash uint64, hashes []uint64) []Match {
|
func Atleast(maxDistance int, searchHash uint64, hashes []uint64) []Match {
|
||||||
matchingHashes := make([]Match, 0, len(hashes)/2) // hope that we don't need all of them
|
matchingHashes := make([]Match, 0, 20) // hope that we don't need all of them
|
||||||
for _, storedHash := range hashes {
|
for _, storedHash := range hashes {
|
||||||
distance := bits.OnesCount64(searchHash ^ storedHash)
|
distance := bits.OnesCount64(searchHash ^ storedHash)
|
||||||
if distance <= maxDistance {
|
if distance <= maxDistance {
|
||||||
@ -178,14 +209,6 @@ func Atleast(maxDistance int, searchHash uint64, hashes []uint64) []Match {
|
|||||||
return matchingHashes
|
return matchingHashes
|
||||||
}
|
}
|
||||||
|
|
||||||
func Insert[S ~[]E, E cmp.Ordered](slice S, item E) S {
|
|
||||||
index, itemFound := slices.BinarySearch(slice, item)
|
|
||||||
if itemFound {
|
|
||||||
return slice
|
|
||||||
}
|
|
||||||
return slices.Insert(slice, index, item)
|
|
||||||
}
|
|
||||||
|
|
||||||
func InsertIdx[S ~[]E, E cmp.Ordered](slice S, item E) (S, int) {
|
func InsertIdx[S ~[]E, E cmp.Ordered](slice S, item E) (S, int) {
|
||||||
index, itemFound := slices.BinarySearch(slice, item)
|
index, itemFound := slices.BinarySearch(slice, item)
|
||||||
if itemFound {
|
if itemFound {
|
||||||
@ -194,10 +217,9 @@ func InsertIdx[S ~[]E, E cmp.Ordered](slice S, item E) (S, int) {
|
|||||||
return slices.Insert(slice, index, item), index
|
return slices.Insert(slice, index, item), index
|
||||||
}
|
}
|
||||||
|
|
||||||
func MemStats() uint64 {
|
func Insert[S ~[]E, E cmp.Ordered](slice S, item E) S {
|
||||||
var m runtime.MemStats
|
slice, _ = InsertIdx(slice, item)
|
||||||
runtime.ReadMemStats(&m)
|
return slice
|
||||||
return m.Alloc
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func HashImage(i Im) ImageHash {
|
func HashImage(i Im) ImageHash {
|
||||||
|
150
map.go
150
map.go
@ -1,150 +0,0 @@
|
|||||||
package ch
|
|
||||||
|
|
||||||
import (
|
|
||||||
"cmp"
|
|
||||||
"fmt"
|
|
||||||
"slices"
|
|
||||||
"sync"
|
|
||||||
)
|
|
||||||
|
|
||||||
type MapStorage struct {
|
|
||||||
basicMapStorage
|
|
||||||
partialHash [3][8]map[uint8][]uint64
|
|
||||||
}
|
|
||||||
|
|
||||||
func (m *MapStorage) GetMatches(hashes []Hash, max int, exactOnly bool) ([]Result, error) {
|
|
||||||
var foundMatches []Result
|
|
||||||
m.hashMutex.RLock()
|
|
||||||
defer m.hashMutex.RUnlock()
|
|
||||||
resetTime()
|
|
||||||
defer logTime("Search Complete")
|
|
||||||
|
|
||||||
if exactOnly { // exact matches are also found by partial matches. Don't bother with exact matches so we don't have to de-duplicate
|
|
||||||
for _, hash := range hashes {
|
|
||||||
hashType := int(hash.Kind) - 1
|
|
||||||
index, hashFound := m.findHash(hashType, hash.Hash)
|
|
||||||
if hashFound {
|
|
||||||
foundMatches = append(foundMatches, Result{
|
|
||||||
Distance: 0,
|
|
||||||
Hash: hash,
|
|
||||||
IDs: ToIDList(*m.hashes[hashType][index].ids),
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// If we have exact matches don't bother with other matches
|
|
||||||
logTime("Search Exact")
|
|
||||||
if len(foundMatches) > 0 && exactOnly {
|
|
||||||
return foundMatches, nil
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
totalPartialHashes := 0
|
|
||||||
for _, searchHash := range hashes {
|
|
||||||
foundHashes := make(map[uint64]struct{})
|
|
||||||
hashType := int(searchHash.Kind) - 1
|
|
||||||
for i, partialHash := range SplitHash(searchHash.Hash) {
|
|
||||||
partialHashes := m.partialHash[hashType][i][partialHash]
|
|
||||||
totalPartialHashes += len(partialHashes)
|
|
||||||
for _, match := range Atleast(max, searchHash.Hash, partialHashes) {
|
|
||||||
_, alreadyMatched := foundHashes[match.Hash]
|
|
||||||
if index, hashFound := m.findHash(hashType, match.Hash); hashFound && !alreadyMatched {
|
|
||||||
foundHashes[match.Hash] = struct{}{}
|
|
||||||
foundMatches = append(foundMatches, Result{IDs: ToIDList(*m.hashes[hashType][index].ids), Distance: match.Distance, Hash: Hash{Hash: match.Hash, Kind: searchHash.Kind}})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
fmt.Println("Total partial hashes tested:", totalPartialHashes)
|
|
||||||
go m.printSizes()
|
|
||||||
return foundMatches, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (m *MapStorage) MapHashes(hash ImageHash) {
|
|
||||||
m.basicMapStorage.MapHashes(hash)
|
|
||||||
for _, hash := range hash.Hashes {
|
|
||||||
hashType := int(hash.Kind) - 1
|
|
||||||
for i, partialHash := range SplitHash(hash.Hash) {
|
|
||||||
m.partialHash[hashType][i][partialHash] = Insert(m.partialHash[hashType][i][partialHash], hash.Hash)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func (m *MapStorage) DecodeHashes(hashes SavedHashes) error {
|
|
||||||
for hashType, sourceHashes := range hashes.Hashes {
|
|
||||||
m.hashes[hashType] = make([]structHash, len(sourceHashes))
|
|
||||||
for savedHash, idlistLocation := range sourceHashes {
|
|
||||||
m.hashes[hashType] = append(m.hashes[hashType], structHash{savedHash, &hashes.IDs[idlistLocation]})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
for hashType := range m.hashes {
|
|
||||||
slices.SortFunc(m.hashes[hashType], func(a, b structHash) int {
|
|
||||||
return cmp.Compare(a.hash, b.hash)
|
|
||||||
})
|
|
||||||
}
|
|
||||||
m.printSizes()
|
|
||||||
for _, partialHashes := range m.partialHash {
|
|
||||||
for _, partMap := range partialHashes {
|
|
||||||
for part, hashes := range partMap {
|
|
||||||
slices.Sort(hashes)
|
|
||||||
partMap[part] = slices.Compact(hashes)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
m.printSizes()
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (m *MapStorage) printSizes() {
|
|
||||||
fmt.Println("Length of hashes:", len(m.hashes[0])+len(m.hashes[1])+len(m.hashes[2]))
|
|
||||||
// fmt.Println("Size of", "hashes:", size.Of(m.hashes)/1024/1024, "MB")
|
|
||||||
// fmt.Println("Size of", "ids:", size.Of(m.ids)/1024/1024, "MB")
|
|
||||||
// fmt.Println("Size of", "MapStorage:", size.Of(m)/1024/1024, "MB")
|
|
||||||
|
|
||||||
}
|
|
||||||
|
|
||||||
func NewMapStorage() (HashStorage, error) {
|
|
||||||
storage := &MapStorage{
|
|
||||||
basicMapStorage: basicMapStorage{
|
|
||||||
hashMutex: sync.RWMutex{},
|
|
||||||
hashes: [3][]structHash{
|
|
||||||
[]structHash{},
|
|
||||||
[]structHash{},
|
|
||||||
[]structHash{},
|
|
||||||
},
|
|
||||||
},
|
|
||||||
partialHash: [3][8]map[uint8][]uint64{
|
|
||||||
{
|
|
||||||
make(map[uint8][]uint64),
|
|
||||||
make(map[uint8][]uint64),
|
|
||||||
make(map[uint8][]uint64),
|
|
||||||
make(map[uint8][]uint64),
|
|
||||||
make(map[uint8][]uint64),
|
|
||||||
make(map[uint8][]uint64),
|
|
||||||
make(map[uint8][]uint64),
|
|
||||||
make(map[uint8][]uint64),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
make(map[uint8][]uint64),
|
|
||||||
make(map[uint8][]uint64),
|
|
||||||
make(map[uint8][]uint64),
|
|
||||||
make(map[uint8][]uint64),
|
|
||||||
make(map[uint8][]uint64),
|
|
||||||
make(map[uint8][]uint64),
|
|
||||||
make(map[uint8][]uint64),
|
|
||||||
make(map[uint8][]uint64),
|
|
||||||
},
|
|
||||||
{
|
|
||||||
make(map[uint8][]uint64),
|
|
||||||
make(map[uint8][]uint64),
|
|
||||||
make(map[uint8][]uint64),
|
|
||||||
make(map[uint8][]uint64),
|
|
||||||
make(map[uint8][]uint64),
|
|
||||||
make(map[uint8][]uint64),
|
|
||||||
make(map[uint8][]uint64),
|
|
||||||
make(map[uint8][]uint64),
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}
|
|
||||||
return storage, nil
|
|
||||||
}
|
|
16
path.go
Normal file
16
path.go
Normal file
@ -0,0 +1,16 @@
|
|||||||
|
package ch
|
||||||
|
|
||||||
|
import (
|
||||||
|
"os"
|
||||||
|
"path/filepath"
|
||||||
|
)
|
||||||
|
|
||||||
|
func RmdirP(path string) error {
|
||||||
|
err := os.Remove(path)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
dir, _ := filepath.Split(path)
|
||||||
|
_ = RmdirP(dir) // We only care about errors for the first directory we always expect atleast one to fail
|
||||||
|
return nil
|
||||||
|
}
|
274
savedHashes.go
Normal file
274
savedHashes.go
Normal file
@ -0,0 +1,274 @@
|
|||||||
|
package ch
|
||||||
|
|
||||||
|
import (
|
||||||
|
"cmp"
|
||||||
|
"errors"
|
||||||
|
"fmt"
|
||||||
|
"slices"
|
||||||
|
"strings"
|
||||||
|
|
||||||
|
// json "github.com/goccy/go-json"
|
||||||
|
json "github.com/json-iterator/go"
|
||||||
|
// "encoding/json"
|
||||||
|
|
||||||
|
"gitea.narnian.us/lordwelch/goimagehash"
|
||||||
|
"github.com/vmihailenco/msgpack"
|
||||||
|
)
|
||||||
|
|
||||||
|
type Format int
|
||||||
|
|
||||||
|
const (
|
||||||
|
Msgpack Format = iota + 1
|
||||||
|
JSON
|
||||||
|
|
||||||
|
CurrentSavedHashesVersion int = 2
|
||||||
|
)
|
||||||
|
|
||||||
|
var versionMap = map[int]versionDecoder{
|
||||||
|
0: DecodeHashesV0,
|
||||||
|
1: DecodeHashesV1,
|
||||||
|
2: DecodeHashesV2,
|
||||||
|
}
|
||||||
|
|
||||||
|
var formatNames = map[Format]string{
|
||||||
|
JSON: "json",
|
||||||
|
Msgpack: "msgpack",
|
||||||
|
}
|
||||||
|
|
||||||
|
var formatValues = map[string]Format{
|
||||||
|
"json": JSON,
|
||||||
|
"msgpack": Msgpack,
|
||||||
|
}
|
||||||
|
|
||||||
|
type OldSavedHashes map[Source]map[string][3]uint64
|
||||||
|
type SavedHashesv1 struct {
|
||||||
|
IDs [][]ID
|
||||||
|
Hashes [3]map[uint64]int
|
||||||
|
}
|
||||||
|
|
||||||
|
// SavedHashes The IDs and Hashes fields have no direct correlation
|
||||||
|
// It is perfectly valid to have an empty IDs or an empty Hashes field
|
||||||
|
// If two covers have identical hashes then they should be two entries in Hashes not a set in IDs with two IDs from the same source
|
||||||
|
type SavedHashes struct {
|
||||||
|
Version int
|
||||||
|
IDs [][]ID // List of sets of IDs that are the same across Sources, should generally only have one Source per set
|
||||||
|
Hashes []SavedHash // List of all known hashes, hashes will be duplicated for each source
|
||||||
|
}
|
||||||
|
|
||||||
|
type SavedHash struct {
|
||||||
|
Hash Hash
|
||||||
|
ID ID
|
||||||
|
}
|
||||||
|
|
||||||
|
type Encoder func(any) ([]byte, error)
|
||||||
|
type Decoder func([]byte, interface{}) error
|
||||||
|
type versionDecoder func(Decoder, []byte) (*SavedHashes, error)
|
||||||
|
|
||||||
|
var NoHashes = errors.New("no hashes")
|
||||||
|
var DecodeError = errors.New("decoder failure")
|
||||||
|
|
||||||
|
func (f Format) String() string {
|
||||||
|
if name, known := formatNames[f]; known {
|
||||||
|
return name
|
||||||
|
}
|
||||||
|
return "Unknown"
|
||||||
|
}
|
||||||
|
|
||||||
|
func (f *Format) Set(s string) error {
|
||||||
|
if format, known := formatValues[strings.ToLower(s)]; known {
|
||||||
|
*f = format
|
||||||
|
} else {
|
||||||
|
return fmt.Errorf("Unknown format: %d", f)
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (h *SavedHash) Clone() SavedHash {
|
||||||
|
return SavedHash{
|
||||||
|
Hash: Hash{
|
||||||
|
Hash: h.Hash.Hash,
|
||||||
|
Kind: h.Hash.Kind,
|
||||||
|
},
|
||||||
|
ID: ID{
|
||||||
|
Domain: NewSource(*h.ID.Domain),
|
||||||
|
ID: strings.Clone(h.ID.ID),
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *SavedHashes) InsertHash(hash SavedHash) {
|
||||||
|
index, itemFound := slices.BinarySearchFunc(s.Hashes, hash, func(existing SavedHash, target SavedHash) int {
|
||||||
|
return cmp.Or(
|
||||||
|
cmp.Compare(existing.Hash.Hash, target.Hash.Hash),
|
||||||
|
cmp.Compare(existing.Hash.Kind, target.Hash.Kind),
|
||||||
|
cmp.Compare(*existing.ID.Domain, *target.ID.Domain),
|
||||||
|
cmp.Compare(existing.ID.ID, target.ID.ID),
|
||||||
|
)
|
||||||
|
})
|
||||||
|
if !itemFound {
|
||||||
|
s.Hashes = slices.Insert(s.Hashes, index, hash)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func ConvertHashesV0(oldHashes OldSavedHashes) *SavedHashes {
|
||||||
|
t := SavedHashes{}
|
||||||
|
idcount := 0
|
||||||
|
for _, ids := range oldHashes {
|
||||||
|
idcount += len(ids)
|
||||||
|
}
|
||||||
|
t.Hashes = make([]SavedHash, 0, idcount)
|
||||||
|
for domain, sourceHashes := range oldHashes {
|
||||||
|
for id, hashes := range sourceHashes {
|
||||||
|
for hashType, hash := range hashes {
|
||||||
|
t.Hashes = append(t.Hashes, SavedHash{
|
||||||
|
Hash: Hash{
|
||||||
|
Kind: goimagehash.Kind(hashType + 1),
|
||||||
|
Hash: hash,
|
||||||
|
},
|
||||||
|
ID: ID{NewSource(domain), id},
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
fmt.Println("Length of hashes", len(t.Hashes))
|
||||||
|
fmt.Println("Length of ID lists", len(t.IDs))
|
||||||
|
return &t
|
||||||
|
}
|
||||||
|
|
||||||
|
func ConvertHashesV1(oldHashes SavedHashesv1) *SavedHashes {
|
||||||
|
t := SavedHashes{}
|
||||||
|
hashCount := 0
|
||||||
|
for _, hashes := range oldHashes.Hashes {
|
||||||
|
hashCount += len(hashes)
|
||||||
|
}
|
||||||
|
t.IDs = oldHashes.IDs
|
||||||
|
t.Hashes = make([]SavedHash, 0, hashCount)
|
||||||
|
for hashType, sourceHashes := range oldHashes.Hashes {
|
||||||
|
for hash, index := range sourceHashes {
|
||||||
|
for _, id := range oldHashes.IDs[index] {
|
||||||
|
t.Hashes = append(t.Hashes, SavedHash{
|
||||||
|
ID: ID{NewSource(*id.Domain), id.ID},
|
||||||
|
Hash: Hash{
|
||||||
|
Kind: goimagehash.Kind(hashType + 1),
|
||||||
|
Hash: hash,
|
||||||
|
},
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
fmt.Println("Length of hashes", len(t.Hashes))
|
||||||
|
fmt.Println("Length of ID lists", len(t.IDs))
|
||||||
|
return &t
|
||||||
|
}
|
||||||
|
|
||||||
|
func DecodeHashesV0(decode Decoder, hashes []byte) (*SavedHashes, error) {
|
||||||
|
loadedHashes := OldSavedHashes{}
|
||||||
|
err := decode(hashes, &loadedHashes)
|
||||||
|
if err != nil {
|
||||||
|
return nil, fmt.Errorf("%w: %w", DecodeError, err)
|
||||||
|
}
|
||||||
|
if len(loadedHashes) == 0 {
|
||||||
|
return nil, NoHashes
|
||||||
|
}
|
||||||
|
fmt.Println("Loaded V0 hashes")
|
||||||
|
return ConvertHashesV0(loadedHashes), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func DecodeHashesV1(decode Decoder, hashes []byte) (*SavedHashes, error) {
|
||||||
|
loadedHashes := SavedHashesv1{}
|
||||||
|
err := decode(hashes, &loadedHashes)
|
||||||
|
if err != nil {
|
||||||
|
return nil, fmt.Errorf("%w: %w", DecodeError, err)
|
||||||
|
}
|
||||||
|
hashesCount := 0
|
||||||
|
for _, hashes := range loadedHashes.Hashes {
|
||||||
|
hashesCount += len(hashes)
|
||||||
|
}
|
||||||
|
if hashesCount < 1 {
|
||||||
|
return nil, NoHashes
|
||||||
|
}
|
||||||
|
fmt.Println("Loaded V1 hashes")
|
||||||
|
return ConvertHashesV1(loadedHashes), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func DecodeHashesV2(decode Decoder, hashes []byte) (*SavedHashes, error) {
|
||||||
|
fmt.Println("Decode v2 hashes")
|
||||||
|
loadedHashes := SavedHashes{}
|
||||||
|
err := decode(hashes, &loadedHashes)
|
||||||
|
if err != nil {
|
||||||
|
return nil, fmt.Errorf("%w: %w", DecodeError, err)
|
||||||
|
}
|
||||||
|
if len(loadedHashes.Hashes) < 1 && len(loadedHashes.IDs) < 1 {
|
||||||
|
return nil, NoHashes
|
||||||
|
}
|
||||||
|
|
||||||
|
fmt.Println("Length of hashes", len(loadedHashes.Hashes))
|
||||||
|
fmt.Println("Length of ID lists", len(loadedHashes.IDs))
|
||||||
|
fmt.Println("Loaded V2 hashes")
|
||||||
|
return &loadedHashes, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func getSavedHashesVersion(decode Decoder, hashes []byte) (int, error) {
|
||||||
|
type version struct {
|
||||||
|
Version int
|
||||||
|
}
|
||||||
|
var savedVersion version
|
||||||
|
err := decode(hashes, &savedVersion)
|
||||||
|
if err != nil {
|
||||||
|
return -1, fmt.Errorf("%w: %w", DecodeError, err)
|
||||||
|
}
|
||||||
|
if savedVersion.Version > 1 {
|
||||||
|
return savedVersion.Version, nil
|
||||||
|
}
|
||||||
|
return -1, nil
|
||||||
|
}
|
||||||
|
func DecodeHashes(format Format, hashes []byte) (*SavedHashes, error) {
|
||||||
|
var decode Decoder
|
||||||
|
switch format {
|
||||||
|
case Msgpack:
|
||||||
|
decode = msgpack.Unmarshal
|
||||||
|
fmt.Println("Decode Msgpack")
|
||||||
|
case JSON:
|
||||||
|
decode = json.Unmarshal
|
||||||
|
fmt.Println("Decode JSON")
|
||||||
|
|
||||||
|
default:
|
||||||
|
return nil, fmt.Errorf("Unknown format: %v", format)
|
||||||
|
}
|
||||||
|
version, err := getSavedHashesVersion(decode, hashes)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
if decodeVersion, knownVersion := versionMap[version]; knownVersion {
|
||||||
|
return decodeVersion(decode, hashes)
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, decodeVersion := range []versionDecoder{
|
||||||
|
DecodeHashesV0,
|
||||||
|
DecodeHashesV1,
|
||||||
|
DecodeHashesV2,
|
||||||
|
} {
|
||||||
|
loadedHashes, err := decodeVersion(decode, hashes)
|
||||||
|
if err == nil {
|
||||||
|
return loadedHashes, nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil, NoHashes
|
||||||
|
}
|
||||||
|
|
||||||
|
func EncodeHashes(hashes *SavedHashes, format Format) ([]byte, error) {
|
||||||
|
var encoder Encoder
|
||||||
|
switch format {
|
||||||
|
case Msgpack:
|
||||||
|
encoder = msgpack.Marshal
|
||||||
|
case JSON:
|
||||||
|
encoder = json.Marshal
|
||||||
|
default:
|
||||||
|
return nil, fmt.Errorf("Unknown format: %v", format)
|
||||||
|
}
|
||||||
|
|
||||||
|
hashes.Version = CurrentSavedHashesVersion
|
||||||
|
return encoder(hashes)
|
||||||
|
}
|
466
sqlite.go
466
sqlite.go
@ -1,466 +0,0 @@
|
|||||||
package ch
|
|
||||||
|
|
||||||
import (
|
|
||||||
"context"
|
|
||||||
"database/sql"
|
|
||||||
"errors"
|
|
||||||
"fmt"
|
|
||||||
"log"
|
|
||||||
"math/bits"
|
|
||||||
"strings"
|
|
||||||
"time"
|
|
||||||
|
|
||||||
"gitea.narnian.us/lordwelch/goimagehash"
|
|
||||||
_ "modernc.org/sqlite"
|
|
||||||
)
|
|
||||||
|
|
||||||
type sqliteStorage struct {
|
|
||||||
db *sql.DB
|
|
||||||
}
|
|
||||||
type sqliteHash struct {
|
|
||||||
hashid int
|
|
||||||
Result
|
|
||||||
}
|
|
||||||
|
|
||||||
func (s *sqliteStorage) findExactHashes(statement *sql.Stmt, items ...interface{}) ([]sqliteHash, error) { // exact matches are also found by partial matches. Don't bother with exact matches so we don't have to de-duplicate
|
|
||||||
hashes := []sqliteHash{}
|
|
||||||
rows, err := statement.Query(items...)
|
|
||||||
if err != nil {
|
|
||||||
return hashes, err
|
|
||||||
}
|
|
||||||
|
|
||||||
for rows.Next() {
|
|
||||||
var (
|
|
||||||
r = sqliteHash{Result: Result{IDs: make(IDList)}}
|
|
||||||
h int64
|
|
||||||
)
|
|
||||||
err = rows.Scan(&r.hashid, &h, &r.Hash.Kind)
|
|
||||||
if err != nil {
|
|
||||||
rows.Close()
|
|
||||||
return hashes, err
|
|
||||||
}
|
|
||||||
r.Hash.Hash = uint64(h)
|
|
||||||
hashes = append(hashes, r)
|
|
||||||
}
|
|
||||||
rows.Close()
|
|
||||||
statement, err = s.db.PrepareContext(context.Background(), `SELECT IDS.domain, IDs.id FROM IDs JOIN id_hash ON IDs.rowid = id_hash.idid WHERE (id_hash.hashid=?) ORDER BY IDs.domain, IDs.ID;`)
|
|
||||||
if err != nil {
|
|
||||||
return hashes, err
|
|
||||||
}
|
|
||||||
for _, hash := range hashes {
|
|
||||||
rows, err := statement.Query(hash.hashid)
|
|
||||||
if err != nil {
|
|
||||||
return hashes, err
|
|
||||||
}
|
|
||||||
for rows.Next() {
|
|
||||||
var source Source
|
|
||||||
var id string
|
|
||||||
err := rows.Scan(&source, &id)
|
|
||||||
if err != nil {
|
|
||||||
return hashes, err
|
|
||||||
}
|
|
||||||
hash.IDs[source] = append(hash.IDs[source], id)
|
|
||||||
}
|
|
||||||
rows.Close()
|
|
||||||
}
|
|
||||||
return hashes, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (s *sqliteStorage) findPartialHashes(max int, search_hash int64, kind goimagehash.Kind) ([]sqliteHash, error) { // exact matches are also found by partial matches. Don't bother with exact matches so we don't have to de-duplicate
|
|
||||||
hashes := []sqliteHash{}
|
|
||||||
statement, err := s.db.PrepareContext(context.Background(), `SELECT rowid,hash,kind FROM Hashes WHERE (kind=?) AND (((hash >> (0 * 8) & 0xFF)=(?2 >> (0 * 8) & 0xFF)) OR ((hash >> (1 * 8) & 0xFF)=(?2 >> (1 * 8) & 0xFF)) OR ((hash >> (2 * 8) & 0xFF)=(?2 >> (2 * 8) & 0xFF)) OR ((hash >> (3 * 8) & 0xFF)=(?2 >> (3 * 8) & 0xFF)) OR ((hash >> (4 * 8) & 0xFF)=(?2 >> (4 * 8) & 0xFF)) OR ((hash >> (5 * 8) & 0xFF)=(?2 >> (5 * 8) & 0xFF)) OR ((hash >> (6 * 8) & 0xFF)=(?2 >> (6 * 8) & 0xFF)) OR ((hash >> (7 * 8) & 0xFF)=(?2 >> (7 * 8) & 0xFF)));`)
|
|
||||||
if err != nil {
|
|
||||||
return hashes, err
|
|
||||||
}
|
|
||||||
rows, err := statement.Query(kind, int64(search_hash))
|
|
||||||
if err != nil {
|
|
||||||
return hashes, err
|
|
||||||
}
|
|
||||||
|
|
||||||
for rows.Next() {
|
|
||||||
var (
|
|
||||||
r = sqliteHash{Result: Result{IDs: make(IDList)}}
|
|
||||||
h int64
|
|
||||||
)
|
|
||||||
err = rows.Scan(&r.hashid, &h, &r.Hash.Kind)
|
|
||||||
if err != nil {
|
|
||||||
rows.Close()
|
|
||||||
return hashes, err
|
|
||||||
}
|
|
||||||
r.Hash.Hash = uint64(h)
|
|
||||||
r.Distance = bits.OnesCount64(uint64(search_hash) ^ r.Hash.Hash)
|
|
||||||
if r.Distance <= max {
|
|
||||||
hashes = append(hashes, r)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
rows.Close()
|
|
||||||
logTime("Filter partial " + kind.String())
|
|
||||||
|
|
||||||
statement, err = s.db.PrepareContext(context.Background(), `SELECT DISTINCT IDS.domain, IDs.id, id_hash.hashid FROM IDs JOIN id_hash ON IDs.rowid = id_hash.idid WHERE (id_hash.hashid in (`+strings.TrimRight(strings.Repeat("?,", len(hashes)), ",")+`)) ORDER BY IDs.domain, IDs.ID;`)
|
|
||||||
if err != nil {
|
|
||||||
return hashes, err
|
|
||||||
}
|
|
||||||
|
|
||||||
var ids []any
|
|
||||||
for _, hash := range hashes {
|
|
||||||
ids = append(ids, hash.hashid)
|
|
||||||
}
|
|
||||||
rows, err = statement.Query(ids...)
|
|
||||||
if err != nil {
|
|
||||||
return hashes, err
|
|
||||||
}
|
|
||||||
for rows.Next() {
|
|
||||||
var source Source
|
|
||||||
var id string
|
|
||||||
var hashid int
|
|
||||||
err := rows.Scan(&source, &id, &hashid)
|
|
||||||
if err != nil {
|
|
||||||
return hashes, err
|
|
||||||
}
|
|
||||||
for _, hash := range hashes {
|
|
||||||
if hash.hashid == hashid {
|
|
||||||
hash.IDs[source] = append(hash.IDs[source], id)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
rows.Close()
|
|
||||||
return hashes, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (s *sqliteStorage) dropIndexes() error {
|
|
||||||
_, err := s.db.Exec(`
|
|
||||||
|
|
||||||
DROP INDEX IF EXISTS hash_index;
|
|
||||||
DROP INDEX IF EXISTS hash_1_index;
|
|
||||||
DROP INDEX IF EXISTS hash_2_index;
|
|
||||||
DROP INDEX IF EXISTS hash_3_index;
|
|
||||||
DROP INDEX IF EXISTS hash_4_index;
|
|
||||||
DROP INDEX IF EXISTS hash_5_index;
|
|
||||||
DROP INDEX IF EXISTS hash_6_index;
|
|
||||||
DROP INDEX IF EXISTS hash_7_index;
|
|
||||||
DROP INDEX IF EXISTS hash_8_index;
|
|
||||||
|
|
||||||
DROP INDEX IF EXISTS id_domain;
|
|
||||||
`)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (s *sqliteStorage) createIndexes() error {
|
|
||||||
_, err := s.db.Exec(`
|
|
||||||
|
|
||||||
CREATE INDEX IF NOT EXISTS hash_index ON Hashes (kind, hash);
|
|
||||||
CREATE INDEX IF NOT EXISTS hash_1_index ON Hashes ((hash >> (0 * 8) & 0xFF));
|
|
||||||
CREATE INDEX IF NOT EXISTS hash_2_index ON Hashes ((hash >> (1 * 8) & 0xFF));
|
|
||||||
CREATE INDEX IF NOT EXISTS hash_3_index ON Hashes ((hash >> (2 * 8) & 0xFF));
|
|
||||||
CREATE INDEX IF NOT EXISTS hash_4_index ON Hashes ((hash >> (3 * 8) & 0xFF));
|
|
||||||
CREATE INDEX IF NOT EXISTS hash_5_index ON Hashes ((hash >> (4 * 8) & 0xFF));
|
|
||||||
CREATE INDEX IF NOT EXISTS hash_6_index ON Hashes ((hash >> (5 * 8) & 0xFF));
|
|
||||||
CREATE INDEX IF NOT EXISTS hash_7_index ON Hashes ((hash >> (6 * 8) & 0xFF));
|
|
||||||
CREATE INDEX IF NOT EXISTS hash_8_index ON Hashes ((hash >> (7 * 8) & 0xFF));
|
|
||||||
|
|
||||||
CREATE INDEX IF NOT EXISTS id_domain ON IDs (domain, id);
|
|
||||||
PRAGMA shrink_memory;
|
|
||||||
ANALYZE;
|
|
||||||
`)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
var (
|
|
||||||
total time.Duration
|
|
||||||
t = time.Now()
|
|
||||||
)
|
|
||||||
|
|
||||||
func resetTime() {
|
|
||||||
total = 0
|
|
||||||
t = time.Now()
|
|
||||||
}
|
|
||||||
|
|
||||||
func logTime(log string) {
|
|
||||||
n := time.Now()
|
|
||||||
s := n.Sub(t)
|
|
||||||
t = n
|
|
||||||
total += s
|
|
||||||
fmt.Printf("total: %v, %s: %v\n", total, log, s)
|
|
||||||
}
|
|
||||||
|
|
||||||
func (s *sqliteStorage) GetMatches(hashes []Hash, max int, exactOnly bool) ([]Result, error) {
|
|
||||||
var (
|
|
||||||
foundMatches []Result
|
|
||||||
)
|
|
||||||
resetTime()
|
|
||||||
|
|
||||||
if exactOnly { // exact matches are also found by partial matches. Don't bother with exact matches so we don't have to de-duplicate
|
|
||||||
|
|
||||||
statement, err := s.db.Prepare(`SELECT rowid,hash,kind FROM Hashes WHERE ` + strings.TrimSuffix(strings.Repeat("(hash=? AND kind=?) OR", len(hashes)), "OR") + `ORDER BY kind,hash;`)
|
|
||||||
if err != nil {
|
|
||||||
logTime("Fail exact")
|
|
||||||
return foundMatches, err
|
|
||||||
}
|
|
||||||
|
|
||||||
args := make([]interface{}, 0, len(hashes)*2)
|
|
||||||
for _, hash := range hashes {
|
|
||||||
if hash.Hash != 0 {
|
|
||||||
args = append(args, int64(hash.Hash), hash.Kind)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
hashes, err := s.findExactHashes(statement, args...)
|
|
||||||
if err != nil {
|
|
||||||
return foundMatches, err
|
|
||||||
}
|
|
||||||
for _, hash := range hashes {
|
|
||||||
foundMatches = append(foundMatches, hash.Result)
|
|
||||||
}
|
|
||||||
|
|
||||||
// If we have exact matches don't bother with other matches
|
|
||||||
if len(foundMatches) > 0 && exactOnly {
|
|
||||||
return foundMatches, nil
|
|
||||||
}
|
|
||||||
logTime("Search Exact")
|
|
||||||
}
|
|
||||||
|
|
||||||
foundHashes := make(map[uint64]struct{})
|
|
||||||
|
|
||||||
for _, hash := range hashes {
|
|
||||||
hashes, err := s.findPartialHashes(max, int64(hash.Hash), hash.Kind)
|
|
||||||
if err != nil {
|
|
||||||
return foundMatches, err
|
|
||||||
}
|
|
||||||
logTime("Search partial " + hash.Kind.String())
|
|
||||||
|
|
||||||
for _, hash := range hashes {
|
|
||||||
if _, alreadyMatched := foundHashes[hash.Hash.Hash]; !alreadyMatched {
|
|
||||||
foundHashes[hash.Hash.Hash] = struct{}{}
|
|
||||||
foundMatches = append(foundMatches, hash.Result)
|
|
||||||
} else {
|
|
||||||
log.Println("Hash already found", hash)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return foundMatches, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (s *sqliteStorage) MapHashes(hash ImageHash) {
|
|
||||||
tx, err := s.db.BeginTx(context.Background(), nil)
|
|
||||||
if err != nil {
|
|
||||||
panic(err)
|
|
||||||
}
|
|
||||||
insertHashes, err := tx.Prepare(`
|
|
||||||
INSERT INTO Hashes (hash,kind) VALUES (?,?) ON CONFLICT DO UPDATE SET hash=?1 RETURNING hashid
|
|
||||||
`)
|
|
||||||
if err != nil {
|
|
||||||
panic(err)
|
|
||||||
}
|
|
||||||
rows, err := tx.Query(`
|
|
||||||
INSERT INTO IDs (domain,id) VALUES (?,?) ON CONFLICT DO UPDATE SET domain=?1 RETURNING idid
|
|
||||||
`, hash.ID.Domain, hash.ID.ID)
|
|
||||||
if err != nil {
|
|
||||||
panic(err)
|
|
||||||
}
|
|
||||||
if !rows.Next() {
|
|
||||||
panic("Unable to insert IDs")
|
|
||||||
}
|
|
||||||
var id_id int64
|
|
||||||
err = rows.Scan(&id_id)
|
|
||||||
if err != nil {
|
|
||||||
panic(err)
|
|
||||||
}
|
|
||||||
rows.Close()
|
|
||||||
hash_ids := []int64{}
|
|
||||||
for _, hash := range hash.Hashes {
|
|
||||||
rows, err := insertHashes.Query(int64(hash.Hash), hash.Kind)
|
|
||||||
if err != nil {
|
|
||||||
panic(err)
|
|
||||||
}
|
|
||||||
|
|
||||||
if !rows.Next() {
|
|
||||||
panic("Unable to insert IDs")
|
|
||||||
}
|
|
||||||
var id int64
|
|
||||||
err = rows.Scan(&id)
|
|
||||||
rows.Close()
|
|
||||||
if err != nil {
|
|
||||||
panic(err)
|
|
||||||
}
|
|
||||||
hash_ids = append(hash_ids, id)
|
|
||||||
}
|
|
||||||
var ids []any
|
|
||||||
for _, hash_id := range hash_ids {
|
|
||||||
ids = append(ids, hash_id, id_id)
|
|
||||||
}
|
|
||||||
_, err = tx.Exec(`INSERT INTO id_hash (hashid,idid) VALUES `+strings.TrimSuffix(strings.Repeat("(?, ?),", len(hash_ids)), ",")+` ON CONFLICT DO NOTHING;`, ids...)
|
|
||||||
if err != nil {
|
|
||||||
panic(fmt.Errorf("Failed inserting: %v,%v: %w", hash.ID.Domain, hash.ID.ID, err))
|
|
||||||
}
|
|
||||||
err = tx.Commit()
|
|
||||||
if err != nil {
|
|
||||||
panic(err)
|
|
||||||
}
|
|
||||||
insertHashes.Close()
|
|
||||||
}
|
|
||||||
|
|
||||||
func (s *sqliteStorage) DecodeHashes(hashes SavedHashes) error {
|
|
||||||
err := s.dropIndexes()
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
for hashType, sourceHashes := range hashes.Hashes {
|
|
||||||
hashKind := goimagehash.Kind(hashType + 1)
|
|
||||||
for hash, idsLocations := range sourceHashes {
|
|
||||||
for _, id := range hashes.IDs[idsLocations] {
|
|
||||||
s.MapHashes(ImageHash{
|
|
||||||
Hashes: []Hash{{hash, hashKind}},
|
|
||||||
ID: id,
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
err = s.createIndexes()
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (s *sqliteStorage) EncodeHashes() (SavedHashes, error) {
|
|
||||||
hashes := SavedHashes{}
|
|
||||||
conn, err := s.db.Conn(context.Background())
|
|
||||||
if err != nil {
|
|
||||||
return hashes, err
|
|
||||||
}
|
|
||||||
defer conn.Close()
|
|
||||||
rows, err := conn.QueryContext(context.Background(), "SELECT IDs.domain,IDs.id,Hashes.hash,Hashes.kind FROM Hashes JOIN id_hash ON id_hash.hashid = hashes.rowid JOIN IDs ON IDs.rowid = id_hash.idid ORDER BY IDs.ID,Hashes.kind,Hashes.hash;")
|
|
||||||
if err != nil {
|
|
||||||
rows.Close()
|
|
||||||
return hashes, err
|
|
||||||
}
|
|
||||||
var (
|
|
||||||
id ID
|
|
||||||
hash Hash
|
|
||||||
)
|
|
||||||
err = rows.Scan(&id.Domain, &id.ID, &hash.Hash, &hash.Kind)
|
|
||||||
if err != nil {
|
|
||||||
return hashes, err
|
|
||||||
}
|
|
||||||
hashes.InsertHash(hash, id)
|
|
||||||
|
|
||||||
return hashes, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (s *sqliteStorage) AssociateIDs(newIDs []NewIDs) error {
|
|
||||||
for _, ids := range newIDs {
|
|
||||||
var oldIDID, newIDID int
|
|
||||||
_, err := s.db.Exec(`INSERT INTO IDs domain,id VALUES (?,?)`, ids.NewID.Domain, ids.NewID.ID)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
rows, err := s.db.Query(`SELECT idid FROM IDs WHERE domain=? AND id=?`, ids.NewID.Domain, ids.NewID.ID)
|
|
||||||
if err != nil && !errors.Is(err, sql.ErrNoRows) {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
if rows.Next() {
|
|
||||||
rows.Scan(&newIDID)
|
|
||||||
} else {
|
|
||||||
return errors.New("Unable to insert New ID into database")
|
|
||||||
}
|
|
||||||
rows.Close()
|
|
||||||
rows, err = s.db.Query(`SELECT idid FROM IDs WHERE domain=? AND id=?`, ids.OldID.Domain, ids.OldID.ID)
|
|
||||||
if err != nil && !errors.Is(err, sql.ErrNoRows) {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
if rows.Next() {
|
|
||||||
rows.Scan(&oldIDID)
|
|
||||||
} else {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
_, err = s.db.Exec(`INSERT INTO id_hash (hashid, id_id) SELECT hashid,? FROM id_hash where id_id=?`, newIDID, oldIDID)
|
|
||||||
if err != nil {
|
|
||||||
return fmt.Errorf("Unable to associate IDs: %w", err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (s *sqliteStorage) GetIDs(id ID) IDList {
|
|
||||||
var idid int
|
|
||||||
rows, err := s.db.Query(`SELECT idid FROM IDs WHERE domain=? AND id=?`, id.Domain, id.ID)
|
|
||||||
if err != nil && !errors.Is(err, sql.ErrNoRows) {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
if rows.Next() {
|
|
||||||
rows.Scan(&idid)
|
|
||||||
} else {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
rows, err = s.db.Query(`SELECT id_hash FROM id_hash WHERE id_id=?`, idid)
|
|
||||||
if err != nil && !errors.Is(err, sql.ErrNoRows) {
|
|
||||||
panic(err)
|
|
||||||
}
|
|
||||||
var hashIDs []interface{}
|
|
||||||
for rows.Next() {
|
|
||||||
var hashID int
|
|
||||||
rows.Scan(&hashID)
|
|
||||||
hashIDs = append(hashIDs, hashID)
|
|
||||||
}
|
|
||||||
rows.Close()
|
|
||||||
|
|
||||||
IDs := make(IDList)
|
|
||||||
rows, err = s.db.Query(`SELECT IDs.domain,IDs.id FROM id_hash JOIN IDs ON id_hash.idid==IDs.idid WHERE hash_id in (`+strings.TrimRight(strings.Repeat("?,", len(hashIDs)), ",")+`)`, hashIDs...)
|
|
||||||
if err != nil && !errors.Is(err, sql.ErrNoRows) {
|
|
||||||
panic(err)
|
|
||||||
}
|
|
||||||
for rows.Next() {
|
|
||||||
var id ID
|
|
||||||
rows.Scan(&id.Domain, id.ID)
|
|
||||||
IDs[id.Domain] = append(IDs[id.Domain], id.ID)
|
|
||||||
}
|
|
||||||
return IDs
|
|
||||||
}
|
|
||||||
|
|
||||||
func NewSqliteStorage(db, path string) (HashStorage, error) {
|
|
||||||
sqlite := &sqliteStorage{}
|
|
||||||
sqlDB, err := sql.Open(db, fmt.Sprintf("file://%s?_pragma=cache_size(-200000)&_pragma=busy_timeout(500)&_pragma=hard_heap_limit(1073741824)&_pragma=journal_mode(wal)&_pragma=soft_heap_limit(314572800)", path))
|
|
||||||
if err != nil {
|
|
||||||
panic(err)
|
|
||||||
}
|
|
||||||
sqlite.db = sqlDB
|
|
||||||
_, err = sqlite.db.Exec(`
|
|
||||||
PRAGMA foreign_keys=ON;
|
|
||||||
CREATE TABLE IF NOT EXISTS Hashes(
|
|
||||||
hashid INTEGER PRIMARY KEY,
|
|
||||||
hash INT NOT NULL,
|
|
||||||
kind int NOT NULL,
|
|
||||||
UNIQUE(kind, hash)
|
|
||||||
);
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS IDs(
|
|
||||||
id TEXT NOT NULL,
|
|
||||||
domain TEXT NOT NULL,
|
|
||||||
idid INTEGER PRIMARY KEY,
|
|
||||||
UNIQUE (domain, id)
|
|
||||||
);
|
|
||||||
CREATE INDEX IF NOT EXISTS id_domain ON IDs (domain, id);
|
|
||||||
|
|
||||||
CREATE TABLE IF NOT EXISTS id_hash(
|
|
||||||
hashid INTEGER,
|
|
||||||
idid INTEGER,
|
|
||||||
FOREIGN KEY(hashid) REFERENCES Hashes(hashid),
|
|
||||||
FOREIGN KEY(idid) REFERENCES IDs(idid)
|
|
||||||
UNIQUE (hashid, idid)
|
|
||||||
);
|
|
||||||
|
|
||||||
`)
|
|
||||||
if err != nil {
|
|
||||||
panic(err)
|
|
||||||
}
|
|
||||||
sqlite.createIndexes()
|
|
||||||
sqlite.db.SetMaxOpenConns(1)
|
|
||||||
return sqlite, nil
|
|
||||||
}
|
|
372
storage/basicmap.go
Normal file
372
storage/basicmap.go
Normal file
@ -0,0 +1,372 @@
|
|||||||
|
package storage
|
||||||
|
|
||||||
|
import (
|
||||||
|
"cmp"
|
||||||
|
"errors"
|
||||||
|
"fmt"
|
||||||
|
"math/bits"
|
||||||
|
"slices"
|
||||||
|
"sync"
|
||||||
|
|
||||||
|
ch "gitea.narnian.us/lordwelch/comic-hasher"
|
||||||
|
"gitea.narnian.us/lordwelch/goimagehash"
|
||||||
|
)
|
||||||
|
|
||||||
|
type basicMapStorage struct {
|
||||||
|
hashMutex *sync.RWMutex
|
||||||
|
|
||||||
|
ids IDMap
|
||||||
|
aHashes []ch.SavedHash
|
||||||
|
dHashes []ch.SavedHash
|
||||||
|
pHashes []ch.SavedHash
|
||||||
|
}
|
||||||
|
type IDs struct {
|
||||||
|
id *ch.ID
|
||||||
|
idList *[]*ch.ID
|
||||||
|
}
|
||||||
|
type IDMap struct {
|
||||||
|
ids []IDs
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m *IDMap) InsertID(id *ch.ID) *ch.ID {
|
||||||
|
return m.insertID(id, &[]*ch.ID{id})
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m *IDMap) insertID(id *ch.ID, idList *[]*ch.ID) *ch.ID {
|
||||||
|
index, found := slices.BinarySearchFunc(m.ids, id, func(id IDs, target *ch.ID) int {
|
||||||
|
return id.id.Compare(*target)
|
||||||
|
})
|
||||||
|
if !found {
|
||||||
|
m.ids = slices.Insert(m.ids, index, IDs{
|
||||||
|
id,
|
||||||
|
idList,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
return m.ids[index].id
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m *IDMap) sort() {
|
||||||
|
slices.SortFunc(m.ids, func(a, b IDs) int {
|
||||||
|
return a.id.Compare(*b.id)
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m *IDMap) FindID(id *ch.ID) (int, bool) {
|
||||||
|
return slices.BinarySearchFunc(m.ids, id, func(id IDs, target *ch.ID) int {
|
||||||
|
return id.id.Compare(*target)
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m *IDMap) GetIDs(id *ch.ID) []ch.ID {
|
||||||
|
index, found := m.FindID(id)
|
||||||
|
|
||||||
|
if !found {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
ids := make([]ch.ID, 0, len(*m.ids[index].idList))
|
||||||
|
for _, id := range *m.ids[index].idList {
|
||||||
|
ids = append(ids, *id)
|
||||||
|
}
|
||||||
|
return ids
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m *IDMap) AssociateIDs(newids []ch.NewIDs) error {
|
||||||
|
for _, newid := range newids {
|
||||||
|
index, found := m.FindID(&newid.OldID)
|
||||||
|
if !found {
|
||||||
|
return ErrIDNotFound
|
||||||
|
}
|
||||||
|
*(m.ids[index].idList) = ch.InsertIDp(*(m.ids[index].idList), &newid.NewID)
|
||||||
|
m.insertID(&newid.NewID, m.ids[index].idList)
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// func (m *IDMap) NewID(domain Source, id string) *ch.ID {
|
||||||
|
// newID := ch.ID{domain, id}
|
||||||
|
// index, found := slices.BinarySearchFunc(m.idList, newID, func(id *ch.ID, target ch.ID) int {
|
||||||
|
// return id.Compare(*target)
|
||||||
|
// })
|
||||||
|
// if !found {
|
||||||
|
// m.idList = slices.Insert(m.idList, index, &newID)
|
||||||
|
// }
|
||||||
|
// return m.idList[index]
|
||||||
|
// }
|
||||||
|
|
||||||
|
var ErrIDNotFound = errors.New("ID not found on this server")
|
||||||
|
|
||||||
|
// atleast must have a read lock before using
|
||||||
|
func (b *basicMapStorage) atleast(kind goimagehash.Kind, maxDistance int, searchHash uint64) []ch.Result {
|
||||||
|
matchingHashes := make([]ch.Result, 0, 20) // hope that we don't need more
|
||||||
|
|
||||||
|
mappedIds := map[int]bool{}
|
||||||
|
storedHash := ch.SavedHash{} // reduces allocations and ensures queries are <1s
|
||||||
|
for _, storedHash = range *b.getCurrentHashes(kind) {
|
||||||
|
distance := bits.OnesCount64(searchHash ^ storedHash.Hash.Hash)
|
||||||
|
if distance <= maxDistance {
|
||||||
|
index, _ := b.ids.FindID(&storedHash.ID)
|
||||||
|
if mappedIds[index] {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
mappedIds[index] = true
|
||||||
|
matchingHashes = append(matchingHashes, ch.Result{
|
||||||
|
Hash: storedHash.Hash,
|
||||||
|
ID: storedHash.ID,
|
||||||
|
Distance: distance,
|
||||||
|
EquivalentIDs: b.ids.GetIDs(&storedHash.ID),
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return matchingHashes
|
||||||
|
}
|
||||||
|
|
||||||
|
func (b *basicMapStorage) exactMatches(hashes []ch.Hash, max int) []ch.Result {
|
||||||
|
var foundMatches []ch.Result
|
||||||
|
for _, hash := range hashes {
|
||||||
|
mappedIds := map[int]bool{}
|
||||||
|
|
||||||
|
index, count := b.findHash(hash)
|
||||||
|
if count > 0 {
|
||||||
|
for _, storedHash := range (*b.getCurrentHashes(hash.Kind))[index : index+count] {
|
||||||
|
index, _ := b.ids.FindID(&storedHash.ID)
|
||||||
|
if mappedIds[index] {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
mappedIds[index] = true
|
||||||
|
|
||||||
|
foundMatches = append(foundMatches, ch.Result{
|
||||||
|
Hash: storedHash.Hash,
|
||||||
|
ID: storedHash.ID,
|
||||||
|
Distance: 0,
|
||||||
|
EquivalentIDs: b.ids.GetIDs(&storedHash.ID),
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
return foundMatches
|
||||||
|
}
|
||||||
|
|
||||||
|
func (b *basicMapStorage) GetMatches(hashes []ch.Hash, max int, exactOnly bool) ([]ch.Result, error) {
|
||||||
|
var (
|
||||||
|
foundMatches []ch.Result
|
||||||
|
tl ch.TimeLog
|
||||||
|
)
|
||||||
|
tl.ResetTime()
|
||||||
|
defer tl.LogTime(fmt.Sprintf("Search Complete: max: %v ExactOnly: %v", max, exactOnly))
|
||||||
|
b.hashMutex.RLock()
|
||||||
|
defer b.hashMutex.RUnlock()
|
||||||
|
|
||||||
|
if exactOnly { // exact matches are also found by partial matches. Don't bother with exact matches so we don't have to de-duplicate
|
||||||
|
foundMatches = b.exactMatches(hashes, max)
|
||||||
|
|
||||||
|
tl.LogTime("Search Exact")
|
||||||
|
if len(foundMatches) > 0 {
|
||||||
|
return foundMatches, nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
foundHashes := make(map[uint64]struct{})
|
||||||
|
totalPartialHashes := 0
|
||||||
|
|
||||||
|
for _, hash := range hashes {
|
||||||
|
foundMatches = append(foundMatches, b.atleast(hash.Kind, max, hash.Hash)...)
|
||||||
|
|
||||||
|
}
|
||||||
|
fmt.Println("Total partial hashes tested:", totalPartialHashes, len(foundHashes))
|
||||||
|
return foundMatches, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// getCurrentHashes must have a read lock before using
|
||||||
|
func (b *basicMapStorage) getCurrentHashes(kind goimagehash.Kind) *[]ch.SavedHash {
|
||||||
|
if kind == goimagehash.AHash {
|
||||||
|
return &b.aHashes
|
||||||
|
}
|
||||||
|
if kind == goimagehash.DHash {
|
||||||
|
return &b.dHashes
|
||||||
|
}
|
||||||
|
if kind == goimagehash.PHash {
|
||||||
|
return &b.pHashes
|
||||||
|
}
|
||||||
|
panic("Unknown hash type: " + kind.String())
|
||||||
|
}
|
||||||
|
|
||||||
|
// findHash must have a read lock before using
|
||||||
|
// return value is index, count
|
||||||
|
// if count < 1 then no results were found
|
||||||
|
func (b *basicMapStorage) findHash(hash ch.Hash) (int, int) {
|
||||||
|
currentHashes := *b.getCurrentHashes(hash.Kind)
|
||||||
|
index, found := slices.BinarySearchFunc(currentHashes, hash, func(existing ch.SavedHash, target ch.Hash) int {
|
||||||
|
return cmp.Compare(existing.Hash.Hash, target.Hash)
|
||||||
|
})
|
||||||
|
if !found {
|
||||||
|
return index, 0
|
||||||
|
}
|
||||||
|
count := 0
|
||||||
|
for i := index + 1; i < len(currentHashes) && currentHashes[i].Hash.Hash == hash.Hash; i++ {
|
||||||
|
count++
|
||||||
|
}
|
||||||
|
return index, count
|
||||||
|
}
|
||||||
|
|
||||||
|
// insertHash must already have a lock
|
||||||
|
func (b *basicMapStorage) insertHash(hash ch.Hash, id ch.ID) {
|
||||||
|
currentHashes := b.getCurrentHashes(hash.Kind)
|
||||||
|
index, count := b.findHash(hash)
|
||||||
|
max := index + count
|
||||||
|
for ; index < max; index++ {
|
||||||
|
if (*currentHashes)[index].ID == id {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
sh := ch.SavedHash{
|
||||||
|
Hash: hash,
|
||||||
|
ID: id,
|
||||||
|
}
|
||||||
|
*currentHashes = slices.Insert(*currentHashes, index, sh)
|
||||||
|
b.ids.InsertID(&sh.ID)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (b *basicMapStorage) MapHashes(hash ch.ImageHash) {
|
||||||
|
b.hashMutex.Lock()
|
||||||
|
defer b.hashMutex.Unlock()
|
||||||
|
for _, ih := range hash.Hashes {
|
||||||
|
b.insertHash(ih, hash.ID)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// DecodeHashes must already have a lock
|
||||||
|
func (b *basicMapStorage) DecodeHashes(hashes *ch.SavedHashes) error {
|
||||||
|
if hashes == nil {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
b.ids.ids = make([]IDs, 0, len(hashes.Hashes))
|
||||||
|
|
||||||
|
// Initialize all the known equal IDs
|
||||||
|
for _, ids := range hashes.IDs {
|
||||||
|
new_ids := make([]*ch.ID, 0, len(ids))
|
||||||
|
for _, id := range ids {
|
||||||
|
new_ids = append(new_ids, &id)
|
||||||
|
}
|
||||||
|
for _, id := range new_ids {
|
||||||
|
b.ids.ids = append(b.ids.ids, IDs{
|
||||||
|
id,
|
||||||
|
&new_ids,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
b.ids.sort()
|
||||||
|
|
||||||
|
slices.SortFunc(hashes.Hashes, func(existing, target ch.SavedHash) int {
|
||||||
|
return cmp.Or(
|
||||||
|
cmp.Compare(*existing.ID.Domain, *target.ID.Domain), // Sorted for id insertion efficiency
|
||||||
|
cmp.Compare(existing.ID.ID, target.ID.ID), // Sorted for id insertion efficiency
|
||||||
|
cmp.Compare(existing.Hash.Kind, target.Hash.Kind),
|
||||||
|
cmp.Compare(existing.Hash.Hash, target.Hash.Hash),
|
||||||
|
)
|
||||||
|
})
|
||||||
|
aHashCount := 0
|
||||||
|
dHashCount := 0
|
||||||
|
pHashCount := 0
|
||||||
|
for _, savedHash := range hashes.Hashes {
|
||||||
|
|
||||||
|
if savedHash.Hash.Kind == goimagehash.AHash {
|
||||||
|
aHashCount += 1
|
||||||
|
}
|
||||||
|
if savedHash.Hash.Kind == goimagehash.DHash {
|
||||||
|
dHashCount += 1
|
||||||
|
}
|
||||||
|
if savedHash.Hash.Kind == goimagehash.PHash {
|
||||||
|
pHashCount += 1
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Assume they are probably fairly equally split between hash types
|
||||||
|
b.aHashes = make([]ch.SavedHash, 0, aHashCount)
|
||||||
|
b.dHashes = make([]ch.SavedHash, 0, dHashCount)
|
||||||
|
b.pHashes = make([]ch.SavedHash, 0, pHashCount)
|
||||||
|
for i := range hashes.Hashes {
|
||||||
|
hash := hashes.Hashes[i].Clone() // Not cloning this will keep strings/slices loaded from json wasting memory
|
||||||
|
if hashes.Hashes[i].Hash.Kind == goimagehash.AHash {
|
||||||
|
b.aHashes = append(b.aHashes, hash)
|
||||||
|
}
|
||||||
|
if hashes.Hashes[i].Hash.Kind == goimagehash.DHash {
|
||||||
|
b.dHashes = append(b.dHashes, hash)
|
||||||
|
}
|
||||||
|
if hashes.Hashes[i].Hash.Kind == goimagehash.PHash {
|
||||||
|
b.pHashes = append(b.pHashes, hash)
|
||||||
|
}
|
||||||
|
|
||||||
|
if hashes.Hashes[i].ID == (ch.ID{}) {
|
||||||
|
fmt.Println("Empty ID detected")
|
||||||
|
panic(hashes.Hashes[i])
|
||||||
|
}
|
||||||
|
// TODO: Make loading this more efficient
|
||||||
|
// All known equal IDs are already mapped we can add any missing ones from hashes
|
||||||
|
b.ids.InsertID(&hash.ID)
|
||||||
|
}
|
||||||
|
|
||||||
|
hashCmp := func(existing, target ch.SavedHash) int {
|
||||||
|
return cmp.Or(
|
||||||
|
cmp.Compare(existing.Hash.Hash, target.Hash.Hash),
|
||||||
|
cmp.Compare(*existing.ID.Domain, *target.ID.Domain),
|
||||||
|
cmp.Compare(existing.ID.ID, target.ID.ID),
|
||||||
|
)
|
||||||
|
}
|
||||||
|
slices.SortFunc(b.aHashes, hashCmp)
|
||||||
|
slices.SortFunc(b.dHashes, hashCmp)
|
||||||
|
slices.SortFunc(b.pHashes, hashCmp)
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// EncodeHashes should already have a lock
|
||||||
|
func (b *basicMapStorage) EncodeHashes() (*ch.SavedHashes, error) {
|
||||||
|
savedHashes := ch.SavedHashes{
|
||||||
|
Hashes: make([]ch.SavedHash, 0, len(b.aHashes)+len(b.dHashes)+len(b.pHashes)),
|
||||||
|
}
|
||||||
|
// savedHashes.Hashes = append(savedHashes.Hashes, b.aHashes...)
|
||||||
|
// savedHashes.Hashes = append(savedHashes.Hashes, b.dHashes...)
|
||||||
|
// savedHashes.Hashes = append(savedHashes.Hashes, b.pHashes...)
|
||||||
|
|
||||||
|
// // Only keep groups len>1 as they are mapped in SavedHashes.Hashes
|
||||||
|
// for _, ids := range b.ids.ids {
|
||||||
|
// if len(*ids.idList) > 1 {
|
||||||
|
// idl := make([]ID, 0, len(*ids.idList))
|
||||||
|
// for _, id := range *ids.idList {
|
||||||
|
// idl = append(idl, *id)
|
||||||
|
// }
|
||||||
|
|
||||||
|
// savedHashes.IDs = append(savedHashes.IDs, idl)
|
||||||
|
// }
|
||||||
|
// }
|
||||||
|
|
||||||
|
return &savedHashes, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (b *basicMapStorage) AssociateIDs(newids []ch.NewIDs) error {
|
||||||
|
b.hashMutex.RLock()
|
||||||
|
defer b.hashMutex.RUnlock()
|
||||||
|
return b.ids.AssociateIDs(newids)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (b *basicMapStorage) GetIDs(id ch.ID) ch.IDList {
|
||||||
|
b.hashMutex.RLock()
|
||||||
|
defer b.hashMutex.RUnlock()
|
||||||
|
ids := b.ids.GetIDs(&id)
|
||||||
|
return ch.ToIDList(ids)
|
||||||
|
}
|
||||||
|
|
||||||
|
func NewBasicMapStorage() (ch.HashStorage, error) {
|
||||||
|
storage := &basicMapStorage{
|
||||||
|
hashMutex: &sync.RWMutex{},
|
||||||
|
ids: IDMap{
|
||||||
|
ids: []IDs{},
|
||||||
|
},
|
||||||
|
aHashes: []ch.SavedHash{},
|
||||||
|
dHashes: []ch.SavedHash{},
|
||||||
|
pHashes: []ch.SavedHash{},
|
||||||
|
}
|
||||||
|
return storage, nil
|
||||||
|
}
|
171
storage/map.go
Normal file
171
storage/map.go
Normal file
@ -0,0 +1,171 @@
|
|||||||
|
package storage
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"slices"
|
||||||
|
"sync"
|
||||||
|
|
||||||
|
ch "gitea.narnian.us/lordwelch/comic-hasher"
|
||||||
|
"gitea.narnian.us/lordwelch/goimagehash"
|
||||||
|
)
|
||||||
|
|
||||||
|
type MapStorage struct {
|
||||||
|
basicMapStorage
|
||||||
|
partialAHash [8]map[uint8][]uint64
|
||||||
|
partialDHash [8]map[uint8][]uint64
|
||||||
|
partialPHash [8]map[uint8][]uint64
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m *MapStorage) GetMatches(hashes []ch.Hash, max int, exactOnly bool) ([]ch.Result, error) {
|
||||||
|
var (
|
||||||
|
foundMatches []ch.Result
|
||||||
|
tl ch.TimeLog
|
||||||
|
)
|
||||||
|
m.hashMutex.RLock()
|
||||||
|
defer m.hashMutex.RUnlock()
|
||||||
|
|
||||||
|
if exactOnly { // exact matches are also found by partial matches. Don't bother with exact matches so we don't have to de-duplicate
|
||||||
|
foundMatches = m.exactMatches(hashes, max)
|
||||||
|
|
||||||
|
tl.LogTime("Search Exact")
|
||||||
|
if len(foundMatches) > 0 {
|
||||||
|
return foundMatches, nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
tl.ResetTime()
|
||||||
|
defer tl.LogTime("Search Complete")
|
||||||
|
|
||||||
|
totalPartialHashes := 0
|
||||||
|
|
||||||
|
for _, searchHash := range hashes {
|
||||||
|
currentHashes, currentPartialHashes := m.getCurrentHashes(searchHash.Kind)
|
||||||
|
potentialMatches := []uint64{}
|
||||||
|
|
||||||
|
for i, partialHash := range ch.SplitHash(searchHash.Hash) {
|
||||||
|
potentialMatches = append(potentialMatches, currentPartialHashes[i][partialHash]...)
|
||||||
|
}
|
||||||
|
|
||||||
|
totalPartialHashes += len(potentialMatches)
|
||||||
|
mappedIds := map[int]bool{}
|
||||||
|
|
||||||
|
for _, match := range ch.Atleast(max, searchHash.Hash, potentialMatches) {
|
||||||
|
matchedHash := ch.Hash{
|
||||||
|
Hash: match.Hash,
|
||||||
|
Kind: searchHash.Kind,
|
||||||
|
}
|
||||||
|
index, count := m.findHash(matchedHash)
|
||||||
|
if count < 1 {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
for _, storedHash := range currentHashes[index : index+count] {
|
||||||
|
idIndex, _ := m.ids.FindID(&storedHash.ID)
|
||||||
|
if mappedIds[idIndex] {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
mappedIds[idIndex] = true
|
||||||
|
|
||||||
|
foundMatches = append(foundMatches, ch.Result{
|
||||||
|
Hash: storedHash.Hash,
|
||||||
|
ID: storedHash.ID,
|
||||||
|
Distance: 0,
|
||||||
|
EquivalentIDs: m.ids.GetIDs(&storedHash.ID),
|
||||||
|
})
|
||||||
|
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
fmt.Println("Total partial hashes tested:", totalPartialHashes)
|
||||||
|
return foundMatches, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// getCurrentHashes must have a read lock before using
|
||||||
|
func (m *MapStorage) getCurrentHashes(kind goimagehash.Kind) ([]ch.SavedHash, [8]map[uint8][]uint64) {
|
||||||
|
if kind == goimagehash.AHash {
|
||||||
|
return m.aHashes, m.partialAHash
|
||||||
|
}
|
||||||
|
if kind == goimagehash.DHash {
|
||||||
|
return m.dHashes, m.partialDHash
|
||||||
|
}
|
||||||
|
if kind == goimagehash.PHash {
|
||||||
|
return m.pHashes, m.partialPHash
|
||||||
|
}
|
||||||
|
panic("Unknown hash type: " + kind.String())
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m *MapStorage) MapHashes(hash ch.ImageHash) {
|
||||||
|
m.basicMapStorage.MapHashes(hash)
|
||||||
|
for _, hash := range hash.Hashes {
|
||||||
|
_, partialHashes := m.getCurrentHashes(hash.Kind)
|
||||||
|
for i, partialHash := range ch.SplitHash(hash.Hash) {
|
||||||
|
partialHashes[i][partialHash] = ch.Insert(partialHashes[i][partialHash], hash.Hash)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m *MapStorage) DecodeHashes(hashes *ch.SavedHashes) error {
|
||||||
|
if hashes == nil {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
if err := m.basicMapStorage.DecodeHashes(hashes); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
mapPartialHashes(m.aHashes, m.partialAHash)
|
||||||
|
mapPartialHashes(m.dHashes, m.partialDHash)
|
||||||
|
mapPartialHashes(m.pHashes, m.partialPHash)
|
||||||
|
|
||||||
|
compactPartialHashes(m.partialAHash)
|
||||||
|
compactPartialHashes(m.partialDHash)
|
||||||
|
compactPartialHashes(m.partialPHash)
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func NewMapStorage() (ch.HashStorage, error) {
|
||||||
|
|
||||||
|
storage := &MapStorage{
|
||||||
|
basicMapStorage: basicMapStorage{
|
||||||
|
hashMutex: &sync.RWMutex{},
|
||||||
|
ids: IDMap{
|
||||||
|
ids: []IDs{},
|
||||||
|
},
|
||||||
|
aHashes: []ch.SavedHash{},
|
||||||
|
dHashes: []ch.SavedHash{},
|
||||||
|
pHashes: []ch.SavedHash{},
|
||||||
|
},
|
||||||
|
partialAHash: newPartialHash(),
|
||||||
|
partialDHash: newPartialHash(),
|
||||||
|
partialPHash: newPartialHash(),
|
||||||
|
}
|
||||||
|
return storage, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func newPartialHash() [8]map[uint8][]uint64 {
|
||||||
|
return [8]map[uint8][]uint64{
|
||||||
|
map[uint8][]uint64{},
|
||||||
|
map[uint8][]uint64{},
|
||||||
|
map[uint8][]uint64{},
|
||||||
|
map[uint8][]uint64{},
|
||||||
|
map[uint8][]uint64{},
|
||||||
|
map[uint8][]uint64{},
|
||||||
|
map[uint8][]uint64{},
|
||||||
|
map[uint8][]uint64{},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func mapPartialHashes(hashes []ch.SavedHash, partialHashMap [8]map[uint8][]uint64) {
|
||||||
|
for _, savedHash := range hashes {
|
||||||
|
for i, partialHash := range ch.SplitHash(savedHash.Hash.Hash) {
|
||||||
|
partialHashMap[i][partialHash] = append(partialHashMap[i][partialHash], savedHash.Hash.Hash)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func compactPartialHashes(partialHashMap [8]map[uint8][]uint64) {
|
||||||
|
for _, partMap := range partialHashMap {
|
||||||
|
for part, hashes := range partMap {
|
||||||
|
slices.Sort(hashes)
|
||||||
|
partMap[part] = slices.Compact(hashes)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
530
storage/sqlite.go
Normal file
530
storage/sqlite.go
Normal file
@ -0,0 +1,530 @@
|
|||||||
|
package storage
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"database/sql"
|
||||||
|
"errors"
|
||||||
|
"fmt"
|
||||||
|
"log"
|
||||||
|
"math/bits"
|
||||||
|
|
||||||
|
ch "gitea.narnian.us/lordwelch/comic-hasher"
|
||||||
|
_ "modernc.org/sqlite"
|
||||||
|
)
|
||||||
|
|
||||||
|
type sqliteStorage struct {
|
||||||
|
db *sql.DB
|
||||||
|
|
||||||
|
hashExactMatchStatement *sql.Stmt
|
||||||
|
hashPartialMatchStatement *sql.Stmt
|
||||||
|
|
||||||
|
idMatchStatement *sql.Stmt
|
||||||
|
|
||||||
|
insertHash *sql.Stmt
|
||||||
|
insertID *sql.Stmt
|
||||||
|
insertEID *sql.Stmt
|
||||||
|
insertIEID *sql.Stmt
|
||||||
|
idExists *sql.Stmt
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *sqliteStorage) findExactHashes(statement *sql.Stmt, hash ch.Hash) (map[ch.ID][]ch.ID, error) {
|
||||||
|
if statement == nil {
|
||||||
|
statement = s.hashExactMatchStatement
|
||||||
|
}
|
||||||
|
hashes := map[ch.ID][]ch.ID{}
|
||||||
|
rows, err := statement.Query(hash.Kind, int64(hash.Hash))
|
||||||
|
if err != nil {
|
||||||
|
return hashes, err
|
||||||
|
}
|
||||||
|
for rows.Next() {
|
||||||
|
var (
|
||||||
|
id ch.ID
|
||||||
|
foundID ch.ID
|
||||||
|
)
|
||||||
|
err = rows.Scan(&foundID.Domain, &foundID.ID, &id.Domain, &id.ID)
|
||||||
|
if err != nil {
|
||||||
|
rows.Close()
|
||||||
|
return hashes, err
|
||||||
|
}
|
||||||
|
hashes[foundID] = append(hashes[foundID], id)
|
||||||
|
}
|
||||||
|
rows.Close()
|
||||||
|
return hashes, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *sqliteStorage) findPartialHashes(tl ch.TimeLog, statement *sql.Stmt, max int, hash ch.Hash) ([]ch.Result, error) {
|
||||||
|
if statement == nil {
|
||||||
|
statement = s.hashPartialMatchStatement
|
||||||
|
}
|
||||||
|
hashResults := []ch.Result{}
|
||||||
|
rows, err := statement.Query(hash.Kind, int64(hash.Hash))
|
||||||
|
if err != nil {
|
||||||
|
return hashResults, err
|
||||||
|
}
|
||||||
|
|
||||||
|
results := map[ch.SavedHash][]ch.ID{}
|
||||||
|
for rows.Next() {
|
||||||
|
var (
|
||||||
|
tmpHash int64
|
||||||
|
sqlHash = ch.SavedHash{
|
||||||
|
Hash: ch.Hash{Kind: hash.Kind},
|
||||||
|
}
|
||||||
|
id ch.ID
|
||||||
|
)
|
||||||
|
err = rows.Scan(&sqlHash.ID.Domain, &sqlHash.ID.ID, &tmpHash, &id.Domain, &id.ID)
|
||||||
|
if err != nil {
|
||||||
|
rows.Close()
|
||||||
|
return hashResults, err
|
||||||
|
}
|
||||||
|
sqlHash.Hash.Hash = uint64(tmpHash)
|
||||||
|
results[sqlHash] = append(results[sqlHash], id)
|
||||||
|
}
|
||||||
|
for sqlHash, ids := range results {
|
||||||
|
res := ch.Result{
|
||||||
|
Hash: sqlHash.Hash,
|
||||||
|
ID: sqlHash.ID,
|
||||||
|
Distance: bits.OnesCount64(hash.Hash ^ sqlHash.Hash.Hash),
|
||||||
|
EquivalentIDs: ids,
|
||||||
|
}
|
||||||
|
if res.Distance <= max {
|
||||||
|
hashResults = append(hashResults, res)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return hashResults, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *sqliteStorage) dropIndexes() error {
|
||||||
|
_, err := s.db.Exec(`
|
||||||
|
DROP INDEX IF EXISTS hash_index;
|
||||||
|
DROP INDEX IF EXISTS hash_1_index;
|
||||||
|
DROP INDEX IF EXISTS hash_2_index;
|
||||||
|
DROP INDEX IF EXISTS hash_3_index;
|
||||||
|
DROP INDEX IF EXISTS hash_4_index;
|
||||||
|
DROP INDEX IF EXISTS hash_5_index;
|
||||||
|
DROP INDEX IF EXISTS hash_6_index;
|
||||||
|
DROP INDEX IF EXISTS hash_7_index;
|
||||||
|
DROP INDEX IF EXISTS hash_8_index;
|
||||||
|
|
||||||
|
DROP INDEX IF EXISTS id_domain;
|
||||||
|
`)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *sqliteStorage) createIndexes() error {
|
||||||
|
_, err := s.db.Exec(`
|
||||||
|
CREATE INDEX IF NOT EXISTS hash_index ON Hashes (kind, hash);
|
||||||
|
CREATE INDEX IF NOT EXISTS hash_1_index ON Hashes ((hash >> (0 * 8) & 0xFF));
|
||||||
|
CREATE INDEX IF NOT EXISTS hash_2_index ON Hashes ((hash >> (1 * 8) & 0xFF));
|
||||||
|
CREATE INDEX IF NOT EXISTS hash_3_index ON Hashes ((hash >> (2 * 8) & 0xFF));
|
||||||
|
CREATE INDEX IF NOT EXISTS hash_4_index ON Hashes ((hash >> (3 * 8) & 0xFF));
|
||||||
|
CREATE INDEX IF NOT EXISTS hash_5_index ON Hashes ((hash >> (4 * 8) & 0xFF));
|
||||||
|
CREATE INDEX IF NOT EXISTS hash_6_index ON Hashes ((hash >> (5 * 8) & 0xFF));
|
||||||
|
CREATE INDEX IF NOT EXISTS hash_7_index ON Hashes ((hash >> (6 * 8) & 0xFF));
|
||||||
|
CREATE INDEX IF NOT EXISTS hash_8_index ON Hashes ((hash >> (7 * 8) & 0xFF));
|
||||||
|
|
||||||
|
CREATE INDEX IF NOT EXISTS id_domain ON IDs (domain, stringid);
|
||||||
|
PRAGMA shrink_memory;
|
||||||
|
ANALYZE;
|
||||||
|
`)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *sqliteStorage) GetMatches(hashes []ch.Hash, max int, exactOnly bool) ([]ch.Result, error) {
|
||||||
|
var (
|
||||||
|
foundMatches []ch.Result
|
||||||
|
tl ch.TimeLog
|
||||||
|
)
|
||||||
|
tl.ResetTime()
|
||||||
|
|
||||||
|
if exactOnly { // exact matches are also found by partial matches. Don't bother with exact matches so we don't have to de-duplicate
|
||||||
|
for _, hash := range hashes {
|
||||||
|
idlist, err := s.findExactHashes(nil, hash)
|
||||||
|
if err != nil {
|
||||||
|
return foundMatches, err
|
||||||
|
}
|
||||||
|
for id, equivalentIDs := range idlist {
|
||||||
|
foundMatches = append(foundMatches, ch.Result{
|
||||||
|
Hash: hash,
|
||||||
|
ID: id,
|
||||||
|
Distance: 0,
|
||||||
|
EquivalentIDs: equivalentIDs,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
tl.LogTime("Search Exact")
|
||||||
|
if len(foundMatches) > 0 {
|
||||||
|
return foundMatches, nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
foundHashes := make(map[uint64]struct{})
|
||||||
|
|
||||||
|
for _, hash := range hashes {
|
||||||
|
results, err := s.findPartialHashes(tl, nil, max, hash)
|
||||||
|
if err != nil {
|
||||||
|
return foundMatches, err
|
||||||
|
}
|
||||||
|
tl.LogTime(fmt.Sprintf("Search partial %v", hash.Kind))
|
||||||
|
|
||||||
|
for _, hash := range results {
|
||||||
|
if _, alreadyMatched := foundHashes[hash.Hash.Hash]; !alreadyMatched {
|
||||||
|
foundHashes[hash.Hash.Hash] = struct{}{}
|
||||||
|
foundMatches = append(foundMatches, hash)
|
||||||
|
} else {
|
||||||
|
log.Println("Hash already found", hash)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return foundMatches, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *sqliteStorage) mapHashes(tx *sql.Tx, hash ch.ImageHash) {
|
||||||
|
var err error
|
||||||
|
insertHash := tx.Stmt(s.insertHash)
|
||||||
|
insertID := tx.Stmt(s.insertID)
|
||||||
|
idExists := tx.Stmt(s.idExists)
|
||||||
|
insertEID := tx.Stmt(s.insertEID)
|
||||||
|
insertIEID := tx.Stmt(s.insertIEID)
|
||||||
|
|
||||||
|
rows, err := insertID.Query(hash.ID.Domain, hash.ID.ID)
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
if !rows.Next() {
|
||||||
|
panic("Unable to insert ID")
|
||||||
|
}
|
||||||
|
|
||||||
|
var id_id int64
|
||||||
|
err = rows.Scan(&id_id)
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
rows.Close()
|
||||||
|
|
||||||
|
for _, hash := range hash.Hashes {
|
||||||
|
_, err := insertHash.Exec(hash.Kind, int64(hash.Hash), id_id)
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
rows.Close()
|
||||||
|
row := idExists.QueryRow(id_id)
|
||||||
|
var count int64
|
||||||
|
err = row.Scan(&count)
|
||||||
|
if err != nil {
|
||||||
|
panic(fmt.Errorf("failed to query id: %w", err))
|
||||||
|
}
|
||||||
|
if count < 1 {
|
||||||
|
row := insertEID.QueryRow()
|
||||||
|
var eid int64
|
||||||
|
err = row.Scan(&eid)
|
||||||
|
if err != nil {
|
||||||
|
panic(fmt.Errorf("failed to insert equivalent id: %w", err))
|
||||||
|
}
|
||||||
|
_, err := insertIEID.Exec(id_id, eid)
|
||||||
|
if err != nil {
|
||||||
|
panic(fmt.Errorf("failed to associate equivalent IDs: %w", err))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
func (s *sqliteStorage) MapHashes(hash ch.ImageHash) {
|
||||||
|
tx, err := s.db.BeginTx(context.Background(), nil)
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
s.mapHashes(tx, hash)
|
||||||
|
err = tx.Commit()
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *sqliteStorage) DecodeHashes(hashes *ch.SavedHashes) error {
|
||||||
|
return nil
|
||||||
|
err := s.dropIndexes()
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
tx, err := s.db.BeginTx(context.Background(), nil)
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
insertID := tx.Stmt(s.insertID)
|
||||||
|
insertEID := tx.Stmt(s.insertEID)
|
||||||
|
insertIEID := tx.Stmt(s.insertIEID)
|
||||||
|
for _, idlist := range hashes.IDs {
|
||||||
|
var eid int64
|
||||||
|
id_ids := make([]int64, 0, len(idlist))
|
||||||
|
for _, id := range idlist {
|
||||||
|
var id_id int64
|
||||||
|
row := insertID.QueryRow(id.Domain, id.ID)
|
||||||
|
err = row.Scan(&id_id)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("failed to insert id: %w", err)
|
||||||
|
}
|
||||||
|
id_ids = append(id_ids, id_id)
|
||||||
|
}
|
||||||
|
row := insertEID.QueryRow()
|
||||||
|
err = row.Scan(&eid)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("failed to insert equivalent id: %w", err)
|
||||||
|
}
|
||||||
|
for _, id_id := range id_ids {
|
||||||
|
_, err = insertIEID.Exec(id_id, eid)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, savedHash := range hashes.Hashes {
|
||||||
|
s.mapHashes(tx, ch.ImageHash{
|
||||||
|
Hashes: []ch.Hash{savedHash.Hash},
|
||||||
|
ID: savedHash.ID,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
err = tx.Commit()
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
err = s.createIndexes()
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *sqliteStorage) EncodeHashes() (*ch.SavedHashes, error) {
|
||||||
|
hashes := ch.SavedHashes{}
|
||||||
|
tx, err := s.db.Begin()
|
||||||
|
if err != nil {
|
||||||
|
return &hashes, err
|
||||||
|
}
|
||||||
|
|
||||||
|
rows, err := tx.Query("SELECT Hashes.kind, Hashes.hash, IDs.domain, IDs.stringid FROM Hashes JOIN IDs ON Hashes.id=IDs.id ORDER BY Hashes.kind, Hashes.hash;")
|
||||||
|
if err != nil {
|
||||||
|
return &hashes, err
|
||||||
|
}
|
||||||
|
for rows.Next() {
|
||||||
|
var (
|
||||||
|
hash ch.SavedHash
|
||||||
|
tmpHash int64
|
||||||
|
)
|
||||||
|
err = rows.Scan(&hash.Hash.Kind, &tmpHash, &hash.ID.Domain, &hash.ID.ID)
|
||||||
|
if err != nil {
|
||||||
|
return &hashes, err
|
||||||
|
}
|
||||||
|
hash.Hash.Hash = uint64(tmpHash)
|
||||||
|
hashes.InsertHash(hash)
|
||||||
|
}
|
||||||
|
rows, err = tx.Query("SELECT IEIDs.equivalentid, IDs.domain, IDs.stringid FROM IDs JOIN IDsToEquivalantIDs AS IEIDs ON IDs.id=IEIDs.idid ORDER BY IEIDs.equivalentid, IDs.domain, IDs.stringid;")
|
||||||
|
if err != nil {
|
||||||
|
return &hashes, err
|
||||||
|
}
|
||||||
|
var (
|
||||||
|
previousEid int64 = -1
|
||||||
|
ids []ch.ID
|
||||||
|
)
|
||||||
|
for rows.Next() {
|
||||||
|
var (
|
||||||
|
id ch.ID
|
||||||
|
newEid int64
|
||||||
|
)
|
||||||
|
err = rows.Scan(&newEid, &id.Domain, &id.Domain)
|
||||||
|
if err != nil {
|
||||||
|
return &hashes, err
|
||||||
|
}
|
||||||
|
if newEid != previousEid {
|
||||||
|
previousEid = newEid
|
||||||
|
// Only keep groups len>1 as they are mapped in SavedHashes.Hashes
|
||||||
|
if len(ids) > 1 {
|
||||||
|
hashes.IDs = append(hashes.IDs, ids)
|
||||||
|
}
|
||||||
|
ids = make([]ch.ID, 0)
|
||||||
|
}
|
||||||
|
ids = append(ids, id)
|
||||||
|
}
|
||||||
|
return &hashes, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *sqliteStorage) AssociateIDs(newIDs []ch.NewIDs) error {
|
||||||
|
tx, err := s.db.BeginTx(context.Background(), nil)
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
insertID := tx.Stmt(s.insertID)
|
||||||
|
insertIEID := tx.Stmt(s.insertIEID)
|
||||||
|
for _, ids := range newIDs {
|
||||||
|
var (
|
||||||
|
newRowid int64
|
||||||
|
oldRowid int64
|
||||||
|
eid int64
|
||||||
|
)
|
||||||
|
rows := tx.QueryRow(`SELECT ITEI.idid, ITEI.equivalentid from IDs JOIN IDsToEquivalantIDs AS ITEI ON IDs.id=ITEI.idid WHERE domain=? AND stringid=?`, ids.OldID.Domain, ids.OldID.ID)
|
||||||
|
|
||||||
|
err := rows.Scan(&oldRowid, &eid)
|
||||||
|
if err != nil {
|
||||||
|
if errors.Is(err, sql.ErrNoRows) {
|
||||||
|
return ErrIDNotFound
|
||||||
|
}
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
rows = insertID.QueryRow(ids.NewID.Domain, ids.NewID.ID)
|
||||||
|
|
||||||
|
err = rows.Scan(&newRowid)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
_, err = insertIEID.Exec(newRowid, eid)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
err = tx.Commit()
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *sqliteStorage) GetIDs(id ch.ID) ch.IDList {
|
||||||
|
var ids []ch.ID
|
||||||
|
rows, err := s.idMatchStatement.Query(id.Domain, id.ID)
|
||||||
|
if err != nil {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
for rows.Next() {
|
||||||
|
var id ch.ID
|
||||||
|
err = rows.Scan(&id.Domain, &id.ID)
|
||||||
|
if err != nil {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
ids = append(ids, id)
|
||||||
|
}
|
||||||
|
return ch.ToIDList(ids)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (s *sqliteStorage) PrepareStatements() error {
|
||||||
|
var err error
|
||||||
|
s.insertHash, err = s.db.Prepare(`INSERT INTO Hashes (kind, hash, id) VALUES (?, ?, ?) ON CONFLICT DO UPDATE SET kind=?1`)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("failed to prepare database statements: %w", err)
|
||||||
|
}
|
||||||
|
s.insertID, err = s.db.Prepare(`INSERT INTO IDs (domain, stringid) VALUES (?,?) ON CONFLICT DO UPDATE SET domain=?1 RETURNING id`)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("failed to prepare database statements: %w", err)
|
||||||
|
}
|
||||||
|
s.insertEID, err = s.db.Prepare(`INSERT INTO EquivalentIDs DEFAULT VALUES RETURNING id;`)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("failed to prepare database statements: %w", err)
|
||||||
|
}
|
||||||
|
s.insertIEID, err = s.db.Prepare(`INSERT INTO IDsToEquivalantIDs (idid, equivalentid) VALUES (?, ?);`)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("failed to prepare database statements: %w", err)
|
||||||
|
}
|
||||||
|
s.idExists, err = s.db.Prepare(`SELECT COUNT(*) from IDsToEquivalantIDs WHERE idid=?`)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("failed to prepare database statements: %w", err)
|
||||||
|
}
|
||||||
|
s.hashExactMatchStatement, err = s.db.Prepare(`
|
||||||
|
select QIDs.domain, QIDs.stringid, IDs.domain, IDs.stringid from IDs
|
||||||
|
join IDsToEquivalantIDs as IEIDs on IDs.id=IEIDs.idid
|
||||||
|
join (
|
||||||
|
select QEIDs.id as id from EquivalentIDs as QEIDs
|
||||||
|
join IDsToEquivalantIDs as QIEIDs on QEIDs.id=QIEIDs.equivalentid
|
||||||
|
join IDs as QIDs on QIDs.id=QIEIDs.idid
|
||||||
|
join Hashes on Hashes.id=QIDs.id
|
||||||
|
where (Hashes.kind=? AND Hashes.hash=?)
|
||||||
|
) as EIDs on EIDs.id=IEIDs.equivalentid;
|
||||||
|
`)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("failed to prepare database statements: %w", err)
|
||||||
|
}
|
||||||
|
s.hashPartialMatchStatement, err = s.db.Prepare(`
|
||||||
|
select QIDs.domain, QIDs.stringid, EIDs.hash, IDs.domain, IDs.stringid from IDs
|
||||||
|
join IDsToEquivalantIDs as IEIDs on IDs.id=IEIDs.idid
|
||||||
|
join (
|
||||||
|
select Hashes.hash as hash, QEIDs.id as id from EquivalentIDs as QEIDs
|
||||||
|
join IDsToEquivalantIDs as QIEIDs on QEIDs.id=QIEIDs.equivalentid
|
||||||
|
join IDs as QIDs on QIDs.id=QIEIDs.idid
|
||||||
|
join Hashes on Hashes.id=QIDs.id
|
||||||
|
where (Hashes.kind=? AND (((Hashes.hash >> (0 * 8) & 0xFF)=(?2 >> (0 * 8) & 0xFF)) OR ((Hashes.hash >> (1 * 8) & 0xFF)=(?2 >> (1 * 8) & 0xFF)) OR ((Hashes.hash >> (2 * 8) & 0xFF)=(?2 >> (2 * 8) & 0xFF)) OR ((Hashes.hash >> (3 * 8) & 0xFF)=(?2 >> (3 * 8) & 0xFF)) OR ((Hashes.hash >> (4 * 8) & 0xFF)=(?2 >> (4 * 8) & 0xFF)) OR ((Hashes.hash >> (5 * 8) & 0xFF)=(?2 >> (5 * 8) & 0xFF)) OR ((Hashes.hash >> (6 * 8) & 0xFF)=(?2 >> (6 * 8) & 0xFF)) OR ((Hashes.hash >> (7 * 8) & 0xFF)=(?2 >> (7 * 8) & 0xFF))))
|
||||||
|
) as EIDs on EIDs.id=IEIDs.equivalentid;
|
||||||
|
`)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("failed to prepare database statements: %w", err)
|
||||||
|
}
|
||||||
|
s.idMatchStatement, err = s.db.Prepare(`
|
||||||
|
select IDs.domain, IDs.stringid from IDs
|
||||||
|
join IDsToEquivalantIDs as IEIDs on IDs.id=IEIDs.idid
|
||||||
|
join (
|
||||||
|
select EIDs.* from EquivalentIDs as EIDs
|
||||||
|
join IDsToEquivalantIDs as QIEIDs on EIDs.id=QIEIDs.equivalentid
|
||||||
|
join IDs as QIDs on QIDs.id=QIEIDs.idid
|
||||||
|
where (QIDs.domain=? AND QIDs.stringid=?)
|
||||||
|
) as EIDs on EIDs.id=IEIDs.equivalentid;
|
||||||
|
`)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("failed to prepare database statements: %w", err)
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func NewSqliteStorage(db, path string) (ch.HashStorage, error) {
|
||||||
|
sqlite := &sqliteStorage{}
|
||||||
|
sqlDB, err := sql.Open(db, fmt.Sprintf("file://%s?_pragma=cache_size(-200000)&_pragma=busy_timeout(500)&_pragma=hard_heap_limit(1073741824)&_pragma=journal_mode(wal)&_pragma=soft_heap_limit(314572800)", path))
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
sqlite.db = sqlDB
|
||||||
|
_, err = sqlite.db.Exec(`
|
||||||
|
PRAGMA foreign_keys=ON;
|
||||||
|
CREATE TABLE IF NOT EXISTS IDs(
|
||||||
|
id INTEGER PRIMARY KEY,
|
||||||
|
stringid TEXT NOT NULL,
|
||||||
|
domain TEXT NOT NULL
|
||||||
|
);
|
||||||
|
|
||||||
|
CREATE TABLE IF NOT EXISTS Hashes(
|
||||||
|
hash INTEGER NOT NULL,
|
||||||
|
kind INTEGER NOT NULL,
|
||||||
|
id INTEGER NOT NULL,
|
||||||
|
|
||||||
|
FOREIGN KEY(id) REFERENCES IDs(id)
|
||||||
|
);
|
||||||
|
|
||||||
|
CREATE TABLE IF NOT EXISTS EquivalentIDs(
|
||||||
|
id integer primary key
|
||||||
|
);
|
||||||
|
|
||||||
|
CREATE TABLE IF NOT EXISTS IDsToEquivalantIDs(
|
||||||
|
idid INTEGER NOT NULL,
|
||||||
|
equivalentid INTEGER NOT NULL,
|
||||||
|
PRIMARY KEY (idid, equivalentid),
|
||||||
|
|
||||||
|
FOREIGN KEY(idid) REFERENCES IDs(id),
|
||||||
|
FOREIGN KEY(equivalentid) REFERENCES EquivalentIDs(id)
|
||||||
|
);
|
||||||
|
`)
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
sqlite.createIndexes()
|
||||||
|
sqlite.db.SetMaxOpenConns(1)
|
||||||
|
err = sqlite.PrepareStatements()
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return sqlite, nil
|
||||||
|
}
|
@ -1,6 +1,6 @@
|
|||||||
//go:build cgo
|
//go:build cgo && !gokrazy
|
||||||
|
|
||||||
package ch
|
package storage
|
||||||
|
|
||||||
import (
|
import (
|
||||||
_ "github.com/mattn/go-sqlite3"
|
_ "github.com/mattn/go-sqlite3"
|
@ -1,6 +1,6 @@
|
|||||||
//go:build !cgo
|
//go:build !cgo && !gokrazy
|
||||||
|
|
||||||
package ch
|
package storage
|
||||||
|
|
||||||
import (
|
import (
|
||||||
_ "github.com/ncruces/go-sqlite3/driver"
|
_ "github.com/ncruces/go-sqlite3/driver"
|
184
storage/vp-tree.go
Normal file
184
storage/vp-tree.go
Normal file
@ -0,0 +1,184 @@
|
|||||||
|
//go:build !gokrazy
|
||||||
|
|
||||||
|
package storage
|
||||||
|
|
||||||
|
import (
|
||||||
|
"errors"
|
||||||
|
"fmt"
|
||||||
|
"math/bits"
|
||||||
|
|
||||||
|
ch "gitea.narnian.us/lordwelch/comic-hasher"
|
||||||
|
"gitea.narnian.us/lordwelch/goimagehash"
|
||||||
|
"gonum.org/v1/gonum/spatial/vptree"
|
||||||
|
)
|
||||||
|
|
||||||
|
type VPTree struct {
|
||||||
|
aTree *vptree.Tree
|
||||||
|
dTree *vptree.Tree
|
||||||
|
pTree *vptree.Tree
|
||||||
|
ids map[ch.ID]*[]ch.ID
|
||||||
|
|
||||||
|
aHashes []vptree.Comparable // temporary, only used for vptree creation
|
||||||
|
dHashes []vptree.Comparable // temporary, only used for vptree creation
|
||||||
|
pHashes []vptree.Comparable // temporary, only used for vptree creation
|
||||||
|
}
|
||||||
|
type VPHash struct {
|
||||||
|
ch.SavedHash
|
||||||
|
}
|
||||||
|
|
||||||
|
func (h *VPHash) Distance(c vptree.Comparable) float64 {
|
||||||
|
h2, ok := c.(*VPHash)
|
||||||
|
if !ok {
|
||||||
|
return -99
|
||||||
|
}
|
||||||
|
return float64(bits.OnesCount64(h.Hash.Hash ^ h2.Hash.Hash))
|
||||||
|
}
|
||||||
|
|
||||||
|
func (v *VPTree) GetMatches(hashes []ch.Hash, max int, exactOnly bool) ([]ch.Result, error) {
|
||||||
|
var (
|
||||||
|
matches []ch.Result
|
||||||
|
exactMatches []ch.Result
|
||||||
|
tl ch.TimeLog
|
||||||
|
)
|
||||||
|
tl.ResetTime()
|
||||||
|
defer tl.LogTime("Search Complete")
|
||||||
|
|
||||||
|
for _, hash := range hashes {
|
||||||
|
results := vptree.NewDistKeeper(float64(max))
|
||||||
|
|
||||||
|
currentTree := v.getCurrentTree(hash.Kind)
|
||||||
|
currentTree.NearestSet(results, &VPHash{ch.SavedHash{Hash: hash}})
|
||||||
|
|
||||||
|
mappedIds := map[*[]ch.ID]bool{}
|
||||||
|
for _, result := range results.Heap {
|
||||||
|
storedHash := result.Comparable.(*VPHash)
|
||||||
|
ids := v.ids[storedHash.ID]
|
||||||
|
if mappedIds[ids] {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
mappedIds[ids] = true
|
||||||
|
if result.Dist == 0 {
|
||||||
|
exactMatches = append(exactMatches, ch.Result{
|
||||||
|
Hash: storedHash.Hash,
|
||||||
|
ID: storedHash.ID,
|
||||||
|
Distance: 0,
|
||||||
|
EquivalentIDs: *v.ids[storedHash.ID],
|
||||||
|
})
|
||||||
|
} else {
|
||||||
|
matches = append(matches, ch.Result{
|
||||||
|
Hash: storedHash.Hash,
|
||||||
|
ID: storedHash.ID,
|
||||||
|
Distance: 0,
|
||||||
|
EquivalentIDs: *v.ids[storedHash.ID],
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if exactOnly && len(exactMatches) > 0 {
|
||||||
|
return exactMatches, nil
|
||||||
|
}
|
||||||
|
exactMatches = append(exactMatches, matches...)
|
||||||
|
return matches, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (v *VPTree) getCurrentTree(kind goimagehash.Kind) *vptree.Tree {
|
||||||
|
if kind == goimagehash.AHash {
|
||||||
|
return v.aTree
|
||||||
|
}
|
||||||
|
if kind == goimagehash.DHash {
|
||||||
|
return v.dTree
|
||||||
|
}
|
||||||
|
if kind == goimagehash.PHash {
|
||||||
|
return v.pTree
|
||||||
|
}
|
||||||
|
panic("Unknown hash type: " + kind.String())
|
||||||
|
}
|
||||||
|
|
||||||
|
func (v *VPTree) MapHashes(ch.ImageHash) {
|
||||||
|
panic("Not Implemented")
|
||||||
|
}
|
||||||
|
|
||||||
|
func (v *VPTree) DecodeHashes(hashes *ch.SavedHashes) error {
|
||||||
|
if hashes == nil {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Initialize all the known equal IDs
|
||||||
|
for _, ids := range hashes.IDs {
|
||||||
|
for _, id := range ids {
|
||||||
|
v.ids[id] = &ids
|
||||||
|
}
|
||||||
|
}
|
||||||
|
var err error
|
||||||
|
for _, savedHash := range hashes.Hashes {
|
||||||
|
if savedHash.Hash.Kind == goimagehash.AHash {
|
||||||
|
v.aHashes = append(v.aHashes, &VPHash{savedHash})
|
||||||
|
}
|
||||||
|
if savedHash.Hash.Kind == goimagehash.DHash {
|
||||||
|
v.dHashes = append(v.dHashes, &VPHash{savedHash})
|
||||||
|
}
|
||||||
|
if savedHash.Hash.Kind == goimagehash.PHash {
|
||||||
|
v.pHashes = append(v.pHashes, &VPHash{savedHash})
|
||||||
|
}
|
||||||
|
|
||||||
|
if savedHash.ID == (ch.ID{}) {
|
||||||
|
fmt.Println("Empty ID detected")
|
||||||
|
panic(savedHash)
|
||||||
|
}
|
||||||
|
// All known equal IDs are already mapped we can add any missing ones from hashes
|
||||||
|
if _, ok := v.ids[savedHash.ID]; !ok {
|
||||||
|
v.ids[savedHash.ID] = &[]ch.ID{savedHash.ID}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
v.aTree, err = vptree.New(v.aHashes, 3, nil)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
v.dTree, err = vptree.New(v.dHashes, 3, nil)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
v.pTree, err = vptree.New(v.pHashes, 3, nil)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
func (v *VPTree) EncodeHashes() (*ch.SavedHashes, error) {
|
||||||
|
return &ch.SavedHashes{}, errors.New("Not Implemented")
|
||||||
|
}
|
||||||
|
|
||||||
|
func (v *VPTree) AssociateIDs(newIDs []ch.NewIDs) error {
|
||||||
|
return errors.New("Not Implemented")
|
||||||
|
}
|
||||||
|
|
||||||
|
func (v *VPTree) GetIDs(id ch.ID) ch.IDList {
|
||||||
|
ids, found := v.ids[id]
|
||||||
|
if !found {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
return ch.ToIDList(*ids)
|
||||||
|
}
|
||||||
|
|
||||||
|
func NewVPStorage() (ch.HashStorage, error) {
|
||||||
|
var err error
|
||||||
|
v := &VPTree{
|
||||||
|
aHashes: []vptree.Comparable{},
|
||||||
|
dHashes: []vptree.Comparable{},
|
||||||
|
pHashes: []vptree.Comparable{},
|
||||||
|
}
|
||||||
|
v.aTree, err = vptree.New(v.aHashes, 3, nil)
|
||||||
|
if err != nil {
|
||||||
|
return v, err
|
||||||
|
}
|
||||||
|
v.dTree, err = vptree.New(v.dHashes, 3, nil)
|
||||||
|
if err != nil {
|
||||||
|
return v, err
|
||||||
|
}
|
||||||
|
v.pTree, err = vptree.New(v.pHashes, 3, nil)
|
||||||
|
if err != nil {
|
||||||
|
return v, err
|
||||||
|
}
|
||||||
|
return v, nil
|
||||||
|
}
|
13
storage/vp-tree_gokrazy.go
Normal file
13
storage/vp-tree_gokrazy.go
Normal file
@ -0,0 +1,13 @@
|
|||||||
|
//go:build gokrazy
|
||||||
|
|
||||||
|
package storage
|
||||||
|
|
||||||
|
import (
|
||||||
|
"errors"
|
||||||
|
|
||||||
|
ch "gitea.narnian.us/lordwelch/comic-hasher"
|
||||||
|
)
|
||||||
|
|
||||||
|
func NewVPStorage() (ch.HashStorage, error) {
|
||||||
|
return nil, errors.New("VPTree not available")
|
||||||
|
}
|
24
timing.go
Normal file
24
timing.go
Normal file
@ -0,0 +1,24 @@
|
|||||||
|
package ch
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"time"
|
||||||
|
)
|
||||||
|
|
||||||
|
type TimeLog struct {
|
||||||
|
total time.Duration
|
||||||
|
last time.Time
|
||||||
|
}
|
||||||
|
|
||||||
|
func (t *TimeLog) ResetTime() {
|
||||||
|
t.total = 0
|
||||||
|
t.last = time.Now()
|
||||||
|
}
|
||||||
|
|
||||||
|
func (t *TimeLog) LogTime(log string) {
|
||||||
|
now := time.Now()
|
||||||
|
diff := now.Sub(t.last)
|
||||||
|
t.last = now
|
||||||
|
t.total += diff
|
||||||
|
fmt.Printf("total: %v, %s: %v\n", t.total, log, diff)
|
||||||
|
}
|
105
vp-tree.go
105
vp-tree.go
@ -1,105 +0,0 @@
|
|||||||
package ch
|
|
||||||
|
|
||||||
import (
|
|
||||||
"errors"
|
|
||||||
"fmt"
|
|
||||||
"math/bits"
|
|
||||||
|
|
||||||
"gitea.narnian.us/lordwelch/goimagehash"
|
|
||||||
"gonum.org/v1/gonum/spatial/vptree"
|
|
||||||
)
|
|
||||||
|
|
||||||
type VPTree struct {
|
|
||||||
trees [3]*vptree.Tree
|
|
||||||
hashes [3][]vptree.Comparable
|
|
||||||
}
|
|
||||||
type VPHash struct {
|
|
||||||
Hash Hash
|
|
||||||
IDs []ID
|
|
||||||
}
|
|
||||||
|
|
||||||
func (h *VPHash) Distance(c vptree.Comparable) float64 {
|
|
||||||
h2, ok := c.(*VPHash)
|
|
||||||
if !ok {
|
|
||||||
return -99
|
|
||||||
}
|
|
||||||
return float64(bits.OnesCount64(h.Hash.Hash ^ h2.Hash.Hash))
|
|
||||||
}
|
|
||||||
|
|
||||||
func (v *VPTree) GetMatches(hashes []Hash, max int, exactOnly bool) ([]Result, error) {
|
|
||||||
var matches []Result
|
|
||||||
var exactMatches []Result
|
|
||||||
fmt.Println(hashes)
|
|
||||||
for _, hash := range hashes {
|
|
||||||
results := vptree.NewDistKeeper(float64(max))
|
|
||||||
hashType := int(hash.Kind) - 1
|
|
||||||
v.trees[hashType].NearestSet(results, &VPHash{Hash: hash})
|
|
||||||
for _, result := range results.Heap {
|
|
||||||
vphash := result.Comparable.(*VPHash)
|
|
||||||
if result.Dist == 0 {
|
|
||||||
exactMatches = append(exactMatches, Result{
|
|
||||||
IDs: ToIDList(vphash.IDs),
|
|
||||||
Distance: int(result.Dist),
|
|
||||||
Hash: vphash.Hash,
|
|
||||||
})
|
|
||||||
} else {
|
|
||||||
matches = append(matches, Result{
|
|
||||||
IDs: ToIDList(vphash.IDs),
|
|
||||||
Distance: int(result.Dist),
|
|
||||||
Hash: vphash.Hash,
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if len(exactMatches) > 0 && exactOnly {
|
|
||||||
return exactMatches, nil
|
|
||||||
}
|
|
||||||
matches = append(exactMatches[:len(exactMatches):len(exactMatches)], matches...)
|
|
||||||
return matches, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (v *VPTree) MapHashes(ImageHash) {
|
|
||||||
panic("Not Implemented")
|
|
||||||
}
|
|
||||||
|
|
||||||
func (v *VPTree) DecodeHashes(hashes SavedHashes) error {
|
|
||||||
var err error
|
|
||||||
for hashType, sourceHashes := range hashes.Hashes {
|
|
||||||
for hash, idsLocation := range sourceHashes {
|
|
||||||
var (
|
|
||||||
hashKind = goimagehash.Kind(hashType + 1)
|
|
||||||
)
|
|
||||||
hash := &VPHash{Hash{hash, hashKind}, hashes.IDs[idsLocation]}
|
|
||||||
v.hashes[hashType] = append(v.hashes[hashType], hash)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
for hashType := range 3 {
|
|
||||||
v.trees[hashType], err = vptree.New(v.hashes[hashType], 3, nil)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
func (v *VPTree) EncodeHashes() (SavedHashes, error) {
|
|
||||||
return SavedHashes{}, errors.New("Not Implemented")
|
|
||||||
}
|
|
||||||
|
|
||||||
func (v *VPTree) AssociateIDs(newIDs []NewIDs) error {
|
|
||||||
return errors.New("Not Implemented")
|
|
||||||
}
|
|
||||||
|
|
||||||
func (v *VPTree) GetIDs(id ID) IDList {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func NewVPStorage() (HashStorage, error) {
|
|
||||||
|
|
||||||
return &VPTree{
|
|
||||||
hashes: [3][]vptree.Comparable{
|
|
||||||
make([]vptree.Comparable, 0, 1_000_000),
|
|
||||||
make([]vptree.Comparable, 0, 1_000_000),
|
|
||||||
make([]vptree.Comparable, 0, 1_000_000),
|
|
||||||
},
|
|
||||||
}, nil
|
|
||||||
}
|
|
Reference in New Issue
Block a user