karl
commited on
Commit
·
80ea4ee
1
Parent(s):
5d3e4b5
tensor deduplication working now
Browse files- _bighash.py +1 -1
- _safetensors.py +24 -20
_bighash.py
CHANGED
@@ -45,7 +45,7 @@ def tensor_python_tuple_hash(items, out_or_in_place):
|
|
45 |
|
46 |
def hash(buffer, *incoming_unhashed_ints):
|
47 |
if len(buffer) < 16:
|
48 |
-
return buffer.__hash__()
|
49 |
|
50 |
# first pass
|
51 |
# - allocate storage
|
|
|
45 |
|
46 |
def hash(buffer, *incoming_unhashed_ints):
|
47 |
if len(buffer) < 16:
|
48 |
+
return bytes(buffer).__hash__()
|
49 |
|
50 |
# first pass
|
51 |
# - allocate storage
|
_safetensors.py
CHANGED
@@ -15,24 +15,25 @@ class WritingSafeTensors:
|
|
15 |
self.weight_map = {}
|
16 |
self.hash_map = {}
|
17 |
def add(self, name, tensor):
|
18 |
-
print(name, '...')
|
19 |
tensor_hash = self.file.add(name, tensor)
|
20 |
-
image_of = self.hash_map.
|
21 |
-
if image_of is not
|
22 |
-
self.file.undo(name, tensor)
|
23 |
imaged_hash = self.weight_map[image_of].add(name, tensor, image_of)
|
24 |
assert imaged_hash == tensor_hash
|
25 |
-
|
26 |
-
|
27 |
-
|
28 |
-
|
29 |
-
|
30 |
-
self.
|
31 |
-
|
32 |
-
|
33 |
-
|
34 |
-
|
35 |
-
|
|
|
|
|
36 |
self.weight_map[name] = self.file
|
37 |
def finalize(self):
|
38 |
if len(self.files) == 1:
|
@@ -114,6 +115,8 @@ class WritingSafeTensors:
|
|
114 |
self.mmapview[start : end],
|
115 |
dtype=tensor.dtype, count=tensor.numel(),
|
116 |
).view(tensor.shape or [1])[:] = tensor
|
|
|
|
|
117 |
else:
|
118 |
image = self.header[image_of]
|
119 |
start, end = image['data_offsets']
|
@@ -121,7 +124,7 @@ class WritingSafeTensors:
|
|
121 |
assert (tensor == torch.frombuffer(
|
122 |
self.mmapview[start : end],
|
123 |
dtype=tensor.dtype, count=tensor.numel(),
|
124 |
-
).view(tensor.shape
|
125 |
|
126 |
tensor.flatten()
|
127 |
tensor_hash = hash(self.mmapview[start : end])
|
@@ -140,12 +143,13 @@ class WritingSafeTensors:
|
|
140 |
'data_offsets':
|
141 |
[start, end],
|
142 |
}
|
143 |
-
self.size = end
|
144 |
return tensor_hash
|
145 |
-
def undo(self, name, tensor):
|
146 |
-
|
|
|
|
|
|
|
147 |
del self.header[name]
|
148 |
-
self.size -= length
|
149 |
def set_metadata(self, **metadata):
|
150 |
m = self.header['__metadata__']
|
151 |
for k, v in metadata.items():
|
|
|
15 |
self.weight_map = {}
|
16 |
self.hash_map = {}
|
17 |
def add(self, name, tensor):
|
|
|
18 |
tensor_hash = self.file.add(name, tensor)
|
19 |
+
image_of = self.hash_map.setdefault(tensor_hash, name)
|
20 |
+
if image_of is not name:
|
21 |
+
self.file.undo(name, tensor, True)
|
22 |
imaged_hash = self.weight_map[image_of].add(name, tensor, image_of)
|
23 |
assert imaged_hash == tensor_hash
|
24 |
+
else:
|
25 |
+
print(name, '...')
|
26 |
+
if self.file.size >= self.file_size:
|
27 |
+
self.file.undo(name, tensor, False)
|
28 |
+
ct = len(self.files)
|
29 |
+
if len(self.files) == 1:
|
30 |
+
self.file.rename(f'{self.name}-{ct:05}.safetensors')
|
31 |
+
self.file.set_metadata(index = str(ct))
|
32 |
+
self.files = {self.file.filename:self.file}
|
33 |
+
ct += 1
|
34 |
+
self.file = self.File(f'{self.name}-{ct:05}.safetensors', index = ct)
|
35 |
+
self.files[self.file.filename] = self.file
|
36 |
+
self.file.add(name, tensor)
|
37 |
self.weight_map[name] = self.file
|
38 |
def finalize(self):
|
39 |
if len(self.files) == 1:
|
|
|
115 |
self.mmapview[start : end],
|
116 |
dtype=tensor.dtype, count=tensor.numel(),
|
117 |
).view(tensor.shape or [1])[:] = tensor
|
118 |
+
assert end >= self.size
|
119 |
+
self.size = end
|
120 |
else:
|
121 |
image = self.header[image_of]
|
122 |
start, end = image['data_offsets']
|
|
|
124 |
assert (tensor == torch.frombuffer(
|
125 |
self.mmapview[start : end],
|
126 |
dtype=tensor.dtype, count=tensor.numel(),
|
127 |
+
).view(tensor.shape)).all()
|
128 |
|
129 |
tensor.flatten()
|
130 |
tensor_hash = hash(self.mmapview[start : end])
|
|
|
143 |
'data_offsets':
|
144 |
[start, end],
|
145 |
}
|
|
|
146 |
return tensor_hash
|
147 |
+
def undo(self, name, tensor, is_image):
|
148 |
+
if not is_image:
|
149 |
+
length = tensor.numel() * tensor.dtype.itemsize
|
150 |
+
assert [self.size - length, self.size] == self.header[name]['data_offsets']
|
151 |
+
self.size -= length
|
152 |
del self.header[name]
|
|
|
153 |
def set_metadata(self, **metadata):
|
154 |
m = self.header['__metadata__']
|
155 |
for k, v in metadata.items():
|