1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
use git_hash::SIZE_OF_SHA1_DIGEST as SHA1_SIZE;
use git_object::owned;
use std::io;
const _TYPE_EXT1: u8 = 0;
const COMMIT: u8 = 1;
const TREE: u8 = 2;
const BLOB: u8 = 3;
const TAG: u8 = 4;
const _TYPE_EXT2: u8 = 5;
const OFS_DELTA: u8 = 6;
const REF_DELTA: u8 = 7;
#[derive(PartialEq, Eq, Debug, Hash, Ord, PartialOrd, Clone)]
#[cfg_attr(feature = "serde1", derive(serde::Serialize, serde::Deserialize))]
pub struct Entry {
pub header: Header,
pub decompressed_size: u64,
pub data_offset: u64,
}
impl Entry {
pub fn base_pack_offset(&self, distance: u64) -> u64 {
let pack_offset = self.data_offset - self.header_size() as u64;
pack_offset.checked_sub(distance).expect("in-bound distance of deltas")
}
pub fn pack_offset(&self) -> u64 {
self.data_offset - self.header_size() as u64
}
pub fn header_size(&self) -> usize {
self.header
.to_write(self.decompressed_size, io::sink())
.expect("io::sink() to never fail")
}
}
impl Entry {
pub fn from_bytes(d: &[u8], pack_offset: u64) -> Entry {
let (type_id, size, mut consumed) = parse_header_info(d);
use self::Header::*;
let object = match type_id {
OFS_DELTA => {
let (distance, leb_bytes) = leb64decode(&d[consumed..]);
let delta = OfsDelta {
base_distance: distance,
};
consumed += leb_bytes;
delta
}
REF_DELTA => {
let delta = RefDelta {
base_id: owned::Id::from_20_bytes(&d[consumed..consumed + SHA1_SIZE]),
};
consumed += SHA1_SIZE;
delta
}
BLOB => Blob,
TREE => Tree,
COMMIT => Commit,
TAG => Tag,
_ => panic!("We currently don't support any V3 features or extensions"),
};
Entry {
header: object,
decompressed_size: size,
data_offset: pack_offset + consumed as u64,
}
}
pub fn from_read(mut r: impl io::Read, pack_offset: u64) -> Result<Entry, io::Error> {
let (type_id, size, mut consumed) = streaming_parse_header_info(&mut r)?;
use self::Header::*;
let object = match type_id {
OFS_DELTA => {
let (distance, leb_bytes) = streaming_leb64decode(&mut r)?;
let delta = OfsDelta {
base_distance: distance,
};
consumed += leb_bytes;
delta
}
REF_DELTA => {
let mut buf = [0u8; SHA1_SIZE];
r.read_exact(&mut buf)?;
let delta = RefDelta {
base_id: owned::Id::new_sha1(buf),
};
consumed += SHA1_SIZE;
delta
}
BLOB => Blob,
TREE => Tree,
COMMIT => Commit,
TAG => Tag,
_ => panic!("We currently don't support any V3 features or extensions"),
};
Ok(Entry {
header: object,
decompressed_size: size,
data_offset: pack_offset + consumed as u64,
})
}
}
#[derive(PartialEq, Eq, Debug, Hash, Ord, PartialOrd, Clone, Copy)]
#[cfg_attr(feature = "serde1", derive(serde::Serialize, serde::Deserialize))]
#[allow(missing_docs)]
pub enum Header {
Commit,
Tree,
Blob,
Tag,
RefDelta { base_id: owned::Id },
OfsDelta { base_distance: u64 },
}
impl Header {
pub fn verified_base_pack_offset(pack_offset: u64, distance: u64) -> Option<u64> {
if distance == 0 {
return None;
}
pack_offset.checked_sub(distance)
}
pub fn to_kind(&self) -> Option<git_object::Kind> {
use git_object::Kind::*;
Some(match self {
Header::Tree => Tree,
Header::Blob => Blob,
Header::Commit => Commit,
Header::Tag => Tag,
Header::RefDelta { .. } | Header::OfsDelta { .. } => return None,
})
}
pub fn to_type_id(&self) -> u8 {
use Header::*;
match self {
Blob => BLOB,
Tree => TREE,
Commit => COMMIT,
Tag => TAG,
OfsDelta { .. } => OFS_DELTA,
RefDelta { .. } => REF_DELTA,
}
}
pub fn is_delta(&self) -> bool {
matches!(self, Header::OfsDelta { .. } | Header::RefDelta { .. })
}
pub fn is_base(&self) -> bool {
!self.is_delta()
}
}
impl Header {
pub fn to_write(&self, decompressed_size_in_bytes: u64, mut out: impl io::Write) -> io::Result<usize> {
let mut size = decompressed_size_in_bytes;
let mut written = 1;
let mut c: u8 = (self.to_type_id() << 4) | (size as u8 & 0b0000_1111);
size >>= 4;
while size != 0 {
out.write_all(&[c | 0b1000_0000])?;
written += 1;
c = size as u8 & 0b0111_1111;
size >>= 7;
}
out.write_all(&[c])?;
use Header::*;
match self {
RefDelta { base_id: oid } => {
out.write_all(oid.as_slice())?;
written += oid.as_slice().len();
}
OfsDelta { mut base_distance } => {
let mut buf = [0u8; 10];
let mut bytes_written = 1;
buf[buf.len() - 1] = base_distance as u8 & 0b0111_1111;
for out in buf.iter_mut().rev().skip(1) {
base_distance >>= 7;
if base_distance == 0 {
break;
}
base_distance -= 1;
*out = 0b1000_0000 | (base_distance as u8 & 0b0111_1111);
bytes_written += 1;
}
out.write_all(&buf[buf.len() - bytes_written..])?;
written += bytes_written;
}
Blob | Tree | Commit | Tag => {}
}
Ok(written)
}
}
#[inline]
fn leb64decode(d: &[u8]) -> (u64, usize) {
let mut i = 0;
let mut c = d[i];
i += 1;
let mut value = c as u64 & 0x7f;
while c & 0x80 != 0 {
c = d[i];
i += 1;
value += 1;
value = (value << 7) + (c as u64 & 0x7f)
}
(value, i)
}
#[inline]
fn streaming_leb64decode(mut r: impl io::Read) -> Result<(u64, usize), io::Error> {
let mut b = [0u8; 1];
let mut i = 0;
r.read_exact(&mut b)?;
i += 1;
let mut value = b[0] as u64 & 0x7f;
while b[0] & 0x80 != 0 {
r.read_exact(&mut b)?;
i += 1;
value += 1;
value = (value << 7) + (b[0] as u64 & 0x7f)
}
Ok((value, i))
}
#[inline]
fn parse_header_info(data: &[u8]) -> (u8, u64, usize) {
let mut c = data[0];
let mut i = 1;
let type_id = (c >> 4) & 0b0000_0111;
let mut size = c as u64 & 0b0000_1111;
let mut s = 4;
while c & 0b1000_0000 != 0 {
c = data[i];
i += 1;
size += ((c & 0b0111_1111) as u64) << s;
s += 7
}
(type_id, size, i)
}
#[inline]
fn streaming_parse_header_info(mut read: impl io::Read) -> Result<(u8, u64, usize), io::Error> {
let mut byte = [0u8; 1];
read.read_exact(&mut byte)?;
let mut c = byte[0];
let mut i = 1;
let type_id = (c >> 4) & 0b0000_0111;
let mut size = c as u64 & 0b0000_1111;
let mut s = 4;
while c & 0b1000_0000 != 0 {
read.read_exact(&mut byte)?;
c = byte[0];
i += 1;
size += ((c & 0b0111_1111) as u64) << s;
s += 7
}
Ok((type_id, size, i))
}