22
22
if (hpfs_sb(s)->sb_chk) if (hpfs_stop_cycles(s, a, &c1, &c2, "hpfs_bplus_lookup")) return -1;
23
23
if (btree->internal) {
24
24
for (i = 0; i < btree->n_used_nodes; i++)
25
if (btree->u.internal[i].file_secno > sec) {
26
a = btree->u.internal[i].down;
25
if (le32_to_cpu(btree->u.internal[i].file_secno) > sec) {
26
a = le32_to_cpu(btree->u.internal[i].down);
28
28
if (!(anode = hpfs_map_anode(s, a, &bh))) return -1;
29
29
btree = &anode->btree;
36
36
for (i = 0; i < btree->n_used_nodes; i++)
37
if (btree->u.external[i].file_secno <= sec &&
38
btree->u.external[i].file_secno + btree->u.external[i].length > sec) {
39
a = btree->u.external[i].disk_secno + sec - btree->u.external[i].file_secno;
37
if (le32_to_cpu(btree->u.external[i].file_secno) <= sec &&
38
le32_to_cpu(btree->u.external[i].file_secno) + le32_to_cpu(btree->u.external[i].length) > sec) {
39
a = le32_to_cpu(btree->u.external[i].disk_secno) + sec - le32_to_cpu(btree->u.external[i].file_secno);
40
40
if (hpfs_sb(s)->sb_chk) if (hpfs_chk_sectors(s, a, 1, "data")) {
45
45
struct hpfs_inode_info *hpfs_inode = hpfs_i(inode);
46
hpfs_inode->i_file_sec = btree->u.external[i].file_secno;
47
hpfs_inode->i_disk_sec = btree->u.external[i].disk_secno;
48
hpfs_inode->i_n_secs = btree->u.external[i].length;
46
hpfs_inode->i_file_sec = le32_to_cpu(btree->u.external[i].file_secno);
47
hpfs_inode->i_disk_sec = le32_to_cpu(btree->u.external[i].disk_secno);
48
hpfs_inode->i_n_secs = le32_to_cpu(btree->u.external[i].length);
97
if (btree->u.external[n].file_secno + btree->u.external[n].length != fsecno) {
97
if (le32_to_cpu(btree->u.external[n].file_secno) + le32_to_cpu(btree->u.external[n].length) != fsecno) {
98
98
hpfs_error(s, "allocated size %08x, trying to add sector %08x, %cnode %08x",
99
btree->u.external[n].file_secno + btree->u.external[n].length, fsecno,
99
le32_to_cpu(btree->u.external[n].file_secno) + le32_to_cpu(btree->u.external[n].length), fsecno,
100
100
fnod?'f':'a', node);
104
if (hpfs_alloc_if_possible(s, se = btree->u.external[n].disk_secno + btree->u.external[n].length)) {
105
btree->u.external[n].length++;
104
if (hpfs_alloc_if_possible(s, se = le32_to_cpu(btree->u.external[n].disk_secno) + le32_to_cpu(btree->u.external[n].length))) {
105
btree->u.external[n].length = cpu_to_le32(le32_to_cpu(btree->u.external[n].length) + 1);
106
106
mark_buffer_dirty(bh);
116
116
se = !fnod ? node : (node + 16384) & ~16383;
118
if (!(se = hpfs_alloc_sector(s, se, 1, fsecno*ALLOC_M>ALLOC_FWD_MAX ? ALLOC_FWD_MAX : fsecno*ALLOC_M<ALLOC_FWD_MIN ? ALLOC_FWD_MIN : fsecno*ALLOC_M, 1))) {
118
if (!(se = hpfs_alloc_sector(s, se, 1, fsecno*ALLOC_M>ALLOC_FWD_MAX ? ALLOC_FWD_MAX : fsecno*ALLOC_M<ALLOC_FWD_MIN ? ALLOC_FWD_MIN : fsecno*ALLOC_M))) {
122
fs = n < 0 ? 0 : btree->u.external[n].file_secno + btree->u.external[n].length;
122
fs = n < 0 ? 0 : le32_to_cpu(btree->u.external[n].file_secno) + le32_to_cpu(btree->u.external[n].length);
123
123
if (!btree->n_free_nodes) {
124
up = a != node ? anode->up : -1;
124
up = a != node ? le32_to_cpu(anode->up) : -1;
125
125
if (!(anode = hpfs_alloc_anode(s, a, &na, &bh1))) {
127
127
hpfs_free_sectors(s, se, 1);
130
130
if (a == node && fnod) {
131
anode->up = cpu_to_le32(node);
132
132
anode->btree.fnode_parent = 1;
133
133
anode->btree.n_used_nodes = btree->n_used_nodes;
134
134
anode->btree.first_free = btree->first_free;
137
137
btree->internal = 1;
138
138
btree->n_free_nodes = 11;
139
139
btree->n_used_nodes = 1;
140
btree->first_free = (char *)&(btree->u.internal[1]) - (char *)btree;
141
btree->u.internal[0].file_secno = -1;
142
btree->u.internal[0].down = na;
140
btree->first_free = cpu_to_le16((char *)&(btree->u.internal[1]) - (char *)btree);
141
btree->u.internal[0].file_secno = cpu_to_le32(-1);
142
btree->u.internal[0].down = cpu_to_le32(na);
143
143
mark_buffer_dirty(bh);
144
144
} else if (!(ranode = hpfs_alloc_anode(s, /*a*/0, &ra, &bh2))) {
153
153
btree = &anode->btree;
155
155
btree->n_free_nodes--; n = btree->n_used_nodes++;
156
btree->first_free += 12;
157
btree->u.external[n].disk_secno = se;
158
btree->u.external[n].file_secno = fs;
159
btree->u.external[n].length = 1;
156
btree->first_free = cpu_to_le16(le16_to_cpu(btree->first_free) + 12);
157
btree->u.external[n].disk_secno = cpu_to_le32(se);
158
btree->u.external[n].file_secno = cpu_to_le32(fs);
159
btree->u.external[n].length = cpu_to_le32(1);
160
160
mark_buffer_dirty(bh);
162
162
if ((a == node && fnod) || na == -1) return se;
164
while (up != (anode_secno)-1) {
165
165
struct anode *new_anode;
166
166
if (hpfs_sb(s)->sb_chk)
167
167
if (hpfs_stop_cycles(s, up, &c1, &c2, "hpfs_add_sector_to_btree #2")) return -1;
175
175
if (btree->n_free_nodes) {
176
176
btree->n_free_nodes--; n = btree->n_used_nodes++;
177
btree->first_free += 8;
178
btree->u.internal[n].file_secno = -1;
179
btree->u.internal[n].down = na;
180
btree->u.internal[n-1].file_secno = fs;
177
btree->first_free = cpu_to_le16(le16_to_cpu(btree->first_free) + 8);
178
btree->u.internal[n].file_secno = cpu_to_le32(-1);
179
btree->u.internal[n].down = cpu_to_le32(na);
180
btree->u.internal[n-1].file_secno = cpu_to_le32(fs);
181
181
mark_buffer_dirty(bh);
184
184
hpfs_free_sectors(s, ra, 1);
185
185
if ((anode = hpfs_map_anode(s, na, &bh))) {
186
anode->up = cpu_to_le32(up);
187
187
anode->btree.fnode_parent = up == node && fnod;
188
188
mark_buffer_dirty(bh);
193
up = up != node ? anode->up : -1;
194
btree->u.internal[btree->n_used_nodes - 1].file_secno = /*fs*/-1;
193
up = up != node ? le32_to_cpu(anode->up) : -1;
194
btree->u.internal[btree->n_used_nodes - 1].file_secno = cpu_to_le32(/*fs*/-1);
195
195
mark_buffer_dirty(bh);
198
198
if ((new_anode = hpfs_alloc_anode(s, a, &na, &bh))) {
199
199
anode = new_anode;
200
/*anode->up = up != -1 ? up : ra;*/
200
/*anode->up = cpu_to_le32(up != -1 ? up : ra);*/
201
201
anode->btree.internal = 1;
202
202
anode->btree.n_used_nodes = 1;
203
203
anode->btree.n_free_nodes = 59;
204
anode->btree.first_free = 16;
205
anode->btree.u.internal[0].down = a;
206
anode->btree.u.internal[0].file_secno = -1;
204
anode->btree.first_free = cpu_to_le16(16);
205
anode->btree.u.internal[0].down = cpu_to_le32(a);
206
anode->btree.u.internal[0].file_secno = cpu_to_le32(-1);
207
207
mark_buffer_dirty(bh);
209
209
if ((anode = hpfs_map_anode(s, a, &bh))) {
210
anode->up = cpu_to_le32(na);
211
211
mark_buffer_dirty(bh);
216
216
if ((anode = hpfs_map_anode(s, na, &bh))) {
217
anode->up = cpu_to_le32(node);
218
218
if (fnod) anode->btree.fnode_parent = 1;
219
219
mark_buffer_dirty(bh);
233
233
btree = &fnode->btree;
236
memcpy(&ranode->btree, btree, btree->first_free);
235
ranode->up = cpu_to_le32(node);
236
memcpy(&ranode->btree, btree, le16_to_cpu(btree->first_free));
237
237
if (fnod) ranode->btree.fnode_parent = 1;
238
238
ranode->btree.n_free_nodes = (ranode->btree.internal ? 60 : 40) - ranode->btree.n_used_nodes;
239
239
if (ranode->btree.internal) for (n = 0; n < ranode->btree.n_used_nodes; n++) {
240
240
struct anode *unode;
241
if ((unode = hpfs_map_anode(s, ranode->u.internal[n].down, &bh1))) {
241
if ((unode = hpfs_map_anode(s, le32_to_cpu(ranode->u.internal[n].down), &bh1))) {
242
unode->up = cpu_to_le32(ra);
243
243
unode->btree.fnode_parent = 0;
244
244
mark_buffer_dirty(bh1);
248
248
btree->internal = 1;
249
249
btree->n_free_nodes = fnod ? 10 : 58;
250
250
btree->n_used_nodes = 2;
251
btree->first_free = (char *)&btree->u.internal[2] - (char *)btree;
252
btree->u.internal[0].file_secno = fs;
253
btree->u.internal[0].down = ra;
254
btree->u.internal[1].file_secno = -1;
255
btree->u.internal[1].down = na;
251
btree->first_free = cpu_to_le16((char *)&btree->u.internal[2] - (char *)btree);
252
btree->u.internal[0].file_secno = cpu_to_le32(fs);
253
btree->u.internal[0].down = cpu_to_le32(ra);
254
btree->u.internal[1].file_secno = cpu_to_le32(-1);
255
btree->u.internal[1].down = cpu_to_le32(na);
256
256
mark_buffer_dirty(bh);
258
258
mark_buffer_dirty(bh2);
281
281
while (btree1->internal) {
282
ano = btree1->u.internal[pos].down;
282
ano = le32_to_cpu(btree1->u.internal[pos].down);
283
283
if (level) brelse(bh);
284
284
if (hpfs_sb(s)->sb_chk)
285
285
if (hpfs_stop_cycles(s, ano, &d1, &d2, "hpfs_remove_btree #1"))
292
292
for (i = 0; i < btree1->n_used_nodes; i++)
293
hpfs_free_sectors(s, btree1->u.external[i].disk_secno, btree1->u.external[i].length);
293
hpfs_free_sectors(s, le32_to_cpu(btree1->u.external[i].disk_secno), le32_to_cpu(btree1->u.external[i].length));
295
295
if (!level) return;
298
298
if (hpfs_stop_cycles(s, ano, &c1, &c2, "hpfs_remove_btree #2")) return;
299
299
hpfs_free_sectors(s, ano, 1);
301
ano = le32_to_cpu(anode->up);
303
303
if (!(anode = hpfs_map_anode(s, ano, &bh))) return;
304
304
btree1 = &anode->btree;
305
305
} else btree1 = btree;
306
306
for (i = 0; i < btree1->n_used_nodes; i++) {
307
if (btree1->u.internal[i].down == oano) {
307
if (le32_to_cpu(btree1->u.internal[i].down) == oano) {
308
308
if ((pos = i + 1) < btree1->n_used_nodes)
421
421
while (btree->internal) {
422
422
nodes = btree->n_used_nodes + btree->n_free_nodes;
423
423
for (i = 0; i < btree->n_used_nodes; i++)
424
if (btree->u.internal[i].file_secno >= secs) goto f;
424
if (le32_to_cpu(btree->u.internal[i].file_secno) >= secs) goto f;
426
426
hpfs_error(s, "internal btree %08x doesn't end with -1", node);
429
429
for (j = i + 1; j < btree->n_used_nodes; j++)
430
hpfs_ea_remove(s, btree->u.internal[j].down, 1, 0);
430
hpfs_ea_remove(s, le32_to_cpu(btree->u.internal[j].down), 1, 0);
431
431
btree->n_used_nodes = i + 1;
432
432
btree->n_free_nodes = nodes - btree->n_used_nodes;
433
btree->first_free = 8 + 8 * btree->n_used_nodes;
433
btree->first_free = cpu_to_le16(8 + 8 * btree->n_used_nodes);
434
434
mark_buffer_dirty(bh);
435
if (btree->u.internal[i].file_secno == secs) {
435
if (btree->u.internal[i].file_secno == cpu_to_le32(secs)) {
439
node = btree->u.internal[i].down;
439
node = le32_to_cpu(btree->u.internal[i].down);
441
441
if (hpfs_sb(s)->sb_chk)
442
442
if (hpfs_stop_cycles(s, node, &c1, &c2, "hpfs_truncate_btree"))
447
447
nodes = btree->n_used_nodes + btree->n_free_nodes;
448
448
for (i = 0; i < btree->n_used_nodes; i++)
449
if (btree->u.external[i].file_secno + btree->u.external[i].length >= secs) goto ff;
449
if (le32_to_cpu(btree->u.external[i].file_secno) + le32_to_cpu(btree->u.external[i].length) >= secs) goto ff;
453
if (secs <= btree->u.external[i].file_secno) {
453
if (secs <= le32_to_cpu(btree->u.external[i].file_secno)) {
454
454
hpfs_error(s, "there is an allocation error in file %08x, sector %08x", f, secs);
457
else if (btree->u.external[i].file_secno + btree->u.external[i].length > secs) {
458
hpfs_free_sectors(s, btree->u.external[i].disk_secno + secs -
459
btree->u.external[i].file_secno, btree->u.external[i].length
460
- secs + btree->u.external[i].file_secno); /* I hope gcc optimizes this :-) */
461
btree->u.external[i].length = secs - btree->u.external[i].file_secno;
457
else if (le32_to_cpu(btree->u.external[i].file_secno) + le32_to_cpu(btree->u.external[i].length) > secs) {
458
hpfs_free_sectors(s, le32_to_cpu(btree->u.external[i].disk_secno) + secs -
459
le32_to_cpu(btree->u.external[i].file_secno), le32_to_cpu(btree->u.external[i].length)
460
- secs + le32_to_cpu(btree->u.external[i].file_secno)); /* I hope gcc optimizes this :-) */
461
btree->u.external[i].length = cpu_to_le32(secs - le32_to_cpu(btree->u.external[i].file_secno));
463
463
for (j = i + 1; j < btree->n_used_nodes; j++)
464
hpfs_free_sectors(s, btree->u.external[j].disk_secno, btree->u.external[j].length);
464
hpfs_free_sectors(s, le32_to_cpu(btree->u.external[j].disk_secno), le32_to_cpu(btree->u.external[j].length));
465
465
btree->n_used_nodes = i + 1;
466
466
btree->n_free_nodes = nodes - btree->n_used_nodes;
467
btree->first_free = 8 + 12 * btree->n_used_nodes;
467
btree->first_free = cpu_to_le16(8 + 12 * btree->n_used_nodes);
468
468
mark_buffer_dirty(bh);
480
480
struct extended_attribute *ea_end;
481
481
if (!(fnode = hpfs_map_fnode(s, fno, &bh))) return;
482
482
if (!fnode->dirflag) hpfs_remove_btree(s, &fnode->btree);
483
else hpfs_remove_dtree(s, fnode->u.external[0].disk_secno);
483
else hpfs_remove_dtree(s, le32_to_cpu(fnode->u.external[0].disk_secno));
484
484
ea_end = fnode_end_ea(fnode);
485
485
for (ea = fnode_ea(fnode); ea < ea_end; ea = next_ea(ea))
486
486
if (ea->indirect)
487
487
hpfs_ea_remove(s, ea_sec(ea), ea->anode, ea_len(ea));
488
hpfs_ea_ext_remove(s, fnode->ea_secno, fnode->ea_anode, fnode->ea_size_l);
488
hpfs_ea_ext_remove(s, le32_to_cpu(fnode->ea_secno), fnode->ea_anode, le32_to_cpu(fnode->ea_size_l));
490
490
hpfs_free_sectors(s, fno, 1);