author | eschrock |
Mon, 30 Jan 2006 21:34:28 -0800 | |
changeset 1354 | 81359ee1ee63 |
parent 873 | adefbfa5f42d |
child 1491 | bdcb30e07e7d |
permissions | -rw-r--r-- |
789 | 1 |
/* |
2 |
* CDDL HEADER START |
|
3 |
* |
|
4 |
* The contents of this file are subject to the terms of the |
|
5 |
* Common Development and Distribution License, Version 1.0 only |
|
6 |
* (the "License"). You may not use this file except in compliance |
|
7 |
* with the License. |
|
8 |
* |
|
9 |
* You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE |
|
10 |
* or http://www.opensolaris.org/os/licensing. |
|
11 |
* See the License for the specific language governing permissions |
|
12 |
* and limitations under the License. |
|
13 |
* |
|
14 |
* When distributing Covered Code, include this CDDL HEADER in each |
|
15 |
* file and include the License file at usr/src/OPENSOLARIS.LICENSE. |
|
16 |
* If applicable, add the following below this CDDL HEADER, with the |
|
17 |
* fields enclosed by brackets "[]" replaced with your own identifying |
|
18 |
* information: Portions Copyright [yyyy] [name of copyright owner] |
|
19 |
* |
|
20 |
* CDDL HEADER END |
|
21 |
*/ |
|
22 |
/* |
|
23 |
* Copyright 2005 Sun Microsystems, Inc. All rights reserved. |
|
24 |
* Use is subject to license terms. |
|
25 |
*/ |
|
26 |
||
27 |
#pragma ident "%Z%%M% %I% %E% SMI" |
|
28 |
||
29 |
#include <sys/dmu.h> |
|
30 |
#include <sys/dmu_impl.h> |
|
31 |
#include <sys/dbuf.h> |
|
32 |
#include <sys/dmu_tx.h> |
|
33 |
#include <sys/dmu_objset.h> |
|
34 |
#include <sys/dsl_dataset.h> /* for dsl_dataset_block_freeable() */ |
|
35 |
#include <sys/dsl_dir.h> /* for dsl_dir_tempreserve_*() */ |
|
36 |
#include <sys/dsl_pool.h> |
|
37 |
#include <sys/zap_impl.h> /* for ZAP_BLOCK_SHIFT */ |
|
38 |
#include <sys/spa.h> |
|
39 |
#include <sys/zfs_context.h> |
|
40 |
||
41 |
#ifdef ZFS_DEBUG |
|
42 |
int dmu_use_tx_debug_bufs = 1; |
|
43 |
#endif |
|
44 |
||
45 |
dmu_tx_t * |
|
46 |
dmu_tx_create_ds(dsl_dir_t *dd) |
|
47 |
{ |
|
48 |
dmu_tx_t *tx = kmem_zalloc(sizeof (dmu_tx_t), KM_SLEEP); |
|
49 |
tx->tx_dir = dd; |
|
50 |
if (dd) |
|
51 |
tx->tx_pool = dd->dd_pool; |
|
52 |
list_create(&tx->tx_holds, sizeof (dmu_tx_hold_t), |
|
53 |
offsetof(dmu_tx_hold_t, dth_node)); |
|
54 |
refcount_create(&tx->tx_space_written); |
|
55 |
refcount_create(&tx->tx_space_freed); |
|
56 |
return (tx); |
|
57 |
} |
|
58 |
||
59 |
dmu_tx_t * |
|
60 |
dmu_tx_create(objset_t *os) |
|
61 |
{ |
|
62 |
dmu_tx_t *tx = dmu_tx_create_ds(os->os->os_dsl_dataset->ds_dir); |
|
63 |
tx->tx_objset = os; |
|
64 |
return (tx); |
|
65 |
} |
|
66 |
||
67 |
dmu_tx_t * |
|
68 |
dmu_tx_create_assigned(struct dsl_pool *dp, uint64_t txg) |
|
69 |
{ |
|
70 |
dmu_tx_t *tx = dmu_tx_create_ds(NULL); |
|
71 |
||
72 |
ASSERT3U(txg, <=, dp->dp_tx.tx_open_txg); |
|
73 |
tx->tx_pool = dp; |
|
74 |
tx->tx_txg = txg; |
|
75 |
tx->tx_anyobj = TRUE; |
|
76 |
||
77 |
return (tx); |
|
78 |
} |
|
79 |
||
80 |
int |
|
81 |
dmu_tx_is_syncing(dmu_tx_t *tx) |
|
82 |
{ |
|
83 |
return (tx->tx_anyobj); |
|
84 |
} |
|
85 |
||
86 |
int |
|
87 |
dmu_tx_private_ok(dmu_tx_t *tx) |
|
88 |
{ |
|
89 |
return (tx->tx_anyobj || tx->tx_privateobj); |
|
90 |
} |
|
91 |
||
92 |
static void |
|
93 |
dmu_tx_hold_object_impl(dmu_tx_t *tx, objset_t *os, uint64_t object, |
|
94 |
enum dmu_tx_hold_type type, dmu_tx_hold_func_t func, |
|
95 |
uint64_t arg1, uint64_t arg2) |
|
96 |
{ |
|
97 |
dmu_tx_hold_t *dth; |
|
98 |
dnode_t *dn = NULL; |
|
99 |
||
100 |
if (object != DMU_NEW_OBJECT) { |
|
101 |
dn = dnode_hold(os->os, object, tx); |
|
102 |
||
103 |
if (tx->tx_txg != 0) { |
|
104 |
mutex_enter(&dn->dn_mtx); |
|
105 |
/* |
|
106 |
* dn->dn_assigned_txg == tx->tx_txg doesn't pose a |
|
107 |
* problem, but there's no way for it to happen (for |
|
108 |
* now, at least). |
|
109 |
*/ |
|
110 |
ASSERT(dn->dn_assigned_txg == 0); |
|
111 |
ASSERT(dn->dn_assigned_tx == NULL); |
|
112 |
dn->dn_assigned_txg = tx->tx_txg; |
|
113 |
dn->dn_assigned_tx = tx; |
|
114 |
(void) refcount_add(&dn->dn_tx_holds, tx); |
|
115 |
mutex_exit(&dn->dn_mtx); |
|
116 |
} |
|
117 |
} |
|
118 |
||
119 |
dth = kmem_zalloc(sizeof (dmu_tx_hold_t), KM_SLEEP); |
|
120 |
dth->dth_dnode = dn; |
|
121 |
dth->dth_type = type; |
|
122 |
dth->dth_func = func; |
|
123 |
dth->dth_arg1 = arg1; |
|
124 |
dth->dth_arg2 = arg2; |
|
125 |
/* |
|
126 |
* XXX Investigate using a different data structure to keep |
|
127 |
* track of dnodes in a tx. Maybe array, since there will |
|
128 |
* generally not be many entries? |
|
129 |
*/ |
|
130 |
list_insert_tail(&tx->tx_holds, dth); |
|
131 |
} |
|
132 |
||
133 |
void |
|
134 |
dmu_tx_add_new_object(dmu_tx_t *tx, objset_t *os, uint64_t object) |
|
135 |
{ |
|
136 |
/* |
|
137 |
* If we're syncing, they can manipulate any object anyhow, and |
|
138 |
* the hold on the dnode_t can cause problems. |
|
139 |
*/ |
|
140 |
if (!dmu_tx_is_syncing(tx)) { |
|
141 |
dmu_tx_hold_object_impl(tx, os, object, THT_NEWOBJECT, |
|
142 |
NULL, 0, 0); |
|
143 |
} |
|
144 |
} |
|
145 |
||
146 |
/* ARGSUSED */ |
|
147 |
static void |
|
148 |
dmu_tx_count_write(dmu_tx_t *tx, dnode_t *dn, uint64_t off, uint64_t len) |
|
149 |
{ |
|
150 |
uint64_t start, end, space; |
|
151 |
int min_bs, max_bs, min_ibs, max_ibs, epbs, bits; |
|
152 |
||
153 |
if (len == 0) |
|
154 |
return; |
|
155 |
||
156 |
min_bs = SPA_MINBLOCKSHIFT; |
|
157 |
max_bs = SPA_MAXBLOCKSHIFT; |
|
158 |
min_ibs = DN_MIN_INDBLKSHIFT; |
|
159 |
max_ibs = DN_MAX_INDBLKSHIFT; |
|
160 |
||
161 |
/* |
|
162 |
* If there's more than one block, the blocksize can't change, |
|
163 |
* so we can make a more precise estimate. Alternatively, |
|
164 |
* if the dnode's ibs is larger than max_ibs, always use that. |
|
165 |
* This ensures that if we reduce DN_MAX_INDBLKSHIFT, |
|
166 |
* the code will still work correctly on existing pools. |
|
167 |
*/ |
|
168 |
if (dn && (dn->dn_maxblkid != 0 || dn->dn_indblkshift > max_ibs)) { |
|
169 |
min_ibs = max_ibs = dn->dn_indblkshift; |
|
170 |
if (dn->dn_datablkshift != 0) |
|
171 |
min_bs = max_bs = dn->dn_datablkshift; |
|
172 |
} |
|
173 |
||
174 |
/* |
|
175 |
* 'end' is the last thing we will access, not one past. |
|
176 |
* This way we won't overflow when accessing the last byte. |
|
177 |
*/ |
|
178 |
start = P2ALIGN(off, 1ULL << max_bs); |
|
179 |
end = P2ROUNDUP(off + len, 1ULL << max_bs) - 1; |
|
180 |
space = end - start + 1; |
|
181 |
||
182 |
start >>= min_bs; |
|
183 |
end >>= min_bs; |
|
184 |
||
185 |
epbs = min_ibs - SPA_BLKPTRSHIFT; |
|
186 |
||
187 |
/* |
|
188 |
* The object contains at most 2^(64 - min_bs) blocks, |
|
189 |
* and each indirect level maps 2^epbs. |
|
190 |
*/ |
|
191 |
for (bits = 64 - min_bs; bits >= 0; bits -= epbs) { |
|
192 |
start >>= epbs; |
|
193 |
end >>= epbs; |
|
194 |
/* |
|
195 |
* If we increase the number of levels of indirection, |
|
196 |
* we'll need new blkid=0 indirect blocks. If start == 0, |
|
197 |
* we're already accounting for that blocks; and if end == 0, |
|
198 |
* we can't increase the number of levels beyond that. |
|
199 |
*/ |
|
200 |
if (start != 0 && end != 0) |
|
201 |
space += 1ULL << max_ibs; |
|
202 |
space += (end - start + 1) << max_ibs; |
|
203 |
} |
|
204 |
||
205 |
ASSERT(space < 2 * DMU_MAX_ACCESS); |
|
206 |
||
207 |
tx->tx_space_towrite += space; |
|
208 |
} |
|
209 |
||
210 |
static void |
|
211 |
dmu_tx_count_dnode(dmu_tx_t *tx, dnode_t *dn) |
|
212 |
{ |
|
213 |
dnode_t *mdn = tx->tx_objset->os->os_meta_dnode; |
|
214 |
uint64_t object = dn ? dn->dn_object : DN_MAX_OBJECT - 1; |
|
215 |
uint64_t pre_write_space; |
|
216 |
||
217 |
ASSERT(object < DN_MAX_OBJECT); |
|
218 |
pre_write_space = tx->tx_space_towrite; |
|
219 |
dmu_tx_count_write(tx, mdn, object << DNODE_SHIFT, 1 << DNODE_SHIFT); |
|
220 |
if (dn && dn->dn_dbuf->db_blkptr && |
|
221 |
dsl_dataset_block_freeable(dn->dn_objset->os_dsl_dataset, |
|
222 |
dn->dn_dbuf->db_blkptr->blk_birth, tx)) { |
|
223 |
tx->tx_space_tooverwrite += |
|
224 |
tx->tx_space_towrite - pre_write_space; |
|
225 |
tx->tx_space_towrite = pre_write_space; |
|
226 |
} |
|
227 |
} |
|
228 |
||
229 |
/* ARGSUSED */ |
|
230 |
static void |
|
231 |
dmu_tx_hold_write_impl(dmu_tx_t *tx, dnode_t *dn, uint64_t off, uint64_t len) |
|
232 |
{ |
|
233 |
dmu_tx_count_write(tx, dn, off, len); |
|
234 |
dmu_tx_count_dnode(tx, dn); |
|
235 |
} |
|
236 |
||
237 |
void |
|
238 |
dmu_tx_hold_write(dmu_tx_t *tx, uint64_t object, uint64_t off, int len) |
|
239 |
{ |
|
240 |
ASSERT(tx->tx_txg == 0); |
|
241 |
ASSERT(len > 0 && len < DMU_MAX_ACCESS); |
|
242 |
ASSERT(UINT64_MAX - off >= len - 1); |
|
243 |
||
244 |
dmu_tx_hold_object_impl(tx, tx->tx_objset, object, THT_WRITE, |
|
245 |
dmu_tx_hold_write_impl, off, len); |
|
246 |
} |
|
247 |
||
248 |
static void |
|
249 |
dmu_tx_count_free(dmu_tx_t *tx, dnode_t *dn, uint64_t off, uint64_t len) |
|
250 |
{ |
|
251 |
uint64_t blkid, nblks; |
|
252 |
uint64_t space = 0; |
|
253 |
dsl_dataset_t *ds = dn->dn_objset->os_dsl_dataset; |
|
254 |
||
255 |
ASSERT(dn->dn_assigned_tx == tx || dn->dn_assigned_tx == NULL); |
|
256 |
||
257 |
if (dn->dn_datablkshift == 0) |
|
258 |
return; |
|
259 |
/* |
|
260 |
* not that the dnode can change, since it isn't dirty, but |
|
261 |
* dbuf_hold_impl() wants us to have the struct_rwlock. |
|
262 |
* also need it to protect dn_maxblkid. |
|
263 |
*/ |
|
264 |
rw_enter(&dn->dn_struct_rwlock, RW_READER); |
|
265 |
blkid = off >> dn->dn_datablkshift; |
|
266 |
nblks = (off + len) >> dn->dn_datablkshift; |
|
267 |
||
268 |
if (blkid >= dn->dn_maxblkid) |
|
269 |
goto out; |
|
270 |
if (blkid + nblks > dn->dn_maxblkid) |
|
271 |
nblks = dn->dn_maxblkid - blkid; |
|
272 |
||
273 |
/* don't bother after the 100,000 blocks */ |
|
274 |
nblks = MIN(nblks, 128*1024); |
|
275 |
||
276 |
if (dn->dn_phys->dn_nlevels == 1) { |
|
277 |
int i; |
|
278 |
for (i = 0; i < nblks; i++) { |
|
279 |
blkptr_t *bp = dn->dn_phys->dn_blkptr; |
|
280 |
ASSERT3U(blkid + i, <, dn->dn_phys->dn_nblkptr); |
|
281 |
bp += blkid + i; |
|
282 |
if (dsl_dataset_block_freeable(ds, bp->blk_birth, tx)) { |
|
283 |
dprintf_bp(bp, "can free old%s", ""); |
|
284 |
space += BP_GET_ASIZE(bp); |
|
285 |
} |
|
286 |
} |
|
287 |
goto out; |
|
288 |
} |
|
289 |
||
290 |
while (nblks) { |
|
291 |
dmu_buf_impl_t *dbuf; |
|
292 |
int err, epbs, blkoff, tochk; |
|
293 |
||
294 |
epbs = dn->dn_indblkshift - SPA_BLKPTRSHIFT; |
|
295 |
blkoff = P2PHASE(blkid, 1<<epbs); |
|
296 |
tochk = MIN((1<<epbs) - blkoff, nblks); |
|
297 |
||
298 |
err = dbuf_hold_impl(dn, 1, blkid >> epbs, TRUE, FTAG, &dbuf); |
|
299 |
if (err == 0) { |
|
300 |
int i; |
|
301 |
blkptr_t *bp; |
|
302 |
||
303 |
dbuf_read_havestruct(dbuf); |
|
304 |
||
305 |
bp = dbuf->db.db_data; |
|
306 |
bp += blkoff; |
|
307 |
||
308 |
for (i = 0; i < tochk; i++) { |
|
309 |
if (dsl_dataset_block_freeable(ds, |
|
310 |
bp[i].blk_birth, tx)) { |
|
311 |
dprintf_bp(&bp[i], |
|
312 |
"can free old%s", ""); |
|
313 |
space += BP_GET_ASIZE(&bp[i]); |
|
314 |
} |
|
315 |
} |
|
316 |
dbuf_remove_ref(dbuf, FTAG); |
|
317 |
} else { |
|
318 |
/* the indirect block is sparse */ |
|
319 |
ASSERT(err == ENOENT); |
|
320 |
} |
|
321 |
||
322 |
blkid += tochk; |
|
323 |
nblks -= tochk; |
|
324 |
} |
|
325 |
out: |
|
326 |
rw_exit(&dn->dn_struct_rwlock); |
|
327 |
||
328 |
tx->tx_space_tofree += space; |
|
329 |
} |
|
330 |
||
331 |
static void |
|
332 |
dmu_tx_hold_free_impl(dmu_tx_t *tx, dnode_t *dn, uint64_t off, uint64_t len) |
|
333 |
{ |
|
334 |
int dirty; |
|
335 |
||
336 |
/* first block */ |
|
337 |
if (off != 0 /* || dn->dn_maxblkid == 0 */) |
|
338 |
dmu_tx_count_write(tx, dn, off, 1); |
|
339 |
/* last block */ |
|
340 |
if (len != DMU_OBJECT_END) |
|
341 |
dmu_tx_count_write(tx, dn, off+len, 1); |
|
342 |
||
343 |
dmu_tx_count_dnode(tx, dn); |
|
344 |
||
345 |
if (off >= (dn->dn_maxblkid+1) * dn->dn_datablksz) |
|
346 |
return; |
|
347 |
if (len == DMU_OBJECT_END) |
|
348 |
len = (dn->dn_maxblkid+1) * dn->dn_datablksz - off; |
|
349 |
||
350 |
/* XXX locking */ |
|
351 |
dirty = dn->dn_dirtyblksz[0] | dn->dn_dirtyblksz[1] | |
|
352 |
dn->dn_dirtyblksz[2] | dn->dn_dirtyblksz[3]; |
|
353 |
if (dn->dn_assigned_tx != NULL && !dirty) |
|
354 |
dmu_tx_count_free(tx, dn, off, len); |
|
355 |
} |
|
356 |
||
357 |
void |
|
358 |
dmu_tx_hold_free(dmu_tx_t *tx, uint64_t object, uint64_t off, uint64_t len) |
|
359 |
{ |
|
360 |
ASSERT(tx->tx_txg == 0); |
|
361 |
||
362 |
dmu_tx_hold_object_impl(tx, tx->tx_objset, object, THT_FREE, |
|
363 |
dmu_tx_hold_free_impl, off, len); |
|
364 |
} |
|
365 |
||
366 |
/* ARGSUSED */ |
|
367 |
static void |
|
368 |
dmu_tx_hold_zap_impl(dmu_tx_t *tx, dnode_t *dn, uint64_t nops, uint64_t cops) |
|
369 |
{ |
|
370 |
uint64_t nblocks; |
|
371 |
int epbs; |
|
372 |
||
373 |
dmu_tx_count_dnode(tx, dn); |
|
374 |
||
375 |
if (dn == NULL) { |
|
376 |
/* |
|
377 |
* Assuming that nops+cops is not super huge, we will be |
|
378 |
* able to fit a new object's entries into one leaf |
|
379 |
* block. So there will be at most 2 blocks total, |
|
380 |
* including the header block. |
|
381 |
*/ |
|
382 |
dmu_tx_count_write(tx, dn, 0, 2 << ZAP_BLOCK_SHIFT); |
|
383 |
return; |
|
384 |
} |
|
385 |
||
386 |
ASSERT3P(dmu_ot[dn->dn_type].ot_byteswap, ==, zap_byteswap); |
|
387 |
||
388 |
if (dn->dn_maxblkid == 0 && nops == 0) { |
|
389 |
/* |
|
390 |
* If there is only one block (i.e. this is a micro-zap) |
|
391 |
* and we are only doing updates, the accounting is simple. |
|
392 |
*/ |
|
393 |
if (dsl_dataset_block_freeable(dn->dn_objset->os_dsl_dataset, |
|
394 |
dn->dn_phys->dn_blkptr[0].blk_birth, tx)) |
|
395 |
tx->tx_space_tooverwrite += dn->dn_datablksz; |
|
396 |
else |
|
397 |
tx->tx_space_towrite += dn->dn_datablksz; |
|
398 |
return; |
|
399 |
} |
|
400 |
||
401 |
/* |
|
402 |
* 3 blocks overwritten per op: target leaf, ptrtbl block, header block |
|
403 |
* 3 new blocks written per op: new split leaf, 2 grown ptrtbl blocks |
|
404 |
*/ |
|
405 |
dmu_tx_count_write(tx, dn, dn->dn_maxblkid * dn->dn_datablksz, |
|
406 |
(nops * 6ULL + cops * 3ULL) << ZAP_BLOCK_SHIFT); |
|
407 |
||
408 |
/* |
|
409 |
* If the modified blocks are scattered to the four winds, |
|
410 |
* we'll have to modify an indirect twig for each. |
|
411 |
*/ |
|
412 |
epbs = dn->dn_indblkshift - SPA_BLKPTRSHIFT; |
|
413 |
for (nblocks = dn->dn_maxblkid >> epbs; nblocks != 0; nblocks >>= epbs) |
|
414 |
tx->tx_space_towrite += |
|
415 |
((nops + cops) * 3ULL) << dn->dn_indblkshift; |
|
416 |
} |
|
417 |
||
418 |
void |
|
419 |
dmu_tx_hold_zap(dmu_tx_t *tx, uint64_t object, int ops) |
|
420 |
{ |
|
421 |
ASSERT(tx->tx_txg == 0); |
|
422 |
||
423 |
dmu_tx_hold_object_impl(tx, tx->tx_objset, object, THT_ZAP, |
|
424 |
dmu_tx_hold_zap_impl, (ops > 0?ops:0), (ops < 0?-ops:0)); |
|
425 |
} |
|
426 |
||
427 |
void |
|
428 |
dmu_tx_hold_bonus(dmu_tx_t *tx, uint64_t object) |
|
429 |
{ |
|
430 |
ASSERT(tx->tx_txg == 0); |
|
431 |
||
432 |
dmu_tx_hold_object_impl(tx, tx->tx_objset, object, THT_BONUS, |
|
433 |
dmu_tx_hold_write_impl, 0, 0); |
|
434 |
} |
|
435 |
||
436 |
||
437 |
/* ARGSUSED */ |
|
438 |
static void |
|
439 |
dmu_tx_hold_space_impl(dmu_tx_t *tx, dnode_t *dn, |
|
440 |
uint64_t space, uint64_t unused) |
|
441 |
{ |
|
442 |
tx->tx_space_towrite += space; |
|
443 |
} |
|
444 |
||
445 |
void |
|
446 |
dmu_tx_hold_space(dmu_tx_t *tx, uint64_t space) |
|
447 |
{ |
|
448 |
ASSERT(tx->tx_txg == 0); |
|
449 |
||
450 |
dmu_tx_hold_object_impl(tx, tx->tx_objset, DMU_NEW_OBJECT, THT_SPACE, |
|
451 |
dmu_tx_hold_space_impl, space, 0); |
|
452 |
} |
|
453 |
||
454 |
int |
|
455 |
dmu_tx_holds(dmu_tx_t *tx, uint64_t object) |
|
456 |
{ |
|
457 |
dmu_tx_hold_t *dth; |
|
458 |
int holds = 0; |
|
459 |
||
460 |
/* |
|
461 |
* By asserting that the tx is assigned, we're counting the |
|
462 |
* number of dn_tx_holds, which is the same as the number of |
|
463 |
* dn_holds. Otherwise, we'd be counting dn_holds, but |
|
464 |
* dn_tx_holds could be 0. |
|
465 |
*/ |
|
466 |
ASSERT(tx->tx_txg != 0); |
|
467 |
||
468 |
/* if (tx->tx_anyobj == TRUE) */ |
|
469 |
/* return (0); */ |
|
470 |
||
471 |
for (dth = list_head(&tx->tx_holds); dth; |
|
472 |
dth = list_next(&tx->tx_holds, dth)) { |
|
473 |
if (dth->dth_dnode && dth->dth_dnode->dn_object == object) |
|
474 |
holds++; |
|
475 |
} |
|
476 |
||
477 |
return (holds); |
|
478 |
} |
|
479 |
||
873
adefbfa5f42d
6347448 non ZFS_DEBUG kernels shouldn't call empty verify functions
ek110237
parents:
789
diff
changeset
|
480 |
#ifdef ZFS_DEBUG |
789 | 481 |
void |
482 |
dmu_tx_dirty_buf(dmu_tx_t *tx, dmu_buf_impl_t *db) |
|
483 |
{ |
|
484 |
dmu_tx_hold_t *dth; |
|
485 |
int match_object = FALSE, match_offset = FALSE; |
|
486 |
dnode_t *dn = db->db_dnode; |
|
487 |
||
488 |
ASSERT(tx->tx_txg != 0); |
|
489 |
ASSERT(tx->tx_objset == NULL || dn->dn_objset == tx->tx_objset->os); |
|
490 |
ASSERT3U(dn->dn_object, ==, db->db.db_object); |
|
491 |
||
492 |
if (tx->tx_anyobj) |
|
493 |
return; |
|
494 |
||
495 |
/* XXX No checking on the meta dnode for now */ |
|
496 |
if (db->db.db_object & DMU_PRIVATE_OBJECT) |
|
497 |
return; |
|
498 |
||
499 |
for (dth = list_head(&tx->tx_holds); dth; |
|
500 |
dth = list_next(&tx->tx_holds, dth)) { |
|
501 |
ASSERT(dn == NULL || dn->dn_assigned_txg == tx->tx_txg); |
|
502 |
if (dth->dth_dnode == dn && dth->dth_type != THT_NEWOBJECT) |
|
503 |
match_object = TRUE; |
|
504 |
if (dth->dth_dnode == NULL || dth->dth_dnode == dn) { |
|
505 |
int datablkshift = dn->dn_datablkshift ? |
|
506 |
dn->dn_datablkshift : SPA_MAXBLOCKSHIFT; |
|
507 |
int epbs = dn->dn_indblkshift - SPA_BLKPTRSHIFT; |
|
508 |
int shift = datablkshift + epbs * db->db_level; |
|
509 |
uint64_t beginblk = shift >= 64 ? 0 : |
|
510 |
(dth->dth_arg1 >> shift); |
|
511 |
uint64_t endblk = shift >= 64 ? 0 : |
|
512 |
((dth->dth_arg1 + dth->dth_arg2 - 1) >> shift); |
|
513 |
uint64_t blkid = db->db_blkid; |
|
514 |
||
515 |
/* XXX dth_arg2 better not be zero... */ |
|
516 |
||
517 |
dprintf("found dth type %x beginblk=%llx endblk=%llx\n", |
|
518 |
dth->dth_type, beginblk, endblk); |
|
519 |
||
520 |
switch (dth->dth_type) { |
|
521 |
case THT_WRITE: |
|
522 |
if (blkid >= beginblk && blkid <= endblk) |
|
523 |
match_offset = TRUE; |
|
524 |
/* |
|
525 |
* We will let this hold work for the bonus |
|
526 |
* buffer so that we don't need to hold it |
|
527 |
* when creating a new object. |
|
528 |
*/ |
|
529 |
if (blkid == DB_BONUS_BLKID) |
|
530 |
match_offset = TRUE; |
|
531 |
/* |
|
532 |
* They might have to increase nlevels, |
|
533 |
* thus dirtying the new TLIBs. Or the |
|
534 |
* might have to change the block size, |
|
535 |
* thus dirying the new lvl=0 blk=0. |
|
536 |
*/ |
|
537 |
if (blkid == 0) |
|
538 |
match_offset = TRUE; |
|
539 |
break; |
|
540 |
case THT_FREE: |
|
541 |
if (blkid == beginblk && |
|
542 |
(dth->dth_arg1 != 0 || |
|
543 |
dn->dn_maxblkid == 0)) |
|
544 |
match_offset = TRUE; |
|
545 |
if (blkid == endblk && |
|
546 |
dth->dth_arg2 != DMU_OBJECT_END) |
|
547 |
match_offset = TRUE; |
|
548 |
break; |
|
549 |
case THT_BONUS: |
|
550 |
if (blkid == DB_BONUS_BLKID) |
|
551 |
match_offset = TRUE; |
|
552 |
break; |
|
553 |
case THT_ZAP: |
|
554 |
match_offset = TRUE; |
|
555 |
break; |
|
556 |
case THT_NEWOBJECT: |
|
557 |
match_object = TRUE; |
|
558 |
break; |
|
559 |
default: |
|
560 |
ASSERT(!"bad dth_type"); |
|
561 |
} |
|
562 |
} |
|
563 |
if (match_object && match_offset) |
|
564 |
return; |
|
565 |
} |
|
566 |
panic("dirtying dbuf obj=%llx lvl=%u blkid=%llx but not tx_held\n", |
|
567 |
(u_longlong_t)db->db.db_object, db->db_level, |
|
568 |
(u_longlong_t)db->db_blkid); |
|
873
adefbfa5f42d
6347448 non ZFS_DEBUG kernels shouldn't call empty verify functions
ek110237
parents:
789
diff
changeset
|
569 |
} |
789 | 570 |
#endif |
571 |
||
572 |
static int |
|
573 |
dmu_tx_try_assign(dmu_tx_t *tx, uint64_t txg_how, dmu_tx_hold_t **last_dth) |
|
574 |
{ |
|
575 |
dmu_tx_hold_t *dth; |
|
576 |
uint64_t lsize, asize, fsize; |
|
577 |
||
578 |
*last_dth = NULL; |
|
579 |
||
580 |
tx->tx_space_towrite = 0; |
|
581 |
tx->tx_space_tofree = 0; |
|
582 |
tx->tx_space_tooverwrite = 0; |
|
583 |
tx->tx_txg = txg_hold_open(tx->tx_pool, &tx->tx_txgh); |
|
584 |
||
585 |
if (txg_how >= TXG_INITIAL && txg_how != tx->tx_txg) |
|
586 |
return (ERESTART); |
|
587 |
||
588 |
for (dth = list_head(&tx->tx_holds); dth; |
|
589 |
*last_dth = dth, dth = list_next(&tx->tx_holds, dth)) { |
|
590 |
dnode_t *dn = dth->dth_dnode; |
|
591 |
if (dn != NULL) { |
|
592 |
mutex_enter(&dn->dn_mtx); |
|
593 |
while (dn->dn_assigned_txg == tx->tx_txg - 1) { |
|
594 |
if (txg_how != TXG_WAIT) { |
|
595 |
mutex_exit(&dn->dn_mtx); |
|
596 |
return (ERESTART); |
|
597 |
} |
|
598 |
cv_wait(&dn->dn_notxholds, &dn->dn_mtx); |
|
599 |
} |
|
600 |
if (dn->dn_assigned_txg == 0) { |
|
601 |
ASSERT(dn->dn_assigned_tx == NULL); |
|
602 |
dn->dn_assigned_txg = tx->tx_txg; |
|
603 |
dn->dn_assigned_tx = tx; |
|
604 |
} else { |
|
605 |
ASSERT(dn->dn_assigned_txg == tx->tx_txg); |
|
606 |
if (dn->dn_assigned_tx != tx) |
|
607 |
dn->dn_assigned_tx = NULL; |
|
608 |
} |
|
609 |
(void) refcount_add(&dn->dn_tx_holds, tx); |
|
610 |
mutex_exit(&dn->dn_mtx); |
|
611 |
} |
|
612 |
if (dth->dth_func) |
|
613 |
dth->dth_func(tx, dn, dth->dth_arg1, dth->dth_arg2); |
|
614 |
} |
|
615 |
||
616 |
/* |
|
617 |
* Convert logical size to worst-case allocated size. |
|
618 |
*/ |
|
619 |
fsize = spa_get_asize(tx->tx_pool->dp_spa, tx->tx_space_tooverwrite) + |
|
620 |
tx->tx_space_tofree; |
|
621 |
lsize = tx->tx_space_towrite + tx->tx_space_tooverwrite; |
|
622 |
asize = spa_get_asize(tx->tx_pool->dp_spa, lsize); |
|
623 |
tx->tx_space_towrite = asize; |
|
624 |
||
625 |
if (tx->tx_dir && asize != 0) { |
|
626 |
int err = dsl_dir_tempreserve_space(tx->tx_dir, |
|
627 |
lsize, asize, fsize, &tx->tx_tempreserve_cookie, tx); |
|
628 |
if (err) |
|
629 |
return (err); |
|
630 |
} |
|
631 |
||
632 |
return (0); |
|
633 |
} |
|
634 |
||
635 |
static uint64_t |
|
636 |
dmu_tx_unassign(dmu_tx_t *tx, dmu_tx_hold_t *last_dth) |
|
637 |
{ |
|
638 |
uint64_t txg = tx->tx_txg; |
|
639 |
dmu_tx_hold_t *dth; |
|
640 |
||
641 |
ASSERT(txg != 0); |
|
642 |
||
643 |
txg_rele_to_quiesce(&tx->tx_txgh); |
|
644 |
||
645 |
for (dth = last_dth; dth; dth = list_prev(&tx->tx_holds, dth)) { |
|
646 |
dnode_t *dn = dth->dth_dnode; |
|
647 |
||
648 |
if (dn == NULL) |
|
649 |
continue; |
|
650 |
mutex_enter(&dn->dn_mtx); |
|
651 |
ASSERT3U(dn->dn_assigned_txg, ==, txg); |
|
652 |
||
653 |
if (refcount_remove(&dn->dn_tx_holds, tx) == 0) { |
|
654 |
dn->dn_assigned_txg = 0; |
|
655 |
dn->dn_assigned_tx = NULL; |
|
656 |
cv_broadcast(&dn->dn_notxholds); |
|
657 |
} |
|
658 |
mutex_exit(&dn->dn_mtx); |
|
659 |
} |
|
660 |
||
661 |
txg_rele_to_sync(&tx->tx_txgh); |
|
662 |
||
663 |
tx->tx_txg = 0; |
|
664 |
return (txg); |
|
665 |
} |
|
666 |
||
667 |
/* |
|
668 |
* Assign tx to a transaction group. txg_how can be one of: |
|
669 |
* |
|
670 |
* (1) TXG_WAIT. If the current open txg is full, waits until there's |
|
671 |
* a new one. This should be used when you're not holding locks. |
|
672 |
* If will only fail if we're truly out of space (or over quota). |
|
673 |
* |
|
674 |
* (2) TXG_NOWAIT. If we can't assign into the current open txg without |
|
675 |
* blocking, returns immediately with ERESTART. This should be used |
|
676 |
* whenever you're holding locks. On an ERESTART error, the caller |
|
677 |
* should drop locks, do a txg_wait_open(dp, 0), and try again. |
|
678 |
* |
|
679 |
* (3) A specific txg. Use this if you need to ensure that multiple |
|
680 |
* transactions all sync in the same txg. Like TXG_NOWAIT, it |
|
681 |
* returns ERESTART if it can't assign you into the requested txg. |
|
682 |
*/ |
|
683 |
int |
|
684 |
dmu_tx_assign(dmu_tx_t *tx, uint64_t txg_how) |
|
685 |
{ |
|
686 |
dmu_tx_hold_t *last_dth; |
|
687 |
int err; |
|
688 |
||
689 |
ASSERT(tx->tx_txg == 0); |
|
690 |
ASSERT(txg_how != 0); |
|
691 |
ASSERT(!dsl_pool_sync_context(tx->tx_pool)); |
|
692 |
ASSERT3U(tx->tx_space_towrite, ==, 0); |
|
693 |
ASSERT3U(tx->tx_space_tofree, ==, 0); |
|
694 |
||
695 |
while ((err = dmu_tx_try_assign(tx, txg_how, &last_dth)) != 0) { |
|
696 |
uint64_t txg = dmu_tx_unassign(tx, last_dth); |
|
697 |
||
698 |
if (err != ERESTART || txg_how != TXG_WAIT) |
|
699 |
return (err); |
|
700 |
||
701 |
txg_wait_open(tx->tx_pool, txg + 1); |
|
702 |
} |
|
703 |
||
704 |
txg_rele_to_quiesce(&tx->tx_txgh); |
|
705 |
||
706 |
return (0); |
|
707 |
} |
|
708 |
||
709 |
void |
|
710 |
dmu_tx_willuse_space(dmu_tx_t *tx, int64_t delta) |
|
711 |
{ |
|
712 |
if (tx->tx_dir == NULL || delta == 0) |
|
713 |
return; |
|
714 |
||
715 |
if (delta > 0) { |
|
716 |
ASSERT3U(refcount_count(&tx->tx_space_written) + delta, <=, |
|
717 |
tx->tx_space_towrite); |
|
718 |
(void) refcount_add_many(&tx->tx_space_written, delta, NULL); |
|
719 |
} else { |
|
720 |
(void) refcount_add_many(&tx->tx_space_freed, -delta, NULL); |
|
721 |
} |
|
722 |
} |
|
723 |
||
724 |
void |
|
725 |
dmu_tx_commit(dmu_tx_t *tx) |
|
726 |
{ |
|
727 |
dmu_tx_hold_t *dth; |
|
728 |
||
729 |
ASSERT(tx->tx_txg != 0); |
|
730 |
||
731 |
while (dth = list_head(&tx->tx_holds)) { |
|
732 |
dnode_t *dn = dth->dth_dnode; |
|
733 |
||
734 |
list_remove(&tx->tx_holds, dth); |
|
735 |
kmem_free(dth, sizeof (dmu_tx_hold_t)); |
|
736 |
if (dn == NULL) |
|
737 |
continue; |
|
738 |
mutex_enter(&dn->dn_mtx); |
|
739 |
ASSERT3U(dn->dn_assigned_txg, ==, tx->tx_txg); |
|
740 |
||
741 |
if (refcount_remove(&dn->dn_tx_holds, tx) == 0) { |
|
742 |
dn->dn_assigned_txg = 0; |
|
743 |
dn->dn_assigned_tx = NULL; |
|
744 |
cv_broadcast(&dn->dn_notxholds); |
|
745 |
} |
|
746 |
mutex_exit(&dn->dn_mtx); |
|
747 |
dnode_rele(dn, tx); |
|
748 |
} |
|
749 |
||
750 |
if (tx->tx_dir && tx->tx_space_towrite > 0) { |
|
751 |
dsl_dir_tempreserve_clear(tx->tx_tempreserve_cookie, tx); |
|
752 |
} |
|
753 |
||
754 |
if (tx->tx_anyobj == FALSE) |
|
755 |
txg_rele_to_sync(&tx->tx_txgh); |
|
756 |
dprintf("towrite=%llu written=%llu tofree=%llu freed=%llu\n", |
|
757 |
tx->tx_space_towrite, refcount_count(&tx->tx_space_written), |
|
758 |
tx->tx_space_tofree, refcount_count(&tx->tx_space_freed)); |
|
759 |
refcount_destroy_many(&tx->tx_space_written, |
|
760 |
refcount_count(&tx->tx_space_written)); |
|
761 |
refcount_destroy_many(&tx->tx_space_freed, |
|
762 |
refcount_count(&tx->tx_space_freed)); |
|
763 |
#ifdef ZFS_DEBUG |
|
764 |
if (tx->tx_debug_buf) |
|
765 |
kmem_free(tx->tx_debug_buf, 4096); |
|
766 |
#endif |
|
767 |
kmem_free(tx, sizeof (dmu_tx_t)); |
|
768 |
} |
|
769 |
||
770 |
void |
|
771 |
dmu_tx_abort(dmu_tx_t *tx) |
|
772 |
{ |
|
773 |
dmu_tx_hold_t *dth; |
|
774 |
||
775 |
ASSERT(tx->tx_txg == 0); |
|
776 |
||
777 |
while (dth = list_head(&tx->tx_holds)) { |
|
778 |
dnode_t *dn = dth->dth_dnode; |
|
779 |
||
780 |
list_remove(&tx->tx_holds, dth); |
|
781 |
kmem_free(dth, sizeof (dmu_tx_hold_t)); |
|
782 |
if (dn != NULL) |
|
783 |
dnode_rele(dn, tx); |
|
784 |
} |
|
785 |
refcount_destroy_many(&tx->tx_space_written, |
|
786 |
refcount_count(&tx->tx_space_written)); |
|
787 |
refcount_destroy_many(&tx->tx_space_freed, |
|
788 |
refcount_count(&tx->tx_space_freed)); |
|
789 |
#ifdef ZFS_DEBUG |
|
790 |
if (tx->tx_debug_buf) |
|
791 |
kmem_free(tx->tx_debug_buf, 4096); |
|
792 |
#endif |
|
793 |
kmem_free(tx, sizeof (dmu_tx_t)); |
|
794 |
} |
|
795 |
||
796 |
uint64_t |
|
797 |
dmu_tx_get_txg(dmu_tx_t *tx) |
|
798 |
{ |
|
799 |
ASSERT(tx->tx_txg != 0); |
|
800 |
return (tx->tx_txg); |
|
801 |
} |