1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 
22 /*
23  * Copyright (c) 2012 by Delphix. All rights reserved.
24  */
25 
26 /*
27  * LibZFS_Core (lzc) is intended to replace most functionality in libzfs.
28  * It has the following characteristics:
29  *
30  *  - Thread Safe.  libzfs_core is accessible concurrently from multiple
31  *  threads.  This is accomplished primarily by avoiding global data
32  *  (e.g. caching).  Since it's thread-safe, there is no reason for a
33  *  process to have multiple libzfs "instances".  Therefore, we store
34  *  our few pieces of data (e.g. the file descriptor) in global
35  *  variables.  The fd is reference-counted so that the libzfs_core
36  *  library can be "initialized" multiple times (e.g. by different
37  *  consumers within the same process).
38  *
39  *  - Committed Interface.  The libzfs_core interface will be committed,
40  *  therefore consumers can compile against it and be confident that
41  *  their code will continue to work on future releases of this code.
42  *  Currently, the interface is Evolving (not Committed), but we intend
43  *  to commit to it once it is more complete and we determine that it
44  *  meets the needs of all consumers.
45  *
46  *  - Programatic Error Handling.  libzfs_core communicates errors with
47  *  defined error numbers, and doesn't print anything to stdout/stderr.
48  *
49  *  - Thin Layer.  libzfs_core is a thin layer, marshaling arguments
50  *  to/from the kernel ioctls.  There is generally a 1:1 correspondence
51  *  between libzfs_core functions and ioctls to /dev/zfs.
52  *
53  *  - Clear Atomicity.  Because libzfs_core functions are generally 1:1
54  *  with kernel ioctls, and kernel ioctls are general atomic, each
55  *  libzfs_core function is atomic.  For example, creating multiple
56  *  snapshots with a single call to lzc_snapshot() is atomic -- it
57  *  can't fail with only some of the requested snapshots created, even
58  *  in the event of power loss or system crash.
59  *
60  *  - Continued libzfs Support.  Some higher-level operations (e.g.
61  *  support for "zfs send -R") are too complicated to fit the scope of
62  *  libzfs_core.  This functionality will continue to live in libzfs.
63  *  Where appropriate, libzfs will use the underlying atomic operations
64  *  of libzfs_core.  For example, libzfs may implement "zfs send -R |
65  *  zfs receive" by using individual "send one snapshot", rename,
66  *  destroy, and "receive one snapshot" operations in libzfs_core.
67  *  /sbin/zfs and /zbin/zpool will link with both libzfs and
68  *  libzfs_core.  Other consumers should aim to use only libzfs_core,
69  *  since that will be the supported, stable interface going forwards.
70  */
71 
72 #include <libzfs_core.h>
73 #include <ctype.h>
74 #include <unistd.h>
75 #include <stdlib.h>
76 #include <string.h>
77 #include <errno.h>
78 #include <fcntl.h>
79 #include <pthread.h>
80 #include <sys/nvpair.h>
81 #include <sys/param.h>
82 #include <sys/types.h>
83 #include <sys/stat.h>
84 #include <sys/zfs_ioctl.h>
85 
86 static int g_fd;
87 static pthread_mutex_t g_lock = PTHREAD_MUTEX_INITIALIZER;
88 static int g_refcount;
89 
90 int
91 libzfs_core_init(void)
92 {
93 	(void) pthread_mutex_lock(&g_lock);
94 	if (g_refcount == 0) {
95 		g_fd = open("/dev/zfs", O_RDWR);
96 		if (g_fd < 0) {
97 			(void) pthread_mutex_unlock(&g_lock);
98 			return (errno);
99 		}
100 	}
101 	g_refcount++;
102 	(void) pthread_mutex_unlock(&g_lock);
103 	return (0);
104 }
105 
106 void
107 libzfs_core_fini(void)
108 {
109 	(void) pthread_mutex_lock(&g_lock);
110 	ASSERT3S(g_refcount, >, 0);
111 	g_refcount--;
112 	if (g_refcount == 0)
113 		(void) close(g_fd);
114 	(void) pthread_mutex_unlock(&g_lock);
115 }
116 
117 static int
118 lzc_ioctl(zfs_ioc_t ioc, const char *name,
119     nvlist_t *source, nvlist_t **resultp)
120 {
121 	zfs_cmd_t zc = { 0 };
122 	int error = 0;
123 	char *packed;
124 	size_t size;
125 
126 	ASSERT3S(g_refcount, >, 0);
127 
128 	(void) strlcpy(zc.zc_name, name, sizeof (zc.zc_name));
129 
130 	packed = fnvlist_pack(source, &size);
131 	zc.zc_nvlist_src = (uint64_t)(uintptr_t)packed;
132 	zc.zc_nvlist_src_size = size;
133 
134 	if (resultp != NULL) {
135 		*resultp = NULL;
136 		zc.zc_nvlist_dst_size = MAX(size * 2, 128 * 1024);
137 		zc.zc_nvlist_dst = (uint64_t)(uintptr_t)
138 		    malloc(zc.zc_nvlist_dst_size);
139 		if (zc.zc_nvlist_dst == NULL) {
140 			error = ENOMEM;
141 			goto out;
142 		}
143 	}
144 
145 	while (ioctl(g_fd, ioc, &zc) != 0) {
146 		if (errno == ENOMEM && resultp != NULL) {
147 			free((void *)(uintptr_t)zc.zc_nvlist_dst);
148 			zc.zc_nvlist_dst_size *= 2;
149 			zc.zc_nvlist_dst = (uint64_t)(uintptr_t)
150 			    malloc(zc.zc_nvlist_dst_size);
151 			if (zc.zc_nvlist_dst == NULL) {
152 				error = ENOMEM;
153 				goto out;
154 			}
155 		} else {
156 			error = errno;
157 			break;
158 		}
159 	}
160 	if (zc.zc_nvlist_dst_filled) {
161 		*resultp = fnvlist_unpack((void *)(uintptr_t)zc.zc_nvlist_dst,
162 		    zc.zc_nvlist_dst_size);
163 	}
164 
165 out:
166 	fnvlist_pack_free(packed, size);
167 	free((void *)(uintptr_t)zc.zc_nvlist_dst);
168 	return (error);
169 }
170 
171 int
172 lzc_create(const char *fsname, dmu_objset_type_t type, nvlist_t *props)
173 {
174 	int error;
175 	nvlist_t *args = fnvlist_alloc();
176 	fnvlist_add_int32(args, "type", type);
177 	if (props != NULL)
178 		fnvlist_add_nvlist(args, "props", props);
179 	error = lzc_ioctl(ZFS_IOC_CREATE, fsname, args, NULL);
180 	nvlist_free(args);
181 	return (error);
182 }
183 
184 int
185 lzc_clone(const char *fsname, const char *origin,
186     nvlist_t *props)
187 {
188 	int error;
189 	nvlist_t *args = fnvlist_alloc();
190 	fnvlist_add_string(args, "origin", origin);
191 	if (props != NULL)
192 		fnvlist_add_nvlist(args, "props", props);
193 	error = lzc_ioctl(ZFS_IOC_CLONE, fsname, args, NULL);
194 	nvlist_free(args);
195 	return (error);
196 }
197 
198 /*
199  * Creates snapshots.
200  *
201  * The keys in the snaps nvlist are the snapshots to be created.
202  * They must all be in the same pool.
203  *
204  * The props nvlist is properties to set.  Currently only user properties
205  * are supported.  { user:prop_name -> string value }
206  *
207  * The returned results nvlist will have an entry for each snapshot that failed.
208  * The value will be the (int32) error code.
209  *
210  * The return value will be 0 if all snapshots were created, otherwise it will
211  * be the errno of a (unspecified) snapshot that failed.
212  */
213 int
214 lzc_snapshot(nvlist_t *snaps, nvlist_t *props, nvlist_t **errlist)
215 {
216 	nvpair_t *elem;
217 	nvlist_t *args;
218 	int error;
219 	char pool[MAXNAMELEN];
220 
221 	*errlist = NULL;
222 
223 	/* determine the pool name */
224 	elem = nvlist_next_nvpair(snaps, NULL);
225 	if (elem == NULL)
226 		return (0);
227 	(void) strlcpy(pool, nvpair_name(elem), sizeof (pool));
228 	pool[strcspn(pool, "/@")] = '\0';
229 
230 	args = fnvlist_alloc();
231 	fnvlist_add_nvlist(args, "snaps", snaps);
232 	if (props != NULL)
233 		fnvlist_add_nvlist(args, "props", props);
234 
235 	error = lzc_ioctl(ZFS_IOC_SNAPSHOT, pool, args, errlist);
236 	nvlist_free(args);
237 
238 	return (error);
239 }
240 
241 /*
242  * Destroys snapshots.
243  *
244  * The keys in the snaps nvlist are the snapshots to be destroyed.
245  * They must all be in the same pool.
246  *
247  * Snapshots that do not exist will be silently ignored.
248  *
249  * If 'defer' is not set, and a snapshot has user holds or clones, the
250  * destroy operation will fail and none of the snapshots will be
251  * destroyed.
252  *
253  * If 'defer' is set, and a snapshot has user holds or clones, it will be
254  * marked for deferred destruction, and will be destroyed when the last hold
255  * or clone is removed/destroyed.
256  *
257  * The return value will be 0 if all snapshots were destroyed (or marked for
258  * later destruction if 'defer' is set) or didn't exist to begin with.
259  *
260  * Otherwise the return value will be the errno of a (unspecified) snapshot
261  * that failed, no snapshots will be destroyed, and the errlist will have an
262  * entry for each snapshot that failed.  The value in the errlist will be
263  * the (int32) error code.
264  */
265 int
266 lzc_destroy_snaps(nvlist_t *snaps, boolean_t defer, nvlist_t **errlist)
267 {
268 	nvpair_t *elem;
269 	nvlist_t *args;
270 	int error;
271 	char pool[MAXNAMELEN];
272 
273 	/* determine the pool name */
274 	elem = nvlist_next_nvpair(snaps, NULL);
275 	if (elem == NULL)
276 		return (0);
277 	(void) strlcpy(pool, nvpair_name(elem), sizeof (pool));
278 	pool[strcspn(pool, "/@")] = '\0';
279 
280 	args = fnvlist_alloc();
281 	fnvlist_add_nvlist(args, "snaps", snaps);
282 	if (defer)
283 		fnvlist_add_boolean(args, "defer");
284 
285 	error = lzc_ioctl(ZFS_IOC_DESTROY_SNAPS, pool, args, errlist);
286 	nvlist_free(args);
287 
288 	return (error);
289 
290 }
291 
292 int
293 lzc_snaprange_space(const char *firstsnap, const char *lastsnap,
294     uint64_t *usedp)
295 {
296 	nvlist_t *args;
297 	nvlist_t *result;
298 	int err;
299 	char fs[MAXNAMELEN];
300 	char *atp;
301 
302 	/* determine the fs name */
303 	(void) strlcpy(fs, firstsnap, sizeof (fs));
304 	atp = strchr(fs, '@');
305 	if (atp == NULL)
306 		return (EINVAL);
307 	*atp = '\0';
308 
309 	args = fnvlist_alloc();
310 	fnvlist_add_string(args, "firstsnap", firstsnap);
311 
312 	err = lzc_ioctl(ZFS_IOC_SPACE_SNAPS, lastsnap, args, &result);
313 	nvlist_free(args);
314 	if (err == 0)
315 		*usedp = fnvlist_lookup_uint64(result, "used");
316 	fnvlist_free(result);
317 
318 	return (err);
319 }
320 
321 boolean_t
322 lzc_exists(const char *dataset)
323 {
324 	/*
325 	 * The objset_stats ioctl is still legacy, so we need to construct our
326 	 * own zfs_cmd_t rather than using zfsc_ioctl().
327 	 */
328 	zfs_cmd_t zc = { 0 };
329 
330 	(void) strlcpy(zc.zc_name, dataset, sizeof (zc.zc_name));
331 	return (ioctl(g_fd, ZFS_IOC_OBJSET_STATS, &zc) == 0);
332 }
333 
334 /*
335  * Create "user holds" on snapshots.  If there is a hold on a snapshot,
336  * the snapshot can not be destroyed.  (However, it can be marked for deletion
337  * by lzc_destroy_snaps(defer=B_TRUE).)
338  *
339  * The keys in the nvlist are snapshot names.
340  * The snapshots must all be in the same pool.
341  * The value is the name of the hold (string type).
342  *
343  * If cleanup_fd is not -1, it must be the result of open("/dev/zfs", O_EXCL).
344  * In this case, when the cleanup_fd is closed (including on process
345  * termination), the holds will be released.  If the system is shut down
346  * uncleanly, the holds will be released when the pool is next opened
347  * or imported.
348  *
349  * The return value will be 0 if all holds were created. Otherwise the return
350  * value will be the errno of a (unspecified) hold that failed, no holds will
351  * be created, and the errlist will have an entry for each hold that
352  * failed (name = snapshot).  The value in the errlist will be the error
353  * code (int32).
354  */
355 int
356 lzc_hold(nvlist_t *holds, int cleanup_fd, nvlist_t **errlist)
357 {
358 	char pool[MAXNAMELEN];
359 	nvlist_t *args;
360 	nvpair_t *elem;
361 	int error;
362 
363 	/* determine the pool name */
364 	elem = nvlist_next_nvpair(holds, NULL);
365 	if (elem == NULL)
366 		return (0);
367 	(void) strlcpy(pool, nvpair_name(elem), sizeof (pool));
368 	pool[strcspn(pool, "/@")] = '\0';
369 
370 	args = fnvlist_alloc();
371 	fnvlist_add_nvlist(args, "holds", holds);
372 	if (cleanup_fd != -1)
373 		fnvlist_add_int32(args, "cleanup_fd", cleanup_fd);
374 
375 	error = lzc_ioctl(ZFS_IOC_HOLD, pool, args, errlist);
376 	nvlist_free(args);
377 	return (error);
378 }
379 
380 /*
381  * Release "user holds" on snapshots.  If the snapshot has been marked for
382  * deferred destroy (by lzc_destroy_snaps(defer=B_TRUE)), it does not have
383  * any clones, and all the user holds are removed, then the snapshot will be
384  * destroyed.
385  *
386  * The keys in the nvlist are snapshot names.
387  * The snapshots must all be in the same pool.
388  * The value is a nvlist whose keys are the holds to remove.
389  *
390  * The return value will be 0 if all holds were removed.
391  * Otherwise the return value will be the errno of a (unspecified) release
392  * that failed, no holds will be released, and the errlist will have an
393  * entry for each snapshot that has failed releases (name = snapshot).
394  * The value in the errlist will be the error code (int32) of a failed release.
395  */
396 int
397 lzc_release(nvlist_t *holds, nvlist_t **errlist)
398 {
399 	char pool[MAXNAMELEN];
400 	nvpair_t *elem;
401 
402 	/* determine the pool name */
403 	elem = nvlist_next_nvpair(holds, NULL);
404 	if (elem == NULL)
405 		return (0);
406 	(void) strlcpy(pool, nvpair_name(elem), sizeof (pool));
407 	pool[strcspn(pool, "/@")] = '\0';
408 
409 	return (lzc_ioctl(ZFS_IOC_RELEASE, pool, holds, errlist));
410 }
411 
412 /*
413  * Retrieve list of user holds on the specified snapshot.
414  *
415  * On success, *holdsp will be set to a nvlist which the caller must free.
416  * The keys are the names of the holds, and the value is the creation time
417  * of the hold (uint64) in seconds since the epoch.
418  */
419 int
420 lzc_get_holds(const char *snapname, nvlist_t **holdsp)
421 {
422 	int error;
423 	nvlist_t *innvl = fnvlist_alloc();
424 	error = lzc_ioctl(ZFS_IOC_GET_HOLDS, snapname, innvl, holdsp);
425 	fnvlist_free(innvl);
426 	return (error);
427 }
428 
429 /*
430  * If fromsnap is NULL, a full (non-incremental) stream will be sent.
431  */
432 int
433 lzc_send(const char *snapname, const char *fromsnap, int fd)
434 {
435 	nvlist_t *args;
436 	int err;
437 
438 	args = fnvlist_alloc();
439 	fnvlist_add_int32(args, "fd", fd);
440 	if (fromsnap != NULL)
441 		fnvlist_add_string(args, "fromsnap", fromsnap);
442 	err = lzc_ioctl(ZFS_IOC_SEND_NEW, snapname, args, NULL);
443 	nvlist_free(args);
444 	return (err);
445 }
446 
447 /*
448  * If fromsnap is NULL, a full (non-incremental) stream will be estimated.
449  */
450 int
451 lzc_send_space(const char *snapname, const char *fromsnap, uint64_t *spacep)
452 {
453 	nvlist_t *args;
454 	nvlist_t *result;
455 	int err;
456 
457 	args = fnvlist_alloc();
458 	if (fromsnap != NULL)
459 		fnvlist_add_string(args, "fromsnap", fromsnap);
460 	err = lzc_ioctl(ZFS_IOC_SEND_SPACE, snapname, args, &result);
461 	nvlist_free(args);
462 	if (err == 0)
463 		*spacep = fnvlist_lookup_uint64(result, "space");
464 	nvlist_free(result);
465 	return (err);
466 }
467 
468 static int
469 recv_read(int fd, void *buf, int ilen)
470 {
471 	char *cp = buf;
472 	int rv;
473 	int len = ilen;
474 
475 	do {
476 		rv = read(fd, cp, len);
477 		cp += rv;
478 		len -= rv;
479 	} while (rv > 0);
480 
481 	if (rv < 0 || len != 0)
482 		return (EIO);
483 
484 	return (0);
485 }
486 
487 /*
488  * The simplest receive case: receive from the specified fd, creating the
489  * specified snapshot.  Apply the specified properties a "received" properties
490  * (which can be overridden by locally-set properties).  If the stream is a
491  * clone, its origin snapshot must be specified by 'origin'.  The 'force'
492  * flag will cause the target filesystem to be rolled back or destroyed if
493  * necessary to receive.
494  *
495  * Return 0 on success or an errno on failure.
496  *
497  * Note: this interface does not work on dedup'd streams
498  * (those with DMU_BACKUP_FEATURE_DEDUP).
499  */
500 int
501 lzc_receive(const char *snapname, nvlist_t *props, const char *origin,
502     boolean_t force, int fd)
503 {
504 	/*
505 	 * The receive ioctl is still legacy, so we need to construct our own
506 	 * zfs_cmd_t rather than using zfsc_ioctl().
507 	 */
508 	zfs_cmd_t zc = { 0 };
509 	char *atp;
510 	char *packed = NULL;
511 	size_t size;
512 	dmu_replay_record_t drr;
513 	int error;
514 
515 	ASSERT3S(g_refcount, >, 0);
516 
517 	/* zc_name is name of containing filesystem */
518 	(void) strlcpy(zc.zc_name, snapname, sizeof (zc.zc_name));
519 	atp = strchr(zc.zc_name, '@');
520 	if (atp == NULL)
521 		return (EINVAL);
522 	*atp = '\0';
523 
524 	/* if the fs does not exist, try its parent. */
525 	if (!lzc_exists(zc.zc_name)) {
526 		char *slashp = strrchr(zc.zc_name, '/');
527 		if (slashp == NULL)
528 			return (ENOENT);
529 		*slashp = '\0';
530 
531 	}
532 
533 	/* zc_value is full name of the snapshot to create */
534 	(void) strlcpy(zc.zc_value, snapname, sizeof (zc.zc_value));
535 
536 	if (props != NULL) {
537 		/* zc_nvlist_src is props to set */
538 		packed = fnvlist_pack(props, &size);
539 		zc.zc_nvlist_src = (uint64_t)(uintptr_t)packed;
540 		zc.zc_nvlist_src_size = size;
541 	}
542 
543 	/* zc_string is name of clone origin (if DRR_FLAG_CLONE) */
544 	if (origin != NULL)
545 		(void) strlcpy(zc.zc_string, origin, sizeof (zc.zc_string));
546 
547 	/* zc_begin_record is non-byteswapped BEGIN record */
548 	error = recv_read(fd, &drr, sizeof (drr));
549 	if (error != 0)
550 		goto out;
551 	zc.zc_begin_record = drr.drr_u.drr_begin;
552 
553 	/* zc_cookie is fd to read from */
554 	zc.zc_cookie = fd;
555 
556 	/* zc guid is force flag */
557 	zc.zc_guid = force;
558 
559 	/* zc_cleanup_fd is unused */
560 	zc.zc_cleanup_fd = -1;
561 
562 	error = ioctl(g_fd, ZFS_IOC_RECV, &zc);
563 	if (error != 0)
564 		error = errno;
565 
566 out:
567 	if (packed != NULL)
568 		fnvlist_pack_free(packed, size);
569 	free((void*)(uintptr_t)zc.zc_nvlist_dst);
570 	return (error);
571 }
572