* 2.6.28 API change,
* Added insert_inode_locked() helper function, prior to this most callers
* used insert_inode_hash(). The older method doesn't check for collisions
- * in the inode_hashtable but it still acceptible for use.
+ * in the inode_hashtable but it still acceptable for use.
*/
#ifndef HAVE_INSERT_INODE_LOCKED
static inline int
/*
* Slab allocation interfaces. The SPL slab differs from the standard
* Linux SLAB or SLUB primarily in that each cache may be backed by slabs
- * allocated from the physical or virtal memory address space. The virtual
+ * allocated from the physical or virtual memory address space. The virtual
* slabs allow for good behavior when allocation large objects of identical
* size. This slab implementation also supports both constructors and
* destructors which the Linux slab does not.
} arc_buf_contents_t;
/*
- * The following breakdows of arc_size exist for kstat only.
+ * The following breakdowns of arc_size exist for kstat only.
*/
typedef enum arc_space_type {
ARC_SPACE_DATA,
* Note that buffers can be in one of 6 states:
* ARC_anon - anonymous (discussed below)
* ARC_mru - recently used, currently cached
- * ARC_mru_ghost - recentely used, no longer in cache
+ * ARC_mru_ghost - recently used, no longer in cache
* ARC_mfu - frequently used, currently cached
* ARC_mfu_ghost - frequently used, no longer in cache
* ARC_l2c_only - exists in L2ARC but not other states
*
* 3. Use avl_destroy_nodes() to quickly process/free up any remaining nodes.
* Note that once you use avl_destroy_nodes(), you can no longer
- * use any routine except avl_destroy_nodes() and avl_destoy().
+ * use any routine except avl_destroy_nodes() and avl_destroy().
*
* 4. Use avl_destroy() to destroy the AVL tree itself.
*
* user data structure which must contain a field of type avl_node_t.
*
* Also assume the user data structures looks like:
- * stuct my_type {
+ * struct my_type {
* ...
* avl_node_t my_link;
* ...
/*
* Set the data blocksize for an object.
*
- * The object cannot have any blocks allcated beyond the first. If
+ * The object cannot have any blocks allocated beyond the first. If
* the first block is allocated already, the new size must be greater
* than the current block size. If these conditions are not met,
* ENOTSUP will be returned.
* checksums, and perform any necessary byte-swapping to the on-disk
* format.
*/
-/* Solaris library abstraction for EFI partitons */
+/* Solaris library abstraction for EFI partitions */
typedef struct dk_part {
diskaddr_t p_start; /* starting LBA */
diskaddr_t p_size; /* size in blocks */
- struct uuid p_guid; /* partion type GUID */
+ struct uuid p_guid; /* partition type GUID */
ushort_t p_tag; /* converted to part'n type GUID */
ushort_t p_flag; /* attributes */
char p_name[EFI_PART_NAME_LEN]; /* partition name */
/* values not stored on disk */
uint64_t pss_pass_exam; /* examined bytes per scan pass */
uint64_t pss_pass_start; /* start time of a scan pass */
- uint64_t pss_pass_scrub_pause; /* pause time of a scurb pass */
+ uint64_t pss_pass_scrub_pause; /* pause time of a scrub pass */
/* cumulative time scrub spent paused, needed for rate calculation */
uint64_t pss_pass_scrub_spent_paused;
uint64_t pss_pass_issued; /* issued bytes per scan pass */
uint64_t vs_fragmentation; /* device fragmentation */
uint64_t vs_initialize_bytes_done; /* bytes initialized */
uint64_t vs_initialize_bytes_est; /* total bytes to initialize */
- uint64_t vs_initialize_state; /* vdev_initialzing_state_t */
+ uint64_t vs_initialize_state; /* vdev_initializing_state_t */
uint64_t vs_initialize_action_time; /* time_t */
uint64_t vs_checkpoint_space; /* checkpoint-consumed space */
uint64_t vs_resilver_deferred; /* resilver deferred */
** a single double value, using NaN values to represent non-number
** values. The trick only works on 32-bit machines (ints and pointers
** are 32-bit values) with numbers represented as IEEE 754-2008 doubles
-** with conventional endianess (12345678 or 87654321), in CPUs that do
+** with conventional endianness (12345678 or 87654321), in CPUs that do
** not produce signaling NaN values (all NaNs are quiet).
*/
typedef struct sa_attr_reg {
char *sa_name; /* attribute name */
uint16_t sa_length;
- sa_bswap_type_t sa_byteswap; /* bswap functon enum */
+ sa_bswap_type_t sa_byteswap; /* bswap function enum */
sa_attr_type_t sa_attr; /* filled in during registration */
} sa_attr_reg_t;
* the DEFINE_DTRACE_PROBE macros.
*
* When adding new DTRACE_PROBEs to zfs source, both a tracepoint event
- * class defintition and a DEFINE_DTRACE_PROBE definition are needed to
+ * class definition and a DEFINE_DTRACE_PROBE definition are needed to
* avoid undefined function errors.
*/
/*
* When tracepoints are not available, a DEFINE_DTRACE_PROBE* macro is
* needed for each DTRACE_PROBE. These will be used to generate stub
- * tracing functions and protoypes for those functions. See
+ * tracing functions and prototypes for those functions. See
* include/sys/trace.h.
*/
* the number of active transaction holds (tc_count). As transactions
* are assigned into a transaction group the appropriate tc_count is
* incremented to indicate that there are pending changes that have yet
- * to quiesce. Consumers evenutally call txg_rele_to_sync() to decrement
+ * to quiesce. Consumers eventually call txg_rele_to_sync() to decrement
* the tc_count. A transaction group is not considered quiesced until all
* tx_cpu structures have reached a tc_count of zero.
*
/*
* The tx_state structure maintains the state information about the different
- * stages of the pool's transcation groups. A per pool tx_state structure
+ * stages of the pool's transaction groups. A per pool tx_state structure
* is used to track this information. The tx_state structure also points to
* an array of tx_cpu structures (described above). Although the tx_sync_lock
* is used to protect the members of this structure, it is not used to
*
* raidz_parity Returns parity of the RAIDZ block
* raidz_ncols Returns number of columns the block spans
- * raidz_nbigcols Returns number of big columns columns
+ * raidz_nbigcols Returns number of big columns
* raidz_col_p Returns pointer to a column
* raidz_col_size Returns size of a column
* raidz_big_size Returns size of big columns
/*
* The name of this argument. For keyword arguments this is the name
* functions will use to set the argument. For positional arguments
- * the name has no programatic meaning, but will appear in error
+ * the name has no programmatic meaning, but will appear in error
* messages and help output.
*/
const char *za_name;
/*
* All ACEs have a common hdr. For
* owner@, group@, and everyone@ this is all
- * thats needed.
+ * that's needed.
*/
typedef struct zfs_ace_hdr {
uint16_t z_type;
/*
* This structure emulates the vfs_t from other platforms. It's purpose
- * is to faciliate the handling of mount options and minimize structural
+ * is to facilitate the handling of mount options and minimize structural
* differences between the platforms.
*/
typedef struct vfs {
list_t z_all_znodes; /* all znodes in the fs */
uint64_t z_nr_znodes; /* number of znodes in the fs */
unsigned long z_rollback_time; /* last online rollback time */
- unsigned long z_snap_defer_time; /* last snapshot unmount deferal */
+ unsigned long z_snap_defer_time; /* last snapshot unmount deferral */
kmutex_t z_znodes_lock; /* lock for z_all_znodes */
arc_prune_t *z_arc_prune; /* called by ARC to prune caches */
struct inode *z_ctldir; /* .zfs directory inode */
* Log blocks are chained together. Originally they were chained at the
* end of the block. For performance reasons the chain was moved to the
* beginning of the block which allows writes for only the data being used.
- * The older position is supported for backwards compatability.
+ * The older position is supported for backwards compatibility.
*
* The zio_eck_t contains a zec_cksum which for the intent log is
* the sequence number of this log block. A seq of 0 is invalid.
/*
* Number of transactions (reads, writes, renames, etc.)
- * that have been commited.
+ * that have been committed.
*/
kstat_named_t zil_itx_count;
/* length of the encryption key */
size_t ci_keylen;
- /* human-readable name of the encryption alforithm */
+ /* human-readable name of the encryption algorithm */
char *ci_name;
} zio_crypt_info_t;
/* buffer for hmac key */
uint8_t zk_hmac_keydata[SHA512_HMAC_KEYLEN];
- /* buffer for currrent encryption key derived from master key */
+ /* buffer for current encryption key derived from master key */
uint8_t zk_current_keydata[MASTER_KEY_MAX_LEN];
/* current 64 bit salt for deriving an encryption key */
/* template of hmac key for illumos crypto api */
crypto_ctx_template_t zk_hmac_tmpl;
- /* lock for changing the salt and dependant values */
+ /* lock for changing the salt and dependent values */
krwlock_t zk_salt_lock;
} zio_crypt_key_t;
*
* NOP Write:
* The NOP write feature is performed by the ZIO_STAGE_NOP_WRITE stage
- * and is added to an existing write pipeline if a crypographically
+ * and is added to an existing write pipeline if a cryptographically
* secure checksum (i.e. SHA256) is enabled and compression is turned on.
* The NOP write stage will compare the checksums of the current data
* on-disk (level-0 blocks only) and the data that is currently being written.