Searched defs:writes (Results 1 - 4 of 4) sorted by relevance
/illumos-gate/usr/src/cmd/stat/vmstat/ |
H A D | vmstat.c | 226 double reads, writes; local 237 writes = new->is_stats.writes - (old ? old->is_stats.writes : 0); 238 adjprintf(" %*.0f", 2, (reads + writes) / hr_etime * NANOSEC);
|
/illumos-gate/usr/src/uts/common/fs/zfs/ |
H A D | vdev_queue.c | 70 * except for the async write class. Asynchronous writes represent the data 73 * periodically so the number of queued async writes will quickly burst up and 114 * must further throttle incoming writes (see dmu_tx_delay() for details). 135 * between reads, writes, and scrubs. E.g., increasing 137 * more quickly, but reads and writes to have higher latency and lower 163 * For read I/Os, we also aggregate across small adjacency gaps; for writes 360 ksio->writes++; 405 int writes; local 431 writes = (dirty - min_bytes) * 436 ASSERT3U(writes, > [all...] |
/illumos-gate/usr/src/uts/common/sys/ |
H A D | kstat.h | 600 uint_t writes; /* number of write operations */ member in struct:kstat_io
|
/illumos-gate/usr/src/uts/sun4u/starcat/io/ |
H A D | schpc.c | 4295 int reg, writes = 0; local 4299 while (writes < save_reg_list[list_entry].number) { 4306 (writes * sizeof (uint64_t)), 4315 (writes * sizeof (uint64_t))), 4324 (writes * sizeof (uint32_t)), 4333 (writes * sizeof (uint32_t))), 4342 (writes * sizeof (uint16_t)), 4351 (writes * sizeof (uint16_t))), 4360 (writes * sizeof (uint8_t)), 4369 (writes * sizeo [all...] |
Completed in 71 milliseconds