Searched defs:writes (Results 1 - 4 of 4) sorted by relevance

/illumos-gate/usr/src/cmd/stat/vmstat/
H A Dvmstat.c226 double reads, writes; local
237 writes = new->is_stats.writes - (old ? old->is_stats.writes : 0);
238 adjprintf(" %*.0f", 2, (reads + writes) / hr_etime * NANOSEC);
/illumos-gate/usr/src/uts/common/fs/zfs/
H A Dvdev_queue.c70 * except for the async write class. Asynchronous writes represent the data
73 * periodically so the number of queued async writes will quickly burst up and
114 * must further throttle incoming writes (see dmu_tx_delay() for details).
135 * between reads, writes, and scrubs. E.g., increasing
137 * more quickly, but reads and writes to have higher latency and lower
163 * For read I/Os, we also aggregate across small adjacency gaps; for writes
360 ksio->writes++;
405 int writes; local
431 writes = (dirty - min_bytes) *
436 ASSERT3U(writes, >
[all...]
/illumos-gate/usr/src/uts/common/sys/
H A Dkstat.h600 uint_t writes; /* number of write operations */ member in struct:kstat_io
/illumos-gate/usr/src/uts/sun4u/starcat/io/
H A Dschpc.c4295 int reg, writes = 0; local
4299 while (writes < save_reg_list[list_entry].number) {
4306 (writes * sizeof (uint64_t)),
4315 (writes * sizeof (uint64_t))),
4324 (writes * sizeof (uint32_t)),
4333 (writes * sizeof (uint32_t))),
4342 (writes * sizeof (uint16_t)),
4351 (writes * sizeof (uint16_t))),
4360 (writes * sizeof (uint8_t)),
4369 (writes * sizeo
[all...]

Completed in 71 milliseconds