amt 1844 drivers/infiniband/hw/hfi1/init.c unsigned amt; amt 1850 drivers/infiniband/hw/hfi1/init.c amt = rcvhdrq_size(rcd); amt 1856 drivers/infiniband/hw/hfi1/init.c rcd->rcvhdrq = dma_alloc_coherent(&dd->pcidev->dev, amt, amt 1863 drivers/infiniband/hw/hfi1/init.c amt, rcd->ctxt); amt 1908 drivers/infiniband/hw/hfi1/init.c dma_free_coherent(&dd->pcidev->dev, amt, rcd->rcvhdrq, amt 1545 drivers/infiniband/hw/qib/qib_init.c unsigned amt; amt 1552 drivers/infiniband/hw/qib/qib_init.c amt = ALIGN(dd->rcvhdrcnt * dd->rcvhdrentsize * amt 1560 drivers/infiniband/hw/qib/qib_init.c &dd->pcidev->dev, amt, &rcd->rcvhdrq_phys, amt 1567 drivers/infiniband/hw/qib/qib_init.c amt, rcd->ctxt); amt 1588 drivers/infiniband/hw/qib/qib_init.c rcd->rcvhdrq_size = amt; amt 1604 drivers/infiniband/hw/qib/qib_init.c dma_free_coherent(&dd->pcidev->dev, amt, rcd->rcvhdrq, amt 1934 drivers/tty/n_tty.c int amt = poll && !TIME_CHAR(tty) && MIN_CHAR(tty) ? MIN_CHAR(tty) : 1; amt 1939 drivers/tty/n_tty.c return ldata->commit_head - ldata->read_tail >= amt; amt 246 drivers/video/fbdev/nvidia/nv_setup.c u32 amt; amt 248 drivers/video/fbdev/nvidia/nv_setup.c pci_read_config_dword(dev, 0x7c, &amt); amt 249 drivers/video/fbdev/nvidia/nv_setup.c par->RamAmountKBytes = (((amt >> 6) & 31) + 1) * 1024; amt 251 drivers/video/fbdev/nvidia/nv_setup.c u32 amt; amt 253 drivers/video/fbdev/nvidia/nv_setup.c pci_read_config_dword(dev, 0x84, &amt); amt 254 drivers/video/fbdev/nvidia/nv_setup.c par->RamAmountKBytes = (((amt >> 4) & 127) + 1) * 1024; amt 161 drivers/video/fbdev/riva/nv_driver.c u32 amt; amt 231 drivers/video/fbdev/riva/nv_driver.c pci_read_config_dword(dev, 0x7C, &amt); amt 233 drivers/video/fbdev/riva/nv_driver.c memlen = (((amt >> 6) & 31) + 1) * 1024; amt 236 drivers/video/fbdev/riva/nv_driver.c pci_read_config_dword(dev, 0x84, &amt); amt 238 drivers/video/fbdev/riva/nv_driver.c memlen = (((amt >> 4) & 127) + 1) * 1024; amt 2114 drivers/video/fbdev/riva/riva_hw.c u32 amt; amt 2127 drivers/video/fbdev/riva/riva_hw.c pci_read_config_dword(dev, 0x7C, &amt); amt 2129 drivers/video/fbdev/riva/riva_hw.c chip->RamAmountKBytes = (((amt >> 6) & 31) + 1) * 1024; amt 2132 drivers/video/fbdev/riva/riva_hw.c pci_read_config_dword(dev, 0x84, &amt); amt 2134 drivers/video/fbdev/riva/riva_hw.c chip->RamAmountKBytes = (((amt >> 4) & 127) + 1) * 1024; amt 558 fs/gfs2/aops.c unsigned amt; amt 563 fs/gfs2/aops.c amt = size - copied; amt 565 fs/gfs2/aops.c amt = PAGE_SIZE - offset; amt 570 fs/gfs2/aops.c memcpy(buf + copied, p + offset, amt); amt 573 fs/gfs2/aops.c copied += amt; amt 1404 fs/gfs2/rgrp.c u64 amt; amt 1450 fs/gfs2/rgrp.c &amt); amt 1455 fs/gfs2/rgrp.c trimmed += amt; amt 12 include/linux/sched/xacct.h static inline void add_rchar(struct task_struct *tsk, ssize_t amt) amt 14 include/linux/sched/xacct.h tsk->ioac.rchar += amt; amt 17 include/linux/sched/xacct.h static inline void add_wchar(struct task_struct *tsk, ssize_t amt) amt 19 include/linux/sched/xacct.h tsk->ioac.wchar += amt; amt 32 include/linux/sched/xacct.h static inline void add_rchar(struct task_struct *tsk, ssize_t amt) amt 36 include/linux/sched/xacct.h static inline void add_wchar(struct task_struct *tsk, ssize_t amt) amt 1278 include/net/sock.h sk_memory_allocated_add(struct sock *sk, int amt) amt 1280 include/net/sock.h return atomic_long_add_return(amt, sk->sk_prot->memory_allocated); amt 1284 include/net/sock.h sk_memory_allocated_sub(struct sock *sk, int amt) amt 1286 include/net/sock.h atomic_long_sub(amt, sk->sk_prot->memory_allocated); amt 1381 include/net/sock.h int __sk_mem_raise_allocated(struct sock *sk, int size, int amt, int kind); amt 1407 include/net/sock.h static inline int sk_mem_pages(int amt) amt 1409 include/net/sock.h return (amt + SK_MEM_QUANTUM - 1) >> SK_MEM_QUANTUM_SHIFT; amt 2180 include/net/sock.h int amt = 0; amt 2183 include/net/sock.h amt = sk->sk_sndbuf - refcount_read(&sk->sk_wmem_alloc); amt 2184 include/net/sock.h if (amt < 0) amt 2185 include/net/sock.h amt = 0; amt 2187 include/net/sock.h return amt; amt 36 kernel/sched/stats.h #define __schedstat_add(var, amt) do { var += (amt); } while (0) amt 37 kernel/sched/stats.h #define schedstat_add(var, amt) do { if (schedstat_enabled()) { var += (amt); } } while (0) amt 50 kernel/sched/stats.h # define __schedstat_add(var, amt) do { } while (0) amt 51 kernel/sched/stats.h # define schedstat_add(var, amt) do { } while (0) amt 2500 net/core/sock.c int __sk_mem_raise_allocated(struct sock *sk, int size, int amt, int kind) amt 2503 net/core/sock.c long allocated = sk_memory_allocated_add(sk, amt); amt 2507 net/core/sock.c !(charged = mem_cgroup_charge_skmem(sk->sk_memcg, amt))) amt 2568 net/core/sock.c sk_memory_allocated_sub(sk, amt); amt 2571 net/core/sock.c mem_cgroup_uncharge_skmem(sk->sk_memcg, amt); amt 2589 net/core/sock.c int ret, amt = sk_mem_pages(size); amt 2591 net/core/sock.c sk->sk_forward_alloc += amt << SK_MEM_QUANTUM_SHIFT; amt 2592 net/core/sock.c ret = __sk_mem_raise_allocated(sk, size, amt, kind); amt 2594 net/core/sock.c sk->sk_forward_alloc -= amt << SK_MEM_QUANTUM_SHIFT; amt 494 net/ipv4/inet_connection_sock.c int amt; amt 504 net/ipv4/inet_connection_sock.c amt = sk_mem_pages(newsk->sk_forward_alloc + amt 507 net/ipv4/inet_connection_sock.c if (newsk->sk_memcg && amt) amt 508 net/ipv4/inet_connection_sock.c mem_cgroup_charge_skmem(newsk->sk_memcg, amt); amt 3126 net/ipv4/tcp_output.c int amt; amt 3130 net/ipv4/tcp_output.c amt = sk_mem_pages(size); amt 3131 net/ipv4/tcp_output.c sk->sk_forward_alloc += amt * SK_MEM_QUANTUM; amt 3132 net/ipv4/tcp_output.c sk_memory_allocated_add(sk, amt); amt 3135 net/ipv4/tcp_output.c mem_cgroup_charge_skmem(sk->sk_memcg, amt); amt 1366 net/ipv4/udp.c int amt; amt 1388 net/ipv4/udp.c amt = (sk->sk_forward_alloc - partial) & ~(SK_MEM_QUANTUM - 1); amt 1389 net/ipv4/udp.c sk->sk_forward_alloc -= amt; amt 1391 net/ipv4/udp.c if (amt) amt 1392 net/ipv4/udp.c __sk_mem_reduce_allocated(sk, amt >> SK_MEM_QUANTUM_SHIFT); amt 1450 net/ipv4/udp.c int rmem, delta, amt, err = -ENOMEM; amt 1484 net/ipv4/udp.c amt = sk_mem_pages(size); amt 1485 net/ipv4/udp.c delta = amt << SK_MEM_QUANTUM_SHIFT; amt 1486 net/ipv4/udp.c if (!__sk_mem_raise_allocated(sk, delta, amt, SK_MEM_RECV)) { amt 161 net/sctp/diag.c int amt; amt 164 net/sctp/diag.c amt = asoc->sndbuf_used; amt 166 net/sctp/diag.c amt = sk_wmem_alloc_get(sk); amt 167 net/sctp/diag.c mem[SK_MEMINFO_WMEM_ALLOC] = amt; amt 169 net/sctp/diag.c amt = atomic_read(&asoc->rmem_alloc); amt 171 net/sctp/diag.c amt = sk_rmem_alloc_get(sk); amt 172 net/sctp/diag.c mem[SK_MEMINFO_RMEM_ALLOC] = amt; amt 508 tools/testing/selftests/kvm/lib/kvm_util.c size_t amt; amt 514 tools/testing/selftests/kvm/lib/kvm_util.c for (uintptr_t offset = 0; offset < len; offset += amt) { amt 527 tools/testing/selftests/kvm/lib/kvm_util.c amt = len - offset; amt 528 tools/testing/selftests/kvm/lib/kvm_util.c if ((ptr1 >> vm->page_shift) != ((ptr1 + amt) >> vm->page_shift)) amt 529 tools/testing/selftests/kvm/lib/kvm_util.c amt = vm->page_size - (ptr1 % vm->page_size); amt 530 tools/testing/selftests/kvm/lib/kvm_util.c if ((ptr2 >> vm->page_shift) != ((ptr2 + amt) >> vm->page_shift)) amt 531 tools/testing/selftests/kvm/lib/kvm_util.c amt = vm->page_size - (ptr2 % vm->page_size); amt 533 tools/testing/selftests/kvm/lib/kvm_util.c assert((ptr1 >> vm->page_shift) == ((ptr1 + amt - 1) >> vm->page_shift)); amt 534 tools/testing/selftests/kvm/lib/kvm_util.c assert((ptr2 >> vm->page_shift) == ((ptr2 + amt - 1) >> vm->page_shift)); amt 541 tools/testing/selftests/kvm/lib/kvm_util.c int ret = memcmp((void *)ptr1, (void *)ptr2, amt);