/linux-4.1.27/fs/dlm/ |
H A D | util.c | 86 void dlm_message_out(struct dlm_message *ms) dlm_message_out() argument 88 header_out(&ms->m_header); dlm_message_out() 90 ms->m_type = cpu_to_le32(ms->m_type); dlm_message_out() 91 ms->m_nodeid = cpu_to_le32(ms->m_nodeid); dlm_message_out() 92 ms->m_pid = cpu_to_le32(ms->m_pid); dlm_message_out() 93 ms->m_lkid = cpu_to_le32(ms->m_lkid); dlm_message_out() 94 ms->m_remid = cpu_to_le32(ms->m_remid); dlm_message_out() 95 ms->m_parent_lkid = cpu_to_le32(ms->m_parent_lkid); dlm_message_out() 96 ms->m_parent_remid = cpu_to_le32(ms->m_parent_remid); dlm_message_out() 97 ms->m_exflags = cpu_to_le32(ms->m_exflags); dlm_message_out() 98 ms->m_sbflags = cpu_to_le32(ms->m_sbflags); dlm_message_out() 99 ms->m_flags = cpu_to_le32(ms->m_flags); dlm_message_out() 100 ms->m_lvbseq = cpu_to_le32(ms->m_lvbseq); dlm_message_out() 101 ms->m_hash = cpu_to_le32(ms->m_hash); dlm_message_out() 102 ms->m_status = cpu_to_le32(ms->m_status); dlm_message_out() 103 ms->m_grmode = cpu_to_le32(ms->m_grmode); dlm_message_out() 104 ms->m_rqmode = cpu_to_le32(ms->m_rqmode); dlm_message_out() 105 ms->m_bastmode = cpu_to_le32(ms->m_bastmode); dlm_message_out() 106 ms->m_asts = cpu_to_le32(ms->m_asts); dlm_message_out() 107 ms->m_result = cpu_to_le32(to_dlm_errno(ms->m_result)); dlm_message_out() 110 void dlm_message_in(struct dlm_message *ms) dlm_message_in() argument 112 header_in(&ms->m_header); dlm_message_in() 114 ms->m_type = le32_to_cpu(ms->m_type); dlm_message_in() 115 ms->m_nodeid = le32_to_cpu(ms->m_nodeid); dlm_message_in() 116 ms->m_pid = le32_to_cpu(ms->m_pid); dlm_message_in() 117 ms->m_lkid = le32_to_cpu(ms->m_lkid); dlm_message_in() 118 ms->m_remid = le32_to_cpu(ms->m_remid); dlm_message_in() 119 ms->m_parent_lkid = le32_to_cpu(ms->m_parent_lkid); dlm_message_in() 120 ms->m_parent_remid = le32_to_cpu(ms->m_parent_remid); dlm_message_in() 121 ms->m_exflags = le32_to_cpu(ms->m_exflags); dlm_message_in() 122 ms->m_sbflags = le32_to_cpu(ms->m_sbflags); dlm_message_in() 123 ms->m_flags = le32_to_cpu(ms->m_flags); dlm_message_in() 124 ms->m_lvbseq = le32_to_cpu(ms->m_lvbseq); dlm_message_in() 125 ms->m_hash = le32_to_cpu(ms->m_hash); dlm_message_in() 126 ms->m_status = le32_to_cpu(ms->m_status); dlm_message_in() 127 ms->m_grmode = le32_to_cpu(ms->m_grmode); dlm_message_in() 128 ms->m_rqmode = le32_to_cpu(ms->m_rqmode); dlm_message_in() 129 ms->m_bastmode = le32_to_cpu(ms->m_bastmode); dlm_message_in() 130 ms->m_asts = le32_to_cpu(ms->m_asts); dlm_message_in() 131 ms->m_result = from_dlm_errno(le32_to_cpu(ms->m_result)); dlm_message_in()
|
H A D | util.h | 16 void dlm_message_out(struct dlm_message *ms); 17 void dlm_message_in(struct dlm_message *ms);
|
H A D | requestqueue.c | 34 void dlm_add_requestqueue(struct dlm_ls *ls, int nodeid, struct dlm_message *ms) dlm_add_requestqueue() argument 37 int length = ms->m_header.h_length - sizeof(struct dlm_message); dlm_add_requestqueue() 47 memcpy(&e->request, ms, ms->m_header.h_length); dlm_add_requestqueue() 68 struct dlm_message *ms; dlm_process_requestqueue() local 82 ms = &e->request; dlm_process_requestqueue() 86 ms->m_type, ms->m_header.h_nodeid, dlm_process_requestqueue() 87 ms->m_lkid, ms->m_remid, ms->m_result, dlm_process_requestqueue() 130 static int purge_request(struct dlm_ls *ls, struct dlm_message *ms, int nodeid) purge_request() argument 132 uint32_t type = ms->m_type; purge_request() 157 struct dlm_message *ms; dlm_purge_requestqueue() local 162 ms = &e->request; dlm_purge_requestqueue() 164 if (purge_request(ls, ms, e->nodeid)) { dlm_purge_requestqueue()
|
H A D | lock.c | 89 struct dlm_message *ms); 90 static int receive_extralen(struct dlm_message *ms); 1512 struct dlm_message *ms) _remove_from_waiters() 1551 is_overlap_cancel(lkb) && ms && !ms->m_result) { _remove_from_waiters() 1569 lkb->lkb_id, ms ? ms->m_header.h_nodeid : 0, lkb->lkb_remid, _remove_from_waiters() 1610 static int remove_from_waiters_ms(struct dlm_lkb *lkb, struct dlm_message *ms) remove_from_waiters_ms() argument 1615 if (ms->m_flags != DLM_IFL_STUB_MS) remove_from_waiters_ms() 1617 error = _remove_from_waiters(lkb, ms->m_type, ms); remove_from_waiters_ms() 1618 if (ms->m_flags != DLM_IFL_STUB_MS) remove_from_waiters_ms() 2029 struct dlm_message *ms) set_lvb_lock_pc() 2041 int len = receive_extralen(ms); set_lvb_lock_pc() 2044 memcpy(lkb->lkb_lvbptr, ms->m_extra, len); set_lvb_lock_pc() 2045 lkb->lkb_lvbseq = ms->m_lvbseq; set_lvb_lock_pc() 2136 struct dlm_message *ms) grant_lock_pc() 2138 set_lvb_lock_pc(r, lkb, ms); grant_lock_pc() 2174 static void munge_altmode(struct dlm_lkb *lkb, struct dlm_message *ms) munge_altmode() argument 2176 if (ms->m_type != DLM_MSG_REQUEST_REPLY && munge_altmode() 2177 ms->m_type != DLM_MSG_GRANT) { munge_altmode() 2179 lkb->lkb_id, ms->m_type); munge_altmode() 3531 struct dlm_message *ms; _create_message() local 3545 ms = (struct dlm_message *) mb; _create_message() 3547 ms->m_header.h_version = (DLM_HEADER_MAJOR | DLM_HEADER_MINOR); _create_message() 3548 ms->m_header.h_lockspace = ls->ls_global_id; _create_message() 3549 ms->m_header.h_nodeid = dlm_our_nodeid(); _create_message() 3550 ms->m_header.h_length = mb_len; _create_message() 3551 ms->m_header.h_cmd = DLM_MSG; _create_message() 3553 ms->m_type = mstype; _create_message() 3556 *ms_ret = ms; _create_message() 3590 static int send_message(struct dlm_mhandle *mh, struct dlm_message *ms) send_message() argument 3592 dlm_message_out(ms); send_message() 3598 struct dlm_message *ms) send_args() 3600 ms->m_nodeid = lkb->lkb_nodeid; send_args() 3601 ms->m_pid = lkb->lkb_ownpid; send_args() 3602 ms->m_lkid = lkb->lkb_id; send_args() 3603 ms->m_remid = lkb->lkb_remid; send_args() 3604 ms->m_exflags = lkb->lkb_exflags; send_args() 3605 ms->m_sbflags = lkb->lkb_sbflags; send_args() 3606 ms->m_flags = lkb->lkb_flags; send_args() 3607 ms->m_lvbseq = lkb->lkb_lvbseq; send_args() 3608 ms->m_status = lkb->lkb_status; send_args() 3609 ms->m_grmode = lkb->lkb_grmode; send_args() 3610 ms->m_rqmode = lkb->lkb_rqmode; send_args() 3611 ms->m_hash = r->res_hash; send_args() 3617 ms->m_asts |= DLM_CB_BAST; send_args() 3619 ms->m_asts |= DLM_CB_CAST; send_args() 3624 switch (ms->m_type) { send_args() 3627 memcpy(ms->m_extra, r->res_name, r->res_length); send_args() 3636 memcpy(ms->m_extra, lkb->lkb_lvbptr, r->res_ls->ls_lvblen); send_args() 3643 struct dlm_message *ms; send_common() local 3653 error = create_message(r, lkb, to_nodeid, mstype, &ms, &mh); send_common() 3657 send_args(r, lkb, ms); send_common() 3659 error = send_message(mh, ms); send_common() 3708 struct dlm_message *ms; send_grant() local 3714 error = create_message(r, lkb, to_nodeid, DLM_MSG_GRANT, &ms, &mh); send_grant() 3718 send_args(r, lkb, ms); send_grant() 3720 ms->m_result = 0; send_grant() 3722 error = send_message(mh, ms); send_grant() 3729 struct dlm_message *ms; send_bast() local 3735 error = create_message(r, NULL, to_nodeid, DLM_MSG_BAST, &ms, &mh); send_bast() 3739 send_args(r, lkb, ms); send_bast() 3741 ms->m_bastmode = mode; send_bast() 3743 error = send_message(mh, ms); send_bast() 3750 struct dlm_message *ms; send_lookup() local 3760 error = create_message(r, NULL, to_nodeid, DLM_MSG_LOOKUP, &ms, &mh); send_lookup() 3764 send_args(r, lkb, ms); send_lookup() 3766 error = send_message(mh, ms); send_lookup() 3778 struct dlm_message *ms; send_remove() local 3784 error = create_message(r, NULL, to_nodeid, DLM_MSG_REMOVE, &ms, &mh); send_remove() 3788 memcpy(ms->m_extra, r->res_name, r->res_length); send_remove() 3789 ms->m_hash = r->res_hash; send_remove() 3791 error = send_message(mh, ms); send_remove() 3799 struct dlm_message *ms; send_common_reply() local 3805 error = create_message(r, lkb, to_nodeid, mstype, &ms, &mh); send_common_reply() 3809 send_args(r, lkb, ms); send_common_reply() 3811 ms->m_result = rv; send_common_reply() 3813 error = send_message(mh, ms); send_common_reply() 3842 struct dlm_message *ms; send_lookup_reply() local 3846 error = create_message(r, NULL, nodeid, DLM_MSG_LOOKUP_REPLY, &ms, &mh); send_lookup_reply() 3850 ms->m_lkid = ms_in->m_lkid; send_lookup_reply() 3851 ms->m_result = rv; send_lookup_reply() 3852 ms->m_nodeid = ret_nodeid; send_lookup_reply() 3854 error = send_message(mh, ms); send_lookup_reply() 3863 static void receive_flags(struct dlm_lkb *lkb, struct dlm_message *ms) receive_flags() argument 3865 lkb->lkb_exflags = ms->m_exflags; receive_flags() 3866 lkb->lkb_sbflags = ms->m_sbflags; receive_flags() 3868 (ms->m_flags & 0x0000FFFF); receive_flags() 3871 static void receive_flags_reply(struct dlm_lkb *lkb, struct dlm_message *ms) receive_flags_reply() argument 3873 if (ms->m_flags == DLM_IFL_STUB_MS) receive_flags_reply() 3876 lkb->lkb_sbflags = ms->m_sbflags; receive_flags_reply() 3878 (ms->m_flags & 0x0000FFFF); receive_flags_reply() 3881 static int receive_extralen(struct dlm_message *ms) receive_extralen() argument 3883 return (ms->m_header.h_length - sizeof(struct dlm_message)); receive_extralen() 3887 struct dlm_message *ms) receive_lvb() 3896 len = receive_extralen(ms); receive_lvb() 3899 memcpy(lkb->lkb_lvbptr, ms->m_extra, len); receive_lvb() 3915 struct dlm_message *ms) receive_request_args() 3917 lkb->lkb_nodeid = ms->m_header.h_nodeid; receive_request_args() 3918 lkb->lkb_ownpid = ms->m_pid; receive_request_args() 3919 lkb->lkb_remid = ms->m_lkid; receive_request_args() 3921 lkb->lkb_rqmode = ms->m_rqmode; receive_request_args() 3923 lkb->lkb_bastfn = (ms->m_asts & DLM_CB_BAST) ? &fake_bastfn : NULL; receive_request_args() 3924 lkb->lkb_astfn = (ms->m_asts & DLM_CB_CAST) ? &fake_astfn : NULL; receive_request_args() 3937 struct dlm_message *ms) receive_convert_args() 3942 if (receive_lvb(ls, lkb, ms)) receive_convert_args() 3945 lkb->lkb_rqmode = ms->m_rqmode; receive_convert_args() 3946 lkb->lkb_lvbseq = ms->m_lvbseq; receive_convert_args() 3952 struct dlm_message *ms) receive_unlock_args() 3954 if (receive_lvb(ls, lkb, ms)) receive_unlock_args() 3962 static void setup_stub_lkb(struct dlm_ls *ls, struct dlm_message *ms) setup_stub_lkb() argument 3965 lkb->lkb_nodeid = ms->m_header.h_nodeid; setup_stub_lkb() 3966 lkb->lkb_remid = ms->m_lkid; setup_stub_lkb() 3972 static int validate_message(struct dlm_lkb *lkb, struct dlm_message *ms) validate_message() argument 3974 int from = ms->m_header.h_nodeid; validate_message() 3977 switch (ms->m_type) { validate_message() 4008 ms->m_type, from, lkb->lkb_id, lkb->lkb_remid, validate_message() 4016 struct dlm_message *ms; send_repeat_remove() local 4056 dir_nodeid, DLM_MSG_REMOVE, &ms, &mh); send_repeat_remove() 4060 memcpy(ms->m_extra, name, len); send_repeat_remove() 4061 ms->m_hash = hash; send_repeat_remove() 4063 send_message(mh, ms); send_repeat_remove() 4071 static int receive_request(struct dlm_ls *ls, struct dlm_message *ms) receive_request() argument 4078 from_nodeid = ms->m_header.h_nodeid; receive_request() 4084 receive_flags(lkb, ms); receive_request() 4086 error = receive_request_args(ls, lkb, ms); receive_request() 4098 namelen = receive_extralen(ms); receive_request() 4100 error = find_rsb(ls, ms->m_extra, namelen, from_nodeid, receive_request() 4151 ms->m_lkid, from_nodeid, error); receive_request() 4155 send_repeat_remove(ls, ms->m_extra, namelen); receive_request() 4159 setup_stub_lkb(ls, ms); receive_request() 4164 static int receive_convert(struct dlm_ls *ls, struct dlm_message *ms) receive_convert() argument 4170 error = find_lkb(ls, ms->m_remid, &lkb); receive_convert() 4174 if (lkb->lkb_remid != ms->m_lkid) { receive_convert() 4178 ms->m_header.h_nodeid, ms->m_lkid); receive_convert() 4188 error = validate_message(lkb, ms); receive_convert() 4192 receive_flags(lkb, ms); receive_convert() 4194 error = receive_convert_args(ls, lkb, ms); receive_convert() 4213 setup_stub_lkb(ls, ms); receive_convert() 4218 static int receive_unlock(struct dlm_ls *ls, struct dlm_message *ms) receive_unlock() argument 4224 error = find_lkb(ls, ms->m_remid, &lkb); receive_unlock() 4228 if (lkb->lkb_remid != ms->m_lkid) { receive_unlock() 4231 ms->m_header.h_nodeid, ms->m_lkid); receive_unlock() 4241 error = validate_message(lkb, ms); receive_unlock() 4245 receive_flags(lkb, ms); receive_unlock() 4247 error = receive_unlock_args(ls, lkb, ms); receive_unlock() 4263 setup_stub_lkb(ls, ms); receive_unlock() 4268 static int receive_cancel(struct dlm_ls *ls, struct dlm_message *ms) receive_cancel() argument 4274 error = find_lkb(ls, ms->m_remid, &lkb); receive_cancel() 4278 receive_flags(lkb, ms); receive_cancel() 4285 error = validate_message(lkb, ms); receive_cancel() 4299 setup_stub_lkb(ls, ms); receive_cancel() 4304 static int receive_grant(struct dlm_ls *ls, struct dlm_message *ms) receive_grant() argument 4310 error = find_lkb(ls, ms->m_remid, &lkb); receive_grant() 4319 error = validate_message(lkb, ms); receive_grant() 4323 receive_flags_reply(lkb, ms); receive_grant() 4325 munge_altmode(lkb, ms); receive_grant() 4326 grant_lock_pc(r, lkb, ms); receive_grant() 4335 static int receive_bast(struct dlm_ls *ls, struct dlm_message *ms) receive_bast() argument 4341 error = find_lkb(ls, ms->m_remid, &lkb); receive_bast() 4350 error = validate_message(lkb, ms); receive_bast() 4354 queue_bast(r, lkb, ms->m_bastmode); receive_bast() 4355 lkb->lkb_highbast = ms->m_bastmode; receive_bast() 4363 static void receive_lookup(struct dlm_ls *ls, struct dlm_message *ms) receive_lookup() argument 4367 from_nodeid = ms->m_header.h_nodeid; receive_lookup() 4370 len = receive_extralen(ms); receive_lookup() 4372 error = dlm_master_lookup(ls, from_nodeid, ms->m_extra, len, 0, receive_lookup() 4377 receive_request(ls, ms); receive_lookup() 4380 send_lookup_reply(ls, ms, ret_nodeid, error); receive_lookup() 4383 static void receive_remove(struct dlm_ls *ls, struct dlm_message *ms) receive_remove() argument 4390 from_nodeid = ms->m_header.h_nodeid; receive_remove() 4392 len = receive_extralen(ms); receive_remove() 4400 dir_nodeid = dlm_hash2nodeid(ls, ms->m_hash); receive_remove() 4417 memcpy(name, ms->m_extra, len); receive_remove() 4471 static void receive_purge(struct dlm_ls *ls, struct dlm_message *ms) receive_purge() argument 4473 do_purge(ls, ms->m_nodeid, ms->m_pid); receive_purge() 4476 static int receive_request_reply(struct dlm_ls *ls, struct dlm_message *ms) receive_request_reply() argument 4481 int from_nodeid = ms->m_header.h_nodeid; receive_request_reply() 4483 error = find_lkb(ls, ms->m_remid, &lkb); receive_request_reply() 4491 error = validate_message(lkb, ms); receive_request_reply() 4499 lkb->lkb_id, from_nodeid, ms->m_lkid, ms->m_result); receive_request_reply() 4513 result = ms->m_result; receive_request_reply() 4526 receive_flags_reply(lkb, ms); receive_request_reply() 4527 lkb->lkb_remid = ms->m_lkid; receive_request_reply() 4529 munge_altmode(lkb, ms); receive_request_reply() 4534 grant_lock_pc(r, lkb, ms); receive_request_reply() 4597 struct dlm_message *ms) __receive_convert_reply() 4600 switch (ms->m_result) { __receive_convert_reply() 4607 receive_flags_reply(lkb, ms); __receive_convert_reply() 4614 receive_flags_reply(lkb, ms); __receive_convert_reply() 4624 receive_flags_reply(lkb, ms); __receive_convert_reply() 4627 grant_lock_pc(r, lkb, ms); __receive_convert_reply() 4633 lkb->lkb_id, ms->m_header.h_nodeid, ms->m_lkid, __receive_convert_reply() 4634 ms->m_result); __receive_convert_reply() 4640 static void _receive_convert_reply(struct dlm_lkb *lkb, struct dlm_message *ms) _receive_convert_reply() argument 4648 error = validate_message(lkb, ms); _receive_convert_reply() 4653 error = remove_from_waiters_ms(lkb, ms); _receive_convert_reply() 4657 __receive_convert_reply(r, lkb, ms); _receive_convert_reply() 4663 static int receive_convert_reply(struct dlm_ls *ls, struct dlm_message *ms) receive_convert_reply() argument 4668 error = find_lkb(ls, ms->m_remid, &lkb); receive_convert_reply() 4672 _receive_convert_reply(lkb, ms); receive_convert_reply() 4677 static void _receive_unlock_reply(struct dlm_lkb *lkb, struct dlm_message *ms) _receive_unlock_reply() argument 4685 error = validate_message(lkb, ms); _receive_unlock_reply() 4690 error = remove_from_waiters_ms(lkb, ms); _receive_unlock_reply() 4696 switch (ms->m_result) { _receive_unlock_reply() 4698 receive_flags_reply(lkb, ms); _receive_unlock_reply() 4706 lkb->lkb_id, ms->m_result); _receive_unlock_reply() 4713 static int receive_unlock_reply(struct dlm_ls *ls, struct dlm_message *ms) receive_unlock_reply() argument 4718 error = find_lkb(ls, ms->m_remid, &lkb); receive_unlock_reply() 4722 _receive_unlock_reply(lkb, ms); receive_unlock_reply() 4727 static void _receive_cancel_reply(struct dlm_lkb *lkb, struct dlm_message *ms) _receive_cancel_reply() argument 4735 error = validate_message(lkb, ms); _receive_cancel_reply() 4740 error = remove_from_waiters_ms(lkb, ms); _receive_cancel_reply() 4746 switch (ms->m_result) { _receive_cancel_reply() 4748 receive_flags_reply(lkb, ms); _receive_cancel_reply() 4756 lkb->lkb_id, ms->m_result); _receive_cancel_reply() 4763 static int receive_cancel_reply(struct dlm_ls *ls, struct dlm_message *ms) receive_cancel_reply() argument 4768 error = find_lkb(ls, ms->m_remid, &lkb); receive_cancel_reply() 4772 _receive_cancel_reply(lkb, ms); receive_cancel_reply() 4777 static void receive_lookup_reply(struct dlm_ls *ls, struct dlm_message *ms) receive_lookup_reply() argument 4784 error = find_lkb(ls, ms->m_lkid, &lkb); receive_lookup_reply() 4786 log_error(ls, "receive_lookup_reply no lkid %x", ms->m_lkid); receive_lookup_reply() 4790 /* ms->m_result is the value returned by dlm_master_lookup on dir node receive_lookup_reply() 4801 ret_nodeid = ms->m_nodeid; receive_lookup_reply() 4813 lkb->lkb_id, ms->m_header.h_nodeid, ret_nodeid, receive_lookup_reply() 4826 lkb->lkb_id, ms->m_header.h_nodeid); receive_lookup_reply() 4855 static void _receive_message(struct dlm_ls *ls, struct dlm_message *ms, _receive_message() argument 4860 if (!dlm_is_member(ls, ms->m_header.h_nodeid)) { _receive_message() 4862 ms->m_type, ms->m_header.h_nodeid, ms->m_lkid, _receive_message() 4863 ms->m_remid, ms->m_result); _receive_message() 4867 switch (ms->m_type) { _receive_message() 4872 error = receive_request(ls, ms); _receive_message() 4876 error = receive_convert(ls, ms); _receive_message() 4880 error = receive_unlock(ls, ms); _receive_message() 4885 error = receive_cancel(ls, ms); _receive_message() 4891 error = receive_request_reply(ls, ms); _receive_message() 4895 error = receive_convert_reply(ls, ms); _receive_message() 4899 error = receive_unlock_reply(ls, ms); _receive_message() 4903 error = receive_cancel_reply(ls, ms); _receive_message() 4910 error = receive_grant(ls, ms); _receive_message() 4915 error = receive_bast(ls, ms); _receive_message() 4921 receive_lookup(ls, ms); _receive_message() 4925 receive_remove(ls, ms); _receive_message() 4931 receive_lookup_reply(ls, ms); _receive_message() 4937 receive_purge(ls, ms); _receive_message() 4941 log_error(ls, "unknown message type %d", ms->m_type); _receive_message() 4957 ms->m_type, ms->m_remid, ms->m_header.h_nodeid, _receive_message() 4958 ms->m_lkid, saved_seq); _receive_message() 4961 ms->m_type, ms->m_remid, ms->m_header.h_nodeid, _receive_message() 4962 ms->m_lkid, saved_seq); _receive_message() 4964 if (ms->m_type == DLM_MSG_CONVERT) _receive_message() 4965 dlm_dump_rsb_hash(ls, ms->m_hash); _receive_message() 4971 ms->m_type, ms->m_header.h_nodeid, _receive_message() 4972 ms->m_lkid, ms->m_remid, saved_seq); _receive_message() 4984 static void dlm_receive_message(struct dlm_ls *ls, struct dlm_message *ms, dlm_receive_message() argument 4993 ms->m_type, nodeid); dlm_receive_message() 4997 dlm_add_requestqueue(ls, nodeid, ms); dlm_receive_message() 5000 _receive_message(ls, ms, 0); dlm_receive_message() 5007 void dlm_receive_message_saved(struct dlm_ls *ls, struct dlm_message *ms, dlm_receive_message_saved() argument 5010 _receive_message(ls, ms, saved_seq); dlm_receive_message_saved() 6273 struct dlm_message *ms; send_purge() local 6278 DLM_MSG_PURGE, &ms, &mh); send_purge() 6281 ms->m_nodeid = nodeid; send_purge() 6282 ms->m_pid = pid; send_purge() 6284 return send_message(mh, ms); send_purge() 1511 _remove_from_waiters(struct dlm_lkb *lkb, int mstype, struct dlm_message *ms) _remove_from_waiters() argument 2028 set_lvb_lock_pc(struct dlm_rsb *r, struct dlm_lkb *lkb, struct dlm_message *ms) set_lvb_lock_pc() argument 2135 grant_lock_pc(struct dlm_rsb *r, struct dlm_lkb *lkb, struct dlm_message *ms) grant_lock_pc() argument 3597 send_args(struct dlm_rsb *r, struct dlm_lkb *lkb, struct dlm_message *ms) send_args() argument 3886 receive_lvb(struct dlm_ls *ls, struct dlm_lkb *lkb, struct dlm_message *ms) receive_lvb() argument 3914 receive_request_args(struct dlm_ls *ls, struct dlm_lkb *lkb, struct dlm_message *ms) receive_request_args() argument 3936 receive_convert_args(struct dlm_ls *ls, struct dlm_lkb *lkb, struct dlm_message *ms) receive_convert_args() argument 3951 receive_unlock_args(struct dlm_ls *ls, struct dlm_lkb *lkb, struct dlm_message *ms) receive_unlock_args() argument 4596 __receive_convert_reply(struct dlm_rsb *r, struct dlm_lkb *lkb, struct dlm_message *ms) __receive_convert_reply() argument
|
H A D | requestqueue.h | 16 void dlm_add_requestqueue(struct dlm_ls *ls, int nodeid, struct dlm_message *ms);
|
/linux-4.1.27/drivers/staging/dgnc/ |
H A D | dgnc_utils.h | 4 int dgnc_ms_sleep(ulong ms);
|
H A D | dgnc_utils.c | 9 * Put the driver to sleep for x ms's 13 int dgnc_ms_sleep(ulong ms) dgnc_ms_sleep() argument 16 schedule_timeout((ms * HZ) / 1000); dgnc_ms_sleep()
|
/linux-4.1.27/drivers/spi/ |
H A D | spi-mpc52xx.c | 83 int (*state)(int irq, struct mpc52xx_spi *ms, u8 status, u8 data); 96 static void mpc52xx_spi_chipsel(struct mpc52xx_spi *ms, int value) mpc52xx_spi_chipsel() argument 100 if (ms->gpio_cs_count > 0) { mpc52xx_spi_chipsel() 101 cs = ms->message->spi->chip_select; mpc52xx_spi_chipsel() 102 gpio_set_value(ms->gpio_cs[cs], value ? 0 : 1); mpc52xx_spi_chipsel() 104 out_8(ms->regs + SPI_PORTDATA, value ? 0 : 0x08); mpc52xx_spi_chipsel() 112 static void mpc52xx_spi_start_transfer(struct mpc52xx_spi *ms) mpc52xx_spi_start_transfer() argument 114 ms->rx_buf = ms->transfer->rx_buf; mpc52xx_spi_start_transfer() 115 ms->tx_buf = ms->transfer->tx_buf; mpc52xx_spi_start_transfer() 116 ms->len = ms->transfer->len; mpc52xx_spi_start_transfer() 119 if (ms->cs_change) mpc52xx_spi_start_transfer() 120 mpc52xx_spi_chipsel(ms, 1); mpc52xx_spi_start_transfer() 121 ms->cs_change = ms->transfer->cs_change; mpc52xx_spi_start_transfer() 124 ms->wcol_tx_timestamp = get_tbl(); mpc52xx_spi_start_transfer() 125 if (ms->tx_buf) mpc52xx_spi_start_transfer() 126 out_8(ms->regs + SPI_DATA, *ms->tx_buf++); mpc52xx_spi_start_transfer() 128 out_8(ms->regs + SPI_DATA, 0); mpc52xx_spi_start_transfer() 132 static int mpc52xx_spi_fsmstate_transfer(int irq, struct mpc52xx_spi *ms, 134 static int mpc52xx_spi_fsmstate_wait(int irq, struct mpc52xx_spi *ms, 144 mpc52xx_spi_fsmstate_idle(int irq, struct mpc52xx_spi *ms, u8 status, u8 data) mpc52xx_spi_fsmstate_idle() argument 151 dev_err(&ms->master->dev, "spurious irq, status=0x%.2x\n", mpc52xx_spi_fsmstate_idle() 155 if (list_empty(&ms->queue)) mpc52xx_spi_fsmstate_idle() 159 ms->message = list_first_entry(&ms->queue, struct spi_message, queue); mpc52xx_spi_fsmstate_idle() 160 list_del_init(&ms->message->queue); mpc52xx_spi_fsmstate_idle() 164 spi = ms->message->spi; mpc52xx_spi_fsmstate_idle() 171 out_8(ms->regs + SPI_CTRL1, ctrl1); mpc52xx_spi_fsmstate_idle() 176 sppr = ((ms->ipb_freq / ms->message->spi->max_speed_hz) + 1) >> 1; mpc52xx_spi_fsmstate_idle() 190 out_8(ms->regs + SPI_BRR, sppr << 4 | spr); /* Set speed */ mpc52xx_spi_fsmstate_idle() 192 ms->cs_change = 1; mpc52xx_spi_fsmstate_idle() 193 ms->transfer = container_of(ms->message->transfers.next, mpc52xx_spi_fsmstate_idle() 196 mpc52xx_spi_start_transfer(ms); mpc52xx_spi_fsmstate_idle() 197 ms->state = mpc52xx_spi_fsmstate_transfer; mpc52xx_spi_fsmstate_idle() 210 static int mpc52xx_spi_fsmstate_transfer(int irq, struct mpc52xx_spi *ms, mpc52xx_spi_fsmstate_transfer() argument 214 return ms->irq0 ? FSM_STOP : FSM_POLL; mpc52xx_spi_fsmstate_transfer() 224 ms->wcol_count++; mpc52xx_spi_fsmstate_transfer() 225 ms->wcol_ticks += get_tbl() - ms->wcol_tx_timestamp; mpc52xx_spi_fsmstate_transfer() 226 ms->wcol_tx_timestamp = get_tbl(); mpc52xx_spi_fsmstate_transfer() 228 if (ms->tx_buf) mpc52xx_spi_fsmstate_transfer() 229 data = *(ms->tx_buf - 1); mpc52xx_spi_fsmstate_transfer() 230 out_8(ms->regs + SPI_DATA, data); /* try again */ mpc52xx_spi_fsmstate_transfer() 233 ms->modf_count++; mpc52xx_spi_fsmstate_transfer() 234 dev_err(&ms->master->dev, "mode fault\n"); mpc52xx_spi_fsmstate_transfer() 235 mpc52xx_spi_chipsel(ms, 0); mpc52xx_spi_fsmstate_transfer() 236 ms->message->status = -EIO; mpc52xx_spi_fsmstate_transfer() 237 if (ms->message->complete) mpc52xx_spi_fsmstate_transfer() 238 ms->message->complete(ms->message->context); mpc52xx_spi_fsmstate_transfer() 239 ms->state = mpc52xx_spi_fsmstate_idle; mpc52xx_spi_fsmstate_transfer() 244 ms->byte_count++; mpc52xx_spi_fsmstate_transfer() 245 if (ms->rx_buf) mpc52xx_spi_fsmstate_transfer() 246 *ms->rx_buf++ = data; mpc52xx_spi_fsmstate_transfer() 249 ms->len--; mpc52xx_spi_fsmstate_transfer() 250 if (ms->len == 0) { mpc52xx_spi_fsmstate_transfer() 251 ms->timestamp = get_tbl(); mpc52xx_spi_fsmstate_transfer() 252 ms->timestamp += ms->transfer->delay_usecs * tb_ticks_per_usec; mpc52xx_spi_fsmstate_transfer() 253 ms->state = mpc52xx_spi_fsmstate_wait; mpc52xx_spi_fsmstate_transfer() 258 ms->wcol_tx_timestamp = get_tbl(); mpc52xx_spi_fsmstate_transfer() 259 if (ms->tx_buf) mpc52xx_spi_fsmstate_transfer() 260 out_8(ms->regs + SPI_DATA, *ms->tx_buf++); mpc52xx_spi_fsmstate_transfer() 262 out_8(ms->regs + SPI_DATA, 0); mpc52xx_spi_fsmstate_transfer() 274 mpc52xx_spi_fsmstate_wait(int irq, struct mpc52xx_spi *ms, u8 status, u8 data) mpc52xx_spi_fsmstate_wait() argument 277 dev_err(&ms->master->dev, "spurious irq, status=0x%.2x\n", mpc52xx_spi_fsmstate_wait() 280 if (((int)get_tbl()) - ms->timestamp < 0) mpc52xx_spi_fsmstate_wait() 283 ms->message->actual_length += ms->transfer->len; mpc52xx_spi_fsmstate_wait() 288 if (ms->transfer->transfer_list.next == &ms->message->transfers) { mpc52xx_spi_fsmstate_wait() 289 ms->msg_count++; mpc52xx_spi_fsmstate_wait() 290 mpc52xx_spi_chipsel(ms, 0); mpc52xx_spi_fsmstate_wait() 291 ms->message->status = 0; mpc52xx_spi_fsmstate_wait() 292 if (ms->message->complete) mpc52xx_spi_fsmstate_wait() 293 ms->message->complete(ms->message->context); mpc52xx_spi_fsmstate_wait() 294 ms->state = mpc52xx_spi_fsmstate_idle; mpc52xx_spi_fsmstate_wait() 300 if (ms->cs_change) mpc52xx_spi_fsmstate_wait() 301 mpc52xx_spi_chipsel(ms, 0); mpc52xx_spi_fsmstate_wait() 303 ms->transfer = container_of(ms->transfer->transfer_list.next, mpc52xx_spi_fsmstate_wait() 305 mpc52xx_spi_start_transfer(ms); mpc52xx_spi_fsmstate_wait() 306 ms->state = mpc52xx_spi_fsmstate_transfer; mpc52xx_spi_fsmstate_wait() 313 * @ms: pointer to mpc52xx_spi driver data 315 static void mpc52xx_spi_fsm_process(int irq, struct mpc52xx_spi *ms) mpc52xx_spi_fsm_process() argument 323 status = in_8(ms->regs + SPI_STATUS); mpc52xx_spi_fsm_process() 324 data = in_8(ms->regs + SPI_DATA); mpc52xx_spi_fsm_process() 325 rc = ms->state(irq, ms, status, data); mpc52xx_spi_fsm_process() 329 schedule_work(&ms->work); mpc52xx_spi_fsm_process() 337 struct mpc52xx_spi *ms = _ms; mpc52xx_spi_irq() local 338 spin_lock(&ms->lock); mpc52xx_spi_irq() 339 mpc52xx_spi_fsm_process(irq, ms); mpc52xx_spi_irq() 340 spin_unlock(&ms->lock); mpc52xx_spi_irq() 349 struct mpc52xx_spi *ms = container_of(work, struct mpc52xx_spi, work); mpc52xx_spi_wq() local 352 spin_lock_irqsave(&ms->lock, flags); mpc52xx_spi_wq() 353 mpc52xx_spi_fsm_process(0, ms); mpc52xx_spi_wq() 354 spin_unlock_irqrestore(&ms->lock, flags); mpc52xx_spi_wq() 363 struct mpc52xx_spi *ms = spi_master_get_devdata(spi->master); mpc52xx_spi_transfer() local 369 spin_lock_irqsave(&ms->lock, flags); mpc52xx_spi_transfer() 370 list_add_tail(&m->queue, &ms->queue); mpc52xx_spi_transfer() 371 spin_unlock_irqrestore(&ms->lock, flags); mpc52xx_spi_transfer() 372 schedule_work(&ms->work); mpc52xx_spi_transfer() 383 struct mpc52xx_spi *ms; mpc52xx_spi_probe() local 417 master = spi_alloc_master(&op->dev, sizeof *ms); mpc52xx_spi_probe() 430 ms = spi_master_get_devdata(master); mpc52xx_spi_probe() 431 ms->master = master; mpc52xx_spi_probe() 432 ms->regs = regs; mpc52xx_spi_probe() 433 ms->irq0 = irq_of_parse_and_map(op->dev.of_node, 0); mpc52xx_spi_probe() 434 ms->irq1 = irq_of_parse_and_map(op->dev.of_node, 1); mpc52xx_spi_probe() 435 ms->state = mpc52xx_spi_fsmstate_idle; mpc52xx_spi_probe() 436 ms->ipb_freq = mpc5xxx_get_bus_frequency(op->dev.of_node); mpc52xx_spi_probe() 437 ms->gpio_cs_count = of_gpio_count(op->dev.of_node); mpc52xx_spi_probe() 438 if (ms->gpio_cs_count > 0) { mpc52xx_spi_probe() 439 master->num_chipselect = ms->gpio_cs_count; mpc52xx_spi_probe() 440 ms->gpio_cs = kmalloc(ms->gpio_cs_count * sizeof(unsigned int), mpc52xx_spi_probe() 442 if (!ms->gpio_cs) { mpc52xx_spi_probe() 447 for (i = 0; i < ms->gpio_cs_count; i++) { mpc52xx_spi_probe() 466 ms->gpio_cs[i] = gpio_cs; mpc52xx_spi_probe() 470 spin_lock_init(&ms->lock); mpc52xx_spi_probe() 471 INIT_LIST_HEAD(&ms->queue); mpc52xx_spi_probe() 472 INIT_WORK(&ms->work, mpc52xx_spi_wq); mpc52xx_spi_probe() 475 if (ms->irq0 && ms->irq1) { mpc52xx_spi_probe() 476 rc = request_irq(ms->irq0, mpc52xx_spi_irq, 0, mpc52xx_spi_probe() 477 "mpc5200-spi-modf", ms); mpc52xx_spi_probe() 478 rc |= request_irq(ms->irq1, mpc52xx_spi_irq, 0, mpc52xx_spi_probe() 479 "mpc5200-spi-spif", ms); mpc52xx_spi_probe() 481 free_irq(ms->irq0, ms); mpc52xx_spi_probe() 482 free_irq(ms->irq1, ms); mpc52xx_spi_probe() 483 ms->irq0 = ms->irq1 = 0; mpc52xx_spi_probe() 487 ms->irq0 = ms->irq1 = 0; mpc52xx_spi_probe() 490 if (!ms->irq0) mpc52xx_spi_probe() 498 dev_info(&ms->master->dev, "registered MPC5200 SPI bus\n"); mpc52xx_spi_probe() 503 dev_err(&ms->master->dev, "initialization failed\n"); mpc52xx_spi_probe() 506 gpio_free(ms->gpio_cs[i]); mpc52xx_spi_probe() 508 kfree(ms->gpio_cs); mpc52xx_spi_probe() 520 struct mpc52xx_spi *ms = spi_master_get_devdata(master); mpc52xx_spi_remove() local 523 free_irq(ms->irq0, ms); mpc52xx_spi_remove() 524 free_irq(ms->irq1, ms); mpc52xx_spi_remove() 526 for (i = 0; i < ms->gpio_cs_count; i++) mpc52xx_spi_remove() 527 gpio_free(ms->gpio_cs[i]); mpc52xx_spi_remove() 529 kfree(ms->gpio_cs); mpc52xx_spi_remove() 531 iounmap(ms->regs); mpc52xx_spi_remove()
|
/linux-4.1.27/drivers/scsi/ |
H A D | mesh.c | 185 static void mesh_done(struct mesh_state *ms, int start_next); 186 static void mesh_interrupt(struct mesh_state *ms); 187 static void cmd_complete(struct mesh_state *ms); 188 static void set_dma_cmds(struct mesh_state *ms, struct scsi_cmnd *cmd); 189 static void halt_dma(struct mesh_state *ms); 190 static void phase_mismatch(struct mesh_state *ms); 212 static void dlog(struct mesh_state *ms, char *fmt, int a) dlog() argument 214 struct mesh_target *tp = &ms->tgts[ms->conn_tgt]; dlog() 218 slp = &ms->log[ms->log_ix]; dlog() 221 tlp->phase = (ms->msgphase << 4) + ms->phase; dlog() 222 tlp->bs0 = ms->mesh->bus_status0; dlog() 223 tlp->bs1 = ms->mesh->bus_status1; dlog() 224 tlp->tgt = ms->conn_tgt; dlog() 231 if (++ms->log_ix >= N_DBG_SLOG) dlog() 232 ms->log_ix = 0; dlog() 233 if (ms->n_log < N_DBG_SLOG) dlog() 234 ++ms->n_log; dlog() 237 static void dumplog(struct mesh_state *ms, int t) dumplog() argument 239 struct mesh_target *tp = &ms->tgts[t]; dumplog() 263 static void dumpslog(struct mesh_state *ms) dumpslog() argument 268 if (ms->n_log == 0) dumpslog() 270 i = ms->log_ix - ms->n_log; dumpslog() 273 ms->n_log = 0; dumpslog() 275 lp = &ms->log[i]; dumpslog() 285 } while (i != ms->log_ix); dumpslog() 290 static inline void dlog(struct mesh_state *ms, char *fmt, int a) dlog() argument 292 static inline void dumplog(struct mesh_state *ms, int tgt) dumplog() argument 294 static inline void dumpslog(struct mesh_state *ms) dumpslog() argument 302 mesh_dump_regs(struct mesh_state *ms) mesh_dump_regs() argument 304 volatile struct mesh_regs __iomem *mr = ms->mesh; mesh_dump_regs() 305 volatile struct dbdma_regs __iomem *md = ms->dma; mesh_dump_regs() 310 ms, mr, md); mesh_dump_regs() 322 ms->phase, ms->msgphase, ms->conn_tgt, ms->data_ptr); mesh_dump_regs() 324 ms->dma_started, ms->dma_count, ms->n_msgout); mesh_dump_regs() 326 tp = &ms->tgts[t]; mesh_dump_regs() 347 static void mesh_completed(struct mesh_state *ms, struct scsi_cmnd *cmd) mesh_completed() argument 357 static void mesh_init(struct mesh_state *ms) mesh_init() argument 359 volatile struct mesh_regs __iomem *mr = ms->mesh; mesh_init() 360 volatile struct dbdma_regs __iomem *md = ms->dma; mesh_init() 373 out_8(&mr->source_id, ms->host->this_id); mesh_init() 374 out_8(&mr->sel_timeout, 25); /* 250ms */ mesh_init() 399 ms->phase = idle; mesh_init() 400 ms->msgphase = msg_none; mesh_init() 404 static void mesh_start_cmd(struct mesh_state *ms, struct scsi_cmnd *cmd) mesh_start_cmd() argument 406 volatile struct mesh_regs __iomem *mr = ms->mesh; mesh_start_cmd() 410 ms->current_req = cmd; mesh_start_cmd() 411 ms->tgts[id].data_goes_out = cmd->sc_data_direction == DMA_TO_DEVICE; mesh_start_cmd() 412 ms->tgts[id].current_req = cmd; mesh_start_cmd() 424 if (ms->dma_started) mesh_start_cmd() 427 ms->phase = arbitrating; mesh_start_cmd() 428 ms->msgphase = msg_none; mesh_start_cmd() 429 ms->data_ptr = 0; mesh_start_cmd() 430 ms->dma_started = 0; mesh_start_cmd() 431 ms->n_msgout = 0; mesh_start_cmd() 432 ms->last_n_msgout = 0; mesh_start_cmd() 433 ms->expect_reply = 0; mesh_start_cmd() 434 ms->conn_tgt = id; mesh_start_cmd() 435 ms->tgts[id].saved_ptr = 0; mesh_start_cmd() 436 ms->stat = DID_OK; mesh_start_cmd() 437 ms->aborting = 0; mesh_start_cmd() 439 ms->tgts[id].n_log = 0; mesh_start_cmd() 440 dlog(ms, "start cmd=%x", (int) cmd); mesh_start_cmd() 444 dlog(ms, "about to arb, intr/exc/err/fc=%.8x", mesh_start_cmd() 456 dlog(ms, "busy b4 arb, intr/exc/err/fc=%.8x", mesh_start_cmd() 463 dlog(ms, "intr b4 arb, intr/exc/err/fc=%.8x", mesh_start_cmd() 466 mesh_interrupt(ms); mesh_start_cmd() 467 if (ms->phase != arbitrating) mesh_start_cmd() 474 ms->stat = DID_BUS_BUSY; mesh_start_cmd() 475 ms->phase = idle; mesh_start_cmd() 476 mesh_done(ms, 0); mesh_start_cmd() 501 dlog(ms, "intr after disresel, intr/exc/err/fc=%.8x", mesh_start_cmd() 504 mesh_interrupt(ms); mesh_start_cmd() 505 if (ms->phase != arbitrating) mesh_start_cmd() 507 dlog(ms, "after intr after disresel, intr/exc/err/fc=%.8x", mesh_start_cmd() 519 dlog(ms, "after arb, intr/exc/err/fc=%.8x", mesh_start_cmd() 524 dlog(ms, "resel? after arb, intr/exc/err/fc=%.8x", mesh_start_cmd() 535 dlog(ms, "tried reset after arb, intr/exc/err/fc=%.8x", mesh_start_cmd() 557 static void mesh_start(struct mesh_state *ms) mesh_start() argument 561 if (ms->phase != idle || ms->current_req != NULL) { mesh_start() 562 printk(KERN_ERR "inappropriate mesh_start (phase=%d, ms=%p)", mesh_start() 563 ms->phase, ms); mesh_start() 567 while (ms->phase == idle) { mesh_start() 569 for (cmd = ms->request_q; ; cmd = (struct scsi_cmnd *) cmd->host_scribble) { mesh_start() 572 if (ms->tgts[cmd->device->id].current_req == NULL) mesh_start() 578 ms->request_q = next; mesh_start() 582 ms->request_qtail = prev; mesh_start() 584 mesh_start_cmd(ms, cmd); mesh_start() 588 static void mesh_done(struct mesh_state *ms, int start_next) mesh_done() argument 591 struct mesh_target *tp = &ms->tgts[ms->conn_tgt]; mesh_done() 593 cmd = ms->current_req; mesh_done() 594 ms->current_req = NULL; mesh_done() 597 cmd->result = (ms->stat << 16) + cmd->SCp.Status; mesh_done() 598 if (ms->stat == DID_OK) mesh_done() 602 cmd->result, ms->data_ptr, scsi_bufflen(cmd)); mesh_done() 613 cmd->SCp.this_residual -= ms->data_ptr; mesh_done() 614 mesh_completed(ms, cmd); mesh_done() 617 out_8(&ms->mesh->sequence, SEQ_ENBRESEL); mesh_done() 618 mesh_flush_io(ms->mesh); mesh_done() 620 ms->phase = idle; mesh_done() 621 mesh_start(ms); mesh_done() 625 static inline void add_sdtr_msg(struct mesh_state *ms) add_sdtr_msg() argument 627 int i = ms->n_msgout; add_sdtr_msg() 629 ms->msgout[i] = EXTENDED_MESSAGE; add_sdtr_msg() 630 ms->msgout[i+1] = 3; add_sdtr_msg() 631 ms->msgout[i+2] = EXTENDED_SDTR; add_sdtr_msg() 632 ms->msgout[i+3] = mesh_sync_period/4; add_sdtr_msg() 633 ms->msgout[i+4] = (ALLOW_SYNC(ms->conn_tgt)? mesh_sync_offset: 0); add_sdtr_msg() 634 ms->n_msgout = i + 5; add_sdtr_msg() 637 static void set_sdtr(struct mesh_state *ms, int period, int offset) set_sdtr() argument 639 struct mesh_target *tp = &ms->tgts[ms->conn_tgt]; set_sdtr() 640 volatile struct mesh_regs __iomem *mr = ms->mesh; set_sdtr() 648 ms->conn_tgt); set_sdtr() 657 v = (ms->clk_freq / 5000) * period; set_sdtr() 662 tr = (ms->clk_freq + 250000) / 500000; set_sdtr() 668 tr = ((ms->clk_freq / (v + 2)) + 199999) / 200000; set_sdtr() 675 ms->conn_tgt, tr/10, tr%10); set_sdtr() 678 static void start_phase(struct mesh_state *ms) start_phase() argument 681 volatile struct mesh_regs __iomem *mr = ms->mesh; start_phase() 682 volatile struct dbdma_regs __iomem *md = ms->dma; start_phase() 683 struct scsi_cmnd *cmd = ms->current_req; start_phase() 684 struct mesh_target *tp = &ms->tgts[ms->conn_tgt]; start_phase() 686 dlog(ms, "start_phase nmo/exc/fc/seq = %.8x", start_phase() 687 MKWORD(ms->n_msgout, mr->exception, mr->fifo_count, mr->sequence)); start_phase() 689 seq = use_active_neg + (ms->n_msgout? SEQ_ATN: 0); start_phase() 690 switch (ms->msgphase) { start_phase() 698 ms->n_msgin = 0; start_phase() 707 if (ms->n_msgout <= 0) { start_phase() 709 ms->n_msgout); start_phase() 710 mesh_dump_regs(ms); start_phase() 711 ms->msgphase = msg_none; start_phase() 714 if (ALLOW_DEBUG(ms->conn_tgt)) { start_phase() 716 ms->n_msgout); start_phase() 717 for (i = 0; i < ms->n_msgout; ++i) start_phase() 718 printk(" %x", ms->msgout[i]); start_phase() 721 dlog(ms, "msgout msg=%.8x", MKWORD(ms->n_msgout, ms->msgout[0], start_phase() 722 ms->msgout[1], ms->msgout[2])); start_phase() 732 dlog(ms, "bus0 was %.2x explicitly asserting ATN", mr->bus_status0); start_phase() 739 dlog(ms,"hace: after explicit ATN bus0=%.2x",mr->bus_status0); start_phase() 741 if (ms->n_msgout == 1) { start_phase() 748 cmd_complete(ms); start_phase() 750 out_8(&mr->count_lo, ms->n_msgout - 1); start_phase() 752 for (i = 0; i < ms->n_msgout - 1; ++i) start_phase() 753 out_8(&mr->fifo, ms->msgout[i]); start_phase() 759 ms->msgphase); start_phase() 762 switch (ms->phase) { start_phase() 764 out_8(&mr->dest_id, ms->conn_tgt); start_phase() 784 if (!ms->dma_started) { start_phase() 785 set_dma_cmds(ms, cmd); start_phase() 786 out_le32(&md->cmdptr, virt_to_phys(ms->dma_cmds)); start_phase() 788 ms->dma_started = 1; start_phase() 790 nb = ms->dma_count; start_phase() 793 ms->dma_count -= nb; start_phase() 794 ms->data_ptr += nb; start_phase() 810 dlog(ms, "enbresel intr/exc/err/fc=%.8x", start_phase() 817 ms->phase); start_phase() 818 dumpslog(ms); start_phase() 823 static inline void get_msgin(struct mesh_state *ms) get_msgin() argument 825 volatile struct mesh_regs __iomem *mr = ms->mesh; get_msgin() 830 i = ms->n_msgin; get_msgin() 831 ms->n_msgin = i + n; get_msgin() 833 ms->msgin[i++] = in_8(&mr->fifo); get_msgin() 837 static inline int msgin_length(struct mesh_state *ms) msgin_length() argument 842 if (ms->n_msgin > 0) { msgin_length() 843 b = ms->msgin[0]; msgin_length() 846 n = ms->n_msgin < 2? 2: ms->msgin[1] + 2; msgin_length() 855 static void reselected(struct mesh_state *ms) reselected() argument 857 volatile struct mesh_regs __iomem *mr = ms->mesh; reselected() 862 switch (ms->phase) { reselected() 866 if ((cmd = ms->current_req) != NULL) { reselected() 868 cmd->host_scribble = (void *) ms->request_q; reselected() 869 if (ms->request_q == NULL) reselected() 870 ms->request_qtail = cmd; reselected() 871 ms->request_q = cmd; reselected() 872 tp = &ms->tgts[cmd->device->id]; reselected() 877 ms->phase = reselecting; reselected() 878 mesh_done(ms, 0); reselected() 884 ms->msgphase, ms->phase, ms->conn_tgt); reselected() 885 dumplog(ms, ms->conn_tgt); reselected() 886 dumpslog(ms); reselected() 889 if (ms->dma_started) { reselected() 891 halt_dma(ms); reselected() 893 ms->current_req = NULL; reselected() 894 ms->phase = dataing; reselected() 895 ms->msgphase = msg_in; reselected() 896 ms->n_msgout = 0; reselected() 897 ms->last_n_msgout = 0; reselected() 898 prev = ms->conn_tgt; reselected() 912 dlog(ms, "extra resel err/exc/fc = %.6x", reselected() 928 ms->conn_tgt = ms->host->this_id; reselected() 934 dlog(ms, "reseldata %x", b); reselected() 937 if ((b & (1 << t)) != 0 && t != ms->host->this_id) reselected() 939 if (b != (1 << t) + (1 << ms->host->this_id)) { reselected() 941 ms->conn_tgt = ms->host->this_id; reselected() 949 ms->conn_tgt = t; reselected() 950 tp = &ms->tgts[t]; reselected() 957 ms->current_req = tp->current_req; reselected() 962 ms->data_ptr = tp->saved_ptr; reselected() 963 dlog(ms, "resel prev tgt=%d", prev); reselected() 964 dlog(ms, "resel err/exc=%.4x", MKWORD(0, 0, mr->error, mr->exception)); reselected() 965 start_phase(ms); reselected() 969 dumplog(ms, ms->conn_tgt); reselected() 970 dumpslog(ms); reselected() 971 ms->data_ptr = 0; reselected() 972 ms->aborting = 1; reselected() 973 start_phase(ms); reselected() 976 static void do_abort(struct mesh_state *ms) do_abort() argument 978 ms->msgout[0] = ABORT; do_abort() 979 ms->n_msgout = 1; do_abort() 980 ms->aborting = 1; do_abort() 981 ms->stat = DID_ABORT; do_abort() 982 dlog(ms, "abort", 0); do_abort() 985 static void handle_reset(struct mesh_state *ms) handle_reset() argument 990 volatile struct mesh_regs __iomem *mr = ms->mesh; handle_reset() 993 tp = &ms->tgts[tgt]; handle_reset() 997 mesh_completed(ms, cmd); handle_reset() 999 ms->tgts[tgt].sdtr_state = do_sdtr; handle_reset() 1000 ms->tgts[tgt].sync_params = ASYNC_PARAMS; handle_reset() 1002 ms->current_req = NULL; handle_reset() 1003 while ((cmd = ms->request_q) != NULL) { handle_reset() 1004 ms->request_q = (struct scsi_cmnd *) cmd->host_scribble; handle_reset() 1006 mesh_completed(ms, cmd); handle_reset() 1008 ms->phase = idle; handle_reset() 1009 ms->msgphase = msg_none; handle_reset() 1021 struct mesh_state *ms = dev_id; do_mesh_interrupt() local 1022 struct Scsi_Host *dev = ms->host; do_mesh_interrupt() 1025 mesh_interrupt(ms); do_mesh_interrupt() 1030 static void handle_error(struct mesh_state *ms) handle_error() argument 1033 volatile struct mesh_regs __iomem *mr = ms->mesh; handle_error() 1038 dlog(ms, "error err/exc/fc/cl=%.8x", handle_error() 1047 handle_reset(ms); handle_error() 1054 reselected(ms); handle_error() 1057 if (!ms->aborting) { handle_error() 1059 ms->conn_tgt); handle_error() 1060 dumplog(ms, ms->conn_tgt); handle_error() 1061 dumpslog(ms); handle_error() 1064 ms->stat = DID_ABORT; handle_error() 1065 mesh_done(ms, 1); handle_error() 1069 if (ms->msgphase == msg_in) { handle_error() 1071 ms->conn_tgt); handle_error() 1072 ms->msgout[0] = MSG_PARITY_ERROR; handle_error() 1073 ms->n_msgout = 1; handle_error() 1074 ms->msgphase = msg_in_bad; handle_error() 1075 cmd_complete(ms); handle_error() 1078 if (ms->stat == DID_OK) { handle_error() 1080 ms->conn_tgt); handle_error() 1081 ms->stat = DID_PARITY; handle_error() 1085 cmd_complete(ms); handle_error() 1098 reselected(ms); handle_error() 1104 phase_mismatch(ms); handle_error() 1112 mesh_dump_regs(ms); handle_error() 1113 dumplog(ms, ms->conn_tgt); handle_error() 1114 if (ms->phase > selecting && (in_8(&mr->bus_status1) & BS1_BSY)) { handle_error() 1116 do_abort(ms); handle_error() 1117 phase_mismatch(ms); handle_error() 1120 ms->stat = DID_ERROR; handle_error() 1121 mesh_done(ms, 1); handle_error() 1124 static void handle_exception(struct mesh_state *ms) handle_exception() argument 1127 volatile struct mesh_regs __iomem *mr = ms->mesh; handle_exception() 1134 reselected(ms); handle_exception() 1137 ms->stat = DID_BUS_BUSY; handle_exception() 1138 mesh_done(ms, 1); handle_exception() 1141 ms->stat = DID_BAD_TARGET; handle_exception() 1142 mesh_done(ms, 1); handle_exception() 1146 phase_mismatch(ms); handle_exception() 1149 mesh_dump_regs(ms); handle_exception() 1150 dumplog(ms, ms->conn_tgt); handle_exception() 1151 do_abort(ms); handle_exception() 1152 phase_mismatch(ms); handle_exception() 1156 static void handle_msgin(struct mesh_state *ms) handle_msgin() argument 1159 struct scsi_cmnd *cmd = ms->current_req; handle_msgin() 1160 struct mesh_target *tp = &ms->tgts[ms->conn_tgt]; handle_msgin() 1162 if (ms->n_msgin == 0) handle_msgin() 1164 code = ms->msgin[0]; handle_msgin() 1165 if (ALLOW_DEBUG(ms->conn_tgt)) { handle_msgin() 1166 printk(KERN_DEBUG "got %d message bytes:", ms->n_msgin); handle_msgin() 1167 for (i = 0; i < ms->n_msgin; ++i) handle_msgin() 1168 printk(" %x", ms->msgin[i]); handle_msgin() 1171 dlog(ms, "msgin msg=%.8x", handle_msgin() 1172 MKWORD(ms->n_msgin, code, ms->msgin[1], ms->msgin[2])); handle_msgin() 1174 ms->expect_reply = 0; handle_msgin() 1175 ms->n_msgout = 0; handle_msgin() 1176 if (ms->n_msgin < msgin_length(ms)) handle_msgin() 1184 switch (ms->msgin[2]) { handle_msgin() 1186 ms->data_ptr += (ms->msgin[3] << 24) + ms->msgin[6] handle_msgin() 1187 + (ms->msgin[4] << 16) + (ms->msgin[5] << 8); handle_msgin() 1192 add_sdtr_msg(ms); handle_msgin() 1195 if (ms->msgout[3] < ms->msgin[3]) handle_msgin() 1196 ms->msgout[3] = ms->msgin[3]; handle_msgin() 1197 if (ms->msgout[4] > ms->msgin[4]) handle_msgin() 1198 ms->msgout[4] = ms->msgin[4]; handle_msgin() 1199 set_sdtr(ms, ms->msgout[3], ms->msgout[4]); handle_msgin() 1200 ms->msgphase = msg_out; handle_msgin() 1202 set_sdtr(ms, ms->msgin[3], ms->msgin[4]); handle_msgin() 1210 tp->saved_ptr = ms->data_ptr; handle_msgin() 1213 ms->data_ptr = tp->saved_ptr; handle_msgin() 1216 ms->phase = disconnecting; handle_msgin() 1222 set_sdtr(ms, 0, 0); handle_msgin() 1229 do_abort(ms); handle_msgin() 1230 ms->msgphase = msg_out; handle_msgin() 1235 cmd->device->lun, ms->conn_tgt); handle_msgin() 1245 ms->conn_tgt); handle_msgin() 1246 for (i = 0; i < ms->n_msgin; ++i) handle_msgin() 1247 printk(" %x", ms->msgin[i]); handle_msgin() 1249 ms->msgout[0] = MESSAGE_REJECT; handle_msgin() 1250 ms->n_msgout = 1; handle_msgin() 1251 ms->msgphase = msg_out; handle_msgin() 1257 static void set_dma_cmds(struct mesh_state *ms, struct scsi_cmnd *cmd) set_dma_cmds() argument 1263 dma_cmd = ms->tgts[ms->conn_tgt].data_goes_out? set_dma_cmds() 1265 dcmds = ms->dma_cmds; set_dma_cmds() 1277 off = ms->data_ptr; set_dma_cmds() 1315 ms->dma_count = dtot; 1318 static void halt_dma(struct mesh_state *ms) halt_dma() argument 1320 volatile struct dbdma_regs __iomem *md = ms->dma; halt_dma() 1321 volatile struct mesh_regs __iomem *mr = ms->mesh; halt_dma() 1322 struct scsi_cmnd *cmd = ms->current_req; halt_dma() 1325 if (!ms->tgts[ms->conn_tgt].data_goes_out) { halt_dma() 1336 dlog(ms, "halt_dma fc/count=%.6x", halt_dma() 1338 if (ms->tgts[ms->conn_tgt].data_goes_out) halt_dma() 1342 ms->data_ptr -= nb; halt_dma() 1343 dlog(ms, "data_ptr %x", ms->data_ptr); halt_dma() 1344 if (ms->data_ptr < 0) { halt_dma() 1345 printk(KERN_ERR "mesh: halt_dma: data_ptr=%d (nb=%d, ms=%p)\n", halt_dma() 1346 ms->data_ptr, nb, ms); halt_dma() 1347 ms->data_ptr = 0; halt_dma() 1349 dumplog(ms, ms->conn_tgt); halt_dma() 1350 dumpslog(ms); halt_dma() 1353 ms->data_ptr > scsi_bufflen(cmd)) { halt_dma() 1356 ms->conn_tgt, ms->data_ptr, scsi_bufflen(cmd), halt_dma() 1357 ms->tgts[ms->conn_tgt].data_goes_out); halt_dma() 1360 ms->dma_started = 0; halt_dma() 1363 static void phase_mismatch(struct mesh_state *ms) phase_mismatch() argument 1365 volatile struct mesh_regs __iomem *mr = ms->mesh; phase_mismatch() 1368 dlog(ms, "phasemm ch/cl/seq/fc=%.8x", phase_mismatch() 1371 if (ms->msgphase == msg_out_xxx && phase == BP_MSGOUT) { phase_mismatch() 1377 out_8(&mr->fifo, ms->msgout[ms->n_msgout-1]); phase_mismatch() 1378 ms->msgphase = msg_out_last; phase_mismatch() 1382 if (ms->msgphase == msg_in) { phase_mismatch() 1383 get_msgin(ms); phase_mismatch() 1384 if (ms->n_msgin) phase_mismatch() 1385 handle_msgin(ms); phase_mismatch() 1388 if (ms->dma_started) phase_mismatch() 1389 halt_dma(ms); phase_mismatch() 1396 ms->msgphase = msg_none; phase_mismatch() 1399 ms->tgts[ms->conn_tgt].data_goes_out = 0; phase_mismatch() 1400 ms->phase = dataing; phase_mismatch() 1403 ms->tgts[ms->conn_tgt].data_goes_out = 1; phase_mismatch() 1404 ms->phase = dataing; phase_mismatch() 1407 ms->phase = commanding; phase_mismatch() 1410 ms->phase = statusing; phase_mismatch() 1413 ms->msgphase = msg_in; phase_mismatch() 1414 ms->n_msgin = 0; phase_mismatch() 1417 ms->msgphase = msg_out; phase_mismatch() 1418 if (ms->n_msgout == 0) { phase_mismatch() 1419 if (ms->aborting) { phase_mismatch() 1420 do_abort(ms); phase_mismatch() 1422 if (ms->last_n_msgout == 0) { phase_mismatch() 1425 ms->msgout[0] = NOP; phase_mismatch() 1426 ms->last_n_msgout = 1; phase_mismatch() 1428 ms->n_msgout = ms->last_n_msgout; phase_mismatch() 1434 ms->stat = DID_ERROR; phase_mismatch() 1435 mesh_done(ms, 1); phase_mismatch() 1439 start_phase(ms); phase_mismatch() 1442 static void cmd_complete(struct mesh_state *ms) cmd_complete() argument 1444 volatile struct mesh_regs __iomem *mr = ms->mesh; cmd_complete() 1445 struct scsi_cmnd *cmd = ms->current_req; cmd_complete() 1446 struct mesh_target *tp = &ms->tgts[ms->conn_tgt]; cmd_complete() 1449 dlog(ms, "cmd_complete fc=%x", mr->fifo_count); cmd_complete() 1450 seq = use_active_neg + (ms->n_msgout? SEQ_ATN: 0); cmd_complete() 1451 switch (ms->msgphase) { cmd_complete() 1454 ms->n_msgin = 0; cmd_complete() 1455 ms->msgphase = msg_in; cmd_complete() 1460 get_msgin(ms); cmd_complete() 1461 n = msgin_length(ms); cmd_complete() 1462 if (ms->n_msgin < n) { cmd_complete() 1463 out_8(&mr->count_lo, n - ms->n_msgin); cmd_complete() 1466 ms->msgphase = msg_none; cmd_complete() 1467 handle_msgin(ms); cmd_complete() 1468 start_phase(ms); cmd_complete() 1497 dlog(ms, "last_mbyte err/exc/fc/cl=%.8x", cmd_complete() 1502 ms->last_n_msgout = ms->n_msgout; cmd_complete() 1503 ms->n_msgout = 0; cmd_complete() 1507 handle_error(ms); cmd_complete() 1516 handle_exception(ms); cmd_complete() 1523 out_8(&mr->fifo, ms->msgout[ms->n_msgout-1]); cmd_complete() 1524 ms->msgphase = msg_out_last; cmd_complete() 1527 ms->msgphase = msg_out_xxx; cmd_complete() 1532 ms->last_n_msgout = ms->n_msgout; cmd_complete() 1533 ms->n_msgout = 0; cmd_complete() 1534 ms->msgphase = ms->expect_reply? msg_in: msg_none; cmd_complete() 1535 start_phase(ms); cmd_complete() 1539 switch (ms->phase) { cmd_complete() 1542 dumpslog(ms); cmd_complete() 1545 dlog(ms, "Selecting phase at command completion",0); cmd_complete() 1546 ms->msgout[0] = IDENTIFY(ALLOW_RESEL(ms->conn_tgt), cmd_complete() 1548 ms->n_msgout = 1; cmd_complete() 1549 ms->expect_reply = 0; cmd_complete() 1550 if (ms->aborting) { cmd_complete() 1551 ms->msgout[0] = ABORT; cmd_complete() 1552 ms->n_msgout++; cmd_complete() 1555 add_sdtr_msg(ms); cmd_complete() 1556 ms->expect_reply = 1; cmd_complete() 1559 ms->msgphase = msg_out; cmd_complete() 1570 dlog(ms, "impatient for req", ms->n_msgout); cmd_complete() 1571 ms->msgphase = msg_none; cmd_complete() 1578 if (ms->dma_count != 0) { cmd_complete() 1579 start_phase(ms); cmd_complete() 1595 halt_dma(ms); cmd_complete() 1604 ms->msgphase = msg_in; cmd_complete() 1607 mesh_done(ms, 1); cmd_complete() 1610 ms->current_req = NULL; cmd_complete() 1611 ms->phase = idle; cmd_complete() 1612 mesh_start(ms); cmd_complete() 1617 ++ms->phase; cmd_complete() 1618 start_phase(ms); cmd_complete() 1630 struct mesh_state *ms; mesh_queue_lck() local 1635 ms = (struct mesh_state *) cmd->device->host->hostdata; mesh_queue_lck() 1637 if (ms->request_q == NULL) mesh_queue_lck() 1638 ms->request_q = cmd; mesh_queue_lck() 1640 ms->request_qtail->host_scribble = (void *) cmd; mesh_queue_lck() 1641 ms->request_qtail = cmd; mesh_queue_lck() 1643 if (ms->phase == idle) mesh_queue_lck() 1644 mesh_start(ms); mesh_queue_lck() 1656 static void mesh_interrupt(struct mesh_state *ms) mesh_interrupt() argument 1658 volatile struct mesh_regs __iomem *mr = ms->mesh; mesh_interrupt() 1662 if (ALLOW_DEBUG(ms->conn_tgt)) mesh_interrupt() 1666 ms->phase, ms->msgphase); mesh_interrupt() 1669 dlog(ms, "interrupt intr/err/exc/seq=%.8x", mesh_interrupt() 1672 handle_error(ms); mesh_interrupt() 1674 handle_exception(ms); mesh_interrupt() 1677 cmd_complete(ms); mesh_interrupt() 1688 struct mesh_state *ms = (struct mesh_state *) cmd->device->host->hostdata; mesh_abort() local 1691 mesh_dump_regs(ms); mesh_abort() 1692 dumplog(ms, cmd->device->id); mesh_abort() 1693 dumpslog(ms); mesh_abort() 1705 struct mesh_state *ms = (struct mesh_state *) cmd->device->host->hostdata; mesh_host_reset() local 1706 volatile struct mesh_regs __iomem *mr = ms->mesh; mesh_host_reset() 1707 volatile struct dbdma_regs __iomem *md = ms->dma; mesh_host_reset() 1712 spin_lock_irqsave(ms->host->host_lock, flags); mesh_host_reset() 1722 out_8(&mr->source_id, ms->host->this_id); mesh_host_reset() 1723 out_8(&mr->sel_timeout, 25); /* 250ms */ mesh_host_reset() 1733 handle_reset(ms); mesh_host_reset() 1735 spin_unlock_irqrestore(ms->host->host_lock, flags); mesh_host_reset() 1739 static void set_mesh_power(struct mesh_state *ms, int state) set_mesh_power() argument 1744 pmac_call_feature(PMAC_FTR_MESH_ENABLE, macio_get_of_node(ms->mdev), 0, 1); set_mesh_power() 1747 pmac_call_feature(PMAC_FTR_MESH_ENABLE, macio_get_of_node(ms->mdev), 0, 0); set_mesh_power() 1756 struct mesh_state *ms = (struct mesh_state *)macio_get_drvdata(mdev); mesh_suspend() local 1767 if (ms->phase == sleeping) mesh_suspend() 1770 scsi_block_requests(ms->host); mesh_suspend() 1771 spin_lock_irqsave(ms->host->host_lock, flags); mesh_suspend() 1772 while(ms->phase != idle) { mesh_suspend() 1773 spin_unlock_irqrestore(ms->host->host_lock, flags); mesh_suspend() 1775 spin_lock_irqsave(ms->host->host_lock, flags); mesh_suspend() 1777 ms->phase = sleeping; mesh_suspend() 1778 spin_unlock_irqrestore(ms->host->host_lock, flags); mesh_suspend() 1779 disable_irq(ms->meshintr); mesh_suspend() 1780 set_mesh_power(ms, 0); mesh_suspend() 1787 struct mesh_state *ms = (struct mesh_state *)macio_get_drvdata(mdev); mesh_resume() local 1790 if (ms->phase != sleeping) mesh_resume() 1793 set_mesh_power(ms, 1); mesh_resume() 1794 mesh_init(ms); mesh_resume() 1795 spin_lock_irqsave(ms->host->host_lock, flags); mesh_resume() 1796 mesh_start(ms); mesh_resume() 1797 spin_unlock_irqrestore(ms->host->host_lock, flags); mesh_resume() 1798 enable_irq(ms->meshintr); mesh_resume() 1799 scsi_unblock_requests(ms->host); mesh_resume() 1813 struct mesh_state *ms = (struct mesh_state *)macio_get_drvdata(mdev); mesh_shutdown() local 1818 spin_lock_irqsave(ms->host->host_lock, flags); mesh_shutdown() 1819 mr = ms->mesh; mesh_shutdown() 1826 spin_unlock_irqrestore(ms->host->host_lock, flags); mesh_shutdown() 1850 struct mesh_state *ms; mesh_probe() local 1889 ms = (struct mesh_state *) mesh_host->hostdata; mesh_probe() 1890 macio_set_drvdata(mdev, ms); mesh_probe() 1891 ms->host = mesh_host; mesh_probe() 1892 ms->mdev = mdev; mesh_probe() 1893 ms->pdev = pdev; mesh_probe() 1895 ms->mesh = ioremap(macio_resource_start(mdev, 0), 0x1000); mesh_probe() 1896 if (ms->mesh == NULL) { mesh_probe() 1900 ms->dma = ioremap(macio_resource_start(mdev, 1), 0x1000); mesh_probe() 1901 if (ms->dma == NULL) { mesh_probe() 1903 iounmap(ms->mesh); mesh_probe() 1907 ms->meshintr = macio_irq(mdev, 0); mesh_probe() 1908 ms->dmaintr = macio_irq(mdev, 1); mesh_probe() 1913 ms->dma_cmd_size = (mesh_host->sg_tablesize + 2) * sizeof(struct dbdma_cmd); mesh_probe() 1919 ms->dma_cmd_size, &dma_cmd_bus); mesh_probe() 1925 ms->dma_cmds = (struct dbdma_cmd *) DBDMA_ALIGN(dma_cmd_space); mesh_probe() 1926 ms->dma_cmd_space = dma_cmd_space; mesh_probe() 1927 ms->dma_cmd_bus = dma_cmd_bus + ((unsigned long)ms->dma_cmds) mesh_probe() 1929 ms->current_req = NULL; mesh_probe() 1931 ms->tgts[tgt].sdtr_state = do_sdtr; mesh_probe() 1932 ms->tgts[tgt].sync_params = ASYNC_PARAMS; mesh_probe() 1933 ms->tgts[tgt].current_req = NULL; mesh_probe() 1937 ms->clk_freq = *cfp; mesh_probe() 1940 ms->clk_freq = 50000000; mesh_probe() 1946 minper = 1000000000 / (ms->clk_freq / 5); /* ns */ mesh_probe() 1951 set_mesh_power(ms, 1); mesh_probe() 1954 mesh_init(ms); mesh_probe() 1957 if (request_irq(ms->meshintr, do_mesh_interrupt, 0, "MESH", ms)) { mesh_probe() 1958 printk(KERN_ERR "MESH: can't get irq %d\n", ms->meshintr); mesh_probe() 1970 free_irq(ms->meshintr, ms); mesh_probe() 1976 set_mesh_power(ms, 0); mesh_probe() 1977 pci_free_consistent(macio_get_pci_dev(mdev), ms->dma_cmd_size, mesh_probe() 1978 ms->dma_cmd_space, ms->dma_cmd_bus); mesh_probe() 1980 iounmap(ms->dma); mesh_probe() 1981 iounmap(ms->mesh); mesh_probe() 1992 struct mesh_state *ms = (struct mesh_state *)macio_get_drvdata(mdev); mesh_remove() local 1993 struct Scsi_Host *mesh_host = ms->host; mesh_remove() 1997 free_irq(ms->meshintr, ms); mesh_remove() 2003 set_mesh_power(ms, 0); mesh_remove() 2006 iounmap(ms->mesh); mesh_remove() 2007 iounmap(ms->dma); mesh_remove() 2010 pci_free_consistent(macio_get_pci_dev(mdev), ms->dma_cmd_size, mesh_remove() 2011 ms->dma_cmd_space, ms->dma_cmd_bus); mesh_remove()
|
/linux-4.1.27/drivers/md/ |
H A D | dm-raid1.c | 42 struct mirror_set *ms; member in struct:mirror 90 struct mirror_set *ms = context; wakeup_mirrord() local 92 queue_work(ms->kmirrord_wq, &ms->kmirrord_work); wakeup_mirrord() 97 struct mirror_set *ms = (struct mirror_set *) data; delayed_wake_fn() local 99 clear_bit(0, &ms->timer_pending); delayed_wake_fn() 100 wakeup_mirrord(ms); delayed_wake_fn() 103 static void delayed_wake(struct mirror_set *ms) delayed_wake() argument 105 if (test_and_set_bit(0, &ms->timer_pending)) delayed_wake() 108 ms->timer.expires = jiffies + HZ / 5; delayed_wake() 109 ms->timer.data = (unsigned long) ms; delayed_wake() 110 ms->timer.function = delayed_wake_fn; delayed_wake() 111 add_timer(&ms->timer); delayed_wake() 119 static void queue_bio(struct mirror_set *ms, struct bio *bio, int rw) queue_bio() argument 125 bl = (rw == WRITE) ? &ms->writes : &ms->reads; queue_bio() 126 spin_lock_irqsave(&ms->lock, flags); queue_bio() 129 spin_unlock_irqrestore(&ms->lock, flags); queue_bio() 132 wakeup_mirrord(ms); queue_bio() 137 struct mirror_set *ms = context; dispatch_bios() local 141 queue_bio(ms, bio, WRITE); dispatch_bios() 171 static struct mirror *get_default_mirror(struct mirror_set *ms) get_default_mirror() argument 173 return &ms->mirror[atomic_read(&ms->default_mirror)]; get_default_mirror() 178 struct mirror_set *ms = m->ms; set_default_mirror() local 179 struct mirror *m0 = &(ms->mirror[0]); set_default_mirror() 181 atomic_set(&ms->default_mirror, m - m0); set_default_mirror() 184 static struct mirror *get_valid_mirror(struct mirror_set *ms) get_valid_mirror() argument 188 for (m = ms->mirror; m < ms->mirror + ms->nr_mirrors; m++) get_valid_mirror() 211 struct mirror_set *ms = m->ms; fail_mirror() local 214 ms->leg_failure = 1; fail_mirror() 226 if (!errors_handled(ms)) fail_mirror() 229 if (m != get_default_mirror(ms)) fail_mirror() 232 if (!ms->in_sync) { fail_mirror() 242 new = get_valid_mirror(ms); fail_mirror() 249 schedule_work(&ms->trigger_event); fail_mirror() 254 struct mirror_set *ms = ti->private; mirror_flush() local 258 struct dm_io_region io[ms->nr_mirrors]; mirror_flush() 264 .client = ms->io_client, mirror_flush() 267 for (i = 0, m = ms->mirror; i < ms->nr_mirrors; i++, m++) { mirror_flush() 274 dm_io(&io_req, ms->nr_mirrors, io, &error_bits); mirror_flush() 276 for (i = 0; i < ms->nr_mirrors; i++) mirror_flush() 278 fail_mirror(ms->mirror + i, mirror_flush() 297 struct mirror_set *ms = dm_rh_region_context(reg); recovery_complete() local 303 fail_mirror(get_default_mirror(ms), DM_RAID1_SYNC_ERROR); recovery_complete() 313 for (m = 0; m < ms->nr_mirrors; m++) { recovery_complete() 314 if (&ms->mirror[m] == get_default_mirror(ms)) recovery_complete() 317 fail_mirror(ms->mirror + m, recovery_complete() 326 static int recover(struct mirror_set *ms, struct dm_region *reg) recover() argument 334 sector_t region_size = dm_rh_get_region_size(ms->rh); recover() 337 m = get_default_mirror(ms); recover() 339 from.sector = m->offset + dm_rh_region_to_sector(ms->rh, key); recover() 340 if (key == (ms->nr_regions - 1)) { recover() 345 from.count = ms->ti->len & (region_size - 1); recover() 352 for (i = 0, dest = to; i < ms->nr_mirrors; i++) { recover() 353 if (&ms->mirror[i] == get_default_mirror(ms)) recover() 356 m = ms->mirror + i; recover() 358 dest->sector = m->offset + dm_rh_region_to_sector(ms->rh, key); recover() 364 if (!errors_handled(ms)) recover() 367 r = dm_kcopyd_copy(ms->kcopyd_client, &from, ms->nr_mirrors - 1, to, recover() 373 static void do_recovery(struct mirror_set *ms) do_recovery() argument 376 struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); do_recovery() 382 dm_rh_recovery_prepare(ms->rh); do_recovery() 387 while ((reg = dm_rh_recovery_start(ms->rh))) { do_recovery() 388 r = recover(ms, reg); do_recovery() 396 if (!ms->in_sync && do_recovery() 397 (log->type->get_sync_count(log) == ms->nr_regions)) { do_recovery() 399 dm_table_event(ms->ti->table); do_recovery() 400 ms->in_sync = 1; do_recovery() 407 static struct mirror *choose_mirror(struct mirror_set *ms, sector_t sector) choose_mirror() argument 409 struct mirror *m = get_default_mirror(ms); choose_mirror() 415 if (m-- == ms->mirror) choose_mirror() 416 m += ms->nr_mirrors; choose_mirror() 417 } while (m != get_default_mirror(ms)); choose_mirror() 424 struct mirror *default_mirror = get_default_mirror(m->ms); default_ok() 429 static int mirror_available(struct mirror_set *ms, struct bio *bio) mirror_available() argument 431 struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); mirror_available() 432 region_t region = dm_rh_bio_to_region(ms->rh, bio); mirror_available() 435 return choose_mirror(ms, bio->bi_iter.bi_sector) ? 1 : 0; mirror_available() 447 return m->offset + dm_target_offset(m->ms->ti, bio->bi_iter.bi_sector); map_sector() 464 static void hold_bio(struct mirror_set *ms, struct bio *bio) hold_bio() argument 470 spin_lock_irq(&ms->lock); hold_bio() 472 if (atomic_read(&ms->suspend)) { hold_bio() 473 spin_unlock_irq(&ms->lock); hold_bio() 478 if (dm_noflush_suspending(ms->ti)) hold_bio() 488 bio_list_add(&ms->holds, bio); hold_bio() 489 spin_unlock_irq(&ms->lock); hold_bio() 510 if (likely(default_ok(m)) || mirror_available(m->ms, bio)) { read_callback() 514 queue_bio(m->ms, bio, bio_rw(bio)); read_callback() 533 .client = m->ms->io_client, read_async_bio() 541 static inline int region_in_sync(struct mirror_set *ms, region_t region, region_in_sync() argument 544 int state = dm_rh_get_state(ms->rh, region, may_block); region_in_sync() 548 static void do_reads(struct mirror_set *ms, struct bio_list *reads) do_reads() argument 555 region = dm_rh_bio_to_region(ms->rh, bio); do_reads() 556 m = get_default_mirror(ms); do_reads() 561 if (likely(region_in_sync(ms, region, 1))) do_reads() 562 m = choose_mirror(ms, bio->bi_iter.bi_sector); do_reads() 589 struct mirror_set *ms; write_callback() local 593 ms = bio_get_m(bio)->ms; write_callback() 616 for (i = 0; i < ms->nr_mirrors; i++) write_callback() 618 fail_mirror(ms->mirror + i, DM_RAID1_WRITE_ERROR); write_callback() 625 spin_lock_irqsave(&ms->lock, flags); write_callback() 626 if (!ms->failures.head) write_callback() 628 bio_list_add(&ms->failures, bio); write_callback() 629 spin_unlock_irqrestore(&ms->lock, flags); write_callback() 631 wakeup_mirrord(ms); write_callback() 634 static void do_write(struct mirror_set *ms, struct bio *bio) do_write() argument 637 struct dm_io_region io[ms->nr_mirrors], *dest = io; do_write() 645 .client = ms->io_client, do_write() 654 for (i = 0, m = ms->mirror; i < ms->nr_mirrors; i++, m++) do_write() 661 bio_set_m(bio, get_default_mirror(ms)); do_write() 663 BUG_ON(dm_io(&io_req, ms->nr_mirrors, io, NULL)); do_write() 666 static void do_writes(struct mirror_set *ms, struct bio_list *writes) do_writes() argument 672 struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); do_writes() 693 region = dm_rh_bio_to_region(ms->rh, bio); do_writes() 701 state = dm_rh_get_state(ms->rh, region, 1); do_writes() 725 spin_lock_irq(&ms->lock); do_writes() 726 bio_list_merge(&ms->writes, &requeue); do_writes() 727 spin_unlock_irq(&ms->lock); do_writes() 728 delayed_wake(ms); do_writes() 736 dm_rh_inc_pending(ms->rh, &sync); do_writes() 737 dm_rh_inc_pending(ms->rh, &nosync); do_writes() 744 ms->log_failure = dm_rh_flush(ms->rh) ? 1 : ms->log_failure; do_writes() 749 if (unlikely(ms->log_failure) && errors_handled(ms)) { do_writes() 750 spin_lock_irq(&ms->lock); do_writes() 751 bio_list_merge(&ms->failures, &sync); do_writes() 752 spin_unlock_irq(&ms->lock); do_writes() 753 wakeup_mirrord(ms); do_writes() 756 do_write(ms, bio); do_writes() 759 dm_rh_delay(ms->rh, bio); do_writes() 762 if (unlikely(ms->leg_failure) && errors_handled(ms)) { do_writes() 763 spin_lock_irq(&ms->lock); do_writes() 764 bio_list_add(&ms->failures, bio); do_writes() 765 spin_unlock_irq(&ms->lock); do_writes() 766 wakeup_mirrord(ms); do_writes() 768 map_bio(get_default_mirror(ms), bio); do_writes() 774 static void do_failures(struct mirror_set *ms, struct bio_list *failures) do_failures() argument 799 if (!ms->log_failure) { do_failures() 800 ms->in_sync = 0; do_failures() 801 dm_rh_mark_nosync(ms->rh, bio); do_failures() 812 if (!get_valid_mirror(ms)) do_failures() 814 else if (errors_handled(ms)) do_failures() 815 hold_bio(ms, bio); do_failures() 823 struct mirror_set *ms = trigger_event() local 826 dm_table_event(ms->ti->table); trigger_event() 834 struct mirror_set *ms = container_of(work, struct mirror_set, do_mirror() local 839 spin_lock_irqsave(&ms->lock, flags); do_mirror() 840 reads = ms->reads; do_mirror() 841 writes = ms->writes; do_mirror() 842 failures = ms->failures; do_mirror() 843 bio_list_init(&ms->reads); do_mirror() 844 bio_list_init(&ms->writes); do_mirror() 845 bio_list_init(&ms->failures); do_mirror() 846 spin_unlock_irqrestore(&ms->lock, flags); do_mirror() 848 dm_rh_update_states(ms->rh, errors_handled(ms)); do_mirror() 849 do_recovery(ms); do_mirror() 850 do_reads(ms, &reads); do_mirror() 851 do_writes(ms, &writes); do_mirror() 852 do_failures(ms, &failures); do_mirror() 864 struct mirror_set *ms = NULL; alloc_context() local 866 len = sizeof(*ms) + (sizeof(ms->mirror[0]) * nr_mirrors); alloc_context() 868 ms = kzalloc(len, GFP_KERNEL); alloc_context() 869 if (!ms) { alloc_context() 874 spin_lock_init(&ms->lock); alloc_context() 875 bio_list_init(&ms->reads); alloc_context() 876 bio_list_init(&ms->writes); alloc_context() 877 bio_list_init(&ms->failures); alloc_context() 878 bio_list_init(&ms->holds); alloc_context() 880 ms->ti = ti; alloc_context() 881 ms->nr_mirrors = nr_mirrors; alloc_context() 882 ms->nr_regions = dm_sector_div_up(ti->len, region_size); alloc_context() 883 ms->in_sync = 0; alloc_context() 884 ms->log_failure = 0; alloc_context() 885 ms->leg_failure = 0; alloc_context() 886 atomic_set(&ms->suspend, 0); alloc_context() 887 atomic_set(&ms->default_mirror, DEFAULT_MIRROR); alloc_context() 889 ms->io_client = dm_io_client_create(); alloc_context() 890 if (IS_ERR(ms->io_client)) { alloc_context() 892 kfree(ms); alloc_context() 896 ms->rh = dm_region_hash_create(ms, dispatch_bios, wakeup_mirrord, alloc_context() 898 ms->ti->begin, MAX_RECOVERY, alloc_context() 899 dl, region_size, ms->nr_regions); alloc_context() 900 if (IS_ERR(ms->rh)) { alloc_context() 902 dm_io_client_destroy(ms->io_client); alloc_context() 903 kfree(ms); alloc_context() 907 return ms; alloc_context() 910 static void free_context(struct mirror_set *ms, struct dm_target *ti, free_context() argument 914 dm_put_device(ti, ms->mirror[m].dev); free_context() 916 dm_io_client_destroy(ms->io_client); free_context() 917 dm_region_hash_destroy(ms->rh); free_context() 918 kfree(ms); free_context() 921 static int get_mirror(struct mirror_set *ms, struct dm_target *ti, get_mirror() argument 933 &ms->mirror[mirror].dev)) { get_mirror() 938 ms->mirror[mirror].ms = ms; get_mirror() 939 atomic_set(&(ms->mirror[mirror].error_count), 0); get_mirror() 940 ms->mirror[mirror].error_type = 0; get_mirror() 941 ms->mirror[mirror].offset = offset; get_mirror() 984 static int parse_features(struct mirror_set *ms, unsigned argc, char **argv, parse_features() argument 988 struct dm_target *ti = ms->ti; parse_features() 1011 ms->features |= DM_RAID1_HANDLE_ERRORS; parse_features() 1038 struct mirror_set *ms; mirror_ctr() local 1064 ms = alloc_context(nr_mirrors, dl->type->get_region_size(dl), ti, dl); mirror_ctr() 1065 if (!ms) { mirror_ctr() 1072 r = get_mirror(ms, ti, m, argv); mirror_ctr() 1074 free_context(ms, ti, m); mirror_ctr() 1081 ti->private = ms; mirror_ctr() 1083 r = dm_set_target_max_io_len(ti, dm_rh_get_region_size(ms->rh)); mirror_ctr() 1092 ms->kmirrord_wq = alloc_workqueue("kmirrord", WQ_MEM_RECLAIM, 0); mirror_ctr() 1093 if (!ms->kmirrord_wq) { mirror_ctr() 1098 INIT_WORK(&ms->kmirrord_work, do_mirror); mirror_ctr() 1099 init_timer(&ms->timer); mirror_ctr() 1100 ms->timer_pending = 0; mirror_ctr() 1101 INIT_WORK(&ms->trigger_event, trigger_event); mirror_ctr() 1103 r = parse_features(ms, argc, argv, &args_used); mirror_ctr() 1125 ms->kcopyd_client = dm_kcopyd_client_create(&dm_kcopyd_throttle); mirror_ctr() 1126 if (IS_ERR(ms->kcopyd_client)) { mirror_ctr() 1127 r = PTR_ERR(ms->kcopyd_client); mirror_ctr() 1131 wakeup_mirrord(ms); mirror_ctr() 1135 destroy_workqueue(ms->kmirrord_wq); mirror_ctr() 1137 free_context(ms, ti, ms->nr_mirrors); mirror_ctr() 1143 struct mirror_set *ms = (struct mirror_set *) ti->private; mirror_dtr() local 1145 del_timer_sync(&ms->timer); mirror_dtr() 1146 flush_workqueue(ms->kmirrord_wq); mirror_dtr() 1147 flush_work(&ms->trigger_event); mirror_dtr() 1148 dm_kcopyd_client_destroy(ms->kcopyd_client); mirror_dtr() 1149 destroy_workqueue(ms->kmirrord_wq); mirror_dtr() 1150 free_context(ms, ti, ms->nr_mirrors); mirror_dtr() 1160 struct mirror_set *ms = ti->private; mirror_map() local 1161 struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); mirror_map() 1169 bio_record->write_region = dm_rh_bio_to_region(ms->rh, bio); mirror_map() 1170 queue_bio(ms, bio, rw); mirror_map() 1174 r = log->type->in_sync(log, dm_rh_bio_to_region(ms->rh, bio), 0); mirror_map() 1185 queue_bio(ms, bio, rw); mirror_map() 1193 m = choose_mirror(ms, bio->bi_iter.bi_sector); mirror_map() 1208 struct mirror_set *ms = (struct mirror_set *) ti->private; mirror_end_io() local 1219 dm_rh_dec(ms->rh, bio_record->write_region); mirror_end_io() 1251 if (default_ok(m) || mirror_available(ms, bio)) { mirror_end_io() 1259 queue_bio(ms, bio, rw); mirror_end_io() 1273 struct mirror_set *ms = (struct mirror_set *) ti->private; mirror_presuspend() local 1274 struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); mirror_presuspend() 1279 atomic_set(&ms->suspend, 1); mirror_presuspend() 1284 * a chance to be added in the hold list because ms->suspend mirror_presuspend() 1287 spin_lock_irq(&ms->lock); mirror_presuspend() 1288 holds = ms->holds; mirror_presuspend() 1289 bio_list_init(&ms->holds); mirror_presuspend() 1290 spin_unlock_irq(&ms->lock); mirror_presuspend() 1293 hold_bio(ms, bio); mirror_presuspend() 1299 dm_rh_stop_recovery(ms->rh); mirror_presuspend() 1302 !dm_rh_recovery_in_flight(ms->rh)); mirror_presuspend() 1314 flush_workqueue(ms->kmirrord_wq); mirror_presuspend() 1319 struct mirror_set *ms = ti->private; mirror_postsuspend() local 1320 struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); mirror_postsuspend() 1329 struct mirror_set *ms = ti->private; mirror_resume() local 1330 struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); mirror_resume() 1332 atomic_set(&ms->suspend, 0); mirror_resume() 1336 dm_rh_start_recovery(ms->rh); mirror_resume() 1368 struct mirror_set *ms = (struct mirror_set *) ti->private; mirror_status() local 1369 struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh); mirror_status() 1370 char buffer[ms->nr_mirrors + 1]; mirror_status() 1374 DMEMIT("%d ", ms->nr_mirrors); mirror_status() 1375 for (m = 0; m < ms->nr_mirrors; m++) { mirror_status() 1376 DMEMIT("%s ", ms->mirror[m].dev->name); mirror_status() 1377 buffer[m] = device_status_char(&(ms->mirror[m])); mirror_status() 1383 (unsigned long long)ms->nr_regions, buffer); mirror_status() 1392 DMEMIT("%d", ms->nr_mirrors); mirror_status() 1393 for (m = 0; m < ms->nr_mirrors; m++) mirror_status() 1394 DMEMIT(" %s %llu", ms->mirror[m].dev->name, mirror_status() 1395 (unsigned long long)ms->mirror[m].offset); mirror_status() 1397 if (ms->features & DM_RAID1_HANDLE_ERRORS) mirror_status() 1405 struct mirror_set *ms = ti->private; mirror_iterate_devices() local 1409 for (i = 0; !ret && i < ms->nr_mirrors; i++) mirror_iterate_devices() 1410 ret = fn(ti, ms->mirror[i].dev, mirror_iterate_devices() 1411 ms->mirror[i].offset, ti->len, data); mirror_iterate_devices()
|
/linux-4.1.27/include/crypto/ |
H A D | gf128mul.h | 103 * ms ls ms ls ms ls ms ls ms ls ms ls ms ls ms ls 107 * ms ls ms ls ms ls ms ls ms ls ms ls ms ls ms ls 113 * ms ls ms ls ms ls ms ls ms ls ms ls ms ls ms ls 117 * ms ls ms ls ms ls ms ls ms ls ms ls ms ls ms ls 139 ms ls ms ls ms ls ms ls ms ls ms ls ms ls ms ls 143 ms ls ms ls ms ls ms ls ms ls ms ls ms ls ms ls 150 ms ls ms ls ms ls ms ls ms ls ms ls ms ls ms ls 154 ms ls ms ls ms ls ms ls ms ls ms ls ms ls ms ls
|
/linux-4.1.27/include/linux/spi/ |
H A D | ad7877.h | 19 u8 first_conversion_delay; /* 0 = 0.5us, 1 = 128us, 2 = 1ms, 3 = 8ms */ 22 u8 pen_down_acc_interval; /* 0 = covert once, 1 = every 0.5 ms, 23 2 = ever 1 ms, 3 = every 8 ms,*/
|
H A D | ad7879.h | 18 * all the way to 9.440ms in steps of 35us. 22 * way to 4.096ms in steps of 128us.
|
/linux-4.1.27/drivers/staging/rts5208/ |
H A D | Makefile | 6 rtsx_card.o general.o sd.o xd.o ms.o spi.o trace.o
|
/linux-4.1.27/arch/mips/kernel/ |
H A D | spinlock_test.c | 93 struct spin_multi_state ms; multi_get() local 96 ms.lock = __RAW_SPIN_LOCK_UNLOCKED("multi_get"); multi_get() 97 ms.loops = 1000000; multi_get() 99 atomic_set(&ms.start_wait, 2); multi_get() 100 atomic_set(&ms.enter_wait, 2); multi_get() 101 atomic_set(&ms.exit_wait, 2); multi_get() 102 t1.state = &ms; multi_get() 103 t2.state = &ms; multi_get()
|
/linux-4.1.27/sound/pci/ice1712/ |
H A D | wm8776.c | 305 .enum_names = { "0.25 ms", "0.5 ms", "1 ms", "2 ms", "4 ms", 306 "8 ms", "16 ms", "32 ms", "64 ms", "128 ms", "256 ms" }, 315 .enum_names = { "1.2 ms", "2.4 ms", "4.8 ms", "9.6 ms", 316 "19.2 ms", "38.4 ms", "76.8 ms", "154 ms", "307 ms", 317 "614 ms", "1.23 s" }, 327 "1 ms", "2 ms", "4 ms" }, 355 .enum_names = { "8.40 ms", "16.8 ms", "33.6 ms", "67.2 ms", 356 "134 ms", "269 ms", "538 ms", "1.08 s", "2.15 s", 366 .enum_names = { "33.5 ms", "67.0 ms", "134 ms", "268 ms", 367 "536 ms", "1.07 s", "2.14 s", "4.29 s", "8.58 s", 397 .enum_names = { "0 ms", "2.67 ms", "5.33 ms", "10.6 ms", 398 "21.3 ms", "42.7 ms", "85.3 ms", "171 ms", "341 ms", 399 "683 ms", "1.37 s", "2.73 s", "5.46 s", "10.9 s",
|
H A D | wm8776.h | 108 #define WM8776_ALC2_HOLD_MASK 0x0f /*0=0ms, 1=2.67ms, 2=5.33ms.. */
|
/linux-4.1.27/sound/soc/codecs/ |
H A D | ssm2518.c | 155 "0 ms", "0.1 ms", "0.19 ms", "0.37 ms", "0.75 ms", "1.5 ms", "3 ms", 156 "6 ms", "12 ms", "24 ms", "48 ms", "96 ms", "192 ms", "384 ms", 157 "768 ms", "1536 ms", 161 "0 ms", "1.5 ms", "3 ms", "6 ms", "12 ms", "24 ms", "48 ms", "96 ms", 162 "192 ms", "384 ms", "768 ms", "1536 ms", "3072 ms", "6144 ms", 163 "12288 ms", "24576 ms" 167 "0 ms", "0.67 ms", "1.33 ms", "2.67 ms", "5.33 ms", "10.66 ms", 168 "21.32 ms", "42.64 ms", "85.28 ms", "170.56 ms", "341.12 ms", 169 "682.24 ms", "1364 ms",
|
H A D | wm8737.c | 127 "0", "2.67ms", "5.33ms", "10.66ms", "21.32ms", "42.64ms", "85.28ms", 128 "170.56ms", "341.12ms", "682.24ms", "1.364s", "2.728s", "5.458s", 136 "8.4ms", "16.8ms", "33.6ms", "67.2ms", "134.4ms", "268.8ms", "537.6ms", 144 "33.6ms", "67.2ms", "134.4ms", "268.8ms", "537.6ms", "1.075s", "2.15s",
|
H A D | wm8993.c | 698 "1.45ms", 699 "2.9ms", 700 "5.8ms", 701 "11.6ms", 702 "23.2ms", 703 "46.4ms", 704 "92.8ms", 705 "185.6ms", 712 "186ms", 713 "372ms", 714 "743ms", 716 "2.97ms", 717 "5.94ms", 718 "11.89ms", 719 "23.78ms", 720 "47.56ms", 735 "0.725ms", 736 "1.45ms", 737 "5.8ms",
|
H A D | wm9081.c | 288 "1.45ms", 289 "2.9ms", 290 "5.8ms", 291 "11.6ms", 292 "23.2ms", 293 "46.4ms", 294 "92.8ms", 295 "185.6ms", 301 "186ms", 302 "372ms", 303 "743ms", 315 "0.725ms", 316 "1.45ms", 317 "5.8ms",
|
/linux-4.1.27/arch/mn10300/include/asm/ |
H A D | rtc-regs.h | 32 #define RTCRA_RS_3_90625ms 0x01 /* - 3.90625ms (1/256s) */ 33 #define RTCRA_RS_7_8125ms 0x02 /* - 7.8125ms (1/128s) */ 38 #define RTCRA_RS_1_953125ms 0x07 /* - 1.953125ms (1/512s) */ 39 #define RTCRA_RS_3_90624ms 0x08 /* - 3.90624ms (1/256s) */ 40 #define RTCRA_RS_7_8125ms_b 0x09 /* - 7.8125ms (1/128s) */ 41 #define RTCRA_RS_15_625ms 0x0a /* - 15.625ms (1/64s) */ 42 #define RTCRA_RS_31_25ms 0x0b /* - 31.25ms (1/32s) */ 43 #define RTCRA_RS_62_5ms 0x0c /* - 62.5ms (1/16s) */ 44 #define RTCRA_RS_125ms 0x0d /* - 125ms (1/8s) */ 45 #define RTCRA_RS_250ms 0x0e /* - 250ms (1/4s) */ 46 #define RTCRA_RS_500ms 0x0f /* - 500ms (1/2s) */
|
/linux-4.1.27/drivers/usb/chipidea/ |
H A D | otg_fsm.h | 23 * TA_VBUS_RISE <= 100ms, section 4.4 34 * and 30000 ms, section 5.5, Table 5-1 37 #define TA_AIDL_BDIS (5000) /* a_suspend min 200 ms, section 5.2.1 42 * 500ms is used for B switch to host 51 #define TB_DATA_PLS (10) /* b_srp_init,continue 5~10ms 59 #define TB_ASE0_BRST (155) /* minimum 155 ms, section:5.3.1 */
|
/linux-4.1.27/include/linux/input/ |
H A D | cyttsp.h | 35 #define CY_ACT_INTRVL_DFLT 0x00 /* ms */ 37 #define CY_TCH_TMOUT_DFLT 0xFF /* ms */ 39 #define CY_LP_INTRVL_DFLT 0x0A /* ms */ 48 u8 act_intrvl; /* Active refresh interval; ms */ 49 u8 tch_tmout; /* Active touch timeout; ms */ 50 u8 lp_intrvl; /* Low power refresh interval; ms */
|
H A D | tps6507x-ts.h | 16 int poll_period; /* ms */
|
/linux-4.1.27/net/mac80211/ |
H A D | rc80211_minstrel_debugfs.c | 60 struct minstrel_debugfs_info *ms; minstrel_stats_read() local 62 ms = file->private_data; minstrel_stats_read() 63 return simple_read_from_buffer(buf, len, ppos, ms->buf, ms->len); minstrel_stats_read() 77 struct minstrel_debugfs_info *ms; minstrel_stats_open() local 81 ms = kmalloc(2048, GFP_KERNEL); minstrel_stats_open() 82 if (!ms) minstrel_stats_open() 85 file->private_data = ms; minstrel_stats_open() 86 p = ms->buf; minstrel_stats_open() 133 ms->len = p - ms->buf; minstrel_stats_open() 135 WARN_ON(ms->len + sizeof(*ms) > 2048); minstrel_stats_open() 152 struct minstrel_debugfs_info *ms; minstrel_stats_csv_open() local 156 ms = kmalloc(2048, GFP_KERNEL); minstrel_stats_csv_open() 157 if (!ms) minstrel_stats_csv_open() 160 file->private_data = ms; minstrel_stats_csv_open() 161 p = ms->buf; minstrel_stats_csv_open() 199 ms->len = p - ms->buf; minstrel_stats_csv_open() 201 WARN_ON(ms->len + sizeof(*ms) > 2048); minstrel_stats_csv_open()
|
H A D | rc80211_minstrel_ht_debugfs.c | 112 struct minstrel_debugfs_info *ms; minstrel_ht_stats_open() local 124 ms = kmalloc(32768, GFP_KERNEL); minstrel_ht_stats_open() 125 if (!ms) minstrel_ht_stats_open() 128 file->private_data = ms; minstrel_ht_stats_open() 129 p = ms->buf; minstrel_ht_stats_open() 152 ms->len = p - ms->buf; minstrel_ht_stats_open() 153 WARN_ON(ms->len + sizeof(*ms) > 32768); minstrel_ht_stats_open() 262 struct minstrel_debugfs_info *ms; minstrel_ht_stats_csv_open() local 274 ms = kmalloc(32768, GFP_KERNEL); minstrel_ht_stats_csv_open() 276 if (!ms) minstrel_ht_stats_csv_open() 279 file->private_data = ms; minstrel_ht_stats_csv_open() 281 p = ms->buf; minstrel_ht_stats_csv_open() 289 ms->len = p - ms->buf; minstrel_ht_stats_csv_open() 290 WARN_ON(ms->len + sizeof(*ms) > 32768); minstrel_ht_stats_csv_open()
|
/linux-4.1.27/kernel/ |
H A D | module_signing.c | 190 struct module_signature ms; mod_verify_sig() local 198 if (modlen <= sizeof(ms)) mod_verify_sig() 201 memcpy(&ms, mod + (modlen - sizeof(ms)), sizeof(ms)); mod_verify_sig() 202 modlen -= sizeof(ms); mod_verify_sig() 204 sig_len = be32_to_cpu(ms.sig_len); mod_verify_sig() 208 if ((size_t)ms.signer_len + ms.key_id_len >= modlen) mod_verify_sig() 210 modlen -= (size_t)ms.signer_len + ms.key_id_len; mod_verify_sig() 216 if (ms.algo != PKEY_ALGO_RSA || mod_verify_sig() 217 ms.id_type != PKEY_ID_X509) mod_verify_sig() 220 if (ms.hash >= PKEY_HASH__LAST || mod_verify_sig() 221 !hash_algo_name[ms.hash]) mod_verify_sig() 224 key = request_asymmetric_key(sig, ms.signer_len, mod_verify_sig() 225 sig + ms.signer_len, ms.key_id_len); mod_verify_sig() 229 pks = mod_make_digest(ms.hash, mod, modlen); mod_verify_sig() 235 ret = mod_extract_mpi_array(pks, sig + ms.signer_len + ms.key_id_len, mod_verify_sig()
|
/linux-4.1.27/drivers/media/tuners/ |
H A D | qm1d1c0042.h | 29 u32 lpf_wait; /* wait in tuning with LPF enabled. [ms] */ 30 u32 fast_srch_wait; /* with fast-search mode, no LPF. [ms] */ 31 u32 normal_srch_wait; /* with no LPF/fast-search mode. [ms] */
|
/linux-4.1.27/samples/bpf/ |
H A D | tracex3_user.c | 70 " |1ms |10ms |100ms |1s |10s\n"); print_banner() 72 printf("|1us |10us |100us |1ms |10ms " print_banner() 73 "|100ms |1s |10s\n"); print_banner()
|
/linux-4.1.27/include/uapi/linux/ |
H A D | tcp_metrics.h | 14 TCP_METRIC_RTT, /* in ms units */ 15 TCP_METRIC_RTTVAR, /* in ms units */
|
H A D | hdlcdrv.h | 26 int tx_delay; /* the transmitter keyup delay in 10ms units */ 27 int tx_tail; /* the transmitter keyoff delay in 10ms units */ 28 int slottime; /* the slottime in 10ms; usually 10 = 100ms */
|
/linux-4.1.27/drivers/media/rc/img-ir/ |
H A D | img-ir-jvc.c | 65 .pulse = { 16 /* 8.44 ms */ }, 66 .space = { 8 /* 4.22 ms */ }, 76 .space = { 3 /* 1.5825 ms +-40 us */ }, 82 .ft_min = 10, /* 5.275 ms */
|
H A D | img-ir-sanyo.c | 93 .pulse = { 16 /* 9ms */ }, 94 .space = { 8 /* 4.5ms */ }, 110 .ft_min = 10, /* 5.625 ms */ 114 .repeat = 108, /* 108 ms */ 118 .space = { 4 /* 2.25 ms */ },
|
H A D | img-ir-sharp.c | 89 .space = { 680 /* 1 ms period */ }, 94 .space = { 1680 /* 2 ms period */ }, 100 .ft_min = 5000, /* 5 ms */
|
H A D | img-ir-nec.c | 122 .pulse = { 16 /* 9ms */ }, 123 .space = { 8 /* 4.5ms */ }, 139 .ft_min = 10, /* 5.625 ms */ 143 .repeat = 108, /* 108 ms */ 147 .space = { 4 /* 2.25 ms */ },
|
H A D | img-ir-sony.c | 131 .pulse = { 4 /* 2.4 ms */ }, 141 .pulse = { 2 /* 1.2 ms */ }, 148 .ft_min = 10, /* 6 ms */
|
/linux-4.1.27/arch/blackfin/mach-bf548/include/mach/ |
H A D | bf54x_keys.h | 18 u32 keyup_test_interval; /* in ms */
|
/linux-4.1.27/arch/m68k/fpsp040/ |
H A D | sgetem.S | 94 movel LOCAL_HI(%a0),%d0 |load ms mant in d0 97 movel %d0,LOCAL_HI(%a0) |put ms mant back on stack 106 | ms mantissa part in d0 111 tstl %d0 |if any bits set in ms mant 113 | ;else no bits set in ms mant 119 exg %d0,%d1 |shift ls mant to ms mant 121 lsll %d3,%d0 |shift first 1 to integer bit in ms mant 128 lsll %d3,%d0 |shift ms mant until j-bit is set 134 | ;be shifted into ms mant 135 orl %d6,%d0 |shift the ls mant bits into the ms mant
|
H A D | x_store.S | 118 | extended -> |s| exp | |1| ms mant | | ls mant | 154 movel LOCAL_HI(%a1),%d1 |get ms mantissa 155 bfextu %d1{#1:#20},%d1 |get upper 20 bits of ms 156 orl %d1,%d0 |put these bits in ms word of double 158 movel LOCAL_HI(%a1),%d1 |get ms mantissa 189 | extended -> |s| exp | |1| ms mant | | ls mant | 222 movel LOCAL_HI(%a1),%d1 |get ms mantissa 223 bfextu %d1{#1:#23},%d1 |get upper 23 bits of ms 224 orl %d1,%d0 |put these bits in ms word of single
|
H A D | round.S | 299 | FP_SCR1 = exponent, ms mantissa part, ls mantissa part 316 | ;in ms mant 327 lsll %d0,%d1 |shift ms mant by count 332 orl %d6,%d1 |shift the ls mant bits into the ms mant 359 bfffo LOCAL_HI(%a0){#0:#32},%d7 |find first 1 in ms mant to d7) 360 beqs lower |branch if ms mant is all 0's 365 movel LOCAL_HI(%a0),%d0 |d0 has ms mant 375 | ;be shifted into ms mant 376 orl %d1,%d0 |shift the ls mant bits into the ms mant 377 movel %d0,LOCAL_HI(%a0) |store ms mant into memory 382 | We get here if ms mant was = 0, and we assume ls mant has bits 388 subw #32,%d0 |account for ms mant being all zeros 391 lsll %d7,%d1 |shift first 1 to integer bit in ms mant 392 movew %d0,LOCAL_EX(%a0) |store ms mant 476 movel #0,LOCAL_HI(%a0) |set d1 = 0 (ms mantissa) 477 movel #0,LOCAL_LO(%a0) |set d2 = 0 (ms mantissa)
|
/linux-4.1.27/drivers/net/wireless/brcm80211/brcmsmac/ |
H A D | brcms_trace_brcmsmac.h | 43 __field(uint, ms) 48 __entry->ms = t->ms; 53 "ms=%u set=%u periodic=%u", 54 __entry->ms, __entry->set, __entry->periodic
|
H A D | mac80211_if.h | 41 uint ms; member in struct:brcms_timer 107 void brcms_add_timer(struct brcms_timer *timer, uint ms, int periodic);
|
H A D | phy_shim.c | 74 wlapi_add_timer(struct wlapi_timer *t, uint ms, int periodic) wlapi_add_timer() argument 76 brcms_add_timer((struct brcms_timer *)t, ms, periodic); wlapi_add_timer()
|
/linux-4.1.27/include/scsi/ |
H A D | scsi_devinfo.h | 20 #define BLIST_MS_SKIP_PAGE_08 0x2000 /* do not send ms page 0x08 */ 21 #define BLIST_MS_SKIP_PAGE_3F 0x4000 /* do not send ms page 0x3f */ 22 #define BLIST_USE_10_BYTE_MS 0x8000 /* use 10 byte ms before 6 byte ms */ 23 #define BLIST_MS_192_BYTES_FOR_3F 0x10000 /* 192 byte ms page 0x3f request */
|
/linux-4.1.27/mm/ |
H A D | sparse.c | 108 int __section_nr(struct mem_section* ms) __section_nr() argument 118 if ((ms >= root) && (ms < (root + SECTIONS_PER_ROOT))) __section_nr() 124 return (root_nr * SECTIONS_PER_ROOT) + (ms - root); __section_nr() 178 struct mem_section *ms; memory_present() local 183 ms = __nr_to_section(section); memory_present() 184 if (!ms->section_mem_map) memory_present() 185 ms->section_mem_map = sparse_encode_early_nid(nid) | memory_present() 232 static int __meminit sparse_init_one_section(struct mem_section *ms, sparse_init_one_section() argument 236 if (!present_section(ms)) sparse_init_one_section() 239 ms->section_mem_map &= ~SECTION_MAP_MASK; sparse_init_one_section() 240 ms->section_mem_map |= sparse_encode_mem_map(mem_map, pnum) | sparse_init_one_section() 242 ms->pageblock_flags = pageblock_bitmap; sparse_init_one_section() 423 struct mem_section *ms; sparse_mem_maps_populate_node() local 430 ms = __nr_to_section(pnum); sparse_mem_maps_populate_node() 433 ms->section_mem_map = 0; sparse_mem_maps_populate_node() 452 struct mem_section *ms = __nr_to_section(pnum); sparse_early_mem_map_alloc() local 453 int nid = sparse_early_nid(ms); sparse_early_mem_map_alloc() 461 ms->section_mem_map = 0; sparse_early_mem_map_alloc() 484 struct mem_section *ms; alloc_usemap_and_memmap() local 488 ms = __nr_to_section(pnum); alloc_usemap_and_memmap() 489 nodeid_begin = sparse_early_nid(ms); alloc_usemap_and_memmap() 495 struct mem_section *ms; alloc_usemap_and_memmap() local 500 ms = __nr_to_section(pnum); alloc_usemap_and_memmap() 501 nodeid = sparse_early_nid(ms); alloc_usemap_and_memmap() 697 struct mem_section *ms; sparse_add_one_section() local 721 ms = __pfn_to_section(start_pfn); sparse_add_one_section() 722 if (ms->section_mem_map & SECTION_MARKED_PRESENT) { sparse_add_one_section() 729 ms->section_mem_map |= SECTION_MARKED_PRESENT; sparse_add_one_section() 731 ret = sparse_init_one_section(ms, section_nr, memmap, usemap); sparse_add_one_section() 791 void sparse_remove_one_section(struct zone *zone, struct mem_section *ms) sparse_remove_one_section() argument 798 if (ms->section_mem_map) { sparse_remove_one_section() 799 usemap = ms->pageblock_flags; sparse_remove_one_section() 800 memmap = sparse_decode_mem_map(ms->section_mem_map, sparse_remove_one_section() 801 __section_nr(ms)); sparse_remove_one_section() 802 ms->section_mem_map = 0; sparse_remove_one_section() 803 ms->pageblock_flags = NULL; sparse_remove_one_section()
|
H A D | page_ext.c | 264 struct mem_section *ms; __free_page_ext() local 267 ms = __pfn_to_section(pfn); __free_page_ext() 268 if (!ms || !ms->page_ext) __free_page_ext() 270 base = ms->page_ext + pfn; __free_page_ext() 272 ms->page_ext = NULL; __free_page_ext()
|
H A D | memory_hotplug.c | 188 struct mem_section *ms; register_page_bootmem_info_section() local 192 ms = __nr_to_section(section_nr); register_page_bootmem_info_section() 195 memmap = sparse_decode_mem_map(ms->section_mem_map, section_nr); register_page_bootmem_info_section() 222 struct mem_section *ms; register_page_bootmem_info_section() local 229 ms = __nr_to_section(section_nr); register_page_bootmem_info_section() 231 memmap = sparse_decode_mem_map(ms->section_mem_map, section_nr); register_page_bootmem_info_section() 527 struct mem_section *ms; find_smallest_section_pfn() local 530 ms = __pfn_to_section(start_pfn); find_smallest_section_pfn() 532 if (unlikely(!valid_section(ms))) find_smallest_section_pfn() 552 struct mem_section *ms; find_biggest_section_pfn() local 558 ms = __pfn_to_section(pfn); find_biggest_section_pfn() 560 if (unlikely(!valid_section(ms))) find_biggest_section_pfn() 582 struct mem_section *ms; shrink_zone_span() local 620 ms = __pfn_to_section(pfn); shrink_zone_span() 622 if (unlikely(!valid_section(ms))) shrink_zone_span() 650 struct mem_section *ms; shrink_pgdat_span() local 688 ms = __pfn_to_section(pfn); shrink_pgdat_span() 690 if (unlikely(!valid_section(ms))) shrink_pgdat_span() 724 static int __remove_section(struct zone *zone, struct mem_section *ms) __remove_section() argument 730 if (!valid_section(ms)) __remove_section() 733 ret = unregister_memory_section(ms); __remove_section() 737 scn_nr = __section_nr(ms); __remove_section() 741 sparse_remove_one_section(zone, ms); __remove_section()
|
/linux-4.1.27/include/linux/mfd/ |
H A D | stmpe.h | 136 * 4-> 1 ms, 5 -> 5 ms, 6 -> 10 ms, 7 -> 50 ms) 139 * (0 -> 10 us, 1 -> 100 us, 2 -> 500 us, 3 -> 1 ms, 140 * 4 -> 5 ms, 5 -> 10 ms, 6 for 50 ms, 7 -> 100 ms)
|
/linux-4.1.27/drivers/net/irda/ |
H A D | ma600-sir.c | 92 /* Hm, 0x01 means 10ms - for >= 1ms we would need 0x07 */ ma600_open() 93 qos->min_turn_time.bits = 0x01; /* Needs at least 1 ms */ ma600_open() 154 * 2. clear RTS, set DTR and wait for 1ms 159 * 5. wait at least 10 ms, new setting (baud rate, etc) takes effect here 163 /* total delays are only about 20ms - let's just sleep for now to 176 /* Set RTS low for 1 ms */ ma600_change_speed() 184 /* Wait at least 10ms: fake wait_until_sent - 10 bits at 9600 baud*/ ma600_change_speed() 185 msleep(15); /* old ma600 uses 15ms */ ma600_change_speed() 207 /* Wait at least 10ms */ ma600_change_speed() 222 * 0. DTR=0, RTS=1 and wait 10 ms 223 * 1. DTR=1, RTS=1 and wait 10 ms 227 /* total delays are only about 20ms - let's just sleep for now to 233 /* Reset the dongle : set DTR low for 10 ms */ ma600_reset()
|
H A D | tekram-sir.c | 35 static int tekram_delay = 150; /* default is 150 ms */ 66 pr_debug("%s - using %d ms delay\n", tekram_sir_init() 83 qos->min_turn_time.bits = 0x01; /* Needs at least 10 ms */ tekram_open() 193 * 0. Clear RTS and DTR, and wait 50 ms (power off the IR-210 ) 195 * 2. set DTR, and wait at least 1 ms 205 /* Should sleep 1 ms */ tekram_reset()
|
H A D | litelink-sir.c | 45 #define MAX_DELAY 10000 /* 1 ms */ 84 qos->min_turn_time.bits = 0x7f; /* Needs 0.01 ms */ litelink_open() 104 * baud rates, pulse RTS low for a few ms.
|
H A D | girbil-sir.c | 191 * 0. set RTS, and wait at least 5 ms 211 /* Sleep at least 5 ms */ girbil_reset()
|
H A D | mcp2120-sir.c | 172 * 1. Clear RTS and wait for device reset timer of 30 ms (max). 196 /* clear RTS and wait for at least 30 ms. */ mcp2120_reset()
|
/linux-4.1.27/net/netfilter/ipvs/ |
H A D | ip_vs_sync.c | 298 sb_dequeue(struct netns_ipvs *ipvs, struct ipvs_master_sync_state *ms) sb_dequeue() argument 303 if (list_empty(&ms->sync_queue)) { sb_dequeue() 307 sb = list_entry(ms->sync_queue.next, struct ip_vs_sync_buff, sb_dequeue() 310 ms->sync_queue_len--; sb_dequeue() 311 if (!ms->sync_queue_len) sb_dequeue() 312 ms->sync_queue_delay = 0; sb_dequeue() 355 struct ipvs_master_sync_state *ms) sb_queue_tail() 357 struct ip_vs_sync_buff *sb = ms->sync_buff; sb_queue_tail() 361 ms->sync_queue_len < sysctl_sync_qlen_max(ipvs)) { sb_queue_tail() 362 if (!ms->sync_queue_len) sb_queue_tail() 363 schedule_delayed_work(&ms->master_wakeup_work, sb_queue_tail() 365 ms->sync_queue_len++; sb_queue_tail() 366 list_add_tail(&sb->list, &ms->sync_queue); sb_queue_tail() 367 if ((++ms->sync_queue_delay) == IPVS_SYNC_WAKEUP_RATE) sb_queue_tail() 368 wake_up_process(ms->master_thread); sb_queue_tail() 379 get_curr_sync_buff(struct netns_ipvs *ipvs, struct ipvs_master_sync_state *ms, get_curr_sync_buff() argument 385 sb = ms->sync_buff; get_curr_sync_buff() 387 ms->sync_buff = NULL; get_curr_sync_buff() 534 struct ipvs_master_sync_state *ms; ip_vs_sync_conn_v0() local 554 ms = &ipvs->ms[id]; ip_vs_sync_conn_v0() 555 buff = ms->sync_buff; ip_vs_sync_conn_v0() 560 sb_queue_tail(ipvs, ms); ip_vs_sync_conn_v0() 561 ms->sync_buff = NULL; ip_vs_sync_conn_v0() 572 ms->sync_buff = buff; ip_vs_sync_conn_v0() 603 sb_queue_tail(ipvs, ms); ip_vs_sync_conn_v0() 604 ms->sync_buff = NULL; ip_vs_sync_conn_v0() 630 struct ipvs_master_sync_state *ms; ip_vs_sync_conn() local 664 ms = &ipvs->ms[id]; ip_vs_sync_conn() 683 buff = ms->sync_buff; ip_vs_sync_conn() 689 sb_queue_tail(ipvs, ms); ip_vs_sync_conn() 690 ms->sync_buff = NULL; ip_vs_sync_conn() 703 ms->sync_buff = buff; ip_vs_sync_conn() 1619 struct ipvs_master_sync_state *ms = master_wakeup_work_handler() local 1622 struct netns_ipvs *ipvs = ms->ipvs; master_wakeup_work_handler() 1625 if (ms->sync_queue_len && master_wakeup_work_handler() 1626 ms->sync_queue_delay < IPVS_SYNC_WAKEUP_RATE) { master_wakeup_work_handler() 1627 ms->sync_queue_delay = IPVS_SYNC_WAKEUP_RATE; master_wakeup_work_handler() 1628 wake_up_process(ms->master_thread); master_wakeup_work_handler() 1635 next_sync_buff(struct netns_ipvs *ipvs, struct ipvs_master_sync_state *ms) next_sync_buff() argument 1639 sb = sb_dequeue(ipvs, ms); next_sync_buff() 1643 return get_curr_sync_buff(ipvs, ms, IPVS_SYNC_FLUSH_TIME); next_sync_buff() 1650 struct ipvs_master_sync_state *ms = &ipvs->ms[tinfo->id]; sync_thread_master() local 1659 sb = next_sync_buff(ipvs, ms); sync_thread_master() 1685 while ((sb = sb_dequeue(ipvs, ms))) sync_thread_master() 1690 sb = get_curr_sync_buff(ipvs, ms, 0); sync_thread_master() 1762 if (ipvs->ms) start_sync_thread() 1784 struct ipvs_master_sync_state *ms; start_sync_thread() local 1786 ipvs->ms = kzalloc(count * sizeof(ipvs->ms[0]), GFP_KERNEL); start_sync_thread() 1787 if (!ipvs->ms) start_sync_thread() 1789 ms = ipvs->ms; start_sync_thread() 1790 for (id = 0; id < count; id++, ms++) { start_sync_thread() 1791 INIT_LIST_HEAD(&ms->sync_queue); start_sync_thread() 1792 ms->sync_queue_len = 0; start_sync_thread() 1793 ms->sync_queue_delay = 0; start_sync_thread() 1794 INIT_DELAYED_WORK(&ms->master_wakeup_work, start_sync_thread() 1796 ms->ipvs = ipvs; start_sync_thread() 1838 ipvs->ms[id].master_thread = task; start_sync_thread() 1868 kthread_stop(ipvs->ms[count].master_thread); start_sync_thread() 1876 kfree(ipvs->ms); start_sync_thread() 1877 ipvs->ms = NULL; start_sync_thread() 1893 if (!ipvs->ms) stop_sync_thread() 1910 struct ipvs_master_sync_state *ms = &ipvs->ms[id]; stop_sync_thread() local 1914 task_pid_nr(ms->master_thread)); stop_sync_thread() 1915 cancel_delayed_work_sync(&ms->master_wakeup_work); stop_sync_thread() 1916 ret = kthread_stop(ms->master_thread); stop_sync_thread() 1920 kfree(ipvs->ms); stop_sync_thread() 1921 ipvs->ms = NULL; stop_sync_thread() 354 sb_queue_tail(struct netns_ipvs *ipvs, struct ipvs_master_sync_state *ms) sb_queue_tail() argument
|
/linux-4.1.27/drivers/net/ethernet/qlogic/qlcnic/ |
H A D | qlcnic_hw.c | 1316 struct qlcnic_ms_reg_ctrl *ms) qlcnic_set_ms_controls() 1318 ms->control = QLCNIC_MS_CTRL; qlcnic_set_ms_controls() 1319 ms->low = QLCNIC_MS_ADDR_LO; qlcnic_set_ms_controls() 1320 ms->hi = QLCNIC_MS_ADDR_HI; qlcnic_set_ms_controls() 1322 ms->wd[0] = QLCNIC_MS_WRTDATA_LO; qlcnic_set_ms_controls() 1323 ms->rd[0] = QLCNIC_MS_RDDATA_LO; qlcnic_set_ms_controls() 1324 ms->wd[1] = QLCNIC_MS_WRTDATA_HI; qlcnic_set_ms_controls() 1325 ms->rd[1] = QLCNIC_MS_RDDATA_HI; qlcnic_set_ms_controls() 1326 ms->wd[2] = QLCNIC_MS_WRTDATA_ULO; qlcnic_set_ms_controls() 1327 ms->wd[3] = QLCNIC_MS_WRTDATA_UHI; qlcnic_set_ms_controls() 1328 ms->rd[2] = QLCNIC_MS_RDDATA_ULO; qlcnic_set_ms_controls() 1329 ms->rd[3] = QLCNIC_MS_RDDATA_UHI; qlcnic_set_ms_controls() 1331 ms->wd[0] = QLCNIC_MS_WRTDATA_ULO; qlcnic_set_ms_controls() 1332 ms->rd[0] = QLCNIC_MS_RDDATA_ULO; qlcnic_set_ms_controls() 1333 ms->wd[1] = QLCNIC_MS_WRTDATA_UHI; qlcnic_set_ms_controls() 1334 ms->rd[1] = QLCNIC_MS_RDDATA_UHI; qlcnic_set_ms_controls() 1335 ms->wd[2] = QLCNIC_MS_WRTDATA_LO; qlcnic_set_ms_controls() 1336 ms->wd[3] = QLCNIC_MS_WRTDATA_HI; qlcnic_set_ms_controls() 1337 ms->rd[2] = QLCNIC_MS_RDDATA_LO; qlcnic_set_ms_controls() 1338 ms->rd[3] = QLCNIC_MS_RDDATA_HI; qlcnic_set_ms_controls() 1341 ms->ocm_window = OCM_WIN_P3P(off); qlcnic_set_ms_controls() 1342 ms->off = GET_MEM_OFFS_2M(off); qlcnic_set_ms_controls() 1349 struct qlcnic_ms_reg_ctrl ms; qlcnic_pci_mem_write_2M() local 1355 memset(&ms, 0, sizeof(struct qlcnic_ms_reg_ctrl)); qlcnic_pci_mem_write_2M() 1362 qlcnic_set_ms_controls(adapter, off, &ms); qlcnic_pci_mem_write_2M() 1365 return qlcnic_pci_mem_access_direct(adapter, ms.ocm_window, qlcnic_pci_mem_write_2M() 1366 ms.off, &data, 1); qlcnic_pci_mem_write_2M() 1372 qlcnic_ind_wr(adapter, ms.low, off8); qlcnic_pci_mem_write_2M() 1373 qlcnic_ind_wr(adapter, ms.hi, 0); qlcnic_pci_mem_write_2M() 1375 qlcnic_ind_wr(adapter, ms.control, TA_CTL_ENABLE); qlcnic_pci_mem_write_2M() 1376 qlcnic_ind_wr(adapter, ms.control, QLCNIC_TA_START_ENABLE); qlcnic_pci_mem_write_2M() 1379 temp = qlcnic_ind_rd(adapter, ms.control); qlcnic_pci_mem_write_2M() 1390 qlcnic_ind_wr(adapter, ms.wd[0], qlcnic_ind_rd(adapter, ms.rd[0])); qlcnic_pci_mem_write_2M() 1391 qlcnic_ind_wr(adapter, ms.wd[1], qlcnic_ind_rd(adapter, ms.rd[1])); qlcnic_pci_mem_write_2M() 1393 qlcnic_ind_wr(adapter, ms.wd[2], data & 0xffffffff); qlcnic_pci_mem_write_2M() 1394 qlcnic_ind_wr(adapter, ms.wd[3], (data >> 32) & 0xffffffff); qlcnic_pci_mem_write_2M() 1396 qlcnic_ind_wr(adapter, ms.control, QLCNIC_TA_WRITE_ENABLE); qlcnic_pci_mem_write_2M() 1397 qlcnic_ind_wr(adapter, ms.control, QLCNIC_TA_WRITE_START); qlcnic_pci_mem_write_2M() 1400 temp = qlcnic_ind_rd(adapter, ms.control); qlcnic_pci_mem_write_2M() 1424 struct qlcnic_ms_reg_ctrl ms; qlcnic_pci_mem_read_2M() local 1435 memset(&ms, 0, sizeof(struct qlcnic_ms_reg_ctrl)); qlcnic_pci_mem_read_2M() 1436 qlcnic_set_ms_controls(adapter, off, &ms); qlcnic_pci_mem_read_2M() 1439 return qlcnic_pci_mem_access_direct(adapter, ms.ocm_window, qlcnic_pci_mem_read_2M() 1440 ms.off, data, 0); qlcnic_pci_mem_read_2M() 1446 qlcnic_ind_wr(adapter, ms.low, off8); qlcnic_pci_mem_read_2M() 1447 qlcnic_ind_wr(adapter, ms.hi, 0); qlcnic_pci_mem_read_2M() 1449 qlcnic_ind_wr(adapter, ms.control, TA_CTL_ENABLE); qlcnic_pci_mem_read_2M() 1450 qlcnic_ind_wr(adapter, ms.control, QLCNIC_TA_START_ENABLE); qlcnic_pci_mem_read_2M() 1453 temp = qlcnic_ind_rd(adapter, ms.control); qlcnic_pci_mem_read_2M() 1465 temp = qlcnic_ind_rd(adapter, ms.rd[3]); qlcnic_pci_mem_read_2M() 1467 val |= qlcnic_ind_rd(adapter, ms.rd[2]); qlcnic_pci_mem_read_2M() 1315 qlcnic_set_ms_controls(struct qlcnic_adapter *adapter, u64 off, struct qlcnic_ms_reg_ctrl *ms) qlcnic_set_ms_controls() argument
|
/linux-4.1.27/drivers/leds/ |
H A D | leds-clevo-mail.c | 118 if (*delay_on == 0 /* ms */ && *delay_off == 0 /* ms */) { clevo_mail_led_blink() 123 *delay_on = 1000; /* ms */ clevo_mail_led_blink() 124 *delay_off = 1000; /* ms */ clevo_mail_led_blink() 128 } else if (*delay_on == 500 /* ms */ && *delay_off == 500 /* ms */) { clevo_mail_led_blink() 133 } else if (*delay_on == 1000 /* ms */ && *delay_off == 1000 /* ms */) { clevo_mail_led_blink()
|
H A D | leds-asic3.c | 26 * period from 62ms (2 clocks) to 125s (4000 clocks). 28 #define MS_TO_CLK(ms) DIV_ROUND_CLOSEST(((ms)*1024), 32000)
|
H A D | leds-ipaq-micro.c | 56 /* Maximum duty cycle in ms 256/10 sec = 25600 ms */
|
H A D | leds-lm3533.c | 186 /* Delay limits for hardware accelerated blinking (in ms). */ 213 * Returns time code corresponding to *delay (in ms) and updates *delay to 219 * [ 16, 999] [0x00, 0x3e] step 16 ms 220 * [ 1130, 9781] [0x3d, 0x7f] step 131 ms 221 * [10306, 76890] [0x80, 0xff] step 524 ms 258 * Set delay register base to *delay (in ms) and update *delay to reflect 342 * 1 - 262 ms 343 * 2 - 524 ms
|
/linux-4.1.27/drivers/iio/common/ssp_sensors/ |
H A D | ssp_iio_sensor.h | 45 /* Converts time in ms to frequency */ ssp_convert_to_freq() 60 /* Converts frequency to time in ms */ ssp_convert_to_time()
|
/linux-4.1.27/drivers/media/platform/exynos4-is/ |
H A D | fimc-is-sensor.h | 19 #define S5K6A3_OPEN_TIMEOUT 2000 /* ms */ 37 /* sensor open timeout in ms */
|
/linux-4.1.27/include/media/ |
H A D | ir-kbd-i2c.h | 6 #define DEFAULT_POLLING_INTERVAL 100 /* ms */ 18 u32 polling_interval; /* in ms */
|
H A D | lm3560.h | 46 * min 32ms, step 32ms, max 1024ms
|
H A D | lm3646.h | 64 * min 50ms, step 50ms, max 400ms
|
/linux-4.1.27/drivers/mmc/core/ |
H A D | core.h | 58 static inline void mmc_delay(unsigned int ms) mmc_delay() argument 60 if (ms < 1000 / HZ) { mmc_delay() 62 mdelay(ms); mmc_delay() 64 msleep(ms); mmc_delay()
|
/linux-4.1.27/include/linux/platform_data/ |
H A D | usb-ohci-pxa27x.h | 22 int power_on_delay; /* Power On to Power Good time - in ms
|
H A D | usb-davinci.h | 53 /* Time from power on to power good (in 2 ms units) */
|
/linux-4.1.27/include/linux/ |
H A D | yam.h | 42 #define YAM_TXDELAY (1 << 5) /* Tx Delay (ms) */ 43 #define YAM_TXTAIL (1 << 6) /* Tx Tail (ms) */ 44 #define YAM_PERSIST (1 << 7) /* Persist (ms) */ 45 #define YAM_SLOTTIME (1 << 8) /* Slottime (ms) */
|
H A D | sunserialcore.h | 3 * Generic SUN serial/kbd/ms layer. Based entirely
|
H A D | serial_max3100.h | 22 * @poll_time: poll time for CTS signal in ms, 0 disables (so no hw
|
H A D | mc146818rtc.h | 74 * reset after update (may take 1.984ms @ 32768Hz RefClock) is complete, 75 * totalling to a max high interval of 2.228 ms.
|
H A D | amifd.h | 39 unsigned int step_delay; /* time (in ms) for delay after step */
|
H A D | gpio_mouse.h | 28 * @scan_ms: integer in ms specifying the scan periode.
|
H A D | mailbox_client.h | 23 * @tx_tout: Max block period in ms before TX is assumed failure
|
/linux-4.1.27/include/linux/i2c/ |
H A D | tsc2007.h | 10 unsigned long poll_period; /* time (in ms) between samples */
|
H A D | lm8323.h | 33 int debounce_time; /* Time to watch for key bouncing, in ms. */ 34 int active_time; /* Idle time until sleep, in ms. */
|
/linux-4.1.27/sound/pci/oxygen/ |
H A D | xonar_wm87x6.c | 548 "0 ms", "2.67 ms", "5.33 ms", "10.6 ms", wm8776_field_enum_info() 549 "21.3 ms", "42.7 ms", "85.3 ms", "171 ms", wm8776_field_enum_info() 550 "341 ms", "683 ms", "1.37 s", "2.73 s", wm8776_field_enum_info() 554 "0.25 ms", "0.5 ms", "1 ms", "2 ms", wm8776_field_enum_info() 555 "4 ms", "8 ms", "16 ms", "32 ms", wm8776_field_enum_info() 556 "64 ms", "128 ms", "256 ms", wm8776_field_enum_info() 559 "8.40 ms", "16.8 ms", "33.6 ms", "67.2 ms", wm8776_field_enum_info() 560 "134 ms", "269 ms", "538 ms", "1.08 s", wm8776_field_enum_info() 564 "1.2 ms", "2.4 ms", "4.8 ms", "9.6 ms", wm8776_field_enum_info() 565 "19.2 ms", "38.4 ms", "76.8 ms", "154 ms", wm8776_field_enum_info() 566 "307 ms", "614 ms", "1.23 s", wm8776_field_enum_info() 569 "33.5 ms", "67.0 ms", "134 ms", "268 ms", wm8776_field_enum_info() 570 "536 ms", "1.07 s", "2.14 s", "4.29 s", wm8776_field_enum_info() 575 "500 us", "1 ms", "2 ms", "4 ms", wm8776_field_enum_info()
|
/linux-4.1.27/arch/x86/math-emu/ |
H A D | polynom_Xsig.S | 30 #define SUM_MS -20(%ebp) /* sum ms long */ 33 #define ACCUM_MS -4(%ebp) /* accum ms long */ 76 mull 4(%esi) /* x ms long */ 88 mull 4(%esi) /* x ms long */ 113 adcl 4(%edi),%eax /* term ms long */
|
H A D | round_Xsig.S | 40 orl %edx,%edx /* ms bits */ 100 orl %edx,%edx /* ms bits */ 109 orl %edx,%edx /* ms bits */
|
H A D | wm_sqrt.S | 34 #define FPU_accum_3 -4(%ebp) /* ms word */ 45 #define FPU_fsqrt_arg_2 -20(%ebp) /* ms word */ 47 #define FPU_fsqrt_arg_0 -28(%ebp) /* ls word, at most the ms bit is set */ 54 .long 0 /* ms word */ 68 .long 0 /* ms word */ 72 .long 0 /* ls word, at most the ms bit is set */ 106 movl %eax,FPU_fsqrt_arg_2 /* ms word of n */ 127 movl FPU_fsqrt_arg_2,%ecx /* ms word */ 169 movl FPU_fsqrt_arg_2,%ecx /* ms word of normalized n */ 261 movl FPU_fsqrt_arg_2,%eax /* ms word of normalized n */ 433 stc /* The ms bit. */
|
H A D | reg_u_div.S | 152 stc /* To set the ms bit */ 229 denom ms dw */ 234 mull SIGH(%ebx) /* mul by the ms dw of the denom */ 303 divl %ecx /* Divide the numerator by the denom ms dw */ 308 mull SIGH(%ebx) /* mul by the ms dw of the denom */ 428 stc /* Will set the ms bit */
|
H A D | reg_norm.S | 34 orl %edx,%edx /* ms bits */ 111 orl %edx,%edx /* ms bits */
|
H A D | div_Xsig.S | 147 denom ms dw */ 152 mull XsigH(%ebx) /* mul by the ms dw of the denom */ 221 divl %ecx /* Divide the numerator by the denom ms dw */ 226 mull XsigH(%ebx) /* mul by the ms dw of the denom */
|
H A D | poly.h | 49 /* Macro to access the 8 ms bytes of an Xsig as a long long */ 58 The answer is the ms word of the product. */
|
/linux-4.1.27/drivers/staging/lustre/include/linux/libcfs/ |
H A D | libcfs_fail.h | 46 int __cfs_fail_timeout_set(__u32 id, __u32 value, int ms, int set); 123 static inline int cfs_fail_timeout_set(__u32 id, __u32 value, int ms, int set) cfs_fail_timeout_set() argument 126 return __cfs_fail_timeout_set(id, value, ms, set); cfs_fail_timeout_set() 134 #define CFS_FAIL_TIMEOUT_MS(id, ms) \ 135 cfs_fail_timeout_set(id, 0, ms, CFS_FAIL_LOC_NOSET) 142 #define CFS_FAIL_TIMEOUT_MS_ORSET(id, value, ms) \ 143 cfs_fail_timeout_set(id, value, ms, CFS_FAIL_LOC_ORSET)
|
/linux-4.1.27/drivers/scsi/bfa/ |
H A D | bfa_fcs_lport.c | 1477 struct bfa_fcs_lport_s *port = fdmi->ms->port; bfa_fcs_lport_fdmi_sm_offline() 1517 struct bfa_fcs_lport_s *port = fdmi->ms->port; bfa_fcs_lport_fdmi_sm_sending_rhba() 1542 struct bfa_fcs_lport_s *port = fdmi->ms->port; bfa_fcs_lport_fdmi_sm_rhba() 1591 struct bfa_fcs_lport_s *port = fdmi->ms->port; bfa_fcs_lport_fdmi_sm_rhba_retry() 1622 struct bfa_fcs_lport_s *port = fdmi->ms->port; bfa_fcs_lport_fdmi_sm_sending_rprt() 1647 struct bfa_fcs_lport_s *port = fdmi->ms->port; bfa_fcs_lport_fdmi_sm_rprt() 1694 struct bfa_fcs_lport_s *port = fdmi->ms->port; bfa_fcs_lport_fdmi_sm_rprt_retry() 1725 struct bfa_fcs_lport_s *port = fdmi->ms->port; bfa_fcs_lport_fdmi_sm_sending_rpa() 1750 struct bfa_fcs_lport_s *port = fdmi->ms->port; bfa_fcs_lport_fdmi_sm_rpa() 1795 struct bfa_fcs_lport_s *port = fdmi->ms->port; bfa_fcs_lport_fdmi_sm_rpa_retry() 1823 struct bfa_fcs_lport_s *port = fdmi->ms->port; bfa_fcs_lport_fdmi_sm_online() 1844 struct bfa_fcs_lport_s *port = fdmi->ms->port; bfa_fcs_lport_fdmi_sm_disabled() 1859 struct bfa_fcs_lport_s *port = fdmi->ms->port; bfa_fcs_lport_fdmi_send_rhba() 1898 struct bfa_fcs_lport_s *port = fdmi->ms->port; bfa_fcs_lport_fdmi_build_rhba_pyld() 2158 struct bfa_fcs_lport_s *port = fdmi->ms->port; bfa_fcs_lport_fdmi_rhba_response() 2192 struct bfa_fcs_lport_s *port = fdmi->ms->port; bfa_fcs_lport_fdmi_send_rprt() 2458 struct bfa_fcs_lport_s *port = fdmi->ms->port; bfa_fcs_lport_fdmi_build_rprt_pyld() 2481 struct bfa_fcs_lport_s *port = fdmi->ms->port; bfa_fcs_lport_fdmi_rprt_response() 2515 struct bfa_fcs_lport_s *port = fdmi->ms->port; bfa_fcs_lport_fdmi_send_rpa() 2552 struct bfa_fcs_lport_s *port = fdmi->ms->port; bfa_fcs_lport_fdmi_build_rpa_pyld() 2573 struct bfa_fcs_lport_s *port = fdmi->ms->port; bfa_fcs_lport_fdmi_rpa_response() 2612 struct bfa_fcs_lport_s *port = fdmi->ms->port; bfa_fcs_fdmi_get_hbaattr() 2668 struct bfa_fcs_lport_s *port = fdmi->ms->port; bfa_fcs_fdmi_get_portattr() 2787 bfa_fcs_lport_fdmi_init(struct bfa_fcs_lport_ms_s *ms) bfa_fcs_lport_fdmi_init() argument 2789 struct bfa_fcs_lport_fdmi_s *fdmi = &ms->fdmi; bfa_fcs_lport_fdmi_init() 2791 fdmi->ms = ms; bfa_fcs_lport_fdmi_init() 2792 if (ms->port->fcs->fdmi_enabled) bfa_fcs_lport_fdmi_init() 2799 bfa_fcs_lport_fdmi_offline(struct bfa_fcs_lport_ms_s *ms) bfa_fcs_lport_fdmi_offline() argument 2801 struct bfa_fcs_lport_fdmi_s *fdmi = &ms->fdmi; bfa_fcs_lport_fdmi_offline() 2803 fdmi->ms = ms; bfa_fcs_lport_fdmi_offline() 2808 bfa_fcs_lport_fdmi_online(struct bfa_fcs_lport_ms_s *ms) bfa_fcs_lport_fdmi_online() argument 2810 struct bfa_fcs_lport_fdmi_s *fdmi = &ms->fdmi; bfa_fcs_lport_fdmi_online() 2812 fdmi->ms = ms; bfa_fcs_lport_fdmi_online() 2867 static void bfa_fcs_lport_ms_sm_offline(struct bfa_fcs_lport_ms_s *ms, 2869 static void bfa_fcs_lport_ms_sm_plogi_sending(struct bfa_fcs_lport_ms_s *ms, 2871 static void bfa_fcs_lport_ms_sm_plogi(struct bfa_fcs_lport_ms_s *ms, 2873 static void bfa_fcs_lport_ms_sm_plogi_retry(struct bfa_fcs_lport_ms_s *ms, 2875 static void bfa_fcs_lport_ms_sm_gmal_sending(struct bfa_fcs_lport_ms_s *ms, 2877 static void bfa_fcs_lport_ms_sm_gmal(struct bfa_fcs_lport_ms_s *ms, 2879 static void bfa_fcs_lport_ms_sm_gmal_retry(struct bfa_fcs_lport_ms_s *ms, 2881 static void bfa_fcs_lport_ms_sm_gfn_sending(struct bfa_fcs_lport_ms_s *ms, 2883 static void bfa_fcs_lport_ms_sm_gfn(struct bfa_fcs_lport_ms_s *ms, 2885 static void bfa_fcs_lport_ms_sm_gfn_retry(struct bfa_fcs_lport_ms_s *ms, 2887 static void bfa_fcs_lport_ms_sm_online(struct bfa_fcs_lport_ms_s *ms, 2893 bfa_fcs_lport_ms_sm_offline(struct bfa_fcs_lport_ms_s *ms, bfa_fcs_lport_ms_sm_offline() argument 2896 bfa_trc(ms->port->fcs, ms->port->port_cfg.pwwn); bfa_fcs_lport_ms_sm_offline() 2897 bfa_trc(ms->port->fcs, event); bfa_fcs_lport_ms_sm_offline() 2901 bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_plogi_sending); bfa_fcs_lport_ms_sm_offline() 2902 bfa_fcs_lport_ms_send_plogi(ms, NULL); bfa_fcs_lport_ms_sm_offline() 2909 bfa_sm_fault(ms->port->fcs, event); bfa_fcs_lport_ms_sm_offline() 2914 bfa_fcs_lport_ms_sm_plogi_sending(struct bfa_fcs_lport_ms_s *ms, bfa_fcs_lport_ms_sm_plogi_sending() argument 2917 bfa_trc(ms->port->fcs, ms->port->port_cfg.pwwn); bfa_fcs_lport_ms_sm_plogi_sending() 2918 bfa_trc(ms->port->fcs, event); bfa_fcs_lport_ms_sm_plogi_sending() 2922 bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_plogi); bfa_fcs_lport_ms_sm_plogi_sending() 2926 bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_offline); bfa_fcs_lport_ms_sm_plogi_sending() 2927 bfa_fcxp_walloc_cancel(BFA_FCS_GET_HAL_FROM_PORT(ms->port), bfa_fcs_lport_ms_sm_plogi_sending() 2928 &ms->fcxp_wqe); bfa_fcs_lport_ms_sm_plogi_sending() 2932 bfa_sm_fault(ms->port->fcs, event); bfa_fcs_lport_ms_sm_plogi_sending() 2937 bfa_fcs_lport_ms_sm_plogi(struct bfa_fcs_lport_ms_s *ms, bfa_fcs_lport_ms_sm_plogi() argument 2940 bfa_trc(ms->port->fcs, ms->port->port_cfg.pwwn); bfa_fcs_lport_ms_sm_plogi() 2941 bfa_trc(ms->port->fcs, event); bfa_fcs_lport_ms_sm_plogi() 2948 bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_plogi_retry); bfa_fcs_lport_ms_sm_plogi() 2949 ms->port->stats.ms_retries++; bfa_fcs_lport_ms_sm_plogi() 2950 bfa_timer_start(BFA_FCS_GET_HAL_FROM_PORT(ms->port), bfa_fcs_lport_ms_sm_plogi() 2951 &ms->timer, bfa_fcs_lport_ms_timeout, ms, bfa_fcs_lport_ms_sm_plogi() local 2959 bfa_fcs_lport_fdmi_online(ms); bfa_fcs_lport_ms_sm_plogi() 2964 if (ms->port->vport) { bfa_fcs_lport_ms_sm_plogi() 2965 bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_online); bfa_fcs_lport_ms_sm_plogi() 2973 bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_gmal_sending); bfa_fcs_lport_ms_sm_plogi() 2974 bfa_fcs_lport_ms_send_gmal(ms, NULL); bfa_fcs_lport_ms_sm_plogi() 2978 bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_offline); bfa_fcs_lport_ms_sm_plogi() 2979 bfa_fcxp_discard(ms->fcxp); bfa_fcs_lport_ms_sm_plogi() 2983 bfa_sm_fault(ms->port->fcs, event); bfa_fcs_lport_ms_sm_plogi() 2988 bfa_fcs_lport_ms_sm_plogi_retry(struct bfa_fcs_lport_ms_s *ms, bfa_fcs_lport_ms_sm_plogi_retry() argument 2991 bfa_trc(ms->port->fcs, ms->port->port_cfg.pwwn); bfa_fcs_lport_ms_sm_plogi_retry() 2992 bfa_trc(ms->port->fcs, event); bfa_fcs_lport_ms_sm_plogi_retry() 2999 bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_plogi_sending); bfa_fcs_lport_ms_sm_plogi_retry() 3000 bfa_fcs_lport_ms_send_plogi(ms, NULL); bfa_fcs_lport_ms_sm_plogi_retry() 3004 bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_offline); bfa_fcs_lport_ms_sm_plogi_retry() 3005 bfa_timer_stop(&ms->timer); bfa_fcs_lport_ms_sm_plogi_retry() 3009 bfa_sm_fault(ms->port->fcs, event); bfa_fcs_lport_ms_sm_plogi_retry() 3014 bfa_fcs_lport_ms_sm_online(struct bfa_fcs_lport_ms_s *ms, bfa_fcs_lport_ms_sm_online() argument 3017 bfa_trc(ms->port->fcs, ms->port->port_cfg.pwwn); bfa_fcs_lport_ms_sm_online() 3018 bfa_trc(ms->port->fcs, event); bfa_fcs_lport_ms_sm_online() 3022 bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_offline); bfa_fcs_lport_ms_sm_online() 3026 bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_gfn_sending); bfa_fcs_lport_ms_sm_online() 3027 ms->retry_cnt = 0; bfa_fcs_lport_ms_sm_online() 3028 bfa_fcs_lport_ms_send_gfn(ms, NULL); bfa_fcs_lport_ms_sm_online() 3032 bfa_sm_fault(ms->port->fcs, event); bfa_fcs_lport_ms_sm_online() 3037 bfa_fcs_lport_ms_sm_gmal_sending(struct bfa_fcs_lport_ms_s *ms, bfa_fcs_lport_ms_sm_gmal_sending() argument 3040 bfa_trc(ms->port->fcs, ms->port->port_cfg.pwwn); bfa_fcs_lport_ms_sm_gmal_sending() 3041 bfa_trc(ms->port->fcs, event); bfa_fcs_lport_ms_sm_gmal_sending() 3045 bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_gmal); bfa_fcs_lport_ms_sm_gmal_sending() 3049 bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_offline); bfa_fcs_lport_ms_sm_gmal_sending() 3050 bfa_fcxp_walloc_cancel(BFA_FCS_GET_HAL_FROM_PORT(ms->port), bfa_fcs_lport_ms_sm_gmal_sending() 3051 &ms->fcxp_wqe); bfa_fcs_lport_ms_sm_gmal_sending() 3055 bfa_sm_fault(ms->port->fcs, event); bfa_fcs_lport_ms_sm_gmal_sending() 3060 bfa_fcs_lport_ms_sm_gmal(struct bfa_fcs_lport_ms_s *ms, bfa_fcs_lport_ms_sm_gmal() argument 3063 bfa_trc(ms->port->fcs, ms->port->port_cfg.pwwn); bfa_fcs_lport_ms_sm_gmal() 3064 bfa_trc(ms->port->fcs, event); bfa_fcs_lport_ms_sm_gmal() 3071 if (ms->retry_cnt++ < BFA_FCS_MS_CMD_MAX_RETRIES) { bfa_fcs_lport_ms_sm_gmal() 3072 bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_gmal_retry); bfa_fcs_lport_ms_sm_gmal() 3073 ms->port->stats.ms_retries++; bfa_fcs_lport_ms_sm_gmal() 3074 bfa_timer_start(BFA_FCS_GET_HAL_FROM_PORT(ms->port), bfa_fcs_lport_ms_sm_gmal() 3075 &ms->timer, bfa_fcs_lport_ms_timeout, ms, bfa_fcs_lport_ms_sm_gmal() local 3078 bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_gfn_sending); bfa_fcs_lport_ms_sm_gmal() 3079 bfa_fcs_lport_ms_send_gfn(ms, NULL); bfa_fcs_lport_ms_sm_gmal() 3080 ms->retry_cnt = 0; bfa_fcs_lport_ms_sm_gmal() 3085 bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_gfn_sending); bfa_fcs_lport_ms_sm_gmal() 3086 bfa_fcs_lport_ms_send_gfn(ms, NULL); bfa_fcs_lport_ms_sm_gmal() 3090 bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_offline); bfa_fcs_lport_ms_sm_gmal() 3091 bfa_fcxp_discard(ms->fcxp); bfa_fcs_lport_ms_sm_gmal() 3095 bfa_sm_fault(ms->port->fcs, event); bfa_fcs_lport_ms_sm_gmal() 3100 bfa_fcs_lport_ms_sm_gmal_retry(struct bfa_fcs_lport_ms_s *ms, bfa_fcs_lport_ms_sm_gmal_retry() argument 3103 bfa_trc(ms->port->fcs, ms->port->port_cfg.pwwn); bfa_fcs_lport_ms_sm_gmal_retry() 3104 bfa_trc(ms->port->fcs, event); bfa_fcs_lport_ms_sm_gmal_retry() 3111 bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_gmal_sending); bfa_fcs_lport_ms_sm_gmal_retry() 3112 bfa_fcs_lport_ms_send_gmal(ms, NULL); bfa_fcs_lport_ms_sm_gmal_retry() 3116 bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_offline); bfa_fcs_lport_ms_sm_gmal_retry() 3117 bfa_timer_stop(&ms->timer); bfa_fcs_lport_ms_sm_gmal_retry() 3121 bfa_sm_fault(ms->port->fcs, event); bfa_fcs_lport_ms_sm_gmal_retry() 3131 struct bfa_fcs_lport_ms_s *ms = ms_cbarg; bfa_fcs_lport_ms_send_gmal() local 3132 bfa_fcs_lport_t *port = ms->port; bfa_fcs_lport_ms_send_gmal() 3142 bfa_fcs_fcxp_alloc_wait(port->fcs->bfa, &ms->fcxp_wqe, bfa_fcs_lport_ms_send_gmal() 3143 bfa_fcs_lport_ms_send_gmal, ms, BFA_TRUE); bfa_fcs_lport_ms_send_gmal() 3146 ms->fcxp = fcxp; bfa_fcs_lport_ms_send_gmal() 3154 bfa_fcs_lport_ms_gmal_response, (void *)ms, bfa_fcs_lport_ms_send_gmal() 3157 bfa_sm_send_event(ms, MSSM_EVENT_FCXP_SENT); bfa_fcs_lport_ms_send_gmal() 3166 struct bfa_fcs_lport_ms_s *ms = (struct bfa_fcs_lport_ms_s *) cbarg; bfa_fcs_lport_ms_gmal_response() local 3167 bfa_fcs_lport_t *port = ms->port; bfa_fcs_lport_ms_gmal_response() 3182 bfa_sm_send_event(ms, MSSM_EVENT_RSP_ERROR); bfa_fcs_lport_ms_gmal_response() 3194 bfa_sm_send_event(ms, MSSM_EVENT_RSP_ERROR); bfa_fcs_lport_ms_gmal_response() 3230 bfa_sm_send_event(ms, MSSM_EVENT_RSP_OK); bfa_fcs_lport_ms_gmal_response() 3236 bfa_sm_send_event(ms, MSSM_EVENT_RSP_ERROR); bfa_fcs_lport_ms_gmal_response() 3240 bfa_fcs_lport_ms_sm_gfn_sending(struct bfa_fcs_lport_ms_s *ms, bfa_fcs_lport_ms_sm_gfn_sending() argument 3243 bfa_trc(ms->port->fcs, ms->port->port_cfg.pwwn); bfa_fcs_lport_ms_sm_gfn_sending() 3244 bfa_trc(ms->port->fcs, event); bfa_fcs_lport_ms_sm_gfn_sending() 3248 bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_gfn); bfa_fcs_lport_ms_sm_gfn_sending() 3252 bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_offline); bfa_fcs_lport_ms_sm_gfn_sending() 3253 bfa_fcxp_walloc_cancel(BFA_FCS_GET_HAL_FROM_PORT(ms->port), bfa_fcs_lport_ms_sm_gfn_sending() 3254 &ms->fcxp_wqe); bfa_fcs_lport_ms_sm_gfn_sending() 3258 bfa_sm_fault(ms->port->fcs, event); bfa_fcs_lport_ms_sm_gfn_sending() 3263 bfa_fcs_lport_ms_sm_gfn(struct bfa_fcs_lport_ms_s *ms, bfa_fcs_lport_ms_sm_gfn() argument 3266 bfa_trc(ms->port->fcs, ms->port->port_cfg.pwwn); bfa_fcs_lport_ms_sm_gfn() 3267 bfa_trc(ms->port->fcs, event); bfa_fcs_lport_ms_sm_gfn() 3274 if (ms->retry_cnt++ < BFA_FCS_MS_CMD_MAX_RETRIES) { bfa_fcs_lport_ms_sm_gfn() 3275 bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_gfn_retry); bfa_fcs_lport_ms_sm_gfn() 3276 ms->port->stats.ms_retries++; bfa_fcs_lport_ms_sm_gfn() 3277 bfa_timer_start(BFA_FCS_GET_HAL_FROM_PORT(ms->port), bfa_fcs_lport_ms_sm_gfn() 3278 &ms->timer, bfa_fcs_lport_ms_timeout, ms, bfa_fcs_lport_ms_sm_gfn() local 3281 bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_online); bfa_fcs_lport_ms_sm_gfn() 3282 ms->retry_cnt = 0; bfa_fcs_lport_ms_sm_gfn() 3287 bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_online); bfa_fcs_lport_ms_sm_gfn() 3291 bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_offline); bfa_fcs_lport_ms_sm_gfn() 3292 bfa_fcxp_discard(ms->fcxp); bfa_fcs_lport_ms_sm_gfn() 3296 bfa_sm_fault(ms->port->fcs, event); bfa_fcs_lport_ms_sm_gfn() 3301 bfa_fcs_lport_ms_sm_gfn_retry(struct bfa_fcs_lport_ms_s *ms, bfa_fcs_lport_ms_sm_gfn_retry() argument 3304 bfa_trc(ms->port->fcs, ms->port->port_cfg.pwwn); bfa_fcs_lport_ms_sm_gfn_retry() 3305 bfa_trc(ms->port->fcs, event); bfa_fcs_lport_ms_sm_gfn_retry() 3312 bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_gfn_sending); bfa_fcs_lport_ms_sm_gfn_retry() 3313 bfa_fcs_lport_ms_send_gfn(ms, NULL); bfa_fcs_lport_ms_sm_gfn_retry() 3317 bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_offline); bfa_fcs_lport_ms_sm_gfn_retry() 3318 bfa_timer_stop(&ms->timer); bfa_fcs_lport_ms_sm_gfn_retry() 3322 bfa_sm_fault(ms->port->fcs, event); bfa_fcs_lport_ms_sm_gfn_retry() 3332 struct bfa_fcs_lport_ms_s *ms = ms_cbarg; bfa_fcs_lport_ms_send_gfn() local 3333 bfa_fcs_lport_t *port = ms->port; bfa_fcs_lport_ms_send_gfn() 3343 bfa_fcs_fcxp_alloc_wait(port->fcs->bfa, &ms->fcxp_wqe, bfa_fcs_lport_ms_send_gfn() 3344 bfa_fcs_lport_ms_send_gfn, ms, BFA_TRUE); bfa_fcs_lport_ms_send_gfn() 3347 ms->fcxp = fcxp; bfa_fcs_lport_ms_send_gfn() 3355 bfa_fcs_lport_ms_gfn_response, (void *)ms, bfa_fcs_lport_ms_send_gfn() 3358 bfa_sm_send_event(ms, MSSM_EVENT_FCXP_SENT); bfa_fcs_lport_ms_send_gfn() 3366 struct bfa_fcs_lport_ms_s *ms = (struct bfa_fcs_lport_ms_s *) cbarg; bfa_fcs_lport_ms_gfn_response() local 3367 bfa_fcs_lport_t *port = ms->port; bfa_fcs_lport_ms_gfn_response() 3379 bfa_sm_send_event(ms, MSSM_EVENT_RSP_ERROR); bfa_fcs_lport_ms_gfn_response() 3393 bfa_sm_send_event(ms, MSSM_EVENT_RSP_OK); bfa_fcs_lport_ms_gfn_response() 3399 bfa_sm_send_event(ms, MSSM_EVENT_RSP_ERROR); bfa_fcs_lport_ms_gfn_response() 3409 struct bfa_fcs_lport_ms_s *ms = ms_cbarg; bfa_fcs_lport_ms_send_plogi() local 3410 struct bfa_fcs_lport_s *port = ms->port; bfa_fcs_lport_ms_send_plogi() 3421 bfa_fcs_fcxp_alloc_wait(port->fcs->bfa, &ms->fcxp_wqe, bfa_fcs_lport_ms_send_plogi() 3422 bfa_fcs_lport_ms_send_plogi, ms, BFA_TRUE); bfa_fcs_lport_ms_send_plogi() 3425 ms->fcxp = fcxp; bfa_fcs_lport_ms_send_plogi() 3436 bfa_fcs_lport_ms_plogi_response, (void *)ms, bfa_fcs_lport_ms_send_plogi() 3440 bfa_sm_send_event(ms, MSSM_EVENT_FCXP_SENT); bfa_fcs_lport_ms_send_plogi() 3448 struct bfa_fcs_lport_ms_s *ms = (struct bfa_fcs_lport_ms_s *) cbarg; bfa_fcs_lport_ms_plogi_response() local 3449 struct bfa_fcs_lport_s *port = ms->port; bfa_fcs_lport_ms_plogi_response() 3462 bfa_sm_send_event(ms, MSSM_EVENT_RSP_ERROR); bfa_fcs_lport_ms_plogi_response() 3474 bfa_sm_send_event(ms, MSSM_EVENT_RSP_ERROR); bfa_fcs_lport_ms_plogi_response() 3478 bfa_sm_send_event(ms, MSSM_EVENT_RSP_OK); bfa_fcs_lport_ms_plogi_response() 3488 bfa_sm_send_event(ms, MSSM_EVENT_RSP_ERROR); bfa_fcs_lport_ms_plogi_response() 3494 bfa_sm_send_event(ms, MSSM_EVENT_RSP_ERROR); bfa_fcs_lport_ms_plogi_response() 3501 struct bfa_fcs_lport_ms_s *ms = (struct bfa_fcs_lport_ms_s *) arg; bfa_fcs_lport_ms_timeout() local 3503 ms->port->stats.ms_timeouts++; bfa_fcs_lport_ms_timeout() 3504 bfa_sm_send_event(ms, MSSM_EVENT_TIMEOUT); bfa_fcs_lport_ms_timeout() 3511 struct bfa_fcs_lport_ms_s *ms = BFA_FCS_GET_MS_FROM_PORT(port); bfa_fcs_lport_ms_init() local 3513 ms->port = port; bfa_fcs_lport_ms_init() 3514 bfa_sm_set_state(ms, bfa_fcs_lport_ms_sm_offline); bfa_fcs_lport_ms_init() 3519 bfa_fcs_lport_fdmi_init(ms); bfa_fcs_lport_ms_init() 3525 struct bfa_fcs_lport_ms_s *ms = BFA_FCS_GET_MS_FROM_PORT(port); bfa_fcs_lport_ms_offline() local 3527 ms->port = port; bfa_fcs_lport_ms_offline() 3528 bfa_sm_send_event(ms, MSSM_EVENT_PORT_OFFLINE); bfa_fcs_lport_ms_offline() 3529 bfa_fcs_lport_fdmi_offline(ms); bfa_fcs_lport_ms_offline() 3535 struct bfa_fcs_lport_ms_s *ms = BFA_FCS_GET_MS_FROM_PORT(port); bfa_fcs_lport_ms_online() local 3537 ms->port = port; bfa_fcs_lport_ms_online() 3538 bfa_sm_send_event(ms, MSSM_EVENT_PORT_ONLINE); bfa_fcs_lport_ms_online() 3543 struct bfa_fcs_lport_ms_s *ms = BFA_FCS_GET_MS_FROM_PORT(port); bfa_fcs_lport_ms_fabric_rscn() local 3546 if (bfa_sm_cmp_state(ms, bfa_fcs_lport_ms_sm_online)) bfa_fcs_lport_ms_fabric_rscn() 3547 bfa_sm_send_event(ms, MSSM_EVENT_PORT_FABRIC_RSCN); bfa_fcs_lport_ms_fabric_rscn()
|
/linux-4.1.27/tools/perf/ui/browsers/ |
H A D | annotate.c | 221 struct map_symbol *ms = ab->b.priv; annotate_browser__draw_current_jump() local 222 struct symbol *sym = ms->sym; annotate_browser__draw_current_jump() 345 struct map_symbol *ms = browser->b.priv; annotate_browser__calc_percent() local 346 struct symbol *sym = ms->sym; annotate_browser__calc_percent() 448 struct map_symbol *ms = browser->b.priv; annotate_browser__callq() local 452 .map = ms->map, annotate_browser__callq() 453 .addr = map__objdump_2mem(ms->map, dl->ops.target.addr), annotate_browser__callq() 480 sym_title(ms->sym, ms->map, title, sizeof(title)); annotate_browser__callq() 489 struct map_symbol *ms = browser->b.priv; annotate_browser__find_offset() local 490 struct symbol *sym = ms->sym; annotate_browser__find_offset() 528 struct map_symbol *ms = browser->b.priv; annotate_browser__find_string() local 529 struct symbol *sym = ms->sym; annotate_browser__find_string() 567 struct map_symbol *ms = browser->b.priv; annotate_browser__find_string_reverse() local 568 struct symbol *sym = ms->sym; annotate_browser__find_string_reverse() 668 struct map_symbol *ms = browser->b.priv; annotate_browser__run() local 669 struct symbol *sym = ms->sym; annotate_browser__run() 675 sym_title(sym, ms->map, title, sizeof(title)); annotate_browser__run() 832 int map_symbol__tui_annotate(struct map_symbol *ms, struct perf_evsel *evsel, map_symbol__tui_annotate() argument 835 return symbol__tui_annotate(ms->sym, ms->map, evsel, hbt); map_symbol__tui_annotate() 841 return map_symbol__tui_annotate(&he->ms, evsel, hbt); hist_entry__tui_annotate() 848 struct map_symbol *ms = browser->b.priv; annotate_browser__mark_jump_targets() local 849 struct symbol *sym = ms->sym; annotate_browser__mark_jump_targets() 894 struct map_symbol ms = { symbol__tui_annotate() local 904 .priv = &ms, symbol__tui_annotate()
|
H A D | hists.c | 63 if (he->ms.unfolded) hist_browser__get_folding() 139 static char map_symbol__folded(const struct map_symbol *ms) map_symbol__folded() argument 141 return ms->has_children ? tree__folded_sign(ms->unfolded) : ' '; map_symbol__folded() 146 return map_symbol__folded(&he->ms); hist_entry__folded() 151 return map_symbol__folded(&cl->ms); callchain_list__folded() 154 static void map_symbol__set_folding(struct map_symbol *ms, bool unfold) map_symbol__set_folding() argument 156 ms->unfolded = unfold ? ms->has_children : false; map_symbol__set_folding() 192 unfolded = chain->ms.unfolded; callchain_node__count_rows() 214 static bool map_symbol__toggle_fold(struct map_symbol *ms) map_symbol__toggle_fold() argument 216 if (!ms) map_symbol__toggle_fold() 219 if (!ms->has_children) map_symbol__toggle_fold() 222 ms->unfolded = !ms->unfolded; map_symbol__toggle_fold() 238 chain->ms.has_children = chain->list.next != &child->val || callchain_node__init_have_children_rb_tree() 241 chain->ms.has_children = chain->list.next == &child->val && callchain_node__init_have_children_rb_tree() 255 chain->ms.has_children = has_sibling; callchain_node__init_have_children() 259 chain->ms.has_children = !RB_EMPTY_ROOT(&node->rb_root); callchain_node__init_have_children() 279 he->ms.has_children = !RB_EMPTY_ROOT(&he->sorted_chain); hist_entry__init_have_children() 294 if (he->ms.unfolded) hist_browser__toggle_fold() 321 map_symbol__set_folding(&chain->ms, unfold); callchain_node__set_folding_rb_tree() 322 has_children = chain->ms.has_children; callchain_node__set_folding_rb_tree() 340 map_symbol__set_folding(&chain->ms, unfold); callchain_node__set_folding() 341 has_children = chain->ms.has_children; callchain_node__set_folding() 366 map_symbol__set_folding(&he->ms, unfold); hist_entry__set_folding() 368 if (he->ms.has_children) { hist_entry__set_folding() 514 bool show_annotated = browser->show_dso && chain->ms.sym && symbol__annotation(chain->ms.sym)->src; hist_browser__show_callchain_entry() 519 browser->selection = &chain->ms; hist_browser__show_callchain_entry() 751 browser->selection = &entry->ms; hist_browser__show_entry() 1019 if (h->ms.unfolded) { ui_browser__hists_seek() 1040 if (h->ms.unfolded) { ui_browser__hists_seek() 1077 if (h->ms.unfolded) ui_browser__hists_seek() 1680 sym = browser->he_selection->ms.sym; perf_evsel__hists_browse() 1709 struct map_symbol ms; perf_evsel__hists_browse() local 1720 ms.map = he->branch_info->from.map; perf_evsel__hists_browse() 1721 ms.sym = he->branch_info->from.sym; perf_evsel__hists_browse() 1723 ms.map = he->branch_info->to.map; perf_evsel__hists_browse() 1724 ms.sym = he->branch_info->to.sym; perf_evsel__hists_browse() 1726 ms = *browser->selection; perf_evsel__hists_browse() 1729 notes = symbol__annotation(ms.sym); perf_evsel__hists_browse() 1733 err = map_symbol__tui_annotate(&ms, evsel, hbt); perf_evsel__hists_browse() 1796 sprintf(script_opt, " -S %s ", browser->he_selection->ms.sym->name); perf_evsel__hists_browse()
|
/linux-4.1.27/drivers/video/fbdev/via/ |
H A D | vt1636.c | 30 /* T1: VDD on - Data on. Each increment is 1 ms. (50ms = 031h) */ 32 /* T2: Data on - Backlight on. Each increment is 2 ms. (210ms = 068h) */ 34 /* T3: Backlight off -Data off. Each increment is 2 ms. (210ms = 068h)*/ 36 /* T4: Data off - VDD off. Each increment is 1 ms. (50ms = 031h) */ 38 /* T5: VDD off - VDD on. Each increment is 100 ms. (500ms = 04h) */
|
/linux-4.1.27/drivers/usb/phy/ |
H A D | phy-fsl-usb.h | 248 #define TA_WAIT_VRISE (100) /* a_wait_vrise 100 ms, section: 6.6.5.1 */ 258 /* It is necessary for this timer to be more than 750 ms because of a bug in OPT 259 * test 5.4 in which B OPT disconnects after 750 ms instead of 75ms as stated 262 #define TA_AIDL_BDIS (5000) /* a_suspend minimum 200 ms, section: 6.6.5.3 */ 265 #define TA_BIDL_ADIS (12) /* 3 to 200 ms */ 271 #define TB_DATA_PLS (10) /* b_srp_init,continue 5~10ms, section:5.3.3 */ 272 #define TB_DATA_PLS_MIN (5) /* minimum 5 ms */ 273 #define TB_DATA_PLS_MAX (10) /* maximum 10 ms */ 276 #define TB_SRP_INIT (100) /* b_srp_init,maximum 100 ms, section:5.3.8 */ 288 /* This time should be less than 10ms. It varies from system to system. */ 292 #define TB_ASE0_BRST (20) /* b_wait_acon, mini 3.125 ms,section:6.8.2.4 */ 299 #define TB_SE0_SRP (2) /* b_idle,minimum 2 ms, section:5.3.2 */
|
/linux-4.1.27/arch/ia64/kernel/ |
H A D | mca.c | 896 const pal_min_state_area_t *ms = sos->pal_min_state; finish_pt_regs() local 903 regs->cr_iip = ms->pmsa_iip; finish_pt_regs() 904 regs->cr_ipsr = ms->pmsa_ipsr; finish_pt_regs() 905 regs->cr_ifs = ms->pmsa_ifs; finish_pt_regs() 907 regs->cr_iip = ms->pmsa_xip; finish_pt_regs() 908 regs->cr_ipsr = ms->pmsa_xpsr; finish_pt_regs() 909 regs->cr_ifs = ms->pmsa_xfs; finish_pt_regs() 911 sos->iip = ms->pmsa_iip; finish_pt_regs() 912 sos->ipsr = ms->pmsa_ipsr; finish_pt_regs() 913 sos->ifs = ms->pmsa_ifs; finish_pt_regs() 915 regs->pr = ms->pmsa_pr; finish_pt_regs() 916 regs->b0 = ms->pmsa_br0; finish_pt_regs() 917 regs->ar_rsc = ms->pmsa_rsc; finish_pt_regs() 918 copy_reg(&ms->pmsa_gr[1-1], ms->pmsa_nat_bits, ®s->r1, nat); finish_pt_regs() 919 copy_reg(&ms->pmsa_gr[2-1], ms->pmsa_nat_bits, ®s->r2, nat); finish_pt_regs() 920 copy_reg(&ms->pmsa_gr[3-1], ms->pmsa_nat_bits, ®s->r3, nat); finish_pt_regs() 921 copy_reg(&ms->pmsa_gr[8-1], ms->pmsa_nat_bits, ®s->r8, nat); finish_pt_regs() 922 copy_reg(&ms->pmsa_gr[9-1], ms->pmsa_nat_bits, ®s->r9, nat); finish_pt_regs() 923 copy_reg(&ms->pmsa_gr[10-1], ms->pmsa_nat_bits, ®s->r10, nat); finish_pt_regs() 924 copy_reg(&ms->pmsa_gr[11-1], ms->pmsa_nat_bits, ®s->r11, nat); finish_pt_regs() 925 copy_reg(&ms->pmsa_gr[12-1], ms->pmsa_nat_bits, ®s->r12, nat); finish_pt_regs() 926 copy_reg(&ms->pmsa_gr[13-1], ms->pmsa_nat_bits, ®s->r13, nat); finish_pt_regs() 927 copy_reg(&ms->pmsa_gr[14-1], ms->pmsa_nat_bits, ®s->r14, nat); finish_pt_regs() 928 copy_reg(&ms->pmsa_gr[15-1], ms->pmsa_nat_bits, ®s->r15, nat); finish_pt_regs() 930 bank = ms->pmsa_bank1_gr; finish_pt_regs() 932 bank = ms->pmsa_bank0_gr; finish_pt_regs() 933 copy_reg(&bank[16-16], ms->pmsa_nat_bits, ®s->r16, nat); finish_pt_regs() 934 copy_reg(&bank[17-16], ms->pmsa_nat_bits, ®s->r17, nat); finish_pt_regs() 935 copy_reg(&bank[18-16], ms->pmsa_nat_bits, ®s->r18, nat); finish_pt_regs() 936 copy_reg(&bank[19-16], ms->pmsa_nat_bits, ®s->r19, nat); finish_pt_regs() 937 copy_reg(&bank[20-16], ms->pmsa_nat_bits, ®s->r20, nat); finish_pt_regs() 938 copy_reg(&bank[21-16], ms->pmsa_nat_bits, ®s->r21, nat); finish_pt_regs() 939 copy_reg(&bank[22-16], ms->pmsa_nat_bits, ®s->r22, nat); finish_pt_regs() 940 copy_reg(&bank[23-16], ms->pmsa_nat_bits, ®s->r23, nat); finish_pt_regs() 941 copy_reg(&bank[24-16], ms->pmsa_nat_bits, ®s->r24, nat); finish_pt_regs() 942 copy_reg(&bank[25-16], ms->pmsa_nat_bits, ®s->r25, nat); finish_pt_regs() 943 copy_reg(&bank[26-16], ms->pmsa_nat_bits, ®s->r26, nat); finish_pt_regs() 944 copy_reg(&bank[27-16], ms->pmsa_nat_bits, ®s->r27, nat); finish_pt_regs() 945 copy_reg(&bank[28-16], ms->pmsa_nat_bits, ®s->r28, nat); finish_pt_regs() 946 copy_reg(&bank[29-16], ms->pmsa_nat_bits, ®s->r29, nat); finish_pt_regs() 947 copy_reg(&bank[30-16], ms->pmsa_nat_bits, ®s->r30, nat); finish_pt_regs() 948 copy_reg(&bank[31-16], ms->pmsa_nat_bits, ®s->r31, nat); finish_pt_regs() 972 const pal_min_state_area_t *ms = sos->pal_min_state; ia64_mca_modify_original_stack() local 982 u64 r12 = ms->pmsa_gr[12-1], r13 = ms->pmsa_gr[13-1]; ia64_mca_modify_original_stack() 996 regs->cr_ipsr = ms->pmsa_ipsr; ia64_mca_modify_original_stack() 1053 if (!mca_recover_range(ms->pmsa_iip)) { ia64_mca_modify_original_stack() 1112 copy_reg(&ms->pmsa_gr[4-1], ms->pmsa_nat_bits, &old_sw->r4, &old_unat); ia64_mca_modify_original_stack() 1113 copy_reg(&ms->pmsa_gr[5-1], ms->pmsa_nat_bits, &old_sw->r5, &old_unat); ia64_mca_modify_original_stack() 1114 copy_reg(&ms->pmsa_gr[6-1], ms->pmsa_nat_bits, &old_sw->r6, &old_unat); ia64_mca_modify_original_stack() 1115 copy_reg(&ms->pmsa_gr[7-1], ms->pmsa_nat_bits, &old_sw->r7, &old_unat); ia64_mca_modify_original_stack() 1117 old_sw->b1 = ms->pmsa_br1; ia64_mca_modify_original_stack()
|
/linux-4.1.27/drivers/iio/light/ |
H A D | cm3323.c | 41 {0, 40000}, /* 40 ms */ 42 {0, 80000}, /* 80 ms */ 43 {0, 160000}, /* 160 ms */ 44 {0, 320000}, /* 320 ms */ 45 {0, 640000}, /* 640 ms */ 46 {1, 280000}, /* 1280 ms */
|
H A D | cm32181.c | 44 #define CM32181_MLUX_PER_BIT 5 /* ALS_SM=01 IT=800ms */ 45 #define CM32181_MLUX_PER_BIT_BASE_IT 800000 /* Based on IT=800ms */ 104 * cm32181_read_als_it() - Get sensor integration time (ms) 135 * Convert integration time (ms) to sensor value.
|
/linux-4.1.27/drivers/staging/rtl8712/ |
H A D | osdep_service.h | 80 static inline void sleep_schedulable(int ms) sleep_schedulable() argument 84 delta = msecs_to_jiffies(ms);/*(ms)*/ sleep_schedulable()
|
/linux-4.1.27/drivers/staging/rtl8188eu/include/ |
H A D | usb_ops_linux.h | 25 #define RTW_USB_CONTROL_MSG_TIMEOUT_TEST 10/* ms */ 26 #define RTW_USB_CONTROL_MSG_TIMEOUT 500/* ms */ 30 #define RTW_USB_BULKOUT_TIME 5000/* ms */
|
/linux-4.1.27/arch/mips/jz4740/ |
H A D | reset.c | 84 * Set minimum wakeup pin assertion time: 100 ms. jz4740_power_off() 96 * Set reset pin low-level assertion time after wakeup: 60 ms. jz4740_power_off() 97 * Range is 0 to 125 ms if RTC is clocked at 32 kHz. jz4740_power_off()
|
/linux-4.1.27/tools/testing/selftests/timers/ |
H A D | mqueue-lat.c | 47 #define TARGET_TIMEOUT 100000000 /* 100ms in nanoseconds */ 48 #define UNRESONABLE_LATENCY 40000000 /* 40ms in nanosecs */ 93 target = timespec_add(now, TARGET_TIMEOUT); /* 100ms */ mqueue_lat_test()
|
/linux-4.1.27/arch/mips/include/asm/ |
H A D | mc146818-time.h | 16 * For check timing call set_rtc_mmss() 500ms; used in timer interrupt. 23 * called 500 ms after the second nowtime has started, because when 25 * jump to the next second precisely 500 ms later. Check the Motorola 78 * update precisely 500 ms later. You won't find this mentioned in mc146818_set_rtc_mmss()
|
/linux-4.1.27/include/net/irda/ |
H A D | timer.h | 46 #define POLL_TIMEOUT (450*HZ/1000) /* Must never exceed 500 ms */ 47 #define FINAL_TIMEOUT (500*HZ/1000) /* Must never exceed 500 ms */ 59 * Slot timer must never exceed 85 ms, and must always be at least 25 ms,
|
/linux-4.1.27/include/linux/power/ |
H A D | generic-adc-battery.h | 19 * status.Default set is 10ms.
|
/linux-4.1.27/tools/perf/tests/ |
H A D | hists_common.c | 174 he->ms.map->dso->short_name, print_hists_in() 175 he->ms.sym->name, he->stat.period); print_hists_in() 201 he->ms.map->dso->short_name, print_hists_out() 202 he->ms.sym->name, he->stat.period, print_hists_out()
|
H A D | hists_link.c | 170 he->thread, he->ms.map, he->ms.sym)) { __validate_match() 222 he->thread, he->ms.map, he->ms.sym) && __validate_link() 225 he->thread, he->ms.map, he->ms.sym)) { __validate_link()
|
/linux-4.1.27/fs/minix/ |
H A D | inode.c | 124 struct minix_super_block * ms; minix_remount() local 127 ms = sbi->s_ms; minix_remount() 131 if (ms->s_state & MINIX_VALID_FS || minix_remount() 136 ms->s_state = sbi->s_mount_state; minix_remount() 141 sbi->s_mount_state = ms->s_state; minix_remount() 142 ms->s_state &= ~MINIX_VALID_FS; minix_remount() 162 struct minix_super_block *ms; minix_fill_super() local 183 ms = (struct minix_super_block *) bh->b_data; minix_fill_super() 184 sbi->s_ms = ms; minix_fill_super() 186 sbi->s_mount_state = ms->s_state; minix_fill_super() 187 sbi->s_ninodes = ms->s_ninodes; minix_fill_super() 188 sbi->s_nzones = ms->s_nzones; minix_fill_super() 189 sbi->s_imap_blocks = ms->s_imap_blocks; minix_fill_super() 190 sbi->s_zmap_blocks = ms->s_zmap_blocks; minix_fill_super() 191 sbi->s_firstdatazone = ms->s_firstdatazone; minix_fill_super() 192 sbi->s_log_zone_size = ms->s_log_zone_size; minix_fill_super() 193 sbi->s_max_size = ms->s_max_size; minix_fill_super() 194 s->s_magic = ms->s_magic; minix_fill_super() 207 sbi->s_nzones = ms->s_zones; minix_fill_super() 213 sbi->s_nzones = ms->s_zones; minix_fill_super() 298 ms->s_state &= ~MINIX_VALID_FS; minix_fill_super()
|
/linux-4.1.27/drivers/scsi/fnic/ |
H A D | fnic_fip.h | 23 #define FCOE_CTLR_START_DELAY 2000 /* ms after first adv. to choose FCF */ 24 #define FCOE_CTLR_FIPVLAN_TOV 2000 /* ms after FIP VLAN disc */
|
/linux-4.1.27/drivers/media/pci/cx88/ |
H A D | cx88-input.c | 289 ir->polling = 50; /* ms */ cx88_ir_init() 318 ir->polling = 50; /* ms */ cx88_ir_init() 329 ir->polling = 1; /* ms */ cx88_ir_init() 336 ir->polling = 5; /* ms */ cx88_ir_init() 351 ir->polling = 10; /* ms */ cx88_ir_init() 360 ir->polling = 1; /* ms */ cx88_ir_init() 367 ir->polling = 1; /* ms */ cx88_ir_init() 374 ir->polling = 50; /* ms */ cx88_ir_init() 381 ir->polling = 1; /* ms */ cx88_ir_init() 389 ir->polling = 50; /* ms */ cx88_ir_init() 416 ir->polling = 50; /* ms */ cx88_ir_init() 422 ir->polling = 50; /* ms */ cx88_ir_init() 432 ir->polling = 100; /* ms */ cx88_ir_init() 485 dev->timeout = 10 * 1000 * 1000; /* 10 ms */ cx88_ir_init()
|
/linux-4.1.27/drivers/media/rc/ |
H A D | ir-sharp-decoder.c | 25 #define SHARP_BIT_0_PERIOD (25 * SHARP_UNIT) /* 1ms (680us space) */ 26 #define SHARP_BIT_1_PERIOD (50 * SHARP_UNIT) /* 2ms (1680ms space) */ 27 #define SHARP_ECHO_SPACE (1000 * SHARP_UNIT) /* 40 ms */ 28 #define SHARP_TRAILER_SPACE (125 * SHARP_UNIT) /* 5 ms (even longer) */
|
H A D | nuvoton-cir.h | 380 /* MCE CIR controller signal length: about 43ms 381 * 43ms / 50us (sample period) * 0.85 (inaccuracy) 385 /* MCE CIR keyboard signal length: about 26ms 386 * 26ms / 50us (sample period) * 0.85 (inaccuracy) 391 /* MCE CIR mouse signal length: about 24ms 392 * 24ms / 50us (sample period) * 0.85 (inaccuracy)
|
/linux-4.1.27/net/hsr/ |
H A D | hsr_main.h | 23 #define HSR_LIFE_CHECK_INTERVAL 2000 /* ms */ 24 #define HSR_NODE_FORGET_TIME 60000 /* ms */ 25 #define HSR_ANNOUNCE_INTERVAL 100 /* ms */ 31 #define MAX_SLAVE_DIFF 3000 /* ms */ 38 #define PRUNE_PERIOD 3000 /* ms */
|
/linux-4.1.27/drivers/scsi/aic94xx/ |
H A D | aic94xx_sds.c | 683 static int asd_validate_ms(struct asd_manuf_sec *ms) asd_validate_ms() argument 685 if (ms->sig[0] != 'S' || ms->sig[1] != 'M') { asd_validate_ms() 687 ms->sig[0], ms->sig[1]); asd_validate_ms() 690 if (ms->maj != 0) { asd_validate_ms() 692 ms->maj); asd_validate_ms() 695 ms->offs_next = le16_to_cpu((__force __le16) ms->offs_next); asd_validate_ms() 696 ms->chksum = le16_to_cpu((__force __le16) ms->chksum); asd_validate_ms() 697 ms->size = le16_to_cpu((__force __le16) ms->size); asd_validate_ms() 699 if (asd_calc_flash_chksum((u16 *)ms, ms->size/2)) { asd_validate_ms() 707 struct asd_manuf_sec *ms) asd_ms_get_sas_addr() 709 memcpy(asd_ha->hw_prof.sas_addr, ms->sas_addr, SAS_ADDR_SIZE); asd_ms_get_sas_addr() 714 struct asd_manuf_sec *ms) asd_ms_get_pcba_sn() 716 memcpy(asd_ha->hw_prof.pcba_sn, ms->pcba_sn, ASD_PCBA_SN_SIZE); asd_ms_get_pcba_sn() 781 ASD_DPRINTK("ms: no phy parameters found\n"); asd_ms_get_phy_params() 782 ASD_DPRINTK("ms: Creating default phy parameters\n"); asd_ms_get_phy_params() 808 ASD_DPRINTK("ms: num_phy_desc: %d\n", phy_param->num_phy_desc); asd_ms_get_phy_params() 814 ASD_DPRINTK("ms: phy%d: HIDDEN\n", i); asd_ms_get_phy_params() 817 ASD_DPRINTK("ms: phy%d: REPORTED\n", i); asd_ms_get_phy_params() 822 ASD_DPRINTK("ms: phy%d: ENABLED\n", i); asd_ms_get_phy_params() 834 ASD_DPRINTK("ms: max_phys:0x%x, num_phys:0x%x\n", asd_ms_get_phy_params() 836 ASD_DPRINTK("ms: enabled_phys:0x%x\n", asd_ha->hw_prof.enabled_phys); asd_ms_get_phy_params() 847 ASD_DPRINTK("ms: no connector map found\n"); asd_ms_get_connector_map() 852 ASD_DPRINTK("ms: unsupported: connector map major version 0x%x" asd_ms_get_connector_map() 921 ASD_DPRINTK("ms: couldn't get phy parameters\n"); asd_process_ms() 927 ASD_DPRINTK("ms: couldn't get connector map\n"); asd_process_ms() 706 asd_ms_get_sas_addr(struct asd_ha_struct *asd_ha, struct asd_manuf_sec *ms) asd_ms_get_sas_addr() argument 713 asd_ms_get_pcba_sn(struct asd_ha_struct *asd_ha, struct asd_manuf_sec *ms) asd_ms_get_pcba_sn() argument
|
/linux-4.1.27/drivers/input/touchscreen/ |
H A D | bcm_iproc_tsc.c | 101 * The settling duration (in ms) is the amount of time the tsc 104 * 0 = 0.008 ms 105 * 1 = 0.01 ms 106 * 2 = 0.02 ms 107 * 3 = 0.04 ms 108 * 4 = 0.08 ms 109 * 5 = 0.16 ms 110 * 6 = 0.32 ms 111 * 7 = 0.64 ms 112 * 8 = 1.28 ms 113 * 9 = 2.56 ms 114 * 10 = 5.12 ms 115 * 11 = 10.24 ms
|
/linux-4.1.27/drivers/staging/lustre/lustre/libcfs/ |
H A D | fail.c | 124 int __cfs_fail_timeout_set(__u32 id, __u32 value, int ms, int set) __cfs_fail_timeout_set() argument 131 id, ms); __cfs_fail_timeout_set() 133 schedule_timeout(cfs_time_seconds(ms) / 1000); __cfs_fail_timeout_set()
|
/linux-4.1.27/drivers/md/bcache/ |
H A D | io.c | 223 int ms = us / 1024; bch_bbio_count_io_errors() local 226 ms = min(ms, CONGESTED_MAX + congested); bch_bbio_count_io_errors() 227 atomic_sub(ms, &c->congested); bch_bbio_count_io_errors()
|
H A D | sysfs.c | 53 sysfs_time_stats_attribute(btree_gc, sec, ms); 55 sysfs_time_stats_attribute(btree_sort, ms, us); 56 sysfs_time_stats_attribute(btree_read, ms, us); 532 sysfs_print_time_stats(&c->btree_gc_time, btree_gc, sec, ms); SHOW() 534 sysfs_print_time_stats(&c->sort.time, btree_sort, ms, us); SHOW() 535 sysfs_print_time_stats(&c->btree_read_time, btree_read, ms, us); SHOW() 706 sysfs_time_stats_attribute_list(btree_gc, sec, ms) 708 sysfs_time_stats_attribute_list(btree_sort, ms, us) 709 sysfs_time_stats_attribute_list(btree_read, ms, us)
|
/linux-4.1.27/drivers/media/platform/vivid/ |
H A D | vivid-rds-gen.c | 76 data[1].lsb |= (rds->ta << 4) | (rds->ms << 3); vivid_rds_generate() 122 data[1].lsb |= (rds->ta << 4) | (rds->ms << 3); vivid_rds_generate() 129 data[3].lsb |= (rds->ta << 4) | (rds->ms << 3); vivid_rds_generate() 155 rds->ms = true; vivid_rds_gen_fill()
|
H A D | vivid-rds-gen.h | 44 bool ms; member in struct:vivid_rds_gen
|
H A D | vivid-radio-common.c | 90 rds->ms = dev->radio_tx_rds_ms->cur.val; vivid_radio_rds_init() 105 v4l2_ctrl_s_ctrl(dev->radio_rx_rds_ms, rds->ms); vivid_radio_rds_init()
|
/linux-4.1.27/drivers/staging/speakup/ |
H A D | serialio.h | 33 /* buffer timeout in ms */
|
/linux-4.1.27/drivers/pcmcia/ |
H A D | bcm63xx_pcmcia.h | 9 /* socket polling rate in ms */
|
H A D | sa1111_lubbock.c | 167 * Note: It takes about 2.5ms for the MAX1602 VCC output lubbock_pcmcia_configure_socket() 182 * It takes about 100ms to turn off Vcc. lubbock_pcmcia_configure_socket()
|
/linux-4.1.27/drivers/input/mouse/ |
H A D | hgpk.h | 26 #define SPEW_WATCH_COUNT 42 /* at 12ms/packet, this is 1/2 second */
|
H A D | gpio_mouse.c | 19 * Timer function which is run every scan_ms ms when the device is opened. 130 dev_dbg(&pdev->dev, "%d ms scan time, buttons: %s%s%s\n", gpio_mouse_probe()
|
/linux-4.1.27/tools/power/cpupower/bench/ |
H A D | benchmark.h | 20 /* load loop, this schould take about 1 to 2ms to complete */
|
/linux-4.1.27/drivers/staging/vt6655/ |
H A D | card.h | 44 #define DEFAULT_MSDU_LIFETIME 512 /* ms */ 47 #define DEFAULT_MGN_LIFETIME 8 /* ms */
|
/linux-4.1.27/drivers/input/serio/ |
H A D | i8042-sparcio.h | 71 struct platform_device *ms = of_find_device_by_node(dp); sparc_i8042_probe() local 72 unsigned int irq = ms->archdata.irqs[0]; sparc_i8042_probe()
|
/linux-4.1.27/drivers/pinctrl/sh-pfc/ |
H A D | sh_pfc.h | 170 #define PINMUX_IPSR_NOGM(ispr, fn, ms) \ 171 PINMUX_DATA(fn##_MARK, FN_##fn, FN_##ms) 172 #define PINMUX_IPSR_NOFN(ipsr, fn, ms) \ 173 PINMUX_DATA(fn##_MARK, FN_##ipsr, FN_##ms) 174 #define PINMUX_IPSR_MSEL(ipsr, fn, ms) \ 175 PINMUX_DATA(fn##_MARK, FN_##fn, FN_##ipsr, FN_##ms) 176 #define PINMUX_IPSR_MODSEL_DATA(ipsr, fn, ms) \ 177 PINMUX_DATA(fn##_MARK, FN_##ms, FN_##ipsr, FN_##fn)
|
/linux-4.1.27/arch/powerpc/include/asm/ |
H A D | delay.h | 26 * is supposed to take 1ms, the hypervisor could sleep our partition for 27 * longer (eg 10ms). With the right timing these errors can add up.
|
/linux-4.1.27/arch/arm/mach-rockchip/ |
H A D | pm.h | 72 /* wait 30ms for OSC stable and 30ms for pmic stable */
|
/linux-4.1.27/kernel/time/ |
H A D | timer_stats.c | 284 unsigned long ms; tstats_show() local 299 ms = period.tv_nsec / 1000000; tstats_show() 302 seq_printf(m, "Sample period: %ld.%03ld s\n", period.tv_sec, ms); tstats_show() 325 ms += period.tv_sec * 1000; tstats_show() 326 if (!ms) tstats_show() 327 ms = 1; tstats_show() 331 events, events * 1000 / ms, tstats_show() 332 (events * 1000000 / ms) % 1000); tstats_show()
|
/linux-4.1.27/drivers/media/usb/gspca/ |
H A D | finepix.c | 31 /* Default timeout, in ms */ 48 * the camera will return old frames. On the 4800Z, 20ms is bad, 25ms 49 * will fail every 4 or 5 frames, but 30ms is perfect. On the A210, 50 * 30ms is bad while 35ms is perfect. */
|
/linux-4.1.27/tools/perf/util/ |
H A D | callchain.c | 464 call->ms.sym = cursor_node->sym; fill_node() 465 call->ms.map = cursor_node->map; fill_node() 493 if (cnode->ms.sym && sym && match_chain() 495 return cnode->ms.sym->start - sym->start; match_chain() 701 list->ms.map, list->ms.sym); merge_chain_branch() 829 if (cl->ms.sym) { callchain_list__sym_name() 831 cl->ms.map && !cl->srcline) callchain_list__sym_name() 832 cl->srcline = get_srcline(cl->ms.map->dso, callchain_list__sym_name() 833 map__rip_2objdump(cl->ms.map, callchain_list__sym_name() 835 cl->ms.sym, false); callchain_list__sym_name() 838 cl->ms.sym->name, cl->srcline); callchain_list__sym_name() 840 printed = scnprintf(bf, bfsize, "%s", cl->ms.sym->name); callchain_list__sym_name() 846 cl->ms.map ? callchain_list__sym_name() 847 cl->ms.map->dso->short_name : callchain_list__sym_name()
|
H A D | hist.c | 67 if (h->ms.sym) { hists__calc_col_len() 68 symlen = h->ms.sym->namelen + 4; hists__calc_col_len() 82 if (h->ms.map) { hists__calc_col_len() 83 len = dso__name_len(h->ms.map->dso); hists__calc_col_len() 319 if (he->ms.map) hist_entry__new() 320 he->ms.map->referenced = true; hist_entry__new() 412 if (he->ms.map != entry->ms.map) { add_hist_entry() 413 he->ms.map = entry->ms.map; add_hist_entry() 414 if (he->ms.map) add_hist_entry() 415 he->ms.map->referenced = true; add_hist_entry() 453 .ms = { __hists__add_entry() 775 .ms = { iter_add_next_cumulative_entry() 1169 h->ms.unfolded = false; hists__remove_entry_filter() 1184 (he->ms.map == NULL || he->ms.map->dso != hists->dso_filter)) { hists__filter_entry_by_dso() 1249 (!he->ms.sym || strstr(he->ms.sym->name, hists__filter_entry_by_symbol()
|
H A D | sort.c | 148 return _sort__dso_cmp(right->ms.map, left->ms.map); sort__dso_cmp() 166 return _hist_entry__dso_snprintf(he->ms.map, bf, size, width); hist_entry__dso_snprintf() 204 if (!left->ms.sym && !right->ms.sym) sort__sym_cmp() 217 return _sort__sym_cmp(left->ms.sym, right->ms.sym); sort__sym_cmp() 223 if (!left->ms.sym || !right->ms.sym) sort__sym_sort() 224 return cmp_null(left->ms.sym, right->ms.sym); sort__sym_sort() 226 return strcmp(right->ms.sym->name, left->ms.sym->name); sort__sym_sort() 271 return _hist_entry__sym_snprintf(he->ms.map, he->ms.sym, he->ip, hist_entry__sym_snprintf() 289 if (!left->ms.map) sort__srcline_cmp() 292 struct map *map = left->ms.map; sort__srcline_cmp() 295 left->ms.sym, true); sort__srcline_cmp() 299 if (!right->ms.map) sort__srcline_cmp() 302 struct map *map = right->ms.map; sort__srcline_cmp() 305 right->ms.sym, true); sort__srcline_cmp()
|
/linux-4.1.27/drivers/media/pci/bt8xx/ |
H A D | bttv-input.c | 215 /* Allow some timer jitter (RC5 is ~24ms anyway so this is ok) */ bttv_rc5_timer_end() 444 ir->polling = 50; // ms bttv_input_init() 452 ir->polling = 50; // ms bttv_input_init() 459 ir->polling = 50; // ms bttv_input_init() 467 ir->polling = 50; // ms bttv_input_init() 484 ir->polling = 50; // ms bttv_input_init() 493 ir->polling = 50; // ms bttv_input_init() 503 ir->polling = 50; /* ms */ bttv_input_init() 509 ir->polling = 50; /* ms */ bttv_input_init() 515 ir->polling = 1; /* ms */ bttv_input_init()
|
/linux-4.1.27/drivers/media/usb/tm6000/ |
H A D | tm6000-input.c | 44 #define URB_SUBMIT_DELAY 100 /* ms - Delay to submit an URB request on retrial and init */ 45 #define URB_INT_LED_DELAY 100 /* ms - Delay to turn led on again on int mode */ 107 * It seems that the 900ms timing for NEC is used to detect a RC-5 tm6000_ir_config() 113 leader = 900; /* ms */ tm6000_ir_config() 114 pulse = 700; /* ms - the actual value would be 562 */ tm6000_ir_config() 118 leader = 900; /* ms - from the NEC decoding */ tm6000_ir_config() 119 pulse = 1780; /* ms - The actual value would be 1776 */ tm6000_ir_config() 223 * So, use the scheduler to do it, in a few ms. tm6000_ir_urb_received() 281 /* Retry in 100 ms */ tm6000_ir_int_work()
|
/linux-4.1.27/drivers/media/i2c/ |
H A D | wm8775.c | 53 #define ALC_HOLD 0x85 /* R17: use zero cross detection, ALC hold time 42.6 ms */ 277 ALC hold time 42.6 ms */ wm8775_probe() 285 /* ALC gain ramp up delay 34 s, ALC gain ramp down delay 33 ms */ wm8775_probe() 290 /* Transient window 4ms, lower PGA gain limit -1dB */ wm8775_probe() 295 /* Transient window 4ms, ALC min gain -5dB */ wm8775_probe()
|
/linux-4.1.27/drivers/media/pci/saa7134/ |
H A D | saa7134-input.c | 619 polling = 50; // ms saa7134_input_init1() 626 polling = 50; // ms saa7134_input_init1() 643 polling = 50; // ms saa7134_input_init1() 665 polling = 50; // ms saa7134_input_init1() 672 polling = 50; /* ms */ saa7134_input_init1() 679 polling = 50; // ms saa7134_input_init1() 686 polling = 50; /* ms */ saa7134_input_init1() 706 polling = 50; /* ms */ saa7134_input_init1() 712 polling = 50; // ms saa7134_input_init1() 718 polling = 50; // ms saa7134_input_init1() 724 polling = 5; /* ms */ saa7134_input_init1() 733 polling = 50; // ms saa7134_input_init1() 739 polling = 50; // ms saa7134_input_init1() 775 polling = 50; // ms saa7134_input_init1() 789 polling = 50; //ms saa7134_input_init1() 795 polling = 50; /* ms */ saa7134_input_init1() 801 polling = 50; /* ms */ saa7134_input_init1() 806 polling = 40; /* ms */ saa7134_input_init1() 817 polling = 50; /* ms */ saa7134_input_init1() 1044 * Wait 15 ms from the start of the first IR event before processing saa7134_raw_decode_irq()
|
/linux-4.1.27/drivers/watchdog/ |
H A D | shwdt.c | 48 * implementation that will deal with this properly), otherwise the 10ms 55 * feasible are the 4096 and the 2048 divisors, which yield 5.25 and 2.62ms 60 * the kernel for clearing out WOVF every 2ms or so (again, this depends on 340 "Clock division ratio. Valid ranges are from 0x5 (1.31ms) " 341 "to 0x7 (5.25ms). (default=" __MODULE_STRING(WTCSR_CKS_4096) ")");
|
H A D | ux500_wdt.c | 26 #define WATCHDOG_MAX28 268435 /* 28 bit resolution in ms == 268435.455 s */ 27 #define WATCHDOG_MAX32 4294967 /* 32 bit resolution in ms == 4294967.295 s */
|
/linux-4.1.27/drivers/net/phy/ |
H A D | smsc.c | 77 /* wait end of reset (max 500 ms) */ smsc_phy_reset() 94 * The LAN8710/LAN8720 requires a minimum of 2 link pulses within 64ms of each 119 /* Sleep 64 ms to allow ~5 link test pulses to be sent */ lan87xx_read_status()
|
/linux-4.1.27/drivers/staging/rtl8723au/include/ |
H A D | rtw_pwrctrl.h | 207 #define _rtw_set_pwr_state_check_timer(pwrctrlpriv, ms) \ 209 msecs_to_jiffies(ms))) 233 void rtw_set_ips_deny23a(struct rtw_adapter *padapter, u32 ms);
|
H A D | usb_ops_linux.h | 20 #define RTW_USB_CONTROL_MSG_TIMEOUT 500/* ms */
|
H A D | wifi.h | 22 #define WiFiNavUpperUs 30000 /* 30 ms */
|
H A D | rtw_mlme_ext.h | 24 /* Increase the SURVEY_TO value from 100 to 150 ( 100ms to 150ms ) */ 25 /* The Realtek 8188CE SoftAP will spend around 100ms to send the probe response after receiving the probe request. */ 579 #define set_survey_timer(mlmeext, ms) \ 580 /*DBG_8723A("%s set_survey_timer(%p, %d)\n", __func__, (mlmeext), (ms));*/ \ 581 mod_timer(&mlmeext->survey_timer, jiffies + msecs_to_jiffies(ms)); 583 #define set_link_timer(mlmeext, ms) \ 584 /*DBG_8723A("%s set_link_timer(%p, %d)\n", __func__, (mlmeext), (ms));*/ \ 585 mod_timer(&mlmeext->link_timer, jiffies + msecs_to_jiffies(ms));
|
/linux-4.1.27/arch/mips/dec/ |
H A D | time.c | 62 * be called 500 ms after the second nowtime has started, because when 64 * jump to the next second precisely 500 ms later. Check the Dallas 115 * update precisely 500 ms later. You won't find this mentioned rtc_mips_set_mmss()
|
/linux-4.1.27/drivers/uwb/ |
H A D | driver.c | 71 * superframe, that'd be 3 * 65ms = 195 ~ 200 ms. Let's give it time 72 * for jitter and stuff and make it 500 ms.
|
/linux-4.1.27/arch/mn10300/kernel/ |
H A D | rtc.c | 43 * ms after the second nowtime has started, because when nowtime is written 45 * precisely 500 ms later. Check the Motorola MC146818A or Dallas DS12887 data 101 * update precisely 500 ms later. You won't find this mentioned in set_rtc_mmss()
|
/linux-4.1.27/net/ipv4/ |
H A D | tcp_hybla.c | 27 /* Hybla reference round trip time (default= 1/40 sec = 25 ms), in ms */ 30 MODULE_PARM_DESC(rtt0, "reference rout trip time (ms)");
|
/linux-4.1.27/drivers/media/usb/cpia2/ |
H A D | cpia2.h | 60 #define USBIF_ISO_1 2 /* 128 bytes/ms */ 61 #define USBIF_ISO_2 3 /* 384 bytes/ms */ 62 #define USBIF_ISO_3 4 /* 640 bytes/ms */ 63 #define USBIF_ISO_4 5 /* 768 bytes/ms */ 64 #define USBIF_ISO_5 6 /* 896 bytes/ms */ 65 #define USBIF_ISO_6 7 /* 1023 bytes/ms */
|
/linux-4.1.27/drivers/clk/rockchip/ |
H A D | clk.h | 216 #define COMPOSITE(_id, cname, pnames, f, mo, ms, mw, mf, ds, dw,\ 226 .mux_shift = ms, \ 274 #define COMPOSITE_NODIV(_id, cname, pnames, f, mo, ms, mw, mf, \ 284 .mux_shift = ms, \ 292 #define COMPOSITE_NOGATE(_id, cname, pnames, f, mo, ms, mw, mf, \ 302 .mux_shift = ms, \
|
/linux-4.1.27/drivers/tty/hvc/ |
H A D | hvsi_lib.c | 247 static void maybe_msleep(unsigned long ms) maybe_msleep() argument 251 mdelay(ms); maybe_msleep() 253 msleep(ms); maybe_msleep() 274 /* Try for up to 200ms */ hvsilib_read_mctrl() 317 /* Try for up to 200ms, there can be a packet to hvsilib_establish() 342 /* Try for up to 400ms */ hvsilib_establish()
|
/linux-4.1.27/tools/perf/ |
H A D | builtin-annotate.c | 110 return symbol__tty_annotate(he->ms.sym, he->ms.map, evsel, hist_entry__tty_annotate() 125 if (he->ms.sym == NULL || he->ms.map->dso->annotate_warned) hists__find_annotations() 128 notes = symbol__annotation(he->ms.sym); hists__find_annotations() 181 * symbol, free he->ms.sym->src to signal we already hists__find_annotations()
|
H A D | builtin-top.c | 96 if (!he || !he->ms.sym) perf_top__parse_source() 99 sym = he->ms.sym; perf_top__parse_source() 100 map = he->ms.map; perf_top__parse_source() 141 struct symbol *sym = he->ms.sym; __zero_source_counters() 180 if (he == NULL || he->ms.sym == NULL || perf_top__record_precise_ip() 182 top->sym_filter_entry->ms.sym != he->ms.sym) && use_browser != 1)) perf_top__record_precise_ip() 185 sym = he->ms.sym; perf_top__record_precise_ip() 191 ip = he->ms.map->map_ip(he->ms.map, ip); perf_top__record_precise_ip() 204 if (err == -ERANGE && !he->ms.map->erange_warned) perf_top__record_precise_ip() 205 ui__warn_map_erange(he->ms.map, sym, ip); perf_top__record_precise_ip() 225 symbol = he->ms.sym; perf_top__show_details() 236 more = symbol__annotate_printf(symbol, he->ms.map, top->sym_evsel, perf_top__show_details() 355 if (n->ms.sym && !strcmp(buf, n->ms.sym->name)) { perf_top__prompt_symbol() 377 struct symbol *sym = top->sym_filter_entry->ms.sym; perf_top__print_mapped_keys()
|
/linux-4.1.27/drivers/staging/iio/light/ |
H A D | tsl2x7x.h | 47 * @prx_time: 5.2ms prox integration time - 48 * decrease in 2.7ms periods
|
/linux-4.1.27/drivers/net/wireless/ti/wl1251/ |
H A D | ps.c | 27 /* in ms */ 98 wl1251_debug(DEBUG_PSM, "wakeup time: %u ms", wl1251_ps_elp_wakeup()
|
/linux-4.1.27/drivers/cpufreq/ |
H A D | s3c2410-cpufreq.c | 92 /* transition latency is about 5ms worst-case, so 93 * set 10ms to be sure */
|
H A D | vexpress-spc-cpufreq.c | 40 return 1000000; /* 1 ms */ ve_spc_get_transition_latency()
|
/linux-4.1.27/arch/s390/kernel/vdso32/ |
H A D | clock_gettime.S | 44 2: ms %r0,__VDSO_TK_MULT(%r5) /* * tk->mult */ 107 12: ms %r0,__VDSO_TK_MULT(%r5) /* * tk->mult */
|
/linux-4.1.27/drivers/net/bonding/ |
H A D | bond_procfs.c | 101 seq_printf(seq, "MII Polling Interval (ms): %d\n", bond->params.miimon); bond_info_show_master() 102 seq_printf(seq, "Up Delay (ms): %d\n", bond_info_show_master() 104 seq_printf(seq, "Down Delay (ms): %d\n", bond_info_show_master() 111 seq_printf(seq, "ARP Polling Interval (ms): %d\n", bond_info_show_master()
|
/linux-4.1.27/drivers/media/pci/cx23885/ |
H A D | netup-eeprom.c | 87 mdelay(10); /* prophylactic delay, datasheet write cycle time = 5 ms */ netup_eeprom_write()
|
/linux-4.1.27/drivers/net/wan/lmc/ |
H A D | lmc_debug.c | 66 unsigned long j = jiffies + 3; /* Wait for 50 ms */ lmc_trace()
|
/linux-4.1.27/drivers/misc/mei/ |
H A D | hw-me.h | 47 #define MEI_ME_RPM_TIMEOUT 500 /* ms */
|
H A D | hw-txe.h | 25 #define MEI_TXI_RPM_TIMEOUT 500 /* ms */
|
/linux-4.1.27/drivers/net/fddi/skfp/h/ |
H A D | sba_def.h | 53 #define TWO_MINUTES 13079 /* 9.175 ms/tick */
|
/linux-4.1.27/drivers/net/wireless/orinoco/ |
H A D | orinoco_pci.c | 65 #define HERMES_PCI_COR_ONT (250) /* ms */ 66 #define HERMES_PCI_COR_OFFT (500) /* ms */ 67 #define HERMES_PCI_COR_BUSYT (500) /* ms */ 77 * need only around 245 iteration of the loop to do 250 ms delay.
|
/linux-4.1.27/drivers/scsi/bnx2fc/ |
H A D | bnx2fc_constants.h | 74 /* E_D_TOV timer resolution in ms */ 81 /* REC timer resolution in ms */ 87 /* E_D_TOV timer default wraparound value (2 sec) in 20 ms resolution */ 91 /* REC_TOV timer default wraparound value (3 sec) in 20 ms resolution */
|
/linux-4.1.27/drivers/staging/iio/accel/ |
H A D | adis16203.h | 4 #define ADIS16203_STARTUP_DELAY 220 /* ms */
|
/linux-4.1.27/drivers/gpu/drm/sti/ |
H A D | sti_compositor.h | 18 #define WAIT_NEXT_VSYNC_MS 50 /*ms*/
|
/linux-4.1.27/drivers/media/dvb-frontends/ |
H A D | stb0899_priv.h | 181 s16 t_derot; /* Derotator time constant (ms) */ 182 s16 t_data; /* Data recovery time constant (ms) */ 185 s16 t_agc1; /* Agc1 time constant (ms) */ 186 s16 t_agc2; /* Agc2 time constant (ms) */
|
/linux-4.1.27/include/sound/ |
H A D | max98095.h | 61 * Default delay is 24 to get a time of 100ms.
|
/linux-4.1.27/arch/s390/include/asm/ |
H A D | nmi.h | 37 __u32 ms : 1; /* 21 psw mask and key validity */ member in struct:mci
|
/linux-4.1.27/arch/mips/mti-sead3/ |
H A D | sead3-time.c | 40 /* wait 1 second (the sampling clock transitions every 10ms) */ estimate_cpu_frequency()
|
/linux-4.1.27/arch/blackfin/include/asm/ |
H A D | time.h | 18 * 10ms is 10,000,000 clock ticks, which fits easy into a 32-bit counter
|
/linux-4.1.27/arch/cris/kernel/ |
H A D | time.c | 10 * fixed 500 ms bug at call to set_rtc_mmss, fixed DS12887
|
/linux-4.1.27/arch/arm/mach-pxa/include/mach/ |
H A D | regs-lcd.h | 193 * and delay ranges from 0ms ~ 255ms 196 #define SMART_DELAY(ms) (SMART_CMD_DELAY | ((ms) & 0xff))
|
/linux-4.1.27/arch/arm/mach-pxa/ |
H A D | reset.c | 83 /* ... in 100 ms */ do_hw_reset()
|
/linux-4.1.27/arch/arm/mach-w90x900/ |
H A D | clksel.c | 45 if (strcmp(dev_id, "nuc900-ms") == 0) clock_source_select()
|
/linux-4.1.27/fs/ext3/ |
H A D | file.c | 18 * (jj@sunsite.ms.mff.cuni.cz)
|
/linux-4.1.27/drivers/net/wireless/rsi/ |
H A D | rsi_91x_sdio_ops.c | 26 * @ms_word: ms word need to be initialized. 114 "%s: Unable to set ms word reg\n", rsi_copy_to_card() 137 "%s: Unable to set ms word to common reg\n", rsi_copy_to_card() 184 "%s: Unable to set ms word to common reg\n", rsi_load_ta_instructions() 193 "%s: Unable to set ms word reg\n", __func__); rsi_load_ta_instructions() 486 rsi_dbg(ERR_ZONE, "%s: Unable to set ms word reg\n", rsi_sdio_device_init() 491 "%s: Setting ms word to 0x41050000\n", __func__); rsi_sdio_device_init()
|
/linux-4.1.27/drivers/usb/host/ |
H A D | ehci-timer.c | 49 * increasing delay values (ranging between 1 ms and 100 ms). 59 * ones that matter most all have a delay of 1 ms, so they will be 60 * handled after 2 ms at most, which is okay. In addition to this, we 61 * allow for an expiration range of 1 ms. 117 /* Poll again later, but give up after about 2-4 ms */ ehci_poll_ASS() 163 /* Poll again later, but give up after about 2-4 ms */ ehci_poll_PSS()
|
/linux-4.1.27/arch/m68k/bvme6000/ |
H A D | config.c | 172 * Set up the RTC timer 1 to mode 2, so T1 output toggles every 5ms 173 * (40000 x 125ns). It will interrupt every 10ms, when T1 goes low. 176 * That gives you the number of 125ns ticks in to the 10ms period, 248 v += 10000; /* Int pending, + 10ms */ bvme6000_gettimeoffset() 313 * Algorithm is to wait for the 10ms register to change, and then to
|
/linux-4.1.27/drivers/net/wireless/ti/wlcore/ |
H A D | conf.h | 118 * Range: 0 - 255 (ms) 127 * Range: 0 - 255 (ms) 136 * Range: 0 - 255 (ms) 145 * Range: 0 - 255 (ms) 154 * Range: 0 - 255 (ms) 163 * Range: 0 - 255 (ms) 263 * Range: 0 - 100000 (ms) 305 * Range: 0 - 255 (ms) 312 * Range: 0 - 255 (ms) 707 /* Time in ms for Tx watchdog timer to expire */ 968 * Specifies the dynamic PS timeout in ms that will be used 982 * frame in ms. 1060 * Range: 0 - 60000 ms 1105 * connected interface. It shouldn't exceed 30000 (~30ms) to avoid
|
/linux-4.1.27/drivers/ide/ |
H A D | ide-eh.c | 163 * every 50ms during an atapi drive reset operation. If the drive has not yet 165 * is restarted for another 50ms. 218 * reset_pollfunc() gets invoked to poll the interface for completion every 50ms 221 * for another 50ms. 332 * we set a timer to poll at 50ms intervals. 405 * recover from reset very quickly, saving us the first 50ms wait time.
|
/linux-4.1.27/sound/core/seq/ |
H A D | seq_midi.c | 280 struct seq_midisynth *msynth, *ms; snd_seq_midisynth_probe() local 343 ms = &msynth[p]; snd_seq_midisynth_probe() 345 if (snd_seq_midisynth_new(ms, card, device, p) < 0) snd_seq_midisynth_probe() 389 pcallbacks.private_data = ms; snd_seq_midisynth_probe() 400 ms->seq_client = client->seq_client; snd_seq_midisynth_probe() 401 ms->seq_port = port->addr.port; snd_seq_midisynth_probe()
|
/linux-4.1.27/drivers/iio/imu/ |
H A D | adis16400.h | 22 #define ADIS16400_STARTUP_DELAY 290 /* ms */ 23 #define ADIS16400_MTEST_DELAY 90 /* ms */
|
/linux-4.1.27/drivers/iio/pressure/ |
H A D | t5403.c | 27 #define T5403_MODE_SHIFT 3 /* conversion time: 2, 8, 16, 66 ms */ 53 int wait_time = 3; /* wakeup time in ms */ t5403_read()
|
/linux-4.1.27/drivers/net/wireless/ath/ath9k/ |
H A D | mci.c | 109 * Scale: wlan_airtime is in ms, aggr_limit is in 0.25 ms. ath_mci_adjust_aggr_limit() 110 * When wlan_airtime is less than 4ms, aggregation limit has to be ath_mci_adjust_aggr_limit() 153 "Single SCO, aggregation limit %d 1/4 ms\n", ath_mci_update_scheme() 163 "Single PAN/FTP bt period %d ms dutycycle %d\n", ath_mci_update_scheme() 170 "aggregation limit 1.5 ms dutycycle 30%%\n"); ath_mci_update_scheme() 177 "Two BT profiles aggr limit 1.5 ms dutycycle %d%%\n", ath_mci_update_scheme() 182 "Three or more profiles aggregation limit 1 ms\n"); ath_mci_update_scheme()
|
/linux-4.1.27/drivers/net/wireless/iwlwifi/dvm/ |
H A D | led.c | 42 /* Throughput OFF time(ms) ON time (ms)
|
/linux-4.1.27/drivers/net/wireless/iwlwifi/ |
H A D | iwl-fw-file.h | 624 * @command_queue: timeout for the command queue in ms 625 * @bss: timeout for the queues of a BSS (except for TDLS queues) in ms 626 * @softap: timeout for the queues of a softAP in ms 627 * @p2p_go: timeout for the queues of a P2P GO in ms 628 * @p2p_client: timeout for the queues of a P2P client in ms 629 * @p2p_device: timeout for the queues of a P2P device in ms 630 * @ibss: timeout for the queues of an IBSS in ms 631 * @tdls: timeout for the queues of a TDLS station in ms
|
/linux-4.1.27/drivers/net/wireless/prism54/ |
H A D | isl_38xx.h | 39 #define ISL38XX_RESET_DELAY 50 /* in ms */ 40 #define ISL38XX_WAIT_CYCLE 10 /* in 10ms */
|
/linux-4.1.27/drivers/net/wireless/rtl818x/rtl8180/ |
H A D | sa2400.c | 184 * we rely on the 3ms delay introduced in write_sa2400 */ sa2400_rf_init() 188 * we rely on the 3ms delay introduced in write_sa2400 */ sa2400_rf_init()
|
/linux-4.1.27/drivers/ssb/ |
H A D | driver_extif.c | 122 u32 ssb_extif_watchdog_timer_set_ms(struct bcm47xx_wdt *wdt, u32 ms) ssb_extif_watchdog_timer_set_ms() argument 125 u32 ticks = (SSB_EXTIF_WATCHDOG_CLK / 1000) * ms; ssb_extif_watchdog_timer_set_ms()
|
/linux-4.1.27/drivers/parport/ |
H A D | ieee1284.c | 145 * parport_wait_peripheral - wait for status lines to change in 35ms 151 * desired values, or until 35ms have elapsed (see IEEE 1284-1994 161 * zero, the fast polling time is 35ms, and this function does 166 * 10ms, waking up if an interrupt occurs. 181 entire 35ms. */ parport_wait_peripheral() 199 /* 40ms of slow polling. */ parport_wait_peripheral() 205 /* Wait for 10ms (or until an interrupt occurs if parport_wait_peripheral() 217 * Wait for another 10ms. */ parport_wait_peripheral()
|
/linux-4.1.27/arch/x86/kernel/ |
H A D | rtc.c | 36 * called 500 ms after the second nowtime has started, because when 38 * jump to the next second precisely 500 ms later. Check the Motorola
|
/linux-4.1.27/drivers/gpu/drm/msm/ |
H A D | msm_gpu.h | 113 #define DRM_MSM_INACTIVE_PERIOD 66 /* in ms (roughly four frames) */ 117 #define DRM_MSM_HANGCHECK_PERIOD 500 /* in ms */
|
/linux-4.1.27/drivers/usb/c67x00/ |
H A D | c67x00-hcd.h | 38 * 1ms deadline will be met. 50 * Bandwidth is described in bit time so with a 12MHz USB clock and 1ms
|
/linux-4.1.27/arch/mips/sgi-ip27/ |
H A D | ip27-init.c | 68 * Set CRB timeout at 5ms, (< PI timeout of 10ms) per_hub_init()
|
/linux-4.1.27/arch/powerpc/kernel/ |
H A D | cpu_setup_fsl_booke.S | 60 #define PW20_WAIT_IDLE_BIT 50 /* 1ms, TB frequency is 41.66MHZ */ 79 #define AV_WAIT_IDLE_BIT 50 /* 1ms, TB frequency is 41.66MHZ */
|
/linux-4.1.27/net/irda/ |
H A D | irsysctl.c | 55 /* IrLAP 6.13.2 says 25ms to 10+70ms - allow higher since some devices
|
H A D | qos.c | 108 static __u32 max_turn_times[] = { 500, 250, 100, 50 }; /* ms */ 112 /* 500 ms 250 ms 100 ms 50 ms (max turn time) */ 363 * Not allowed to use a max turn time less than 500 ms if the baudrate irlap_adjust_qos_settings() 369 pr_debug("%s(), adjusting max turn time from %d to 500 ms\n", irlap_adjust_qos_settings() 448 pr_debug("Setting MAX_TURN_TIME to %d ms.\n", irlap_qos_negotiate()
|