Lines Matching refs:scmrq
24 void __scm_free_rq_cluster(struct scm_request *scmrq) in __scm_free_rq_cluster() argument
28 if (!scmrq->cluster.buf) in __scm_free_rq_cluster()
32 free_page((unsigned long) scmrq->cluster.buf[i]); in __scm_free_rq_cluster()
34 kfree(scmrq->cluster.buf); in __scm_free_rq_cluster()
37 int __scm_alloc_rq_cluster(struct scm_request *scmrq) in __scm_alloc_rq_cluster() argument
41 scmrq->cluster.buf = kzalloc(sizeof(void *) * 2 * write_cluster_size, in __scm_alloc_rq_cluster()
43 if (!scmrq->cluster.buf) in __scm_alloc_rq_cluster()
47 scmrq->cluster.buf[i] = (void *) get_zeroed_page(GFP_DMA); in __scm_alloc_rq_cluster()
48 if (!scmrq->cluster.buf[i]) in __scm_alloc_rq_cluster()
51 INIT_LIST_HEAD(&scmrq->cluster.list); in __scm_alloc_rq_cluster()
55 void scm_request_cluster_init(struct scm_request *scmrq) in scm_request_cluster_init() argument
57 scmrq->cluster.state = CLUSTER_NONE; in scm_request_cluster_init()
75 bool scm_reserve_cluster(struct scm_request *scmrq) in scm_reserve_cluster() argument
77 struct request *req = scmrq->request[scmrq->aob->request.msb_count]; in scm_reserve_cluster()
78 struct scm_blk_dev *bdev = scmrq->bdev; in scm_reserve_cluster()
87 if (iter == scmrq) { in scm_reserve_cluster()
105 list_add(&scmrq->cluster.list, &bdev->cluster_list); in scm_reserve_cluster()
111 void scm_release_cluster(struct scm_request *scmrq) in scm_release_cluster() argument
113 struct scm_blk_dev *bdev = scmrq->bdev; in scm_release_cluster()
120 list_del(&scmrq->cluster.list); in scm_release_cluster()
130 static int scm_prepare_cluster_request(struct scm_request *scmrq) in scm_prepare_cluster_request() argument
132 struct scm_blk_dev *bdev = scmrq->bdev; in scm_prepare_cluster_request()
134 struct request *req = scmrq->request[0]; in scm_prepare_cluster_request()
135 struct msb *msb = &scmrq->aob->msb[0]; in scm_prepare_cluster_request()
142 switch (scmrq->cluster.state) { in scm_prepare_cluster_request()
144 scmrq->cluster.state = CLUSTER_READ; in scm_prepare_cluster_request()
160 aidaw = scm_aidaw_fetch(scmrq, msb->blk_count * PAGE_SIZE); in scm_prepare_cluster_request()
164 scmrq->aob->request.msb_count = 1; in scm_prepare_cluster_request()
167 aidaw->data_addr = (u64) scmrq->cluster.buf[i]; in scm_prepare_cluster_request()
179 aidaw->data_addr = (u64) scmrq->cluster.buf[i]; in scm_prepare_cluster_request()
189 aidaw->data_addr = (u64) scmrq->cluster.buf[i]; in scm_prepare_cluster_request()
197 bool scm_need_cluster_request(struct scm_request *scmrq) in scm_need_cluster_request() argument
199 int pos = scmrq->aob->request.msb_count; in scm_need_cluster_request()
201 if (rq_data_dir(scmrq->request[pos]) == READ) in scm_need_cluster_request()
204 return blk_rq_bytes(scmrq->request[pos]) < CLUSTER_SIZE; in scm_need_cluster_request()
208 void scm_initiate_cluster_request(struct scm_request *scmrq) in scm_initiate_cluster_request() argument
210 if (scm_prepare_cluster_request(scmrq)) in scm_initiate_cluster_request()
212 if (eadm_start_aob(scmrq->aob)) in scm_initiate_cluster_request()
216 scm_request_requeue(scmrq); in scm_initiate_cluster_request()
219 bool scm_test_cluster_request(struct scm_request *scmrq) in scm_test_cluster_request() argument
221 return scmrq->cluster.state != CLUSTER_NONE; in scm_test_cluster_request()
224 void scm_cluster_request_irq(struct scm_request *scmrq) in scm_cluster_request_irq() argument
226 struct scm_blk_dev *bdev = scmrq->bdev; in scm_cluster_request_irq()
229 switch (scmrq->cluster.state) { in scm_cluster_request_irq()
234 if (scmrq->error) { in scm_cluster_request_irq()
235 scm_request_finish(scmrq); in scm_cluster_request_irq()
238 scmrq->cluster.state = CLUSTER_WRITE; in scm_cluster_request_irq()
240 scm_initiate_cluster_request(scmrq); in scm_cluster_request_irq()
244 scm_request_finish(scmrq); in scm_cluster_request_irq()