1/* SPDX-License-Identifier: GPL-2.0-only */
2/*
3 * HighPoint RR3xxx/4xxx controller driver for Linux
4 * Copyright (C) 2006-2015 HighPoint Technologies, Inc. All Rights Reserved.
5 *
6 * Please report bugs/comments/suggestions to linux@highpoint-tech.com
7 *
8 * For more information, visit http://www.highpoint-tech.com
9 */
10#ifndef _HPTIOP_H_
11#define _HPTIOP_H_
12
13struct hpt_iopmu_itl {
14 __le32 resrved0[4];
15 __le32 inbound_msgaddr0;
16 __le32 inbound_msgaddr1;
17 __le32 outbound_msgaddr0;
18 __le32 outbound_msgaddr1;
19 __le32 inbound_doorbell;
20 __le32 inbound_intstatus;
21 __le32 inbound_intmask;
22 __le32 outbound_doorbell;
23 __le32 outbound_intstatus;
24 __le32 outbound_intmask;
25 __le32 reserved1[2];
26 __le32 inbound_queue;
27 __le32 outbound_queue;
28};
29
30#define IOPMU_QUEUE_EMPTY 0xffffffff
31#define IOPMU_QUEUE_MASK_HOST_BITS 0xf0000000
32#define IOPMU_QUEUE_ADDR_HOST_BIT 0x80000000
33#define IOPMU_QUEUE_REQUEST_SIZE_BIT 0x40000000
34#define IOPMU_QUEUE_REQUEST_RESULT_BIT 0x40000000
35
36#define IOPMU_OUTBOUND_INT_MSG0 1
37#define IOPMU_OUTBOUND_INT_MSG1 2
38#define IOPMU_OUTBOUND_INT_DOORBELL 4
39#define IOPMU_OUTBOUND_INT_POSTQUEUE 8
40#define IOPMU_OUTBOUND_INT_PCI 0x10
41
42#define IOPMU_INBOUND_INT_MSG0 1
43#define IOPMU_INBOUND_INT_MSG1 2
44#define IOPMU_INBOUND_INT_DOORBELL 4
45#define IOPMU_INBOUND_INT_ERROR 8
46#define IOPMU_INBOUND_INT_POSTQUEUE 0x10
47
48#define MVIOP_QUEUE_LEN 512
49
50struct hpt_iopmu_mv {
51 __le32 inbound_head;
52 __le32 inbound_tail;
53 __le32 outbound_head;
54 __le32 outbound_tail;
55 __le32 inbound_msg;
56 __le32 outbound_msg;
57 __le32 reserve[10];
58 __le64 inbound_q[MVIOP_QUEUE_LEN];
59 __le64 outbound_q[MVIOP_QUEUE_LEN];
60};
61
62struct hpt_iopmv_regs {
63 __le32 reserved[0x20400 / 4];
64 __le32 inbound_doorbell;
65 __le32 inbound_intmask;
66 __le32 outbound_doorbell;
67 __le32 outbound_intmask;
68};
69
70#pragma pack(1)
71struct hpt_iopmu_mvfrey {
72 __le32 reserved0[(0x4000 - 0) / 4];
73 __le32 inbound_base;
74 __le32 inbound_base_high;
75 __le32 reserved1[(0x4018 - 0x4008) / 4];
76 __le32 inbound_write_ptr;
77 __le32 reserved2[(0x402c - 0x401c) / 4];
78 __le32 inbound_conf_ctl;
79 __le32 reserved3[(0x4050 - 0x4030) / 4];
80 __le32 outbound_base;
81 __le32 outbound_base_high;
82 __le32 outbound_shadow_base;
83 __le32 outbound_shadow_base_high;
84 __le32 reserved4[(0x4088 - 0x4060) / 4];
85 __le32 isr_cause;
86 __le32 isr_enable;
87 __le32 reserved5[(0x1020c - 0x4090) / 4];
88 __le32 pcie_f0_int_enable;
89 __le32 reserved6[(0x10400 - 0x10210) / 4];
90 __le32 f0_to_cpu_msg_a;
91 __le32 reserved7[(0x10420 - 0x10404) / 4];
92 __le32 cpu_to_f0_msg_a;
93 __le32 reserved8[(0x10480 - 0x10424) / 4];
94 __le32 f0_doorbell;
95 __le32 f0_doorbell_enable;
96};
97
98struct mvfrey_inlist_entry {
99 dma_addr_t addr;
100 __le32 intrfc_len;
101 __le32 reserved;
102};
103
104struct mvfrey_outlist_entry {
105 __le32 val;
106};
107#pragma pack()
108
109#define MVIOP_MU_QUEUE_ADDR_HOST_MASK (~(0x1full))
110#define MVIOP_MU_QUEUE_ADDR_HOST_BIT 4
111
112#define MVIOP_MU_QUEUE_ADDR_IOP_HIGH32 0xffffffff
113#define MVIOP_MU_QUEUE_REQUEST_RESULT_BIT 1
114#define MVIOP_MU_QUEUE_REQUEST_RETURN_CONTEXT 2
115
116#define MVIOP_MU_INBOUND_INT_MSG 1
117#define MVIOP_MU_INBOUND_INT_POSTQUEUE 2
118#define MVIOP_MU_OUTBOUND_INT_MSG 1
119#define MVIOP_MU_OUTBOUND_INT_POSTQUEUE 2
120
121#define CL_POINTER_TOGGLE 0x00004000
122#define CPU_TO_F0_DRBL_MSG_BIT 0x02000000
123
124enum hpt_iopmu_message {
125 /* host-to-iop messages */
126 IOPMU_INBOUND_MSG0_NOP = 0,
127 IOPMU_INBOUND_MSG0_RESET,
128 IOPMU_INBOUND_MSG0_FLUSH,
129 IOPMU_INBOUND_MSG0_SHUTDOWN,
130 IOPMU_INBOUND_MSG0_STOP_BACKGROUND_TASK,
131 IOPMU_INBOUND_MSG0_START_BACKGROUND_TASK,
132 IOPMU_INBOUND_MSG0_RESET_COMM,
133 IOPMU_INBOUND_MSG0_MAX = 0xff,
134 /* iop-to-host messages */
135 IOPMU_OUTBOUND_MSG0_REGISTER_DEVICE_0 = 0x100,
136 IOPMU_OUTBOUND_MSG0_REGISTER_DEVICE_MAX = 0x1ff,
137 IOPMU_OUTBOUND_MSG0_UNREGISTER_DEVICE_0 = 0x200,
138 IOPMU_OUTBOUND_MSG0_UNREGISTER_DEVICE_MAX = 0x2ff,
139 IOPMU_OUTBOUND_MSG0_REVALIDATE_DEVICE_0 = 0x300,
140 IOPMU_OUTBOUND_MSG0_REVALIDATE_DEVICE_MAX = 0x3ff,
141};
142
143struct hpt_iop_request_header {
144 __le32 size;
145 __le32 type;
146 __le32 flags;
147 __le32 result;
148 __le32 context; /* host context */
149 __le32 context_hi32;
150};
151
152#define IOP_REQUEST_FLAG_SYNC_REQUEST 1
153#define IOP_REQUEST_FLAG_BIST_REQUEST 2
154#define IOP_REQUEST_FLAG_REMAPPED 4
155#define IOP_REQUEST_FLAG_OUTPUT_CONTEXT 8
156#define IOP_REQUEST_FLAG_ADDR_BITS 0x40 /* flags[31:16] is phy_addr[47:32] */
157
158enum hpt_iop_request_type {
159 IOP_REQUEST_TYPE_GET_CONFIG = 0,
160 IOP_REQUEST_TYPE_SET_CONFIG,
161 IOP_REQUEST_TYPE_BLOCK_COMMAND,
162 IOP_REQUEST_TYPE_SCSI_COMMAND,
163 IOP_REQUEST_TYPE_IOCTL_COMMAND,
164 IOP_REQUEST_TYPE_MAX
165};
166
167enum hpt_iop_result_type {
168 IOP_RESULT_PENDING = 0,
169 IOP_RESULT_SUCCESS,
170 IOP_RESULT_FAIL,
171 IOP_RESULT_BUSY,
172 IOP_RESULT_RESET,
173 IOP_RESULT_INVALID_REQUEST,
174 IOP_RESULT_BAD_TARGET,
175 IOP_RESULT_CHECK_CONDITION,
176};
177
178struct hpt_iop_request_get_config {
179 struct hpt_iop_request_header header;
180 __le32 interface_version;
181 __le32 firmware_version;
182 __le32 max_requests;
183 __le32 request_size;
184 __le32 max_sg_count;
185 __le32 data_transfer_length;
186 __le32 alignment_mask;
187 __le32 max_devices;
188 __le32 sdram_size;
189};
190
191struct hpt_iop_request_set_config {
192 struct hpt_iop_request_header header;
193 __le32 iop_id;
194 __le16 vbus_id;
195 __le16 max_host_request_size;
196 __le32 reserve[6];
197};
198
199struct hpt_iopsg {
200 __le32 size;
201 __le32 eot; /* non-zero: end of table */
202 __le64 pci_address;
203};
204
205struct hpt_iop_request_block_command {
206 struct hpt_iop_request_header header;
207 u8 channel;
208 u8 target;
209 u8 lun;
210 u8 pad1;
211 __le16 command; /* IOP_BLOCK_COMMAND_{READ,WRITE} */
212 __le16 sectors;
213 __le64 lba;
214 struct hpt_iopsg sg_list[1];
215};
216
217#define IOP_BLOCK_COMMAND_READ 1
218#define IOP_BLOCK_COMMAND_WRITE 2
219#define IOP_BLOCK_COMMAND_VERIFY 3
220#define IOP_BLOCK_COMMAND_FLUSH 4
221#define IOP_BLOCK_COMMAND_SHUTDOWN 5
222
223struct hpt_iop_request_scsi_command {
224 struct hpt_iop_request_header header;
225 u8 channel;
226 u8 target;
227 u8 lun;
228 u8 pad1;
229 u8 cdb[16];
230 __le32 dataxfer_length;
231 struct hpt_iopsg sg_list[];
232};
233
234struct hpt_iop_request_ioctl_command {
235 struct hpt_iop_request_header header;
236 __le32 ioctl_code;
237 __le32 inbuf_size;
238 __le32 outbuf_size;
239 __le32 bytes_returned;
240 u8 buf[];
241 /* out data should be put at buf[(inbuf_size+3)&~3] */
242};
243
244#define HPTIOP_MAX_REQUESTS 256u
245
246struct hptiop_request {
247 struct hptiop_request *next;
248 void *req_virt;
249 u32 req_shifted_phy;
250 struct scsi_cmnd *scp;
251 int index;
252};
253
254struct hpt_cmd_priv {
255 int mapped;
256 int sgcnt;
257 dma_addr_t dma_handle;
258};
259
260#define HPT_SCP(scp) ((struct hpt_cmd_priv *)scsi_cmd_priv(scp))
261
262enum hptiop_family {
263 UNKNOWN_BASED_IOP,
264 INTEL_BASED_IOP,
265 MV_BASED_IOP,
266 MVFREY_BASED_IOP
267} ;
268
269struct hptiop_hba {
270 struct hptiop_adapter_ops *ops;
271 union {
272 struct {
273 struct hpt_iopmu_itl __iomem *iop;
274 void __iomem *plx;
275 } itl;
276 struct {
277 struct hpt_iopmv_regs *regs;
278 struct hpt_iopmu_mv __iomem *mu;
279 void *internal_req;
280 dma_addr_t internal_req_phy;
281 } mv;
282 struct {
283 struct hpt_iop_request_get_config __iomem *config;
284 struct hpt_iopmu_mvfrey __iomem *mu;
285
286 int internal_mem_size;
287 struct hptiop_request internal_req;
288 int list_count;
289 struct mvfrey_inlist_entry *inlist;
290 dma_addr_t inlist_phy;
291 __le32 inlist_wptr;
292 struct mvfrey_outlist_entry *outlist;
293 dma_addr_t outlist_phy;
294 __le32 *outlist_cptr; /* copy pointer shadow */
295 dma_addr_t outlist_cptr_phy;
296 __le32 outlist_rptr;
297 } mvfrey;
298 } u;
299
300 struct Scsi_Host *host;
301 struct pci_dev *pcidev;
302
303 /* IOP config info */
304 u32 interface_version;
305 u32 firmware_version;
306 u32 sdram_size;
307 u32 max_devices;
308 u32 max_requests;
309 u32 max_request_size;
310 u32 max_sg_descriptors;
311
312 u32 req_size; /* host-allocated request buffer size */
313
314 u32 iopintf_v2: 1;
315 u32 initialized: 1;
316 u32 msg_done: 1;
317
318 struct hptiop_request * req_list;
319 struct hptiop_request reqs[HPTIOP_MAX_REQUESTS];
320
321 /* used to free allocated dma area */
322 void *dma_coherent[HPTIOP_MAX_REQUESTS];
323 dma_addr_t dma_coherent_handle[HPTIOP_MAX_REQUESTS];
324
325 atomic_t reset_count;
326 atomic_t resetting;
327
328 wait_queue_head_t reset_wq;
329 wait_queue_head_t ioctl_wq;
330};
331
332struct hpt_ioctl_k {
333 struct hptiop_hba * hba;
334 u32 ioctl_code;
335 u32 inbuf_size;
336 u32 outbuf_size;
337 void *inbuf;
338 void *outbuf;
339 u32 *bytes_returned;
340 void (*done)(struct hpt_ioctl_k *);
341 int result; /* HPT_IOCTL_RESULT_ */
342};
343
344struct hptiop_adapter_ops {
345 enum hptiop_family family;
346 int (*iop_wait_ready)(struct hptiop_hba *hba, u32 millisec);
347 int (*internal_memalloc)(struct hptiop_hba *hba);
348 int (*internal_memfree)(struct hptiop_hba *hba);
349 int (*map_pci_bar)(struct hptiop_hba *hba);
350 void (*unmap_pci_bar)(struct hptiop_hba *hba);
351 void (*enable_intr)(struct hptiop_hba *hba);
352 void (*disable_intr)(struct hptiop_hba *hba);
353 int (*get_config)(struct hptiop_hba *hba,
354 struct hpt_iop_request_get_config *config);
355 int (*set_config)(struct hptiop_hba *hba,
356 struct hpt_iop_request_set_config *config);
357 int (*iop_intr)(struct hptiop_hba *hba);
358 void (*post_msg)(struct hptiop_hba *hba, u32 msg);
359 void (*post_req)(struct hptiop_hba *hba, struct hptiop_request *_req);
360 int hw_dma_bit_mask;
361 int (*reset_comm)(struct hptiop_hba *hba);
362 __le64 host_phy_flag;
363};
364
365#define HPT_IOCTL_RESULT_OK 0
366#define HPT_IOCTL_RESULT_FAILED (-1)
367
368#if 0
369#define dprintk(fmt, args...) do { printk(fmt, ##args); } while(0)
370#else
371#define dprintk(fmt, args...)
372#endif
373
374#endif
375

source code of linux/drivers/scsi/hptiop.h