1
/****************************************************************************
2
* Driver for Solarflare Solarstorm network controllers and boards
3
* Copyright 2008-2011 Solarflare Communications Inc.
5
* This program is free software; you can redistribute it and/or modify it
6
* under the terms of the GNU General Public License version 2 as published
7
* by the Free Software Foundation, incorporated herein by reference.
10
#include <linux/delay.h>
11
#include "net_driver.h"
15
#include "mcdi_pcol.h"
18
/**************************************************************************
20
* Management-Controller-to-Driver Interface
22
**************************************************************************
25
/* Software-defined structure to the shared-memory */
26
#define CMD_NOTIFY_PORT0 0
27
#define CMD_NOTIFY_PORT1 4
28
#define CMD_PDU_PORT0 0x008
29
#define CMD_PDU_PORT1 0x108
30
#define REBOOT_FLAG_PORT0 0x3f8
31
#define REBOOT_FLAG_PORT1 0x3fc
33
#define MCDI_RPC_TIMEOUT 10 /*seconds */
35
#define MCDI_PDU(efx) \
36
(efx_port_num(efx) ? CMD_PDU_PORT1 : CMD_PDU_PORT0)
37
#define MCDI_DOORBELL(efx) \
38
(efx_port_num(efx) ? CMD_NOTIFY_PORT1 : CMD_NOTIFY_PORT0)
39
#define MCDI_REBOOT_FLAG(efx) \
40
(efx_port_num(efx) ? REBOOT_FLAG_PORT1 : REBOOT_FLAG_PORT0)
43
EFX_MASK32(EFX_WIDTH(MCDI_HEADER_SEQ))
45
static inline struct efx_mcdi_iface *efx_mcdi(struct efx_nic *efx)
47
struct siena_nic_data *nic_data;
48
EFX_BUG_ON_PARANOID(efx_nic_rev(efx) < EFX_REV_SIENA_A0);
49
nic_data = efx->nic_data;
50
return &nic_data->mcdi;
53
void efx_mcdi_init(struct efx_nic *efx)
55
struct efx_mcdi_iface *mcdi;
57
if (efx_nic_rev(efx) < EFX_REV_SIENA_A0)
61
init_waitqueue_head(&mcdi->wq);
62
spin_lock_init(&mcdi->iface_lock);
63
atomic_set(&mcdi->state, MCDI_STATE_QUIESCENT);
64
mcdi->mode = MCDI_MODE_POLL;
66
(void) efx_mcdi_poll_reboot(efx);
69
static void efx_mcdi_copyin(struct efx_nic *efx, unsigned cmd,
70
const u8 *inbuf, size_t inlen)
72
struct efx_mcdi_iface *mcdi = efx_mcdi(efx);
73
unsigned pdu = FR_CZ_MC_TREG_SMEM + MCDI_PDU(efx);
74
unsigned doorbell = FR_CZ_MC_TREG_SMEM + MCDI_DOORBELL(efx);
79
BUG_ON(atomic_read(&mcdi->state) == MCDI_STATE_QUIESCENT);
80
BUG_ON(inlen & 3 || inlen >= 0x100);
82
seqno = mcdi->seqno & SEQ_MASK;
84
if (mcdi->mode == MCDI_MODE_EVENTS)
85
xflags |= MCDI_HEADER_XFLAGS_EVREQ;
87
EFX_POPULATE_DWORD_6(hdr,
88
MCDI_HEADER_RESPONSE, 0,
89
MCDI_HEADER_RESYNC, 1,
90
MCDI_HEADER_CODE, cmd,
91
MCDI_HEADER_DATALEN, inlen,
92
MCDI_HEADER_SEQ, seqno,
93
MCDI_HEADER_XFLAGS, xflags);
95
efx_writed(efx, &hdr, pdu);
97
for (i = 0; i < inlen; i += 4)
98
_efx_writed(efx, *((__le32 *)(inbuf + i)), pdu + 4 + i);
100
/* Ensure the payload is written out before the header */
103
/* ring the doorbell with a distinctive value */
104
_efx_writed(efx, (__force __le32) 0x45789abc, doorbell);
107
static void efx_mcdi_copyout(struct efx_nic *efx, u8 *outbuf, size_t outlen)
109
struct efx_mcdi_iface *mcdi = efx_mcdi(efx);
110
unsigned int pdu = FR_CZ_MC_TREG_SMEM + MCDI_PDU(efx);
113
BUG_ON(atomic_read(&mcdi->state) == MCDI_STATE_QUIESCENT);
114
BUG_ON(outlen & 3 || outlen >= 0x100);
116
for (i = 0; i < outlen; i += 4)
117
*((__le32 *)(outbuf + i)) = _efx_readd(efx, pdu + 4 + i);
120
static int efx_mcdi_poll(struct efx_nic *efx)
122
struct efx_mcdi_iface *mcdi = efx_mcdi(efx);
123
unsigned int time, finish;
124
unsigned int respseq, respcmd, error;
125
unsigned int pdu = FR_CZ_MC_TREG_SMEM + MCDI_PDU(efx);
126
unsigned int rc, spins;
129
/* Check for a reboot atomically with respect to efx_mcdi_copyout() */
130
rc = -efx_mcdi_poll_reboot(efx);
134
/* Poll for completion. Poll quickly (once a us) for the 1st jiffy,
135
* because generally mcdi responses are fast. After that, back off
136
* and poll once a jiffy (approximately)
139
finish = get_seconds() + MCDI_RPC_TIMEOUT;
146
schedule_timeout_uninterruptible(1);
149
time = get_seconds();
152
efx_readd(efx, ®, pdu);
154
/* All 1's indicates that shared memory is in reset (and is
155
* not a valid header). Wait for it to come out reset before
156
* completing the command */
157
if (EFX_DWORD_FIELD(reg, EFX_DWORD_0) != 0xffffffff &&
158
EFX_DWORD_FIELD(reg, MCDI_HEADER_RESPONSE))
165
mcdi->resplen = EFX_DWORD_FIELD(reg, MCDI_HEADER_DATALEN);
166
respseq = EFX_DWORD_FIELD(reg, MCDI_HEADER_SEQ);
167
respcmd = EFX_DWORD_FIELD(reg, MCDI_HEADER_CODE);
168
error = EFX_DWORD_FIELD(reg, MCDI_HEADER_ERROR);
170
if (error && mcdi->resplen == 0) {
171
netif_err(efx, hw, efx->net_dev, "MC rebooted\n");
173
} else if ((respseq ^ mcdi->seqno) & SEQ_MASK) {
174
netif_err(efx, hw, efx->net_dev,
175
"MC response mismatch tx seq 0x%x rx seq 0x%x\n",
176
respseq, mcdi->seqno);
179
efx_readd(efx, ®, pdu + 4);
180
switch (EFX_DWORD_FIELD(reg, EFX_DWORD_0)) {
181
#define TRANSLATE_ERROR(name) \
182
case MC_CMD_ERR_ ## name: \
185
TRANSLATE_ERROR(ENOENT);
186
TRANSLATE_ERROR(EINTR);
187
TRANSLATE_ERROR(EACCES);
188
TRANSLATE_ERROR(EBUSY);
189
TRANSLATE_ERROR(EINVAL);
190
TRANSLATE_ERROR(EDEADLK);
191
TRANSLATE_ERROR(ENOSYS);
192
TRANSLATE_ERROR(ETIME);
193
#undef TRANSLATE_ERROR
206
/* Return rc=0 like wait_event_timeout() */
210
/* Test and clear MC-rebooted flag for this port/function */
211
int efx_mcdi_poll_reboot(struct efx_nic *efx)
213
unsigned int addr = FR_CZ_MC_TREG_SMEM + MCDI_REBOOT_FLAG(efx);
217
if (efx_nic_rev(efx) < EFX_REV_SIENA_A0)
220
efx_readd(efx, ®, addr);
221
value = EFX_DWORD_FIELD(reg, EFX_DWORD_0);
227
efx_writed(efx, ®, addr);
229
if (value == MC_STATUS_DWORD_ASSERT)
235
static void efx_mcdi_acquire(struct efx_mcdi_iface *mcdi)
237
/* Wait until the interface becomes QUIESCENT and we win the race
238
* to mark it RUNNING. */
240
atomic_cmpxchg(&mcdi->state,
241
MCDI_STATE_QUIESCENT,
243
== MCDI_STATE_QUIESCENT);
246
static int efx_mcdi_await_completion(struct efx_nic *efx)
248
struct efx_mcdi_iface *mcdi = efx_mcdi(efx);
250
if (wait_event_timeout(
252
atomic_read(&mcdi->state) == MCDI_STATE_COMPLETED,
253
msecs_to_jiffies(MCDI_RPC_TIMEOUT * 1000)) == 0)
256
/* Check if efx_mcdi_set_mode() switched us back to polled completions.
257
* In which case, poll for completions directly. If efx_mcdi_ev_cpl()
258
* completed the request first, then we'll just end up completing the
259
* request again, which is safe.
261
* We need an smp_rmb() to synchronise with efx_mcdi_mode_poll(), which
262
* wait_event_timeout() implicitly provides.
264
if (mcdi->mode == MCDI_MODE_POLL)
265
return efx_mcdi_poll(efx);
270
static bool efx_mcdi_complete(struct efx_mcdi_iface *mcdi)
272
/* If the interface is RUNNING, then move to COMPLETED and wake any
273
* waiters. If the interface isn't in RUNNING then we've received a
274
* duplicate completion after we've already transitioned back to
275
* QUIESCENT. [A subsequent invocation would increment seqno, so would
276
* have failed the seqno check].
278
if (atomic_cmpxchg(&mcdi->state,
280
MCDI_STATE_COMPLETED) == MCDI_STATE_RUNNING) {
288
static void efx_mcdi_release(struct efx_mcdi_iface *mcdi)
290
atomic_set(&mcdi->state, MCDI_STATE_QUIESCENT);
294
static void efx_mcdi_ev_cpl(struct efx_nic *efx, unsigned int seqno,
295
unsigned int datalen, unsigned int errno)
297
struct efx_mcdi_iface *mcdi = efx_mcdi(efx);
300
spin_lock(&mcdi->iface_lock);
302
if ((seqno ^ mcdi->seqno) & SEQ_MASK) {
304
/* The request has been cancelled */
307
netif_err(efx, hw, efx->net_dev,
308
"MC response mismatch tx seq 0x%x rx "
309
"seq 0x%x\n", seqno, mcdi->seqno);
311
mcdi->resprc = errno;
312
mcdi->resplen = datalen;
317
spin_unlock(&mcdi->iface_lock);
320
efx_mcdi_complete(mcdi);
323
/* Issue the given command by writing the data into the shared memory PDU,
324
* ring the doorbell and wait for completion. Copyout the result. */
325
int efx_mcdi_rpc(struct efx_nic *efx, unsigned cmd,
326
const u8 *inbuf, size_t inlen, u8 *outbuf, size_t outlen,
327
size_t *outlen_actual)
329
struct efx_mcdi_iface *mcdi = efx_mcdi(efx);
331
BUG_ON(efx_nic_rev(efx) < EFX_REV_SIENA_A0);
333
efx_mcdi_acquire(mcdi);
335
/* Serialise with efx_mcdi_ev_cpl() and efx_mcdi_ev_death() */
336
spin_lock_bh(&mcdi->iface_lock);
338
spin_unlock_bh(&mcdi->iface_lock);
340
efx_mcdi_copyin(efx, cmd, inbuf, inlen);
342
if (mcdi->mode == MCDI_MODE_POLL)
343
rc = efx_mcdi_poll(efx);
345
rc = efx_mcdi_await_completion(efx);
348
/* Close the race with efx_mcdi_ev_cpl() executing just too late
349
* and completing a request we've just cancelled, by ensuring
350
* that the seqno check therein fails.
352
spin_lock_bh(&mcdi->iface_lock);
355
spin_unlock_bh(&mcdi->iface_lock);
357
netif_err(efx, hw, efx->net_dev,
358
"MC command 0x%x inlen %d mode %d timed out\n",
359
cmd, (int)inlen, mcdi->mode);
363
/* At the very least we need a memory barrier here to ensure
364
* we pick up changes from efx_mcdi_ev_cpl(). Protect against
365
* a spurious efx_mcdi_ev_cpl() running concurrently by
366
* acquiring the iface_lock. */
367
spin_lock_bh(&mcdi->iface_lock);
369
resplen = mcdi->resplen;
370
spin_unlock_bh(&mcdi->iface_lock);
373
efx_mcdi_copyout(efx, outbuf,
374
min(outlen, mcdi->resplen + 3) & ~0x3);
375
if (outlen_actual != NULL)
376
*outlen_actual = resplen;
377
} else if (cmd == MC_CMD_REBOOT && rc == -EIO)
378
; /* Don't reset if MC_CMD_REBOOT returns EIO */
379
else if (rc == -EIO || rc == -EINTR) {
380
netif_err(efx, hw, efx->net_dev, "MC fatal error %d\n",
382
efx_schedule_reset(efx, RESET_TYPE_MC_FAILURE);
384
netif_dbg(efx, hw, efx->net_dev,
385
"MC command 0x%x inlen %d failed rc=%d\n",
386
cmd, (int)inlen, -rc);
389
efx_mcdi_release(mcdi);
393
void efx_mcdi_mode_poll(struct efx_nic *efx)
395
struct efx_mcdi_iface *mcdi;
397
if (efx_nic_rev(efx) < EFX_REV_SIENA_A0)
400
mcdi = efx_mcdi(efx);
401
if (mcdi->mode == MCDI_MODE_POLL)
404
/* We can switch from event completion to polled completion, because
405
* mcdi requests are always completed in shared memory. We do this by
406
* switching the mode to POLL'd then completing the request.
407
* efx_mcdi_await_completion() will then call efx_mcdi_poll().
409
* We need an smp_wmb() to synchronise with efx_mcdi_await_completion(),
410
* which efx_mcdi_complete() provides for us.
412
mcdi->mode = MCDI_MODE_POLL;
414
efx_mcdi_complete(mcdi);
417
void efx_mcdi_mode_event(struct efx_nic *efx)
419
struct efx_mcdi_iface *mcdi;
421
if (efx_nic_rev(efx) < EFX_REV_SIENA_A0)
424
mcdi = efx_mcdi(efx);
426
if (mcdi->mode == MCDI_MODE_EVENTS)
429
/* We can't switch from polled to event completion in the middle of a
430
* request, because the completion method is specified in the request.
431
* So acquire the interface to serialise the requestors. We don't need
432
* to acquire the iface_lock to change the mode here, but we do need a
433
* write memory barrier ensure that efx_mcdi_rpc() sees it, which
434
* efx_mcdi_acquire() provides.
436
efx_mcdi_acquire(mcdi);
437
mcdi->mode = MCDI_MODE_EVENTS;
438
efx_mcdi_release(mcdi);
441
static void efx_mcdi_ev_death(struct efx_nic *efx, int rc)
443
struct efx_mcdi_iface *mcdi = efx_mcdi(efx);
445
/* If there is an outstanding MCDI request, it has been terminated
446
* either by a BADASSERT or REBOOT event. If the mcdi interface is
447
* in polled mode, then do nothing because the MC reboot handler will
448
* set the header correctly. However, if the mcdi interface is waiting
449
* for a CMDDONE event it won't receive it [and since all MCDI events
450
* are sent to the same queue, we can't be racing with
453
* There's a race here with efx_mcdi_rpc(), because we might receive
454
* a REBOOT event *before* the request has been copied out. In polled
455
* mode (during startup) this is irrelevant, because efx_mcdi_complete()
456
* is ignored. In event mode, this condition is just an edge-case of
457
* receiving a REBOOT event after posting the MCDI request. Did the mc
458
* reboot before or after the copyout? The best we can do always is
459
* just return failure.
461
spin_lock(&mcdi->iface_lock);
462
if (efx_mcdi_complete(mcdi)) {
463
if (mcdi->mode == MCDI_MODE_EVENTS) {
469
/* Nobody was waiting for an MCDI request, so trigger a reset */
470
efx_schedule_reset(efx, RESET_TYPE_MC_FAILURE);
472
spin_unlock(&mcdi->iface_lock);
475
static unsigned int efx_mcdi_event_link_speed[] = {
476
[MCDI_EVENT_LINKCHANGE_SPEED_100M] = 100,
477
[MCDI_EVENT_LINKCHANGE_SPEED_1G] = 1000,
478
[MCDI_EVENT_LINKCHANGE_SPEED_10G] = 10000,
482
static void efx_mcdi_process_link_change(struct efx_nic *efx, efx_qword_t *ev)
484
u32 flags, fcntl, speed, lpa;
486
speed = EFX_QWORD_FIELD(*ev, MCDI_EVENT_LINKCHANGE_SPEED);
487
EFX_BUG_ON_PARANOID(speed >= ARRAY_SIZE(efx_mcdi_event_link_speed));
488
speed = efx_mcdi_event_link_speed[speed];
490
flags = EFX_QWORD_FIELD(*ev, MCDI_EVENT_LINKCHANGE_LINK_FLAGS);
491
fcntl = EFX_QWORD_FIELD(*ev, MCDI_EVENT_LINKCHANGE_FCNTL);
492
lpa = EFX_QWORD_FIELD(*ev, MCDI_EVENT_LINKCHANGE_LP_CAP);
494
/* efx->link_state is only modified by efx_mcdi_phy_get_link(),
495
* which is only run after flushing the event queues. Therefore, it
496
* is safe to modify the link state outside of the mac_lock here.
498
efx_mcdi_phy_decode_link(efx, &efx->link_state, speed, flags, fcntl);
500
efx_mcdi_phy_check_fcntl(efx, lpa);
502
efx_link_status_changed(efx);
505
static const char *sensor_names[] = {
506
[MC_CMD_SENSOR_CONTROLLER_TEMP] = "Controller temp. sensor",
507
[MC_CMD_SENSOR_PHY_COMMON_TEMP] = "PHY shared temp. sensor",
508
[MC_CMD_SENSOR_CONTROLLER_COOLING] = "Controller cooling",
509
[MC_CMD_SENSOR_PHY0_TEMP] = "PHY 0 temp. sensor",
510
[MC_CMD_SENSOR_PHY0_COOLING] = "PHY 0 cooling",
511
[MC_CMD_SENSOR_PHY1_TEMP] = "PHY 1 temp. sensor",
512
[MC_CMD_SENSOR_PHY1_COOLING] = "PHY 1 cooling",
513
[MC_CMD_SENSOR_IN_1V0] = "1.0V supply sensor",
514
[MC_CMD_SENSOR_IN_1V2] = "1.2V supply sensor",
515
[MC_CMD_SENSOR_IN_1V8] = "1.8V supply sensor",
516
[MC_CMD_SENSOR_IN_2V5] = "2.5V supply sensor",
517
[MC_CMD_SENSOR_IN_3V3] = "3.3V supply sensor",
518
[MC_CMD_SENSOR_IN_12V0] = "12V supply sensor"
521
static const char *sensor_status_names[] = {
522
[MC_CMD_SENSOR_STATE_OK] = "OK",
523
[MC_CMD_SENSOR_STATE_WARNING] = "Warning",
524
[MC_CMD_SENSOR_STATE_FATAL] = "Fatal",
525
[MC_CMD_SENSOR_STATE_BROKEN] = "Device failure",
528
static void efx_mcdi_sensor_event(struct efx_nic *efx, efx_qword_t *ev)
530
unsigned int monitor, state, value;
531
const char *name, *state_txt;
532
monitor = EFX_QWORD_FIELD(*ev, MCDI_EVENT_SENSOREVT_MONITOR);
533
state = EFX_QWORD_FIELD(*ev, MCDI_EVENT_SENSOREVT_STATE);
534
value = EFX_QWORD_FIELD(*ev, MCDI_EVENT_SENSOREVT_VALUE);
535
/* Deal gracefully with the board having more drivers than we
536
* know about, but do not expect new sensor states. */
537
name = (monitor >= ARRAY_SIZE(sensor_names))
538
? "No sensor name available" :
539
sensor_names[monitor];
540
EFX_BUG_ON_PARANOID(state >= ARRAY_SIZE(sensor_status_names));
541
state_txt = sensor_status_names[state];
543
netif_err(efx, hw, efx->net_dev,
544
"Sensor %d (%s) reports condition '%s' for raw value %d\n",
545
monitor, name, state_txt, value);
548
/* Called from falcon_process_eventq for MCDI events */
549
void efx_mcdi_process_event(struct efx_channel *channel,
552
struct efx_nic *efx = channel->efx;
553
int code = EFX_QWORD_FIELD(*event, MCDI_EVENT_CODE);
554
u32 data = EFX_QWORD_FIELD(*event, MCDI_EVENT_DATA);
557
case MCDI_EVENT_CODE_BADSSERT:
558
netif_err(efx, hw, efx->net_dev,
559
"MC watchdog or assertion failure at 0x%x\n", data);
560
efx_mcdi_ev_death(efx, EINTR);
563
case MCDI_EVENT_CODE_PMNOTICE:
564
netif_info(efx, wol, efx->net_dev, "MCDI PM event.\n");
567
case MCDI_EVENT_CODE_CMDDONE:
569
MCDI_EVENT_FIELD(*event, CMDDONE_SEQ),
570
MCDI_EVENT_FIELD(*event, CMDDONE_DATALEN),
571
MCDI_EVENT_FIELD(*event, CMDDONE_ERRNO));
574
case MCDI_EVENT_CODE_LINKCHANGE:
575
efx_mcdi_process_link_change(efx, event);
577
case MCDI_EVENT_CODE_SENSOREVT:
578
efx_mcdi_sensor_event(efx, event);
580
case MCDI_EVENT_CODE_SCHEDERR:
581
netif_info(efx, hw, efx->net_dev,
582
"MC Scheduler error address=0x%x\n", data);
584
case MCDI_EVENT_CODE_REBOOT:
585
netif_info(efx, hw, efx->net_dev, "MC Reboot\n");
586
efx_mcdi_ev_death(efx, EIO);
588
case MCDI_EVENT_CODE_MAC_STATS_DMA:
589
/* MAC stats are gather lazily. We can ignore this. */
593
netif_err(efx, hw, efx->net_dev, "Unknown MCDI event 0x%x\n",
598
/**************************************************************************
600
* Specific request functions
602
**************************************************************************
605
void efx_mcdi_print_fwver(struct efx_nic *efx, char *buf, size_t len)
607
u8 outbuf[ALIGN(MC_CMD_GET_VERSION_V1_OUT_LEN, 4)];
609
const __le16 *ver_words;
612
BUILD_BUG_ON(MC_CMD_GET_VERSION_IN_LEN != 0);
614
rc = efx_mcdi_rpc(efx, MC_CMD_GET_VERSION, NULL, 0,
615
outbuf, sizeof(outbuf), &outlength);
619
if (outlength < MC_CMD_GET_VERSION_V1_OUT_LEN) {
624
ver_words = (__le16 *)MCDI_PTR(outbuf, GET_VERSION_OUT_VERSION);
625
snprintf(buf, len, "%u.%u.%u.%u",
626
le16_to_cpu(ver_words[0]), le16_to_cpu(ver_words[1]),
627
le16_to_cpu(ver_words[2]), le16_to_cpu(ver_words[3]));
631
netif_err(efx, probe, efx->net_dev, "%s: failed rc=%d\n", __func__, rc);
635
int efx_mcdi_drv_attach(struct efx_nic *efx, bool driver_operating,
638
u8 inbuf[MC_CMD_DRV_ATTACH_IN_LEN];
639
u8 outbuf[MC_CMD_DRV_ATTACH_OUT_LEN];
643
MCDI_SET_DWORD(inbuf, DRV_ATTACH_IN_NEW_STATE,
644
driver_operating ? 1 : 0);
645
MCDI_SET_DWORD(inbuf, DRV_ATTACH_IN_UPDATE, 1);
647
rc = efx_mcdi_rpc(efx, MC_CMD_DRV_ATTACH, inbuf, sizeof(inbuf),
648
outbuf, sizeof(outbuf), &outlen);
651
if (outlen < MC_CMD_DRV_ATTACH_OUT_LEN) {
656
if (was_attached != NULL)
657
*was_attached = MCDI_DWORD(outbuf, DRV_ATTACH_OUT_OLD_STATE);
661
netif_err(efx, probe, efx->net_dev, "%s: failed rc=%d\n", __func__, rc);
665
int efx_mcdi_get_board_cfg(struct efx_nic *efx, u8 *mac_address,
666
u16 *fw_subtype_list)
668
uint8_t outbuf[MC_CMD_GET_BOARD_CFG_OUT_LEN];
670
int port_num = efx_port_num(efx);
674
BUILD_BUG_ON(MC_CMD_GET_BOARD_CFG_IN_LEN != 0);
676
rc = efx_mcdi_rpc(efx, MC_CMD_GET_BOARD_CFG, NULL, 0,
677
outbuf, sizeof(outbuf), &outlen);
681
if (outlen < MC_CMD_GET_BOARD_CFG_OUT_LEN) {
687
? MC_CMD_GET_BOARD_CFG_OUT_MAC_ADDR_BASE_PORT1_OFST
688
: MC_CMD_GET_BOARD_CFG_OUT_MAC_ADDR_BASE_PORT0_OFST;
690
memcpy(mac_address, outbuf + offset, ETH_ALEN);
692
memcpy(fw_subtype_list,
693
outbuf + MC_CMD_GET_BOARD_CFG_OUT_FW_SUBTYPE_LIST_OFST,
694
MC_CMD_GET_BOARD_CFG_OUT_FW_SUBTYPE_LIST_LEN);
699
netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d len=%d\n",
700
__func__, rc, (int)outlen);
705
int efx_mcdi_log_ctrl(struct efx_nic *efx, bool evq, bool uart, u32 dest_evq)
707
u8 inbuf[MC_CMD_LOG_CTRL_IN_LEN];
712
dest |= MC_CMD_LOG_CTRL_IN_LOG_DEST_UART;
714
dest |= MC_CMD_LOG_CTRL_IN_LOG_DEST_EVQ;
716
MCDI_SET_DWORD(inbuf, LOG_CTRL_IN_LOG_DEST, dest);
717
MCDI_SET_DWORD(inbuf, LOG_CTRL_IN_LOG_DEST_EVQ, dest_evq);
719
BUILD_BUG_ON(MC_CMD_LOG_CTRL_OUT_LEN != 0);
721
rc = efx_mcdi_rpc(efx, MC_CMD_LOG_CTRL, inbuf, sizeof(inbuf),
729
netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d\n", __func__, rc);
733
int efx_mcdi_nvram_types(struct efx_nic *efx, u32 *nvram_types_out)
735
u8 outbuf[MC_CMD_NVRAM_TYPES_OUT_LEN];
739
BUILD_BUG_ON(MC_CMD_NVRAM_TYPES_IN_LEN != 0);
741
rc = efx_mcdi_rpc(efx, MC_CMD_NVRAM_TYPES, NULL, 0,
742
outbuf, sizeof(outbuf), &outlen);
745
if (outlen < MC_CMD_NVRAM_TYPES_OUT_LEN) {
750
*nvram_types_out = MCDI_DWORD(outbuf, NVRAM_TYPES_OUT_TYPES);
754
netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d\n",
759
int efx_mcdi_nvram_info(struct efx_nic *efx, unsigned int type,
760
size_t *size_out, size_t *erase_size_out,
763
u8 inbuf[MC_CMD_NVRAM_INFO_IN_LEN];
764
u8 outbuf[MC_CMD_NVRAM_INFO_OUT_LEN];
768
MCDI_SET_DWORD(inbuf, NVRAM_INFO_IN_TYPE, type);
770
rc = efx_mcdi_rpc(efx, MC_CMD_NVRAM_INFO, inbuf, sizeof(inbuf),
771
outbuf, sizeof(outbuf), &outlen);
774
if (outlen < MC_CMD_NVRAM_INFO_OUT_LEN) {
779
*size_out = MCDI_DWORD(outbuf, NVRAM_INFO_OUT_SIZE);
780
*erase_size_out = MCDI_DWORD(outbuf, NVRAM_INFO_OUT_ERASESIZE);
781
*protected_out = !!(MCDI_DWORD(outbuf, NVRAM_INFO_OUT_FLAGS) &
782
(1 << MC_CMD_NVRAM_PROTECTED_LBN));
786
netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d\n", __func__, rc);
790
int efx_mcdi_nvram_update_start(struct efx_nic *efx, unsigned int type)
792
u8 inbuf[MC_CMD_NVRAM_UPDATE_START_IN_LEN];
795
MCDI_SET_DWORD(inbuf, NVRAM_UPDATE_START_IN_TYPE, type);
797
BUILD_BUG_ON(MC_CMD_NVRAM_UPDATE_START_OUT_LEN != 0);
799
rc = efx_mcdi_rpc(efx, MC_CMD_NVRAM_UPDATE_START, inbuf, sizeof(inbuf),
807
netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d\n", __func__, rc);
811
int efx_mcdi_nvram_read(struct efx_nic *efx, unsigned int type,
812
loff_t offset, u8 *buffer, size_t length)
814
u8 inbuf[MC_CMD_NVRAM_READ_IN_LEN];
815
u8 outbuf[MC_CMD_NVRAM_READ_OUT_LEN(EFX_MCDI_NVRAM_LEN_MAX)];
819
MCDI_SET_DWORD(inbuf, NVRAM_READ_IN_TYPE, type);
820
MCDI_SET_DWORD(inbuf, NVRAM_READ_IN_OFFSET, offset);
821
MCDI_SET_DWORD(inbuf, NVRAM_READ_IN_LENGTH, length);
823
rc = efx_mcdi_rpc(efx, MC_CMD_NVRAM_READ, inbuf, sizeof(inbuf),
824
outbuf, sizeof(outbuf), &outlen);
828
memcpy(buffer, MCDI_PTR(outbuf, NVRAM_READ_OUT_READ_BUFFER), length);
832
netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d\n", __func__, rc);
836
int efx_mcdi_nvram_write(struct efx_nic *efx, unsigned int type,
837
loff_t offset, const u8 *buffer, size_t length)
839
u8 inbuf[MC_CMD_NVRAM_WRITE_IN_LEN(EFX_MCDI_NVRAM_LEN_MAX)];
842
MCDI_SET_DWORD(inbuf, NVRAM_WRITE_IN_TYPE, type);
843
MCDI_SET_DWORD(inbuf, NVRAM_WRITE_IN_OFFSET, offset);
844
MCDI_SET_DWORD(inbuf, NVRAM_WRITE_IN_LENGTH, length);
845
memcpy(MCDI_PTR(inbuf, NVRAM_WRITE_IN_WRITE_BUFFER), buffer, length);
847
BUILD_BUG_ON(MC_CMD_NVRAM_WRITE_OUT_LEN != 0);
849
rc = efx_mcdi_rpc(efx, MC_CMD_NVRAM_WRITE, inbuf,
850
ALIGN(MC_CMD_NVRAM_WRITE_IN_LEN(length), 4),
858
netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d\n", __func__, rc);
862
int efx_mcdi_nvram_erase(struct efx_nic *efx, unsigned int type,
863
loff_t offset, size_t length)
865
u8 inbuf[MC_CMD_NVRAM_ERASE_IN_LEN];
868
MCDI_SET_DWORD(inbuf, NVRAM_ERASE_IN_TYPE, type);
869
MCDI_SET_DWORD(inbuf, NVRAM_ERASE_IN_OFFSET, offset);
870
MCDI_SET_DWORD(inbuf, NVRAM_ERASE_IN_LENGTH, length);
872
BUILD_BUG_ON(MC_CMD_NVRAM_ERASE_OUT_LEN != 0);
874
rc = efx_mcdi_rpc(efx, MC_CMD_NVRAM_ERASE, inbuf, sizeof(inbuf),
882
netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d\n", __func__, rc);
886
int efx_mcdi_nvram_update_finish(struct efx_nic *efx, unsigned int type)
888
u8 inbuf[MC_CMD_NVRAM_UPDATE_FINISH_IN_LEN];
891
MCDI_SET_DWORD(inbuf, NVRAM_UPDATE_FINISH_IN_TYPE, type);
893
BUILD_BUG_ON(MC_CMD_NVRAM_UPDATE_FINISH_OUT_LEN != 0);
895
rc = efx_mcdi_rpc(efx, MC_CMD_NVRAM_UPDATE_FINISH, inbuf, sizeof(inbuf),
903
netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d\n", __func__, rc);
907
static int efx_mcdi_nvram_test(struct efx_nic *efx, unsigned int type)
909
u8 inbuf[MC_CMD_NVRAM_TEST_IN_LEN];
910
u8 outbuf[MC_CMD_NVRAM_TEST_OUT_LEN];
913
MCDI_SET_DWORD(inbuf, NVRAM_TEST_IN_TYPE, type);
915
rc = efx_mcdi_rpc(efx, MC_CMD_NVRAM_TEST, inbuf, sizeof(inbuf),
916
outbuf, sizeof(outbuf), NULL);
920
switch (MCDI_DWORD(outbuf, NVRAM_TEST_OUT_RESULT)) {
921
case MC_CMD_NVRAM_TEST_PASS:
922
case MC_CMD_NVRAM_TEST_NOTSUPP:
929
int efx_mcdi_nvram_test_all(struct efx_nic *efx)
935
rc = efx_mcdi_nvram_types(efx, &nvram_types);
940
while (nvram_types != 0) {
941
if (nvram_types & 1) {
942
rc = efx_mcdi_nvram_test(efx, type);
953
netif_err(efx, hw, efx->net_dev, "%s: failed type=%u\n",
956
netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d\n", __func__, rc);
960
static int efx_mcdi_read_assertion(struct efx_nic *efx)
962
u8 inbuf[MC_CMD_GET_ASSERTS_IN_LEN];
963
u8 outbuf[MC_CMD_GET_ASSERTS_OUT_LEN];
964
unsigned int flags, index, ofst;
970
/* Attempt to read any stored assertion state before we reboot
971
* the mcfw out of the assertion handler. Retry twice, once
972
* because a boot-time assertion might cause this command to fail
973
* with EINTR. And once again because GET_ASSERTS can race with
974
* MC_CMD_REBOOT running on the other port. */
977
MCDI_SET_DWORD(inbuf, GET_ASSERTS_IN_CLEAR, 1);
978
rc = efx_mcdi_rpc(efx, MC_CMD_GET_ASSERTS,
979
inbuf, MC_CMD_GET_ASSERTS_IN_LEN,
980
outbuf, sizeof(outbuf), &outlen);
981
} while ((rc == -EINTR || rc == -EIO) && retry-- > 0);
985
if (outlen < MC_CMD_GET_ASSERTS_OUT_LEN)
988
/* Print out any recorded assertion state */
989
flags = MCDI_DWORD(outbuf, GET_ASSERTS_OUT_GLOBAL_FLAGS);
990
if (flags == MC_CMD_GET_ASSERTS_FLAGS_NO_FAILS)
993
reason = (flags == MC_CMD_GET_ASSERTS_FLAGS_SYS_FAIL)
994
? "system-level assertion"
995
: (flags == MC_CMD_GET_ASSERTS_FLAGS_THR_FAIL)
996
? "thread-level assertion"
997
: (flags == MC_CMD_GET_ASSERTS_FLAGS_WDOG_FIRED)
999
: "unknown assertion";
1000
netif_err(efx, hw, efx->net_dev,
1001
"MCPU %s at PC = 0x%.8x in thread 0x%.8x\n", reason,
1002
MCDI_DWORD(outbuf, GET_ASSERTS_OUT_SAVED_PC_OFFS),
1003
MCDI_DWORD(outbuf, GET_ASSERTS_OUT_THREAD_OFFS));
1005
/* Print out the registers */
1006
ofst = MC_CMD_GET_ASSERTS_OUT_GP_REGS_OFFS_OFST;
1007
for (index = 1; index < 32; index++) {
1008
netif_err(efx, hw, efx->net_dev, "R%.2d (?): 0x%.8x\n", index,
1009
MCDI_DWORD2(outbuf, ofst));
1010
ofst += sizeof(efx_dword_t);
1016
static void efx_mcdi_exit_assertion(struct efx_nic *efx)
1018
u8 inbuf[MC_CMD_REBOOT_IN_LEN];
1020
/* Atomically reboot the mcfw out of the assertion handler */
1021
BUILD_BUG_ON(MC_CMD_REBOOT_OUT_LEN != 0);
1022
MCDI_SET_DWORD(inbuf, REBOOT_IN_FLAGS,
1023
MC_CMD_REBOOT_FLAGS_AFTER_ASSERTION);
1024
efx_mcdi_rpc(efx, MC_CMD_REBOOT, inbuf, MC_CMD_REBOOT_IN_LEN,
1028
int efx_mcdi_handle_assertion(struct efx_nic *efx)
1032
rc = efx_mcdi_read_assertion(efx);
1036
efx_mcdi_exit_assertion(efx);
1041
void efx_mcdi_set_id_led(struct efx_nic *efx, enum efx_led_mode mode)
1043
u8 inbuf[MC_CMD_SET_ID_LED_IN_LEN];
1046
BUILD_BUG_ON(EFX_LED_OFF != MC_CMD_LED_OFF);
1047
BUILD_BUG_ON(EFX_LED_ON != MC_CMD_LED_ON);
1048
BUILD_BUG_ON(EFX_LED_DEFAULT != MC_CMD_LED_DEFAULT);
1050
BUILD_BUG_ON(MC_CMD_SET_ID_LED_OUT_LEN != 0);
1052
MCDI_SET_DWORD(inbuf, SET_ID_LED_IN_STATE, mode);
1054
rc = efx_mcdi_rpc(efx, MC_CMD_SET_ID_LED, inbuf, sizeof(inbuf),
1057
netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d\n",
1061
int efx_mcdi_reset_port(struct efx_nic *efx)
1063
int rc = efx_mcdi_rpc(efx, MC_CMD_PORT_RESET, NULL, 0, NULL, 0, NULL);
1065
netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d\n",
1070
int efx_mcdi_reset_mc(struct efx_nic *efx)
1072
u8 inbuf[MC_CMD_REBOOT_IN_LEN];
1075
BUILD_BUG_ON(MC_CMD_REBOOT_OUT_LEN != 0);
1076
MCDI_SET_DWORD(inbuf, REBOOT_IN_FLAGS, 0);
1077
rc = efx_mcdi_rpc(efx, MC_CMD_REBOOT, inbuf, sizeof(inbuf),
1079
/* White is black, and up is down */
1084
netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d\n", __func__, rc);
1088
static int efx_mcdi_wol_filter_set(struct efx_nic *efx, u32 type,
1089
const u8 *mac, int *id_out)
1091
u8 inbuf[MC_CMD_WOL_FILTER_SET_IN_LEN];
1092
u8 outbuf[MC_CMD_WOL_FILTER_SET_OUT_LEN];
1096
MCDI_SET_DWORD(inbuf, WOL_FILTER_SET_IN_WOL_TYPE, type);
1097
MCDI_SET_DWORD(inbuf, WOL_FILTER_SET_IN_FILTER_MODE,
1098
MC_CMD_FILTER_MODE_SIMPLE);
1099
memcpy(MCDI_PTR(inbuf, WOL_FILTER_SET_IN_MAGIC_MAC), mac, ETH_ALEN);
1101
rc = efx_mcdi_rpc(efx, MC_CMD_WOL_FILTER_SET, inbuf, sizeof(inbuf),
1102
outbuf, sizeof(outbuf), &outlen);
1106
if (outlen < MC_CMD_WOL_FILTER_SET_OUT_LEN) {
1111
*id_out = (int)MCDI_DWORD(outbuf, WOL_FILTER_SET_OUT_FILTER_ID);
1117
netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d\n", __func__, rc);
1124
efx_mcdi_wol_filter_set_magic(struct efx_nic *efx, const u8 *mac, int *id_out)
1126
return efx_mcdi_wol_filter_set(efx, MC_CMD_WOL_TYPE_MAGIC, mac, id_out);
1130
int efx_mcdi_wol_filter_get_magic(struct efx_nic *efx, int *id_out)
1132
u8 outbuf[MC_CMD_WOL_FILTER_GET_OUT_LEN];
1136
rc = efx_mcdi_rpc(efx, MC_CMD_WOL_FILTER_GET, NULL, 0,
1137
outbuf, sizeof(outbuf), &outlen);
1141
if (outlen < MC_CMD_WOL_FILTER_GET_OUT_LEN) {
1146
*id_out = (int)MCDI_DWORD(outbuf, WOL_FILTER_GET_OUT_FILTER_ID);
1152
netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d\n", __func__, rc);
1157
int efx_mcdi_wol_filter_remove(struct efx_nic *efx, int id)
1159
u8 inbuf[MC_CMD_WOL_FILTER_REMOVE_IN_LEN];
1162
MCDI_SET_DWORD(inbuf, WOL_FILTER_REMOVE_IN_FILTER_ID, (u32)id);
1164
rc = efx_mcdi_rpc(efx, MC_CMD_WOL_FILTER_REMOVE, inbuf, sizeof(inbuf),
1172
netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d\n", __func__, rc);
1177
int efx_mcdi_wol_filter_reset(struct efx_nic *efx)
1181
rc = efx_mcdi_rpc(efx, MC_CMD_WOL_FILTER_RESET, NULL, 0, NULL, 0, NULL);
1188
netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d\n", __func__, rc);