blob: 053de11e7258b2204782193d2bda03a19879adbc [file] [log] [blame]
/* SPDX-License-Identifier: Apache-2.0 OR GPL-2.0-or-later */
#if HAVE_CONFIG_H
#include "config.h"
#endif
#if HAVE_ENDIAN_H
#include <endian.h>
#endif
#include <assert.h>
#include <err.h>
#include <errno.h>
#include <inttypes.h>
#include <stdbool.h>
#include <stdlib.h>
#include <string.h>
#define pr_fmt(x) "astlpc: " x
#include "container_of.h"
#include "crc32.h"
#include "libmctp.h"
#include "libmctp-alloc.h"
#include "libmctp-log.h"
#include "libmctp-astlpc.h"
#include "range.h"
#ifdef MCTP_HAVE_FILEIO
#include <unistd.h>
#include <fcntl.h>
#include <sys/ioctl.h>
#include <sys/mman.h>
#include <linux/aspeed-lpc-ctrl.h>
/* kernel interface */
static const char *kcs_path = "/dev/mctp0";
static const char *lpc_path = "/dev/aspeed-lpc-ctrl";
#endif
struct mctp_astlpc_buffer {
uint32_t offset;
uint32_t size;
};
struct mctp_astlpc_layout {
struct mctp_astlpc_buffer rx;
struct mctp_astlpc_buffer tx;
};
struct mctp_astlpc_protocol {
uint16_t version;
uint32_t (*packet_size)(uint32_t body);
uint32_t (*body_size)(uint32_t packet);
void (*pktbuf_protect)(struct mctp_pktbuf *pkt);
bool (*pktbuf_validate)(struct mctp_pktbuf *pkt);
};
struct mctp_binding_astlpc {
struct mctp_binding binding;
void *lpc_map;
struct mctp_astlpc_layout layout;
uint8_t mode;
uint32_t requested_mtu;
const struct mctp_astlpc_protocol *proto;
/* direct ops data */
struct mctp_binding_astlpc_ops ops;
void *ops_data;
/* fileio ops data */
int kcs_fd;
uint8_t kcs_status;
bool running;
};
#define binding_to_astlpc(b) \
container_of(b, struct mctp_binding_astlpc, binding)
#define astlpc_prlog(ctx, lvl, fmt, ...) \
do { \
bool __bmc = ((ctx)->mode == MCTP_BINDING_ASTLPC_MODE_BMC); \
mctp_prlog(lvl, pr_fmt("%s: " fmt), __bmc ? "bmc" : "host", \
##__VA_ARGS__); \
} while (0)
#define astlpc_prerr(ctx, fmt, ...) \
astlpc_prlog(ctx, MCTP_LOG_ERR, fmt, ##__VA_ARGS__)
#define astlpc_prwarn(ctx, fmt, ...) \
astlpc_prlog(ctx, MCTP_LOG_WARNING, fmt, ##__VA_ARGS__)
#define astlpc_prinfo(ctx, fmt, ...) \
astlpc_prlog(ctx, MCTP_LOG_INFO, fmt, ##__VA_ARGS__)
#define astlpc_prdebug(ctx, fmt, ...) \
astlpc_prlog(ctx, MCTP_LOG_DEBUG, fmt, ##__VA_ARGS__)
/* clang-format off */
#define ASTLPC_MCTP_MAGIC 0x4d435450
#define ASTLPC_VER_BAD 0
#define ASTLPC_VER_MIN 1
/* Support testing of new binding protocols */
#ifndef ASTLPC_VER_CUR
#define ASTLPC_VER_CUR 3
#endif
/* clang-format on */
#ifndef ARRAY_SIZE
#define ARRAY_SIZE(a) (sizeof(a) / sizeof(a[0]))
#endif
static uint32_t astlpc_packet_size_v1(uint32_t body)
{
assert((body + 4) > body);
return body + 4;
}
static uint32_t astlpc_body_size_v1(uint32_t packet)
{
assert((packet - 4) < packet);
return packet - 4;
}
void astlpc_pktbuf_protect_v1(struct mctp_pktbuf *pkt)
{
(void)pkt;
}
bool astlpc_pktbuf_validate_v1(struct mctp_pktbuf *pkt)
{
(void)pkt;
return true;
}
static uint32_t astlpc_packet_size_v3(uint32_t body)
{
assert((body + 4 + 4) > body);
return body + 4 + 4;
}
static uint32_t astlpc_body_size_v3(uint32_t packet)
{
assert((packet - 4 - 4) < packet);
return packet - 4 - 4;
}
void astlpc_pktbuf_protect_v3(struct mctp_pktbuf *pkt)
{
uint32_t code;
code = htobe32(crc32(mctp_pktbuf_hdr(pkt), mctp_pktbuf_size(pkt)));
mctp_prdebug("%s: 0x%" PRIx32, __func__, code);
mctp_pktbuf_push(pkt, &code, 4);
}
bool astlpc_pktbuf_validate_v3(struct mctp_pktbuf *pkt)
{
uint32_t code;
void *check;
code = be32toh(crc32(mctp_pktbuf_hdr(pkt), mctp_pktbuf_size(pkt) - 4));
mctp_prdebug("%s: 0x%" PRIx32, __func__, code);
check = mctp_pktbuf_pop(pkt, 4);
return check && !memcmp(&code, check, 4);
}
static const struct mctp_astlpc_protocol astlpc_protocol_version[] = {
[0] = {
.version = 0,
.packet_size = NULL,
.body_size = NULL,
.pktbuf_protect = NULL,
.pktbuf_validate = NULL,
},
[1] = {
.version = 1,
.packet_size = astlpc_packet_size_v1,
.body_size = astlpc_body_size_v1,
.pktbuf_protect = astlpc_pktbuf_protect_v1,
.pktbuf_validate = astlpc_pktbuf_validate_v1,
},
[2] = {
.version = 2,
.packet_size = astlpc_packet_size_v1,
.body_size = astlpc_body_size_v1,
.pktbuf_protect = astlpc_pktbuf_protect_v1,
.pktbuf_validate = astlpc_pktbuf_validate_v1,
},
[3] = {
.version = 3,
.packet_size = astlpc_packet_size_v3,
.body_size = astlpc_body_size_v3,
.pktbuf_protect = astlpc_pktbuf_protect_v3,
.pktbuf_validate = astlpc_pktbuf_validate_v3,
},
};
struct mctp_lpcmap_hdr {
uint32_t magic;
uint16_t bmc_ver_min;
uint16_t bmc_ver_cur;
uint16_t host_ver_min;
uint16_t host_ver_cur;
uint16_t negotiated_ver;
uint16_t pad0;
struct {
uint32_t rx_offset;
uint32_t rx_size;
uint32_t tx_offset;
uint32_t tx_size;
} layout;
} __attribute__((packed));
static const uint32_t control_size = 0x100;
#define LPC_WIN_SIZE (1 * 1024 * 1024)
#define KCS_STATUS_BMC_READY 0x80
#define KCS_STATUS_CHANNEL_ACTIVE 0x40
#define KCS_STATUS_IBF 0x02
#define KCS_STATUS_OBF 0x01
static inline int mctp_astlpc_kcs_write(struct mctp_binding_astlpc *astlpc,
enum mctp_binding_astlpc_kcs_reg reg,
uint8_t val)
{
return astlpc->ops.kcs_write(astlpc->ops_data, reg, val);
}
static inline int mctp_astlpc_kcs_read(struct mctp_binding_astlpc *astlpc,
enum mctp_binding_astlpc_kcs_reg reg,
uint8_t *val)
{
return astlpc->ops.kcs_read(astlpc->ops_data, reg, val);
}
static inline int mctp_astlpc_lpc_write(struct mctp_binding_astlpc *astlpc,
const void *buf, long offset,
size_t len)
{
astlpc_prdebug(astlpc, "%s: %zu bytes to 0x%lx", __func__, len, offset);
assert(offset >= 0);
/* Indirect access */
if (astlpc->ops.lpc_write) {
void *data = astlpc->ops_data;
return astlpc->ops.lpc_write(data, buf, offset, len);
}
/* Direct mapping */
assert(astlpc->lpc_map);
memcpy(&((char *)astlpc->lpc_map)[offset], buf, len);
return 0;
}
static inline int mctp_astlpc_lpc_read(struct mctp_binding_astlpc *astlpc,
void *buf, long offset, size_t len)
{
astlpc_prdebug(astlpc, "%s: %zu bytes from 0x%lx", __func__, len,
offset);
assert(offset >= 0);
/* Indirect access */
if (astlpc->ops.lpc_read) {
void *data = astlpc->ops_data;
return astlpc->ops.lpc_read(data, buf, offset, len);
}
/* Direct mapping */
assert(astlpc->lpc_map);
memcpy(buf, &((char *)astlpc->lpc_map)[offset], len);
return 0;
}
static int mctp_astlpc_kcs_set_status(struct mctp_binding_astlpc *astlpc,
uint8_t status)
{
uint8_t data;
int rc;
/* Since we're setting the status register, we want the other endpoint
* to be interrupted. However, some hardware may only raise a host-side
* interrupt on an ODR event.
* So, write a dummy value of 0xff to ODR, which will ensure that an
* interrupt is triggered, and can be ignored by the host.
*/
data = 0xff;
rc = mctp_astlpc_kcs_write(astlpc, MCTP_ASTLPC_KCS_REG_STATUS, status);
if (rc) {
astlpc_prwarn(astlpc, "KCS status write failed");
return -1;
}
rc = mctp_astlpc_kcs_write(astlpc, MCTP_ASTLPC_KCS_REG_DATA, data);
if (rc) {
astlpc_prwarn(astlpc, "KCS dummy data write failed");
return -1;
}
return 0;
}
static int mctp_astlpc_layout_read(struct mctp_binding_astlpc *astlpc,
struct mctp_astlpc_layout *layout)
{
struct mctp_lpcmap_hdr hdr;
int rc;
rc = mctp_astlpc_lpc_read(astlpc, &hdr, 0, sizeof(hdr));
if (rc < 0)
return rc;
/* Flip the buffers as the names are defined in terms of the host */
if (astlpc->mode == MCTP_BINDING_ASTLPC_MODE_BMC) {
layout->rx.offset = be32toh(hdr.layout.tx_offset);
layout->rx.size = be32toh(hdr.layout.tx_size);
layout->tx.offset = be32toh(hdr.layout.rx_offset);
layout->tx.size = be32toh(hdr.layout.rx_size);
} else {
assert(astlpc->mode == MCTP_BINDING_ASTLPC_MODE_HOST);
layout->rx.offset = be32toh(hdr.layout.rx_offset);
layout->rx.size = be32toh(hdr.layout.rx_size);
layout->tx.offset = be32toh(hdr.layout.tx_offset);
layout->tx.size = be32toh(hdr.layout.tx_size);
}
return 0;
}
static int mctp_astlpc_layout_write(struct mctp_binding_astlpc *astlpc,
struct mctp_astlpc_layout *layout)
{
uint32_t rx_size_be;
if (astlpc->mode == MCTP_BINDING_ASTLPC_MODE_BMC) {
struct mctp_lpcmap_hdr hdr;
/*
* Flip the buffers as the names are defined in terms of the
* host
*/
hdr.layout.rx_offset = htobe32(layout->tx.offset);
hdr.layout.rx_size = htobe32(layout->tx.size);
hdr.layout.tx_offset = htobe32(layout->rx.offset);
hdr.layout.tx_size = htobe32(layout->rx.size);
return mctp_astlpc_lpc_write(astlpc, &hdr.layout,
offsetof(struct mctp_lpcmap_hdr, layout),
sizeof(hdr.layout));
}
assert(astlpc->mode == MCTP_BINDING_ASTLPC_MODE_HOST);
/*
* As of v2 we only need to write rx_size - the offsets are controlled
* by the BMC, as is the BMC's rx_size (host tx_size).
*/
rx_size_be = htobe32(layout->rx.size);
return mctp_astlpc_lpc_write(astlpc, &rx_size_be,
offsetof(struct mctp_lpcmap_hdr, layout.rx_size),
sizeof(rx_size_be));
}
static bool
mctp_astlpc_buffer_validate(const struct mctp_binding_astlpc *astlpc,
const struct mctp_astlpc_buffer *buf,
const char *name)
{
/* Check for overflow */
if (buf->offset + buf->size < buf->offset) {
mctp_prerr(
"%s packet buffer parameters overflow: offset: 0x%" PRIx32
", size: %" PRIu32,
name, buf->offset, buf->size);
return false;
}
/* Check that the buffers are contained within the allocated space */
if (buf->offset + buf->size > LPC_WIN_SIZE) {
mctp_prerr(
"%s packet buffer parameters exceed %uM window size: offset: 0x%" PRIx32
", size: %" PRIu32,
name, (LPC_WIN_SIZE / (1024 * 1024)), buf->offset,
buf->size);
return false;
}
/* Check that the baseline transmission unit is supported */
if (buf->size < astlpc->proto->packet_size(MCTP_PACKET_SIZE(MCTP_BTU))) {
mctp_prerr(
"%s packet buffer too small: Require %" PRIu32 " bytes to support the %u byte baseline transmission unit, found %" PRIu32,
name,
astlpc->proto->packet_size(MCTP_PACKET_SIZE(MCTP_BTU)),
MCTP_BTU, buf->size);
return false;
}
/* Check for overlap with the control space */
if (buf->offset < control_size) {
mctp_prerr(
"%s packet buffer overlaps control region {0x%" PRIx32
", %" PRIu32 "}: Rx {0x%" PRIx32 ", %" PRIu32 "}",
name, 0U, control_size, buf->offset, buf->size);
return false;
}
return true;
}
static bool
mctp_astlpc_layout_validate(const struct mctp_binding_astlpc *astlpc,
const struct mctp_astlpc_layout *layout)
{
const struct mctp_astlpc_buffer *rx = &layout->rx;
const struct mctp_astlpc_buffer *tx = &layout->tx;
bool rx_valid, tx_valid;
rx_valid = mctp_astlpc_buffer_validate(astlpc, rx, "Rx");
tx_valid = mctp_astlpc_buffer_validate(astlpc, tx, "Tx");
if (!(rx_valid && tx_valid))
return false;
/* Check that the buffers are disjoint */
if ((rx->offset <= tx->offset && rx->offset + rx->size > tx->offset) ||
(tx->offset <= rx->offset && tx->offset + tx->size > rx->offset)) {
mctp_prerr("Rx and Tx packet buffers overlap: Rx {0x%" PRIx32
", %" PRIu32 "}, Tx {0x%" PRIx32 ", %" PRIu32 "}",
rx->offset, rx->size, tx->offset, tx->size);
return false;
}
return true;
}
static int mctp_astlpc_init_bmc(struct mctp_binding_astlpc *astlpc)
{
struct mctp_lpcmap_hdr hdr = { 0 };
uint8_t status;
uint32_t sz;
/*
* The largest buffer size is half of the allocated MCTP space
* excluding the control space.
*/
sz = ((LPC_WIN_SIZE - control_size) / 2);
/*
* Trim the MTU to a multiple of 16 to meet the requirements of 12.17
* Query Hop in DSP0236 v1.3.0.
*/
sz = MCTP_BODY_SIZE(astlpc->proto->body_size(sz));
sz &= ~0xfUL;
sz = astlpc->proto->packet_size(MCTP_PACKET_SIZE(sz));
if (astlpc->requested_mtu) {
uint32_t rpkt, rmtu;
rmtu = astlpc->requested_mtu;
rpkt = astlpc->proto->packet_size(MCTP_PACKET_SIZE(rmtu));
sz = MIN(sz, rpkt);
}
/* Flip the buffers as the names are defined in terms of the host */
astlpc->layout.tx.offset = control_size;
astlpc->layout.tx.size = sz;
astlpc->layout.rx.offset =
astlpc->layout.tx.offset + astlpc->layout.tx.size;
astlpc->layout.rx.size = sz;
if (!mctp_astlpc_layout_validate(astlpc, &astlpc->layout)) {
astlpc_prerr(astlpc, "Cannot support an MTU of %" PRIu32, sz);
return -EINVAL;
}
hdr = (struct mctp_lpcmap_hdr){
.magic = htobe32(ASTLPC_MCTP_MAGIC),
.bmc_ver_min = htobe16(ASTLPC_VER_MIN),
.bmc_ver_cur = htobe16(ASTLPC_VER_CUR),
/* Flip the buffers back as we're now describing the host's
* configuration to the host */
.layout.rx_offset = htobe32(astlpc->layout.tx.offset),
.layout.rx_size = htobe32(astlpc->layout.tx.size),
.layout.tx_offset = htobe32(astlpc->layout.rx.offset),
.layout.tx_size = htobe32(astlpc->layout.rx.size),
};
mctp_astlpc_lpc_write(astlpc, &hdr, 0, sizeof(hdr));
/*
* Set status indicating that the BMC is now active. Be explicit about
* clearing OBF; we're reinitialising the binding and so any previous
* buffer state is irrelevant.
*/
status = KCS_STATUS_BMC_READY & ~KCS_STATUS_OBF;
return mctp_astlpc_kcs_set_status(astlpc, status);
}
static int mctp_binding_astlpc_start_bmc(struct mctp_binding *b)
{
struct mctp_binding_astlpc *astlpc =
container_of(b, struct mctp_binding_astlpc, binding);
astlpc->proto = &astlpc_protocol_version[ASTLPC_VER_CUR];
return mctp_astlpc_init_bmc(astlpc);
}
static bool mctp_astlpc_validate_version(uint16_t bmc_ver_min,
uint16_t bmc_ver_cur,
uint16_t host_ver_min,
uint16_t host_ver_cur)
{
if (!(bmc_ver_min && bmc_ver_cur && host_ver_min && host_ver_cur)) {
mctp_prerr("Invalid version present in [%" PRIu16 ", %" PRIu16
"], [%" PRIu16 ", %" PRIu16 "]",
bmc_ver_min, bmc_ver_cur, host_ver_min,
host_ver_cur);
return false;
} else if (bmc_ver_min > bmc_ver_cur) {
mctp_prerr("Invalid bmc version range [%" PRIu16 ", %" PRIu16
"]",
bmc_ver_min, bmc_ver_cur);
return false;
} else if (host_ver_min > host_ver_cur) {
mctp_prerr("Invalid host version range [%" PRIu16 ", %" PRIu16
"]",
host_ver_min, host_ver_cur);
return false;
} else if ((host_ver_cur < bmc_ver_min) ||
(host_ver_min > bmc_ver_cur)) {
mctp_prerr(
"Unable to satisfy version negotiation with ranges [%" PRIu16
", %" PRIu16 "] and [%" PRIu16 ", %" PRIu16 "]",
bmc_ver_min, bmc_ver_cur, host_ver_min, host_ver_cur);
return false;
}
return true;
}
static int mctp_astlpc_negotiate_layout_host(struct mctp_binding_astlpc *astlpc)
{
struct mctp_astlpc_layout layout;
uint32_t rmtu;
uint32_t sz;
int rc;
rc = mctp_astlpc_layout_read(astlpc, &layout);
if (rc < 0)
return rc;
if (!mctp_astlpc_layout_validate(astlpc, &layout)) {
astlpc_prerr(
astlpc,
"BMC provided invalid buffer layout: Rx {0x%" PRIx32
", %" PRIu32 "}, Tx {0x%" PRIx32 ", %" PRIu32 "}",
layout.rx.offset, layout.rx.size, layout.tx.offset,
layout.tx.size);
return -EINVAL;
}
astlpc_prinfo(astlpc, "Desire an MTU of %" PRIu32 " bytes",
astlpc->requested_mtu);
rmtu = astlpc->requested_mtu;
sz = astlpc->proto->packet_size(MCTP_PACKET_SIZE(rmtu));
layout.rx.size = sz;
if (!mctp_astlpc_layout_validate(astlpc, &layout)) {
astlpc_prerr(
astlpc,
"Generated invalid buffer layout with size %" PRIu32
": Rx {0x%" PRIx32 ", %" PRIu32 "}, Tx {0x%" PRIx32
", %" PRIu32 "}",
sz, layout.rx.offset, layout.rx.size, layout.tx.offset,
layout.tx.size);
return -EINVAL;
}
astlpc_prinfo(astlpc, "Requesting MTU of %" PRIu32 " bytes",
astlpc->requested_mtu);
return mctp_astlpc_layout_write(astlpc, &layout);
}
static uint16_t mctp_astlpc_negotiate_version(uint16_t bmc_ver_min,
uint16_t bmc_ver_cur,
uint16_t host_ver_min,
uint16_t host_ver_cur)
{
if (!mctp_astlpc_validate_version(bmc_ver_min, bmc_ver_cur,
host_ver_min, host_ver_cur))
return ASTLPC_VER_BAD;
if (bmc_ver_cur < host_ver_cur)
return bmc_ver_cur;
return host_ver_cur;
}
static int mctp_astlpc_init_host(struct mctp_binding_astlpc *astlpc)
{
const uint16_t ver_min_be = htobe16(ASTLPC_VER_MIN);
const uint16_t ver_cur_be = htobe16(ASTLPC_VER_CUR);
uint16_t bmc_ver_min, bmc_ver_cur, negotiated;
struct mctp_lpcmap_hdr hdr;
uint8_t status;
int rc;
rc = mctp_astlpc_kcs_read(astlpc, MCTP_ASTLPC_KCS_REG_STATUS, &status);
if (rc) {
mctp_prwarn("KCS status read failed");
return rc;
}
astlpc->kcs_status = status;
if (!(status & KCS_STATUS_BMC_READY))
return -EHOSTDOWN;
mctp_astlpc_lpc_read(astlpc, &hdr, 0, sizeof(hdr));
bmc_ver_min = be16toh(hdr.bmc_ver_min);
bmc_ver_cur = be16toh(hdr.bmc_ver_cur);
/* Calculate the expected value of negotiated_ver */
negotiated = mctp_astlpc_negotiate_version(bmc_ver_min, bmc_ver_cur,
ASTLPC_VER_MIN,
ASTLPC_VER_CUR);
if (!negotiated) {
astlpc_prerr(astlpc, "Cannot negotiate with invalid versions");
return -EINVAL;
}
/* Assign protocol ops so we can calculate the packet buffer sizes */
assert(negotiated < ARRAY_SIZE(astlpc_protocol_version));
astlpc->proto = &astlpc_protocol_version[negotiated];
/* Negotiate packet buffers in v2 style if the BMC supports it */
if (negotiated >= 2) {
rc = mctp_astlpc_negotiate_layout_host(astlpc);
if (rc < 0)
return rc;
}
/* Advertise the host's supported protocol versions */
mctp_astlpc_lpc_write(astlpc, &ver_min_be,
offsetof(struct mctp_lpcmap_hdr, host_ver_min),
sizeof(ver_min_be));
mctp_astlpc_lpc_write(astlpc, &ver_cur_be,
offsetof(struct mctp_lpcmap_hdr, host_ver_cur),
sizeof(ver_cur_be));
/* Send channel init command */
rc = mctp_astlpc_kcs_write(astlpc, MCTP_ASTLPC_KCS_REG_DATA, 0x0);
if (rc) {
astlpc_prwarn(astlpc, "KCS write failed");
}
/*
* Configure the host so `astlpc->proto->version == 0` holds until we
* receive a subsequent status update from the BMC. Until then,
* `astlpc->proto->version == 0` indicates that we're yet to complete
* the channel initialisation handshake.
*
* When the BMC provides a status update with KCS_STATUS_CHANNEL_ACTIVE
* set we will assign the appropriate protocol ops struct in accordance
* with `negotiated_ver`.
*/
astlpc->proto = &astlpc_protocol_version[ASTLPC_VER_BAD];
return rc;
}
static int mctp_binding_astlpc_start_host(struct mctp_binding *b)
{
struct mctp_binding_astlpc *astlpc =
container_of(b, struct mctp_binding_astlpc, binding);
return mctp_astlpc_init_host(astlpc);
}
static bool __mctp_astlpc_kcs_ready(struct mctp_binding_astlpc *astlpc,
uint8_t status, bool is_write)
{
bool is_bmc;
bool ready_state;
uint8_t flag;
is_bmc = (astlpc->mode == MCTP_BINDING_ASTLPC_MODE_BMC);
flag = (is_bmc ^ is_write) ? KCS_STATUS_IBF : KCS_STATUS_OBF;
ready_state = is_write ? 0 : 1;
return !!(status & flag) == ready_state;
}
static inline bool
mctp_astlpc_kcs_read_ready(struct mctp_binding_astlpc *astlpc, uint8_t status)
{
return __mctp_astlpc_kcs_ready(astlpc, status, false);
}
static inline bool
mctp_astlpc_kcs_write_ready(struct mctp_binding_astlpc *astlpc, uint8_t status)
{
return __mctp_astlpc_kcs_ready(astlpc, status, true);
}
static int mctp_astlpc_kcs_send(struct mctp_binding_astlpc *astlpc,
uint8_t data)
{
uint8_t status;
int rc;
for (;;) {
rc = mctp_astlpc_kcs_read(astlpc, MCTP_ASTLPC_KCS_REG_STATUS,
&status);
if (rc) {
astlpc_prwarn(astlpc, "KCS status read failed");
return -1;
}
if (mctp_astlpc_kcs_write_ready(astlpc, status))
break;
/* todo: timeout */
}
rc = mctp_astlpc_kcs_write(astlpc, MCTP_ASTLPC_KCS_REG_DATA, data);
if (rc) {
astlpc_prwarn(astlpc, "KCS data write failed");
return -1;
}
return 0;
}
static int mctp_binding_astlpc_tx(struct mctp_binding *b,
struct mctp_pktbuf *pkt)
{
struct mctp_binding_astlpc *astlpc = binding_to_astlpc(b);
uint32_t len, len_be;
struct mctp_hdr *hdr;
hdr = mctp_pktbuf_hdr(pkt);
len = mctp_pktbuf_size(pkt);
astlpc_prdebug(astlpc,
"%s: Transmitting %" PRIu32
"-byte packet (%hhu, %hhu, 0x%hhx)",
__func__, len, hdr->src, hdr->dest, hdr->flags_seq_tag);
if (len > astlpc->proto->body_size(astlpc->layout.tx.size)) {
astlpc_prwarn(astlpc, "invalid TX len %" PRIu32 ": %" PRIu32, len,
astlpc->proto->body_size(astlpc->layout.tx.size));
return -1;
}
len_be = htobe32(len);
mctp_astlpc_lpc_write(astlpc, &len_be, astlpc->layout.tx.offset,
sizeof(len_be));
astlpc->proto->pktbuf_protect(pkt);
len = mctp_pktbuf_size(pkt);
mctp_astlpc_lpc_write(astlpc, hdr, astlpc->layout.tx.offset + 4, len);
mctp_binding_set_tx_enabled(b, false);
mctp_astlpc_kcs_send(astlpc, 0x1);
return 0;
}
static uint32_t mctp_astlpc_calculate_mtu(struct mctp_binding_astlpc *astlpc,
struct mctp_astlpc_layout *layout)
{
uint32_t low, high, limit, rpkt;
/* Derive the largest MTU the BMC _can_ support */
low = MIN(astlpc->layout.rx.offset, astlpc->layout.tx.offset);
high = MAX(astlpc->layout.rx.offset, astlpc->layout.tx.offset);
limit = high - low;
/* Determine the largest MTU the BMC _wants_ to support */
if (astlpc->requested_mtu) {
uint32_t rmtu = astlpc->requested_mtu;
rpkt = astlpc->proto->packet_size(MCTP_PACKET_SIZE(rmtu));
limit = MIN(limit, rpkt);
}
/* Determine the accepted MTU, applied both directions by convention */
rpkt = MIN(limit, layout->tx.size);
return MCTP_BODY_SIZE(astlpc->proto->body_size(rpkt));
}
static int mctp_astlpc_negotiate_layout_bmc(struct mctp_binding_astlpc *astlpc)
{
struct mctp_astlpc_layout proposed, pending;
uint32_t sz, mtu;
int rc;
/* Do we have a valid protocol version? */
if (!astlpc->proto->version)
return -EINVAL;
/* Extract the host's proposed layout */
rc = mctp_astlpc_layout_read(astlpc, &proposed);
if (rc < 0)
return rc;
/* Do we have a reasonable layout? */
if (!mctp_astlpc_layout_validate(astlpc, &proposed))
return -EINVAL;
/* Negotiate the MTU */
mtu = mctp_astlpc_calculate_mtu(astlpc, &proposed);
sz = astlpc->proto->packet_size(MCTP_PACKET_SIZE(mtu));
/*
* Use symmetric MTUs by convention and to pass constraints in rx/tx
* functions
*/
pending = astlpc->layout;
pending.tx.size = sz;
pending.rx.size = sz;
if (mctp_astlpc_layout_validate(astlpc, &pending)) {
/* We found a sensible Rx MTU, so honour it */
astlpc->layout = pending;
/* Enforce the negotiated MTU */
rc = mctp_astlpc_layout_write(astlpc, &astlpc->layout);
if (rc < 0)
return rc;
astlpc_prinfo(astlpc, "Negotiated an MTU of %" PRIu32 " bytes",
mtu);
} else {
astlpc_prwarn(astlpc, "MTU negotiation failed");
return -EINVAL;
}
if (astlpc->proto->version >= 2)
astlpc->binding.pkt_size = MCTP_PACKET_SIZE(mtu);
return 0;
}
static void mctp_astlpc_init_channel(struct mctp_binding_astlpc *astlpc)
{
uint16_t negotiated, negotiated_be;
struct mctp_lpcmap_hdr hdr;
uint8_t status;
int rc;
mctp_astlpc_lpc_read(astlpc, &hdr, 0, sizeof(hdr));
/* Version negotiation */
negotiated =
mctp_astlpc_negotiate_version(ASTLPC_VER_MIN, ASTLPC_VER_CUR,
be16toh(hdr.host_ver_min),
be16toh(hdr.host_ver_cur));
/* MTU negotiation requires knowing which protocol we'll use */
assert(negotiated < ARRAY_SIZE(astlpc_protocol_version));
astlpc->proto = &astlpc_protocol_version[negotiated];
/* Host Rx MTU negotiation: Failure terminates channel init */
rc = mctp_astlpc_negotiate_layout_bmc(astlpc);
if (rc < 0)
negotiated = ASTLPC_VER_BAD;
/* Populate the negotiated version */
negotiated_be = htobe16(negotiated);
mctp_astlpc_lpc_write(astlpc, &negotiated_be,
offsetof(struct mctp_lpcmap_hdr, negotiated_ver),
sizeof(negotiated_be));
/* Finalise the configuration */
status = KCS_STATUS_BMC_READY | KCS_STATUS_OBF;
if (negotiated > 0) {
astlpc_prinfo(astlpc, "Negotiated binding version %" PRIu16,
negotiated);
status |= KCS_STATUS_CHANNEL_ACTIVE;
} else {
astlpc_prerr(astlpc, "Failed to initialise channel");
}
mctp_astlpc_kcs_set_status(astlpc, status);
mctp_binding_set_tx_enabled(&astlpc->binding,
status & KCS_STATUS_CHANNEL_ACTIVE);
}
static void mctp_astlpc_rx_start(struct mctp_binding_astlpc *astlpc)
{
struct mctp_pktbuf *pkt;
uint32_t body, packet;
mctp_astlpc_lpc_read(astlpc, &body, astlpc->layout.rx.offset,
sizeof(body));
body = be32toh(body);
if (body > astlpc->proto->body_size(astlpc->layout.rx.size)) {
astlpc_prwarn(astlpc, "invalid RX len 0x%x", body);
return;
}
assert(astlpc->binding.pkt_size >= 0);
if (body > (uint32_t)astlpc->binding.pkt_size) {
astlpc_prwarn(astlpc, "invalid RX len 0x%x", body);
return;
}
/* Eliminate the medium-specific header that we just read */
packet = astlpc->proto->packet_size(body) - 4;
pkt = mctp_pktbuf_alloc(&astlpc->binding, packet);
if (!pkt)
goto out_complete;
/*
* Read payload and medium-specific trailer from immediately after the
* medium-specific header.
*/
mctp_astlpc_lpc_read(astlpc, mctp_pktbuf_hdr(pkt),
astlpc->layout.rx.offset + 4, packet);
/*
* v3 will validate the CRC32 in the medium-specific trailer and adjust
* the packet size accordingly. On older protocols validation is a no-op
* that always returns true.
*/
if (astlpc->proto->pktbuf_validate(pkt)) {
mctp_bus_rx(&astlpc->binding, pkt);
} else {
/* TODO: Drop any associated assembly */
mctp_pktbuf_free(pkt);
astlpc_prdebug(astlpc, "Dropped corrupt packet");
}
out_complete:
mctp_astlpc_kcs_send(astlpc, 0x2);
}
static void mctp_astlpc_tx_complete(struct mctp_binding_astlpc *astlpc)
{
mctp_binding_set_tx_enabled(&astlpc->binding, true);
}
static int mctp_astlpc_finalise_channel(struct mctp_binding_astlpc *astlpc)
{
struct mctp_astlpc_layout layout;
uint16_t negotiated;
int rc;
rc = mctp_astlpc_lpc_read(astlpc, &negotiated,
offsetof(struct mctp_lpcmap_hdr,
negotiated_ver),
sizeof(negotiated));
if (rc < 0)
return rc;
negotiated = be16toh(negotiated);
astlpc_prerr(astlpc, "Version negotiation got: %u", negotiated);
if (negotiated == ASTLPC_VER_BAD || negotiated < ASTLPC_VER_MIN ||
negotiated > ASTLPC_VER_CUR) {
astlpc_prerr(astlpc, "Failed to negotiate version, got: %u\n",
negotiated);
return -EINVAL;
}
assert(negotiated < ARRAY_SIZE(astlpc_protocol_version));
astlpc->proto = &astlpc_protocol_version[negotiated];
rc = mctp_astlpc_layout_read(astlpc, &layout);
if (rc < 0)
return rc;
if (!mctp_astlpc_layout_validate(astlpc, &layout)) {
mctp_prerr("BMC proposed invalid buffer parameters");
return -EINVAL;
}
astlpc->layout = layout;
if (negotiated >= 2)
astlpc->binding.pkt_size =
astlpc->proto->body_size(astlpc->layout.tx.size);
return 0;
}
static int mctp_astlpc_update_channel(struct mctp_binding_astlpc *astlpc,
uint8_t status)
{
uint8_t updated;
int rc = 0;
assert(astlpc->mode == MCTP_BINDING_ASTLPC_MODE_HOST);
updated = astlpc->kcs_status ^ status;
astlpc_prdebug(astlpc, "%s: status: 0x%x, update: 0x%x", __func__,
status, updated);
if (updated & KCS_STATUS_BMC_READY) {
if (status & KCS_STATUS_BMC_READY) {
astlpc->kcs_status = status;
return astlpc->binding.start(&astlpc->binding);
} else {
mctp_binding_set_tx_enabled(&astlpc->binding, false);
}
}
if (astlpc->proto->version == 0 ||
updated & KCS_STATUS_CHANNEL_ACTIVE) {
bool enable;
rc = mctp_astlpc_finalise_channel(astlpc);
enable = (status & KCS_STATUS_CHANNEL_ACTIVE) && rc == 0;
mctp_binding_set_tx_enabled(&astlpc->binding, enable);
}
astlpc->kcs_status = status;
return rc;
}
int mctp_astlpc_poll(struct mctp_binding_astlpc *astlpc)
{
uint8_t status, data;
int rc;
rc = mctp_astlpc_kcs_read(astlpc, MCTP_ASTLPC_KCS_REG_STATUS, &status);
if (rc) {
astlpc_prwarn(astlpc, "KCS read error");
return -1;
}
astlpc_prdebug(astlpc, "%s: status: 0x%hhx", __func__, status);
if (!mctp_astlpc_kcs_read_ready(astlpc, status))
return 0;
rc = mctp_astlpc_kcs_read(astlpc, MCTP_ASTLPC_KCS_REG_DATA, &data);
if (rc) {
astlpc_prwarn(astlpc, "KCS data read error");
return -1;
}
astlpc_prdebug(astlpc, "%s: data: 0x%hhx", __func__, data);
if (!astlpc->proto->version && !(data == 0x0 || data == 0xff)) {
astlpc_prwarn(astlpc, "Invalid message for binding state: 0x%x",
data);
return 0;
}
switch (data) {
case 0x0:
mctp_astlpc_init_channel(astlpc);
break;
case 0x1:
mctp_astlpc_rx_start(astlpc);
break;
case 0x2:
mctp_astlpc_tx_complete(astlpc);
break;
case 0xff:
/* No responsibilities for the BMC on 0xff */
if (astlpc->mode == MCTP_BINDING_ASTLPC_MODE_HOST) {
rc = mctp_astlpc_update_channel(astlpc, status);
if (rc < 0)
return rc;
}
break;
default:
astlpc_prwarn(astlpc, "unknown message 0x%x", data);
}
/* Handle silent loss of bmc-ready */
if (astlpc->mode == MCTP_BINDING_ASTLPC_MODE_HOST) {
if (!(status & KCS_STATUS_BMC_READY && data == 0xff))
return mctp_astlpc_update_channel(astlpc, status);
}
return rc;
}
/* allocate and basic initialisation */
static struct mctp_binding_astlpc *__mctp_astlpc_init(uint8_t mode,
uint32_t mtu)
{
struct mctp_binding_astlpc *astlpc;
assert((mode == MCTP_BINDING_ASTLPC_MODE_BMC) ||
(mode == MCTP_BINDING_ASTLPC_MODE_HOST));
astlpc = __mctp_alloc(sizeof(*astlpc));
if (!astlpc)
return NULL;
memset(astlpc, 0, sizeof(*astlpc));
astlpc->mode = mode;
astlpc->lpc_map = NULL;
astlpc->requested_mtu = mtu;
astlpc->binding.name = "astlpc";
astlpc->binding.version = 1;
astlpc->binding.pkt_size =
MCTP_PACKET_SIZE(mtu > MCTP_BTU ? mtu : MCTP_BTU);
astlpc->binding.pkt_header = 4;
astlpc->binding.pkt_trailer = 4;
astlpc->binding.tx = mctp_binding_astlpc_tx;
if (mode == MCTP_BINDING_ASTLPC_MODE_BMC)
astlpc->binding.start = mctp_binding_astlpc_start_bmc;
else if (mode == MCTP_BINDING_ASTLPC_MODE_HOST)
astlpc->binding.start = mctp_binding_astlpc_start_host;
else {
astlpc_prerr(astlpc, "%s: Invalid mode: %d\n", __func__, mode);
__mctp_free(astlpc);
return NULL;
}
return astlpc;
}
struct mctp_binding *mctp_binding_astlpc_core(struct mctp_binding_astlpc *b)
{
return &b->binding;
}
struct mctp_binding_astlpc *
mctp_astlpc_init(uint8_t mode, uint32_t mtu, void *lpc_map,
const struct mctp_binding_astlpc_ops *ops, void *ops_data)
{
struct mctp_binding_astlpc *astlpc;
if (!(mode == MCTP_BINDING_ASTLPC_MODE_BMC ||
mode == MCTP_BINDING_ASTLPC_MODE_HOST)) {
mctp_prerr("Unknown binding mode: %u", mode);
return NULL;
}
astlpc = __mctp_astlpc_init(mode, mtu);
if (!astlpc)
return NULL;
memcpy(&astlpc->ops, ops, sizeof(astlpc->ops));
astlpc->ops_data = ops_data;
astlpc->lpc_map = lpc_map;
astlpc->mode = mode;
return astlpc;
}
struct mctp_binding_astlpc *
mctp_astlpc_init_ops(const struct mctp_binding_astlpc_ops *ops, void *ops_data,
void *lpc_map)
{
return mctp_astlpc_init(MCTP_BINDING_ASTLPC_MODE_BMC, MCTP_BTU, lpc_map,
ops, ops_data);
}
void mctp_astlpc_destroy(struct mctp_binding_astlpc *astlpc)
{
/* Clear channel-active and bmc-ready */
if (astlpc->mode == MCTP_BINDING_ASTLPC_MODE_BMC)
mctp_astlpc_kcs_set_status(astlpc, KCS_STATUS_OBF);
__mctp_free(astlpc);
}
#ifdef MCTP_HAVE_FILEIO
static int mctp_astlpc_init_fileio_lpc(struct mctp_binding_astlpc *astlpc)
{
struct aspeed_lpc_ctrl_mapping map = {
.window_type = ASPEED_LPC_CTRL_WINDOW_MEMORY,
.window_id = 0, /* There's only one */
.flags = 0,
.addr = 0,
.offset = 0,
.size = 0
};
void *lpc_map_base;
int fd, rc;
fd = open(lpc_path, O_RDWR | O_SYNC);
if (fd < 0) {
astlpc_prwarn(astlpc, "LPC open (%s) failed", lpc_path);
return -1;
}
rc = ioctl(fd, ASPEED_LPC_CTRL_IOCTL_GET_SIZE, &map);
if (rc) {
astlpc_prwarn(astlpc, "LPC GET_SIZE failed");
close(fd);
return -1;
}
lpc_map_base =
mmap(NULL, map.size, PROT_READ | PROT_WRITE, MAP_SHARED, fd, 0);
if (lpc_map_base == MAP_FAILED) {
astlpc_prwarn(astlpc, "LPC mmap failed");
rc = -1;
} else {
astlpc->lpc_map = lpc_map_base + map.size - LPC_WIN_SIZE;
}
close(fd);
return rc;
}
static int mctp_astlpc_init_fileio_kcs(struct mctp_binding_astlpc *astlpc)
{
astlpc->kcs_fd = open(kcs_path, O_RDWR);
if (astlpc->kcs_fd < 0)
return -1;
return 0;
}
static int __mctp_astlpc_fileio_kcs_read(void *arg,
enum mctp_binding_astlpc_kcs_reg reg, uint8_t *val)
{
struct mctp_binding_astlpc *astlpc = arg;
off_t offset = reg;
int rc;
rc = pread(astlpc->kcs_fd, val, 1, offset);
return rc == 1 ? 0 : -1;
}
static int __mctp_astlpc_fileio_kcs_write(void *arg,
enum mctp_binding_astlpc_kcs_reg reg, uint8_t val)
{
struct mctp_binding_astlpc *astlpc = arg;
off_t offset = reg;
int rc;
rc = pwrite(astlpc->kcs_fd, &val, 1, offset);
return rc == 1 ? 0 : -1;
}
int mctp_astlpc_get_fd(struct mctp_binding_astlpc *astlpc)
{
return astlpc->kcs_fd;
}
struct mctp_binding_astlpc *mctp_astlpc_init_fileio(void)
{
struct mctp_binding_astlpc *astlpc;
int rc;
/*
* If we're doing file IO then we're very likely not running
* freestanding, so lets assume that we're on the BMC side.
*
* Requesting an MTU of 0 requests the largest possible MTU, whatever
* value that might take.
*/
astlpc = __mctp_astlpc_init(MCTP_BINDING_ASTLPC_MODE_BMC, 0);
if (!astlpc)
return NULL;
/* Set internal operations for kcs. We use direct accesses to the lpc
* map area */
astlpc->ops.kcs_read = __mctp_astlpc_fileio_kcs_read;
astlpc->ops.kcs_write = __mctp_astlpc_fileio_kcs_write;
astlpc->ops_data = astlpc;
rc = mctp_astlpc_init_fileio_lpc(astlpc);
if (rc) {
free(astlpc);
return NULL;
}
rc = mctp_astlpc_init_fileio_kcs(astlpc);
if (rc) {
free(astlpc);
return NULL;
}
return astlpc;
}
#else
struct mctp_binding_astlpc * __attribute__((const))
mctp_astlpc_init_fileio(void)
{
astlpc_prerr(astlpc, "Missing support for file IO");
return NULL;
}
int __attribute__((const)) mctp_astlpc_get_fd(
struct mctp_binding_astlpc *astlpc __attribute__((unused)))
{
astlpc_prerr(astlpc, "Missing support for file IO");
return -1;
}
#endif