linux/drivers/usb/gadget/function/f_tcm.c
Chen Yufeng 1538740103 usb: potential integer overflow in usbg_make_tpg()
The variable tpgt in usbg_make_tpg() is defined as unsigned long and is
assigned to tpgt->tport_tpgt, which is defined as u16. This may cause an
integer overflow when tpgt is greater than USHRT_MAX (65535). I
haven't tried to trigger it myself, but it is possible to trigger it
by calling usbg_make_tpg() with a large value for tpgt.

I modified the type of tpgt to match tpgt->tport_tpgt and adjusted the
relevant code accordingly.

This patch is similar to commit 59c816c1f2 ("vhost/scsi: potential
memory corruption").

Signed-off-by: Chen Yufeng <chenyufeng@iie.ac.cn>
Link: https://lore.kernel.org/r/20250415065857.1619-1-chenyufeng@iie.ac.cn
Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
2025-04-15 14:29:33 +02:00

2641 lines
64 KiB
C

// SPDX-License-Identifier: GPL-2.0
/* Target based USB-Gadget
*
* UAS protocol handling, target callbacks, configfs handling,
* BBB (USB Mass Storage Class Bulk-Only (BBB) and Transport protocol handling.
*
* Author: Sebastian Andrzej Siewior <bigeasy at linutronix dot de>
*/
#include <linux/kernel.h>
#include <linux/module.h>
#include <linux/types.h>
#include <linux/string.h>
#include <linux/configfs.h>
#include <linux/ctype.h>
#include <linux/delay.h>
#include <linux/usb/ch9.h>
#include <linux/usb/composite.h>
#include <linux/usb/gadget.h>
#include <linux/usb/storage.h>
#include <scsi/scsi_tcq.h>
#include <target/target_core_base.h>
#include <target/target_core_fabric.h>
#include <linux/unaligned.h>
#include "tcm.h"
#include "u_tcm.h"
#include "configfs.h"
#define TPG_INSTANCES 1
struct tpg_instance {
struct usb_function_instance *func_inst;
struct usbg_tpg *tpg;
};
static struct tpg_instance tpg_instances[TPG_INSTANCES];
static DEFINE_MUTEX(tpg_instances_lock);
static inline struct f_uas *to_f_uas(struct usb_function *f)
{
return container_of(f, struct f_uas, function);
}
/* Start bot.c code */
static int bot_enqueue_cmd_cbw(struct f_uas *fu)
{
int ret;
if (fu->flags & USBG_BOT_CMD_PEND)
return 0;
ret = usb_ep_queue(fu->ep_out, fu->cmd[0].req, GFP_ATOMIC);
if (!ret)
fu->flags |= USBG_BOT_CMD_PEND;
return ret;
}
static void bot_status_complete(struct usb_ep *ep, struct usb_request *req)
{
struct usbg_cmd *cmd = req->context;
struct f_uas *fu = cmd->fu;
transport_generic_free_cmd(&cmd->se_cmd, 0);
if (req->status == -ESHUTDOWN)
return;
if (req->status < 0)
pr_err("ERR %s(%d)\n", __func__, __LINE__);
/* CSW completed, wait for next CBW */
bot_enqueue_cmd_cbw(fu);
}
static void bot_enqueue_sense_code(struct f_uas *fu, struct usbg_cmd *cmd)
{
struct bulk_cs_wrap *csw = &fu->bot_status.csw;
int ret;
unsigned int csw_stat;
csw_stat = cmd->csw_code;
csw->Tag = cmd->bot_tag;
csw->Status = csw_stat;
fu->bot_status.req->context = cmd;
ret = usb_ep_queue(fu->ep_in, fu->bot_status.req, GFP_ATOMIC);
if (ret)
pr_err("%s(%d) ERR: %d\n", __func__, __LINE__, ret);
}
static void bot_err_compl(struct usb_ep *ep, struct usb_request *req)
{
struct usbg_cmd *cmd = req->context;
struct f_uas *fu = cmd->fu;
if (req->status < 0)
pr_err("ERR %s(%d)\n", __func__, __LINE__);
if (cmd->data_len) {
if (cmd->data_len > ep->maxpacket) {
req->length = ep->maxpacket;
cmd->data_len -= ep->maxpacket;
} else {
req->length = cmd->data_len;
cmd->data_len = 0;
}
usb_ep_queue(ep, req, GFP_ATOMIC);
return;
}
bot_enqueue_sense_code(fu, cmd);
}
static void bot_send_bad_status(struct usbg_cmd *cmd)
{
struct f_uas *fu = cmd->fu;
struct bulk_cs_wrap *csw = &fu->bot_status.csw;
struct usb_request *req;
struct usb_ep *ep;
csw->Residue = cpu_to_le32(cmd->data_len);
if (cmd->data_len) {
if (cmd->is_read) {
ep = fu->ep_in;
req = fu->bot_req_in;
} else {
ep = fu->ep_out;
req = fu->bot_req_out;
}
if (cmd->data_len > fu->ep_in->maxpacket) {
req->length = ep->maxpacket;
cmd->data_len -= ep->maxpacket;
} else {
req->length = cmd->data_len;
cmd->data_len = 0;
}
req->complete = bot_err_compl;
req->context = cmd;
req->buf = fu->cmd[0].buf;
usb_ep_queue(ep, req, GFP_KERNEL);
} else {
bot_enqueue_sense_code(fu, cmd);
}
}
static int bot_send_status(struct usbg_cmd *cmd, bool moved_data)
{
struct f_uas *fu = cmd->fu;
struct bulk_cs_wrap *csw = &fu->bot_status.csw;
int ret;
if (cmd->se_cmd.scsi_status == SAM_STAT_GOOD) {
if (!moved_data && cmd->data_len) {
/*
* the host wants to move data, we don't. Fill / empty
* the pipe and then send the csw with reside set.
*/
cmd->csw_code = US_BULK_STAT_OK;
bot_send_bad_status(cmd);
return 0;
}
csw->Tag = cmd->bot_tag;
csw->Residue = cpu_to_le32(0);
csw->Status = US_BULK_STAT_OK;
fu->bot_status.req->context = cmd;
ret = usb_ep_queue(fu->ep_in, fu->bot_status.req, GFP_KERNEL);
if (ret)
pr_err("%s(%d) ERR: %d\n", __func__, __LINE__, ret);
} else {
cmd->csw_code = US_BULK_STAT_FAIL;
bot_send_bad_status(cmd);
}
return 0;
}
/*
* Called after command (no data transfer) or after the write (to device)
* operation is completed
*/
static int bot_send_status_response(struct usbg_cmd *cmd)
{
bool moved_data = false;
if (!cmd->is_read)
moved_data = true;
return bot_send_status(cmd, moved_data);
}
/* Read request completed, now we have to send the CSW */
static void bot_read_compl(struct usb_ep *ep, struct usb_request *req)
{
struct usbg_cmd *cmd = req->context;
if (req->status < 0)
pr_err("ERR %s(%d)\n", __func__, __LINE__);
if (req->status == -ESHUTDOWN) {
transport_generic_free_cmd(&cmd->se_cmd, 0);
return;
}
bot_send_status(cmd, true);
}
static int bot_send_read_response(struct usbg_cmd *cmd)
{
struct f_uas *fu = cmd->fu;
struct se_cmd *se_cmd = &cmd->se_cmd;
struct usb_gadget *gadget = fuas_to_gadget(fu);
int ret;
if (!cmd->data_len) {
cmd->csw_code = US_BULK_STAT_PHASE;
bot_send_bad_status(cmd);
return 0;
}
if (!gadget->sg_supported) {
cmd->data_buf = kmalloc(se_cmd->data_length, GFP_ATOMIC);
if (!cmd->data_buf)
return -ENOMEM;
sg_copy_to_buffer(se_cmd->t_data_sg,
se_cmd->t_data_nents,
cmd->data_buf,
se_cmd->data_length);
fu->bot_req_in->buf = cmd->data_buf;
} else {
fu->bot_req_in->buf = NULL;
fu->bot_req_in->num_sgs = se_cmd->t_data_nents;
fu->bot_req_in->sg = se_cmd->t_data_sg;
}
fu->bot_req_in->complete = bot_read_compl;
fu->bot_req_in->length = se_cmd->data_length;
fu->bot_req_in->context = cmd;
ret = usb_ep_queue(fu->ep_in, fu->bot_req_in, GFP_ATOMIC);
if (ret)
pr_err("%s(%d)\n", __func__, __LINE__);
return 0;
}
static void usbg_data_write_cmpl(struct usb_ep *, struct usb_request *);
static int usbg_prepare_w_request(struct usbg_cmd *, struct usb_request *);
static int bot_send_write_request(struct usbg_cmd *cmd)
{
struct f_uas *fu = cmd->fu;
int ret;
cmd->fu = fu;
if (!cmd->data_len) {
cmd->csw_code = US_BULK_STAT_PHASE;
return -EINVAL;
}
ret = usbg_prepare_w_request(cmd, fu->bot_req_out);
if (ret)
goto cleanup;
ret = usb_ep_queue(fu->ep_out, fu->bot_req_out, GFP_KERNEL);
if (ret)
pr_err("%s(%d)\n", __func__, __LINE__);
cleanup:
return ret;
}
static int bot_submit_command(struct f_uas *, void *, unsigned int);
static void bot_cmd_complete(struct usb_ep *ep, struct usb_request *req)
{
struct f_uas *fu = req->context;
int ret;
if (req->status == -ESHUTDOWN)
return;
fu->flags &= ~USBG_BOT_CMD_PEND;
if (req->status < 0) {
struct usb_gadget *gadget = fuas_to_gadget(fu);
dev_err(&gadget->dev, "BOT command req err (%d)\n", req->status);
bot_enqueue_cmd_cbw(fu);
return;
}
ret = bot_submit_command(fu, req->buf, req->actual);
if (ret) {
pr_err("%s(%d): %d\n", __func__, __LINE__, ret);
if (!(fu->flags & USBG_BOT_WEDGED))
usb_ep_set_wedge(fu->ep_in);
fu->flags |= USBG_BOT_WEDGED;
bot_enqueue_cmd_cbw(fu);
} else if (fu->flags & USBG_BOT_WEDGED) {
fu->flags &= ~USBG_BOT_WEDGED;
usb_ep_clear_halt(fu->ep_in);
}
}
static int bot_prepare_reqs(struct f_uas *fu)
{
int ret;
fu->bot_req_in = usb_ep_alloc_request(fu->ep_in, GFP_KERNEL);
if (!fu->bot_req_in)
goto err;
fu->bot_req_out = usb_ep_alloc_request(fu->ep_out, GFP_KERNEL);
if (!fu->bot_req_out)
goto err_out;
fu->cmd[0].req = usb_ep_alloc_request(fu->ep_out, GFP_KERNEL);
if (!fu->cmd[0].req)
goto err_cmd;
fu->bot_status.req = usb_ep_alloc_request(fu->ep_in, GFP_KERNEL);
if (!fu->bot_status.req)
goto err_sts;
fu->bot_status.req->buf = &fu->bot_status.csw;
fu->bot_status.req->length = US_BULK_CS_WRAP_LEN;
fu->bot_status.req->complete = bot_status_complete;
fu->bot_status.csw.Signature = cpu_to_le32(US_BULK_CS_SIGN);
fu->cmd[0].buf = kmalloc(fu->ep_out->maxpacket, GFP_KERNEL);
if (!fu->cmd[0].buf)
goto err_buf;
fu->cmd[0].req->complete = bot_cmd_complete;
fu->cmd[0].req->buf = fu->cmd[0].buf;
fu->cmd[0].req->length = fu->ep_out->maxpacket;
fu->cmd[0].req->context = fu;
ret = bot_enqueue_cmd_cbw(fu);
if (ret)
goto err_queue;
return 0;
err_queue:
kfree(fu->cmd[0].buf);
fu->cmd[0].buf = NULL;
err_buf:
usb_ep_free_request(fu->ep_in, fu->bot_status.req);
err_sts:
usb_ep_free_request(fu->ep_out, fu->cmd[0].req);
fu->cmd[0].req = NULL;
err_cmd:
usb_ep_free_request(fu->ep_out, fu->bot_req_out);
fu->bot_req_out = NULL;
err_out:
usb_ep_free_request(fu->ep_in, fu->bot_req_in);
fu->bot_req_in = NULL;
err:
pr_err("BOT: endpoint setup failed\n");
return -ENOMEM;
}
static void bot_cleanup_old_alt(struct f_uas *fu)
{
if (!(fu->flags & USBG_ENABLED))
return;
usb_ep_disable(fu->ep_in);
usb_ep_disable(fu->ep_out);
if (!fu->bot_req_in)
return;
usb_ep_free_request(fu->ep_in, fu->bot_req_in);
usb_ep_free_request(fu->ep_out, fu->bot_req_out);
usb_ep_free_request(fu->ep_out, fu->cmd[0].req);
usb_ep_free_request(fu->ep_in, fu->bot_status.req);
kfree(fu->cmd[0].buf);
fu->bot_req_in = NULL;
fu->bot_req_out = NULL;
fu->cmd[0].req = NULL;
fu->bot_status.req = NULL;
fu->cmd[0].buf = NULL;
}
static void bot_set_alt(struct f_uas *fu)
{
struct usb_function *f = &fu->function;
struct usb_gadget *gadget = f->config->cdev->gadget;
int ret;
fu->flags = USBG_IS_BOT;
config_ep_by_speed_and_alt(gadget, f, fu->ep_in, USB_G_ALT_INT_BBB);
ret = usb_ep_enable(fu->ep_in);
if (ret)
goto err_b_in;
config_ep_by_speed_and_alt(gadget, f, fu->ep_out, USB_G_ALT_INT_BBB);
ret = usb_ep_enable(fu->ep_out);
if (ret)
goto err_b_out;
ret = bot_prepare_reqs(fu);
if (ret)
goto err_wq;
fu->flags |= USBG_ENABLED;
pr_info("Using the BOT protocol\n");
return;
err_wq:
usb_ep_disable(fu->ep_out);
err_b_out:
usb_ep_disable(fu->ep_in);
err_b_in:
fu->flags = USBG_IS_BOT;
}
static int usbg_bot_setup(struct usb_function *f,
const struct usb_ctrlrequest *ctrl)
{
struct f_uas *fu = to_f_uas(f);
struct usb_composite_dev *cdev = f->config->cdev;
u16 w_value = le16_to_cpu(ctrl->wValue);
u16 w_length = le16_to_cpu(ctrl->wLength);
int luns;
u8 *ret_lun;
switch (ctrl->bRequest) {
case US_BULK_GET_MAX_LUN:
if (ctrl->bRequestType != (USB_DIR_IN | USB_TYPE_CLASS |
USB_RECIP_INTERFACE))
return -ENOTSUPP;
if (w_length < 1)
return -EINVAL;
if (w_value != 0)
return -EINVAL;
luns = atomic_read(&fu->tpg->tpg_port_count);
if (!luns) {
pr_err("No LUNs configured?\n");
return -EINVAL;
}
luns--;
if (luns > US_BULK_MAX_LUN_LIMIT) {
pr_info_once("Limiting the number of luns to 16\n");
luns = US_BULK_MAX_LUN_LIMIT;
}
ret_lun = cdev->req->buf;
*ret_lun = luns;
cdev->req->length = 1;
return usb_ep_queue(cdev->gadget->ep0, cdev->req, GFP_ATOMIC);
case US_BULK_RESET_REQUEST:
/* XXX maybe we should remove previous requests for IN + OUT */
if (fu->flags & USBG_BOT_WEDGED) {
fu->flags &= ~USBG_BOT_WEDGED;
usb_ep_clear_halt(fu->ep_in);
}
bot_enqueue_cmd_cbw(fu);
return 0;
}
return -ENOTSUPP;
}
/* Start uas.c code */
static int tcm_to_uasp_response(enum tcm_tmrsp_table code)
{
switch (code) {
case TMR_FUNCTION_FAILED:
return RC_TMF_FAILED;
case TMR_FUNCTION_COMPLETE:
case TMR_TASK_DOES_NOT_EXIST:
return RC_TMF_COMPLETE;
case TMR_LUN_DOES_NOT_EXIST:
return RC_INCORRECT_LUN;
case TMR_FUNCTION_REJECTED:
case TMR_TASK_MGMT_FUNCTION_NOT_SUPPORTED:
default:
return RC_TMF_NOT_SUPPORTED;
}
}
static unsigned char uasp_to_tcm_func(int code)
{
switch (code) {
case TMF_ABORT_TASK:
return TMR_ABORT_TASK;
case TMF_ABORT_TASK_SET:
return TMR_ABORT_TASK_SET;
case TMF_CLEAR_TASK_SET:
return TMR_CLEAR_TASK_SET;
case TMF_LOGICAL_UNIT_RESET:
return TMR_LUN_RESET;
case TMF_CLEAR_ACA:
return TMR_CLEAR_ACA;
case TMF_I_T_NEXUS_RESET:
case TMF_QUERY_TASK:
case TMF_QUERY_TASK_SET:
case TMF_QUERY_ASYNC_EVENT:
default:
return TMR_UNKNOWN;
}
}
static void uasp_cleanup_one_stream(struct f_uas *fu, struct uas_stream *stream)
{
/* We have either all three allocated or none */
if (!stream->req_in)
return;
usb_ep_free_request(fu->ep_in, stream->req_in);
usb_ep_free_request(fu->ep_out, stream->req_out);
usb_ep_free_request(fu->ep_status, stream->req_status);
stream->req_in = NULL;
stream->req_out = NULL;
stream->req_status = NULL;
}
static void uasp_free_cmdreq(struct f_uas *fu)
{
int i;
for (i = 0; i < USBG_NUM_CMDS; i++) {
usb_ep_free_request(fu->ep_cmd, fu->cmd[i].req);
kfree(fu->cmd[i].buf);
fu->cmd[i].req = NULL;
fu->cmd[i].buf = NULL;
}
}
static void uasp_cleanup_old_alt(struct f_uas *fu)
{
int i;
if (!(fu->flags & USBG_ENABLED))
return;
usb_ep_disable(fu->ep_in);
usb_ep_disable(fu->ep_out);
usb_ep_disable(fu->ep_status);
usb_ep_disable(fu->ep_cmd);
for (i = 0; i < USBG_NUM_CMDS; i++)
uasp_cleanup_one_stream(fu, &fu->stream[i]);
uasp_free_cmdreq(fu);
}
static void uasp_status_data_cmpl(struct usb_ep *ep, struct usb_request *req);
static int uasp_prepare_r_request(struct usbg_cmd *cmd)
{
struct se_cmd *se_cmd = &cmd->se_cmd;
struct f_uas *fu = cmd->fu;
struct usb_gadget *gadget = fuas_to_gadget(fu);
struct uas_stream *stream = &fu->stream[se_cmd->map_tag];
if (!gadget->sg_supported) {
cmd->data_buf = kmalloc(se_cmd->data_length, GFP_ATOMIC);
if (!cmd->data_buf)
return -ENOMEM;
sg_copy_to_buffer(se_cmd->t_data_sg,
se_cmd->t_data_nents,
cmd->data_buf,
se_cmd->data_length);
stream->req_in->buf = cmd->data_buf;
} else {
stream->req_in->buf = NULL;
stream->req_in->num_sgs = se_cmd->t_data_nents;
stream->req_in->sg = se_cmd->t_data_sg;
}
stream->req_in->is_last = 1;
stream->req_in->stream_id = cmd->tag;
stream->req_in->complete = uasp_status_data_cmpl;
stream->req_in->length = se_cmd->data_length;
stream->req_in->context = cmd;
cmd->state = UASP_SEND_STATUS;
return 0;
}
static void uasp_prepare_status(struct usbg_cmd *cmd)
{
struct se_cmd *se_cmd = &cmd->se_cmd;
struct sense_iu *iu = &cmd->sense_iu;
struct uas_stream *stream = &cmd->fu->stream[se_cmd->map_tag];
cmd->state = UASP_QUEUE_COMMAND;
iu->iu_id = IU_ID_STATUS;
iu->tag = cpu_to_be16(cmd->tag);
/*
* iu->status_qual = cpu_to_be16(STATUS QUALIFIER SAM-4. Where R U?);
*/
iu->len = cpu_to_be16(se_cmd->scsi_sense_length);
iu->status = se_cmd->scsi_status;
stream->req_status->is_last = 1;
stream->req_status->stream_id = cmd->tag;
stream->req_status->context = cmd;
stream->req_status->length = se_cmd->scsi_sense_length + 16;
stream->req_status->buf = iu;
stream->req_status->complete = uasp_status_data_cmpl;
}
static void uasp_prepare_response(struct usbg_cmd *cmd)
{
struct se_cmd *se_cmd = &cmd->se_cmd;
struct response_iu *rsp_iu = &cmd->response_iu;
struct uas_stream *stream = &cmd->fu->stream[se_cmd->map_tag];
cmd->state = UASP_QUEUE_COMMAND;
rsp_iu->iu_id = IU_ID_RESPONSE;
rsp_iu->tag = cpu_to_be16(cmd->tag);
if (cmd->tmr_rsp != RC_RESPONSE_UNKNOWN)
rsp_iu->response_code = cmd->tmr_rsp;
else
rsp_iu->response_code =
tcm_to_uasp_response(se_cmd->se_tmr_req->response);
/*
* The UASP driver must support all the task management functions listed
* in Table 20 of UAS-r04. To remain compliant while indicate that the
* TMR did not go through, report RC_TMF_FAILED instead of
* RC_TMF_NOT_SUPPORTED and print a warning to the user.
*/
switch (cmd->tmr_func) {
case TMF_ABORT_TASK:
case TMF_ABORT_TASK_SET:
case TMF_CLEAR_TASK_SET:
case TMF_LOGICAL_UNIT_RESET:
case TMF_CLEAR_ACA:
case TMF_I_T_NEXUS_RESET:
case TMF_QUERY_TASK:
case TMF_QUERY_TASK_SET:
case TMF_QUERY_ASYNC_EVENT:
if (rsp_iu->response_code == RC_TMF_NOT_SUPPORTED) {
struct usb_gadget *gadget = fuas_to_gadget(cmd->fu);
dev_warn(&gadget->dev, "TMF function %d not supported\n",
cmd->tmr_func);
rsp_iu->response_code = RC_TMF_FAILED;
}
break;
default:
break;
}
stream->req_status->is_last = 1;
stream->req_status->stream_id = cmd->tag;
stream->req_status->context = cmd;
stream->req_status->length = sizeof(struct response_iu);
stream->req_status->buf = rsp_iu;
stream->req_status->complete = uasp_status_data_cmpl;
}
static void usbg_release_cmd(struct se_cmd *se_cmd);
static int uasp_send_tm_response(struct usbg_cmd *cmd);
static void uasp_status_data_cmpl(struct usb_ep *ep, struct usb_request *req)
{
struct usbg_cmd *cmd = req->context;
struct f_uas *fu = cmd->fu;
struct uas_stream *stream = &fu->stream[cmd->se_cmd.map_tag];
int ret;
if (req->status == -ESHUTDOWN)
goto cleanup;
switch (cmd->state) {
case UASP_SEND_DATA:
ret = uasp_prepare_r_request(cmd);
if (ret)
goto cleanup;
ret = usb_ep_queue(fu->ep_in, stream->req_in, GFP_ATOMIC);
if (ret)
pr_err("%s(%d) => %d\n", __func__, __LINE__, ret);
break;
case UASP_RECEIVE_DATA:
ret = usbg_prepare_w_request(cmd, stream->req_out);
if (ret)
goto cleanup;
ret = usb_ep_queue(fu->ep_out, stream->req_out, GFP_ATOMIC);
if (ret)
pr_err("%s(%d) => %d\n", __func__, __LINE__, ret);
break;
case UASP_SEND_STATUS:
uasp_prepare_status(cmd);
ret = usb_ep_queue(fu->ep_status, stream->req_status,
GFP_ATOMIC);
if (ret)
pr_err("%s(%d) => %d\n", __func__, __LINE__, ret);
break;
case UASP_QUEUE_COMMAND:
/*
* Overlapped command detected and cancelled.
* So send overlapped attempted status.
*/
if (cmd->tmr_rsp == RC_OVERLAPPED_TAG &&
req->status == -ECONNRESET) {
uasp_send_tm_response(cmd);
return;
}
hash_del(&stream->node);
/*
* If no command submitted to target core here, just free the
* bitmap index. This is for the cases where f_tcm handles
* status response instead of the target core.
*/
if (cmd->tmr_rsp != RC_OVERLAPPED_TAG &&
cmd->tmr_rsp != RC_RESPONSE_UNKNOWN) {
struct se_session *se_sess;
se_sess = fu->tpg->tpg_nexus->tvn_se_sess;
sbitmap_queue_clear(&se_sess->sess_tag_pool,
cmd->se_cmd.map_tag,
cmd->se_cmd.map_cpu);
} else {
transport_generic_free_cmd(&cmd->se_cmd, 0);
}
usb_ep_queue(fu->ep_cmd, cmd->req, GFP_ATOMIC);
complete(&stream->cmd_completion);
break;
default:
BUG();
}
return;
cleanup:
hash_del(&stream->node);
transport_generic_free_cmd(&cmd->se_cmd, 0);
}
static int uasp_send_status_response(struct usbg_cmd *cmd)
{
struct f_uas *fu = cmd->fu;
struct uas_stream *stream = &fu->stream[cmd->se_cmd.map_tag];
struct sense_iu *iu = &cmd->sense_iu;
iu->tag = cpu_to_be16(cmd->tag);
cmd->fu = fu;
uasp_prepare_status(cmd);
return usb_ep_queue(fu->ep_status, stream->req_status, GFP_ATOMIC);
}
static int uasp_send_tm_response(struct usbg_cmd *cmd)
{
struct f_uas *fu = cmd->fu;
struct uas_stream *stream = &fu->stream[cmd->se_cmd.map_tag];
struct response_iu *iu = &cmd->response_iu;
iu->tag = cpu_to_be16(cmd->tag);
cmd->fu = fu;
uasp_prepare_response(cmd);
return usb_ep_queue(fu->ep_status, stream->req_status, GFP_ATOMIC);
}
static int uasp_send_read_response(struct usbg_cmd *cmd)
{
struct f_uas *fu = cmd->fu;
struct uas_stream *stream = &fu->stream[cmd->se_cmd.map_tag];
struct sense_iu *iu = &cmd->sense_iu;
int ret;
cmd->fu = fu;
iu->tag = cpu_to_be16(cmd->tag);
if (fu->flags & USBG_USE_STREAMS) {
ret = uasp_prepare_r_request(cmd);
if (ret)
goto out;
ret = usb_ep_queue(fu->ep_in, stream->req_in, GFP_ATOMIC);
if (ret) {
pr_err("%s(%d) => %d\n", __func__, __LINE__, ret);
kfree(cmd->data_buf);
cmd->data_buf = NULL;
}
} else {
iu->iu_id = IU_ID_READ_READY;
iu->tag = cpu_to_be16(cmd->tag);
stream->req_status->complete = uasp_status_data_cmpl;
stream->req_status->context = cmd;
cmd->state = UASP_SEND_DATA;
stream->req_status->buf = iu;
stream->req_status->length = sizeof(struct iu);
ret = usb_ep_queue(fu->ep_status, stream->req_status,
GFP_ATOMIC);
if (ret)
pr_err("%s(%d) => %d\n", __func__, __LINE__, ret);
}
out:
return ret;
}
static int uasp_send_write_request(struct usbg_cmd *cmd)
{
struct f_uas *fu = cmd->fu;
struct se_cmd *se_cmd = &cmd->se_cmd;
struct uas_stream *stream = &fu->stream[se_cmd->map_tag];
struct sense_iu *iu = &cmd->sense_iu;
int ret;
cmd->fu = fu;
iu->tag = cpu_to_be16(cmd->tag);
if (fu->flags & USBG_USE_STREAMS) {
ret = usbg_prepare_w_request(cmd, stream->req_out);
if (ret)
goto cleanup;
ret = usb_ep_queue(fu->ep_out, stream->req_out, GFP_ATOMIC);
if (ret)
pr_err("%s(%d)\n", __func__, __LINE__);
} else {
iu->iu_id = IU_ID_WRITE_READY;
iu->tag = cpu_to_be16(cmd->tag);
stream->req_status->complete = uasp_status_data_cmpl;
stream->req_status->context = cmd;
cmd->state = UASP_RECEIVE_DATA;
stream->req_status->buf = iu;
stream->req_status->length = sizeof(struct iu);
ret = usb_ep_queue(fu->ep_status, stream->req_status,
GFP_ATOMIC);
if (ret)
pr_err("%s(%d)\n", __func__, __LINE__);
}
cleanup:
return ret;
}
static int usbg_submit_command(struct f_uas *, struct usb_request *);
static void uasp_cmd_complete(struct usb_ep *ep, struct usb_request *req)
{
struct f_uas *fu = req->context;
if (req->status == -ESHUTDOWN)
return;
if (req->status < 0) {
usb_ep_queue(fu->ep_cmd, req, GFP_ATOMIC);
return;
}
usbg_submit_command(fu, req);
}
static int uasp_alloc_stream_res(struct f_uas *fu, struct uas_stream *stream)
{
init_completion(&stream->cmd_completion);
stream->req_in = usb_ep_alloc_request(fu->ep_in, GFP_KERNEL);
if (!stream->req_in)
goto out;
stream->req_out = usb_ep_alloc_request(fu->ep_out, GFP_KERNEL);
if (!stream->req_out)
goto err_out;
stream->req_status = usb_ep_alloc_request(fu->ep_status, GFP_KERNEL);
if (!stream->req_status)
goto err_sts;
return 0;
err_sts:
usb_ep_free_request(fu->ep_out, stream->req_out);
stream->req_out = NULL;
err_out:
usb_ep_free_request(fu->ep_in, stream->req_in);
stream->req_in = NULL;
out:
return -ENOMEM;
}
static int uasp_alloc_cmd(struct f_uas *fu, int i)
{
fu->cmd[i].req = usb_ep_alloc_request(fu->ep_cmd, GFP_KERNEL);
if (!fu->cmd[i].req)
goto err;
fu->cmd[i].buf = kmalloc(fu->ep_cmd->maxpacket, GFP_KERNEL);
if (!fu->cmd[i].buf)
goto err_buf;
fu->cmd[i].req->complete = uasp_cmd_complete;
fu->cmd[i].req->buf = fu->cmd[i].buf;
fu->cmd[i].req->length = fu->ep_cmd->maxpacket;
fu->cmd[i].req->context = fu;
return 0;
err_buf:
usb_ep_free_request(fu->ep_cmd, fu->cmd[i].req);
err:
return -ENOMEM;
}
static int uasp_prepare_reqs(struct f_uas *fu)
{
int ret;
int i;
for (i = 0; i < USBG_NUM_CMDS; i++) {
ret = uasp_alloc_stream_res(fu, &fu->stream[i]);
if (ret)
goto err_cleanup;
}
for (i = 0; i < USBG_NUM_CMDS; i++) {
ret = uasp_alloc_cmd(fu, i);
if (ret)
goto err_free_stream;
ret = usb_ep_queue(fu->ep_cmd, fu->cmd[i].req, GFP_ATOMIC);
if (ret)
goto err_free_stream;
}
return 0;
err_free_stream:
uasp_free_cmdreq(fu);
err_cleanup:
if (i) {
do {
uasp_cleanup_one_stream(fu, &fu->stream[i - 1]);
i--;
} while (i);
}
pr_err("UASP: endpoint setup failed\n");
return ret;
}
static void uasp_set_alt(struct f_uas *fu)
{
struct usb_function *f = &fu->function;
struct usb_gadget *gadget = f->config->cdev->gadget;
int ret;
fu->flags = USBG_IS_UAS;
if (gadget->speed >= USB_SPEED_SUPER)
fu->flags |= USBG_USE_STREAMS;
config_ep_by_speed_and_alt(gadget, f, fu->ep_in, USB_G_ALT_INT_UAS);
ret = usb_ep_enable(fu->ep_in);
if (ret)
goto err_b_in;
config_ep_by_speed_and_alt(gadget, f, fu->ep_out, USB_G_ALT_INT_UAS);
ret = usb_ep_enable(fu->ep_out);
if (ret)
goto err_b_out;
config_ep_by_speed_and_alt(gadget, f, fu->ep_cmd, USB_G_ALT_INT_UAS);
ret = usb_ep_enable(fu->ep_cmd);
if (ret)
goto err_cmd;
config_ep_by_speed_and_alt(gadget, f, fu->ep_status, USB_G_ALT_INT_UAS);
ret = usb_ep_enable(fu->ep_status);
if (ret)
goto err_status;
ret = uasp_prepare_reqs(fu);
if (ret)
goto err_wq;
fu->flags |= USBG_ENABLED;
pr_info("Using the UAS protocol\n");
return;
err_wq:
usb_ep_disable(fu->ep_status);
err_status:
usb_ep_disable(fu->ep_cmd);
err_cmd:
usb_ep_disable(fu->ep_out);
err_b_out:
usb_ep_disable(fu->ep_in);
err_b_in:
fu->flags = 0;
}
static int get_cmd_dir(const unsigned char *cdb)
{
int ret;
switch (cdb[0]) {
case READ_6:
case READ_10:
case READ_12:
case READ_16:
case INQUIRY:
case MODE_SENSE:
case MODE_SENSE_10:
case SERVICE_ACTION_IN_16:
case MAINTENANCE_IN:
case PERSISTENT_RESERVE_IN:
case SECURITY_PROTOCOL_IN:
case ACCESS_CONTROL_IN:
case REPORT_LUNS:
case READ_BLOCK_LIMITS:
case READ_POSITION:
case READ_CAPACITY:
case READ_TOC:
case READ_FORMAT_CAPACITIES:
case REQUEST_SENSE:
case ATA_12:
case ATA_16:
ret = DMA_FROM_DEVICE;
break;
case WRITE_6:
case WRITE_10:
case WRITE_12:
case WRITE_16:
case MODE_SELECT:
case MODE_SELECT_10:
case WRITE_VERIFY:
case WRITE_VERIFY_12:
case PERSISTENT_RESERVE_OUT:
case MAINTENANCE_OUT:
case SECURITY_PROTOCOL_OUT:
case ACCESS_CONTROL_OUT:
ret = DMA_TO_DEVICE;
break;
case ALLOW_MEDIUM_REMOVAL:
case TEST_UNIT_READY:
case SYNCHRONIZE_CACHE:
case START_STOP:
case ERASE:
case REZERO_UNIT:
case SEEK_10:
case SPACE:
case VERIFY:
case WRITE_FILEMARKS:
ret = DMA_NONE;
break;
default:
#define CMD_DIR_MSG "target: Unknown data direction for SCSI Opcode 0x%02x\n"
pr_warn(CMD_DIR_MSG, cdb[0]);
#undef CMD_DIR_MSG
ret = -EINVAL;
}
return ret;
}
static void usbg_data_write_cmpl(struct usb_ep *ep, struct usb_request *req)
{
struct usbg_cmd *cmd = req->context;
struct se_cmd *se_cmd = &cmd->se_cmd;
cmd->state = UASP_QUEUE_COMMAND;
if (req->status == -ESHUTDOWN) {
struct uas_stream *stream = &cmd->fu->stream[se_cmd->map_tag];
hash_del(&stream->node);
target_put_sess_cmd(se_cmd);
transport_generic_free_cmd(&cmd->se_cmd, 0);
return;
}
if (req->status) {
pr_err("%s() state %d transfer failed\n", __func__, cmd->state);
goto cleanup;
}
if (req->num_sgs == 0) {
sg_copy_from_buffer(se_cmd->t_data_sg,
se_cmd->t_data_nents,
cmd->data_buf,
se_cmd->data_length);
}
cmd->flags |= USBG_CMD_PENDING_DATA_WRITE;
queue_work(cmd->fu->tpg->workqueue, &cmd->work);
return;
cleanup:
target_put_sess_cmd(se_cmd);
/* Command was aborted due to overlapped tag */
if (cmd->state == UASP_QUEUE_COMMAND &&
cmd->tmr_rsp == RC_OVERLAPPED_TAG) {
uasp_send_tm_response(cmd);
return;
}
transport_send_check_condition_and_sense(se_cmd,
TCM_CHECK_CONDITION_ABORT_CMD, 0);
}
static int usbg_prepare_w_request(struct usbg_cmd *cmd, struct usb_request *req)
{
struct se_cmd *se_cmd = &cmd->se_cmd;
struct f_uas *fu = cmd->fu;
struct usb_gadget *gadget = fuas_to_gadget(fu);
if (!gadget->sg_supported) {
cmd->data_buf = kmalloc(se_cmd->data_length, GFP_ATOMIC);
if (!cmd->data_buf)
return -ENOMEM;
req->buf = cmd->data_buf;
} else {
req->buf = NULL;
req->num_sgs = se_cmd->t_data_nents;
req->sg = se_cmd->t_data_sg;
}
req->is_last = 1;
req->stream_id = cmd->tag;
req->complete = usbg_data_write_cmpl;
req->length = se_cmd->data_length;
req->context = cmd;
cmd->state = UASP_SEND_STATUS;
return 0;
}
static int usbg_send_status_response(struct se_cmd *se_cmd)
{
struct usbg_cmd *cmd = container_of(se_cmd, struct usbg_cmd,
se_cmd);
struct f_uas *fu = cmd->fu;
if (fu->flags & USBG_IS_BOT)
return bot_send_status_response(cmd);
else
return uasp_send_status_response(cmd);
}
static int usbg_send_write_request(struct se_cmd *se_cmd)
{
struct usbg_cmd *cmd = container_of(se_cmd, struct usbg_cmd,
se_cmd);
struct f_uas *fu = cmd->fu;
if (fu->flags & USBG_IS_BOT)
return bot_send_write_request(cmd);
else
return uasp_send_write_request(cmd);
}
static int usbg_send_read_response(struct se_cmd *se_cmd)
{
struct usbg_cmd *cmd = container_of(se_cmd, struct usbg_cmd,
se_cmd);
struct f_uas *fu = cmd->fu;
if (fu->flags & USBG_IS_BOT)
return bot_send_read_response(cmd);
else
return uasp_send_read_response(cmd);
}
static void usbg_aborted_task(struct se_cmd *se_cmd);
static void usbg_submit_tmr(struct usbg_cmd *cmd)
{
struct se_session *se_sess;
struct se_cmd *se_cmd;
int flags = TARGET_SCF_ACK_KREF;
se_cmd = &cmd->se_cmd;
se_sess = cmd->fu->tpg->tpg_nexus->tvn_se_sess;
target_submit_tmr(se_cmd, se_sess,
cmd->response_iu.add_response_info,
cmd->unpacked_lun, NULL, uasp_to_tcm_func(cmd->tmr_func),
GFP_ATOMIC, cmd->tag, flags);
}
static void usbg_submit_cmd(struct usbg_cmd *cmd)
{
struct se_cmd *se_cmd;
struct tcm_usbg_nexus *tv_nexus;
struct usbg_tpg *tpg;
int dir, flags = (TARGET_SCF_UNKNOWN_SIZE | TARGET_SCF_ACK_KREF);
/*
* Note: each command will spawn its own process, and each stage of the
* command is processed sequentially. Should this no longer be the case,
* locking is needed.
*/
if (cmd->flags & USBG_CMD_PENDING_DATA_WRITE) {
target_execute_cmd(&cmd->se_cmd);
cmd->flags &= ~USBG_CMD_PENDING_DATA_WRITE;
return;
}
se_cmd = &cmd->se_cmd;
tpg = cmd->fu->tpg;
tv_nexus = tpg->tpg_nexus;
dir = get_cmd_dir(cmd->cmd_buf);
if (dir < 0)
goto out;
target_submit_cmd(se_cmd, tv_nexus->tvn_se_sess, cmd->cmd_buf,
cmd->sense_iu.sense, cmd->unpacked_lun, 0,
cmd->prio_attr, dir, flags);
return;
out:
__target_init_cmd(se_cmd,
tv_nexus->tvn_se_sess->se_tpg->se_tpg_tfo,
tv_nexus->tvn_se_sess, cmd->data_len, DMA_NONE,
cmd->prio_attr, cmd->sense_iu.sense,
cmd->unpacked_lun, NULL);
transport_send_check_condition_and_sense(se_cmd,
TCM_UNSUPPORTED_SCSI_OPCODE, 0);
}
static void usbg_cmd_work(struct work_struct *work)
{
struct usbg_cmd *cmd = container_of(work, struct usbg_cmd, work);
/*
* Failure is detected by f_tcm here. Skip submitting the command to the
* target core if we already know the failing response and send the usb
* response to the host directly.
*/
if (cmd->tmr_rsp != RC_RESPONSE_UNKNOWN)
goto skip;
if (cmd->tmr_func)
usbg_submit_tmr(cmd);
else
usbg_submit_cmd(cmd);
return;
skip:
if (cmd->tmr_rsp == RC_OVERLAPPED_TAG) {
struct f_uas *fu = cmd->fu;
struct se_session *se_sess;
struct uas_stream *stream = NULL;
struct hlist_node *tmp;
struct usbg_cmd *active_cmd = NULL;
se_sess = cmd->fu->tpg->tpg_nexus->tvn_se_sess;
hash_for_each_possible_safe(fu->stream_hash, stream, tmp, node, cmd->tag) {
int i = stream - &fu->stream[0];
active_cmd = &((struct usbg_cmd *)se_sess->sess_cmd_map)[i];
if (active_cmd->tag == cmd->tag)
break;
}
/* Sanity check */
if (!stream || (active_cmd && active_cmd->tag != cmd->tag)) {
usbg_submit_command(cmd->fu, cmd->req);
return;
}
reinit_completion(&stream->cmd_completion);
/*
* A UASP command consists of the command, data, and status
* stages, each operating sequentially from different endpoints.
*
* Each USB endpoint operates independently, and depending on
* hardware implementation, a completion callback for a transfer
* from one endpoint may not reflect the order of completion on
* the wire. This is particularly true for devices with
* endpoints that have independent interrupts and event buffers.
*
* The driver must still detect misbehaving hosts and respond
* with an overlap status. To reduce false overlap failures,
* allow the active and matching stream ID a brief 1ms to
* complete before responding with an overlap command failure.
* Overlap failure should be rare.
*/
wait_for_completion_timeout(&stream->cmd_completion, msecs_to_jiffies(1));
/* If the previous stream is completed, retry the command. */
if (!hash_hashed(&stream->node)) {
usbg_submit_command(cmd->fu, cmd->req);
return;
}
/*
* The command isn't submitted to the target core, so we're safe
* to remove the bitmap index from the session tag pool.
*/
sbitmap_queue_clear(&se_sess->sess_tag_pool,
cmd->se_cmd.map_tag,
cmd->se_cmd.map_cpu);
/*
* Overlap command tag detected. Cancel any pending transfer of
* the command submitted to target core.
*/
active_cmd->tmr_rsp = RC_OVERLAPPED_TAG;
usbg_aborted_task(&active_cmd->se_cmd);
/* Send the response after the transfer is aborted. */
return;
}
uasp_send_tm_response(cmd);
}
static struct usbg_cmd *usbg_get_cmd(struct f_uas *fu,
struct tcm_usbg_nexus *tv_nexus, u32 scsi_tag)
{
struct se_session *se_sess = tv_nexus->tvn_se_sess;
struct usbg_cmd *cmd;
int tag, cpu;
tag = sbitmap_queue_get(&se_sess->sess_tag_pool, &cpu);
if (tag < 0)
return ERR_PTR(-ENOMEM);
cmd = &((struct usbg_cmd *)se_sess->sess_cmd_map)[tag];
memset(cmd, 0, sizeof(*cmd));
cmd->se_cmd.map_tag = tag;
cmd->se_cmd.map_cpu = cpu;
cmd->se_cmd.cpuid = cpu;
cmd->se_cmd.tag = cmd->tag = scsi_tag;
cmd->fu = fu;
return cmd;
}
static void usbg_release_cmd(struct se_cmd *);
static int usbg_submit_command(struct f_uas *fu, struct usb_request *req)
{
struct iu *iu = req->buf;
struct usbg_cmd *cmd;
struct usbg_tpg *tpg = fu->tpg;
struct tcm_usbg_nexus *tv_nexus;
struct uas_stream *stream;
struct hlist_node *tmp;
struct command_iu *cmd_iu;
u32 cmd_len;
u16 scsi_tag;
tv_nexus = tpg->tpg_nexus;
if (!tv_nexus) {
pr_err("Missing nexus, ignoring command\n");
return -EINVAL;
}
scsi_tag = be16_to_cpup(&iu->tag);
cmd = usbg_get_cmd(fu, tv_nexus, scsi_tag);
if (IS_ERR(cmd)) {
pr_err("usbg_get_cmd failed\n");
return -ENOMEM;
}
cmd->req = req;
cmd->fu = fu;
cmd->tag = scsi_tag;
cmd->se_cmd.tag = scsi_tag;
cmd->tmr_func = 0;
cmd->tmr_rsp = RC_RESPONSE_UNKNOWN;
cmd->flags = 0;
cmd_iu = (struct command_iu *)iu;
/* Command and Task Management IUs share the same LUN offset */
cmd->unpacked_lun = scsilun_to_int(&cmd_iu->lun);
if (iu->iu_id != IU_ID_COMMAND && iu->iu_id != IU_ID_TASK_MGMT) {
cmd->tmr_rsp = RC_INVALID_INFO_UNIT;
goto skip;
}
hash_for_each_possible_safe(fu->stream_hash, stream, tmp, node, scsi_tag) {
struct usbg_cmd *active_cmd;
struct se_session *se_sess;
int i = stream - &fu->stream[0];
se_sess = cmd->fu->tpg->tpg_nexus->tvn_se_sess;
active_cmd = &((struct usbg_cmd *)se_sess->sess_cmd_map)[i];
if (active_cmd->tag == scsi_tag) {
cmd->tmr_rsp = RC_OVERLAPPED_TAG;
goto skip;
}
}
stream = &fu->stream[cmd->se_cmd.map_tag];
hash_add(fu->stream_hash, &stream->node, scsi_tag);
if (iu->iu_id == IU_ID_TASK_MGMT) {
struct task_mgmt_iu *tm_iu;
tm_iu = (struct task_mgmt_iu *)iu;
cmd->tmr_func = tm_iu->function;
goto skip;
}
cmd_len = (cmd_iu->len & ~0x3) + 16;
if (cmd_len > USBG_MAX_CMD) {
target_free_tag(tv_nexus->tvn_se_sess, &cmd->se_cmd);
hash_del(&stream->node);
return -EINVAL;
}
memcpy(cmd->cmd_buf, cmd_iu->cdb, cmd_len);
switch (cmd_iu->prio_attr & 0x7) {
case UAS_HEAD_TAG:
cmd->prio_attr = TCM_HEAD_TAG;
break;
case UAS_ORDERED_TAG:
cmd->prio_attr = TCM_ORDERED_TAG;
break;
case UAS_ACA:
cmd->prio_attr = TCM_ACA_TAG;
break;
default:
pr_debug_once("Unsupported prio_attr: %02x.\n",
cmd_iu->prio_attr);
fallthrough;
case UAS_SIMPLE_TAG:
cmd->prio_attr = TCM_SIMPLE_TAG;
break;
}
skip:
INIT_WORK(&cmd->work, usbg_cmd_work);
queue_work(tpg->workqueue, &cmd->work);
return 0;
}
static void bot_cmd_work(struct work_struct *work)
{
struct usbg_cmd *cmd = container_of(work, struct usbg_cmd, work);
struct se_cmd *se_cmd;
struct tcm_usbg_nexus *tv_nexus;
struct usbg_tpg *tpg;
int flags = TARGET_SCF_ACK_KREF;
int dir;
/*
* Note: each command will spawn its own process, and each stage of the
* command is processed sequentially. Should this no longer be the case,
* locking is needed.
*/
if (cmd->flags & USBG_CMD_PENDING_DATA_WRITE) {
target_execute_cmd(&cmd->se_cmd);
cmd->flags &= ~USBG_CMD_PENDING_DATA_WRITE;
return;
}
se_cmd = &cmd->se_cmd;
tpg = cmd->fu->tpg;
tv_nexus = tpg->tpg_nexus;
dir = get_cmd_dir(cmd->cmd_buf);
if (dir < 0)
goto out;
target_submit_cmd(se_cmd, tv_nexus->tvn_se_sess,
cmd->cmd_buf, cmd->sense_iu.sense, cmd->unpacked_lun,
cmd->data_len, cmd->prio_attr, dir, flags);
return;
out:
__target_init_cmd(se_cmd,
tv_nexus->tvn_se_sess->se_tpg->se_tpg_tfo,
tv_nexus->tvn_se_sess, cmd->data_len, DMA_NONE,
cmd->prio_attr, cmd->sense_iu.sense,
cmd->unpacked_lun, NULL);
transport_send_check_condition_and_sense(se_cmd,
TCM_UNSUPPORTED_SCSI_OPCODE, 0);
}
static int bot_submit_command(struct f_uas *fu,
void *cmdbuf, unsigned int len)
{
struct bulk_cb_wrap *cbw = cmdbuf;
struct usbg_cmd *cmd;
struct usbg_tpg *tpg = fu->tpg;
struct tcm_usbg_nexus *tv_nexus;
u32 cmd_len;
if (cbw->Signature != cpu_to_le32(US_BULK_CB_SIGN)) {
pr_err("Wrong signature on CBW\n");
return -EINVAL;
}
if (len != 31) {
pr_err("Wrong length for CBW\n");
return -EINVAL;
}
cmd_len = cbw->Length;
if (cmd_len < 1 || cmd_len > 16)
return -EINVAL;
tv_nexus = tpg->tpg_nexus;
if (!tv_nexus) {
pr_err("Missing nexus, ignoring command\n");
return -ENODEV;
}
cmd = usbg_get_cmd(fu, tv_nexus, cbw->Tag);
if (IS_ERR(cmd)) {
pr_err("usbg_get_cmd failed\n");
return -ENOMEM;
}
memcpy(cmd->cmd_buf, cbw->CDB, cmd_len);
cmd->bot_tag = cbw->Tag;
cmd->prio_attr = TCM_SIMPLE_TAG;
cmd->unpacked_lun = cbw->Lun;
cmd->is_read = cbw->Flags & US_BULK_FLAG_IN ? 1 : 0;
cmd->data_len = le32_to_cpu(cbw->DataTransferLength);
cmd->se_cmd.tag = le32_to_cpu(cmd->bot_tag);
cmd->flags = 0;
INIT_WORK(&cmd->work, bot_cmd_work);
queue_work(tpg->workqueue, &cmd->work);
return 0;
}
/* Start fabric.c code */
static int usbg_check_true(struct se_portal_group *se_tpg)
{
return 1;
}
static char *usbg_get_fabric_wwn(struct se_portal_group *se_tpg)
{
struct usbg_tpg *tpg = container_of(se_tpg,
struct usbg_tpg, se_tpg);
struct usbg_tport *tport = tpg->tport;
return &tport->tport_name[0];
}
static u16 usbg_get_tag(struct se_portal_group *se_tpg)
{
struct usbg_tpg *tpg = container_of(se_tpg,
struct usbg_tpg, se_tpg);
return tpg->tport_tpgt;
}
static void usbg_release_cmd(struct se_cmd *se_cmd)
{
struct usbg_cmd *cmd = container_of(se_cmd, struct usbg_cmd,
se_cmd);
struct se_session *se_sess = se_cmd->se_sess;
cmd->tag = 0;
kfree(cmd->data_buf);
target_free_tag(se_sess, se_cmd);
}
static void usbg_queue_tm_rsp(struct se_cmd *se_cmd)
{
struct usbg_cmd *cmd = container_of(se_cmd, struct usbg_cmd, se_cmd);
uasp_send_tm_response(cmd);
}
static void usbg_aborted_task(struct se_cmd *se_cmd)
{
struct usbg_cmd *cmd = container_of(se_cmd, struct usbg_cmd, se_cmd);
struct f_uas *fu = cmd->fu;
struct usb_gadget *gadget = fuas_to_gadget(fu);
struct uas_stream *stream = &fu->stream[se_cmd->map_tag];
int ret = 0;
if (stream->req_out->status == -EINPROGRESS)
ret = usb_ep_dequeue(fu->ep_out, stream->req_out);
else if (stream->req_in->status == -EINPROGRESS)
ret = usb_ep_dequeue(fu->ep_in, stream->req_in);
else if (stream->req_status->status == -EINPROGRESS)
ret = usb_ep_dequeue(fu->ep_status, stream->req_status);
if (ret)
dev_err(&gadget->dev, "Failed to abort cmd tag %d, (%d)\n",
cmd->tag, ret);
cmd->state = UASP_QUEUE_COMMAND;
}
static const char *usbg_check_wwn(const char *name)
{
const char *n;
unsigned int len;
n = strstr(name, "naa.");
if (!n)
return NULL;
n += 4;
len = strlen(n);
if (len == 0 || len > USBG_NAMELEN - 1)
return NULL;
return n;
}
static int usbg_init_nodeacl(struct se_node_acl *se_nacl, const char *name)
{
if (!usbg_check_wwn(name))
return -EINVAL;
return 0;
}
static struct se_portal_group *usbg_make_tpg(struct se_wwn *wwn,
const char *name)
{
struct usbg_tport *tport = container_of(wwn, struct usbg_tport,
tport_wwn);
struct usbg_tpg *tpg;
u16 tpgt;
int ret;
struct f_tcm_opts *opts;
unsigned i;
if (strstr(name, "tpgt_") != name)
return ERR_PTR(-EINVAL);
if (kstrtou16(name + 5, 0, &tpgt))
return ERR_PTR(-EINVAL);
ret = -ENODEV;
mutex_lock(&tpg_instances_lock);
for (i = 0; i < TPG_INSTANCES; ++i)
if (tpg_instances[i].func_inst && !tpg_instances[i].tpg)
break;
if (i == TPG_INSTANCES)
goto unlock_inst;
opts = container_of(tpg_instances[i].func_inst, struct f_tcm_opts,
func_inst);
mutex_lock(&opts->dep_lock);
if (!opts->ready)
goto unlock_dep;
if (opts->has_dep) {
if (!try_module_get(opts->dependent))
goto unlock_dep;
} else {
ret = configfs_depend_item_unlocked(
wwn->wwn_group.cg_subsys,
&opts->func_inst.group.cg_item);
if (ret)
goto unlock_dep;
}
tpg = kzalloc(sizeof(struct usbg_tpg), GFP_KERNEL);
ret = -ENOMEM;
if (!tpg)
goto unref_dep;
mutex_init(&tpg->tpg_mutex);
atomic_set(&tpg->tpg_port_count, 0);
tpg->workqueue = alloc_workqueue("tcm_usb_gadget",
WQ_UNBOUND, WQ_UNBOUND_MAX_ACTIVE);
if (!tpg->workqueue)
goto free_tpg;
tpg->tport = tport;
tpg->tport_tpgt = tpgt;
/*
* SPC doesn't assign a protocol identifier for USB-SCSI, so we
* pretend to be SAS..
*/
ret = core_tpg_register(wwn, &tpg->se_tpg, SCSI_PROTOCOL_SAS);
if (ret < 0)
goto free_workqueue;
tpg_instances[i].tpg = tpg;
tpg->fi = tpg_instances[i].func_inst;
mutex_unlock(&opts->dep_lock);
mutex_unlock(&tpg_instances_lock);
return &tpg->se_tpg;
free_workqueue:
destroy_workqueue(tpg->workqueue);
free_tpg:
kfree(tpg);
unref_dep:
if (opts->has_dep)
module_put(opts->dependent);
else
configfs_undepend_item_unlocked(&opts->func_inst.group.cg_item);
unlock_dep:
mutex_unlock(&opts->dep_lock);
unlock_inst:
mutex_unlock(&tpg_instances_lock);
return ERR_PTR(ret);
}
static int tcm_usbg_drop_nexus(struct usbg_tpg *);
static void usbg_drop_tpg(struct se_portal_group *se_tpg)
{
struct usbg_tpg *tpg = container_of(se_tpg,
struct usbg_tpg, se_tpg);
unsigned i;
struct f_tcm_opts *opts;
tcm_usbg_drop_nexus(tpg);
core_tpg_deregister(se_tpg);
destroy_workqueue(tpg->workqueue);
mutex_lock(&tpg_instances_lock);
for (i = 0; i < TPG_INSTANCES; ++i)
if (tpg_instances[i].tpg == tpg)
break;
if (i < TPG_INSTANCES) {
tpg_instances[i].tpg = NULL;
opts = container_of(tpg_instances[i].func_inst,
struct f_tcm_opts, func_inst);
mutex_lock(&opts->dep_lock);
if (opts->has_dep)
module_put(opts->dependent);
else
configfs_undepend_item_unlocked(
&opts->func_inst.group.cg_item);
mutex_unlock(&opts->dep_lock);
}
mutex_unlock(&tpg_instances_lock);
kfree(tpg);
}
static struct se_wwn *usbg_make_tport(
struct target_fabric_configfs *tf,
struct config_group *group,
const char *name)
{
struct usbg_tport *tport;
const char *wnn_name;
u64 wwpn = 0;
wnn_name = usbg_check_wwn(name);
if (!wnn_name)
return ERR_PTR(-EINVAL);
tport = kzalloc(sizeof(struct usbg_tport), GFP_KERNEL);
if (!(tport))
return ERR_PTR(-ENOMEM);
tport->tport_wwpn = wwpn;
snprintf(tport->tport_name, sizeof(tport->tport_name), "%s", wnn_name);
return &tport->tport_wwn;
}
static void usbg_drop_tport(struct se_wwn *wwn)
{
struct usbg_tport *tport = container_of(wwn,
struct usbg_tport, tport_wwn);
kfree(tport);
}
/*
* If somebody feels like dropping the version property, go ahead.
*/
static ssize_t usbg_wwn_version_show(struct config_item *item, char *page)
{
return sprintf(page, "usb-gadget fabric module\n");
}
CONFIGFS_ATTR_RO(usbg_wwn_, version);
static struct configfs_attribute *usbg_wwn_attrs[] = {
&usbg_wwn_attr_version,
NULL,
};
static int usbg_attach(struct usbg_tpg *);
static void usbg_detach(struct usbg_tpg *);
static int usbg_enable_tpg(struct se_portal_group *se_tpg, bool enable)
{
struct usbg_tpg *tpg = container_of(se_tpg, struct usbg_tpg, se_tpg);
int ret = 0;
if (enable)
ret = usbg_attach(tpg);
else
usbg_detach(tpg);
if (ret)
return ret;
tpg->gadget_connect = enable;
return 0;
}
static ssize_t tcm_usbg_tpg_nexus_show(struct config_item *item, char *page)
{
struct se_portal_group *se_tpg = to_tpg(item);
struct usbg_tpg *tpg = container_of(se_tpg, struct usbg_tpg, se_tpg);
struct tcm_usbg_nexus *tv_nexus;
ssize_t ret;
mutex_lock(&tpg->tpg_mutex);
tv_nexus = tpg->tpg_nexus;
if (!tv_nexus) {
ret = -ENODEV;
goto out;
}
ret = sysfs_emit(page, "%s\n",
tv_nexus->tvn_se_sess->se_node_acl->initiatorname);
out:
mutex_unlock(&tpg->tpg_mutex);
return ret;
}
static int usbg_alloc_sess_cb(struct se_portal_group *se_tpg,
struct se_session *se_sess, void *p)
{
struct usbg_tpg *tpg = container_of(se_tpg,
struct usbg_tpg, se_tpg);
tpg->tpg_nexus = p;
return 0;
}
static int tcm_usbg_make_nexus(struct usbg_tpg *tpg, char *name)
{
struct tcm_usbg_nexus *tv_nexus;
int ret = 0;
mutex_lock(&tpg->tpg_mutex);
if (tpg->tpg_nexus) {
ret = -EEXIST;
pr_debug("tpg->tpg_nexus already exists\n");
goto out_unlock;
}
tv_nexus = kzalloc(sizeof(*tv_nexus), GFP_KERNEL);
if (!tv_nexus) {
ret = -ENOMEM;
goto out_unlock;
}
tv_nexus->tvn_se_sess = target_setup_session(&tpg->se_tpg,
USB_G_DEFAULT_SESSION_TAGS,
sizeof(struct usbg_cmd),
TARGET_PROT_NORMAL, name,
tv_nexus, usbg_alloc_sess_cb);
if (IS_ERR(tv_nexus->tvn_se_sess)) {
#define MAKE_NEXUS_MSG "core_tpg_check_initiator_node_acl() failed for %s\n"
pr_debug(MAKE_NEXUS_MSG, name);
#undef MAKE_NEXUS_MSG
ret = PTR_ERR(tv_nexus->tvn_se_sess);
kfree(tv_nexus);
}
out_unlock:
mutex_unlock(&tpg->tpg_mutex);
return ret;
}
static int tcm_usbg_drop_nexus(struct usbg_tpg *tpg)
{
struct se_session *se_sess;
struct tcm_usbg_nexus *tv_nexus;
int ret = -ENODEV;
mutex_lock(&tpg->tpg_mutex);
tv_nexus = tpg->tpg_nexus;
if (!tv_nexus)
goto out;
se_sess = tv_nexus->tvn_se_sess;
if (!se_sess)
goto out;
if (atomic_read(&tpg->tpg_port_count)) {
ret = -EPERM;
#define MSG "Unable to remove Host I_T Nexus with active TPG port count: %d\n"
pr_err(MSG, atomic_read(&tpg->tpg_port_count));
#undef MSG
goto out;
}
pr_debug("Removing I_T Nexus to Initiator Port: %s\n",
tv_nexus->tvn_se_sess->se_node_acl->initiatorname);
/*
* Release the SCSI I_T Nexus to the emulated vHost Target Port
*/
target_remove_session(se_sess);
tpg->tpg_nexus = NULL;
kfree(tv_nexus);
ret = 0;
out:
mutex_unlock(&tpg->tpg_mutex);
return ret;
}
static ssize_t tcm_usbg_tpg_nexus_store(struct config_item *item,
const char *page, size_t count)
{
struct se_portal_group *se_tpg = to_tpg(item);
struct usbg_tpg *tpg = container_of(se_tpg, struct usbg_tpg, se_tpg);
unsigned char i_port[USBG_NAMELEN], *ptr;
int ret;
if (!strncmp(page, "NULL", 4)) {
ret = tcm_usbg_drop_nexus(tpg);
return (!ret) ? count : ret;
}
if (strlen(page) >= USBG_NAMELEN) {
#define NEXUS_STORE_MSG "Emulated NAA Sas Address: %s, exceeds max: %d\n"
pr_err(NEXUS_STORE_MSG, page, USBG_NAMELEN);
#undef NEXUS_STORE_MSG
return -EINVAL;
}
snprintf(i_port, USBG_NAMELEN, "%s", page);
ptr = strstr(i_port, "naa.");
if (!ptr) {
pr_err("Missing 'naa.' prefix\n");
return -EINVAL;
}
if (i_port[strlen(i_port) - 1] == '\n')
i_port[strlen(i_port) - 1] = '\0';
ret = tcm_usbg_make_nexus(tpg, &i_port[0]);
if (ret < 0)
return ret;
return count;
}
CONFIGFS_ATTR(tcm_usbg_tpg_, nexus);
static struct configfs_attribute *usbg_base_attrs[] = {
&tcm_usbg_tpg_attr_nexus,
NULL,
};
static int usbg_port_link(struct se_portal_group *se_tpg, struct se_lun *lun)
{
struct usbg_tpg *tpg = container_of(se_tpg, struct usbg_tpg, se_tpg);
atomic_inc(&tpg->tpg_port_count);
smp_mb__after_atomic();
return 0;
}
static void usbg_port_unlink(struct se_portal_group *se_tpg,
struct se_lun *se_lun)
{
struct usbg_tpg *tpg = container_of(se_tpg, struct usbg_tpg, se_tpg);
atomic_dec(&tpg->tpg_port_count);
smp_mb__after_atomic();
}
static int usbg_check_stop_free(struct se_cmd *se_cmd)
{
return target_put_sess_cmd(se_cmd);
}
static const struct target_core_fabric_ops usbg_ops = {
.module = THIS_MODULE,
.fabric_name = "usb_gadget",
.tpg_get_wwn = usbg_get_fabric_wwn,
.tpg_get_tag = usbg_get_tag,
.tpg_check_demo_mode = usbg_check_true,
.release_cmd = usbg_release_cmd,
.sess_get_initiator_sid = NULL,
.write_pending = usbg_send_write_request,
.queue_data_in = usbg_send_read_response,
.queue_status = usbg_send_status_response,
.queue_tm_rsp = usbg_queue_tm_rsp,
.aborted_task = usbg_aborted_task,
.check_stop_free = usbg_check_stop_free,
.fabric_make_wwn = usbg_make_tport,
.fabric_drop_wwn = usbg_drop_tport,
.fabric_make_tpg = usbg_make_tpg,
.fabric_enable_tpg = usbg_enable_tpg,
.fabric_drop_tpg = usbg_drop_tpg,
.fabric_post_link = usbg_port_link,
.fabric_pre_unlink = usbg_port_unlink,
.fabric_init_nodeacl = usbg_init_nodeacl,
.tfc_wwn_attrs = usbg_wwn_attrs,
.tfc_tpg_base_attrs = usbg_base_attrs,
.default_submit_type = TARGET_DIRECT_SUBMIT,
.direct_submit_supp = 1,
};
/* Start gadget.c code */
static struct usb_interface_descriptor bot_intf_desc = {
.bLength = sizeof(bot_intf_desc),
.bDescriptorType = USB_DT_INTERFACE,
.bNumEndpoints = 2,
.bAlternateSetting = USB_G_ALT_INT_BBB,
.bInterfaceClass = USB_CLASS_MASS_STORAGE,
.bInterfaceSubClass = USB_SC_SCSI,
.bInterfaceProtocol = USB_PR_BULK,
};
static struct usb_interface_descriptor uasp_intf_desc = {
.bLength = sizeof(uasp_intf_desc),
.bDescriptorType = USB_DT_INTERFACE,
.bNumEndpoints = 4,
.bAlternateSetting = USB_G_ALT_INT_UAS,
.bInterfaceClass = USB_CLASS_MASS_STORAGE,
.bInterfaceSubClass = USB_SC_SCSI,
.bInterfaceProtocol = USB_PR_UAS,
};
static struct usb_endpoint_descriptor uasp_bi_desc = {
.bLength = USB_DT_ENDPOINT_SIZE,
.bDescriptorType = USB_DT_ENDPOINT,
.bEndpointAddress = USB_DIR_IN,
.bmAttributes = USB_ENDPOINT_XFER_BULK,
.wMaxPacketSize = cpu_to_le16(512),
};
static struct usb_endpoint_descriptor uasp_fs_bi_desc = {
.bLength = USB_DT_ENDPOINT_SIZE,
.bDescriptorType = USB_DT_ENDPOINT,
.bEndpointAddress = USB_DIR_IN,
.bmAttributes = USB_ENDPOINT_XFER_BULK,
};
static struct usb_pipe_usage_descriptor uasp_bi_pipe_desc = {
.bLength = sizeof(uasp_bi_pipe_desc),
.bDescriptorType = USB_DT_PIPE_USAGE,
.bPipeID = DATA_IN_PIPE_ID,
};
static struct usb_endpoint_descriptor uasp_ss_bi_desc = {
.bLength = USB_DT_ENDPOINT_SIZE,
.bDescriptorType = USB_DT_ENDPOINT,
.bEndpointAddress = USB_DIR_IN,
.bmAttributes = USB_ENDPOINT_XFER_BULK,
.wMaxPacketSize = cpu_to_le16(1024),
};
static struct usb_ss_ep_comp_descriptor uasp_bi_ep_comp_desc = {
.bLength = sizeof(uasp_bi_ep_comp_desc),
.bDescriptorType = USB_DT_SS_ENDPOINT_COMP,
.bMaxBurst = 15,
.bmAttributes = UASP_SS_EP_COMP_LOG_STREAMS,
.wBytesPerInterval = 0,
};
static struct usb_ss_ep_comp_descriptor bot_bi_ep_comp_desc = {
.bLength = sizeof(bot_bi_ep_comp_desc),
.bDescriptorType = USB_DT_SS_ENDPOINT_COMP,
.bMaxBurst = 15,
};
static struct usb_endpoint_descriptor uasp_bo_desc = {
.bLength = USB_DT_ENDPOINT_SIZE,
.bDescriptorType = USB_DT_ENDPOINT,
.bEndpointAddress = USB_DIR_OUT,
.bmAttributes = USB_ENDPOINT_XFER_BULK,
.wMaxPacketSize = cpu_to_le16(512),
};
static struct usb_endpoint_descriptor uasp_fs_bo_desc = {
.bLength = USB_DT_ENDPOINT_SIZE,
.bDescriptorType = USB_DT_ENDPOINT,
.bEndpointAddress = USB_DIR_OUT,
.bmAttributes = USB_ENDPOINT_XFER_BULK,
};
static struct usb_pipe_usage_descriptor uasp_bo_pipe_desc = {
.bLength = sizeof(uasp_bo_pipe_desc),
.bDescriptorType = USB_DT_PIPE_USAGE,
.bPipeID = DATA_OUT_PIPE_ID,
};
static struct usb_endpoint_descriptor uasp_ss_bo_desc = {
.bLength = USB_DT_ENDPOINT_SIZE,
.bDescriptorType = USB_DT_ENDPOINT,
.bEndpointAddress = USB_DIR_OUT,
.bmAttributes = USB_ENDPOINT_XFER_BULK,
.wMaxPacketSize = cpu_to_le16(0x400),
};
static struct usb_ss_ep_comp_descriptor uasp_bo_ep_comp_desc = {
.bLength = sizeof(uasp_bo_ep_comp_desc),
.bDescriptorType = USB_DT_SS_ENDPOINT_COMP,
.bMaxBurst = 15,
.bmAttributes = UASP_SS_EP_COMP_LOG_STREAMS,
};
static struct usb_ss_ep_comp_descriptor bot_bo_ep_comp_desc = {
.bLength = sizeof(bot_bo_ep_comp_desc),
.bDescriptorType = USB_DT_SS_ENDPOINT_COMP,
.bMaxBurst = 15,
};
static struct usb_endpoint_descriptor uasp_status_desc = {
.bLength = USB_DT_ENDPOINT_SIZE,
.bDescriptorType = USB_DT_ENDPOINT,
.bEndpointAddress = USB_DIR_IN,
.bmAttributes = USB_ENDPOINT_XFER_BULK,
.wMaxPacketSize = cpu_to_le16(512),
};
static struct usb_endpoint_descriptor uasp_fs_status_desc = {
.bLength = USB_DT_ENDPOINT_SIZE,
.bDescriptorType = USB_DT_ENDPOINT,
.bEndpointAddress = USB_DIR_IN,
.bmAttributes = USB_ENDPOINT_XFER_BULK,
};
static struct usb_pipe_usage_descriptor uasp_status_pipe_desc = {
.bLength = sizeof(uasp_status_pipe_desc),
.bDescriptorType = USB_DT_PIPE_USAGE,
.bPipeID = STATUS_PIPE_ID,
};
static struct usb_endpoint_descriptor uasp_ss_status_desc = {
.bLength = USB_DT_ENDPOINT_SIZE,
.bDescriptorType = USB_DT_ENDPOINT,
.bEndpointAddress = USB_DIR_IN,
.bmAttributes = USB_ENDPOINT_XFER_BULK,
.wMaxPacketSize = cpu_to_le16(1024),
};
static struct usb_ss_ep_comp_descriptor uasp_status_in_ep_comp_desc = {
.bLength = sizeof(uasp_status_in_ep_comp_desc),
.bDescriptorType = USB_DT_SS_ENDPOINT_COMP,
.bmAttributes = UASP_SS_EP_COMP_LOG_STREAMS,
};
static struct usb_endpoint_descriptor uasp_cmd_desc = {
.bLength = USB_DT_ENDPOINT_SIZE,
.bDescriptorType = USB_DT_ENDPOINT,
.bEndpointAddress = USB_DIR_OUT,
.bmAttributes = USB_ENDPOINT_XFER_BULK,
.wMaxPacketSize = cpu_to_le16(512),
};
static struct usb_endpoint_descriptor uasp_fs_cmd_desc = {
.bLength = USB_DT_ENDPOINT_SIZE,
.bDescriptorType = USB_DT_ENDPOINT,
.bEndpointAddress = USB_DIR_OUT,
.bmAttributes = USB_ENDPOINT_XFER_BULK,
};
static struct usb_pipe_usage_descriptor uasp_cmd_pipe_desc = {
.bLength = sizeof(uasp_cmd_pipe_desc),
.bDescriptorType = USB_DT_PIPE_USAGE,
.bPipeID = CMD_PIPE_ID,
};
static struct usb_endpoint_descriptor uasp_ss_cmd_desc = {
.bLength = USB_DT_ENDPOINT_SIZE,
.bDescriptorType = USB_DT_ENDPOINT,
.bEndpointAddress = USB_DIR_OUT,
.bmAttributes = USB_ENDPOINT_XFER_BULK,
.wMaxPacketSize = cpu_to_le16(1024),
};
static struct usb_ss_ep_comp_descriptor uasp_cmd_comp_desc = {
.bLength = sizeof(uasp_cmd_comp_desc),
.bDescriptorType = USB_DT_SS_ENDPOINT_COMP,
};
static struct usb_descriptor_header *uasp_fs_function_desc[] = {
(struct usb_descriptor_header *) &bot_intf_desc,
(struct usb_descriptor_header *) &uasp_fs_bi_desc,
(struct usb_descriptor_header *) &uasp_fs_bo_desc,
(struct usb_descriptor_header *) &uasp_intf_desc,
(struct usb_descriptor_header *) &uasp_fs_bi_desc,
(struct usb_descriptor_header *) &uasp_bi_pipe_desc,
(struct usb_descriptor_header *) &uasp_fs_bo_desc,
(struct usb_descriptor_header *) &uasp_bo_pipe_desc,
(struct usb_descriptor_header *) &uasp_fs_status_desc,
(struct usb_descriptor_header *) &uasp_status_pipe_desc,
(struct usb_descriptor_header *) &uasp_fs_cmd_desc,
(struct usb_descriptor_header *) &uasp_cmd_pipe_desc,
NULL,
};
static struct usb_descriptor_header *uasp_hs_function_desc[] = {
(struct usb_descriptor_header *) &bot_intf_desc,
(struct usb_descriptor_header *) &uasp_bi_desc,
(struct usb_descriptor_header *) &uasp_bo_desc,
(struct usb_descriptor_header *) &uasp_intf_desc,
(struct usb_descriptor_header *) &uasp_bi_desc,
(struct usb_descriptor_header *) &uasp_bi_pipe_desc,
(struct usb_descriptor_header *) &uasp_bo_desc,
(struct usb_descriptor_header *) &uasp_bo_pipe_desc,
(struct usb_descriptor_header *) &uasp_status_desc,
(struct usb_descriptor_header *) &uasp_status_pipe_desc,
(struct usb_descriptor_header *) &uasp_cmd_desc,
(struct usb_descriptor_header *) &uasp_cmd_pipe_desc,
NULL,
};
static struct usb_descriptor_header *uasp_ss_function_desc[] = {
(struct usb_descriptor_header *) &bot_intf_desc,
(struct usb_descriptor_header *) &uasp_ss_bi_desc,
(struct usb_descriptor_header *) &bot_bi_ep_comp_desc,
(struct usb_descriptor_header *) &uasp_ss_bo_desc,
(struct usb_descriptor_header *) &bot_bo_ep_comp_desc,
(struct usb_descriptor_header *) &uasp_intf_desc,
(struct usb_descriptor_header *) &uasp_ss_bi_desc,
(struct usb_descriptor_header *) &uasp_bi_ep_comp_desc,
(struct usb_descriptor_header *) &uasp_bi_pipe_desc,
(struct usb_descriptor_header *) &uasp_ss_bo_desc,
(struct usb_descriptor_header *) &uasp_bo_ep_comp_desc,
(struct usb_descriptor_header *) &uasp_bo_pipe_desc,
(struct usb_descriptor_header *) &uasp_ss_status_desc,
(struct usb_descriptor_header *) &uasp_status_in_ep_comp_desc,
(struct usb_descriptor_header *) &uasp_status_pipe_desc,
(struct usb_descriptor_header *) &uasp_ss_cmd_desc,
(struct usb_descriptor_header *) &uasp_cmd_comp_desc,
(struct usb_descriptor_header *) &uasp_cmd_pipe_desc,
NULL,
};
static struct usb_string tcm_us_strings[] = {
[USB_G_STR_INT_UAS].s = "USB Attached SCSI",
[USB_G_STR_INT_BBB].s = "Bulk Only Transport",
{ },
};
static struct usb_gadget_strings tcm_stringtab = {
.language = 0x0409,
.strings = tcm_us_strings,
};
static struct usb_gadget_strings *tcm_strings[] = {
&tcm_stringtab,
NULL,
};
static int tcm_bind(struct usb_configuration *c, struct usb_function *f)
{
struct f_uas *fu = to_f_uas(f);
struct usb_string *us;
struct usb_gadget *gadget = c->cdev->gadget;
struct usb_ep *ep;
struct f_tcm_opts *opts;
int iface;
int ret;
opts = container_of(f->fi, struct f_tcm_opts, func_inst);
mutex_lock(&opts->dep_lock);
if (!opts->can_attach) {
mutex_unlock(&opts->dep_lock);
return -ENODEV;
}
mutex_unlock(&opts->dep_lock);
us = usb_gstrings_attach(c->cdev, tcm_strings,
ARRAY_SIZE(tcm_us_strings));
if (IS_ERR(us))
return PTR_ERR(us);
bot_intf_desc.iInterface = us[USB_G_STR_INT_BBB].id;
uasp_intf_desc.iInterface = us[USB_G_STR_INT_UAS].id;
iface = usb_interface_id(c, f);
if (iface < 0)
return iface;
bot_intf_desc.bInterfaceNumber = iface;
uasp_intf_desc.bInterfaceNumber = iface;
fu->iface = iface;
ep = usb_ep_autoconfig(gadget, &uasp_fs_bi_desc);
if (!ep)
goto ep_fail;
fu->ep_in = ep;
ep = usb_ep_autoconfig(gadget, &uasp_fs_bo_desc);
if (!ep)
goto ep_fail;
fu->ep_out = ep;
ep = usb_ep_autoconfig(gadget, &uasp_fs_status_desc);
if (!ep)
goto ep_fail;
fu->ep_status = ep;
ep = usb_ep_autoconfig(gadget, &uasp_fs_cmd_desc);
if (!ep)
goto ep_fail;
fu->ep_cmd = ep;
/* Assume endpoint addresses are the same for both speeds */
uasp_bi_desc.bEndpointAddress = uasp_fs_bi_desc.bEndpointAddress;
uasp_bo_desc.bEndpointAddress = uasp_fs_bo_desc.bEndpointAddress;
uasp_status_desc.bEndpointAddress =
uasp_fs_status_desc.bEndpointAddress;
uasp_cmd_desc.bEndpointAddress = uasp_fs_cmd_desc.bEndpointAddress;
uasp_ss_bi_desc.bEndpointAddress = uasp_fs_bi_desc.bEndpointAddress;
uasp_ss_bo_desc.bEndpointAddress = uasp_fs_bo_desc.bEndpointAddress;
uasp_ss_status_desc.bEndpointAddress =
uasp_fs_status_desc.bEndpointAddress;
uasp_ss_cmd_desc.bEndpointAddress = uasp_fs_cmd_desc.bEndpointAddress;
ret = usb_assign_descriptors(f, uasp_fs_function_desc,
uasp_hs_function_desc, uasp_ss_function_desc,
uasp_ss_function_desc);
if (ret)
goto ep_fail;
return 0;
ep_fail:
pr_err("Can't claim all required eps\n");
return -ENOTSUPP;
}
struct guas_setup_wq {
struct work_struct work;
struct f_uas *fu;
unsigned int alt;
};
static void tcm_delayed_set_alt(struct work_struct *wq)
{
struct guas_setup_wq *work = container_of(wq, struct guas_setup_wq,
work);
struct f_uas *fu = work->fu;
int alt = work->alt;
kfree(work);
if (fu->flags & USBG_IS_BOT)
bot_cleanup_old_alt(fu);
if (fu->flags & USBG_IS_UAS)
uasp_cleanup_old_alt(fu);
if (alt == USB_G_ALT_INT_BBB)
bot_set_alt(fu);
else if (alt == USB_G_ALT_INT_UAS)
uasp_set_alt(fu);
usb_composite_setup_continue(fu->function.config->cdev);
}
static int tcm_get_alt(struct usb_function *f, unsigned intf)
{
struct f_uas *fu = to_f_uas(f);
if (fu->iface != intf)
return -EOPNOTSUPP;
if (fu->flags & USBG_IS_BOT)
return USB_G_ALT_INT_BBB;
else if (fu->flags & USBG_IS_UAS)
return USB_G_ALT_INT_UAS;
return -EOPNOTSUPP;
}
static int tcm_set_alt(struct usb_function *f, unsigned intf, unsigned alt)
{
struct f_uas *fu = to_f_uas(f);
if (fu->iface != intf)
return -EOPNOTSUPP;
if ((alt == USB_G_ALT_INT_BBB) || (alt == USB_G_ALT_INT_UAS)) {
struct guas_setup_wq *work;
work = kmalloc(sizeof(*work), GFP_ATOMIC);
if (!work)
return -ENOMEM;
INIT_WORK(&work->work, tcm_delayed_set_alt);
work->fu = fu;
work->alt = alt;
schedule_work(&work->work);
return USB_GADGET_DELAYED_STATUS;
}
return -EOPNOTSUPP;
}
static void tcm_disable(struct usb_function *f)
{
struct f_uas *fu = to_f_uas(f);
if (fu->flags & USBG_IS_UAS)
uasp_cleanup_old_alt(fu);
else if (fu->flags & USBG_IS_BOT)
bot_cleanup_old_alt(fu);
fu->flags = 0;
}
static int tcm_setup(struct usb_function *f,
const struct usb_ctrlrequest *ctrl)
{
struct f_uas *fu = to_f_uas(f);
if (!(fu->flags & USBG_IS_BOT))
return -EOPNOTSUPP;
return usbg_bot_setup(f, ctrl);
}
static inline struct f_tcm_opts *to_f_tcm_opts(struct config_item *item)
{
return container_of(to_config_group(item), struct f_tcm_opts,
func_inst.group);
}
static void tcm_attr_release(struct config_item *item)
{
struct f_tcm_opts *opts = to_f_tcm_opts(item);
usb_put_function_instance(&opts->func_inst);
}
static struct configfs_item_operations tcm_item_ops = {
.release = tcm_attr_release,
};
static const struct config_item_type tcm_func_type = {
.ct_item_ops = &tcm_item_ops,
.ct_owner = THIS_MODULE,
};
static void tcm_free_inst(struct usb_function_instance *f)
{
struct f_tcm_opts *opts;
unsigned i;
opts = container_of(f, struct f_tcm_opts, func_inst);
mutex_lock(&tpg_instances_lock);
for (i = 0; i < TPG_INSTANCES; ++i)
if (tpg_instances[i].func_inst == f)
break;
if (i < TPG_INSTANCES)
tpg_instances[i].func_inst = NULL;
mutex_unlock(&tpg_instances_lock);
kfree(opts);
}
static int tcm_register_callback(struct usb_function_instance *f)
{
struct f_tcm_opts *opts = container_of(f, struct f_tcm_opts, func_inst);
mutex_lock(&opts->dep_lock);
opts->can_attach = true;
mutex_unlock(&opts->dep_lock);
return 0;
}
static void tcm_unregister_callback(struct usb_function_instance *f)
{
struct f_tcm_opts *opts = container_of(f, struct f_tcm_opts, func_inst);
mutex_lock(&opts->dep_lock);
unregister_gadget_item(opts->
func_inst.group.cg_item.ci_parent->ci_parent);
opts->can_attach = false;
mutex_unlock(&opts->dep_lock);
}
static int usbg_attach(struct usbg_tpg *tpg)
{
struct usb_function_instance *f = tpg->fi;
struct f_tcm_opts *opts = container_of(f, struct f_tcm_opts, func_inst);
if (opts->tcm_register_callback)
return opts->tcm_register_callback(f);
return 0;
}
static void usbg_detach(struct usbg_tpg *tpg)
{
struct usb_function_instance *f = tpg->fi;
struct f_tcm_opts *opts = container_of(f, struct f_tcm_opts, func_inst);
if (opts->tcm_unregister_callback)
opts->tcm_unregister_callback(f);
}
static int tcm_set_name(struct usb_function_instance *f, const char *name)
{
struct f_tcm_opts *opts = container_of(f, struct f_tcm_opts, func_inst);
pr_debug("tcm: Activating %s\n", name);
mutex_lock(&opts->dep_lock);
opts->ready = true;
mutex_unlock(&opts->dep_lock);
return 0;
}
static struct usb_function_instance *tcm_alloc_inst(void)
{
struct f_tcm_opts *opts;
int i;
opts = kzalloc(sizeof(*opts), GFP_KERNEL);
if (!opts)
return ERR_PTR(-ENOMEM);
mutex_lock(&tpg_instances_lock);
for (i = 0; i < TPG_INSTANCES; ++i)
if (!tpg_instances[i].func_inst)
break;
if (i == TPG_INSTANCES) {
mutex_unlock(&tpg_instances_lock);
kfree(opts);
return ERR_PTR(-EBUSY);
}
tpg_instances[i].func_inst = &opts->func_inst;
mutex_unlock(&tpg_instances_lock);
mutex_init(&opts->dep_lock);
opts->func_inst.set_inst_name = tcm_set_name;
opts->func_inst.free_func_inst = tcm_free_inst;
opts->tcm_register_callback = tcm_register_callback;
opts->tcm_unregister_callback = tcm_unregister_callback;
config_group_init_type_name(&opts->func_inst.group, "",
&tcm_func_type);
return &opts->func_inst;
}
static void tcm_free(struct usb_function *f)
{
struct f_uas *tcm = to_f_uas(f);
kfree(tcm);
}
static void tcm_unbind(struct usb_configuration *c, struct usb_function *f)
{
usb_free_all_descriptors(f);
}
static struct usb_function *tcm_alloc(struct usb_function_instance *fi)
{
struct f_uas *fu;
unsigned i;
mutex_lock(&tpg_instances_lock);
for (i = 0; i < TPG_INSTANCES; ++i)
if (tpg_instances[i].func_inst == fi)
break;
if (i == TPG_INSTANCES) {
mutex_unlock(&tpg_instances_lock);
return ERR_PTR(-ENODEV);
}
fu = kzalloc(sizeof(*fu), GFP_KERNEL);
if (!fu) {
mutex_unlock(&tpg_instances_lock);
return ERR_PTR(-ENOMEM);
}
fu->function.name = "Target Function";
fu->function.bind = tcm_bind;
fu->function.unbind = tcm_unbind;
fu->function.set_alt = tcm_set_alt;
fu->function.get_alt = tcm_get_alt;
fu->function.setup = tcm_setup;
fu->function.disable = tcm_disable;
fu->function.free_func = tcm_free;
fu->tpg = tpg_instances[i].tpg;
hash_init(fu->stream_hash);
mutex_unlock(&tpg_instances_lock);
return &fu->function;
}
DECLARE_USB_FUNCTION(tcm, tcm_alloc_inst, tcm_alloc);
static int __init tcm_init(void)
{
int ret;
ret = usb_function_register(&tcmusb_func);
if (ret)
return ret;
ret = target_register_template(&usbg_ops);
if (ret)
usb_function_unregister(&tcmusb_func);
return ret;
}
module_init(tcm_init);
static void __exit tcm_exit(void)
{
target_unregister_template(&usbg_ops);
usb_function_unregister(&tcmusb_func);
}
module_exit(tcm_exit);
MODULE_DESCRIPTION("Target based USB-Gadget");
MODULE_LICENSE("GPL");
MODULE_AUTHOR("Sebastian Andrzej Siewior");