xiaozhengsheng 6df0f7d96e 初始版本
2025-08-19 09:49:41 +08:00

400 lines
12 KiB
C

/**
* Copyright (c) 2016 - 2020, Nordic Semiconductor ASA
*
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without modification,
* are permitted provided that the following conditions are met:
*
* 1. Redistributions of source code must retain the above copyright notice, this
* list of conditions and the following disclaimer.
*
* 2. Redistributions in binary form, except as embedded into a Nordic
* Semiconductor ASA integrated circuit in a product or a software update for
* such product, must reproduce the above copyright notice, this list of
* conditions and the following disclaimer in the documentation and/or other
* materials provided with the distribution.
*
* 3. Neither the name of Nordic Semiconductor ASA nor the names of its
* contributors may be used to endorse or promote products derived from this
* software without specific prior written permission.
*
* 4. This software, with or without modification, must only be used with a
* Nordic Semiconductor ASA integrated circuit.
*
* 5. Any software provided in binary form under this license must not be reverse
* engineered, decompiled, modified and/or disassembled.
*
* THIS SOFTWARE IS PROVIDED BY NORDIC SEMICONDUCTOR ASA "AS IS" AND ANY EXPRESS
* OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
* OF MERCHANTABILITY, NONINFRINGEMENT, AND FITNESS FOR A PARTICULAR PURPOSE ARE
* DISCLAIMED. IN NO EVENT SHALL NORDIC SEMICONDUCTOR ASA OR CONTRIBUTORS BE
* LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE
* GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT
* OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*
*/
#include "nrf_block_dev_sdc.h"
/**@file
*
* @ingroup nrf_block_dev_sdc
* @{
*
* @brief This module implements block device API. It should be used as a reference block device.
*/
static volatile sdc_result_t m_last_result;
/**
* @brief Active SDC block device handle. Only one instance.
* */
static nrf_block_dev_sdc_t const * m_active_sdc_dev;
static ret_code_t block_dev_sdc_uninit(nrf_block_dev_t const * p_blk_dev);
static void wait_func(void)
{
}
static void sdc_wait()
{
while (app_sdc_busy_check())
{
wait_func();
}
}
static void sdc_handler(sdc_evt_t const * p_event)
{
m_last_result = p_event->result;
nrf_block_dev_sdc_t const * p_sdc_dev = m_active_sdc_dev;
nrf_block_dev_sdc_work_t * p_work = p_sdc_dev->p_work;
switch (p_event->type)
{
case SDC_EVT_INIT:
{
if (p_event->result != SDC_SUCCESS)
{
nrf_block_dev_t const * block_dev = &(p_sdc_dev->block_dev);
ret_code_t err_code = block_dev_sdc_uninit(block_dev);
APP_ERROR_CHECK(err_code);
}
p_work->geometry.blk_count = app_sdc_info_get()->num_blocks;
p_work->geometry.blk_size = SDC_SECTOR_SIZE;
if (m_active_sdc_dev->p_work->ev_handler)
{
const nrf_block_dev_event_t ev = {
NRF_BLOCK_DEV_EVT_INIT,
((p_event->result == SDC_SUCCESS) ? \
NRF_BLOCK_DEV_RESULT_SUCCESS : NRF_BLOCK_DEV_RESULT_IO_ERROR),
NULL,
p_work->p_context
};
p_work->ev_handler(&p_sdc_dev->block_dev, &ev);
}
}
break;
case SDC_EVT_READ:
if (m_active_sdc_dev->p_work->ev_handler)
{
const nrf_block_dev_event_t ev = {
NRF_BLOCK_DEV_EVT_BLK_READ_DONE,
((p_event->result == SDC_SUCCESS) ? \
NRF_BLOCK_DEV_RESULT_SUCCESS : NRF_BLOCK_DEV_RESULT_IO_ERROR),
&p_work->req,
p_work->p_context
};
p_work->ev_handler(&p_sdc_dev->block_dev, &ev);
}
break;
case SDC_EVT_WRITE:
if (m_active_sdc_dev->p_work->ev_handler)
{
const nrf_block_dev_event_t ev = {
NRF_BLOCK_DEV_EVT_BLK_WRITE_DONE,
((p_event->result == SDC_SUCCESS) ? \
NRF_BLOCK_DEV_RESULT_SUCCESS : NRF_BLOCK_DEV_RESULT_IO_ERROR),
&p_work->req,
p_work->p_context
};
p_work->ev_handler(&p_sdc_dev->block_dev, &ev);
}
break;
default:
APP_ERROR_CHECK(NRF_ERROR_INTERNAL);
return;
}
}
static ret_code_t block_dev_sdc_init(nrf_block_dev_t const * p_blk_dev,
nrf_block_dev_ev_handler ev_handler,
void const * p_context)
{
ASSERT(p_blk_dev);
nrf_block_dev_sdc_t const * p_sdc_dev =
CONTAINER_OF(p_blk_dev, nrf_block_dev_sdc_t, block_dev);
nrf_block_dev_sdc_work_t * p_work = p_sdc_dev->p_work;
if (p_sdc_dev->sdc_bdev_config.block_size != SDC_SECTOR_SIZE)
{
/* Unsupported block size. */
return NRF_ERROR_NOT_SUPPORTED;
}
if (m_active_sdc_dev)
{
/* SDC instance is busy. */
return NRF_ERROR_BUSY;
}
p_work->p_context = p_context;
p_work->ev_handler = ev_handler;
m_active_sdc_dev = p_sdc_dev;
ret_code_t err_code = NRF_SUCCESS;
err_code = app_sdc_init(&p_sdc_dev->sdc_bdev_config.sdc_config, sdc_handler);
if (err_code == NRF_SUCCESS)
{
if (!ev_handler)
{
/* Synchronous mode - wait for the card. */
sdc_wait();
err_code = ((m_last_result == SDC_SUCCESS) ? NRF_SUCCESS : NRF_ERROR_TIMEOUT);
}
}
if (err_code != NRF_SUCCESS)
{
m_active_sdc_dev = NULL;
if (ev_handler)
{
/* Call the user handler with an error status. */
const nrf_block_dev_event_t ev = {
NRF_BLOCK_DEV_EVT_INIT,
NRF_BLOCK_DEV_RESULT_IO_ERROR,
NULL,
p_work->p_context
};
p_work->ev_handler(p_blk_dev, &ev);
}
}
return err_code;
}
static ret_code_t block_dev_sdc_uninit(nrf_block_dev_t const * p_blk_dev)
{
ASSERT(p_blk_dev);
nrf_block_dev_sdc_t const * p_sdc_dev =
CONTAINER_OF(p_blk_dev, nrf_block_dev_sdc_t, block_dev);
nrf_block_dev_sdc_work_t * p_work = p_sdc_dev->p_work;
if (m_active_sdc_dev != p_sdc_dev)
{
/* SDC instance is busy. */
return NRF_ERROR_BUSY;
}
if (app_sdc_busy_check())
{
/* Previous asynchronous operation in progress. */
return NRF_ERROR_BUSY;
}
ret_code_t err_code = app_sdc_uninit();
if (err_code == NRF_SUCCESS)
{
/* Free the instance on success. */
m_active_sdc_dev = NULL;
}
if (p_work->ev_handler)
{
/* SDC uninitialization is a synchronous operation. Call event handler. */
const nrf_block_dev_event_t ev = {
NRF_BLOCK_DEV_EVT_UNINIT,
((err_code == NRF_SUCCESS) ? \
NRF_BLOCK_DEV_RESULT_SUCCESS : NRF_BLOCK_DEV_RESULT_IO_ERROR),
NULL,
p_work->p_context
};
p_work->ev_handler(p_blk_dev, &ev);
}
return err_code;
}
static ret_code_t block_dev_sdc_read_req(nrf_block_dev_t const * p_blk_dev,
nrf_block_req_t const * p_blk)
{
ASSERT(p_blk_dev);
ASSERT(p_blk);
nrf_block_dev_sdc_t const * p_sdc_dev =
CONTAINER_OF(p_blk_dev, nrf_block_dev_sdc_t, block_dev);
nrf_block_dev_sdc_work_t * p_work = p_sdc_dev->p_work;
ret_code_t err_code = NRF_SUCCESS;
if (m_active_sdc_dev != p_sdc_dev)
{
/* SDC instance is busy. */
return NRF_ERROR_BUSY;
}
if (app_sdc_busy_check())
{
/* Previous asynchronous operation in progress. */
return NRF_ERROR_BUSY;
}
p_work->req = *p_blk;
err_code = app_sdc_block_read(p_blk->p_buff, p_blk->blk_id, p_blk->blk_count);
if (err_code == NRF_SUCCESS)
{
if (!p_work->ev_handler)
{
/* Synchronous mode - wait for the card. */
sdc_wait();
err_code = ((m_last_result == SDC_SUCCESS) ? NRF_SUCCESS : NRF_ERROR_TIMEOUT);
}
}
if ((p_work->ev_handler) && (err_code != NRF_SUCCESS))
{
/* Call the user handler with an error status. */
const nrf_block_dev_event_t ev = {
NRF_BLOCK_DEV_EVT_BLK_READ_DONE,
NRF_BLOCK_DEV_RESULT_IO_ERROR,
&p_work->req,
p_work->p_context
};
p_work->ev_handler(p_blk_dev, &ev);
}
return err_code;
}
static ret_code_t block_dev_sdc_write_req(nrf_block_dev_t const * p_blk_dev,
nrf_block_req_t const * p_blk)
{
ASSERT(p_blk_dev);
ASSERT(p_blk);
nrf_block_dev_sdc_t const * p_sdc_dev =
CONTAINER_OF(p_blk_dev, nrf_block_dev_sdc_t, block_dev);
nrf_block_dev_sdc_work_t * p_work = p_sdc_dev->p_work;
ret_code_t err_code = NRF_SUCCESS;
if (m_active_sdc_dev != p_sdc_dev)
{
/* SDC instance is busy. */
return NRF_ERROR_BUSY;
}
if (app_sdc_busy_check())
{
/* Previous asynchronous operation in progress. */
return NRF_ERROR_BUSY;
}
p_work->req = *p_blk;
err_code = app_sdc_block_write(p_blk->p_buff, p_blk->blk_id, p_blk->blk_count);
if (err_code == NRF_SUCCESS)
{
if (!p_work->ev_handler)
{
/* Synchronous mode - wait for the card. */
sdc_wait();
err_code = ((m_last_result == SDC_SUCCESS) ? NRF_SUCCESS : NRF_ERROR_TIMEOUT);
}
}
if ((p_work->ev_handler) && (err_code != NRF_SUCCESS))
{
/* Call the user handler with an error status. */
const nrf_block_dev_event_t ev = {
NRF_BLOCK_DEV_EVT_BLK_READ_DONE,
NRF_BLOCK_DEV_RESULT_IO_ERROR,
&p_work->req,
p_work->p_context
};
p_work->ev_handler(p_blk_dev, &ev);
}
return err_code;
}
static ret_code_t block_dev_sdc_ioctl(nrf_block_dev_t const * p_blk_dev,
nrf_block_dev_ioctl_req_t req,
void * p_data)
{
nrf_block_dev_sdc_t const * p_sdc_dev =
CONTAINER_OF(p_blk_dev, nrf_block_dev_sdc_t, block_dev);
switch (req)
{
case NRF_BLOCK_DEV_IOCTL_REQ_CACHE_FLUSH:
{
bool * p_flushing = p_data;
if (p_flushing)
{
*p_flushing = false;
}
return NRF_SUCCESS;
}
case NRF_BLOCK_DEV_IOCTL_REQ_INFO_STRINGS:
{
if (p_data == NULL)
{
return NRF_ERROR_INVALID_PARAM;
}
nrf_block_dev_info_strings_t const * * pp_strings = p_data;
*pp_strings = &p_sdc_dev->info_strings;
return NRF_SUCCESS;
}
default:
break;
}
return NRF_ERROR_NOT_SUPPORTED;
}
static nrf_block_dev_geometry_t const * block_dev_sdc_geometry(nrf_block_dev_t const * p_blk_dev)
{
ASSERT(p_blk_dev);
nrf_block_dev_sdc_t const * p_sdc_dev =
CONTAINER_OF(p_blk_dev, nrf_block_dev_sdc_t, block_dev);
nrf_block_dev_sdc_work_t const * p_work = p_sdc_dev->p_work;
return &p_work->geometry;
}
const nrf_block_dev_ops_t nrf_block_device_sdc_ops = {
.init = block_dev_sdc_init,
.uninit = block_dev_sdc_uninit,
.read_req = block_dev_sdc_read_req,
.write_req = block_dev_sdc_write_req,
.ioctl = block_dev_sdc_ioctl,
.geometry = block_dev_sdc_geometry,
};
/** @} */