Quellcode durchsuchen

[dm][nvmem] support nvmem

Support for NVMEM(Non Volatile Memory) devices like EEPROM, EFUSES...

Signed-off-by: GuEe-GUI <2991707448@qq.com>
GuEe-GUI vor 1 Monat
Ursprung
Commit
22a77f2694

+ 1 - 0
components/drivers/Kconfig

@@ -33,6 +33,7 @@ rsource "regulator/Kconfig"
 rsource "reset/Kconfig"
 rsource "thermal/Kconfig"
 rsource "virtio/Kconfig"
+rsource "nvmem/Kconfig"
 rsource "dma/Kconfig"
 rsource "mfd/Kconfig"
 rsource "ofw/Kconfig"

+ 81 - 0
components/drivers/include/drivers/nvmem.h

@@ -0,0 +1,81 @@
+/*
+ * Copyright (c) 2006-2023, RT-Thread Development Team
+ *
+ * SPDX-License-Identifier: Apache-2.0
+ *
+ * Change Logs:
+ * Date           Author       Notes
+ * 2023-02-25     GuEe-GUI     the first version
+ */
+
+#ifndef __NVMEM_H__
+#define __NVMEM_H__
+
+#include <ref.h>
+
+#include <drivers/ofw.h>
+#include <drivers/core/dm.h>
+#include <drivers/core/driver.h>
+
+struct rt_nvmem_device
+{
+    struct rt_device parent;
+
+    int cells_nr;
+    rt_list_t cell_nodes;
+
+    rt_ssize_t (*reg_read)(struct rt_nvmem_device *, int offset, void *val, rt_size_t bytes);
+    rt_ssize_t (*reg_write)(struct rt_nvmem_device *, int offset, void *val, rt_size_t bytes);
+
+    rt_ssize_t size;
+    int word_size;
+    int stride;
+
+    rt_bool_t read_only;
+    rt_bool_t ignore_wp;
+    rt_base_t wp_pin;
+    rt_uint8_t wp_pin_active;
+
+    struct rt_ref ref;
+    struct rt_spinlock spinlock;
+
+    void *priv;
+};
+
+struct rt_nvmem_cell
+{
+    rt_list_t list;
+
+    int index;
+    const char *id;
+    const rt_bool_t free_able;
+
+    rt_uint32_t offset;
+    rt_uint32_t bytes;
+    rt_uint32_t bit_offset;
+    rt_uint32_t nbits;
+
+    struct rt_ref ref;
+
+    struct rt_ofw_node *np;
+    struct rt_nvmem_device *nvmem;
+};
+
+rt_err_t rt_nvmem_device_register(struct rt_nvmem_device *ndev);
+rt_err_t rt_nvmem_device_unregister(struct rt_nvmem_device *ndev);
+
+rt_err_t rt_nvmem_device_append_cell(struct rt_nvmem_device *ndev, struct rt_nvmem_cell *cell);
+
+rt_ssize_t rt_nvmem_cell_read(struct rt_nvmem_cell *cell, void *buffer, rt_size_t len);
+rt_ssize_t rt_nvmem_cell_write(struct rt_nvmem_cell *cell, void *buffer, rt_size_t len);
+
+rt_ssize_t rt_nvmem_cell_read_u8(struct rt_nvmem_cell *cell, rt_uint8_t *out_val);
+rt_ssize_t rt_nvmem_cell_read_u16(struct rt_nvmem_cell *cell, rt_uint16_t *out_val);
+rt_ssize_t rt_nvmem_cell_read_u32(struct rt_nvmem_cell *cell, rt_uint32_t *out_val);
+rt_ssize_t rt_nvmem_cell_read_u64(struct rt_nvmem_cell *cell, rt_uint64_t *out_val);
+
+struct rt_nvmem_cell *rt_nvmem_get_cell_by_index(struct rt_device *dev, int index);
+struct rt_nvmem_cell *rt_nvmem_get_cell_by_name(struct rt_device *dev, const char *id);
+void rt_nvmem_put_cell(struct rt_nvmem_cell *cell);
+
+#endif /* __NVMEM_H__ */

+ 4 - 0
components/drivers/include/rtdevice.h

@@ -125,6 +125,10 @@ extern "C" {
 #ifdef RT_USING_THERMAL
 #include "drivers/thermal.h"
 #endif /* RT_USING_THERMAL */
+
+#ifdef RT_USING_NVMEM
+#include "drivers/nvmem.h"
+#endif /* RT_USING_NVMEM */
 #endif /* RT_USING_DM */
 
 #ifdef RT_USING_RTC

+ 12 - 0
components/drivers/nvmem/Kconfig

@@ -0,0 +1,12 @@
+menuconfig RT_USING_NVMEM
+    bool "Using Non Volatile Memory (NVMEM) device drivers"
+    depends on RT_USING_DM
+    depends on RT_USING_OFW
+    depends on RT_USING_PIN
+    select RT_USING_ADT
+    select RT_USING_ADT_REF
+    default n
+
+if RT_USING_NVMEM
+    osource "$(SOC_DM_NVMEM_DIR)/Kconfig"
+endif

+ 15 - 0
components/drivers/nvmem/SConscript

@@ -0,0 +1,15 @@
+from building import *
+
+group = []
+
+if not GetDepend(['RT_USING_NVMEM']):
+    Return('group')
+
+cwd = GetCurrentDir()
+CPPPATH = [cwd + '/../include']
+
+src = ['nvmem.c']
+
+group = DefineGroup('DeviceDrivers', src, depend = [''], CPPPATH = CPPPATH)
+
+Return('group')

+ 548 - 0
components/drivers/nvmem/nvmem.c

@@ -0,0 +1,548 @@
+/*
+ * Copyright (c) 2006-2023, RT-Thread Development Team
+ *
+ * SPDX-License-Identifier: Apache-2.0
+ *
+ * Change Logs:
+ * Date           Author       Notes
+ * 2023-02-25     GuEe-GUI     the first version
+ */
+
+#include <rtthread.h>
+#include <rtdevice.h>
+
+#define DBG_TAG "rtdm.nvmem"
+#define DBG_LVL DBG_INFO
+#include <rtdbg.h>
+
+rt_err_t rt_nvmem_device_register(struct rt_nvmem_device *ndev)
+{
+    struct rt_ofw_node *np;
+
+    if (!ndev)
+    {
+        return -RT_EINVAL;
+    }
+
+    np = ndev->parent.ofw_node;
+
+    if (!ndev->ignore_wp)
+    {
+        rt_uint8_t mode;
+
+        ndev->wp_pin = rt_pin_get_named_pin(&ndev->parent, "wp", 0,
+                &mode, &ndev->wp_pin_active);
+
+        if (ndev->wp_pin < 0 && ndev->wp_pin != PIN_NONE)
+        {
+            return -RT_EINVAL;
+        }
+        else if (ndev->wp_pin >= 0)
+        {
+            rt_pin_mode(ndev->wp_pin, mode);
+        }
+    }
+
+    if (!ndev->cells_nr)
+    {
+        rt_list_init(&ndev->cell_nodes);
+    }
+
+    rt_ref_init(&ndev->ref);
+
+    ndev->read_only = rt_dm_dev_prop_read_bool(&ndev->parent, "read-only") ||
+            ndev->read_only || !ndev->reg_write;
+
+    if (np)
+    {
+        rt_ofw_data(np) = ndev;
+    }
+
+    return RT_EOK;
+}
+
+rt_err_t rt_nvmem_device_unregister(struct rt_nvmem_device *ndev)
+{
+    if (!ndev)
+    {
+        return -RT_EINVAL;
+    }
+
+    if (rt_ref_read(&ndev->ref) != 1)
+    {
+        return -RT_EBUSY;
+    }
+
+    return RT_EOK;
+}
+
+rt_err_t rt_nvmem_device_append_cell(struct rt_nvmem_device *ndev,
+        struct rt_nvmem_cell *cell)
+{
+    rt_ubase_t level;
+
+    if (!ndev)
+    {
+        return -RT_EINVAL;
+    }
+
+    if (!ndev->cells_nr)
+    {
+        rt_list_init(&ndev->cell_nodes);
+    }
+
+    rt_list_init(&cell->list);
+
+    level = rt_spin_lock_irqsave(&ndev->spinlock);
+    rt_list_insert_before(&ndev->cell_nodes, &cell->list);
+    ++ndev->cells_nr;
+    rt_spin_unlock_irqrestore(&ndev->spinlock, level);
+
+    rt_ref_get(&ndev->ref);
+
+    return RT_EOK;
+}
+
+rt_ssize_t rt_nvmem_cell_read(struct rt_nvmem_cell *cell, void *buffer,
+        rt_size_t len)
+{
+    rt_ssize_t res;
+    struct rt_nvmem_device *nvmem;
+
+    if (!cell || !buffer || !len)
+    {
+        return -RT_EINVAL;
+    }
+
+    nvmem = cell->nvmem;
+
+    if (len > nvmem->size || len > cell->bytes)
+    {
+        return -RT_EINVAL;
+    }
+
+    if (!nvmem->reg_read)
+    {
+        return -RT_ENOSYS;
+    }
+
+    if ((res = nvmem->reg_read(nvmem, cell->offset, buffer, len)) < 0)
+    {
+        return res;
+    }
+
+    if (cell->bit_offset || cell->nbits)
+    {
+        /* Shift buffer */
+        rt_uint8_t *p, *b;
+        int extra, bit_offset = cell->bit_offset;
+
+        p = b = buffer;
+
+        if (bit_offset)
+        {
+            /* First shift */
+            *b++ >>= bit_offset;
+
+            /* Setup rest of the bytes if any */
+            for (int i = 1; i < cell->bytes; ++i)
+            {
+                /* Get bits from next byte and shift them towards msb */
+                *p |= *b << (RT_BITS_PER_BYTE - bit_offset);
+
+                p = b;
+                *b++ >>= bit_offset;
+            }
+        }
+        else
+        {
+            /* Point to the msb */
+            p += cell->bytes - 1;
+        }
+
+        /* Result fits in less bytes */
+        extra = cell->bytes - RT_DIV_ROUND_UP(cell->nbits, RT_BITS_PER_BYTE);
+        while (--extra >= 0)
+        {
+            *p-- = 0;
+        }
+
+        /* Clear msb bits if any leftover in the last byte */
+        if (cell->nbits % RT_BITS_PER_BYTE)
+        {
+            *p &= RT_GENMASK((cell->nbits % RT_BITS_PER_BYTE) - 1, 0);
+        }
+    }
+
+    return res;
+}
+
+rt_ssize_t rt_nvmem_cell_write(struct rt_nvmem_cell *cell, void *buffer,
+        rt_size_t len)
+{
+    rt_ssize_t res = 0;
+    struct rt_nvmem_device *nvmem;
+
+    if (!cell || !buffer || !len)
+    {
+        return -RT_EINVAL;
+    }
+
+    nvmem = cell->nvmem;
+
+    if (len > nvmem->size || len > cell->bytes)
+    {
+        return -RT_EINVAL;
+    }
+
+    if (!nvmem->reg_write)
+    {
+        return -RT_ENOSYS;
+    }
+
+    if (cell->bit_offset || cell->nbits)
+    {
+        /* Shift buffer */
+        int nbits, bit_offset = cell->bit_offset;
+        rt_uint8_t v, *p, *buf, *b, pbyte, pbits;
+
+        nbits = cell->nbits;
+        buf = rt_calloc(1, cell->bytes);
+
+        if (!buf)
+        {
+            return -RT_ENOMEM;
+        }
+
+        rt_memcpy(buf, buffer, len);
+        p = b = buf;
+
+        if (bit_offset)
+        {
+            pbyte = *b;
+            *b <<= bit_offset;
+
+            /* Setup the first byte with lsb bits from nvmem */
+            if ((res = nvmem->reg_read(nvmem, cell->offset, &v, 1)) < 0)
+            {
+                goto _end;
+            }
+            *b++ |= RT_GENMASK(bit_offset - 1, 0) & v;
+
+            /* Setup rest of the byte if any */
+            for (int i = 1; i < cell->bytes; ++i)
+            {
+                /* Get last byte bits and shift them towards lsb */
+                pbits = pbyte >> (RT_BITS_PER_BYTE - 1 - bit_offset);
+                pbyte = *b;
+                p = b;
+                *b <<= bit_offset;
+                *b++ |= pbits;
+            }
+        }
+
+        /* If it's not end on byte boundary */
+        if ((nbits + bit_offset) % RT_BITS_PER_BYTE)
+        {
+            /* Setup the last byte with msb bits from nvmem */
+            if ((res = nvmem->reg_read(nvmem, cell->offset + cell->bytes - 1, &v, 1)) < 0)
+            {
+                goto _end;
+            }
+
+            *p |= RT_GENMASK(7, (nbits + bit_offset) % RT_BITS_PER_BYTE) & v;
+        }
+
+        buffer = buf;
+
+    _end:
+        if (res < 0)
+        {
+            rt_free(buf);
+            return res;
+        }
+    }
+
+    if (nvmem->wp_pin >= 0)
+    {
+        rt_pin_write(nvmem->wp_pin, !nvmem->wp_pin_active);
+    }
+
+    res = nvmem->reg_write(nvmem, cell->offset, buffer, len);
+
+    if (nvmem->wp_pin >= 0)
+    {
+        rt_pin_write(nvmem->wp_pin, nvmem->wp_pin_active);
+    }
+
+    if (cell->bit_offset || cell->nbits)
+    {
+        rt_free(buffer);
+    }
+
+    return res;
+}
+
+rt_ssize_t rt_nvmem_cell_read_u8(struct rt_nvmem_cell *cell, rt_uint8_t *out_val)
+{
+    return rt_nvmem_cell_read(cell, out_val, sizeof(*out_val));
+}
+
+rt_ssize_t rt_nvmem_cell_read_u16(struct rt_nvmem_cell *cell, rt_uint16_t *out_val)
+{
+    return rt_nvmem_cell_read(cell, out_val, sizeof(*out_val));
+}
+
+rt_ssize_t rt_nvmem_cell_read_u32(struct rt_nvmem_cell *cell, rt_uint32_t *out_val)
+{
+    return rt_nvmem_cell_read(cell, out_val, sizeof(*out_val));
+}
+
+rt_ssize_t rt_nvmem_cell_read_u64(struct rt_nvmem_cell *cell, rt_uint64_t *out_val)
+{
+    return rt_nvmem_cell_read(cell, out_val, sizeof(*out_val));
+}
+
+static struct rt_nvmem_cell *ofw_nvmem_get_cell(struct rt_ofw_node *np,
+        int index, const char *const_id)
+{
+    rt_ubase_t level;
+    rt_ssize_t length;
+    const fdt32_t *addr;
+    struct rt_nvmem_device *nvmem;
+    struct rt_nvmem_cell *cell, *cell_tmp;
+    struct rt_ofw_cell_args cell_args;
+    struct rt_ofw_node *cell_np = RT_NULL, *nvmem_np = RT_NULL;
+
+    if (rt_ofw_parse_phandle_cells(np,
+            "nvmem-cells", "#nvmem-cell-cells", index, &cell_args))
+    {
+        return RT_NULL;
+    }
+
+    cell_np = cell_args.data;
+    index = cell_args.args_count ? cell_args.args[0] : 0;
+
+    if (!cell_np)
+    {
+        cell = rt_err_ptr(-RT_ERROR);
+        goto _put_node;
+    }
+
+    /* 1.find in ofw node */
+    if (!rt_ofw_data(cell_np))
+    {
+        rt_platform_ofw_request(cell_np);
+    }
+
+    cell = rt_ofw_data(cell_np);
+
+    if (cell && rt_ref_read(&cell->ref) > 0)
+    {
+        rt_ref_get(&cell->ref);
+        goto _put_node;
+    }
+
+    cell = rt_err_ptr(-RT_ERROR);
+    nvmem_np = rt_ofw_get_parent(cell_np);
+
+    if (!nvmem_np)
+    {
+        goto _put_node;
+    }
+
+    nvmem = rt_ofw_data(nvmem_np);
+
+    if (!nvmem)
+    {
+        goto _put_node;
+    }
+
+    level = rt_spin_lock_irqsave(&nvmem->spinlock);
+
+    /* 2.find in const node */
+    rt_list_for_each_entry(cell_tmp, &nvmem->cell_nodes, list)
+    {
+        if (cell_tmp->index == index)
+        {
+            if (const_id && cell_tmp->id && rt_strcmp(const_id, cell_tmp->id))
+            {
+                continue;
+            }
+
+            cell = cell_tmp;
+
+            if (rt_ref_read(&cell->ref))
+            {
+                rt_ref_get(&cell->ref);
+            }
+            else
+            {
+                cell = RT_NULL;
+            }
+
+            break;
+        }
+    }
+
+    rt_spin_unlock_irqrestore(&nvmem->spinlock, level);
+
+    if (!rt_is_err_or_null(cell))
+    {
+        goto _put_node;
+    }
+
+    /* 3.create a new one */
+    cell = rt_calloc(1, sizeof(*cell));
+
+    if (!cell)
+    {
+        cell = rt_err_ptr(-RT_ENOMEM);
+        LOG_E("No memory to create cell: %s (%d)", const_id, index);
+        goto _put_node;
+    }
+
+    cell->index = index;
+    cell->id = const_id;
+
+    *(rt_bool_t *)&cell->free_able = RT_TRUE;
+
+    addr = rt_ofw_prop_read_raw(cell_np, "reg", &length);
+    if (!addr || length < 2 * sizeof(rt_uint32_t))
+    {
+        LOG_E("%s Invalid reg", rt_ofw_node_full_name(cell_np));
+        goto _fail;
+    }
+
+    cell->offset = fdt32_to_cpu(*addr++);
+    cell->bytes = fdt32_to_cpu(*addr);
+
+    addr = rt_ofw_prop_read_raw(cell_np, "reg", &length);
+    if (addr && length == 2 * sizeof(rt_uint32_t))
+    {
+        cell->bit_offset = fdt32_to_cpu(*addr++);
+        cell->nbits = fdt32_to_cpu(*addr);
+    }
+
+    /* user ref is '1' */
+    rt_ref_init(&cell->ref);
+
+    cell->np = cell_np;
+    cell->nvmem = nvmem;
+
+    rt_nvmem_device_append_cell(nvmem, cell);
+
+    rt_ofw_node_get(cell_np);
+    rt_ofw_data(cell_np) = cell;
+
+    goto _put_node;
+
+_fail:
+    rt_free(cell);
+    cell = rt_err_ptr(-RT_EINVAL);
+
+_put_node:
+    rt_ofw_node_put(cell_np);
+    rt_ofw_node_put(nvmem_np);
+
+    return cell;
+}
+
+struct rt_nvmem_cell *rt_nvmem_get_cell_by_index(struct rt_device *dev,
+        int index)
+{
+    if (!dev || index < 0)
+    {
+        return RT_NULL;
+    }
+
+    if (dev->ofw_node)
+    {
+        return ofw_nvmem_get_cell(dev->ofw_node, index, RT_NULL);
+    }
+
+    return rt_err_ptr(-RT_ENOSYS);
+}
+
+struct rt_nvmem_cell *rt_nvmem_get_cell_by_name(struct rt_device *dev,
+        const char *id)
+{
+    struct rt_ofw_node *np;
+
+    if (!dev || !id)
+    {
+        return RT_NULL;
+    }
+
+    np = dev->ofw_node;
+
+    if (np)
+    {
+        int index = 0;
+        const char *const_id;
+        struct rt_ofw_prop *prop;
+
+        rt_ofw_foreach_prop_string(np, "nvmem-cell-names", prop, const_id)
+        {
+            if (!rt_strcmp(id, const_id))
+            {
+                return ofw_nvmem_get_cell(np, index, const_id);
+            }
+
+            ++index;
+        }
+    }
+
+    return rt_err_ptr(-RT_ENOSYS);
+}
+
+static void nvmem_release(struct rt_ref *r)
+{
+    struct rt_nvmem_device *ndev = rt_container_of(r, struct rt_nvmem_device, ref);
+
+    if (ndev->parent.ofw_node)
+    {
+        LOG_E("%s device is release", rt_ofw_node_full_name(ndev->parent.ofw_node));
+    }
+
+    RT_ASSERT(0);
+}
+
+static void cell_release(struct rt_ref *r)
+{
+    rt_ubase_t level;
+    struct rt_nvmem_cell *cell = rt_container_of(r, struct rt_nvmem_cell, ref);
+    struct rt_nvmem_device *nvmem = cell->nvmem;
+
+    if (!cell->free_able)
+    {
+        /* only freeable cells can enter */
+        LOG_E("%s cell is release", cell->id);
+        RT_ASSERT(0);
+    }
+
+    if (cell->np)
+    {
+        rt_ofw_data(cell->np) = RT_NULL;
+        rt_ofw_node_put(cell->np);
+    }
+
+    level = rt_spin_lock_irqsave(&nvmem->spinlock);
+    rt_list_remove(&cell->list);
+    --nvmem->cells_nr;
+    rt_spin_unlock_irqrestore(&nvmem->spinlock, level);
+
+    rt_ref_put(&nvmem->ref, &nvmem_release);
+
+    rt_free(cell);
+}
+
+void rt_nvmem_put_cell(struct rt_nvmem_cell *cell)
+{
+    if (!cell)
+    {
+        return;
+    }
+
+    rt_ref_put(&cell->ref, &cell_release);
+}