|
1 | 1 | /*
|
2 |
| - * SPDX-FileCopyrightText: 2023-2024 Espressif Systems (Shanghai) CO LTD |
| 2 | + * SPDX-FileCopyrightText: 2023-2025 Espressif Systems (Shanghai) CO LTD |
3 | 3 | *
|
4 | 4 | * SPDX-License-Identifier: Apache-2.0
|
5 | 5 | */
|
|
24 | 24 | static const char *TAG = "dma_utils";
|
25 | 25 |
|
26 | 26 | #define ALIGN_UP_BY(num, align) (((num) + ((align) - 1)) & ~((align) - 1))
|
27 |
| -#define ALIGN_DOWN_BY(num, align) ((num) & (~((align) - 1))) |
28 | 27 |
|
29 |
| -esp_err_t esp_dma_split_buffer_to_aligned(void *input_buffer, size_t input_buffer_len, void *stash_buffer, size_t stash_buffer_len, size_t split_alignment, dma_buffer_split_array_t *align_array) |
| 28 | +esp_err_t esp_dma_split_rx_buffer_to_cache_aligned(void *rx_buffer, size_t buffer_len, dma_buffer_split_array_t *align_buf_array, uint8_t** ret_stash_buffer) |
30 | 29 | {
|
31 |
| - esp_err_t ret = ESP_OK; |
32 |
| - ESP_RETURN_ON_FALSE(align_array && input_buffer && input_buffer_len && stash_buffer && split_alignment && !(split_alignment & (split_alignment - 1) |
33 |
| - && (stash_buffer_len >= 2 * split_alignment)), ESP_ERR_INVALID_ARG, TAG, "invalid argument"); |
34 |
| - ESP_RETURN_ON_FALSE(!((uintptr_t)stash_buffer % split_alignment), ESP_ERR_INVALID_ARG, TAG, "extra buffer is not aligned"); |
35 |
| - |
36 |
| - // calculate head_overflow_len |
37 |
| - size_t head_overflow_len = (uintptr_t)input_buffer % split_alignment; |
38 |
| - head_overflow_len = head_overflow_len ? split_alignment - head_overflow_len : 0; |
39 |
| - ESP_LOGD(TAG, "head_addr:%p split_alignment:%zu head_overflow_len:%zu", input_buffer, split_alignment, head_overflow_len); |
40 |
| - // calculate tail_overflow_len |
41 |
| - size_t tail_overflow_len = ((uintptr_t)input_buffer + input_buffer_len) % split_alignment; |
42 |
| - ESP_LOGD(TAG, "tail_addr:%p split_alignment:%zu tail_overflow_len:%zu", input_buffer + input_buffer_len - tail_overflow_len, split_alignment, tail_overflow_len); |
43 |
| - |
44 |
| - uint32_t extra_buf_count = 0; |
45 |
| - input_buffer = (uint8_t*)input_buffer; |
46 |
| - stash_buffer = (uint8_t*)stash_buffer; |
47 |
| - align_array->buf.head.recovery_address = input_buffer; |
48 |
| - align_array->buf.head.aligned_buffer = stash_buffer + split_alignment * extra_buf_count++; |
49 |
| - align_array->buf.head.length = head_overflow_len; |
50 |
| - align_array->buf.body.recovery_address = input_buffer + head_overflow_len; |
51 |
| - align_array->buf.body.aligned_buffer = input_buffer + head_overflow_len; |
52 |
| - align_array->buf.body.length = input_buffer_len - head_overflow_len - tail_overflow_len; |
53 |
| - align_array->buf.tail.recovery_address = input_buffer + input_buffer_len - tail_overflow_len; |
54 |
| - align_array->buf.tail.aligned_buffer = stash_buffer + split_alignment * extra_buf_count++; |
55 |
| - align_array->buf.tail.length = tail_overflow_len; |
56 |
| - |
57 |
| - // special handling when input_buffer length is no more than buffer alignment |
58 |
| - if(head_overflow_len >= input_buffer_len || tail_overflow_len >= input_buffer_len) |
59 |
| - { |
60 |
| - align_array->buf.head.length = input_buffer_len ; |
61 |
| - align_array->buf.body.length = 0 ; |
62 |
| - align_array->buf.tail.length = 0 ; |
| 30 | + ESP_RETURN_ON_FALSE(rx_buffer && buffer_len && align_buf_array, ESP_ERR_INVALID_ARG, TAG, "invalid argument"); |
| 31 | + |
| 32 | + // read the cache line size of internal and external memory, we also use this information to check if a given memory is behind the cache |
| 33 | + size_t int_mem_cache_line_size = cache_hal_get_cache_line_size(CACHE_LL_LEVEL_INT_MEM, CACHE_TYPE_DATA); |
| 34 | + size_t ext_mem_cache_line_size = cache_hal_get_cache_line_size(CACHE_LL_LEVEL_EXT_MEM, CACHE_TYPE_DATA); |
| 35 | + |
| 36 | + size_t split_line_size = 0; |
| 37 | + if (esp_ptr_external_ram(rx_buffer)) { |
| 38 | + split_line_size = ext_mem_cache_line_size; |
| 39 | + } else if (esp_ptr_internal(rx_buffer)) { |
| 40 | + split_line_size = int_mem_cache_line_size; |
| 41 | + } |
| 42 | + ESP_LOGV(TAG, "split_line_size:%zu", split_line_size); |
| 43 | + |
| 44 | + // allocate the stash buffer from internal RAM |
| 45 | + // Note, the split_line_size can be 0, in this case, the stash_buffer is also NULL, which is fine |
| 46 | + uint8_t* stash_buffer = heap_caps_calloc(2, split_line_size, MALLOC_CAP_DMA | MALLOC_CAP_INTERNAL | MALLOC_CAP_8BIT); |
| 47 | + ESP_RETURN_ON_FALSE(!(split_line_size && !stash_buffer), ESP_ERR_NO_MEM, TAG, "no mem for stash buffer"); |
| 48 | + |
| 49 | + // clear align_array to avoid garbage data |
| 50 | + memset(align_buf_array, 0, sizeof(dma_buffer_split_array_t)); |
| 51 | + bool need_cache_sync[3] = {false}; |
| 52 | + |
| 53 | + // if split_line_size is non-zero, split the buffer into head, body and tail |
| 54 | + if (split_line_size > 0) { |
| 55 | + // calculate head_overflow_len |
| 56 | + size_t head_overflow_len = (uintptr_t)rx_buffer % split_line_size; |
| 57 | + head_overflow_len = head_overflow_len ? split_line_size - head_overflow_len : 0; |
| 58 | + ESP_LOGV(TAG, "head_addr:%p head_overflow_len:%zu", rx_buffer, head_overflow_len); |
| 59 | + // calculate tail_overflow_len |
| 60 | + size_t tail_overflow_len = ((uintptr_t)rx_buffer + buffer_len) % split_line_size; |
| 61 | + ESP_LOGV(TAG, "tail_addr:%p tail_overflow_len:%zu", rx_buffer + buffer_len - tail_overflow_len, tail_overflow_len); |
| 62 | + |
| 63 | + uint8_t extra_buf_count = 0; |
| 64 | + uint8_t* input_buffer = (uint8_t*)rx_buffer; |
| 65 | + align_buf_array->buf.head.recovery_address = input_buffer; |
| 66 | + align_buf_array->buf.head.aligned_buffer = stash_buffer + split_line_size * extra_buf_count++; |
| 67 | + align_buf_array->buf.head.length = head_overflow_len; |
| 68 | + need_cache_sync[0] = int_mem_cache_line_size > 0; |
| 69 | + align_buf_array->buf.body.recovery_address = input_buffer + head_overflow_len; |
| 70 | + align_buf_array->buf.body.aligned_buffer = input_buffer + head_overflow_len; |
| 71 | + align_buf_array->buf.body.length = buffer_len - head_overflow_len - tail_overflow_len; |
| 72 | + need_cache_sync[1] = true; |
| 73 | + align_buf_array->buf.tail.recovery_address = input_buffer + buffer_len - tail_overflow_len; |
| 74 | + align_buf_array->buf.tail.aligned_buffer = stash_buffer + split_line_size * extra_buf_count++; |
| 75 | + align_buf_array->buf.tail.length = tail_overflow_len; |
| 76 | + need_cache_sync[2] = int_mem_cache_line_size > 0; |
| 77 | + |
| 78 | + // special handling when input_buffer length is no more than buffer alignment |
| 79 | + if (head_overflow_len >= buffer_len || tail_overflow_len >= buffer_len) { |
| 80 | + align_buf_array->buf.head.length = buffer_len ; |
| 81 | + align_buf_array->buf.body.length = 0 ; |
| 82 | + align_buf_array->buf.tail.length = 0 ; |
| 83 | + } |
| 84 | + } else { |
| 85 | + align_buf_array->buf.body.aligned_buffer = rx_buffer; |
| 86 | + align_buf_array->buf.body.recovery_address = rx_buffer; |
| 87 | + align_buf_array->buf.body.length = buffer_len; |
| 88 | + need_cache_sync[1] = false; |
63 | 89 | }
|
64 | 90 |
|
65 |
| - for(int i = 0; i < 3; i++) { |
66 |
| - if(!align_array->aligned_buffer[i].length) { |
67 |
| - align_array->aligned_buffer[i].aligned_buffer = NULL; |
68 |
| - align_array->aligned_buffer[i].recovery_address = NULL; |
| 91 | + for (int i = 0; i < 3; i++) { |
| 92 | + if (align_buf_array->aligned_buffer[i].length == 0) { |
| 93 | + align_buf_array->aligned_buffer[i].aligned_buffer = NULL; |
| 94 | + align_buf_array->aligned_buffer[i].recovery_address = NULL; |
| 95 | + need_cache_sync[i] = false; |
69 | 96 | }
|
70 | 97 | }
|
71 | 98 |
|
72 |
| - return ret; |
| 99 | + // invalidate the aligned buffer if necessary |
| 100 | + for (int i = 0; i < 3; i++) { |
| 101 | + if (need_cache_sync[i]) { |
| 102 | + esp_cache_msync(align_buf_array->aligned_buffer[i].aligned_buffer, align_buf_array->aligned_buffer[i].length, ESP_CACHE_MSYNC_FLAG_DIR_M2C); |
| 103 | + } |
| 104 | + } |
| 105 | + |
| 106 | + *ret_stash_buffer = stash_buffer; |
| 107 | + return ESP_OK; |
73 | 108 | }
|
74 | 109 |
|
75 |
| -esp_err_t esp_dma_merge_aligned_buffers(dma_buffer_split_array_t *align_array) |
| 110 | +esp_err_t esp_dma_merge_aligned_rx_buffers(dma_buffer_split_array_t *align_array) |
76 | 111 | {
|
77 |
| - esp_err_t ret = ESP_OK; |
78 |
| - ESP_RETURN_ON_FALSE(align_array, ESP_ERR_INVALID_ARG, TAG, "invalid argument"); |
| 112 | + ESP_RETURN_ON_FALSE_ISR(align_array, ESP_ERR_INVALID_ARG, TAG, "invalid argument"); |
79 | 113 |
|
80 | 114 | // only need to copy the head and tail buffer
|
81 |
| - if(align_array->buf.head.length) { |
| 115 | + if (align_array->buf.head.length) { |
82 | 116 | memcpy(align_array->buf.head.recovery_address, align_array->buf.head.aligned_buffer, align_array->buf.head.length);
|
83 | 117 | }
|
84 |
| - if(align_array->buf.tail.length) { |
| 118 | + if (align_array->buf.tail.length) { |
85 | 119 | memcpy(align_array->buf.tail.recovery_address, align_array->buf.tail.aligned_buffer, align_array->buf.tail.length);
|
86 | 120 | }
|
87 |
| - |
88 |
| - return ret; |
| 121 | + return ESP_OK; |
89 | 122 | }
|
90 | 123 |
|
91 | 124 | esp_err_t esp_dma_capable_malloc(size_t size, const esp_dma_mem_info_t *dma_mem_info, void **out_ptr, size_t *actual_size)
|
|
0 commit comments