mirror of
				https://github.com/projectacrn/acrn-hypervisor.git
				synced 2025-10-31 01:09:28 +00:00 
			
		
		
		
	internal commit: 14ac2bc2299032fa6714d1fefa7cf0987b3e3085 Signed-off-by: Eddie Dong <eddie.dong@intel.com>
		
			
				
	
	
		
			195 lines
		
	
	
		
			5.2 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			195 lines
		
	
	
		
			5.2 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * SHARED BUFFER
 | |
|  *
 | |
|  * Copyright (C) 2017 Intel Corporation. All rights reserved.
 | |
|  *
 | |
|  * Redistribution and use in source and binary forms, with or without
 | |
|  * modification, are permitted provided that the following conditions
 | |
|  * are met:
 | |
|  *
 | |
|  *   * Redistributions of source code must retain the above copyright
 | |
|  *     notice, this list of conditions and the following disclaimer.
 | |
|  *   * Redistributions in binary form must reproduce the above copyright
 | |
|  *     notice, this list of conditions and the following disclaimer in
 | |
|  *     the documentation and/or other materials provided with the
 | |
|  *     distribution.
 | |
|  *   * Neither the name of Intel Corporation nor the names of its
 | |
|  *     contributors may be used to endorse or promote products derived
 | |
|  *     from this software without specific prior written permission.
 | |
|  *
 | |
|  * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
 | |
|  * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
 | |
|  * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
 | |
|  * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
 | |
|  * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
 | |
|  * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
 | |
|  * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
 | |
|  * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
 | |
|  * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
 | |
|  * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
 | |
|  * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
 | |
|  *
 | |
|  * Li Fei <fei1.li@intel.com>
 | |
|  *
 | |
|  */
 | |
| 
 | |
| #include <hv_lib.h>
 | |
| #include <acrn_common.h>
 | |
| #include <hv_arch.h>
 | |
| #include <hv_debug.h>
 | |
| 
 | |
| DEFINE_CPU_DATA(uint64_t * [ACRN_SBUF_ID_MAX], sbuf);
 | |
| 
 | |
| static inline bool sbuf_is_empty(struct shared_buf *sbuf)
 | |
| {
 | |
| 	return (sbuf->head == sbuf->tail);
 | |
| }
 | |
| 
 | |
| static inline uint32_t sbuf_next_ptr(uint32_t pos,
 | |
| 		uint32_t span, uint32_t scope)
 | |
| {
 | |
| 	pos += span;
 | |
| 	pos = (pos >= scope) ? (pos - scope) : pos;
 | |
| 	return pos;
 | |
| }
 | |
| 
 | |
| static inline uint32_t sbuf_calculate_allocate_size(uint32_t ele_num,
 | |
| 						uint32_t ele_size)
 | |
| {
 | |
| 	uint64_t sbuf_allocate_size;
 | |
| 
 | |
| 	sbuf_allocate_size = ele_num * ele_size;
 | |
| 	sbuf_allocate_size +=  SBUF_HEAD_SIZE;
 | |
| 	if (sbuf_allocate_size > SBUF_MAX_SIZE) {
 | |
| 		pr_err("%s, num=0x%x, size=0x%x exceed 0x%x",
 | |
| 			__func__, ele_num, ele_size, SBUF_MAX_SIZE);
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	return sbuf_allocate_size;
 | |
| }
 | |
| 
 | |
| struct shared_buf *sbuf_allocate(uint32_t ele_num, uint32_t ele_size)
 | |
| {
 | |
| 	struct shared_buf *sbuf;
 | |
| 	uint32_t sbuf_allocate_size;
 | |
| 
 | |
| 	if (!ele_num || !ele_size) {
 | |
| 		pr_err("%s invalid parameter!", __func__);
 | |
| 		return NULL;
 | |
| 	}
 | |
| 
 | |
| 	sbuf_allocate_size = sbuf_calculate_allocate_size(ele_num, ele_size);
 | |
| 	if (!sbuf_allocate_size)
 | |
| 		return NULL;
 | |
| 
 | |
| 	sbuf = malloc(sbuf_allocate_size);
 | |
| 	if (sbuf == NULL) {
 | |
| 		pr_err("%s no memory!", __func__);
 | |
| 		return NULL;
 | |
| 	}
 | |
| 
 | |
| 	memset(sbuf, 0, SBUF_HEAD_SIZE);
 | |
| 	sbuf->ele_num = ele_num;
 | |
| 	sbuf->ele_size = ele_size;
 | |
| 	sbuf->size = ele_num * ele_size;
 | |
| 	sbuf->magic = SBUF_MAGIC;
 | |
| 	pr_info("%s ele_num=0x%x, ele_size=0x%x allocated",
 | |
| 			__func__, ele_num, ele_size);
 | |
| 	return sbuf;
 | |
| }
 | |
| 
 | |
| void sbuf_free(struct shared_buf *sbuf)
 | |
| {
 | |
| 	if ((sbuf == NULL) || sbuf->magic != SBUF_MAGIC) {
 | |
| 		pr_err("%s invalid parameter!", __func__);
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	sbuf->magic = 0;
 | |
| 	free(sbuf);
 | |
| }
 | |
| 
 | |
| int sbuf_get(struct shared_buf *sbuf, uint8_t *data)
 | |
| {
 | |
| 	const void *from;
 | |
| 
 | |
| 	if ((sbuf == NULL) || (data == NULL))
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	if (sbuf_is_empty(sbuf)) {
 | |
| 		/* no data available */
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	from = (void *)sbuf + SBUF_HEAD_SIZE + sbuf->head;
 | |
| 
 | |
| 	memcpy_s((void *)data, sbuf->ele_size, from, sbuf->ele_size);
 | |
| 
 | |
| 	sbuf->head = sbuf_next_ptr(sbuf->head, sbuf->ele_size, sbuf->size);
 | |
| 
 | |
| 	return sbuf->ele_size;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * The high caller should guarantee each time there must have
 | |
|  * sbuf->ele_size data can be write form data and this function
 | |
|  * should guarantee execution atomically.
 | |
|  *
 | |
|  * flag:
 | |
|  * If OVERWRITE_EN set, buf can store (ele_num - 1) elements at most.
 | |
|  * Should use lock to guarantee that only one read or write at
 | |
|  * the same time.
 | |
|  * if OVERWRITE_EN not set, buf can store (ele_num - 1) elements
 | |
|  * at most. Shouldn't modify the sbuf->head.
 | |
|  *
 | |
|  * return:
 | |
|  * ele_size:	write succeeded.
 | |
|  * 0:		no write, buf is full
 | |
|  * negative:	failed.
 | |
|  */
 | |
| 
 | |
| int sbuf_put(struct shared_buf *sbuf, uint8_t *data)
 | |
| {
 | |
| 	void *to;
 | |
| 	uint32_t next_tail;
 | |
| 	bool trigger_overwrite = false;
 | |
| 
 | |
| 	if ((sbuf == NULL) || (data == NULL))
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	next_tail = sbuf_next_ptr(sbuf->tail, sbuf->ele_size, sbuf->size);
 | |
| 	/* if this write would trigger overrun */
 | |
| 	if (next_tail == sbuf->head) {
 | |
| 		/* accumulate overrun count if necessary */
 | |
| 		sbuf->overrun_cnt += sbuf->flags & OVERRUN_CNT_EN;
 | |
| 		if (!(sbuf->flags & OVERWRITE_EN)) {
 | |
| 			/* if not enable over write, return here. */
 | |
| 			return 0;
 | |
| 		}
 | |
| 		trigger_overwrite = true;
 | |
| 	}
 | |
| 
 | |
| 	to = (void *)sbuf + SBUF_HEAD_SIZE + sbuf->tail;
 | |
| 
 | |
| 	memcpy_s(to, sbuf->ele_size, data, sbuf->ele_size);
 | |
| 
 | |
| 	if (trigger_overwrite) {
 | |
| 		sbuf->head = sbuf_next_ptr(sbuf->head,
 | |
| 				sbuf->ele_size, sbuf->size);
 | |
| 	}
 | |
| 	sbuf->tail = next_tail;
 | |
| 
 | |
| 	return sbuf->ele_size;
 | |
| }
 | |
| 
 | |
| int sbuf_share_setup(uint32_t pcpu_id, uint32_t sbuf_id, uint64_t *hva)
 | |
| {
 | |
| 	if (pcpu_id >= (uint32_t) phy_cpu_num ||
 | |
| 			sbuf_id >= ACRN_SBUF_ID_MAX)
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	per_cpu(sbuf, pcpu_id)[sbuf_id] = hva;
 | |
| 	return 0;
 | |
| }
 |