FreeCalypso > hg > freecalypso-sw
diff gsm-fw/riviera/rvm/rvm_swe_hdlr.c @ 143:afceeeb2cba1
Our nuc-fw is destined to become gsm-fw, so I went ahead and did the big hg mv
author | Michael Spacefalcon <msokolov@ivan.Harhan.ORG> |
---|---|
date | Tue, 12 Nov 2013 05:35:48 +0000 |
parents | nuc-fw/riviera/rvm/rvm_swe_hdlr.c@15e972110527 |
children |
line wrap: on
line diff
--- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/gsm-fw/riviera/rvm/rvm_swe_hdlr.c Tue Nov 12 05:35:48 2013 +0000 @@ -0,0 +1,1251 @@ +/** + * + * @file rvm_swe_hdlr.c + * + * This file contains the functions related to SWEs management within RVM. + * + * @author David Lamy-Charrier (d-lamy@ti.com) + * @version 0.1 + * + */ + +/* + * Revision History: + * + * 10/26/2001 David Lamy-Charrier Create for Riviera 1.6. + * + * (C) Copyright 2001 by Texas Instruments Incorporated, All Rights Reserved + */ +#include <stdio.h> +#include "../rvf/rvf_env.h" +#include "rvm_gen.h" +#include "rvm_api.h" +#include "rvm_i.h" + +#include "../rvf/rvf_i.h" /* ET2 rvf private invocation API */ +#include "../rvf/rvf_api.h" /* A-M-E-N-D-E-D! */ + +/* temporary inclusion for HCI pb on WINDOWS */ +/* TO DO: remove it. */ +#include "rvm_use_id_list.h" + +#include <string.h> + +extern T_RVM_CONST_SWE_INFO RVM_SWE_GET_INFO_ARRAY[]; + +extern T_RVM_USE_ID * RVM_TYPE2_SWE_GROUPS[]; + +extern BOOLEAN rvm_allocated_task_id [MAX_RVF_TASKS]; + +extern T_RVM_KNOWN_SWE * rvm_swe_array; + + +/* private */ +T_RVM_RETURN _fatal(T_RVM_PROCESSING_SWE* appli, UINT8 rm); + +/*********************************************************************** +* Function _resolve_t2_grouping (private) +* +* Description resolves number of group directives & ret. group count +*************************************************************************/ +UINT8 _resolve_t2_grouping(T_RVM_PROCESSING_SWE* appli, T_RVM_GROUP_DIRECTIVE* gd) { + T_RVM_INFO_SWE swe_info; + T_RVM_PROCESSING_SWE* cur_swe = appli; + UINT8 i=0, j=0, k=0; + + for(; cur_swe != NULL;) { + UINT8 swe_index = cur_swe->swe_id; + + rvm_swe_array[swe_index].swe_get_info(&swe_info); + + if (rvm_swe_array[swe_index].swe_state !=SWE_RUNNING && //== SWE_NOT_STARTED && + swe_info.swe_type==RVM_SWE_TYPE_2) { + + for(i=0; i<MAX_GRPS; i++) { + if(swe_info.type_info.type2.swe_group_directive == gd[i].group_directive) { + for(k=0; gd[i].hosted_swe_db_index[k]!=0; k++); + if(k<MAX_COMPOSITES) { + gd[i].hosted_swe_db_index[k]=swe_index; + } else { + /* TO DO ... ERROR !!! */ + } +// RVM_TRACE_WARNING_PARAM("rvm.SweHndlr.resolve_t2_grouping(), appended to grp entry , nb=",\ +// (UINT32)swe_index); +//printf("rvm.SweHndlr.resolve_t2_grouping(): appended %d to group: %d\n",gd[i].hosted_swe_db_index[k], gd[i].host_task_addr); + + break; + } else if( swe_info.type_info.type2.swe_group_directive != gd[i].group_directive && + gd[i].host_task_addr==0 ) { + + /* Constraint! Expects all group priorites and stack sz to be equal + * Additional method must be used to set highest entity pri. or resolve */ + gd[i].host_task_addr=RVF_INVALID_ADDR_ID; //rvm_allocate_task_id(1); + gd[i].group_directive=swe_info.type_info.type2.swe_group_directive; + gd[i].task_priority=swe_info.type_info.type2.priority; + gd[i].stack_size=swe_info.type_info.type2.stack_size; + + gd[i].hosted_swe_db_index[0]=swe_index; + j++; +// RVM_TRACE_WARNING_PARAM("rvm.SweHndlr.resolve_t2_grouping(), created grp entry , nb=",\ +// (UINT32)swe_index); +//printf("rvm.SweHndlr.resolve_t2_grouping(): created host group: %d AND append %d\n",gd[i].host_task_addr, gd[i].hosted_swe_db_index[0]); + break; + } + } + + } else RVM_TRACE_WARNING_PARAM("rvm.SweHndlr.resolve_t2_grouping(), SWE Not type 2: ", rvm_swe_array[swe_index].swe_use_id); + cur_swe = cur_swe->next_swe; /* process next SWE */ + } +//printf("rvm.SweHndlr.resolve_t2_grouping(): total group count: %d\n", j); + +//for(i=0; i<j; i++) /* de'bugger only!! */ +// for(k=0; k<MAX_COMPOSITES && gd[i].hosted_swe_db_index[k]!=0; k++) +// printf("host addr: %d, T2 swe_db_index %d\n", +// gd[i].host_task_addr, gd[i].hosted_swe_db_index[k]); + + return j; +} + +/******************************************************************************* +** Function rvm_allocate_task_id +** +** Description Internal function which allocate the first available +** task id to a SWE in creation +*******************************************************************************/ +T_RVM_TASK_ID rvm_allocate_task_id(UINT8 isRealTask) { +/* UINT8 i=0; */ + + /* Find the 1st free task id + If we reach the max: all task ids are allocated => not possible to start SWE.*/ + /* while (rvm_allocated_task_id[i] == TRUE) + { + i++; + if (i == MAX_RVF_TASKS) + return RVF_INVALID_TASK; + }*/ + + /* Lock task id and return its value. */ + /* rvm_allocated_task_id[i] = TRUE; */ + /* return ((T_RVM_TASK_ID) i); */ + return (T_RVM_TASK_ID) rvf_allocate_task_id(isRealTask); /* A-M-E-N-D-E-D! */ +} + + +/******************************************************************************* +** +** Function rvm_set_swe_info +** +** Description This function call the set_info function of each SWEs required +** to start a specified SWE. +** +** Parameters: T_RVM_PROCESSING_SWE * appli: list of required SWEs with their parameters. +** +** Returns T_RVM_RETURN: RVM_OK if successful. +** +*******************************************************************************/ +T_RVM_RETURN rvm_set_swe_info(T_RVM_PROCESSING_SWE * appli) +{ + T_RVM_PROCESSING_SWE * cur_swe = appli; + UINT8 i; + T_RVF_MB_ID _bk_id_table[RVM_MAX_NB_MEM_BK]; + + /* for each SWE in the list */ + while( cur_swe != NULL ) + { + UINT8 swe_index = cur_swe->swe_id; + + if (rvm_swe_array[swe_index].swe_state != SWE_RUNNING) + /* Call the set_info function for only those for which MB were just created */ + { + /* First build return path */ + T_RVM_INFO_SWE swe_info; + T_RV_RETURN_PATH return_path[RVM_MAX_NB_LINKED_SWE]; + T_RVM_USE_ID linked_swe_use_id[RVM_MAX_NB_LINKED_SWE]; + UINT8 nb_linked_swe = 0; + + rvm_swe_array[swe_index].swe_get_info(&swe_info); + + switch( swe_info.swe_type) + { + case(RVM_SWE_TYPE_1): + { nb_linked_swe = swe_info.type_info.type1.nb_linked_swe; + memcpy( linked_swe_use_id, swe_info.type_info.type1.linked_swe_id, RVM_MAX_NB_LINKED_SWE * sizeof(T_RVM_USE_ID) ); + if(rvm_swe_array[swe_index].swe_state != SWE_NOT_STARTED) { + for(i=0;i<swe_info.type_info.type1.nb_mem_bank; i++) { + rvf_get_mb_id((char*)&swe_info.type_info.type1.mem_bank[i], + &_bk_id_table[i]); + } + } + break; + } + case(RVM_SWE_TYPE_2): + { nb_linked_swe = swe_info.type_info.type2.nb_linked_swe; + memcpy( linked_swe_use_id, swe_info.type_info.type2.linked_swe_id, RVM_MAX_NB_LINKED_SWE * sizeof(T_RVM_USE_ID) ); + if((rvm_swe_array[cur_swe->swe_id].swe_addr_id = rvm_allocate_task_id(0))==RVF_INVALID_ADDR_ID) { + return RVM_INTERNAL_ERR; + } + if(rvm_swe_array[swe_index].swe_state != SWE_NOT_STARTED) { + for(i=0;i<swe_info.type_info.type2.nb_mem_bank; i++) { + rvf_get_mb_id((char*)&swe_info.type_info.type2.mem_bank[i], + &_bk_id_table[i]); + } + } + break; + } + case(RVM_SWE_TYPE_3): + { nb_linked_swe = swe_info.type_info.type3.nb_linked_swe; + memcpy( linked_swe_use_id, swe_info.type_info.type3.linked_swe_id, RVM_MAX_NB_LINKED_SWE * sizeof(T_RVM_USE_ID) ); + if((rvm_swe_array[cur_swe->swe_id].swe_addr_id = rvm_allocate_task_id(1))==RVF_INVALID_ADDR_ID) { + return RVM_INTERNAL_ERR; + } + if(rvm_swe_array[swe_index].swe_state != SWE_NOT_STARTED) { + for(i=0;i<swe_info.type_info.type3.nb_mem_bank; i++) { + rvf_get_mb_id((char*)&swe_info.type_info.type3.mem_bank[i], + &_bk_id_table[i]); + } + } + break; + } + case(RVM_SWE_TYPE_4): + { nb_linked_swe = swe_info.type_info.type4.nb_linked_swe; + memcpy( linked_swe_use_id, swe_info.type_info.type4.linked_swe_id, RVM_MAX_NB_LINKED_SWE * sizeof(T_RVM_USE_ID) ); + if((rvm_swe_array[cur_swe->swe_id].swe_addr_id = rvm_allocate_task_id(1))==RVF_INVALID_ADDR_ID) { + return RVM_INTERNAL_ERR; + } + if(rvm_swe_array[swe_index].swe_state != SWE_NOT_STARTED) { + for(i=0;i<swe_info.type_info.type4.nb_mem_bank; i++) { + rvf_get_mb_id((char*)&swe_info.type_info.type4.mem_bank[i], + &_bk_id_table[i]); + } + } + break; + } + } + rvm_swe_array[cur_swe->swe_id].swe_return_path.addr_id=rvm_swe_array[cur_swe->swe_id].swe_addr_id; + + for (i=0; i < nb_linked_swe; i++) + { + UINT8 linked_swe_index; + if (rvm_get_swe_index(&linked_swe_index, linked_swe_use_id[i]) != RVM_OK) + { + return RVM_INTERNAL_ERR; + } + return_path[i].callback_func = rvm_swe_array[linked_swe_index].swe_return_path.callback_func; + /* TO DO: manage addr_id for GROUP_MEMBER SWEs */ + return_path[i].addr_id = rvm_swe_array[linked_swe_index].swe_addr_id; + } + + + if (cur_swe->rvm_functions.set_info != NULL ) { + if(rvm_swe_array[swe_index].swe_state == SWE_NOT_STARTED) { + cur_swe->rvm_functions.set_info(rvm_swe_array[cur_swe->swe_id].swe_addr_id, \ + return_path, \ + cur_swe->bk_id_table, \ + rvm_error); + } else { + cur_swe->rvm_functions.set_info(rvm_swe_array[cur_swe->swe_id].swe_addr_id, \ + return_path, \ + _bk_id_table, \ + rvm_error); + } + } + } + + cur_swe = cur_swe->next_swe; /* process next SWE */ + } + return RVM_OK; +} + + +/******************************************************************************* +** +** Function rvm_initialize_swe +** +** Description This function initialize all the required SWEs which are not running. +** It also creates the tasks in a suspend state. +** Then it resumes the tasks and call the start function of each SWE. +** +** Parameters: T_RVM_PROCESSING_SWE * appli: list of required SWEs with their parameters. +** +** Returns T_RVM_RETURN: RVM_OK if successful. +** +*******************************************************************************/ +T_RVM_RETURN rvm_initialize_swe( T_RVM_PROCESSING_SWE * appli, + T_RVM_GROUP_DIRECTIVE* gd, + UINT8 t2cnt) { + T_RVM_PROCESSING_SWE * cur_swe = appli; + UINT8 i=0, j=0; + UINT16 tuid=0; + T_RVF_BUFFER* stack_ptr=NULL; + T_RVM_INFO_SWE swe_info; + +#ifdef _WINDOWS + BOOLEAN hci_started = FALSE; +#endif + + /* for each SWE in the list, initialize it */ + while( cur_swe != NULL ) + { + UINT8 swe_index = cur_swe->swe_id; + + if ( rvm_swe_array[swe_index].swe_state != SWE_RUNNING) + { + /* call its init function */ + if (cur_swe->rvm_functions.init) + { + if (cur_swe->rvm_functions.init() != RVM_OK) + { + rvf_send_trace("RVM: Error Calling init function of swe nb ", 43, \ + (UINT32)swe_index, RV_TRACE_LEVEL_ERROR, RVM_USE_ID ); + } + } + } + cur_swe = cur_swe->next_swe; + } + + + /* for each SWE in the list, create the task if necessary. */ + cur_swe = appli; + while( cur_swe != NULL ) + { + UINT8 swe_index = cur_swe->swe_id; + + if ( rvm_swe_array[swe_index].swe_state != SWE_RUNNING) { + /* start the task if necessary in SUSPEND mode */ + + if ( cur_swe->swe_type == RVM_SWE_TYPE_4) { + /* allocate a buffer for the stack */ + if ( rvm_allocate_stack_buffer( cur_swe->stack_size, + &rvm_swe_array[swe_index].stack_ptr) != RVM_OK) { + + rvf_send_trace("RVM: Error allocating stack nb:", 28, (UINT32)rvm_swe_array[swe_index].swe_addr_id, RV_TRACE_LEVEL_ERROR, RVM_USE_ID); + + /* TO DO: manage the error case */ + return RVF_MEMORY_ERR; + } + + /* start the task in suspend mode */ + if (rvf_create_task((TASKPTR) cur_swe->rvm_functions.core, \ + (UINT8)rvm_swe_array[swe_index].swe_addr_id,\ + rvm_swe_array[swe_index].swe_name, \ + rvm_swe_array[swe_index].stack_ptr, \ + cur_swe->stack_size, \ + cur_swe->priority, \ + ET4_TASK,\ + DEFAULT_TIME_SLICING, \ + SUSPEND ) != RV_OK) { + + rvf_send_trace("RVM: Error Creating Task nb:", 28, (UINT32)rvm_swe_array[swe_index].swe_addr_id, RV_TRACE_LEVEL_ERROR, RVM_USE_ID); + } + + rvf_setRtAddrSweIndex(rvm_swe_array[swe_index].swe_addr_id, + swe_index); + + rvf_send_trace("RVM: Created task nb ", 21, (UINT32)rvm_swe_array[swe_index].swe_addr_id, RV_TRACE_LEVEL_DEBUG_LOW, RVM_USE_ID); + + } else if (cur_swe->swe_type == RVM_SWE_TYPE_3) { + /* allocate a buffer for the stack */ + if ( rvm_allocate_stack_buffer( cur_swe->stack_size, + &rvm_swe_array[swe_index].stack_ptr) != RVM_OK) { + + rvf_send_trace("RVM: Error allocating stack nb:", 28, (UINT32)rvm_swe_array[swe_index].swe_addr_id, RV_TRACE_LEVEL_ERROR, RVM_USE_ID); + /* TO DO: manage the error case */ + return RVF_MEMORY_ERR; + } + + /* start the task in suspend mode */ + if (rvf_create_task((TASKPTR)rvm_t3_proxy, \ + (UINT8)rvm_swe_array[swe_index].swe_addr_id,\ + rvm_swe_array[swe_index].swe_name, \ + rvm_swe_array[swe_index].stack_ptr, \ + cur_swe->stack_size, \ + cur_swe->priority, \ + ET3_TASK,\ + DEFAULT_TIME_SLICING, \ + SUSPEND ) != RV_OK) { + + rvf_send_trace("RVM: Error Creating E3 Task nb:", 28, (UINT32)rvm_swe_array[swe_index].swe_addr_id, RV_TRACE_LEVEL_ERROR, RVM_USE_ID); + } + + rvf_register_t3_handlers(rvm_swe_array[swe_index].swe_addr_id, + cur_swe->rvm_functions.handle_message, /* traverse list hence: cur_swe->rvm_functions */ + cur_swe->rvm_functions.handle_timer ); + + rvf_setRtAddrSweIndex(rvm_swe_array[swe_index].swe_addr_id, + swe_index); + + rvf_send_trace("RVM: Created task nb ", 21, (UINT32)rvm_swe_array[swe_index].swe_addr_id, RV_TRACE_LEVEL_DEBUG_LOW, RVM_USE_ID); + + + } + } + cur_swe = cur_swe->next_swe; /* process next SWE */ + } + /* resolve T2 grouping */ + for(i=0; i<t2cnt; i++) { + gd[i].host_task_addr=rvf_resolveHostingAddrId(gd[i]); + if( gd[i].host_task_addr==RVF_INVALID_ADDR_ID) { + + if ( rvm_allocate_stack_buffer( gd[i].stack_size, &stack_ptr) != RVM_OK){ + /* TO DO: manage the error case - ABORT & Clean-up if one or more linked Ent. fail */ + //break; + return RVF_MEMORY_ERR; + } + + gd[i].host_task_addr=rvm_allocate_task_id(1); + rvf_create_task((TASKPTR)rvm_t2_proxy, + gd[i].host_task_addr, // + "hosting_task", + stack_ptr, + gd[i].stack_size, + gd[i].task_priority, + ET2_HOST_TASK, + DEFAULT_TIME_SLICING, + SUSPEND); + + rvf_associateGrpToHost(gd[i].host_task_addr, gd[i].group_directive); + } + + for(j=0; j<MAX_COMPOSITES && gd[i].hosted_swe_db_index[j]!=0; j++) { + /* create virtual task for each "hosted_swe_db_index[]" */ + rvm_swe_array[gd[i].hosted_swe_db_index[j]].swe_get_info(&swe_info); + + rvf_create_virtual_task(swe_info.type_info.type2.handle_message, + swe_info.type_info.type2.handle_timer, + rvm_swe_array[gd[i].hosted_swe_db_index[j]].swe_addr_id, + gd[i].host_task_addr, + rvm_swe_array[gd[i].hosted_swe_db_index[j]].swe_name, + rvm_swe_array[gd[i].hosted_swe_db_index[j]].swe_priority, + ET2_VTASK); + rvf_setRtAddrSweIndex(rvm_swe_array[gd[i].hosted_swe_db_index[j]].swe_addr_id, + gd[i].hosted_swe_db_index[j]); + + /* register each with associate host */ + rvf_registerToHost( gd[i].host_task_addr, + rvm_swe_array[gd[i].hosted_swe_db_index[j]].swe_addr_id); + } + + } + + /* resume all hosting tasks... */ + for(i=0; i<t2cnt; i++) rvf_resume_task((UINT8)gd[i].host_task_addr); + + /* start composites or virtual tasks */ + for(i=0; i<t2cnt; i++) { + rvm_start_group_req((UINT8)gd[i].host_task_addr, + gd[i].hosted_swe_db_index); + } + + /* for each SWE in the list, start it if necessary. */ + for(cur_swe = appli; cur_swe != NULL; ) { + UINT8 swe_index = cur_swe->swe_id; + + if ( rvm_swe_array[swe_index].swe_state != SWE_RUNNING) { + /* if the SWE is a task, resume it */ + if ( (cur_swe->swe_type == RVM_SWE_TYPE_3) + || (cur_swe->swe_type == RVM_SWE_TYPE_4) ) { + + /* TO DO: check the return value */ + if(rvf_resume_task((UINT8)rvm_swe_array[swe_index].swe_addr_id )!=RVF_OK) { + RVM_TRACE_WARNING("RVM: ERROR! UNABLE TO RESUME SWE"); + return RVF_INTERNAL_ERR; + } + rvf_send_trace("RVM: Resumed task nb ", 21, (UINT32)rvm_swe_array[swe_index].swe_addr_id, RV_TRACE_LEVEL_DEBUG_LOW, RVM_USE_ID); + rvf_send_trace("RVM: Resumed SWE ", 17, (UINT32)rvm_swe_array[swe_index].swe_use_id, RV_TRACE_LEVEL_DEBUG_LOW, RVM_USE_ID); + +#ifdef _WINDOWS + if (rvm_swe_array[swe_index].swe_use_id == HCI_USE_ID ) { + hci_started = TRUE; + } +#endif + + } else if(cur_swe->swe_type==RVM_SWE_TYPE_1) { /* A-M-E-N-D-E-D! */ + + /* call its init function */ + if (cur_swe->rvm_functions.start) { + if (cur_swe->rvm_functions.start() != RVM_OK) { + rvf_send_trace("RVM: Error Calling start function of swe nb ", 44, \ + (UINT32)swe_index, RV_TRACE_LEVEL_ERROR, RVM_USE_ID); + } + } + } + } + + /* increment the number of using swe and points to the using appli */ + /* DOES NOT DEPEND ON THE STATE */ + /*rvm_swe_array[swe_index].swe_get_info(&swe_info); + switch( swe_info.swe_type) { + case RVM_SWE_TYPE_1: + if(!swe_info.type_info.type1.nb_linked_swe) rvm_swe_array[swe_index].nb_using_appli=0; + break; + case RVM_SWE_TYPE_2: + if(!swe_info.type_info.type2.nb_linked_swe) rvm_swe_array[swe_index].nb_using_appli=0; + break; + case RVM_SWE_TYPE_3: + if(!swe_info.type_info.type3.nb_linked_swe) rvm_swe_array[swe_index].nb_using_appli=0; + break; + case RVM_SWE_TYPE_4: + if(!swe_info.type_info.type4.nb_linked_swe) rvm_swe_array[swe_index].nb_using_appli=0; + break; + default: rvm_swe_array[swe_index].nb_using_appli=0; + }*/ + +// if(rvm_swe_array[swe_index].nb_using_appli) { +// rvm_swe_array[swe_index].using_appli[rvm_swe_array[swe_index].nb_using_appli++] = appli->swe_id; +// +// } + + if(rvm_swe_array[appli->swe_id].nb_using_appli<RVM_MAX_SWE_USING ) { + rvm_swe_array[appli->swe_id].using_appli[rvm_swe_array[appli->swe_id].nb_using_appli++]=swe_index; + } else { + RVM_TRACE_WARNING_PARAM("RVM: Unable to track 'Using Appli' list is full nb=", appli->swe_id); + } + + cur_swe = cur_swe->next_swe; /* process next SWE */ + } + + for(cur_swe=appli; cur_swe!=NULL; ) { + rvm_swe_array[cur_swe->swe_id].swe_state = SWE_RUNNING; + cur_swe = cur_swe->next_swe; + } + +#ifdef _WINDOWS + if (hci_started == TRUE) { + rvf_delay(RVF_MS_TO_TICKS(1000)); + } +#endif + + return RVM_OK; +} + + +/******************************************************************************* +** +** Function rvm_stop_swe_list +** +** Description This function will call the stop functions when possible. +** +** Parameters: T_RVM_PROCESSING_SWE * appli: list of required SWEs with their parameters. +** +** Returns T_RVM_OK if all allocation are successful, +** else T_RVM_INTERNAL_ERR (then some SWE are not stopped. +** +*******************************************************************************/ +T_RVM_RETURN rvm_stop_swe_list( T_RVM_PROCESSING_SWE * appli, T_RV_HDR* hdr) +{ + T_RVM_PROCESSING_SWE * cur_swe = appli; + T_RVM_INFO_SWE swe_info; + volatile T_RVM_RETURN rvm_ret_value = RVM_OK; + T_RVM_STOP_MSG* p_msg=NULL; + UINT8 i=0; + + + /* for each SWE in the list */ + while (cur_swe != NULL ) + { + UINT8 swe_index = cur_swe->swe_id; + + /* If nb_using_appli > 1, SWE cannot be stopped */ +/* if (rvm_swe_array[swe_index].nb_using_appli > 1) { + cur_swe = cur_swe->next_swe; + continue; + } + // If nb_using_appli == 1 but using_appli != appli, SWE cannot be stopped + if ((rvm_swe_array[swe_index].nb_using_appli == 1) && \ + (rvm_swe_array[swe_index].using_appli[0] != appli->swe_id)) { + cur_swe = cur_swe->next_swe; + continue; + } +*/ + if (cur_swe->swe_type==RVM_SWE_TYPE_1) { //cater for de-init of lib + if(cur_swe->rvm_functions.stop1)cur_swe->rvm_functions.stop1(); + if(cur_swe->rvm_functions.kill)cur_swe->rvm_functions.kill(); + cur_swe = cur_swe->next_swe; + continue; + } + if (cur_swe->swe_type==RVM_SWE_TYPE_4) { // etype 4 restriction + RVM_TRACE_WARNING_PARAM("RVM: Stop & Kill is not applicable to Type 4 entities, nb=", (UINT32)swe_index); + for (rvm_swe_array[swe_index].nb_using_appli=0,i=0; i<RVM_MAX_SWE_USING; i++) { //reset using appli - workaround! + rvm_swe_array[swe_index].using_appli[i] = RVM_INVALID_SWE_INDEX; + } + cur_swe = cur_swe->next_swe; + continue; + } + /* Retrieve stop function with a get_info */ + if (rvm_swe_array[swe_index].swe_get_info == NULL) + { + RVM_TRACE_WARNING_PARAM("RVM: SWE with no get info, cannot be stopped, nb=", (UINT32)swe_index); + cur_swe = cur_swe->next_swe; + rvm_ret_value = RVM_INTERNAL_ERR; + continue; + } + rvm_swe_array[swe_index].swe_get_info( &swe_info); + + if (cur_swe->rvm_functions.stop == NULL) { + RVM_TRACE_WARNING_PARAM("RVM: SWE with no stop function, cannot be stopped, nb=", (UINT32)swe_index); + cur_swe = cur_swe->next_swe; + continue; + } + + if (rvf_get_buf( rvm_mem_bank, sizeof(T_RVM_STOP_MSG), (void **)&p_msg) == RVF_RED ) { + RVM_TRACE_WARNING_PARAM("RVM: Unable to create STOP msg, nb=", (UINT32)swe_index); + cur_swe = cur_swe->next_swe; + continue; + } + + p_msg->header.msg_id = RVM_STOP_MSG; + p_msg->header.src_addr_id = hdr->src_addr_id; + p_msg->header.dest_addr_id = hdr->dest_addr_id; +// p_msg->header.callback_func = hdr->callback_func; + p_msg->rp.callback_func = ((T_RVM_STOP_MSG*)hdr)->rp.callback_func; + p_msg->status = SWE_STOPPING; + p_msg->swe_num = swe_index; //((T_RVM_STOP_MSG*)hdr)->swe_num; + + if ( rvf_send_msg( rvm_swe_array[swe_index].swe_addr_id, p_msg) != RVF_OK) { + rvm_ret_value = RVM_INTERNAL_ERR; + cur_swe = cur_swe->next_swe; + continue; + } + + rvm_swe_array[swe_index].swe_state=SWE_STOPPING; + +/*printf("SHUTDOWN: SWE %s nb %d USING APPLI= %d\n",rvm_swe_array[swe_index].swe_name, swe_index, rvm_swe_array[swe_index].nb_using_appli); +for(i=0; i<rvm_swe_array[swe_index].nb_using_appli; i++)printf(" %d, ", rvm_swe_array[swe_index].using_appli[i]); +printf("\n");*/ + + for (rvm_swe_array[swe_index].nb_using_appli=0,i=0; i<RVM_MAX_SWE_USING; i++) { //reset using appli - workaround! + rvm_swe_array[swe_index].using_appli[i] = RVM_INVALID_SWE_INDEX; + } + +/*printf("SHUTDOWN: SWE %s nb %d USING APPLI= %d\n",rvm_swe_array[swe_index].swe_name, swe_index, rvm_swe_array[swe_index].nb_using_appli); +for(i=0; i<rvm_swe_array[swe_index].nb_using_appli; i++)printf(" %d, ", rvm_swe_array[swe_index].using_appli[i]); +printf("\n");*/ + + /* Stop SWE - amended to ASYNC */ + /* TO DO: for type 2 and 3 SWEs, send a message to the host to call the stop function */ + //cur_swe->rvm_functions.stop(NULL); + + /* Proceed to the next SWE */ + cur_swe = cur_swe->next_swe; + } + + return rvm_ret_value; +} + + +/******************************************************************************* +** +** Function rvm_suspend_swe_tasks +** +** Description This function will suspend all SWE that are tasks. +** +** Parameters: T_RVM_PROCESSING_SWE * appli: list of required SWEs with their parameters. +** +** Returns T_RVM_OK if all allocation are successful, +** else T_RVM_INTERNAL_ERR (then some SWE are not stopped. +** +*******************************************************************************/ +T_RVM_RETURN rvm_suspend_swe_tasks( T_RVM_PROCESSING_SWE * appli) +{ + T_RVM_PROCESSING_SWE * cur_swe = appli; + T_RVM_INFO_SWE swe_info; + volatile T_RVM_RETURN rvm_ret_value = RVM_OK; + + /* for each SWE in the list */ + while (cur_swe != NULL ) + { + UINT8 swe_index = cur_swe->swe_id; + + /* If nb_using_appli > 1, SWE cannot be stopped */ + if (rvm_swe_array[swe_index].nb_using_appli > 1) + { + cur_swe = cur_swe->next_swe; + continue; + } + /* If nb_using_appli == 1 but using_appli != appli, SWE cannot be stopped */ + if ((rvm_swe_array[swe_index].nb_using_appli == 1) && \ + (rvm_swe_array[swe_index].using_appli[0] != appli->swe_id)) + { + cur_swe = cur_swe->next_swe; + continue; + } + + /* Retrieve task info with a get_info */ + if (rvm_swe_array[swe_index].swe_get_info == NULL) + { + RVM_TRACE_WARNING_PARAM("RVM: SWE with no get info, cannot be stopped, nb=", (UINT32)swe_index); + cur_swe = cur_swe->next_swe; + rvm_ret_value = RVM_INTERNAL_ERR; + continue; + } + rvm_swe_array[swe_index].swe_get_info( &swe_info); + + /* If SWE is not a task, continue */ + /* TO DO: manage group member SWEs */ + if ( (swe_info.swe_type == RVM_SWE_TYPE_1) || + (swe_info.swe_type == RVM_SWE_TYPE_2) ) + { + cur_swe = cur_swe->next_swe; + continue; + } + + /* Suspend SWE task */ + rvf_suspend_task( (UINT8)rvm_swe_array[swe_index].swe_return_path.addr_id); + RVM_TRACE_DEBUG_LOW_PARAM("RVM: Suspended task nb ", (UINT32) (rvm_swe_array[swe_index].swe_return_path.addr_id & 0x000000FF) ); + + /* Proceed to the next SWE */ + cur_swe = cur_swe->next_swe; + } + + return rvm_ret_value; +} + + +/******************************************************************************* +** +** Function rvm_kill_swe_list +** +** Description This function will call the kill functions when possible. +** It will also delete the task, the stack and the used MBs. +** +** Parameters: T_RVM_PROCESSING_SWE * appli: list of required SWEs with their parameters. +** +** Returns T_RVM_OK if everything is successful, +** else T_RVM_INTERNAL_ERR (then some SWE are not killed). +** +*******************************************************************************/ +T_RVM_RETURN rvm_kill_swe_list( T_RVM_PROCESSING_SWE * appli) +{ + T_RVM_PROCESSING_SWE * cur_swe = appli; + T_RVM_INFO_SWE swe_info; + volatile T_RVM_RETURN rvm_ret_value = RVM_OK; + + /* for each SWE in the list */ + while (cur_swe != NULL ) + { + UINT8 swe_index = cur_swe->swe_id; + + /* If nb_using_appli > 1, SWE cannot be killed */ + if (rvm_swe_array[swe_index].nb_using_appli > 1) + { + cur_swe = cur_swe->next_swe; + continue; + } + + /* If nb_using_appli == 1 but using_appli != appli, SWE cannot be killed */ + if ((rvm_swe_array[swe_index].nb_using_appli == 1) && \ + (rvm_swe_array[swe_index].using_appli[0] != appli->swe_id)) + { + cur_swe = cur_swe->next_swe; + continue; + } + + /* Retrieve kill function with a get_info */ + if (rvm_swe_array[swe_index].swe_get_info == NULL) + { + RVM_TRACE_WARNING_PARAM("RVM: SWE with no get info, cannot be killed, nb=", (UINT32)swe_index); + cur_swe = cur_swe->next_swe; + rvm_ret_value = RVM_INTERNAL_ERR; + continue; + } + rvm_swe_array[swe_index].swe_get_info( &swe_info); + + if (cur_swe->rvm_functions.kill == NULL) + { + RVM_TRACE_WARNING_PARAM("RVM: SWE with no kill function, cannot be killed, nb=", (UINT32)swe_index); + cur_swe = cur_swe->next_swe; + rvm_ret_value = RVM_INTERNAL_ERR; + continue; + } + + /* Kill SWE */ + cur_swe->rvm_functions.kill(); + + /* TO DO: manage group member SWEs */ + /* If the SWE is a task, the task should be deleted, as well as its stack */ + if ( (swe_info.swe_type == RVM_SWE_TYPE_3) || + (swe_info.swe_type == RVM_SWE_TYPE_4) ) + { + rvf_exit_task((UINT8)(rvm_swe_array[swe_index].swe_return_path.addr_id)); + rvf_free_buf(rvm_swe_array[swe_index].stack_ptr); + RVM_TRACE_DEBUG_LOW_PARAM("RVM: Deleted task nb ", (UINT32)(rvm_swe_array[swe_index].swe_return_path.addr_id & 0x000000FF)); + rvf_free_sys_resources(rvm_swe_array[swe_index].swe_addr_id, 2); + + } else if(swe_info.swe_type == RVM_SWE_TYPE_2) { + rvf_free_sys_resources(rvm_swe_array[swe_index].swe_addr_id, 0); + } + + /* Proceed to the next SWE */ + cur_swe = cur_swe->next_swe; + } + + return rvm_ret_value; +} + + +/******************************************************************************* +** +** Function rvm_launch_appli +** +** Description Called by the main RVM task to start a specified known application +** +** Parameters: T_RVM_MSG msg: containing the return path and the index of the +** application to start in the array of known SWEs. +** +** Returns None +** +*******************************************************************************/ +void rvm_launch_appli( T_RVM_MSG * msg_Ptr) { + T_RVM_GROUP_DIRECTIVE GroupDirectives[MAX_GRPS]; + UINT8 gdCount=0; + T_RVM_PROCESSING_SWE * appli = NULL; /* pointer to the first element of the list */ + T_RV_RETURN_PATH appli_return_path; + UINT8 i,j=0; + + for(i=0; i<MAX_GRPS; i++) { + GroupDirectives[i].group_directive=0; + GroupDirectives[i].host_task_addr=0; + GroupDirectives[i].stack_size=0; + memset(&GroupDirectives[i].hosted_swe_db_index, 0, (sizeof(UINT8)*MAX_COMPOSITES)); + } + + /* store the return path of the caller */ + appli_return_path.callback_func = msg_Ptr->rp.callback_func; + appli_return_path.addr_id = msg_Ptr->header.src_addr_id; + + /* recursively call all get_info functions and build the list of running swe */ + if ( rvm_build_swe_list( &appli, msg_Ptr->swe_num, 0) != RVM_OK ) + { + /* Display error message + error case: use the return_path to inform the caller that an error occurs*/ + rvm_snd_msg_to_upper(RVM_START_APPLI, RVM_INVALID_PARAMETER, msg_Ptr->swe_num, appli_return_path); + RVM_TRACE_ERROR("RVM: SWE list built error"); + return; + } + + gdCount=_resolve_t2_grouping(appli, GroupDirectives); + + if(!appli) { + // error case: use return_path to inform the caller about memory lack + // Unlock state of SWE and free memory + RVM_TRACE_WARNING_PARAM("RVM: ABORTED, Stand-alone ENTITY start request!", (UINT32)msg_Ptr->swe_num); + rvm_snd_msg_to_upper(RVM_START_APPLI, RVM_NOT_READY, msg_Ptr->swe_num, appli_return_path); + rvm_delete_used_memory (appli); + return; + } + + RVM_TRACE_DEBUG_HIGH("RVM: SWE list built success"); + RVM_TRACE_DEBUG_HIGH_PARAM("RVM: trying to launch SWE", rvm_swe_array[appli->swe_id].swe_use_id); + + /* check if there is enough available memory */ + if ( rvm_verify_memory_requirement( appli, GroupDirectives, gdCount) != RVM_OK) + { + /* error case: use return_path to inform the caller about memory lack */ + /* Unlock state of SWE and free memory */ + RVM_TRACE_WARNING_PARAM("RVM: SWE not enough memory: unable to launch Appli nb", (UINT32)appli->swe_id); + rvm_snd_msg_to_upper(RVM_START_APPLI, RVM_MEMORY_ERR, msg_Ptr->swe_num, appli_return_path); + rvm_delete_used_memory (appli); + return; + } + + /* allocates memory banks */ + if ( rvm_allocate_mb( appli) != RVM_OK ) + { /* error case: use return_path to inform the caller about memory lack */ + rvm_delete_used_memory (appli); + rvm_snd_msg_to_upper(RVM_START_APPLI, RVM_MEMORY_ERR, msg_Ptr->swe_num, appli_return_path); + RVM_TRACE_WARNING("RVM: SWE memory bank allocation error - launch aborted!"); + return; + } + RVM_TRACE_DEBUG_LOW("RVM: SWE memory bank allocation success"); + + /* call set_info function for each SWE */ + if ( rvm_set_swe_info( appli) != RVM_OK) + { /* error case: use return_path to inform the caller that an error occurs */ + RVM_TRACE_WARNING("RVM: SWE set info functions error"); + _fatal(appli, 0); + rvm_delete_created_mb(appli); + rvm_delete_used_memory (appli); + rvm_snd_msg_to_upper(RVM_START_APPLI, RVM_INTERNAL_ERR, msg_Ptr->swe_num, appli_return_path); + return; + } + RVM_TRACE_DEBUG_LOW("RVM: SWE set info functions called"); + + + /* call the init and start functions */ + if ( rvm_initialize_swe( appli, GroupDirectives, gdCount) != RVM_OK) + { /* error case: use return_path to inform the caller that an error occurs */ + RVM_TRACE_WARNING("RVM: SWE initialization error"); + rvm_snd_msg_to_upper(RVM_START_APPLI, RVM_INTERNAL_ERR, msg_Ptr->swe_num, appli_return_path); + _fatal(appli, 2); + rvm_delete_created_mb(appli); + rvm_delete_used_memory (appli); + return; + } + RVM_TRACE_DEBUG_LOW("RVM: SWE initialization success"); + + /* build a message and send the response to the caller */ + /* send a result using the return_path */ + rvm_snd_msg_to_upper(RVM_START_APPLI, RVM_OK, msg_Ptr->swe_num, appli_return_path); + + + /* and store the return_path */ + rvm_swe_array[ msg_Ptr->swe_num ].mmi_return_path.callback_func = msg_Ptr->rp.callback_func; + rvm_swe_array[ msg_Ptr->swe_num ].mmi_return_path.addr_id = msg_Ptr->header.src_addr_id; + + /* Once Everything is back in stand-by, release used memory */ + rvm_delete_used_memory (appli); +} + + +/******************************************************************************* +** +** Function rvm_shut_down_appli +** +** Description Called by the main RVM task to stop a specified known application +** +** Parameters: T_RVM_MSG msg: containing the return path and the index of the +** application to stop in the array of known SWEs. +** +** Returns None +** +*******************************************************************************/ +void rvm_stop_appli( T_RVM_STOP_MSG* msg_Ptr) { + T_RVM_PROCESSING_SWE * appli = NULL; /* pointer to the first element of the list */ + T_RVM_RETURN ret_value; + UINT8 swe_idx = 200; + T_RV_RETURN_PATH appli_return_path; + + appli_return_path.callback_func = msg_Ptr->rp.callback_func; + appli_return_path.addr_id = msg_Ptr->header.src_addr_id; + + + RVM_TRACE_DEBUG_HIGH_PARAM("RVM: trying to stop Appli nb ", (UINT32)swe_idx); + + if (rvm_swe_array[msg_Ptr->swe_num].nb_using_appli > 1) { + RVM_TRACE_WARNING_PARAM("RVM: SWE has dependencies, nb=", (UINT32)msg_Ptr->swe_num); + return; + } + // ??? If nb_using_appli == 1 but using_appli != appli, SWE cannot be stopped + if ((rvm_swe_array[msg_Ptr->swe_num].nb_using_appli == 1) && \ + (rvm_swe_array[msg_Ptr->swe_num].using_appli[0] != msg_Ptr->swe_num)) { + RVM_TRACE_WARNING_PARAM("RVM: SWE has dependencies, nb=", (UINT32)msg_Ptr->swe_num); + return; + } + + + /* TO DO : REBUILD SWE LIST !!!! */ + if ( rvm_build_swe_list( &appli, msg_Ptr->swe_num, 1) != RVM_OK ) + { + /* Display error message + error case: use the return_path to inform the caller that an error occurs*/ + rvm_snd_msg_to_upper(RVM_START_APPLI, RVM_INVALID_PARAMETER, msg_Ptr->swe_num, appli_return_path); + RVM_TRACE_ERROR("RVM: SWE list built error"); + return; + } + + /* Stop all swe in the list that are used only once */ + if ((ret_value = rvm_stop_swe_list(appli, (T_RV_HDR*)msg_Ptr)) != RVM_OK ) + { + /* Display error message + TO DO: error case: use the return_path to inform the caller that an error occurs */ + RVM_TRACE_WARNING_PARAM("RVM: Error in SWE stop", (UINT32)ret_value); + return; + } + + rvm_delete_used_memory (appli); + + RVM_TRACE_DEBUG_LOW("RVM: SWE stop broadcast!"); + +} + +// NOTE: presently no timeout exists, if the ENT. fails to reply +// to stop with rvm_swe_stopped() RVM doesn't kill it. +void rvm_swe_has_stopped(T_RVM_STOP_MSG* msg) { + T_RVM_STOP_MSG* p_msg=(T_RVM_STOP_MSG*)msg; + T_RV_RETURN_PATH appli_return_path; + + appli_return_path.callback_func = msg->rp.callback_func; + appli_return_path.addr_id = msg->header.src_addr_id; + + if(msg->status!=SWE_STOPPING) { + // inform upper of problem + rvm_snd_msg_to_upper(RVM_STOP_APPLI, RVM_INVALID_PARAMETER, msg->swe_num, appli_return_path); + RVM_TRACE_ERROR("RVM: Entity declines STOP REQ"); + rvf_free_msg((T_RV_HDR*)msg); + return; + } + + // cont. with shutdown - MUST DO ERROR CASE ! + rvm_shutdown_swe(p_msg->swe_num); + + // set stopped status + rvm_swe_array[p_msg->swe_num].swe_state=SWE_KILLED; //SWE_STOPPING; + + /* build a message and send the response to the caller */ + /* send a result using the return_path */ + if(rvm_get_mb_level(p_msg->swe_num) ){ + rvm_snd_msg_to_upper(RVM_STOP_APPLI, RV_MEMORY_REMAINING, msg->swe_num, appli_return_path); + } else { + rvm_snd_msg_to_upper(RVM_STOP_APPLI, RVM_OK, msg->swe_num, appli_return_path); + } + + /* and store the return_path */ + rvm_swe_array[ msg->swe_num ].mmi_return_path.callback_func = msg->rp.callback_func; + rvm_swe_array[ msg->swe_num ].mmi_return_path.addr_id = msg->header.src_addr_id; + +} + +void rvm_shutdown_swe(UINT8 index) { //should ret. ok or fail + rvm_suspend_swe(index); + rvm_kill_swe(index); +} + +void rvm_suspend_swe(UINT8 swe_index) { + volatile T_RVM_RETURN rvm_ret_value = RVM_OK; + T_RVM_INFO_SWE swe_info; + + /* ??? If nb_using_appli > 1, SWE cannot be stopped + if (rvm_swe_array[swe_index].nb_using_appli > 1) { + RVM_TRACE_WARNING_PARAM("RVM-SUSPEND: SWE has dependencies, nb=", (UINT32)swe_index); + } + // ??? If nb_using_appli == 1 but using_appli != appli, SWE cannot be stopped + if ((rvm_swe_array[swe_index].nb_using_appli == 1) && \ + (rvm_swe_array[swe_index].using_appli[0] != swe_index)) { + RVM_TRACE_WARNING_PARAM("RVM-SUSPEND: SWE has dependencies, nb=", (UINT32)swe_index); + }*/ + + /* Retrieve task info with a get_info */ + if (rvm_swe_array[swe_index].swe_get_info == NULL) { + RVM_TRACE_WARNING_PARAM("RVM: SWE with no get info, cannot be stopped, nb=", (UINT32)swe_index); + rvm_ret_value = RVM_INTERNAL_ERR; + return; + } + rvm_swe_array[swe_index].swe_get_info( &swe_info); + + /* If SWE is not a task, continue */ + /* TO DO: manage group member SWEs */ + if ( (swe_info.swe_type == RVM_SWE_TYPE_1) || + (swe_info.swe_type == RVM_SWE_TYPE_2) ) { + return; + } + + /* Suspend SWE task */ + rvf_suspend_task( (UINT8)rvm_swe_array[swe_index].swe_return_path.addr_id); + RVM_TRACE_DEBUG_LOW_PARAM("RVM: Suspended task nb ", (UINT32) (rvm_swe_array[swe_index].swe_return_path.addr_id & 0x000000FF) ); +} + +T_RVM_RETURN rvm_kill_swe(UINT8 swe_index) { + T_RVM_INFO_SWE swe_info; + volatile T_RVM_RETURN rvm_ret_value = RVM_OK; + UINT8 isVirtual=0; + T_RVF_G_ADDR_ID gid=RVF_INVALID_ADDR_ID; + UINT8 isIdle=0; + UINT8 i=0; + + /* If nb_using_appli > 1, SWE cannot be killed + if (rvm_swe_array[swe_index].nb_using_appli > 1) return rvm_ret_value; + + // If nb_using_appli == 1 but using_appli != appli, SWE cannot be killed + if ((rvm_swe_array[swe_index].nb_using_appli == 1) && \ + (rvm_swe_array[swe_index].using_appli[0] != swe_index)) { + RVM_TRACE_WARNING_PARAM("RVM-KILL: SWE has dependencies, nb=", (UINT32)swe_index); + return rvm_ret_value; + }*/ + + /* Retrieve kill function with a get_info */ + if (rvm_swe_array[swe_index].swe_get_info == NULL){ + RVM_TRACE_WARNING_PARAM("RVM-KILL: SWE has no kill function defined, nb=", (UINT32)swe_index); + rvm_ret_value = RVM_INTERNAL_ERR; + } + + rvm_swe_array[swe_index].swe_get_info(&swe_info); + switch( swe_info.swe_type) { + case RVM_SWE_TYPE_1: +// if(swe_info.type_info.type1.kill) swe_info.type_info.type1.kill() ; + isVirtual=1; + break; + case RVM_SWE_TYPE_2: + gid=resolveHostAddrId(rvm_swe_array[swe_index].swe_addr_id); + rvf_unregisterFromHost(gid, rvm_swe_array[swe_index].swe_addr_id); + rvf_isHostingTaskIdle(gid, &isIdle); + if(isIdle) { // Defered suspend of hosting task: + rvf_suspend_task(gid); + rvf_exit_task(gid); + rvf_free_sys_resources(gid, 2); + } + if(swe_info.type_info.type2.kill) swe_info.type_info.type2.kill(); + isVirtual=1; + break; + case RVM_SWE_TYPE_3: + if(swe_info.type_info.type3.kill) swe_info.type_info.type3.kill(); + break; + case RVM_SWE_TYPE_4: + if(swe_info.type_info.type4.kill) swe_info.type_info.type4.kill(); + break; + default: + RVM_TRACE_WARNING_PARAM("RVM: SWE with no kill function, cannot be killed, nb=", (UINT32)swe_index); + } + + if(!isVirtual) { + rvf_exit_task((UINT8)(rvm_swe_array[swe_index].swe_return_path.addr_id)); + rvf_free_buf(rvm_swe_array[swe_index].stack_ptr); + RVM_TRACE_DEBUG_LOW_PARAM("RVM: Deleted task nb ", (UINT32)(rvm_swe_array[swe_index].swe_return_path.addr_id & 0x000000FF)); + rvf_free_sys_resources(rvm_swe_array[swe_index].swe_addr_id, 2); + } else { + rvf_free_sys_resources(rvm_swe_array[swe_index].swe_addr_id, 0); + } + + return rvm_ret_value; +} + +UINT8 rvm_get_mb_level(UINT8 swe_index) { + T_RVM_INFO_SWE swe_info; + INT8 i=0; + UINT8 isUsed=0; + + rvm_swe_array[swe_index].swe_get_info(&swe_info); + switch( swe_info.swe_type) { + case RVM_SWE_TYPE_1: + if(swe_info.type_info.type1.nb_mem_bank!=0) + for(i=0; i<swe_info.type_info.type1.nb_mem_bank; i++) { + rvf_mb_is_used(swe_info.type_info.type1.mem_bank[i].bank_name, &isUsed); + if(isUsed) return isUsed; + } + return isUsed; + case RVM_SWE_TYPE_2: + if(swe_info.type_info.type2.nb_mem_bank!=0) + for(i=0; i<swe_info.type_info.type2.nb_mem_bank; i++) { + rvf_mb_is_used(swe_info.type_info.type2.mem_bank[i].bank_name, &isUsed); + if(isUsed) return isUsed; + } + return isUsed; + case RVM_SWE_TYPE_3: + if(swe_info.type_info.type3.nb_mem_bank!=0) + for(i=0; i<swe_info.type_info.type3.nb_mem_bank; i++) { + rvf_mb_is_used(swe_info.type_info.type3.mem_bank[i].bank_name, &isUsed); + if(isUsed) return isUsed; + } + return isUsed; + case RVM_SWE_TYPE_4: + if(swe_info.type_info.type4.nb_mem_bank!=0) + for(i=0; i<swe_info.type_info.type4.nb_mem_bank; i++) { + rvf_mb_is_used(swe_info.type_info.type4.mem_bank[i].bank_name, &isUsed); + if(isUsed) return isUsed; + } + return isUsed; + default: RVM_TRACE_DEBUG_LOW("RVM: Error rvm_get_mb_level()"); + return isUsed; + } +} + +void rvm_shut_down_appli( T_RVM_MSG * msg_Ptr) { + T_RVM_PROCESSING_SWE * appli = NULL; /* pointer to the first element of the list */ + T_RVM_RETURN ret_value; + UINT8 swe_idx = 200; + T_RV_RETURN_PATH appli_return_path; + + + appli_return_path.callback_func = msg_Ptr->rp.callback_func; + appli_return_path.addr_id = msg_Ptr->header.src_addr_id; + + + RVM_TRACE_DEBUG_HIGH_PARAM("RVM: trying to stop Appli nb ", (UINT32)swe_idx); + + /* TO DO : REBUILD SWE LIST !!!! */ + if ( rvm_build_swe_list( &appli, msg_Ptr->swe_num, 1) != RVM_OK ) + { + /* Display error message + error case: use the return_path to inform the caller that an error occurs*/ + rvm_snd_msg_to_upper(RVM_START_APPLI, RVM_INVALID_PARAMETER, msg_Ptr->swe_num, appli_return_path); + RVM_TRACE_ERROR("RVM: SWE list built error"); + return; + } + + /* Stop all swe in the list that are used only once */ + if ((ret_value = rvm_stop_swe_list(appli, (T_RV_HDR*)msg_Ptr)) != RVM_OK ) + { + /* Display error message + TO DO: error case: use the return_path to inform the caller that an error occurs */ + RVM_TRACE_WARNING_PARAM("RVM: Error in SWE stop", (UINT32)ret_value); + return; + } + RVM_TRACE_DEBUG_LOW("RVM: SWE stop success"); + + + /* Suspend all swe that are tasks */ + if ((ret_value = rvm_suspend_swe_tasks(appli)) != RVM_OK ) + { + /* Display error message + TO DO: error case: use the return_path to inform the caller that an error occurs */ + RVM_TRACE_WARNING_PARAM("RVM: Error in tasks suspension", (UINT32)ret_value); + return; + } + RVM_TRACE_DEBUG_LOW("RVM: SWE task supsended"); + + /* Kill all SWEs */ + if ((ret_value = rvm_kill_swe_list(appli)) != RVM_OK) + { + /* Display error message + TO DO: error case: use the return_path to inform the caller that an error occurs */ + RVM_TRACE_WARNING_PARAM("RVM: Error in SWE killing", (UINT32)ret_value); + return; + } + RVM_TRACE_DEBUG_LOW("RVM: SWE kill success"); + + + /* Delete the swe Memory Banks */ + rvm_delete_created_mb(appli); + + /* Delete memory used and restore NOT_STARTED states */ + if ((ret_value = rvm_clean_env(appli)) != RVM_OK) + { + /* Display error message + TO DO: error case: use the return_path to inform the caller that an error occurs */ + RVM_TRACE_WARNING_PARAM("RVM: Error in Memory cleaning", (UINT32)ret_value); + return; + } + RVM_TRACE_DEBUG_LOW("RVM: Memory cleaning success"); + + + /* build a message and send the response to the caller */ + /* send a result using the return_path */ + rvm_snd_msg_to_upper(RVM_STOP_APPLI, RVM_OK, msg_Ptr->swe_num, appli_return_path); + + + /* and store the return_path */ + rvm_swe_array[ msg_Ptr->swe_num ].mmi_return_path.callback_func = msg_Ptr->rp.callback_func; + rvm_swe_array[ msg_Ptr->swe_num ].mmi_return_path.addr_id = msg_Ptr->header.src_addr_id; +} + +T_RVM_RETURN _fatal( T_RVM_PROCESSING_SWE * appli, UINT8 rm) { + T_RVM_PROCESSING_SWE * cur_swe = NULL; +// T_RVM_INFO_SWE swe_info; + + RVM_TRACE_DEBUG_LOW("RVM: Fatality handler: reclaiming system resources!"); + /* free all appli's system resources */ + for (cur_swe = appli; cur_swe!=NULL; ) { + if(rvm_swe_array[cur_swe->swe_id].swe_state!=SWE_RUNNING) + rvf_free_sys_resources(rvm_swe_array[cur_swe->swe_id].swe_addr_id, rm); + } + + + return RVM_OK; +} + +/******************************************************************************* +** +** Function rvm_generic_swe_core +** +** Description This is the main task core used for GROUP_MEMBER SWEs hosting +** and for SINGLE SWEs. +** +** Parameters: useless, may be for future evolutions if Nucleus really +** supports it. +** +** Returns None +** +*******************************************************************************/ +T_RVM_RETURN rvm_generic_swe_core(void) +{ + return RVM_OK; +}