mirror of
https://github.com/yuzu-emu/mbedtls.git
synced 2024-11-30 20:04:20 +01:00
80448aae2c
GCC 11 generated the warnings because the parameter `ret_buf` was declared as `const char[10]`, but some of the arguments provided in `run_test_snprintf` are shorter literals, like "". Now the type of `ret_buf` is `const char *`. Both implementations of `test_snprintf` were fixed. Signed-off-by: Rodrigo Dias Correa <rodrigo@correas.us>
821 lines
25 KiB
Plaintext
821 lines
25 KiB
Plaintext
#line 2 "suites/host_test.function"
|
|
|
|
/**
|
|
* \brief Verifies that string is in string parameter format i.e. "<str>"
|
|
* It also strips enclosing '"' from the input string.
|
|
*
|
|
* \param str String parameter.
|
|
*
|
|
* \return 0 if success else 1
|
|
*/
|
|
int verify_string( char **str )
|
|
{
|
|
if( ( *str )[0] != '"' ||
|
|
( *str )[strlen( *str ) - 1] != '"' )
|
|
{
|
|
mbedtls_fprintf( stderr,
|
|
"Expected string (with \"\") for parameter and got: %s\n", *str );
|
|
return( -1 );
|
|
}
|
|
|
|
( *str )++;
|
|
( *str )[strlen( *str ) - 1] = '\0';
|
|
|
|
return( 0 );
|
|
}
|
|
|
|
/**
|
|
* \brief Verifies that string is an integer. Also gives the converted
|
|
* integer value.
|
|
*
|
|
* \param str Input string.
|
|
* \param value Pointer to int for output value.
|
|
*
|
|
* \return 0 if success else 1
|
|
*/
|
|
int verify_int( char *str, int *value )
|
|
{
|
|
size_t i;
|
|
int minus = 0;
|
|
int digits = 1;
|
|
int hex = 0;
|
|
|
|
for( i = 0; i < strlen( str ); i++ )
|
|
{
|
|
if( i == 0 && str[i] == '-' )
|
|
{
|
|
minus = 1;
|
|
continue;
|
|
}
|
|
|
|
if( ( ( minus && i == 2 ) || ( !minus && i == 1 ) ) &&
|
|
str[i - 1] == '0' && ( str[i] == 'x' || str[i] == 'X' ) )
|
|
{
|
|
hex = 1;
|
|
continue;
|
|
}
|
|
|
|
if( ! ( ( str[i] >= '0' && str[i] <= '9' ) ||
|
|
( hex && ( ( str[i] >= 'a' && str[i] <= 'f' ) ||
|
|
( str[i] >= 'A' && str[i] <= 'F' ) ) ) ) )
|
|
{
|
|
digits = 0;
|
|
break;
|
|
}
|
|
}
|
|
|
|
if( digits )
|
|
{
|
|
if( hex )
|
|
*value = strtol( str, NULL, 16 );
|
|
else
|
|
*value = strtol( str, NULL, 10 );
|
|
|
|
return( 0 );
|
|
}
|
|
|
|
mbedtls_fprintf( stderr,
|
|
"Expected integer for parameter and got: %s\n", str );
|
|
return( KEY_VALUE_MAPPING_NOT_FOUND );
|
|
}
|
|
|
|
|
|
/**
|
|
* \brief Usage string.
|
|
*
|
|
*/
|
|
#define USAGE \
|
|
"Usage: %s [OPTIONS] files...\n\n" \
|
|
" Command line arguments:\n" \
|
|
" files... One or more test data files. If no file is\n" \
|
|
" specified the following default test case\n" \
|
|
" file is used:\n" \
|
|
" %s\n\n" \
|
|
" Options:\n" \
|
|
" -v | --verbose Display full information about each test\n" \
|
|
" -h | --help Display this information\n\n", \
|
|
argv[0], \
|
|
"TESTCASE_FILENAME"
|
|
|
|
|
|
/**
|
|
* \brief Read a line from the passed file pointer.
|
|
*
|
|
* \param f FILE pointer
|
|
* \param buf Pointer to memory to hold read line.
|
|
* \param len Length of the buf.
|
|
*
|
|
* \return 0 if success else -1
|
|
*/
|
|
int get_line( FILE *f, char *buf, size_t len )
|
|
{
|
|
char *ret;
|
|
int i = 0, str_len = 0, has_string = 0;
|
|
|
|
/* Read until we get a valid line */
|
|
do
|
|
{
|
|
ret = fgets( buf, len, f );
|
|
if( ret == NULL )
|
|
return( -1 );
|
|
|
|
str_len = strlen( buf );
|
|
|
|
/* Skip empty line and comment */
|
|
if ( str_len == 0 || buf[0] == '#' )
|
|
continue;
|
|
has_string = 0;
|
|
for ( i = 0; i < str_len; i++ )
|
|
{
|
|
char c = buf[i];
|
|
if ( c != ' ' && c != '\t' && c != '\n' &&
|
|
c != '\v' && c != '\f' && c != '\r' )
|
|
{
|
|
has_string = 1;
|
|
break;
|
|
}
|
|
}
|
|
} while( !has_string );
|
|
|
|
/* Strip new line and carriage return */
|
|
ret = buf + strlen( buf );
|
|
if( ret-- > buf && *ret == '\n' )
|
|
*ret = '\0';
|
|
if( ret-- > buf && *ret == '\r' )
|
|
*ret = '\0';
|
|
|
|
return( 0 );
|
|
}
|
|
|
|
/**
|
|
* \brief Splits string delimited by ':'. Ignores '\:'.
|
|
*
|
|
* \param buf Input string
|
|
* \param len Input string length
|
|
* \param params Out params found
|
|
* \param params_len Out params array len
|
|
*
|
|
* \return Count of strings found.
|
|
*/
|
|
static int parse_arguments( char *buf, size_t len, char **params,
|
|
size_t params_len )
|
|
{
|
|
size_t cnt = 0, i;
|
|
char *cur = buf;
|
|
char *p = buf, *q;
|
|
|
|
params[cnt++] = cur;
|
|
|
|
while( *p != '\0' && p < ( buf + len ) )
|
|
{
|
|
if( *p == '\\' )
|
|
{
|
|
p++;
|
|
p++;
|
|
continue;
|
|
}
|
|
if( *p == ':' )
|
|
{
|
|
if( p + 1 < buf + len )
|
|
{
|
|
cur = p + 1;
|
|
TEST_HELPER_ASSERT( cnt < params_len );
|
|
params[cnt++] = cur;
|
|
}
|
|
*p = '\0';
|
|
}
|
|
|
|
p++;
|
|
}
|
|
|
|
/* Replace newlines, question marks and colons in strings */
|
|
for( i = 0; i < cnt; i++ )
|
|
{
|
|
p = params[i];
|
|
q = params[i];
|
|
|
|
while( *p != '\0' )
|
|
{
|
|
if( *p == '\\' && *( p + 1 ) == 'n' )
|
|
{
|
|
p += 2;
|
|
*( q++ ) = '\n';
|
|
}
|
|
else if( *p == '\\' && *( p + 1 ) == ':' )
|
|
{
|
|
p += 2;
|
|
*( q++ ) = ':';
|
|
}
|
|
else if( *p == '\\' && *( p + 1 ) == '?' )
|
|
{
|
|
p += 2;
|
|
*( q++ ) = '?';
|
|
}
|
|
else
|
|
*( q++ ) = *( p++ );
|
|
}
|
|
*q = '\0';
|
|
}
|
|
|
|
return( cnt );
|
|
}
|
|
|
|
/**
|
|
* \brief Converts parameters into test function consumable parameters.
|
|
* Example: Input: {"int", "0", "char*", "Hello",
|
|
* "hex", "abef", "exp", "1"}
|
|
* Output: {
|
|
* 0, // Verified int
|
|
* "Hello", // Verified string
|
|
* 2, { 0xab, 0xef },// Converted len,hex pair
|
|
* 9600 // Evaluated expression
|
|
* }
|
|
*
|
|
*
|
|
* \param cnt Parameter array count.
|
|
* \param params Out array of found parameters.
|
|
* \param int_params_store Memory for storing processed integer parameters.
|
|
*
|
|
* \return 0 for success else 1
|
|
*/
|
|
static int convert_params( size_t cnt , char ** params , int * int_params_store )
|
|
{
|
|
char ** cur = params;
|
|
char ** out = params;
|
|
int ret = DISPATCH_TEST_SUCCESS;
|
|
|
|
while ( cur < params + cnt )
|
|
{
|
|
char * type = *cur++;
|
|
char * val = *cur++;
|
|
|
|
if ( strcmp( type, "char*" ) == 0 )
|
|
{
|
|
if ( verify_string( &val ) == 0 )
|
|
{
|
|
*out++ = val;
|
|
}
|
|
else
|
|
{
|
|
ret = ( DISPATCH_INVALID_TEST_DATA );
|
|
break;
|
|
}
|
|
}
|
|
else if ( strcmp( type, "int" ) == 0 )
|
|
{
|
|
if ( verify_int( val, int_params_store ) == 0 )
|
|
{
|
|
*out++ = (char *) int_params_store++;
|
|
}
|
|
else
|
|
{
|
|
ret = ( DISPATCH_INVALID_TEST_DATA );
|
|
break;
|
|
}
|
|
}
|
|
else if ( strcmp( type, "hex" ) == 0 )
|
|
{
|
|
if ( verify_string( &val ) == 0 )
|
|
{
|
|
size_t len;
|
|
|
|
TEST_HELPER_ASSERT(
|
|
mbedtls_test_unhexify( (unsigned char *) val, strlen( val ),
|
|
val, &len ) == 0 );
|
|
|
|
*int_params_store = len;
|
|
*out++ = val;
|
|
*out++ = (char *)(int_params_store++);
|
|
}
|
|
else
|
|
{
|
|
ret = ( DISPATCH_INVALID_TEST_DATA );
|
|
break;
|
|
}
|
|
}
|
|
else if ( strcmp( type, "exp" ) == 0 )
|
|
{
|
|
int exp_id = strtol( val, NULL, 10 );
|
|
if ( get_expression ( exp_id, int_params_store ) == 0 )
|
|
{
|
|
*out++ = (char *)int_params_store++;
|
|
}
|
|
else
|
|
{
|
|
ret = ( DISPATCH_INVALID_TEST_DATA );
|
|
break;
|
|
}
|
|
}
|
|
else
|
|
{
|
|
ret = ( DISPATCH_INVALID_TEST_DATA );
|
|
break;
|
|
}
|
|
}
|
|
return( ret );
|
|
}
|
|
|
|
/**
|
|
* \brief Tests snprintf implementation with test input.
|
|
*
|
|
* \note
|
|
* At high optimization levels (e.g. gcc -O3), this function may be
|
|
* inlined in run_test_snprintf. This can trigger a spurious warning about
|
|
* potential misuse of snprintf from gcc -Wformat-truncation (observed with
|
|
* gcc 7.2). This warning makes tests in run_test_snprintf redundant on gcc
|
|
* only. They are still valid for other compilers. Avoid this warning by
|
|
* forbidding inlining of this function by gcc.
|
|
*
|
|
* \param n Buffer test length.
|
|
* \param ref_buf Expected buffer.
|
|
* \param ref_ret Expected snprintf return value.
|
|
*
|
|
* \return 0 for success else 1
|
|
*/
|
|
#if defined(__GNUC__)
|
|
__attribute__((__noinline__))
|
|
#endif
|
|
static int test_snprintf( size_t n, const char *ref_buf, int ref_ret )
|
|
{
|
|
int ret;
|
|
char buf[10] = "xxxxxxxxx";
|
|
const char ref[10] = "xxxxxxxxx";
|
|
|
|
if( n >= sizeof( buf ) )
|
|
return( -1 );
|
|
ret = mbedtls_snprintf( buf, n, "%s", "123" );
|
|
if( ret < 0 || (size_t) ret >= n )
|
|
ret = -1;
|
|
|
|
if( strncmp( ref_buf, buf, sizeof( buf ) ) != 0 ||
|
|
ref_ret != ret ||
|
|
memcmp( buf + n, ref + n, sizeof( buf ) - n ) != 0 )
|
|
{
|
|
return( 1 );
|
|
}
|
|
|
|
return( 0 );
|
|
}
|
|
|
|
/**
|
|
* \brief Tests snprintf implementation.
|
|
*
|
|
* \param none
|
|
*
|
|
* \return 0 for success else 1
|
|
*/
|
|
static int run_test_snprintf( void )
|
|
{
|
|
return( test_snprintf( 0, "xxxxxxxxx", -1 ) != 0 ||
|
|
test_snprintf( 1, "", -1 ) != 0 ||
|
|
test_snprintf( 2, "1", -1 ) != 0 ||
|
|
test_snprintf( 3, "12", -1 ) != 0 ||
|
|
test_snprintf( 4, "123", 3 ) != 0 ||
|
|
test_snprintf( 5, "123", 3 ) != 0 );
|
|
}
|
|
|
|
/** \brief Write the description of the test case to the outcome CSV file.
|
|
*
|
|
* \param outcome_file The file to write to.
|
|
* If this is \c NULL, this function does nothing.
|
|
* \param argv0 The test suite name.
|
|
* \param test_case The test case description.
|
|
*/
|
|
static void write_outcome_entry( FILE *outcome_file,
|
|
const char *argv0,
|
|
const char *test_case )
|
|
{
|
|
/* The non-varying fields are initialized on first use. */
|
|
static const char *platform = NULL;
|
|
static const char *configuration = NULL;
|
|
static const char *test_suite = NULL;
|
|
|
|
if( outcome_file == NULL )
|
|
return;
|
|
|
|
if( platform == NULL )
|
|
{
|
|
platform = getenv( "MBEDTLS_TEST_PLATFORM" );
|
|
if( platform == NULL )
|
|
platform = "unknown";
|
|
}
|
|
if( configuration == NULL )
|
|
{
|
|
configuration = getenv( "MBEDTLS_TEST_CONFIGURATION" );
|
|
if( configuration == NULL )
|
|
configuration = "unknown";
|
|
}
|
|
if( test_suite == NULL )
|
|
{
|
|
test_suite = strrchr( argv0, '/' );
|
|
if( test_suite != NULL )
|
|
test_suite += 1; // skip the '/'
|
|
else
|
|
test_suite = argv0;
|
|
}
|
|
|
|
/* Write the beginning of the outcome line.
|
|
* Ignore errors: writing the outcome file is on a best-effort basis. */
|
|
mbedtls_fprintf( outcome_file, "%s;%s;%s;%s;",
|
|
platform, configuration, test_suite, test_case );
|
|
}
|
|
|
|
/** \brief Write the result of the test case to the outcome CSV file.
|
|
*
|
|
* \param outcome_file The file to write to.
|
|
* If this is \c NULL, this function does nothing.
|
|
* \param unmet_dep_count The number of unmet dependencies.
|
|
* \param unmet_dependencies The array of unmet dependencies.
|
|
* \param missing_unmet_dependencies Non-zero if there was a problem tracking
|
|
* all unmet dependencies, 0 otherwise.
|
|
* \param ret The test dispatch status (DISPATCH_xxx).
|
|
* \param test_info A pointer to the test info structure.
|
|
*/
|
|
static void write_outcome_result( FILE *outcome_file,
|
|
size_t unmet_dep_count,
|
|
int unmet_dependencies[],
|
|
int missing_unmet_dependencies,
|
|
int ret,
|
|
const test_info_t *info )
|
|
{
|
|
if( outcome_file == NULL )
|
|
return;
|
|
|
|
/* Write the end of the outcome line.
|
|
* Ignore errors: writing the outcome file is on a best-effort basis. */
|
|
switch( ret )
|
|
{
|
|
case DISPATCH_TEST_SUCCESS:
|
|
if( unmet_dep_count > 0 )
|
|
{
|
|
size_t i;
|
|
mbedtls_fprintf( outcome_file, "SKIP" );
|
|
for( i = 0; i < unmet_dep_count; i++ )
|
|
{
|
|
mbedtls_fprintf( outcome_file, "%c%d",
|
|
i == 0 ? ';' : ':',
|
|
unmet_dependencies[i] );
|
|
}
|
|
if( missing_unmet_dependencies )
|
|
mbedtls_fprintf( outcome_file, ":..." );
|
|
break;
|
|
}
|
|
switch( info->result )
|
|
{
|
|
case TEST_RESULT_SUCCESS:
|
|
mbedtls_fprintf( outcome_file, "PASS;" );
|
|
break;
|
|
case TEST_RESULT_SKIPPED:
|
|
mbedtls_fprintf( outcome_file, "SKIP;Runtime skip" );
|
|
break;
|
|
default:
|
|
mbedtls_fprintf( outcome_file, "FAIL;%s:%d:%s",
|
|
info->filename, info->line_no,
|
|
info->test );
|
|
break;
|
|
}
|
|
break;
|
|
case DISPATCH_TEST_FN_NOT_FOUND:
|
|
mbedtls_fprintf( outcome_file, "FAIL;Test function not found" );
|
|
break;
|
|
case DISPATCH_INVALID_TEST_DATA:
|
|
mbedtls_fprintf( outcome_file, "FAIL;Invalid test data" );
|
|
break;
|
|
case DISPATCH_UNSUPPORTED_SUITE:
|
|
mbedtls_fprintf( outcome_file, "SKIP;Unsupported suite" );
|
|
break;
|
|
default:
|
|
mbedtls_fprintf( outcome_file, "FAIL;Unknown cause" );
|
|
break;
|
|
}
|
|
mbedtls_fprintf( outcome_file, "\n" );
|
|
fflush( outcome_file );
|
|
}
|
|
|
|
/**
|
|
* \brief Desktop implementation of execute_tests().
|
|
* Parses command line and executes tests from
|
|
* supplied or default data file.
|
|
*
|
|
* \param argc Command line argument count.
|
|
* \param argv Argument array.
|
|
*
|
|
* \return Program exit status.
|
|
*/
|
|
int execute_tests( int argc , const char ** argv )
|
|
{
|
|
/* Local Configurations and options */
|
|
const char *default_filename = "DATA_FILE";
|
|
const char *test_filename = NULL;
|
|
const char **test_files = NULL;
|
|
size_t testfile_count = 0;
|
|
int option_verbose = 0;
|
|
size_t function_id = 0;
|
|
|
|
/* Other Local variables */
|
|
int arg_index = 1;
|
|
const char *next_arg;
|
|
size_t testfile_index, i, cnt;
|
|
int ret;
|
|
unsigned total_errors = 0, total_tests = 0, total_skipped = 0;
|
|
FILE *file;
|
|
char buf[5000];
|
|
char *params[50];
|
|
/* Store for proccessed integer params. */
|
|
int int_params[50];
|
|
void *pointer;
|
|
#if defined(__unix__) || (defined(__APPLE__) && defined(__MACH__))
|
|
int stdout_fd = -1;
|
|
#endif /* __unix__ || __APPLE__ __MACH__ */
|
|
const char *outcome_file_name = getenv( "MBEDTLS_TEST_OUTCOME_FILE" );
|
|
FILE *outcome_file = NULL;
|
|
|
|
#if defined(MBEDTLS_MEMORY_BUFFER_ALLOC_C) && \
|
|
!defined(TEST_SUITE_MEMORY_BUFFER_ALLOC)
|
|
unsigned char alloc_buf[1000000];
|
|
mbedtls_memory_buffer_alloc_init( alloc_buf, sizeof( alloc_buf ) );
|
|
#endif
|
|
|
|
/*
|
|
* The C standard doesn't guarantee that all-bits-0 is the representation
|
|
* of a NULL pointer. We do however use that in our code for initializing
|
|
* structures, which should work on every modern platform. Let's be sure.
|
|
*/
|
|
memset( &pointer, 0, sizeof( void * ) );
|
|
if( pointer != NULL )
|
|
{
|
|
mbedtls_fprintf( stderr, "all-bits-zero is not a NULL pointer\n" );
|
|
return( 1 );
|
|
}
|
|
|
|
/*
|
|
* Make sure we have a snprintf that correctly zero-terminates
|
|
*/
|
|
if( run_test_snprintf() != 0 )
|
|
{
|
|
mbedtls_fprintf( stderr, "the snprintf implementation is broken\n" );
|
|
return( 1 );
|
|
}
|
|
|
|
if( outcome_file_name != NULL && *outcome_file_name != '\0' )
|
|
{
|
|
outcome_file = fopen( outcome_file_name, "a" );
|
|
if( outcome_file == NULL )
|
|
{
|
|
mbedtls_fprintf( stderr, "Unable to open outcome file. Continuing anyway.\n" );
|
|
}
|
|
}
|
|
|
|
while( arg_index < argc )
|
|
{
|
|
next_arg = argv[arg_index];
|
|
|
|
if( strcmp( next_arg, "--verbose" ) == 0 ||
|
|
strcmp( next_arg, "-v" ) == 0 )
|
|
{
|
|
option_verbose = 1;
|
|
}
|
|
else if( strcmp(next_arg, "--help" ) == 0 ||
|
|
strcmp(next_arg, "-h" ) == 0 )
|
|
{
|
|
mbedtls_fprintf( stdout, USAGE );
|
|
mbedtls_exit( EXIT_SUCCESS );
|
|
}
|
|
else
|
|
{
|
|
/* Not an option, therefore treat all further arguments as the file
|
|
* list.
|
|
*/
|
|
test_files = &argv[ arg_index ];
|
|
testfile_count = argc - arg_index;
|
|
}
|
|
|
|
arg_index++;
|
|
}
|
|
|
|
/* If no files were specified, assume a default */
|
|
if ( test_files == NULL || testfile_count == 0 )
|
|
{
|
|
test_files = &default_filename;
|
|
testfile_count = 1;
|
|
}
|
|
|
|
/* Initialize the struct that holds information about the last test */
|
|
memset( &test_info, 0, sizeof( test_info ) );
|
|
|
|
/* Now begin to execute the tests in the testfiles */
|
|
for ( testfile_index = 0;
|
|
testfile_index < testfile_count;
|
|
testfile_index++ )
|
|
{
|
|
size_t unmet_dep_count = 0;
|
|
int unmet_dependencies[20];
|
|
int missing_unmet_dependencies = 0;
|
|
|
|
test_filename = test_files[ testfile_index ];
|
|
|
|
file = fopen( test_filename, "r" );
|
|
if( file == NULL )
|
|
{
|
|
mbedtls_fprintf( stderr, "Failed to open test file: %s\n",
|
|
test_filename );
|
|
if( outcome_file != NULL )
|
|
fclose( outcome_file );
|
|
return( 1 );
|
|
}
|
|
|
|
while( !feof( file ) )
|
|
{
|
|
if( unmet_dep_count > 0 )
|
|
{
|
|
mbedtls_fprintf( stderr,
|
|
"FATAL: Dep count larger than zero at start of loop\n" );
|
|
mbedtls_exit( MBEDTLS_EXIT_FAILURE );
|
|
}
|
|
unmet_dep_count = 0;
|
|
missing_unmet_dependencies = 0;
|
|
|
|
if( ( ret = get_line( file, buf, sizeof(buf) ) ) != 0 )
|
|
break;
|
|
mbedtls_fprintf( stdout, "%s%.66s",
|
|
test_info.result == TEST_RESULT_FAILED ? "\n" : "", buf );
|
|
mbedtls_fprintf( stdout, " " );
|
|
for( i = strlen( buf ) + 1; i < 67; i++ )
|
|
mbedtls_fprintf( stdout, "." );
|
|
mbedtls_fprintf( stdout, " " );
|
|
fflush( stdout );
|
|
write_outcome_entry( outcome_file, argv[0], buf );
|
|
|
|
total_tests++;
|
|
|
|
if( ( ret = get_line( file, buf, sizeof( buf ) ) ) != 0 )
|
|
break;
|
|
cnt = parse_arguments( buf, strlen( buf ), params,
|
|
sizeof( params ) / sizeof( params[0] ) );
|
|
|
|
if( strcmp( params[0], "depends_on" ) == 0 )
|
|
{
|
|
for( i = 1; i < cnt; i++ )
|
|
{
|
|
int dep_id = strtol( params[i], NULL, 10 );
|
|
if( dep_check( dep_id ) != DEPENDENCY_SUPPORTED )
|
|
{
|
|
if( unmet_dep_count <
|
|
ARRAY_LENGTH( unmet_dependencies ) )
|
|
{
|
|
unmet_dependencies[unmet_dep_count] = dep_id;
|
|
unmet_dep_count++;
|
|
}
|
|
else
|
|
{
|
|
missing_unmet_dependencies = 1;
|
|
}
|
|
}
|
|
}
|
|
|
|
if( ( ret = get_line( file, buf, sizeof( buf ) ) ) != 0 )
|
|
break;
|
|
cnt = parse_arguments( buf, strlen( buf ), params,
|
|
sizeof( params ) / sizeof( params[0] ) );
|
|
}
|
|
|
|
// If there are no unmet dependencies execute the test
|
|
if( unmet_dep_count == 0 )
|
|
{
|
|
test_info.result = TEST_RESULT_SUCCESS;
|
|
test_info.step = (unsigned long)( -1 );
|
|
|
|
#if defined(__unix__) || (defined(__APPLE__) && defined(__MACH__))
|
|
/* Suppress all output from the library unless we're verbose
|
|
* mode
|
|
*/
|
|
if( !option_verbose )
|
|
{
|
|
stdout_fd = redirect_output( stdout, "/dev/null" );
|
|
if( stdout_fd == -1 )
|
|
{
|
|
/* Redirection has failed with no stdout so exit */
|
|
exit( 1 );
|
|
}
|
|
}
|
|
#endif /* __unix__ || __APPLE__ __MACH__ */
|
|
|
|
function_id = strtoul( params[0], NULL, 10 );
|
|
if ( (ret = check_test( function_id )) == DISPATCH_TEST_SUCCESS )
|
|
{
|
|
ret = convert_params( cnt - 1, params + 1, int_params );
|
|
if ( DISPATCH_TEST_SUCCESS == ret )
|
|
{
|
|
ret = dispatch_test( function_id, (void **)( params + 1 ) );
|
|
}
|
|
}
|
|
|
|
#if defined(__unix__) || (defined(__APPLE__) && defined(__MACH__))
|
|
if( !option_verbose && restore_output( stdout, stdout_fd ) )
|
|
{
|
|
/* Redirection has failed with no stdout so exit */
|
|
exit( 1 );
|
|
}
|
|
#endif /* __unix__ || __APPLE__ __MACH__ */
|
|
|
|
}
|
|
|
|
write_outcome_result( outcome_file,
|
|
unmet_dep_count, unmet_dependencies,
|
|
missing_unmet_dependencies,
|
|
ret, &test_info );
|
|
if( unmet_dep_count > 0 || ret == DISPATCH_UNSUPPORTED_SUITE )
|
|
{
|
|
total_skipped++;
|
|
mbedtls_fprintf( stdout, "----" );
|
|
|
|
if( 1 == option_verbose && ret == DISPATCH_UNSUPPORTED_SUITE )
|
|
{
|
|
mbedtls_fprintf( stdout, "\n Test Suite not enabled" );
|
|
}
|
|
|
|
if( 1 == option_verbose && unmet_dep_count > 0 )
|
|
{
|
|
mbedtls_fprintf( stdout, "\n Unmet dependencies: " );
|
|
for( i = 0; i < unmet_dep_count; i++ )
|
|
{
|
|
mbedtls_fprintf( stdout, "%d ",
|
|
unmet_dependencies[i] );
|
|
}
|
|
if( missing_unmet_dependencies )
|
|
mbedtls_fprintf( stdout, "..." );
|
|
}
|
|
mbedtls_fprintf( stdout, "\n" );
|
|
fflush( stdout );
|
|
|
|
unmet_dep_count = 0;
|
|
missing_unmet_dependencies = 0;
|
|
}
|
|
else if( ret == DISPATCH_TEST_SUCCESS )
|
|
{
|
|
if( test_info.result == TEST_RESULT_SUCCESS )
|
|
{
|
|
mbedtls_fprintf( stdout, "PASS\n" );
|
|
}
|
|
else if( test_info.result == TEST_RESULT_SKIPPED )
|
|
{
|
|
mbedtls_fprintf( stdout, "----\n" );
|
|
total_skipped++;
|
|
}
|
|
else
|
|
{
|
|
total_errors++;
|
|
mbedtls_fprintf( stdout, "FAILED\n" );
|
|
mbedtls_fprintf( stdout, " %s\n at ",
|
|
test_info.test );
|
|
if( test_info.step != (unsigned long)( -1 ) )
|
|
{
|
|
mbedtls_fprintf( stdout, "step %lu, ",
|
|
test_info.step );
|
|
}
|
|
mbedtls_fprintf( stdout, "line %d, %s",
|
|
test_info.line_no, test_info.filename );
|
|
}
|
|
fflush( stdout );
|
|
}
|
|
else if( ret == DISPATCH_INVALID_TEST_DATA )
|
|
{
|
|
mbedtls_fprintf( stderr, "FAILED: FATAL PARSE ERROR\n" );
|
|
fclose( file );
|
|
mbedtls_exit( 2 );
|
|
}
|
|
else if( ret == DISPATCH_TEST_FN_NOT_FOUND )
|
|
{
|
|
mbedtls_fprintf( stderr, "FAILED: FATAL TEST FUNCTION NOT FOUND\n" );
|
|
fclose( file );
|
|
mbedtls_exit( 2 );
|
|
}
|
|
else
|
|
total_errors++;
|
|
}
|
|
fclose( file );
|
|
}
|
|
|
|
if( outcome_file != NULL )
|
|
fclose( outcome_file );
|
|
|
|
mbedtls_fprintf( stdout, "\n----------------------------------------------------------------------------\n\n");
|
|
if( total_errors == 0 )
|
|
mbedtls_fprintf( stdout, "PASSED" );
|
|
else
|
|
mbedtls_fprintf( stdout, "FAILED" );
|
|
|
|
mbedtls_fprintf( stdout, " (%u / %u tests (%u skipped))\n",
|
|
total_tests - total_errors, total_tests, total_skipped );
|
|
|
|
#if defined(MBEDTLS_MEMORY_BUFFER_ALLOC_C) && \
|
|
!defined(TEST_SUITE_MEMORY_BUFFER_ALLOC)
|
|
#if defined(MBEDTLS_MEMORY_DEBUG)
|
|
mbedtls_memory_buffer_alloc_status();
|
|
#endif
|
|
mbedtls_memory_buffer_alloc_free();
|
|
#endif
|
|
|
|
return( total_errors != 0 );
|
|
}
|