MOAB: Mesh Oriented datABase
(version 5.4.1)
|
#include <iostream>
#include <cstdlib>
#include <vector>
#include <map>
#include <string>
#include <cstdio>
#include <iomanip>
#include <fstream>
#include "moab/MOABConfig.h"
#include <termios.h>
#include <sys/ioctl.h>
#include <cmath>
#include <cassert>
#include <cfloat>
#include "moab/Core.hpp"
#include "moab/Range.hpp"
#include "MBTagConventions.hpp"
#include "moab/Interface.hpp"
#include "moab/verdict/VerdictWrapper.hpp"
Go to the source code of this file.
Functions | |
static void | print_usage (const char *name, std::ostream &stream) |
int | main (int argc, char *argv[]) |
int main | ( | int | argc, |
char * | argv[] | ||
) |
Definition at line 36 of file quality.cpp.
References moab::VerdictWrapper::all_quality_measures(), moab::Range::begin(), moab::Range::clear(), moab::Range::end(), entities, moab::VerdictWrapper::entity_type_name(), ErrorCode, moab::ParallelComm::filter_pstatus(), moab::Core::get_adjacencies(), moab::Core::get_entities_by_handle(), moab::ParallelComm::get_pcomm(), moab::Core::id_from_handle(), moab::Core::load_file(), mb, moab::MB_EDGE_RATIO, moab::MB_QUALITY_COUNT, MB_SUCCESS, MBEDGE, MBENTITYSET, moab::Range::merge(), MPI_COMM_WORLD, moab::VerdictWrapper::num_qualities(), ofile, out_file, moab::VerdictWrapper::possible_quality(), print_usage(), PSTATUS_NOT, PSTATUS_NOT_OWNED, moab::VerdictWrapper::quality_name(), read_options, moab::VerdictWrapper::set_size(), size, moab::Range::size(), moab::Range::subset_by_dimension(), moab::Range::subset_by_type(), and moab::Interface::UNION.
{ int proc_id = 0, size = 1; #ifdef MOAB_HAVE_MPI MPI_Init( &argc, &argv ); MPI_Comm_rank( MPI_COMM_WORLD, &proc_id ); MPI_Comm_size( MPI_COMM_WORLD, &size ); #endif if( argc < 2 && 0 == proc_id ) { print_usage( argv[0], std::cerr ); #ifdef MOAB_HAVE_MPI MPI_Finalize(); #endif return 1; } Core mb; std::string read_options; if( size > 1 ) read_options = "PARALLEL=READ_PART;PARTITION=PARALLEL_PARTITION;PARALLEL_RESOLVE_SHARED_ENTS"; if( size > 1 && argc > 2 ) { std::cerr << " cannot use verbose option in parallel \n"; #ifdef MOAB_HAVE_MPI MPI_Finalize(); #endif return 1; } char* out_file = NULL; std::ofstream ofile; if( argc == 3 ) { std::cout << " write verbose output to a CSV file " << argv[2] << "\n"; out_file = argv[2]; ofile.open( out_file ); } if( MB_SUCCESS != mb.load_file( argv[1], 0, read_options.c_str() ) ) { fprintf( stderr, "Error reading file: %s\n", argv[1] ); #ifdef MOAB_HAVE_MPI MPI_Finalize(); #endif return 1; } VerdictWrapper vw( &mb ); vw.set_size( 1. ); // for relative size measures; maybe it should be user input Range entities; ErrorCode rval = mb.get_entities_by_handle( 0, entities ); if( MB_SUCCESS != rval ) { fprintf( stderr, "Error getting entities from file %s\n", argv[1] ); #ifdef MOAB_HAVE_MPI MPI_Finalize(); #endif return 1; } // get all edges and faces, force them to be created Range allfaces, alledges; Range cells = entities.subset_by_dimension( 3 ); rval = mb.get_adjacencies( cells, 2, true, allfaces, Interface::UNION ); if( MB_SUCCESS != rval ) { fprintf( stderr, "Error getting all faces" ); #ifdef MOAB_HAVE_MPI MPI_Finalize(); #endif return 1; } rval = mb.get_adjacencies( allfaces, 1, true, alledges, Interface::UNION ); if( MB_SUCCESS != rval ) { fprintf( stderr, "Error getting all edges" ); #ifdef MOAB_HAVE_MPI MPI_Finalize(); #endif return 1; } entities.merge( allfaces ); entities.merge( alledges ); for( EntityType et = MBENTITYSET; et >= MBEDGE; et-- ) { int num_qualities = vw.num_qualities( et ); if( !num_qualities ) continue; Range owned = entities.subset_by_type( et ); std::map< QualityType, double > qualities, minq, maxq; int ne_local = (int)owned.size(); int ne_global = ne_local; #ifdef MOAB_HAVE_MPI int mpi_err; if( size > 1 ) { // filter the entities not owned, so we do not process them more than once ParallelComm* pcomm = ParallelComm::get_pcomm( &mb, 0 ); Range current = owned; owned.clear(); rval = pcomm->filter_pstatus( current, PSTATUS_NOT_OWNED, PSTATUS_NOT, -1, &owned ); if( rval != MB_SUCCESS ) { MPI_Finalize(); return 1; } ne_local = (int)owned.size(); mpi_err = MPI_Reduce( &ne_local, &ne_global, 1, MPI_INT, MPI_SUM, 0, MPI_COMM_WORLD ); if( mpi_err ) { MPI_Finalize(); return 1; } } #endif if( ne_global > 0 ) { if( ne_local > 0 ) { Range::iterator it = owned.begin(); rval = vw.all_quality_measures( *it, qualities ); if( MB_SUCCESS != rval ) { fprintf( stderr, "Error getting quality for entity type %d with id %ld \n", et, (long)mb.id_from_handle( *it ) ); #ifdef MOAB_HAVE_MPI MPI_Finalize(); #endif return 1; } if( ofile.is_open() ) { // write first header or this entity type, then the first values, separated by // commas ofile << " There are " << ne_local << " entities of type " << vw.entity_type_name( et ) << " with " << qualities.size() << " qualities:\n" << " Entity id "; for( std::map< QualityType, double >::iterator qit = qualities.begin(); qit != qualities.end(); ++qit ) { ofile << ", " << vw.quality_name( qit->first ); } ofile << "\n"; ofile << mb.id_from_handle( *it ); for( std::map< QualityType, double >::iterator qit = qualities.begin(); qit != qualities.end(); ++qit ) { ofile << ", " << qit->second; } ofile << "\n"; } minq = qualities; maxq = qualities; ++it; for( ; it != owned.end(); ++it ) { rval = vw.all_quality_measures( *it, qualities ); if( MB_SUCCESS != rval ) { fprintf( stderr, "Error getting quality for entity type %d with id %ld \n", et, (long)mb.id_from_handle( *it ) ); #ifdef MOAB_HAVE_MPI MPI_Finalize(); #endif return 1; } if( ofile.is_open() ) { ofile << mb.id_from_handle( *it ); for( std::map< QualityType, double >::iterator qit = qualities.begin(); qit != qualities.end(); ++qit ) { ofile << ", " << qit->second; } ofile << "\n"; } std::map< QualityType, double >::iterator minit = minq.begin(); std::map< QualityType, double >::iterator maxit = maxq.begin(); for( std::map< QualityType, double >::iterator mit = qualities.begin(); mit != qualities.end(); ++mit, ++minit, ++maxit ) { if( mit->second > maxit->second ) maxit->second = mit->second; if( mit->second < minit->second ) minit->second = mit->second; } } } if( 0 == proc_id ) { std::cout << " \n\n " << ne_global << " entities of type " << vw.entity_type_name( et ) << "\n"; std::cout << std::setw( 30 ) << "Quality Name" << std::setw( 15 ) << " MIN" << std::setw( 15 ) << " MAX" << "\n"; } QualityType quality_type = MB_EDGE_RATIO; for( int i = 0; i < num_qualities; i++, quality_type = (QualityType)( quality_type + 1 ) ) { while( !( vw.possible_quality( et, quality_type ) ) && quality_type < MB_QUALITY_COUNT ) quality_type = (QualityType)( quality_type + 1 ); // will get them in order const char* name_q = vw.quality_name( quality_type ); double local_min, global_min; double local_max, global_max; if( ne_local > 0 ) { local_min = minq[quality_type]; local_max = maxq[quality_type]; } else { local_min = 1.e38; // so this task has no entities of this type local_max = -1.e38; // it can get here only in parallel } #ifdef MOAB_HAVE_MPI mpi_err = MPI_Reduce( &local_min, &global_min, 1, MPI_DOUBLE, MPI_MIN, 0, MPI_COMM_WORLD ); if( mpi_err ) { MPI_Finalize(); return 1; } mpi_err = MPI_Reduce( &local_max, &global_max, 1, MPI_DOUBLE, MPI_MAX, 0, MPI_COMM_WORLD ); if( mpi_err ) { MPI_Finalize(); return 1; } #else global_min = local_min; global_max = local_max; #endif if( 0 == proc_id ) { std::cout << std::setw( 30 ) << name_q << std::setw( 15 ) << global_min << std::setw( 15 ) << global_max << "\n"; } } } } // etype if( ofile.is_open() ) { ofile.close(); } #ifdef MOAB_HAVE_MPI MPI_Finalize(); #endif return 0; }
static void print_usage | ( | const char * | name, |
std::ostream & | stream | ||
) | [static] |
Definition at line 31 of file quality.cpp.
{ stream << "Usage: " << name << " <input_file> [<output_file>]" << std::endl; }