#!/usr/bin/env bash # Turn on extended globbing shopt -s extglob # Bail on any error set -e prog=$(basename $0) print_help() { cat < | ... ] DESCRIPTION Extracts backtraces and lock tables from Asterisk coredump files. For each coredump found, 4 new result files are created: - .brief.txt: The output of "thread apply all bt". - .thread1.txt: The output of "thread apply 1 bt full". - .full.txt: The output of "thread apply all bt full". - .locks.txt: If asterisk was compiled with "DEBUG_THREADS", this file will contain a dump of the locks table similar to doing a "core show locks" from the asterisk CLI. Optional features: - The running asterisk process can be suspended and dumped. - The coredumps can be merged into a tarball. - The coredumps can be deleted after processing. - The results files can be merged into a tarball. - The results files can be deleted after processing. Options: --help Print this help. --running Create a coredump from the running asterisk instance and process it along with any other coredumps found (if any). WARNING: This WILL interrupt call processing. You will be asked to confirm. The coredump will be written to /tmp if $OUTPUTDIR is not defined. --RUNNING Same as --running but without the confirmation prompt. DANGEROUS!! --latest Process only the latest coredump from those specified (based on last-modified time). If a dump of the running process was requested, it is always included in addition to the latest from the existing coredumps. --tarball-coredumps Creates a gzipped tarball of coredumps processed, their results txt files and copies of /etc/os-release, /usr/sbin/asterisk, /usr/lib(64)/libasterisk* and /usr/lib(64)/asterisk as those files are needed to properly examine the coredump. The file will be named $OUTPUTDIR/asterisk..coredumps.tar.gz or $OUTPUTDIR/asterisk-.coredumps.tar.gz if --tarball-uniqueid was specified. WARNING: This file could 1gb in size! Mutually exclusive with --tartball-results --delete-coredumps-after Deletes all processed coredumps regardless of whether a tarball was created. --tarball-results Creates a gzipped tarball of all result files produced. The tarball name will be: $OUTPUTDIR/asterisk..results.tar.gz Mutually exclusive with --tartball-coredumps --delete-results-after Deletes all processed results regardless of whether a tarball was created. It probably doesn't make sense to use this option unless you have also specified --tarball-results. --tarball-config Adds the contents of /etc/asterisk to the tarball created with --tarball-coredumps or --tarball-results. --tarball-uniqueid="" Normally DATEFORMAT is used to make the tarballs unique but you can use your own unique id in the tarball names such as the Jira issue id. --no-default-search Ignore COREDUMPS from the config files and process only coredumps listed on the command line (if any) and/or the running asterisk instance (if requested). --append-coredumps Append any coredumps specified on the command line to the config file specified ones instead of overriding them. --asterisk-binary Path to the asterisk binary. Default: look for asterisk in the PATH. | A list of coredumps or coredump search patterns. Unless --append-coredumps was specified, these entries will override those specified in the config files. Any resulting file that isn't actually a coredump is silently ignored. If your patterns contains spaces be sure to only quote the portion of the pattern that DOESN'T contain wildcard expressions. If you quote the whole pattern, it won't be expanded. If --no-default-search is specified and no files are specified on the command line, then the only the running asterisk process will be dumped (if requested). Otherwise if no files are specified on the command line the value of COREDUMPS from ast_debug_tools.conf will be used. Failing that, the following patterns will be used: /tmp/core[-._]asterisk!(*.txt) /tmp/core[-._]\$(hostname)!(*.txt) NOTES You must be root to use $prog. $OUTPUTDIR can be read from the current environment or from the ast_debug_tools.conf file described below. If not specified, work products are placed in the same directory as the core file. The script relies on not only bash, but also recent GNU date and gdb with python support. *BSD operating systems may require installation of the 'coreutils' and 'devel/gdb' packagess and minor tweaking of the ast_debug_tools.conf file. Any files output will have ':' characters changed to '-'. This is to facilitate uploading those files to Jira which doesn't like the colons. FILES /etc/asterisk/ast_debug_tools.conf ~/ast_debug_tools.conf ./ast_debug_tools.conf # # This file is used by the Asterisk debug tools. # Unlike other Asterisk config files, this one is # "sourced" by bash and must adhere to bash semantics. # # A list of coredumps and/or coredump search patterns. # Bash extended globs are enabled and any resulting files # that aren't actually coredumps are silently ignored # so you can be liberal with the globs. # # If your patterns contains spaces be sure to only quote # the portion of the pattern that DOESN'T contain wildcard # expressions. If you quote the whole pattern, it won't # be expanded and the glob characters will be treated as # literals. # # The exclusion of files ending ".txt" is just for # demonstration purposes as non-coredumps will be ignored # anyway. COREDUMPS=(/tmp/core[-._]asterisk!(*.txt) /tmp/core[-._]\$(hostname)!(*.txt)) # The directory to contain output files and work directories. # For output from existing core files, the default is the # directory that the core file is found in. For core files # produced from a running process, the default is /tmp. OUTPUTDIR=/some/directory # Date command for the "running" coredump and tarballs. # DATEFORMAT will be executed to get the timestamp. # Don't put quotes around the format string or they'll be # treated as literal characters. Also be aware of colons # in the output as you can't upload files with colons in # the name to Jira. # # Unix timestamp #DATEFORMAT='date +%s.%N' # # *BSD/MacOS doesn't support %N but after installing GNU # coreutils... #DATEFORMAT='gdate +%s.%N' # # Readable GMT #DATEFORMAT='date -u +%FT%H-%M-%S%z' # # Readable Local time DATEFORMAT='date +%FT%H-%M-%S%z' EOF exit 1 } if [ $EUID -ne 0 ] ; then echo "You must be root to use $prog." exit 1 fi running=false RUNNING=false latest=false tarball_coredumps=false tarball_config=false delete_coredumps_after=false tarball_results=false delete_results_after=false append_coredumps=false declare -a COREDUMPS declare -a ARGS_COREDUMPS # readconf reads a bash-sourceable file and sets variables # that havn't already been set. This allows variables set # on the command line or that are already in the environment # to take precedence over those read from the file. # # Setting the values can't be done in a subshell so you can't # just pipe the output of sed into the while. readconf() { while read line ; do v=${line%%=*} [ -z "${!v}" ] && eval $line || : done </dev/null || : ) if [[ "$result" =~ ^hello$ ]] ; then GDB=$g break fi done if [ -z "$GDB" ] ; then echo "No suitable gdb was found in $PATH" exit 1 fi if [ -n "$OUTPUTDIR" ] ; then if [ ! -d "$OUTPUTDIR" ] ; then echo "OUTPUTDIR $OUTPUTDIR doesn't exists or is not a directory" exit 1 fi fi if [ ${#COREDUMPS[@]} -eq 0 ] ; then COREDUMPS+=(/tmp/core[-._]asterisk!(*.txt) /tmp/core[-._]$(hostname)!(*.txt)) fi DATEFORMAT=${DATEFORMAT:-'date +%FT%H-%M-%S%z'} # Use "$@" (with the quotes) so spaces in patterns or # file names are preserved. # Later on when we have to iterate over COREDUMPS, we always # use the indexes rather than trying to expand the values of COREDUMPS # just in case. for a in "$@" ; do case "$a" in --running) running=true ;; --RUNNING) RUNNING=true ;; --no-default-search) # Clean out COREDUMPS from config files COREDUMPS=() ;; --latest) latest=true ;; --tarball-coredumps) tarball_coredumps=true ;; --tarball-config) tarball_config=true ;; --delete-coredumps-after) delete_coredumps_after=true ;; --tarball-results) tarball_results=true ;; --delete-results-after) delete_results_after=true ;; --append-coredumps) append_coredumps=true ;; --tarball-uniqueid=*) tarball_uniqueid=${a#*=} ;; --asterisk-bin=*) asterisk_bin=${a#*=} ;; --help|-*) print_help ;; *) ARGS_COREDUMPS+=("$a") # If any files are specified on the command line, ignore those # specified in the config files unless append-coredumps was specified. if ! $append_coredumps ; then COREDUMPS=() fi esac done # append coredumps/patterns specified as command line arguments to COREDUMPS. for i in ${!ARGS_COREDUMPS[@]} ; do COREDUMPS+=("${ARGS_COREDUMPS[$i]}") done # At this point, all glob entries that match files should be expanded. # Any entries that don't exist are probably globs that didn't match anything # and need to be pruned. Any non coredumps are also pruned. for i in ${!COREDUMPS[@]} ; do if [ ! -f "${COREDUMPS[$i]}" ] ; then unset COREDUMPS[$i] continue fi # Some versions of 'file' don't allow only the first n bytes of the # file to be processed so we use dd to grab just the first 32 bytes. mimetype=$(dd if="${COREDUMPS[$i]}" bs=32 count=1 2>/dev/null | file -bi -) if [[ ! "$mimetype" =~ coredump ]] ; then unset COREDUMPS[$i] continue fi done # Sort and weed out any dups IFS=$'\x0a' readarray -t COREDUMPS < <(echo -n "${COREDUMPS[*]}" | sort -u ) unset IFS # If --latest, get the last modified timestamp of each file, # sort them, then return the latest. if [ ${#COREDUMPS[@]} -gt 0 ] && $latest ; then lf=$(find "${COREDUMPS[@]}" -printf '%T@ %p\n' | sort -n | tail -1) COREDUMPS=("${lf#* }") fi # Timestamp to use for output files df=${tarball_uniqueid:-$(${DATEFORMAT})} if [ x"$asterisk_bin" = x ]; then asterisk_bin=$(which asterisk) fi if $running || $RUNNING ; then # We need to go through some gyrations to find the pid of the running # MAIN asterisk process and not someone or something running asterisk -r. unset pid # Simplest case first... pids=$(pgrep -f "$asterisk_bin" || : ) pidcount=$(echo $pids | wc -w) # Single process, great. if [ $pidcount -eq 1 ] ; then pid=$pids echo "Found a single asterisk instance running as process $pid" fi # More than 1 asterisk process running if [ x"$pid" = x ] ; then # More than 1 process running, let's try asking asterisk for it's # pidfile pidfile=$("$asterisk_bin" -rx "core show settings" 2>/dev/null | sed -n -r -e "s/^\s*pid file:\s+(.*)/\1/gpi") # We found it if [ x"$pidfile" != x -a -f "$pidfile" ] ; then pid=$(cat "$pidfile") echo "Found pidfile $pidfile with process $pid" fi fi # It's possible that asterisk was started with the -C option which means the # control socket and pidfile might not be where we expect. We're going to # have to parse the process arguments to see if -C was specified. # The first process that has a -C argument determines which config # file to use to find the pidfile of the main process. # NOTE: The ps command doesn't quote command line arguments that it # displays so we need to look in /proc//cmdline. if [ x"$pid" = x ] ; then # BSDs might not mount /proc by default :( mounted_proc=0 if uname -o | grep -qi "bsd" ; then if ! mount | grep -qi "/proc" ; then echo "Temporarily mounting /proc" mounted_proc=1 mount -t procfs proc /proc fi fi for p in $pids ; do # Fields in cmdline are delimited by NULLs astetcconf=$(sed -n -r -e "s/.*\x00-C\x00([^\x00]+).*/\1/gp" /proc/$p/cmdline) if [ x"$astetcconf" != x ] ; then pidfile=$("$asterisk_bin" -C "$astetcconf" -rx "core show settings" 2>/dev/null | sed -n -r -e "s/^\s*pid file:\s+(.*)/\1/gpi") if [ x"$pidfile" != x -a -f "$pidfile" ] ; then pid=$(cat "$pidfile") echo "Found pidfile $pidfile the hard way with process $pid" break fi fi done if [ $mounted_proc -eq 1 ] ; then echo "Unmounting /proc" umount /proc fi fi if [ x"$pid" = x ] ; then >&2 echo "Can't determine pid of the running asterisk instance" exit 1 fi if $RUNNING ; then answer=Y else read -p "WARNING: Taking a core dump of the running asterisk instance will suspend call processing while the dump is saved. Do you wish to continue? (y/N) " answer fi if [[ "$answer" =~ ^[Yy] ]] ; then cf="${OUTPUTDIR:-/tmp}/core-asterisk-running-$df" echo "Dumping running asterisk process to $cf" ${GDB} ${asterisk_bin} -p $pid -q --batch --ex "gcore $cf" >/dev/null 2>&1 COREDUMPS+=("$cf") else echo "Skipping dump of running process" fi fi if [ "${#COREDUMPS[@]}" -eq 0 ] ; then echo "No coredumps found" print_help fi # Extract the gdb scripts from the end of this script # and save them to /tmp/.gdbinit gdbinit=${OUTPUTDIR:-/tmp}/.ast_coredumper.gdbinit trap "rm $gdbinit" EXIT ss=`egrep -n "^#@@@SCRIPTSTART@@@" $0 |cut -f1 -d:` tail -n +${ss} $0 >$gdbinit # Now iterate over the coredumps and dump the debugging info for i in ${!COREDUMPS[@]} ; do cf=$(readlink -ne ${COREDUMPS[$i]}) echo "Processing $cf" cfdir=`dirname ${cf}` cfname=`basename ${cf}` outputdir=${OUTPUTDIR:-${cfdir}} ${GDB} -n --batch -q --ex "source $gdbinit" "$asterisk_bin" "$cf" 2>/dev/null | ( of=/dev/null while IFS= read line ; do if [[ "$line" =~ !@!@!@!\ ([^\ ]+)\ !@!@!@! ]] ; then of=${outputdir}/${cfname}-${BASH_REMATCH[1]} of=${of//:/-} rm -f "$of" echo "Creating $of" fi echo -e $"$line" >> "$of" done ) if $tarball_coredumps ; then cfname=${cfname//:/-} tf=${outputdir}/${cfname}.tar.gz echo "Creating ${tf}" dest=${outputdir}/${cfname}.output rm -rf ${dest} 2>/dev/null || : libdir=usr/lib [ -d /usr/lib64 ] && libdir+=64 mkdir -p ${dest}/tmp ${dest}/${libdir}/asterisk ${dest}/etc ${dest}/usr/sbin ln -s ${cf} ${dest}/tmp/${cfname} cp ${outputdir}/${cfname}*.txt ${dest}/tmp/ [ -f /etc/os-release ] && cp /etc/os-release ${dest}/etc/ if $tarball_config ; then cp -a /etc/asterisk ${dest}/etc/ fi cp -a /${libdir}/libasterisk* ${dest}/${libdir}/ cp -a /${libdir}/asterisk/* ${dest}/${libdir}/asterisk/ cp -a /usr/sbin/asterisk ${dest}/usr/sbin rm -rf ${tf} tar -chzf ${tf} --transform="s/^[.]/${cfname}.output/" -C ${dest} . sleep 3 rm -rf ${dest} echo "Created $tf" elif $tarball_results ; then cfname=${cfname//:/-} tf=${outputdir}/${cfname}.tar.gz echo "Creating ${tf}" dest=${outputdir}/${cfname}.output rm -rf ${dest} 2>/dev/null || : mkdir -p ${dest} cp ${outputdir}/${cfname}*.txt ${dest}/ if $tarball_config ; then mkdir -p ${dest}/etc cp -a /etc/asterisk ${dest}/etc/ fi tar -chzf ${tf} --transform="s/^[.]/${cfname}/" -C ${dest} . rm -rf ${dest} echo "Created $tf" fi if $delete_coredumps_after ; then rm -rf "${cf}" fi if $delete_results_after ; then rm -rf "${cf//:/-}"-{brief,full,thread1,locks,info}.txt fi done exit # Be careful editng the inline scripts. # They're space-indented. # We need the python bit because lock_infos isn't # a valid symbol in asterisk unless DEBUG_THREADS was # used during the compile. Also, interrupt and continue # are only valid for a running program. #@@@SCRIPTSTART@@@ python import datetime def timeval_to_datetime(value): """Convert a timeval struct to a python datetime object Args: value: A gdb Value representing a C timeval Return: A python datetime object """ sec = int(value['tv_sec']) usec = int(value['tv_usec']) return datetime.datetime.fromtimestamp(sec + usec / float(1000000)) def s_strip(value): """Convert the given value to a string, and strip any leading/trailing spaces and/or quotes. Args: name: The gdb Value to convert and strip Return: The stripped value as a string """ if value == None: return "None" try: if 'char *' not in str(value.type) and 'char [' not in str(value.type): # Use the string method for everything but string pointers (only # points to first letter) and non-string values in general return value.string().strip('" ') or "" except: pass return str(value).strip('" ') or "" def get(name): """Retrieve a named variable's value as a string using GDB. Args: name: The name of the variable to look up Return: The variable's value as a string """ return s_strip(gdb.parse_and_eval(name)) def get_container_hash_objects(name, type, on_object=None): """Retrieve a list of objects from an ao2_container_hash. Expected on_object signature: res, stop = on_object(GDB Value) The given callback, on_object, is called for each object found in the container. The callback is passed a dereferenced GDB Value object and expects an object to be returned, which is then appended to a list of objects to be returned by this function. Iteration can be stopped by returning "True" for the second return value. If on_object is not specified then the dereferenced GDB value is instead added directly to the returned list. Args: name: The name of the ao2_container type: The type of objects stored in the container on_object: Optional function called on each object found Return: A list of container objects """ objs = [] try: container = gdb.parse_and_eval(name).cast( gdb.lookup_type('struct ao2_container_hash').pointer()) # Loop over every bucket searching for hash bucket nodes for n in range(container['n_buckets']): node = container['buckets'][n]['list']['last'] while node: # Each node holds the needed object obj = node.dereference()['common']['obj'].cast( gdb.lookup_type(type).pointer()).dereference() res, stop = on_object(obj) if on_object else (obj, False) if res: objs.append(res) if stop: return objs node = node.dereference()['links']['last'] except Exception as e: print("{0} - {1}".format(name, e)) pass return objs def get_container_rbtree_objects(name, type, on_object=None): """Retrieve a list of objects from an ao2_container_rbtree. Expected on_object signature: res, stop = on_object(GDB Value) The given callback, on_object, is called for each object found in the container. The callback is passed a dereferenced GDB Value object and expects an object to be returned, which is then appended to a list of objects to be returned by this function. Iteration can be stopped by returning "True" for the second return value. If on_object is not specified then the dereferenced GDB value is instead added directly to the returned list. Args: name: The name of the ao2_container type: The type of objects stored in the container on_object: Optional function called on each object found Return: A list of container objects """ objs = [] def handle_node(node): if not node: return True # Each node holds the needed object obj = node.dereference()['common']['obj'].cast( gdb.lookup_type(type).pointer()).dereference() res, stop = on_object(obj) if on_object else (obj, False) if res: objs.append(res) return not stop and (handle_node(node['left']) and handle_node(node['right'])) try: container = gdb.parse_and_eval(name).cast( gdb.lookup_type('struct ao2_container_rbtree').pointer()) handle_node(container['root']) except Exception as e: print("{0} - {1}".format(name, e)) pass return objs def build_info(): try: return ("Asterisk {0} built by {1} @ {2} on a {3} running {4} on {5}" .format(get("asterisk_version"), get("ast_build_user"), get("ast_build_hostname"), get("ast_build_machine"), get("ast_build_os"), get("ast_build_date"))) except: return "Unable to retrieve build info" def build_opts(): try: return get("asterisk_build_opts") except: return "Unable to retrieve build options" def uptime(): try: started = timeval_to_datetime(gdb.parse_and_eval("ast_startuptime")) loaded = timeval_to_datetime(gdb.parse_and_eval("ast_lastreloadtime")) return ("System started: {0}\n" "Last reload: {1}".format(started, loaded)) except: return "Unable to retrieve uptime" class TaskProcessor(object): template = ("{name:70} {processed:>10} {in_queue:>10} {max_depth:>10} " "{low_water:>10} {high_water:>10}") header = {'name': 'Processor', 'processed': 'Processed', 'in_queue': 'In Queue', 'max_depth': 'Max Depth', 'low_water': 'Low water', 'high_water': 'High water'} @staticmethod def objects(): try: objs = get_container_hash_objects('tps_singletons', 'struct ast_taskprocessor', TaskProcessor.from_value) objs.sort(key=lambda x: x.name.lower()) except Exception as e: return [] return objs @staticmethod def from_value(value): return TaskProcessor( value['name'], value['stats']['_tasks_processed_count'], value['tps_queue_size'], value['stats']['max_qsize'], value['tps_queue_low'], value['tps_queue_high']), False def __init__(self, name, processed, in_queue, max_depth, low_water, high_water): self.name = s_strip(name) self.processed = int(processed) self.in_queue = int(in_queue) self.max_depth = int(max_depth) self.low_water = int(low_water) self.high_water = int(high_water) class Channel(object): template = ("{name:30} {context:>20} {exten:>20} {priority:>10} {state:>25} " "{app:>20} {data:>30} {caller_id:>15} {created:>30} " "{account_code:>15} {peer_account:>15} {bridge_id:>38}") header = {'name': 'Channel', 'context': 'Context', 'exten': 'Extension', 'priority': 'Priority', 'state': "State", 'app': 'Application', 'data': 'Data', 'caller_id': 'CallerID', 'created': 'Created', 'account_code': 'Accountcode', 'peer_account': 'PeerAccount', 'bridge_id': 'BridgeID'} @staticmethod def objects(): try: objs = get_container_hash_objects('channels', 'struct ast_channel', Channel.from_value) objs.sort(key=lambda x: x.name.lower()) except: return [] return objs @staticmethod def from_value(value): bridge_id = None if value['bridge']: bridge_id = value['bridge']['uniqueid'] return Channel( value['name'], value['context'], value['exten'], value['priority'], value['state'], value['appl'], value['data'], value['caller']['id']['number']['str'], timeval_to_datetime(value['creationtime']), value['accountcode'], value['peeraccount'], bridge_id), False @staticmethod def summary(): try: return ("{0} active channels\n" "{1} active calls\n" "{2} calls processed".format( int(gdb.parse_and_eval( 'channels').dereference()['elements']), get("countcalls"), get("totalcalls"))) except: return "Unable to retrieve channel summary" def __init__(self, name, context=None, exten=None, priority=None, state=None, app=None, data=None, caller_id=None, created=None, account_code=None, peer_account=None, bridge_id=None): self.name = s_strip(name) self.context = s_strip(context) self.exten = s_strip(exten) self.priority = int(priority) self.state = s_strip(state) self.app = s_strip(app) self.data = s_strip(data) self.caller_id = s_strip(caller_id) self.created = s_strip(created) self.account_code = s_strip(account_code) self.peer_account = s_strip(peer_account) self.bridge_id = s_strip(bridge_id) class Bridge(object): template = ("{uniqueid:38} {num_channels:>15} {subclass:>10} {tech:>20} " "{created:>30}") header = {'uniqueid': 'Bridge-ID', 'num_channels': 'Chans', 'subclass': 'Type', 'tech': 'Technology', 'created': 'Created'} @staticmethod def objects(): try: objs = get_container_rbtree_objects('bridges', 'struct ast_bridge', Bridge.from_value) objs.sort(key=lambda x: x.uniqueid.lower()) except: return [] return objs @staticmethod def from_value(value): return Bridge( value['uniqueid'], value['num_channels'], timeval_to_datetime(value['creationtime']), value['v_table']['name'], value['technology']['name']), False def __init__(self, uniqueid, num_channels=None, created=None, subclass=None, tech=None): self.uniqueid = s_strip(uniqueid) self.num_channels = int(num_channels) self.created = s_strip(created) self.subclass = s_strip(subclass) self.tech = s_strip(tech) class DumpAsteriskCommand(gdb.Command): def __init__(self): super(DumpAsteriskCommand, self).__init__ ("dump-asterisk", gdb.COMMAND_OBSCURE, gdb.COMPLETE_COMMAND) def print_table(self, type): plural = "{0}s".format(type.__name__) objs = type.objects() if not len(objs): print("{0} not found\n".format(plural)) return print("{0} ({1}):\n".format(plural, len(objs))) print(type.template.format(**type.header)) for obj in objs: print(type.template.format(**vars(obj))) print("\n") def invoke(self, arg, from_tty): try: gdb.execute("interrupt", from_tty) except: pass print("!@!@!@! thread1.txt !@!@!@!\n") try: gdb.execute("p $_siginfo", from_tty) gdb.execute("info signal $_siginfo.si_signo") except: pass try: gdb.execute("thread apply 1 bt full", from_tty) except: pass print("!@!@!@! brief.txt !@!@!@!\n") try: gdb.execute("p $_siginfo", from_tty) gdb.execute("info signal $_siginfo.si_signo") except: pass try: gdb.execute("thread apply all bt", from_tty) except: pass print("!@!@!@! full.txt !@!@!@!\n") try: gdb.execute("p $_siginfo", from_tty) gdb.execute("info signal $_siginfo.si_signo") except: pass try: gdb.execute("thread apply all bt full", from_tty) except: pass print("!@!@!@! locks.txt !@!@!@!\n") try: gdb.execute("p $_siginfo", from_tty) gdb.execute("info signal $_siginfo.si_signo") except: pass try: gdb.execute("show_locks", from_tty) except: pass print("!@!@!@! info.txt !@!@!@!\n") gdb.execute('set print addr off') try: print("{0}\n".format(build_info())) print("{0}\n".format(uptime())) print("Build options = {0}\n".format(build_opts())) self.print_table(TaskProcessor) self.print_table(Bridge) self.print_table(Channel) print(Channel.summary()) except: pass finally: gdb.execute('set print addr on') try: gdb.execute("continue", from_tty) except: pass DumpAsteriskCommand () end define show_locks set $n = lock_infos.first if $argc == 0 printf " where_held count-|\n" printf " suspended-| |\n" printf " type- | times locked-| | |\n" printf "thread status file line function lock name | lock addr | | |\n" else printf "thread,status,file,line,function,lock_name,lock_type,lock_addr,times_locked,suspended,where_held_count,where_held_file,where_held_line,where_held_function,there_held_thread\n" end while $n if $n->num_locks > 0 set $i = 0 while $i < $n->num_locks if $n->locks[$i]->suspended == 0 if ((ast_mutex_t *)$n->locks[$i]->lock_addr)->tracking if $n->locks[$i]->type > 0 set $track = ((ast_rwlock_t *)$n->locks[$i]->lock_addr)->track else set $track = ((ast_mutex_t *)$n->locks[$i]->lock_addr)->track end end set $reentrancy = $track->reentrancy set $pending = $n->locks[$i]->pending if $argc > 0 printf "%p,%d,%s,%d,%s,%s,%d,%p,%d,%d,%d",\ $n->thread_id, $n->locks[$i]->pending, $n->locks[$i]->file, $n->locks[$i]->line_num, $n->locks[$i]->func,\ $n->locks[$i]->lock_name, $n->locks[$i]->type, $n->locks[$i]->lock_addr, $n->locks[$i]->times_locked,\ $n->locks[$i]->suspended, $track->reentrancy if $reentrancy if $pending printf ",%s,%d,%s,%p", $track->file[0], $track->lineno[0], $track->func[0], $track->thread[0] end end else if $n->locks[$i]->pending < 0 printf "%p failed %-20s %6d %-36s %-20s %d %14p %3d %d %d",\ $n->thread_id,\ $n->locks[$i]->file, $n->locks[$i]->line_num, $n->locks[$i]->func,\ $n->locks[$i]->lock_name, $n->locks[$i]->type, $n->locks[$i]->lock_addr, $n->locks[$i]->times_locked,\ $n->locks[$i]->suspended, $track->reentrancy end if $n->locks[$i]->pending == 0 printf "%p holding %-20s %6d %-36s %-20s %d %14p %3d %d %d",\ $n->thread_id,\ $n->locks[$i]->file, $n->locks[$i]->line_num, $n->locks[$i]->func,\ $n->locks[$i]->lock_name, $n->locks[$i]->type, $n->locks[$i]->lock_addr, $n->locks[$i]->times_locked,\ $n->locks[$i]->suspended, $track->reentrancy end if $n->locks[$i]->pending > 0 printf "%p waiting %-20s %6d %-36s %-20s %d %14p %3d %d %d",\ $n->thread_id,\ $n->locks[$i]->file, $n->locks[$i]->line_num, $n->locks[$i]->func,\ $n->locks[$i]->lock_name, $n->locks[$i]->type, $n->locks[$i]->lock_addr, $n->locks[$i]->times_locked,\ $n->locks[$i]->suspended, $track->reentrancy end if $reentrancy if $pending printf "\n held at: %-20s %6d %-36s by 0x%08lx", $track->file[0], $track->lineno[0], $track->func[0], $track->thread_id[0] end end end printf "\n" end set $i = $i + 1 end end set $n = $n->entry->next end end dump-asterisk