Dre4m Shell
Server IP : 85.214.239.14  /  Your IP : 3.133.120.150
Web Server : Apache/2.4.62 (Debian)
System : Linux h2886529.stratoserver.net 4.9.0 #1 SMP Mon Sep 30 15:36:27 MSK 2024 x86_64
User : www-data ( 33)
PHP Version : 7.4.18
Disable Function : pcntl_alarm,pcntl_fork,pcntl_waitpid,pcntl_wait,pcntl_wifexited,pcntl_wifstopped,pcntl_wifsignaled,pcntl_wifcontinued,pcntl_wexitstatus,pcntl_wtermsig,pcntl_wstopsig,pcntl_signal,pcntl_signal_get_handler,pcntl_signal_dispatch,pcntl_get_last_error,pcntl_strerror,pcntl_sigprocmask,pcntl_sigwaitinfo,pcntl_sigtimedwait,pcntl_exec,pcntl_getpriority,pcntl_setpriority,pcntl_async_signals,pcntl_unshare,
MySQL : OFF  |  cURL : OFF  |  WGET : ON  |  Perl : ON  |  Python : ON  |  Sudo : ON  |  Pkexec : OFF
Directory :  /proc/3/cwd/usr/lib/node_modules/pm2/node_modules/@pm2/agent/src/push/

Upload File :
current_dir [ Writeable ] document_root [ Writeable ]

 

Command :


[ HOME SHELL ]     

Current File : /proc/3/cwd/usr/lib/node_modules/pm2/node_modules/@pm2/agent/src/push/TransactionAggregator.js
'use strict'

const cst = require('../../constants.js')
const log = require('debug')('interactor:aggregator')
const Utility = require('../Utility.js')
const fclone = require('fclone')
const Histogram = require('../utils/probes/Histogram')

/**
 *
 * # Data structure sent to interactor
 *
 * {
 *  'process_name': {
 *    process : {},         // PM2 process meta data
 *    data : {
 *      routes : [          // array of all routes ordered by count
 *        {
 *          path: '/',       // path of the route
 *          meta: {
 *            count: 50,     // count of this route
 *            max: 300,      // max latency of this route
 *            min: 50,       // min latency of this route
 *            mean: 120      // mean latency of this route
 *          }
 *          variances:  [{  // array of variance order by count
 *           spans : [
 *              ...         // transactions
 *           ],
 *           count: 50,     // count of this variance
 *           max: 300,      // max latency of this variance
 *           min: 50,       // min latency of this variance
 *           mean: 120      // mean latency of this variance
 *          }]
 *        }
 *      ],
 *      meta : {
 *        trace_count : 50,  // trace number
 *        mean_latency      : 40,  // global app latency in ms
 *        http_meter        : 30,  // global app req per minutes
 *        db_meter          : 20,  // number of database transaction per min
 *      }
 *    }
 *   }
 * }
 */

module.exports = class TransactionAggregator {
  constructor (pushInteractor) {
    this.processes = {}
    this.stackParser = pushInteractor._stackParser
    this.pushInteractor = pushInteractor

    this.LABELS = {
      'HTTP_RESPONSE_CODE_LABEL_KEY': 'http/status_code',
      'HTTP_URL_LABEL_KEY': 'http/url',
      'HTTP_METHOD_LABEL_KEY': 'http/method',
      'HTTP_RESPONSE_SIZE_LABEL_KEY': 'http/response/size',
      'STACK_TRACE_DETAILS_KEY': 'stacktrace',
      'ERROR_DETAILS_NAME': 'error/name',
      'ERROR_DETAILS_MESSAGE': 'error/message',
      'HTTP_SOURCE_IP': 'http/source/ip',
      'HTTP_PATH_LABEL_KEY': 'http/path'
    }
    this.SPANS_DB = ['redis', 'mysql', 'pg', 'mongo', 'outbound_http']
    this.REGEX_JSON_CLEANUP = /":(?!\[|{)\\"[^"]*\\"|":(["'])(?:(?=(\\?))\2.)*?\1|":(?!\[|{)[^,}\]]*|":\[[^{]*]/g

    this.init()
  }

  /**
   * First method to call in real environment
   * - Listen to restart event for initialization period
   * - Clear aggregation on process stop
   * - Launch worker to attach data to be pushed to KM
   */
  init () {
    // New Process Online, reset & wait a bit before processing
    this.pushInteractor._ipm2.bus.on('process:event', (data) => {
      if (data.event !== 'online' || !this.processes[data.process.name]) return false

      let rev = (data.process.versioning && data.process.versioning.revision)
        ? data.process.versioning.revision : null

      this.resetAggregation(data.process.name, {
        rev: rev,
        server: this.pushInteractor.opts.MACHINE_NAME
      })
    })

    // Process getting offline, delete aggregation
    this.pushInteractor._ipm2.bus.on('process:event', (data) => {
      if (data.event !== 'stop' || !this.processes[data.process.name]) return false
      log('Deleting aggregation for %s', data.process.name)
      delete this.processes[data.process.name]
    })

    this.launchWorker()
  }

  /**
   * Reset aggregation for target app_name
   */
  resetAggregation (appName, meta) {
    log('Reseting agg for app:%s meta:%j', appName, meta)

    if (this.processes[appName].initialization_timeout) {
      log('Reseting initialization timeout app:%s', appName)
      clearTimeout(this.processes[appName].initialization_timeout)
      clearInterval(this.processes[appName].notifier)
      this.processes[appName].notifier = null
    }

    this.processes[appName] = this.initializeRouteMeta({
      name: appName,
      rev: meta.rev,
      server: meta.server
    })

    let start = Date.now()
    this.processes[appName].notifier = setInterval(_ => {
      let elapsed = Date.now() - start
      // failsafe
      if (elapsed / 1000 > cst.AGGREGATION_DURATION) {
        clearInterval(this.processes[appName].notifier)
        this.processes[appName].notifier = null
      }

      let msg = {
        data: {
          learning_duration: cst.AGGREGATION_DURATION,
          elapsed: elapsed
        },
        process: this.processes[appName].process
      }
      this.pushInteractor && this.pushInteractor.transport.send('axm:transaction:learning', msg)
    }, 5000)

    this.processes[appName].initialization_timeout = setTimeout(_ => {
      log('Initialization timeout finished for app:%s', appName)
      clearInterval(this.processes[appName].notifier)
      this.processes[appName].notifier = null
      this.processes[appName].initialization_timeout = null
    }, cst.AGGREGATION_DURATION)
  }

  /**
   * Clear aggregated data for all process
   */
  clearData () {
    Object.keys(this.processes).forEach((process) => {
      this.resetAggregation(process, this.processes[process].process)
    })
  }

  /**
   * Generate new entry for application
   *
   * @param {Object} process process meta
   */
  initializeRouteMeta (process) {
    if (process.pm_id) delete process.pm_id

    return {
      routes: {},
      meta: {
        trace_count: 0,
        http_meter: new Utility.EWMA(),
        db_meter: new Utility.EWMA(),
        histogram: new Histogram({ measurement: 'median' }),
        db_histograms: {}
      },
      process: process
    }
  }

  getAggregation () {
    return this.processes
  }

  validateData (packet) {
    if (!packet || !packet.data) {
      log('Packet malformated', packet)
      return false
    }

    if (!packet.process) {
      log('Got packet without process: %j', packet)
      return false
    }

    if (!packet.data.spans || !packet.data.spans[0]) {
      log('Trace without spans: %s', Object.keys(packet.data))
      return false
    }

    if (!packet.data.spans[0].labels) {
      log('Trace spans without labels: %s', Object.keys(packet.data.spans))
      return false
    }

    return true
  }

  /**
   * Main method to aggregate and compute stats for traces
   *
   * @param {Object} packet
   * @param {Object} packet.process  process metadata
   * @param {Object} packet.data     trace
   */
  aggregate (packet) {
    if (this.validateData(packet) === false) return false

    const newTrace = packet.data
    const appName = packet.process.name

    if (!this.processes[appName]) {
      this.processes[appName] = this.initializeRouteMeta(packet.process)
    }

    let process = this.processes[appName]

    // Get http path of current span
    let path = newTrace.spans[0].labels[this.LABELS.HTTP_PATH_LABEL_KEY]
    if (!path) return false

    // Cleanup spans
    this.censorSpans(newTrace.spans)

    // remove spans with startTime == endTime
    newTrace.spans = newTrace.spans.filter((span) => {
      return span.endTime !== span.startTime
    })

    // compute duration of child spans
    newTrace.spans.forEach((span) => {
      span.mean = Math.round(new Date(span.endTime) - new Date(span.startTime))
      delete span.endTime
    })

    // Update app meta (mean_latency, http_meter, db_meter, trace_count)
    newTrace.spans.forEach((span) => {
      if (!span.name || !span.kind) return false

      if (span.kind === 'RPC_SERVER') {
        process.meta.histogram.update(span.mean)
        return process.meta.http_meter.update()
      }

      // Override outbount http queries for processing
      if (span.labels && span.labels['http/method'] && span.labels['http/status_code']) {
        span.labels['service'] = span.name
        span.name = 'outbound_http'
      }

      for (let i = 0, len = this.SPANS_DB.length; i < len; i++) {
        if (span.name.indexOf(this.SPANS_DB[i]) > -1) {
          process.meta.db_meter.update()
          if (!process.meta.db_histograms[this.SPANS_DB[i]]) {
            process.meta.db_histograms[this.SPANS_DB[i]] = new Histogram({ measurement: 'mean' })
          }
          process.meta.db_histograms[this.SPANS_DB[i]].update(span.mean)
          break
        }
      }
    })

    process.meta.trace_count++

    /**
     * Handle traces aggregation
     */
    if (path[0] === '/' && path !== '/') {
      path = path.substr(1, path.length - 1)
    }

    let matched = this.matchPath(path, process.routes)

    if (!matched) {
      process.routes[path] = []
      this.mergeTrace(process.routes[path], newTrace, process)
    } else {
      this.mergeTrace(process.routes[matched], newTrace, process)
    }

    return this.processes
  }

  /**
   * Merge new trace and compute mean, min, max, count
   *
   * @param {Object}  aggregated previous aggregated route
   * @param {Object}  trace
   */
  mergeTrace (aggregated, trace, process) {
    if (!aggregated || !trace) return

    // if the trace doesn't any spans stop aggregation here
    if (trace.spans.length === 0) return

    // create data structure if needed
    if (!aggregated.variances) aggregated.variances = []
    if (!aggregated.meta) {
      aggregated.meta = {
        histogram: new Histogram({ measurement: 'median' }),
        meter: new Utility.EWMA()
      }
    }

    aggregated.meta.histogram.update(trace.spans[0].mean)
    aggregated.meta.meter.update()

    const merge = (variance) => {
      // no variance found so its a new one
      if (variance == null) {
        delete trace.projectId
        delete trace.traceId
        trace.histogram = new Histogram({ measurement: 'median' })
        trace.histogram.update(trace.spans[0].mean)

        trace.spans.forEach((span) => {
          span.histogram = new Histogram({ measurement: 'median' })
          span.histogram.update(span.mean)
          delete span.mean
        })

        // parse strackrace
        this.parseStacktrace(trace.spans)
        aggregated.variances.push(trace)
      } else {
        // check to see if request is anormally slow, if yes send it as inquisitor
        if (trace.spans[0].mean > variance.histogram.percentiles([0.95])[0.95] &&
          typeof pushInteractor !== 'undefined' && !process.initialization_timeout) {
          // serialize and add metadata
          this.parseStacktrace(trace.spans)
          let data = {
            trace: fclone(trace.spans),
            variance: fclone(variance.spans.map((span) => {
              return {
                labels: span.labels,
                kind: span.kind,
                name: span.name,
                startTime: span.startTime,
                percentiles: {
                  p5: variance.histogram.percentiles([0.5])[0.5],
                  p95: variance.histogram.percentiles([0.95])[0.95]
                }
              }
            })),
            meta: {
              value: trace.spans[0].mean,
              percentiles: {
                p5: variance.histogram.percentiles([0.5])[0.5],
                p75: variance.histogram.percentiles([0.75])[0.75],
                p95: variance.histogram.percentiles([0.95])[0.95],
                p99: variance.histogram.percentiles([0.99])[0.99]
              },
              min: variance.histogram.getMin(),
              max: variance.histogram.getMax(),
              count: variance.histogram.getCount()
            },
            process: process.process
          }
          this.pushInteractor.transport.send('axm:transaction:outlier', data)
        }

        // variance found, merge spans
        variance.histogram.update(trace.spans[0].mean)

        // update duration of spans to be mean
        this.updateSpanDuration(variance.spans, trace.spans, variance.count)

        // delete stacktrace before merging
        trace.spans.forEach((span) => {
          delete span.labels.stacktrace
        })
      }
    }

    // for every variance, check spans same variance
    for (let i = 0; i < aggregated.variances.length; i++) {
      if (this.compareList(aggregated.variances[i].spans, trace.spans)) {
        return merge(aggregated.variances[i])
      }
    }
    // else its a new variance
    return merge(null)
  }

  /**
   * Parkour simultaneously both spans list to update value of the first one using value of the second one
   * The first should be variance already aggregated for which we want to merge the second one
   * The second one is a new trace, so we need to re-compute mean/min/max time for each spans
   */
  updateSpanDuration (spans, newSpans) {
    for (let i = 0, len = spans.length; i < len; i++) {
      if (!newSpans[i]) continue
      spans[i].histogram.update(newSpans[i].mean)
    }
  }

  /**
   * Compare two spans list by going down on each span and comparing child and attribute
   */
  compareList (one, two) {
    if (one.length !== two.length) return false

    for (let i = 0, len = one; i < len; i++) {
      if (one[i].name !== two[i].name) return false
      if (one[i].kind !== two[i].kind) return false
      if (!one[i].labels && two[i].labels) return false
      if (one[i].labels && !two[i].labels) return false
      if (one[i].labels.length !== two[i].labels.length) return false
    }
    return true
  }

  /**
   * Will return the route if we found an already matched route
   */
  matchPath (path, routes) {
    // empty route is / without the fist slash
    if (!path || !routes) return false
    if (path === '/') return routes[path] ? path : null

    // remove the last slash if exist
    if (path[path.length - 1] === '/') {
      path = path.substr(0, path.length - 1)
    }

    // split to get array of segment
    path = path.split('/')

    // if the path has only one segment, we just need to compare the key
    if (path.length === 1) return routes[path[0]] ? routes[path[0]] : null

    // check in routes already stored for match
    let keys = Object.keys(routes)
    for (let i = 0, len = keys.length; i < len; i++) {
      let route = keys[i]
      let segments = route.split('/')

      if (segments.length !== path.length) continue

      for (let j = path.length - 1; j >= 0; j--) {
        // different segment, try to find if new route or not
        if (path[j] !== segments[j]) {
          // if the aggregator already have matched that segment with a wildcard and the next segment is the same
          if (this.isIdentifier(path[j]) && segments[j] === '*' && path[j - 1] === segments[j - 1]) {
            return segments.join('/')
          // case a let in url match, so we continue because they must be other let in url
          } else if (path[j - 1] !== undefined && path[j - 1] === segments[j - 1] && this.isIdentifier(path[j]) && this.isIdentifier(segments[j])) {
            segments[j] = '*'
            // update routes in cache
            routes[segments.join('/')] = routes[route]
            delete routes[keys[i]]
            return segments.join('/')
          } else {
            break
          }
        }

        // if finish to iterate over segment of path, we must be on the same route
        if (j === 0) return segments.join('/')
      }
    }
  }

  launchWorker () {
    this._worker = setInterval(_ => {
      let normalized = this.prepareAggregationforShipping()
      Object.keys(normalized).forEach((key) => {
        this.pushInteractor.transport.send('axm:transaction', normalized[key])
      })
    }, cst.TRACE_FLUSH_INTERVAL)
  }

  /**
   * Normalize aggregation
   */
  prepareAggregationforShipping () {
    let normalized = {}

    // Iterate each applications
    Object.keys(this.processes).forEach((appName) => {
      let process = this.processes[appName]
      let routes = process.routes

      if (process.initialization_timeout) {
        log('Waiting for app %s to be initialized', appName)
        return null
      }

      normalized[appName] = {
        data: {
          routes: [],
          meta: fclone({
            trace_count: process.meta.trace_count,
            http_meter: Math.round(process.meta.http_meter.rate(1000) * 100) / 100,
            db_meter: Math.round(process.meta.db_meter.rate(1000) * 100) / 100,
            http_percentiles: {
              median: process.meta.histogram.percentiles([0.5])[0.5],
              p95: process.meta.histogram.percentiles([0.95])[0.95],
              p99: process.meta.histogram.percentiles([0.99])[0.99]
            },
            db_percentiles: {}
          })
        },
        process: process.process
      }

      // compute percentiles for each db spans if they exist
      this.SPANS_DB.forEach((name) => {
        let histogram = process.meta.db_histograms[name]
        if (!histogram) return
        normalized[appName].data.meta.db_percentiles[name] = fclone(histogram.percentiles([0.5])[0.5])
      })

      Object.keys(routes).forEach((path) => {
        let data = routes[path]

        // hard check for invalid data
        if (!data.variances || data.variances.length === 0) return

        // get top 5 variances of the same route
        const variances = data.variances.sort((a, b) => {
          return b.count - a.count
        }).slice(0, 5)

        // create a copy without reference to stored one
        let routeCopy = {
          path: path === '/' ? '/' : '/' + path,
          meta: fclone({
            min: data.meta.histogram.getMin(),
            max: data.meta.histogram.getMax(),
            count: data.meta.histogram.getCount(),
            meter: Math.round(data.meta.meter.rate(1000) * 100) / 100,
            median: data.meta.histogram.percentiles([0.5])[0.5],
            p95: data.meta.histogram.percentiles([0.95])[0.95]
          }),
          variances: []
        }

        variances.forEach((variance) => {
          // hard check for invalid data
          if (!variance.spans || variance.spans.length === 0) return

          // deep copy of variances data
          let tmp = fclone({
            spans: [],
            count: variance.histogram.getCount(),
            min: variance.histogram.getMin(),
            max: variance.histogram.getMax(),
            median: variance.histogram.percentiles([0.5])[0.5],
            p95: variance.histogram.percentiles([0.95])[0.95]
          })

          // get data for each span
          variance.spans.forEach((span) => {
            tmp.spans.push(fclone({
              name: span.name,
              labels: span.labels,
              kind: span.kind,
              startTime: span.startTime,
              min: span.histogram.getMin(),
              max: span.histogram.getMax(),
              median: span.histogram.percentiles([0.5])[0.5]
            }))
          })
          // push serialized into normalized data
          routeCopy.variances.push(tmp)
        })
        // push the route into normalized data
        normalized[appName].data.routes.push(routeCopy)
      })
    })

    return normalized
  }

  /**
   * Check if the string can be a id of some sort
   *
   * @param {String} id
   */
  isIdentifier (id) {
    id = typeof (id) !== 'string' ? id + '' : id

    // uuid v1/v4 with/without dash
    if (id.match(/[0-9a-f]{8}-[0-9a-f]{4}-[14][0-9a-f]{3}-[0-9a-f]{4}-[0-9a-f]{12}|[0-9a-f]{12}[14][0-9a-f]{19}/i)) {
      return true
    // if number
    } else if (id.match(/\d+/)) {
      return true
    // if suit of nbr/letters
    } else if (id.match(/[0-9]+[a-z]+|[a-z]+[0-9]+/)) {
      return true
    // if match pattern with multiple char spaced by . - _ @
    } else if (id.match(/((?:[0-9a-zA-Z]+[@\-_.][0-9a-zA-Z]+|[0-9a-zA-Z]+[@\-_.]|[@\-_.][0-9a-zA-Z]+)+)/)) {
      return true
    }
    return false
  }

  /**
   * Cleanup trace data
   * - delete result(s)
   * - replace labels value with a question mark
   *
   * @param {Object} spans list of span for a trace
   */
  censorSpans (spans) {
    if (!spans) return log('spans is null')
    if (cst.DEBUG) return

    spans.forEach((span) => {
      if (!span.labels) return

      delete span.labels.results
      delete span.labels.result
      delete span.spanId
      delete span.parentSpanId
      delete span.labels.values

      Object.keys(span.labels).forEach((key) => {
        if (typeof (span.labels[key]) === 'string' && key !== 'stacktrace') {
          span.labels[key] = span.labels[key].replace(this.REGEX_JSON_CLEANUP, '\": \"?\"') // eslint-disable-line
        }
      })
    })
  }

  /**
   * Parse stackrace of spans to extract and normalize data
   *
   * @param {Object} spans list of span for a trace
   */
  parseStacktrace (spans) {
    if (!spans) return log('spans is null')

    spans.forEach((span) => {
      // if empty make sure that it doesnt exist
      if (!span ||
          !span.labels ||
          !span.labels.stacktrace ||
          typeof (span.labels.stacktrace) !== 'string') return

      // you never know what come through that door
      try {
        span.labels.stacktrace = JSON.parse(span.labels.stacktrace)
      } catch (e) {
        return
      }

      if (!span.labels.stacktrace || !(span.labels.stacktrace.stack_frame instanceof Array)) return
      // parse the stacktrace
      let result = this.stackParser.parse(span.labels.stacktrace.stack_frame)
      if (result) {
        span.labels['source/file'] = result.callsite || undefined
        span.labels['source/context'] = result.context || undefined
      }
    })

    spans.forEach((span) => {
      if (!span || !span.labels) return
      delete span.labels.stacktrace
    })
  }
}

Anon7 - 2022
AnonSec Team