github.com/aquanetwork/aquachain@v1.7.8/common/metrics/influxdb/influxdb.go (about) 1 package influxdb 2 3 import ( 4 "fmt" 5 "log" 6 uurl "net/url" 7 "time" 8 9 "github.com/influxdata/influxdb/client" 10 "gitlab.com/aquachain/aquachain/common/metrics" 11 ) 12 13 type reporter struct { 14 reg metrics.Registry 15 interval time.Duration 16 17 url uurl.URL 18 database string 19 username string 20 password string 21 namespace string 22 tags map[string]string 23 24 client *client.Client 25 26 cache map[string]int64 27 } 28 29 // InfluxDB starts a InfluxDB reporter which will post the from the given metrics.Registry at each d interval. 30 func InfluxDB(r metrics.Registry, d time.Duration, url, database, username, password, namespace string) { 31 InfluxDBWithTags(r, d, url, database, username, password, namespace, nil) 32 } 33 34 // InfluxDBWithTags starts a InfluxDB reporter which will post the from the given metrics.Registry at each d interval with the specified tags 35 func InfluxDBWithTags(r metrics.Registry, d time.Duration, url, database, username, password, namespace string, tags map[string]string) { 36 u, err := uurl.Parse(url) 37 if err != nil { 38 log.Printf("unable to parse InfluxDB url %s. err=%v", url, err) 39 return 40 } 41 42 rep := &reporter{ 43 reg: r, 44 interval: d, 45 url: *u, 46 database: database, 47 username: username, 48 password: password, 49 namespace: namespace, 50 tags: tags, 51 cache: make(map[string]int64), 52 } 53 if err := rep.makeClient(); err != nil { 54 log.Printf("unable to make InfluxDB client. err=%v", err) 55 return 56 } 57 58 rep.run() 59 } 60 61 func (r *reporter) makeClient() (err error) { 62 r.client, err = client.NewClient(client.Config{ 63 URL: r.url, 64 Username: r.username, 65 Password: r.password, 66 }) 67 68 return 69 } 70 71 func (r *reporter) run() { 72 intervalTicker := time.Tick(r.interval) 73 pingTicker := time.Tick(time.Second * 5) 74 75 for { 76 select { 77 case <-intervalTicker: 78 if err := r.send(); err != nil { 79 log.Printf("unable to send to InfluxDB. err=%v", err) 80 } 81 case <-pingTicker: 82 _, _, err := r.client.Ping() 83 if err != nil { 84 log.Printf("got error while sending a ping to InfluxDB, trying to recreate client. err=%v", err) 85 86 if err = r.makeClient(); err != nil { 87 log.Printf("unable to make InfluxDB client. err=%v", err) 88 } 89 } 90 } 91 } 92 } 93 94 func (r *reporter) send() error { 95 var pts []client.Point 96 97 r.reg.Each(func(name string, i interface{}) { 98 now := time.Now() 99 namespace := r.namespace 100 101 switch metric := i.(type) { 102 case metrics.Counter: 103 v := metric.Count() 104 l := r.cache[name] 105 pts = append(pts, client.Point{ 106 Measurement: fmt.Sprintf("%s%s.count", namespace, name), 107 Tags: r.tags, 108 Fields: map[string]interface{}{ 109 "value": v - l, 110 }, 111 Time: now, 112 }) 113 r.cache[name] = v 114 case metrics.Gauge: 115 ms := metric.Snapshot() 116 pts = append(pts, client.Point{ 117 Measurement: fmt.Sprintf("%s%s.gauge", namespace, name), 118 Tags: r.tags, 119 Fields: map[string]interface{}{ 120 "value": ms.Value(), 121 }, 122 Time: now, 123 }) 124 case metrics.GaugeFloat64: 125 ms := metric.Snapshot() 126 pts = append(pts, client.Point{ 127 Measurement: fmt.Sprintf("%s%s.gauge", namespace, name), 128 Tags: r.tags, 129 Fields: map[string]interface{}{ 130 "value": ms.Value(), 131 }, 132 Time: now, 133 }) 134 case metrics.Histogram: 135 ms := metric.Snapshot() 136 ps := ms.Percentiles([]float64{0.5, 0.75, 0.95, 0.99, 0.999, 0.9999}) 137 pts = append(pts, client.Point{ 138 Measurement: fmt.Sprintf("%s%s.histogram", namespace, name), 139 Tags: r.tags, 140 Fields: map[string]interface{}{ 141 "count": ms.Count(), 142 "max": ms.Max(), 143 "mean": ms.Mean(), 144 "min": ms.Min(), 145 "stddev": ms.StdDev(), 146 "variance": ms.Variance(), 147 "p50": ps[0], 148 "p75": ps[1], 149 "p95": ps[2], 150 "p99": ps[3], 151 "p999": ps[4], 152 "p9999": ps[5], 153 }, 154 Time: now, 155 }) 156 case metrics.Meter: 157 ms := metric.Snapshot() 158 pts = append(pts, client.Point{ 159 Measurement: fmt.Sprintf("%s%s.meter", namespace, name), 160 Tags: r.tags, 161 Fields: map[string]interface{}{ 162 "count": ms.Count(), 163 "m1": ms.Rate1(), 164 "m5": ms.Rate5(), 165 "m15": ms.Rate15(), 166 "mean": ms.RateMean(), 167 }, 168 Time: now, 169 }) 170 case metrics.Timer: 171 ms := metric.Snapshot() 172 ps := ms.Percentiles([]float64{0.5, 0.75, 0.95, 0.99, 0.999, 0.9999}) 173 pts = append(pts, client.Point{ 174 Measurement: fmt.Sprintf("%s%s.timer", namespace, name), 175 Tags: r.tags, 176 Fields: map[string]interface{}{ 177 "count": ms.Count(), 178 "max": ms.Max(), 179 "mean": ms.Mean(), 180 "min": ms.Min(), 181 "stddev": ms.StdDev(), 182 "variance": ms.Variance(), 183 "p50": ps[0], 184 "p75": ps[1], 185 "p95": ps[2], 186 "p99": ps[3], 187 "p999": ps[4], 188 "p9999": ps[5], 189 "m1": ms.Rate1(), 190 "m5": ms.Rate5(), 191 "m15": ms.Rate15(), 192 "meanrate": ms.RateMean(), 193 }, 194 Time: now, 195 }) 196 case metrics.ResettingTimer: 197 t := metric.Snapshot() 198 199 if len(t.Values()) > 0 { 200 ps := t.Percentiles([]float64{50, 95, 99}) 201 val := t.Values() 202 pts = append(pts, client.Point{ 203 Measurement: fmt.Sprintf("%s%s.span", namespace, name), 204 Tags: r.tags, 205 Fields: map[string]interface{}{ 206 "count": len(val), 207 "max": val[len(val)-1], 208 "mean": t.Mean(), 209 "min": val[0], 210 "p50": ps[0], 211 "p95": ps[1], 212 "p99": ps[2], 213 }, 214 Time: now, 215 }) 216 } 217 } 218 }) 219 220 bps := client.BatchPoints{ 221 Points: pts, 222 Database: r.database, 223 } 224 225 _, err := r.client.Write(bps) 226 return err 227 }