xref: /openbmc/qemu/tests/migration/guestperf/plot.py (revision 0ed93f4c)
1#
2# Migration test graph plotting
3#
4# Copyright (c) 2016 Red Hat, Inc.
5#
6# This library is free software; you can redistribute it and/or
7# modify it under the terms of the GNU Lesser General Public
8# License as published by the Free Software Foundation; either
9# version 2 of the License, or (at your option) any later version.
10#
11# This library is distributed in the hope that it will be useful,
12# but WITHOUT ANY WARRANTY; without even the implied warranty of
13# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
14# Lesser General Public License for more details.
15#
16# You should have received a copy of the GNU Lesser General Public
17# License along with this library; if not, see <http://www.gnu.org/licenses/>.
18#
19
20import sys
21
22
23class Plot(object):
24
25    # Generated using
26    # http://tools.medialab.sciences-po.fr/iwanthue/
27    COLORS = ["#CD54D0",
28              "#79D94C",
29              "#7470CD",
30              "#D2D251",
31              "#863D79",
32              "#76DDA6",
33              "#D4467B",
34              "#61923D",
35              "#CB9CCA",
36              "#D98F36",
37              "#8CC8DA",
38              "#CE4831",
39              "#5E7693",
40              "#9B803F",
41              "#412F4C",
42              "#CECBA6",
43              "#6D3229",
44              "#598B73",
45              "#C8827C",
46              "#394427"]
47
48    def __init__(self,
49                 reports,
50                 migration_iters,
51                 total_guest_cpu,
52                 split_guest_cpu,
53                 qemu_cpu,
54                 vcpu_cpu):
55
56        self._reports = reports
57        self._migration_iters = migration_iters
58        self._total_guest_cpu = total_guest_cpu
59        self._split_guest_cpu = split_guest_cpu
60        self._qemu_cpu = qemu_cpu
61        self._vcpu_cpu = vcpu_cpu
62        self._color_idx = 0
63
64    def _next_color(self):
65        color = self.COLORS[self._color_idx]
66        self._color_idx += 1
67        if self._color_idx >= len(self.COLORS):
68            self._color_idx = 0
69        return color
70
71    def _get_progress_label(self, progress):
72        if progress:
73            return "\n\n" + "\n".join(
74                ["Status: %s" % progress._status,
75                 "Iteration: %d" % progress._ram._iterations,
76                 "Throttle: %02d%%" % progress._throttle_pcent,
77                 "Dirty rate: %dMB/s" % (progress._ram._dirty_rate_pps * 4 / 1024.0)])
78        else:
79            return "\n\n" + "\n".join(
80                ["Status: %s" % "none",
81                 "Iteration: %d" % 0])
82
83    def _find_start_time(self, report):
84        startqemu = report._qemu_timings._records[0]._timestamp
85        startguest = report._guest_timings._records[0]._timestamp
86        if startqemu < startguest:
87            return startqemu
88        else:
89            return stasrtguest
90
91    def _get_guest_max_value(self, report):
92        maxvalue = 0
93        for record in report._guest_timings._records:
94            if record._value > maxvalue:
95                maxvalue = record._value
96        return maxvalue
97
98    def _get_qemu_max_value(self, report):
99        maxvalue = 0
100        oldvalue = None
101        oldtime = None
102        for record in report._qemu_timings._records:
103            if oldvalue is not None:
104                cpudelta = (record._value - oldvalue) / 1000.0
105                timedelta = record._timestamp - oldtime
106                if timedelta == 0:
107                    continue
108                util = cpudelta / timedelta * 100.0
109            else:
110                util = 0
111            oldvalue = record._value
112            oldtime = record._timestamp
113
114            if util > maxvalue:
115                maxvalue = util
116        return maxvalue
117
118    def _get_total_guest_cpu_graph(self, report, starttime):
119        xaxis = []
120        yaxis = []
121        labels = []
122        progress_idx = -1
123        for record in report._guest_timings._records:
124            while ((progress_idx + 1) < len(report._progress_history) and
125                   report._progress_history[progress_idx + 1]._now < record._timestamp):
126                progress_idx = progress_idx + 1
127
128            if progress_idx >= 0:
129                progress = report._progress_history[progress_idx]
130            else:
131                progress = None
132
133            xaxis.append(record._timestamp - starttime)
134            yaxis.append(record._value)
135            labels.append(self._get_progress_label(progress))
136
137        from plotly import graph_objs as go
138        return go.Scatter(x=xaxis,
139                          y=yaxis,
140                          name="Guest PIDs: %s" % report._scenario._name,
141                          mode='lines',
142                          line={
143                              "dash": "solid",
144                              "color": self._next_color(),
145                              "shape": "linear",
146                              "width": 1
147                          },
148                          text=labels)
149
150    def _get_split_guest_cpu_graphs(self, report, starttime):
151        threads = {}
152        for record in report._guest_timings._records:
153            if record._tid in threads:
154                continue
155            threads[record._tid] = {
156                "xaxis": [],
157                "yaxis": [],
158                "labels": [],
159            }
160
161        progress_idx = -1
162        for record in report._guest_timings._records:
163            while ((progress_idx + 1) < len(report._progress_history) and
164                   report._progress_history[progress_idx + 1]._now < record._timestamp):
165                progress_idx = progress_idx + 1
166
167            if progress_idx >= 0:
168                progress = report._progress_history[progress_idx]
169            else:
170                progress = None
171
172            threads[record._tid]["xaxis"].append(record._timestamp - starttime)
173            threads[record._tid]["yaxis"].append(record._value)
174            threads[record._tid]["labels"].append(self._get_progress_label(progress))
175
176
177        graphs = []
178        from plotly import graph_objs as go
179        for tid in threads.keys():
180            graphs.append(
181                go.Scatter(x=threads[tid]["xaxis"],
182                           y=threads[tid]["yaxis"],
183                           name="PID %s: %s" % (tid, report._scenario._name),
184                           mode="lines",
185                           line={
186                               "dash": "solid",
187                               "color": self._next_color(),
188                               "shape": "linear",
189                               "width": 1
190                           },
191                           text=threads[tid]["labels"]))
192        return graphs
193
194    def _get_migration_iters_graph(self, report, starttime):
195        xaxis = []
196        yaxis = []
197        labels = []
198        for progress in report._progress_history:
199            xaxis.append(progress._now - starttime)
200            yaxis.append(0)
201            labels.append(self._get_progress_label(progress))
202
203        from plotly import graph_objs as go
204        return go.Scatter(x=xaxis,
205                          y=yaxis,
206                          text=labels,
207                          name="Migration iterations",
208                          mode="markers",
209                          marker={
210                              "color": self._next_color(),
211                              "symbol": "star",
212                              "size": 5
213                          })
214
215    def _get_qemu_cpu_graph(self, report, starttime):
216        xaxis = []
217        yaxis = []
218        labels = []
219        progress_idx = -1
220
221        first = report._qemu_timings._records[0]
222        abstimestamps = [first._timestamp]
223        absvalues = [first._value]
224
225        for record in report._qemu_timings._records[1:]:
226            while ((progress_idx + 1) < len(report._progress_history) and
227                   report._progress_history[progress_idx + 1]._now < record._timestamp):
228                progress_idx = progress_idx + 1
229
230            if progress_idx >= 0:
231                progress = report._progress_history[progress_idx]
232            else:
233                progress = None
234
235            oldvalue = absvalues[-1]
236            oldtime = abstimestamps[-1]
237
238            cpudelta = (record._value - oldvalue) / 1000.0
239            timedelta = record._timestamp - oldtime
240            if timedelta == 0:
241                continue
242            util = cpudelta / timedelta * 100.0
243
244            abstimestamps.append(record._timestamp)
245            absvalues.append(record._value)
246
247            xaxis.append(record._timestamp - starttime)
248            yaxis.append(util)
249            labels.append(self._get_progress_label(progress))
250
251        from plotly import graph_objs as go
252        return go.Scatter(x=xaxis,
253                          y=yaxis,
254                          yaxis="y2",
255                          name="QEMU: %s" % report._scenario._name,
256                          mode='lines',
257                          line={
258                              "dash": "solid",
259                              "color": self._next_color(),
260                              "shape": "linear",
261                              "width": 1
262                          },
263                          text=labels)
264
265    def _get_vcpu_cpu_graphs(self, report, starttime):
266        threads = {}
267        for record in report._vcpu_timings._records:
268            if record._tid in threads:
269                continue
270            threads[record._tid] = {
271                "xaxis": [],
272                "yaxis": [],
273                "labels": [],
274                "absvalue": [record._value],
275                "abstime": [record._timestamp],
276            }
277
278        progress_idx = -1
279        for record in report._vcpu_timings._records:
280            while ((progress_idx + 1) < len(report._progress_history) and
281                   report._progress_history[progress_idx + 1]._now < record._timestamp):
282                progress_idx = progress_idx + 1
283
284            if progress_idx >= 0:
285                progress = report._progress_history[progress_idx]
286            else:
287                progress = None
288
289            oldvalue = threads[record._tid]["absvalue"][-1]
290            oldtime = threads[record._tid]["abstime"][-1]
291
292            cpudelta = (record._value - oldvalue) / 1000.0
293            timedelta = record._timestamp - oldtime
294            if timedelta == 0:
295                continue
296            util = cpudelta / timedelta * 100.0
297            if util > 100:
298                util = 100
299
300            threads[record._tid]["absvalue"].append(record._value)
301            threads[record._tid]["abstime"].append(record._timestamp)
302
303            threads[record._tid]["xaxis"].append(record._timestamp - starttime)
304            threads[record._tid]["yaxis"].append(util)
305            threads[record._tid]["labels"].append(self._get_progress_label(progress))
306
307
308        graphs = []
309        from plotly import graph_objs as go
310        for tid in threads.keys():
311            graphs.append(
312                go.Scatter(x=threads[tid]["xaxis"],
313                           y=threads[tid]["yaxis"],
314                           yaxis="y2",
315                           name="VCPU %s: %s" % (tid, report._scenario._name),
316                           mode="lines",
317                           line={
318                               "dash": "solid",
319                               "color": self._next_color(),
320                               "shape": "linear",
321                               "width": 1
322                           },
323                           text=threads[tid]["labels"]))
324        return graphs
325
326    def _generate_chart_report(self, report):
327        graphs = []
328        starttime = self._find_start_time(report)
329        if self._total_guest_cpu:
330            graphs.append(self._get_total_guest_cpu_graph(report, starttime))
331        if self._split_guest_cpu:
332            graphs.extend(self._get_split_guest_cpu_graphs(report, starttime))
333        if self._qemu_cpu:
334            graphs.append(self._get_qemu_cpu_graph(report, starttime))
335        if self._vcpu_cpu:
336            graphs.extend(self._get_vcpu_cpu_graphs(report, starttime))
337        if self._migration_iters:
338            graphs.append(self._get_migration_iters_graph(report, starttime))
339        return graphs
340
341    def _generate_annotation(self, starttime, progress):
342        return {
343            "text": progress._status,
344            "x": progress._now - starttime,
345            "y": 10,
346        }
347
348    def _generate_annotations(self, report):
349        starttime = self._find_start_time(report)
350        annotations = {}
351        started = False
352        for progress in report._progress_history:
353            if progress._status == "setup":
354                continue
355            if progress._status not in annotations:
356                annotations[progress._status] = self._generate_annotation(starttime, progress)
357
358        return annotations.values()
359
360    def _generate_chart(self):
361        from plotly.offline import plot
362        from plotly import graph_objs as go
363
364        graphs = []
365        yaxismax = 0
366        yaxismax2 = 0
367        for report in self._reports:
368            graphs.extend(self._generate_chart_report(report))
369
370            maxvalue = self._get_guest_max_value(report)
371            if maxvalue > yaxismax:
372                yaxismax = maxvalue
373
374            maxvalue = self._get_qemu_max_value(report)
375            if maxvalue > yaxismax2:
376                yaxismax2 = maxvalue
377
378        yaxismax += 100
379        if not self._qemu_cpu:
380            yaxismax2 = 110
381        yaxismax2 += 10
382
383        annotations = []
384        if self._migration_iters:
385            for report in self._reports:
386                annotations.extend(self._generate_annotations(report))
387
388        layout = go.Layout(title="Migration comparison",
389                           xaxis={
390                               "title": "Wallclock time (secs)",
391                               "showgrid": False,
392                           },
393                           yaxis={
394                               "title": "Memory update speed (ms/GB)",
395                               "showgrid": False,
396                               "range": [0, yaxismax],
397                           },
398                           yaxis2={
399                               "title": "Hostutilization (%)",
400                               "overlaying": "y",
401                               "side": "right",
402                               "range": [0, yaxismax2],
403                               "showgrid": False,
404                           },
405                           annotations=annotations)
406
407        figure = go.Figure(data=graphs, layout=layout)
408
409        return plot(figure,
410                    show_link=False,
411                    include_plotlyjs=False,
412                    output_type="div")
413
414
415    def _generate_report(self):
416        pieces = []
417        for report in self._reports:
418            pieces.append("""
419<h3>Report %s</h3>
420<table>
421""" % report._scenario._name)
422
423            pieces.append("""
424  <tr class="subhead">
425    <th colspan="2">Test config</th>
426  </tr>
427  <tr>
428    <th>Emulator:</th>
429    <td>%s</td>
430  </tr>
431  <tr>
432    <th>Kernel:</th>
433    <td>%s</td>
434  </tr>
435  <tr>
436    <th>Ramdisk:</th>
437    <td>%s</td>
438  </tr>
439  <tr>
440    <th>Transport:</th>
441    <td>%s</td>
442  </tr>
443  <tr>
444    <th>Host:</th>
445    <td>%s</td>
446  </tr>
447""" % (report._binary, report._kernel,
448       report._initrd, report._transport, report._dst_host))
449
450            hardware = report._hardware
451            pieces.append("""
452  <tr class="subhead">
453    <th colspan="2">Hardware config</th>
454  </tr>
455  <tr>
456    <th>CPUs:</th>
457    <td>%d</td>
458  </tr>
459  <tr>
460    <th>RAM:</th>
461    <td>%d GB</td>
462  </tr>
463  <tr>
464    <th>Source CPU bind:</th>
465    <td>%s</td>
466  </tr>
467  <tr>
468    <th>Source RAM bind:</th>
469    <td>%s</td>
470  </tr>
471  <tr>
472    <th>Dest CPU bind:</th>
473    <td>%s</td>
474  </tr>
475  <tr>
476    <th>Dest RAM bind:</th>
477    <td>%s</td>
478  </tr>
479  <tr>
480    <th>Preallocate RAM:</th>
481    <td>%s</td>
482  </tr>
483  <tr>
484    <th>Locked RAM:</th>
485    <td>%s</td>
486  </tr>
487  <tr>
488    <th>Huge pages:</th>
489    <td>%s</td>
490  </tr>
491""" % (hardware._cpus, hardware._mem,
492       ",".join(hardware._src_cpu_bind),
493       ",".join(hardware._src_mem_bind),
494       ",".join(hardware._dst_cpu_bind),
495       ",".join(hardware._dst_mem_bind),
496       "yes" if hardware._prealloc_pages else "no",
497       "yes" if hardware._locked_pages else "no",
498       "yes" if hardware._huge_pages else "no"))
499
500            scenario = report._scenario
501            pieces.append("""
502  <tr class="subhead">
503    <th colspan="2">Scenario config</th>
504  </tr>
505  <tr>
506    <th>Max downtime:</th>
507    <td>%d milli-sec</td>
508  </tr>
509  <tr>
510    <th>Max bandwidth:</th>
511    <td>%d MB/sec</td>
512  </tr>
513  <tr>
514    <th>Max iters:</th>
515    <td>%d</td>
516  </tr>
517  <tr>
518    <th>Max time:</th>
519    <td>%d secs</td>
520  </tr>
521  <tr>
522    <th>Pause:</th>
523    <td>%s</td>
524  </tr>
525  <tr>
526    <th>Pause iters:</th>
527    <td>%d</td>
528  </tr>
529  <tr>
530    <th>Post-copy:</th>
531    <td>%s</td>
532  </tr>
533  <tr>
534    <th>Post-copy iters:</th>
535    <td>%d</td>
536  </tr>
537  <tr>
538    <th>Auto-converge:</th>
539    <td>%s</td>
540  </tr>
541  <tr>
542    <th>Auto-converge iters:</th>
543    <td>%d</td>
544  </tr>
545  <tr>
546    <th>MT compression:</th>
547    <td>%s</td>
548  </tr>
549  <tr>
550    <th>MT compression threads:</th>
551    <td>%d</td>
552  </tr>
553  <tr>
554    <th>XBZRLE compression:</th>
555    <td>%s</td>
556  </tr>
557  <tr>
558    <th>XBZRLE compression cache:</th>
559    <td>%d%% of RAM</td>
560  </tr>
561""" % (scenario._downtime, scenario._bandwidth,
562       scenario._max_iters, scenario._max_time,
563       "yes" if scenario._pause else "no", scenario._pause_iters,
564       "yes" if scenario._post_copy else "no", scenario._post_copy_iters,
565       "yes" if scenario._auto_converge else "no", scenario._auto_converge_step,
566       "yes" if scenario._compression_mt else "no", scenario._compression_mt_threads,
567       "yes" if scenario._compression_xbzrle else "no", scenario._compression_xbzrle_cache))
568
569            pieces.append("""
570</table>
571""")
572
573        return "\n".join(pieces)
574
575    def _generate_style(self):
576        return """
577#report table tr th {
578    text-align: right;
579}
580#report table tr td {
581    text-align: left;
582}
583#report table tr.subhead th {
584    background: rgb(192, 192, 192);
585    text-align: center;
586}
587
588"""
589
590    def generate_html(self, fh):
591        print("""<html>
592  <head>
593    <script type="text/javascript" src="plotly.min.js">
594    </script>
595    <style type="text/css">
596%s
597    </style>
598    <title>Migration report</title>
599  </head>
600  <body>
601    <h1>Migration report</h1>
602    <h2>Chart summary</h2>
603    <div id="chart">
604""" % self._generate_style(), file=fh)
605        print(self._generate_chart(), file=fh)
606        print("""
607    </div>
608    <h2>Report details</h2>
609    <div id="report">
610""", file=fh)
611        print(self._generate_report(), file=fh)
612        print("""
613    </div>
614  </body>
615</html>
616""", file=fh)
617
618    def generate(self, filename):
619        if filename is None:
620            self.generate_html(sys.stdout)
621        else:
622            with open(filename, "w") as fh:
623                self.generate_html(fh)
624