1=======================
2Display Core Next (DCN)
3=======================
4
5To equip our readers with the basic knowledge of how AMD Display Core Next
6(DCN) works, we need to start with an overview of the hardware pipeline. Below
7you can see a picture that provides a DCN overview, keep in mind that this is a
8generic diagram, and we have variations per ASIC.
9
10.. kernel-figure:: dc_pipeline_overview.svg
11
12Based on this diagram, we can pass through each block and briefly describe
13them:
14
15* **Display Controller Hub (DCHUB)**: This is the gateway between the Scalable
16  Data Port (SDP) and DCN. This component has multiple features, such as memory
17  arbitration, rotation, and cursor manipulation.
18
19* **Display Pipe and Plane (DPP)**: This block provides pre-blend pixel
20  processing such as color space conversion, linearization of pixel data, tone
21  mapping, and gamut mapping.
22
23* **Multiple Pipe/Plane Combined (MPC)**: This component performs blending of
24  multiple planes, using global or per-pixel alpha.
25
26* **Output Pixel Processing (OPP)**: Process and format pixels to be sent to
27  the display.
28
29* **Output Pipe Timing Combiner (OPTC)**: It generates time output to combine
30  streams or divide capabilities. CRC values are generated in this block.
31
32* **Display Output (DIO)**: Codify the output to the display connected to our
33  GPU.
34
35* **Display Writeback (DWB)**: It provides the ability to write the output of
36  the display pipe back to memory as video frames.
37
38* **Multi-Media HUB (MMHUBBUB)**: Memory controller interface for DMCUB and DWB
39  (Note that DWB is not hooked yet).
40
41* **DCN Management Unit (DMU)**: It provides registers with access control and
42  interrupts the controller to the SOC host interrupt unit. This block includes
43  the Display Micro-Controller Unit - version B (DMCUB), which is handled via
44  firmware.
45
46* **DCN Clock Generator Block (DCCG)**: It provides the clocks and resets
47  for all of the display controller clock domains.
48
49* **Azalia (AZ)**: Audio engine.
50
51The above diagram is an architecture generalization of DCN, which means that
52every ASIC has variations around this base model. Notice that the display
53pipeline is connected to the Scalable Data Port (SDP) via DCHUB; you can see
54the SDP as the element from our Data Fabric that feeds the display pipe.
55
56Always approach the DCN architecture as something flexible that can be
57configured and reconfigured in multiple ways; in other words, each block can be
58setup or ignored accordingly with userspace demands. For example, if we
59want to drive an 8k@60Hz with a DSC enabled, our DCN may require 4 DPP and 2
60OPP. It is DC's responsibility to drive the best configuration for each
61specific scenario. Orchestrate all of these components together requires a
62sophisticated communication interface which is highlighted in the diagram by
63the edges that connect each block; from the chart, each connection between
64these blocks represents:
65
661. Pixel data interface (red): Represents the pixel data flow;
672. Global sync signals (green): It is a set of synchronization signals composed
68   by VStartup, VUpdate, and VReady;
693. Config interface: Responsible to configure blocks;
704. Sideband signals: All other signals that do not fit the previous one.
71
72These signals are essential and play an important role in DCN. Nevertheless,
73the Global Sync deserves an extra level of detail described in the next
74section.
75
76All of these components are represented by a data structure named dc_state.
77From DCHUB to MPC, we have a representation called dc_plane; from MPC to OPTC,
78we have dc_stream, and the output (DIO) is handled by dc_link. Keep in mind
79that HUBP accesses a surface using a specific format read from memory, and our
80dc_plane should work to convert all pixels in the plane to something that can
81be sent to the display via dc_stream and dc_link.
82
83Front End and Back End
84----------------------
85
86Display pipeline can be broken down into two components that are usually
87referred as **Front End (FE)** and **Back End (BE)**, where FE consists of:
88
89* DCHUB (Mainly referring to a subcomponent named HUBP)
90* DPP
91* MPC
92
93On the other hand, BE consist of
94
95* OPP
96* OPTC
97* DIO (DP/HDMI stream encoder and link encoder)
98
99OPP and OPTC are two joining blocks between FE and BE. On a side note, this is
100a one-to-one mapping of the link encoder to PHY, but we can configure the DCN
101to choose which link encoder to connect to which PHY. FE's main responsibility
102is to change, blend and compose pixel data, while BE's job is to frame a
103generic pixel stream to a specific display's pixel stream.
104
105Data Flow
106---------
107
108Initially, data is passed in from VRAM through Data Fabric (DF) in native pixel
109formats. Such data format stays through till HUBP in DCHUB, where HUBP unpacks
110different pixel formats and outputs them to DPP in uniform streams through 4
111channels (1 for alpha + 3 for colors).
112
113The Converter and Cursor (CNVC) in DPP would then normalize the data
114representation and convert them to a DCN specific floating-point format (i.e.,
115different from the IEEE floating-point format). In the process, CNVC also
116applies a degamma function to transform the data from non-linear to linear
117space to relax the floating-point calculations following. Data would stay in
118this floating-point format from DPP to OPP.
119
120Starting OPP, because color transformation and blending have been completed
121(i.e alpha can be dropped), and the end sinks do not require the precision and
122dynamic range that floating points provide (i.e. all displays are in integer
123depth format), bit-depth reduction/dithering would kick in. In OPP, we would
124also apply a regamma function to introduce the gamma removed earlier back.
125Eventually, we output data in integer format at DIO.
126
127AMD Hardware Pipeline
128---------------------
129
130When discussing graphics on Linux, the **pipeline** term can sometimes be
131overloaded with multiple meanings, so it is important to define what we mean
132when we say **pipeline**. In the DCN driver, we use the term **hardware
133pipeline** or **pipeline** or just **pipe** as an abstraction to indicate a
134sequence of DCN blocks instantiated to address some specific configuration. DC
135core treats DCN blocks as individual resources, meaning we can build a pipeline
136by taking resources for all individual hardware blocks to compose one pipeline.
137In actuality, we can't connect an arbitrary block from one pipe to a block from
138another pipe; they are routed linearly, except for DSC, which can be
139arbitrarily assigned as needed. We have this pipeline concept for trying to
140optimize bandwidth utilization.
141
142.. kernel-figure:: pipeline_4k_no_split.svg
143
144Additionally, let's take a look at parts of the DTN log (see
145'Documentation/gpu/amdgpu/display/dc-debug.rst' for more information) since
146this log can help us to see part of this pipeline behavior in real-time::
147
148 HUBP:  format  addr_hi  width  height ...
149 [ 0]:      8h      81h   3840    2160
150 [ 1]:      0h       0h      0       0
151 [ 2]:      0h       0h      0       0
152 [ 3]:      0h       0h      0       0
153 [ 4]:      0h       0h      0       0
154 ...
155 MPCC:  OPP  DPP ...
156 [ 0]:   0h   0h ...
157
158The first thing to notice from the diagram and DTN log it is the fact that we
159have different clock domains for each part of the DCN blocks. In this example,
160we have just a single **pipeline** where the data flows from DCHUB to DIO, as
161we intuitively expect. Nonetheless, DCN is flexible, as mentioned before, and
162we can split this single pipe differently, as described in the below diagram:
163
164.. kernel-figure:: pipeline_4k_split.svg
165
166Now, if we inspect the DTN log again we can see some interesting changes::
167
168 HUBP:  format  addr_hi  width  height ...
169 [ 0]:      8h      81h   1920    2160 ...
170 ...
171 [ 4]:      0h       0h      0       0 ...
172 [ 5]:      8h      81h   1920    2160 ...
173 ...
174 MPCC:  OPP  DPP ...
175 [ 0]:   0h   0h ...
176 [ 5]:   0h   5h ...
177
178From the above example, we now split the display pipeline into two vertical
179parts of 1920x2160 (i.e., 3440x2160), and as a result, we could reduce the
180clock frequency in the DPP part. This is not only useful for saving power but
181also to better handle the required throughput. The idea to keep in mind here is
182that the pipe configuration can vary a lot according to the display
183configuration, and it is the DML's responsibility to set up all required
184configuration parameters for multiple scenarios supported by our hardware.
185
186Global Sync
187-----------
188
189Many DCN registers are double buffered, most importantly the surface address.
190This allows us to update DCN hardware atomically for page flips, as well as
191for most other updates that don't require enabling or disabling of new pipes.
192
193(Note: There are many scenarios when DC will decide to reserve extra pipes
194in order to support outputs that need a very high pixel clock, or for
195power saving purposes.)
196
197These atomic register updates are driven by global sync signals in DCN. In
198order to understand how atomic updates interact with DCN hardware, and how DCN
199signals page flip and vblank events it is helpful to understand how global sync
200is programmed.
201
202Global sync consists of three signals, VSTARTUP, VUPDATE, and VREADY. These are
203calculated by the Display Mode Library - DML (drivers/gpu/drm/amd/display/dc/dml)
204based on a large number of parameters and ensure our hardware is able to feed
205the DCN pipeline without underflows or hangs in any given system configuration.
206The global sync signals always happen during VBlank, are independent from the
207VSync signal, and do not overlap each other.
208
209VUPDATE is the only signal that is of interest to the rest of the driver stack
210or userspace clients as it signals the point at which hardware latches to
211atomically programmed (i.e. double buffered) registers. Even though it is
212independent of the VSync signal we use VUPDATE to signal the VSync event as it
213provides the best indication of how atomic commits and hardware interact.
214
215Since DCN hardware is double-buffered the DC driver is able to program the
216hardware at any point during the frame.
217
218The below picture illustrates the global sync signals:
219
220.. kernel-figure:: global_sync_vblank.svg
221
222These signals affect core DCN behavior. Programming them incorrectly will lead
223to a number of negative consequences, most of them quite catastrophic.
224
225The following picture shows how global sync allows for a mailbox style of
226updates, i.e. it allows for multiple re-configurations between VUpdate
227events where only the last configuration programmed before the VUpdate signal
228becomes effective.
229
230.. kernel-figure:: config_example.svg
231