开心的滑板 · idea clean后报 错误: ...· 3 天前 · |
急躁的甜瓜 · bootstrap jquery 拖动排序 ...· 4 天前 · |
快乐的大蒜 · 3. Nsight Compute — ...· 5 天前 · |
大气的木耳 · Querydsl 参考指南 - 南石 - 博客园· 6 天前 · |
帅气的投影仪 · [email protected] ...· 1 周前 · |
儒雅的手电筒 · 什么是:局部异常因子 (LOF) - ...· 2 周前 · |
小眼睛的刺猬 · PlaceholderImageFactor ...· 1 月前 · |
细心的毛巾 · Windows中从Git Bash运行jq· 1 月前 · |
不羁的槟榔 · python tkinter ...· 2 月前 · |
霸气的花卷 · 未能登录 · Issue #115 · ...· 1 年前 · |
For users migrating from Visual Profiler to NVIDIA Nsight Compute, please see the Visual Profiler Transition Guide for comparison of features and workflows.
This document is a user guide to the next-generation NVIDIA Nsight Compute profiling tools. NVIDIA Nsight Compute is an interactive kernel profiler for CUDA applications. It provides detailed performance metrics and API debugging via a user interface and command line tool. In addition, its baseline feature allows users to compare results within the tool. NVIDIA Nsight Compute provides a customizable and data-driven user interface and metric collection and can be extended with analysis scripts for post-processing results.
Important Features
Interactive kernel profiler and API debugger
Graphical profile report
Result comparison across one or multiple reports within the tool
Fast Data Collection
UI and Command Line interface
Fully customizable reports and analysis rules
The following sections provide brief step-by-step guides of how to setup and run NVIDIA Nsight Compute to collect profile information. All directories are relative to the base directory of NVIDIA Nsight Compute, unless specified otherwise.
The UI executable is called ncu-ui. A shortcut with this name is located in the base directory of the NVIDIA Nsight Compute installation. The actual executable is located in the folder
host\windows-desktop-win7-x64
on Windows or
host/linux-desktop-glibc_2_11_3-x64
on Linux. By default, when installing from a Linux
.run
file, NVIDIA Nsight Compute is located in
/usr/local/cuda-<cuda-version>/nsight-compute-<version>
. When installing from a
.deb
or
.rpm
package, it is located in
/opt/nvidia/nsight-compute/<version>
to be consistent with
Nsight Systems
. In Windows, the default path is
C:\Program
Files\NVIDIA
Corporation\Nsight
Compute
<version>
.
After starting NVIDIA Nsight Compute, by default the Welcome Page is opened. The Start section allows the user to start a new activity, open an existing report, create a new project or load an existing project. The Continue section provides links to recently opened reports and projects. The Explore section provides information about what is new in the latest release, as well as links to additional training. See Environment on how to change the start-up action.
Launch the target application from NVIDIA Nsight Compute
When starting NVIDIA Nsight Compute, the
Welcome Page
will appear. Click on
Quick Launch
to open the
Connection
dialog. If the
Connection
dialog doesn’t appear, you can open it using the
Connect
button from the main toolbar, as long as you are not currently connected. Select your target platform on the left-hand side and your connection target (machine) from the
Connection
drop down. If you have your local target platform selected,
localhost
will become available as a connection. Use the + button to add a new connection target. Then, continue by filling in the details in the
Launch
tab. In the
Activity
panel, select the Interactive Profile activity to initiate a session that allows controlling the execution of the target application and selecting the kernels of interest interactively. Press
Launch
to start the session.
Launch the target application with tools instrumentation from the command line
The ncu can act as a simple wrapper that forces the target application to load the necessary libraries for tools instrumentation. The parameter
--mode=launch
specifies that the target application should be launched and suspended before the first instrumented API call. That way the application waits until we connect with the UI.
$ ncu --mode=launch CuVectorAddDrv.exe
Launch NVIDIA Nsight Compute and connect to target application
Select the target machine at the top of the dialog to connect and update the list of attachable applications. By default, localhost is pre-selected if the target matches your current local platform. Select the Attach tab and the target application of interest and press Attach. Once connected, the layout of NVIDIA Nsight Compute changes into stepping mode that allows you to control the execution of any calls into the instrumented API. When connected, the API Stream window indicates that the target application waits before the very first API call.
Control application execution
Use the API Stream window to step the calls into the instrumented API. The dropdown at the top allows switching between different CPU threads of the application. Step In (F11), Step Over (F10), and Step Out (Shift + F11) are available from the Debug menu or the corresponding toolbar buttons. While stepping, function return values and function parameters are captured.
Use Resume (F5) and Pause to allow the program to run freely. Freeze control is available to define the behavior of threads currently not in focus, i.e. selected in the thread drop down. By default, the API Stream stops on any API call that returns an error code. This can be toggled in the Debug menu by Break On API Error.
Isolate a kernel launch
To quickly isolate a kernel launch for profiling, use the Run to Next Kernel button in the toolbar of the API Stream window to jump to the next kernel launch. The execution will stop before the kernel launch is executed.
Profile a kernel launch
Once the execution of the target application is suspended at a kernel launch, additional actions become available in the UI. These actions are either available from the menu or from the toolbar. Please note that the actions are disabled, if the API stream is not at a qualifying state (not at a kernel launch or launching on an unsupported GPU). To profile, press Profile Kernel and wait until the result is shown in the Profiler Report. Profiling progress is reported in the lower right corner status bar.
Instead of manually selecting Profile, it is also possible to enable Auto Profile from the Profile menu. If enabled, each kernel matching the current kernel filter (if any) will be profiled using the current section configuration. This is especially useful if an application is to be profiled unattended, or the number of kernel launches to be profiled is very large. Sections can be enabled or disabled using the Metric Selection tool window.
Profile Series allows to configure the collection of a set of profile results at once. Each result in the set is profiled with varying parameters. Series are useful to investigate the behavior of a kernel across a large set of parameters without the need to recompile and rerun the application many times.
For a detailed description of the options available in this activity, see Interactive Profile Activity.
3.2.2. Non-Interactive Profile Activity
Launch the target application from NVIDIA Nsight Compute
When starting NVIDIA Nsight Compute, the Welcome Page will appear. Click on Start Activity to open the Start Activity dialog. If the Start Activity dialog doesn’t appear, you can open it using the Start Activity button from the main toolbar, as long as you are not currently connected. Select your target platform on the left-hand side and your localhost from the Connection drop down. Then, fill in the launch details. In the Activity panel, select the Profile activity to initiate a session that pre-configures the profile session and launches the command line profiler to collect the data. Provide the Output File name to enable starting the session with the Launch button.
Additional Launch Options
For more details on these options, see Command Line Options. The options are grouped into tabs: The Filter tab exposes the options to specify which kernels should be profiled. Options include the kernel regex filter, the number of launches to skip, and the total number of launches to profile. The Sections tab allows you to select which sections should be collected for each kernel launch. Hover over a section to see its description as a tool-tip. To change the sections that are enabled by default, use the Metric Selection tool window. The Sampling tab allows you to configure sampling options for each kernel launch. The Other tab includes the option to collect NVTX information or custom metrics via the --metrics
option.
For a detailed description of the options available in this activity, see Profile Activity.
3.2.3. System Trace Activity
Launch the target application from NVIDIA Nsight Compute
When starting NVIDIA Nsight Compute, the Welcome Page will appear. Click on Start Activity to open the Start Activity dialog. If the Start Activity dialog doesn’t appear, you can open it using the Start Activity button from the main toolbar, as long as you are not currently connected. Select your local target platform on the left-hand side and your localhost from the Connection drop down. Then, fill in the launch details. In the Activity panel, select the System Trace activity to initiate a session with pre-configured settings. Press Launch to start the session.
Once the session is completed, the Nsight Systems report is opened in a new document. By default, the timeline view is shown. It provides detailed information of the activity of the CPU and GPUs and helps understanding the overall behavior and performance of application. Once a CUDA kernel is identified to be on the critical path and not meeting the performance expectations, right click on the kernel launch on timeline and select Profile Kernel from the context menu. A new Start Activity opens up that is already preconfigured to profile the selected kernel launch. Proceed with optimizing the selected kernel using Non-Interactive Profile Activity
You can switch between different Report Pages using the tab bar on the top-left of the report.
You can also use Ctrl + Shift + N and Ctrl + Shift + P shortcut keys or corresponding toolbar button to navigate next and previous pages, respectively.
A report can contain any number of results. The Current dropdown allows switching between the different results in a report.
Diffing multiple results
On the Details page, use the Compare - Add Baseline button for the current result to become the baseline all other results from this report and any other report opened in the same instance of NVIDIA Nsight Compute get compared to.
When a baseline is set, every element on the Details page shows two values: The current value of the result in focus and the corresponding value of the baseline or the percentage of change from the corresponding baseline value.
Use the Clear Baselines entry from the same group button, the Profile menu or the corresponding toolbar button to remove all baselines. For more information see Baselines.
Following rules
On the Details page, many sections provide rules with valuable information on detected problems and optimization suggestions.
Rules can be user-defined too. For more information, see the Customization Guide.
3.3. Start Activity Dialog
Use the Start Activity Dialog to launch and attach to applications on your local and remote platforms. Start by selecting the Target Platform for profiling. By default (and if supported) your local platform will be selected. Select the platform on which you would like to start the target application or connect to a running process.
When using a remote platform, you will be asked to select or create a Connection in the top drop down. To create a new connection, select + and enter your connection details. When using the local platform, localhost will be selected as the default and no further connection settings are required. You can still create or select a remote connection, if profiling will be on a remote system of the same platform.
Depending on your target platform, select either Launch or Remote Launch to launch an application for profiling on the target. Note that Remote Launch will only be available if supported on the target platform.
Fill in the following launch details for the application:
Application Executable: Specifies the root application to launch. Note that this may not be the final application that you wish to profile. It can be a script or launcher that creates other processes.
Working Directory: The directory in which the application will be launched.
Command Line Arguments: Specify the arguments to pass to the application executable.
Environment: The environment variables to set for the launched application.
Select Attach to attach the profiler to an application already running on the target platform. This application must have been started using another NVIDIA Nsight Compute CLI instance. The list will show all application processes running on the target system which can be attached. Select the refresh button to re-create this list.
Finally, select the Activity to be run on the target for the launched or attached application. Note that not all activities are necessarily compatible with all targets and connection options. Currently, the following activities exist:
3.3.1. Remote Connections
Remote devices that support SSH can also be configured as a target in the Start Activity Dialog. To configure a remote device, ensure an SSH-capable Target Platform is selected, then press the + button. The following configuration dialog will be presented.
NVIDIA Nsight Compute supports both password and private key authentication methods. In this dialog, select the authentication method and enter the following information:
Password
IP/Host Name: The IP address or host name of the target device.
User Name: The user name to be used for the SSH connection.
Password: The user password to be used for the SSH connection.
Port: The port to be used for the SSH connection. (The default value is 22)
Deployment Directory: The directory to use on the target device to deploy supporting files. The specified user must have write permissions to this location.
Connection Name: The name of the remote connection that will show up in the Start Activity Dialog. If not set, it will default to <User>@<Host>:<Port>.
Private Key
IP/Host Name: The IP address or host name of the target device.
User Name: The user name to be used for the SSH connection.
SSH Private Key: The private key that is used to authenticate to SSH server.
SSH Key Passphrase: The passphrase for your private key.
Port: The port to be used for the SSH connection. (The default value is 22)
Deployment Directory: The directory to use on the target device to deploy supporting files. The specified user must have write permissions to this location.
Connection Name: The name of the remote connection that will show up in the Start Activity Dialog. If not set, it will default to <User>@<Host>:<Port>.
In addition to keyfiles specified by path and plain password authentication, NVIDIA Nsight Compute supports keyboard-interactive authentication, standard keyfile path searching and SSH agents.
When all information is entered, click the Add button to make use of this new connection.
When a remote connection is selected in the Start Activity Dialog, the Application Executable file browser will browse the remote file system using the configured SSH connection, allowing the user to select the target application on the remote device.
When an activity is launched on a remote device, the following steps are taken:
The command line profiler and supporting files are copied into the Deployment Directory on the remote device. (Only files that do not exist or are out of date are copied.)
Communication channels are opened to prepare for the traffic between the UI and the Application Executable.
For Interactive Profile activities, a SOCKS proxy is started on the host machine.
For Non-Interactive Profile activities, a remote forwarding channel is opened on the target machine to tunnel profiling information back to the host.
The Application Executable is executed on the remote device.
For Interactive Profile activities, a connection is established to the remote application and the profiling session begins.
For Non-Interactive Profile activities, the remote application is executed under the command line profiler and the specified report file is generated.
For non-interactive profiling activities, the generated report file is copied back to the host, and opened.
The progress of each of these steps is presented in the Progress Log.
Note that once either activity type has been launched remotely, the tools necessary for further profiling sessions can be found in the Deployment Directory on the remote device.
On Linux and Mac host platforms, NVIDIA Nsight Compute supports SSH remote profiling on target machines which are not directly addressable from the machine the UI is running on through the ProxyJump
and ProxyCommand
SSH options.
These options can be used to specify intermediate hosts to connect to or actual commands to run to obtain a socket connected to the SSH server on the target host and can be added to your SSH configuration file.
Note that for both options, NVIDIA Nsight Compute runs external commands and does not implement any mechanism to authenticate to the intermediate hosts using the credentials entered in the Start Activity Dialog. These credentials will only be used to authenticate to the final target in the chain of machines.
When using the ProxyJump
option NVIDIA Nsight Compute uses the OpenSSH client to establish the connection to the intermediate hosts. This means that in order to use ProxyJump
or ProxyCommand
, a version of OpenSSH supporting these options must be installed on the host machine.
A common way to authenticate to the intermediate hosts in this case is to use a SSH agent and have it hold the private keys used for authentication.
Since the OpenSSH SSH client is used, you can also use the SSH askpass mechanism to handle these authentications in an interactive manner.
It might happen on slow networks that connections used for remote profiling through SSH time out. If this is the case, the ConnectTimeout
option can be used to set the desired timeout value.
A known limitation of the remote profiling through SSH is that problems may arise if NVIDIA Nsight Compute tries to do remote profiling through SSH by connecting to the same machine it is running on. In this case, the workaround is to do local profiling through localhost
.
For more information about available options for the OpenSSH client and the ecosystem of tools it can be used with for authentication refer to the official manual pages.
3.3.2. Interactive Profile Activity
The Interactive Profile activity allows you to initiate a session that controls the execution of the target application, similar to a debugger. You can step API calls and workloads (CUDA kernels), pause and resume, and interactively select the kernels of interest and which metrics to collect.
This activity does currently not support profiling or attaching to child processes.
Enable CPU Call Stack
Collect the CPU-sided Call Stack at the location of each profiled kernel launch.
CPU Call Stack Types
If “Enable CPU Call Stack” is set to “Yes”, the type(s) of call stack may be selected here.
Enable NVTX Support
Collect NVTX information provided by the application or its libraries. Required to support stepping to specific NVTX contexts.
Disable Profiling Start/Stop
Ignore calls to cu(da)ProfilerStart
or cu(da)ProfilerStop
made by the application.
Enable Profiling From Start
Enables profiling from the application start. Disabling this is useful if the application calls cu(da)ProfilerStart
and kernels before the first call to this API should not be profiled. Note that disabling this does not prevent you from manually profiling kernels.
Cache Control
Control the behavior of the GPU caches during profiling. Allowed values: For Flush All, all GPU caches are flushed before each kernel replay iteration during profiling. While metric values in the execution environment of the application might be slightly different without invalidating the caches, this mode offers the most reproducible metric results across the replay passes and also across multiple runs of the target application.
For Flush None, no GPU caches are flushed during profiling. This can improve performance and better replicates the application behavior if only a single kernel replay pass is necessary for metric collection. However, some metric results will vary depending on prior GPU work, and between replay iterations. This can lead to inconsistent and out-of-bounds metric values.
Clock Control
Control the behavior of the GPU clocks during profiling. Allowed values: For Base, GPC and memory clocks are locked to their respective base frequency during profiling. This has no impact on thermal throttling. For None, no GPC or memory frequencies are changed during profiling.
Import Source
Enables permanently importing available source files into the report. Missing source files are searched in Source Lookup folders. Source information must be embedded in the executable, e.g. via the -lineinfo
compiler option. Imported files are used in the CUDA-C view on the Source Page.
Graph Profiling
Set if CUDA graphs should be stepped and profiled as individual Nodes or as complete Graphs. See the Kernel Profiling Guide for more information on this mode.
3.3.3. Profile Activity
The Profile activity provides a traditional, pre-configurable profiler. After configuring which kernels to profile, which metrics to collect, etc, the application is run under the profiler without interactive control. The activity completes once the application terminates. For applications that normally do not terminate on their own, e.g. interactive user interfaces, you can cancel the activity once all expected kernels are profiled.
This activity does not support attaching to processes previously launched via NVIDIA Nsight Compute. These processes will be shown grayed out in the Attach tab.
Output File
Path to report file where the collected profile should be stored. If not present, the report extension .ncu-rep
is added automatically. The placeholder %i
is supported for the filename component. It is replaced by a sequentially increasing number to create a unique filename. This maps to the --export
command line option.
Force Overwrite
If set, existing report file are overwritten. This maps to the --force-overwrite
command line option.
Target Processes
Select the processes you want to profile. In mode Application Only, only the root application process is profiled. In mode all, the root application process and all its child processes are profiled. This maps to the --target-processes
command line option.
Replay Mode
Select the method for replaying kernel launches multiple times. In mode Kernel, individual kernel launches are replayed transparently during the single execution of the target application. In mode Application, the entire target application is relaunched multiple times. In each iteration, additional data for the target kernel launches is collected. Application replay requires the program execution to be deterministic. This maps to the --replay-mode
command line option. See the Kernel Profiling Guide for more details on the replay modes.
Graph Profiling
Set if CUDA graphs should be profiled as individual Nodes or as complete Graphs.
Additional Options
All remaining options map to their command line profiler equivalents. See the Command Line Options for details.
3.3.4. Reset
Entries in the connection dialog are saved as part of the current project. When working in a custom project, simply close the project to reset the dialog.
When not working in a custom project, entries are stored as part of the default project. You can delete all information from the default project by closing NVIDIA Nsight Compute and then deleting the project file from disk.
3.4. Main Menu and Toolbar
Information on the main menu and toolbar.
New Project Create new profiling Projects with the New Project Dialog.
Open Project Open an existing profiling project.
Recent Projects Open an existing profiling project from the list of recently used projects.
Save Project Save the current profiling project.
Save Project As Save the current profiling project with a new filename.
Close Project Close the current profiling project.
New File Create a new file.
Open File Open an existing file.
Open Remote File
Download an existing file from a remote host and open it locally. The opened file will only exist in memory and will not be written to the local machine’s disk unless the user explicitly saves it. For more information concerning the selection of a remote host to download the file from, see the section about Remote Connections.
Only a subset of file types that are supported locally can be opened from a remote target. The following table lists file types that can be opened remotely.
Save Save the current file
Save As Save a copy of the current file with a different name or type or in a different location.
Save All Files Save all open files.
Close Close the current file.
Close All Files Close all open files.
Recent Files Open an existing file from the list of recently used files.
Exit Exit Nsight Compute.
Connection
Start Activity Open the Start Activity Dialog to launch or attach to a target application. Disabled when already connected.
Disconnect Disconnect from the current target application, allows the application to continue normally and potentially re-attach.
Terminate Disconnect from and terminate the current target application immediately.
Debug
Pause Pause the target application at the next intercepted API call or launch.
Resume Resume the target application.
Step In Step into the current API call or launch to the next nested call, if any, or the subsequent API call, otherwise.
Step Over Step over the current API call or launch and suspend at the next, non-nested API call or launch.
Step Out Step out of the current nested API call or launch to the next, non-parent API call or launch one level above.
Freeze API
When disabled, all CPU threads are enabled and continue to run during stepping or resume, and all threads stop as soon as at least one thread arrives at the next API call or launch. This also means that during stepping or resume the currently selected thread might change as the old selected thread makes no forward progress and the API Stream automatically switches to the thread with a new API call or launch. When enabled, only the currently selected CPU thread is enabled. All other threads are disabled and blocked.
Stepping now completes if the current thread arrives at the next API call or launch. The selected thread never changes. However, if the selected thread does not call any further API calls or waits at a barrier for another thread to make progress, stepping may not complete and hang indefinitely. In this case, pause, select another thread, and continue stepping until the original thread is unblocked. In this mode, only the selected thread will ever make forward progress.
Break On API Error When enabled, during resume or stepping, execution is suspended as soon as an API call returns an error code.
Run to Next Kernel See API Stream tool window.
Run to Next API Call See API Stream tool window.
Run to Next Range Start See API Stream tool window.
Run to Next Range End See API Stream tool window.
API Statistics Opens the API Statistics tool window
API Stream Opens the API Stream tool window
Resources Opens the Resources tool window
NVTX Opens the NVTX tool window
Profile
Profile Kernel When suspended at a kernel launch, select the profile using the current configuration.
Profile Series When suspended at a kernel launch, open the Profile Series configuration dialog to setup and collect a series of profile results.
Auto Profile Enable or disable auto profiling. If enabled, each kernel matching the current kernel filter (if any) will be profiled using the current section configuration.
Baselines Opens the Baselines tool window.
Clear Baselines Clear all current baselines.
Import Source Permanently import resolved source files into the report. Existing content may be overwritten.
Section/Rules Info Opens the Metric Selection tool window.
Tools
Project Explorer Opens the Project Explorer tool window.
Output Messages Opens the Output Messages tool window.
Options Opens the Options dialog.
Window
Save Window Layout Allows you to specify a name for the current layout. The layouts are saved to a Layouts folder in the documents directory as named “.nvlayout” files.
Apply Window Layout Once you have saved a layout, you can restore them by using the “Apply Window Layout” menu entry. Simply select the entry from sub-menu you want to apply.
Manage Window Layout Allows you to delete or rename old layouts.
Restore Default Layout Restore views to their original size and position.
Show Welcome Page Opens the Welcome Page.
Documentation Opens the latest documentation for NVIDIA Nsight Compute online.
Documentation (local) Opens the local HTML documentation for NVIDIA Nsight Compute that has shipped with the tool.
Check For Updates Checks online if a newer version of NVIDIA Nsight Compute is available for download.
Reset Application Data Reset all NVIDIA Nsight Compute configuration data saved on disk, including option settings, default paths, recent project references etc. This will not delete saved reports.
Send Feedback Opens a dialog that allows you to send bug reports and suggestions for features. Optionally, the feedback includes basic system information, screenshots, or additional files (such as profile reports).
About Opens the About dialog with information about the version of NVIDIA Nsight Compute.
3.4.2. Main Toolbar
The main toolbar shows commonly used operations from the main menu. See Main Menu for their description.
3.4.3. Status Banners
Status banners are used to display important messages, such as profiler errors. The message can be dismissed by clicking the ‘X’ button. The number of banners shown at the same time is limited and old messages can get dismissed automatically if new ones appear. Use the Output Messages window to see the complete message history.
The API Statistics window is available when NVIDIA Nsight Compute is connected to a target application. It opens by default as soon as the connection is established. It can be re-opened using Debug > API Statistics from the main menu.
Whenever the target application is suspended, it shows a summary of tracked API calls with some statistical information, such as the number of calls, their total, average, minimum and maximum duration. Note that this view cannot be used as a replacement for Nsight Systems when trying to optimize CPU performance of your application.
The Reset button deletes all statistics collected to the current point and starts a new collection. Use the Export to CSV button to export the current statistics to a CSV file.
3.5.2. API Stream
The API Stream window is available when NVIDIA Nsight Compute is connected to a target application. It opens by default as soon as the connection is established. It can be re-opened using Debug > API Stream from the main menu.
Whenever the target application is suspended, the window shows the history of API calls and traced kernel launches. The currently suspended API call or kernel launch (activity) is marked with a yellow arrow. If the suspension is at a subcall, the parent call is marked with a green arrow. The API call or kernel is suspended before being executed.
For each activity, further information is shown such as the kernel name or the function parameters (Func Parameters) and return value (Func Return). Note that the function return value will only become available once you step out or over the API call.
Use the Current Thread dropdown to switch between the active threads. The dropdown shows the thread ID followed by the current API name. One of several options can be chosen in the trigger dropdown, which are executed by the adjacent >> button. Run to Next Kernel resumes execution until the next kernel launch is found in any enabled thread. Run to Next API Call resumes execution until the next API call matching Next Trigger is found in any enabled thread. Run to Next Range Start resumes execution until the next start of an active profiler range is found. Profiler ranges are defined by using the cu(da)ProfilerStart/Stop
API calls. Run to Next Range Stop resumes execution until the next stop of an active profiler range is found. The API Level dropdown changes which API levels are shown in the stream. The Export to CSV button exports the currently visible stream to a CSV file.
3.5.3. Baselines
The Baselines tool window can be opened by clicking the Baselines entry in the Profile menu. It provides a centralized place from which to manage configured baselines. (Refer to Baselines, for information on how to create baselines from profile results.)
The baseline visibility can be controlled by clicking on the check box in a table row. When the check box is checked, the baseline will be visible in the summary header as well as all graphs in all sections. When unchecked the baseline will be hidden and will not contribute to metric difference calculations.
The baseline color can be changed by double-clicking on the color swatch in the table row. The color dialog which is opened provides the ability to choose an arbitrary color as well as offers a palette of predefined colors associated with the stock baseline color rotation.
The baseline name can be changed by double-clicking on the Name column in the table row. The name must not be empty and must be less than the Maximum Baseline Name Length as specified in the options dialog.
The z-order of a selected baseline can be changed by clicking the Move Baseline Up and Move Baseline Down buttons in the tool bar. When a baseline is moved up or down its new position will be reflected in the report header as well as in each graph. Currently, only one baseline may be moved at a time.
The selected baselines may be removed by clicking on the Clear Selected Baselines button in the tool bar. All baselines can be removed at once by clicking on the Clear All Baselines button, from either the global tool bar or the tool window tool bar.
The configured baselines can be saved to a file by clicking on the Save Baselines button in the tool bar. By default baseline files use the .ncu-bln
extension. Baseline files can be opened locally and/or shared with other users.
Baseline information can be loaded by clicking on the Load Baselines button in the tool bar. When a baseline file is loaded, currently configured baselines will be replaced. A dialog will be presented to the user to confirm this operation when necessary.
Differences between the current result and the baselines can be visualized with graphical bars for metrics in Details page section headers. Use the Difference Bars drop down to select the visualization mode. Bars are extending from left to right and have a fixed maximum.
3.5.4. Metric Details
The Metric Details tool window can be opened using the Metric Details entry in the Profile menu or the respective tool bar button. When a report and the tool window are open, a metric can be selected in the report to display additional information in the tool window. It also contains a search bar to look up metrics in the focused report.
Report metrics can be selected on the Details Page or on the Raw Page. The window will show basic information (name, unit and raw value of the metric) as well as additional information, such as its extended description.
For results collected for Green Contexts applications,
the Metric Details tool window also shows an additional row with the Attribution level of
the selected performance metric (e.g., Context or Green Context, if applicable).
The search bar can be used to open metrics in the focused report. It shows available matches as you type. The entered string must match from the start of the metric name.
By default, selecting or searching for a new metric updates the current Default Tab. You can click the pin button located in the upper-left corner of the tab control to create a copy of the default tab, unless the same metric is already pinned. This makes it possible to save multiple tabs and quickly switch between them to compare values.
Some metrics contain Instance Values. When available, they are listed in the tool window. Instance values can have a Correlation ID that allows correlating the individual value with its associated entity, e.g. a function address or instruction name.
For metrics collected with PM sampling, the correlation ID is the GPU timestamp in nanoseconds. It is shown relative to the first timestamp for this metric.
Note that the instance values for PM sampling metrics are currently not context-switched , even if this is enabled in the corresponding timeline UI.
3.5.5. Launch Details
The Launch Details tool window can be opened using the Launch Details entry
in the Profile menu or the respective tool bar button. When a result
containing multiple sub-launches is selected and this tool window is open, it
will display information about each sub-launch contained in the result.
This tool window is split into two sections:
a header displaying information applying to the result as a whole
a body displaying information specific to the viewed sub-launch
Header
On the left side of its header, this tool window displays the selected result’s
name and the number of sub-launches it is comprised of.
The right side contains a combo box that allows selection of the sub-launch the
body should represent. Each element of the combo box contains an index for the
sub-launch as well as the name of the function that it launched if available.
Body
The body of this tool window displays a table with sub-launch-specific metrics.
This table has four columns:
Metric Name: the name of the metric
Metric Unit: the unit for metric values
Instance Value: the value of this metric for the selected sub-launch
Aggregate Value: the aggregate value for this metric over all sub-launches
in the selected result
3.5.6. NVTX
The NVTX window is available when NVIDIA Nsight Compute is connected to a target application. If closed, it can be re-opened using Debug > NVTX from the main menu. Whenever the target application is suspended, the window shows the state of all active NVTX domains and ranges in the currently selected thread. Note that NVTX information is only tracked if the launching command line profiler instance was started with --nvtx
or NVTX was enabled in the NVIDIA Nsight Compute launch dialog.
Use the Current Thread dropdown in the API Stream window to change the currently selected thread. NVIDIA Nsight Compute supports NVTX named resources, such as threads, CUDA devices, CUDA contexts, etc. If a resource is named using NVTX, the appropriate UI elements will be updated.
3.5.7. CPU Call Stack
The CPU Call Stack window is available when NVIDIA Nsight Compute is connected to a target application.
If closed, it can be re-opened using Debug > CPU Call Stack from the main menu.
Whenever the target application is suspended, the window shows all enabled CPU call stacks for the currently selected thread.
Use the Call Stack Type dropdown menu to switch between stack types in case multiple stack types were enabled (e.g., Native, Python).
Note that Python call stack collection requires CPython version 3.9 or later.
3.5.8. Resources
The Resources window is available when NVIDIA Nsight Compute is connected to a target application. It shows information about the currently known resources, such as CUDA devices, CUDA streams or kernels. The window is updated every time the target application is suspended. If closed, it can be re-opened using Debug > Resources from the main menu.
Using the dropdown on the top, different views can be selected, where each view is specific to one kind of resource (context, stream, kernel, …). The Filter edit allows you to create filter expressions using the column headers of the currently selected resource.
The resource table shows all information for each resource instance. Each instance has a unique ID, the API Call ID when this resource was created, its handle, associated handles, and further parameters. When a resource is destroyed, it is removed from its table.
Memory Allocations
When using the asynchronous malloc/free APIs, the resource view for Memory Allocation will also include the memory objects created in this manner. These memory objects have a non-zero memory pool handle. The Mode column will indicate which code path was taken during the allocation of the corresponding object. The modes are:
REUSE_STREAM_SUBPOOL: The memory object was allocated in memory that was previously freed. The memory was backed by the memory pool set as current for the stream on which the allocation was made.
USE_EXISTING_POOL_MEMORY: The memory object was allocated in memory that was previously freed. The memory is backed by the default memory pool of the stream on which the allocation was made.
REUSE_EVENT_DEPENDENCIES: The memory object was allocated in memory that was previously freed in another stream of the same context. A stream ordering dependency of the allocating stream on the free action existed. Cuda events and null stream interactions can create the required stream ordered dependencies.
REUSE_OPPORTUNISTIC: The memory object was allocated in memory that was previously freed in another stream of the same context. However, no dependency between the free and allocation existed. This mode requires that the free be already committed at the time the allocation is requested. Changes in execution behavior might result in different modes for multiple runs of the application.
REUSE_INTERNAL_DEPENDENCIES: The memory object was allocated in memory that was previously freed in another stream of the same context. New internal stream dependencies may have been added in order to establish the stream ordering required to reuse a piece of memory previously released.
REQUEST_NEW_ALLOCATION: New memory had to be allocated for this memory object as no viable reusable pool memory was found. The allocation performance is comparable to using the non-asynchronous malloc/free APIs.
Graphviz DOT and SVG exports
Some of the shown Resources can also be exported to GraphViz DOT or SVG* files using the Export to GraphViz
or Export to SVG
buttons.
When exporting OptiX traversable handles, the traversable graph node types will be encoded using shapes and colors as described in the following table.
3.5.9. Metric Selection
The Metric Selection window can be opened from the main menu using Profile > Metric Selection. It tracks all metric sets, sections and rules currently loaded in NVIDIA Nsight Compute, independent from a specific connection or report. The directory to load those files from can be configured in the Profile options dialog. It is used to inspect available sets, sections and rules, as well as to configure which should be collected, and which rules should be applied. You can also specify a comma separated list of individual metrics, that should be collected. The window has two views, which can be selected using the dropdown in its header.
The Metric Sets view shows all available metric sets. Each set is associated with a number of metrics sections. You can choose a set appropriate to the level of detail for which you want to collect performance metrics. Sets which collect more detailed information normally incur higher runtime overhead during profiling.
When enabling a set in this view, the associated metric sections are enabled in the Metric Sections/Rules view. When disabling a set in this view, the associated sections in the Metric Sections/Rules view are disabled. If no set is enabled, or if sections are manually enabled/disabled in the Metric Sections/Rules view, the <custom> entry is marked active to represent that no section set is currently enabled. Note that the basic set is enabled by default.
Whenever a kernel is profiled manually, or when auto-profiling is enabled, only sections enabled in the Metric Sections/Rules view and individual metrics specified in input box are collected. Similarly, whenever rules are applied, only rules enabled in this view are active.
The enabled states of sections and rules are persisted across NVIDIA Nsight Compute launches. The Reload button reloads all sections and rules from disk again. If a new section or rule is found, it will be enabled if possible. If any errors occur while loading a rule, they will be listed in an extra entry with a warning icon and a description of the error.
Use the Enable All and Disable All checkboxes to enable or disable all sections and rules at once. The Filter text box can be used to filter what is currently shown in the view. It does not alter activation of any entry.
The table shows sections and rules with their activation status, their relationship and further parameters, such as associated metrics or the original file on disk. Rules associated with a section are shown as children of their section entry. Rules independent of any section are shown under an additional Independent Rules entry.
Clicking an entry in the table’s Filename column opens this file as a document. It can be edited and saved (ctrl + s) directly in NVIDIA Nsight Compute. After editing the file, Reload must be selected to apply those changes. Document also supports text search (ctrl + f), zoom in (ctrl + mouse scroll down), zoom out (ctrl + mouse scroll up) functionalities.
When a section or rule file is modified, the entry in the State column will show User Modified to reflect that it has been modified from its default state. When a User Modified row is selected, the Restore button will be enabled. Clicking the Restore button will restore the entry to its default state and automatically Reload the sections and rules.
Similarly, when a stock section or rule file is removed from the configured Sections Directory (specified in the Profile options dialog), the State column will show User Deleted. User Deleted files can also be restored using the Restore button.
Section and rule files that are created by the user (and not shipped with NVIDIA Nsight Compute) will show up as User Created in the state column.
See the Sections and Rules for the list of default sections for NVIDIA Nsight Compute.
3.6. Profiler Report
The profiler report contains all the information collected during profiling for each kernel launch. In the user interface, it consists of a header with general information, as well as controls to switch between report pages or individual collected launches.
3.6.1. Header
The top of the report shows a table with information about the selected profile result (as Current) and potentially additional baselines.
For many values in this table, tooltips provide additional information or data, e.g., the tooltip of the column Attributes provides additional information about the context type and resources used for the launch.
The Result dropdown can be used to switch between all collected kernel launches. The information displayed in each page commonly represents the selected launch instance. On some pages (e.g. Raw), information for all launches is shown and the selected instance is highlighted. You can type in this dropdown to quickly filter and find a kernel launch.
The Apply Filters button opens the filter dialog. You can use more than one filter to narrow down your results. On the filter dialog, enter your filter parameters and press OK button. The Launch dropdown, Summary Page table, and Raw Page table will be filtered accordingly. Select the arrow dropdown to access the Clear Filters button, which removes all filters.
Underneath the current and baseline results are the tabs for switching between the report pages. The pages themselves are explained in detail in the next section.
Each group button to the right of the page tabs opens a context menu that features related actions.
Some actions may be enabled only when the related report page is selected.
Compare
Add Baseline promotes the current result in focus to become the baseline of all other results from this report and any other report opened in the same instance of NVIDIA Nsight Compute.
Clear Baselines removes all currently active baselines. You may also use the Baselines tool window to manage baseline for comparison.
Source Comparison navigates to the Source Comparison document in case at least two profile results are available for comparison.
Tools
Occupancy Calculator opens the Occupancy Calculator in a new document.
Metric Details Windows opens the Metric Details tool window. When the window is open and a metric is selected elsewhere in the report, it shows detailed information about it.
Launch Details Windows opens the Launch Details tool window. When the window is open and a result containing multiple sub-launches is selected, it displays information about each sub-launch in the result.
Show/Hide Rules Output toggles the visibility of rule results.
Show/Hide Section Descriptions toggles the visibility of section descriptions on the Details page.
Show/Hide Green Context Markers toggles the visibility of markers for attributable Green Context metrics.
Expand Sections expands all sections to show their body contents, not only header and rule output. Note that sections may have multiple bodies and the visible one can be chosen using the dropdown in the section header.
Collapse Sections collapses all sections to show only their header and rule output.
Export
Copy as Image - Copies the contents of the page to the clipboard as an image.
Save as Image - Saves the contents of the page to a file as an image.
Save as PDF - Saves the contents of the page to a file as a PDF.
Export to CSV - Exports the contents of the page to CSV format.
More (three bars icon)
Apply Rules applies all rules available for this report. If rules had been applied previously, those results will be replaced. By default, rules are applied immediately once the kernel launch has been profiled. This can be changed in the options under Tools > Options > Profile > Report UI > Apply Applicable Rules Automatically.
Reset to Default resets the page to a default state by removing any persisted settings.
3.6.2. Report Pages
Use the Page dropdown in the header to switch between the report pages.
By default, when opening a report with a single profile result, the Details Page is shown. When opening a report with multiple results, the Summary Page is selected instead. You can change the default report page in the Profile options.
Summary Page
The Summary page shows a table of all collected results in the report, as well as a list of the most important rule outputs (Prioritized Rules) which are ordered by the estimated speedup that could potential be obtained by following their guidance. Prioritized Rules are shown by default and can be toggled with the [R] button on the upper right of the page.
The Summary Table gives you a quick comparison overview across all profiled workloads. It contains a number of important, pre-selected metrics which can be customized as explained below. Its columns can be sorted by clicking the column header. You can transpose the table with the Transpose button. Aggregate of all results per each counter metric is shown in the table header along with the column name. You can change the aggregated values by selecting the desired results for multiple metrics simultaneously. When selecting any entry by single-click, a list of its Prioritized Rules will be shown below the table. Double-click any entry to make the result the currently active one and switch to the Details Page page to inspect its performance data.
By default, kernel demangled names are simplified, renamed and shown in an optimized manner. This behavior can be changed with Rename Demangled Names option. If an auto-simplified name is not useful, you can rename it through a configuration file.
You can also persist the updated names directly in the report by double-clicking on the name, renaming and saving the report.
Use Rename Kernels Config Path option to specify the configuration file which should be used while importing renamed kernels or exporting demangled names with mappings to rename them. To export names to a new file, click Export button and use Rename Kernels Config option.
See Kernel Renaming for more details on configuration file usage.
You can configure the list of metrics included in this table in the Profile options dialog. If a metric has multiple instance values, the number of instances is shown after its standard value. A metric with ten instance values could for example look like this: 35.48 {10}
. In the Profile options dialog, you can select that all instance values should be shown individually. You can also inspect the instances values of a metric result in the Metric Details tool window.
In addition to metrics, you can also configure the table to include any of the following properties:
Properties
property__api_call_id
ID of the API call associated with this profile result.
property__block_size
Block Size. If the result contains multiple launches, this will contain the maximum value for each dimension of the block.
property__creation_time
Local collection time.
property__demangled_name
Kernel demangled name, potentially renamed.
property__device_name
GPU device name.
property__estimated_speedup
Maximal relative speedup achievable for this profile result as estimated by the guided analysis rules.
property__function_name
Kernel function name or range name.
property__grid_dimensions
Grid Dimensions. If the result contains multiple launches, this will contain the maximum value for each dimension of the grid.
property__grid_offset
Grid Offset.
property__grid_size
Grid Size. If the result contains multiple launches, this will contain the maximum value for each dimension of the grid.
property__issues_detected
Number of issues detected by guided analysis rules for this profile result.
property__kernel_id
Kernel ID.
property__mangled_name
Kernel mangled name.
property__original_demangled_name
Original kernel demangled name without any renaming.
property__process_name
Process name.
property__runtime_improvement
Runtime improvement corresponding to the estimated speedup.
property__series_id
ID of the profile series.
property__series_parameters
Profile series parameters.
property__thread_id
CPU thread ID.
For Range Replay reports, a smaller set of columns is shown by default, as not all apply to such results.
For the currently selected metric result the Prioritized Rules show the most impactful rule results with respect to the estimated potential speedup. Clicking on any of the rule names on the left allows you to easily navigate to the containing section on the details page. With the downward-facing arrow on the right a table with the relevant key performance indicators can be toggled. This table contains the metrics which should be tracked when optimizing performance according to the rule guidance.
Details Page
Overview
The Details page is the main page for all metric data collected during a kernel launch. The page is split into individual sections.
Each section consists of a header table and an optional body that can be expanded. You can expand or collapse the body of each section by clicking on its respective header.
The sections are completely user defined and can be changed easily by updating their respective files.
For more information on customizing sections, see the Customization Guide.
For a list of sections shipped with NVIDIA Nsight Compute, see Sections and Rules.
By default, once a new profile result is collected, all applicable rules are applied. Any rule results will be shown as Recommendations on this page. Most rule results will contain an optimization advice along with an estimate of the improvement that could be achieved when successfully implementing this advice. Other rule results will be purely informative or have a warning icon to indicate a problem that occurred during execution (e.g., an optional metric that could not be collected). Results with error icons typically indicate an error while applying the rule.
Estimates of potential improvement are shown below the rule result’s name and exist in two types. Global estimates (“Est. Speedup”) are an approximation of the decrease in workload runtime, whereas local estimates (“Est. Local Speedup”) are an approximation of the increase in efficiency of the hardware utilization of the particular performance problem the rule addresses.
If a rule result references another report section, it will appear as a link in the recommendation. Select the link to scroll to the respective section. If the section was not collected in the same profile result, enable it in the Metric Selection tool window.
You can add or edit comments in each section of the Details view by clicking on the comment button (speech bubble). The comment icon will be highlighted in sections that contain a comment. Comments are persisted in the report and are summarized in the Comments Page.
Besides their header, sections typically have one or more bodies with additional charts or tables. Click the triangle Expander icon in the top-left corner of each section to show or hide those. If a section has multiple bodies, a dropdown in their top-right corner allows you to switch between them.
Memory
If enabled, the Memory Workload Analysis section contains a Memory chart that visualizes data transfers, cache hit rates, instructions and memory requests. More information on how to use and read this chart can be found in the Kernel Profiling Guide.
Occupancy
You can open the Occupancy Calculator by clicking on the calculator button in the report header or in the header of the Occupancy Section.
Range Replay
Note that for Range Replay results some UI elements, analysis rules, metrics or section body items such as charts or tables might not be available, as they only apply to kernel launch-based results. The filters can be checked in the corresponding section files.
Rooflines
If enabled, the GPU Speed Of Light Roofline Chart section contains a Roofline chart that is particularly helpful for visualizing kernel performance at a glance. (To enable roofline charts in the report, ensure that the section is enabled when profiling.) More information on how to use and read this chart can be found in Roofline Charts. NVIDIA Nsight Compute ships with several different definitions for roofline charts, including hierarchical rooflines. These additional rooflines are defined in different section files. While not part of the full section set, a new section set called roofline was added to collect and show all rooflines in one report. The idea of hierarchical rooflines is that they define multiple ceilings that represent the limiters of a hardware hierarchy. For example, a hierarchical roofline focusing on the memory hierarchy could have ceilings for the throughputs of the L1 cache, L2 cache and device memory. If the achieved performance of a kernel is limited by one of the ceilings of a hierarchical roofline, it can indicate that the corresponding unit of the hierarchy is a potential bottleneck.
The roofline chart can be zoomed and panned for more effective data analysis, using the controls in the table below.
When the Metric Details tool window is open, you can click on an achieved value in the chart to see its metric formula in the tool window.
Click the Zoom In button in the top right corner of the chart.
Click the left mouse button and drag to create a rectangle that bounds the area of interest.
Press the plus (+) key.
Use Ctrl + MouseWheel (Windows and Linux only)
Click the Zoom Out button in the top right corner of the chart.
Click the right mouse button.
Press the minus (-) key.
Use Ctrl + MouseWheel (Windows and Linux only)
Click the Zoom Reset button in the top right corner of the chart.
Press the Escape (Esc) key.
Use Ctrl (Command on Mac) + LeftMouseButton to grab the chart, then move the mouse.
Use the cursor keys.
Source
Sections such as Source Counters can contain source hot spot tables. These tables indicate the N highest or lowest values of one or more metrics in your kernel source code. Select the location links to navigate directly to this location in the Source Page. Hover the mouse over a value to see which metrics contribute to it.
Timelines
When collecting metrics with PM sampling, they can be viewed in a timeline. The timeline shows metrics selected in the respective section file or on the command line with their labels/names and their values over time.
Different metrics may be collected in different passes (replays) of the workload, as only a limited number of them can be sampled in the same pass. Context switch trace is used to filter the collected data to only include samples from the profiled contexts and to align it in the timeline.
You can hover the mouse over a metric row label to see further information on the metrics in the row. Hovering over a sample on the timeline shows the metric values at that timestamp within the current row. With the Metric Details tool window open, click to select a value on the timeline and show the metric and all its raw timestamps (absolute and relative) correlated values in the tool window.
You can also use the Metric Details tool window to inspect profiler metrics generated during PM sampling. These provide information about the used sampling intervals, buffer sizes, dropped samples and other properties for each collection pass. A detailed list can be found in the metrics reference.
The timeline has a context menu for further actions regarding copying content and zooming.
In addition, the Enable/Disable Context Switch Filter option can be used to enable or disable the filtering of the timeline data with context switch information, if it is available.
When the context switch filter is enabled (the default), samples from each pass group are only shown for the active contexts.
When the context switch filter is disabled, the raw collected sampling data is shown along with a separate row for each pass group’s context switch trace.
When the context menu option is not available, the report does not include context switch trace data.
In this case, the option Enable/Disable Trim Filter is shown instead, which, when enabled, tries to align based on the first non-zero value in any sampling metric in this pass group.
However, this fallback does not take into account actual context switches.
The timeline row Workload Execution shows each kernel’s start and end timestamp. When the context switch filter is enabled, kernel execution is only shown for one of the passes for the active contexts. When the context switch filter is disabled, kernel execution is shown for all the passes.
Source Page
The Source page correlates assembly (SASS) with high-level code such as CUDA-C, Python or PTX. In addition, it displays instruction-correlated metrics to help pinpoint performance problems in your code.
The page can be switched between different Views to focus on a specific source layer or see two layers side-by-side. This includes SASS, PTX and Source (CUDA-C, Fortran, Python, …), as well as their combinations. Which options are available depends on the source information embedded into the executable.
The high-level Source (CUDA-C) view is available if the application was built with the -lineinfo
or --generate-line-info
nvcc flag to correlate SASS and source. When using separate linking at the ELF level, there is no PTX available in the ELF that would correspond to the final SASS. As such, NVIDIA Nsight Compute does not show any PTX even though it would be available statically in the executable and could be shown with cuobjdump -all -lptx
. However, this is a pre-linked version of the PTX and cannot be reliably used for correlation.
Navigation
The View dropdown can be used to select different code (correlation) options: SASS, PTX and Source (CUDA-C, Fortran, Python, …).
In side-by-side views, when selecting a line in the left-hand- or right-hand-side, any correlated lines in the opposite view are highlighted. However, when the Show Single File For Multi-File Sources option is set to Yes, the target file or source object must already be selected in the respective view for those correlated lines to be shown.
In side-by-side views, the view which has blue border is called focused view. Focused view can be changed by clicking anywhere in the other view. Control groups which are bordered with blue color will work for a focused view only.
The Source dropdown allows you to switch between the files or functions that provide the content in the view. When a different source entry is selected, the view scrolls to the start of this file or function.
If a view contains multiple source files or functions, [+] and [-] buttons are shown. These can be used to expand or collapse the view, thereby showing or hiding the file or function content except for its header. These will work for both the views if that group is in the linked state otherwise in an unlinked state it will work for a focused view. If collapsed, all metrics are shown aggregated to provide a quick overview.
You can use the Find (source code) line edit to search the Source column of a focused view. Enter the text to search and use the associated buttons to find the next or previous occurrence in this column. While the line edit is selected, you can also use the Enter or Shift*+*Enter keys to search for the next or previous occurrence, respectively.
The SASS view is filtered to only show functions that were executed in the launch. You can toggle the Show Only Executed Functions option to change this, but performance of this page may be negatively affected for large binaries. It is possible that some SASS instructions are shown as N/A. Those instructions are not currently exposed publicly.
In side-by-side views, the Navigate By dropdowns are linked with each other by default, thereby changing column names from one dropdown will change it in the other view only if column is available. These dropdowns can be unlinked with the link-unlink button provided just before it.
Only filenames are shown in the view, together with a File Not Found error, if the source files cannot be found in their original location. This can occur, for example, if the report was moved to a different system. Select a filename and click the Resolve button above to specify where this source can be found on the local filesystem. However, the view always shows the source files if the import source option was selected during profiling, and the files were available at that time. If a file is found in its original or any source lookup location, but its attributes don’t match, a File Mismatch error is shown. See the Source Lookup options for changing file lookup behavior.
If the report was collected using remote profiling, and automatic resolution of remote files is enabled in the Profile options, NVIDIA Nsight Compute will attempt to load the source from the remote target. If the connection credentials are not yet available in the current NVIDIA Nsight Compute instance, they are prompted in a dialog. Loading from a remote target is currently only available for Linux x86_64 targets and Linux and Windows hosts.
Metrics
Metrics Correlation
The page is most useful when inspecting performance information and metrics correlated with your code. Metrics are shown in columns, which can be enabled or disabled using the Column Chooser accessible using the column header right click menu.
To not move out of view when scrolling horizontally, columns can be fixed. By default, the Source column is fixed to the left, enabling easy inspection of all metrics correlated to a source line. To change fixing of columns, right click the column header and select Freeze or Unfreeze, respectively.
The heatmap on the right-hand side of each view can be used to quickly identify locations with high metric values of the currently selected metric in the dropdown. The heatmap uses a black-body radiation color scale where black denotes the lowest mapped value and white the highest, respectively. The current scale is shown when clicking and holding the heatmap with the right mouse button.
By default, applicable metrics are shown as percentage values relative to their sum across the launch. A bar is filling from left to right to indicate the value at a specific source location relative to this metric’s maximum within the launch. The [%] and [+-] buttons can be used to switch the display from relative to absolute and from abbreviated absolute to full-precision absolute, respectively. For relative values and bars, the [circle/pie] button can be used to switch the display between relative to global (launch) and relative to local (function/file) scope. This button is disabled when the view is collapsed, as percentages are always relative to the global launch scope in this case.
Pre-Defined Source Metrics
Live Registers
Number of registers that need to be kept valid by the compiler. A high value indicates that many registers are required at this code location, potentially increasing the register pressure and the maximum number of register required by the kernel.
The total number of registers reported as launch__registers_per_thread
may be significantly higher than the maximum live registers. The compiler may need to allocate specific registers that can creates holes in the allocation, thereby affecting launch__registers_per_thread
, even if the maximum live registers is smaller. This may happen due to ABI restrictions, or restrictions enforced by particular hardware instructions. The compiler may not have a complete picture of which registers may be used in either callee or caller and has to obey ABI conventions, thereby allocating different registers even if some register could have theoretically been re-used.
Warp Stall Sampling (All Samples)1
The number of samples from the Statistical Sampler at this program location.
Warp Stall Sampling (Not-issued Samples)2
The number of samples from the Statistical Sampler at this program location on cycles the warp scheduler issued no instructions. Note that (Not Issued) samples may be taken on a different profiling pass than (All) samples mentioned above, so their values do not strictly correlate.
This metric is only available on devices with compute capability 7.0 or higher.
Instructions Executed
Number of times the source (instruction) was executed per individual warp, independent of the number of participating threads within each warp.
Thread Instructions Executed
Number of times the source (instruction) was executed by any thread, regardless of predicate presence or evaluation.
Predicated-On Thread Instructions Executed
Number of times the source (instruction) was executed by any active, predicated-on thread. For instructions that are executed unconditionally (i.e. without predicate), this is the number of active threads in the warp, multiplied with the respective Instructions Executed value.
Avg. Threads Executed
Average number of thread-level executed instructions per warp, regardless of their predicate.
Avg. Predicated-On Threads Executed
Average number of predicated-on thread-level executed instructions per warp.
Divergent Branches
Number of divergent branch targets, including fallthrough. Incremented only when there are two or more active threads with divergent targets. Divergent branches can lead to warp stalls due to resolving the branch or instruction cache misses.
Information on Memory Operations
L1 Tag Requests Global
memory_l1_tag_requests_global
Number of L1 tag requests generated by global memory instructions.
L1 Conflicts Shared N-Way
derived__memory_l1_conflicts_shared_nway
Average N-way conflict in L1 per shared memory instruction. A 1-way access has no conflicts and resolves in a single pass. Note: This is a derived metric which can not be collected directly.
L1 Wavefronts Shared Excessive
derived__memory_l1_wavefronts_shared_excessive
Excessive number of wavefronts in L1 from shared memory instructions, because not all not predicated-off threads performed the operation. Note: This is a derived metric which can not be collected directly.
L1 Wavefronts Shared
memory_l1_wavefronts_shared
Number of wavefronts in L1 from shared memory instructions.
L1 Wavefronts Shared Ideal
memory_l1_wavefronts_shared_ideal
Ideal number of wavefronts in L1 from shared memory instructions, assuming each not predicated-off thread performed the operation.
L2 Theoretical Sectors Global Excessive
derived__memory_l2_theoretical_sectors_global_excessive
Excessive theoretical number of sectors requested in L2 from global memory instructions, because not all not predicated-off threads performed the operation. Note: This is a derived metric which can not be collected directly.
L2 Theoretical Sectors Global
memory_l2_theoretical_sectors_global
Theoretical number of sectors requested in L2 from global memory instructions.
L2 Theoretical Sectors Global Ideal
memory_l2_theoretical_sectors_global_ideal
Ideal number of sectors requested in L2 from global memory instructions, assuming each not predicated-off thread performed the operation.
L2 Theoretical Sectors Local
memory_l2_theoretical_sectors_local
Theoretical number of sectors requested in L2 from local memory instructions.
All L1/L2 Sectors/Wavefronts/Requests metrics give the number of achieved (actually required), ideal, and excessive (achieved - ideal) sectors/wavefronts/requests. Ideal metrics indicate the number that would needed, given each not predicated-off thread performed the operation of given width. Excessive metrics indicate the required surplus over the ideal case. Reducing divergence between threads can reduce the excess amount and result in less work for the respective HW units.
Several of the above metrics on memory operations were renamed in version 2021.2 as follows:
L2 Explicit Evict Policy Metrics
Starting with the NVIDIA Ampere architecture the eviction policy of the L2 cache can be tuned to match the kernel’s access pattern. The eviction policy can be either set implicitly for a memory window (for more details see CUaccessProperty) or set explicitly per executed memory instruction. If set explicitly, the desired eviction behavior for the cases of an L2 cache hit or miss are passed as input to the instruction. For more details refer to CUDA’s Cache Eviction Priority Hints.
L2 Explicit Evict Policies
smsp__inst_executed_memdesc_explicit_evict_type
Comma separated list of configured explicit eviction policies. As the policies can be set dynamically at runtime, this list includes all policies that were part of any executed instruction.
L2 Explicit Hit Policy Evict First
smsp__inst_executed_memdesc_explicit_hitprop_evict_first
Number of times a memory instruction was executed by any warp which had the evict_first
policy set in case the access leads to a cache hit in L2. Data cached with this policy will be first in the eviction priority order and will likely be evicted when cache eviction is required. This policy is suitable for streaming data.
L2 Explicit Hit Policy Evict Last
smsp__inst_executed_memdesc_explicit_hitprop_evict_last
Number of times a memory instruction was executed by any warp which had the evict_last
policy set in case the access leads to a cache hit in L2. Data cached with this policy will be last in the eviction priority order and will likely be evicted only after other data with evict_normal
or evict_first
eviction policy is already evicted. This policy is suitable for data that should remain persistent in cache.
L2 Explicit Hit Policy Evict Normal
smsp__inst_executed_memdesc_explicit_hitprop_evict_normal
Number of times a memory instruction was executed by any warp which had the evict_normal
(default) policy set in case the access leads to a cache hit in L2.
L2 Explicit Hit Policy Evict Normal Demote
smsp__inst_executed_memdesc_explicit_hitprop_evict_normal_demote
Number of times a memory instruction was executed by any warp which had the evict_normal_demote
policy set in case the access leads to a cache hit in L2.
L2 Explicit Miss Policy Evict First
smsp__inst_executed_memdesc_explicit_missprop_evict_first
Number of times a memory instruction was executed by any warp which had the evict_first
policy set in case the access leads to a cache miss in L2. Data cached with this policy will be first in the eviction priority order and will likely be evicted cache eviction is required. This policy is suitable for streaming data.
L2 Explicit Miss Policy Evict Normal
smsp__inst_executed_memdesc_explicit_missprop_evict_normal
Number of times a memory instruction was executed by any warp which had the evict_normal
(default) policy set in case the access leads to a cache miss in L2.
Individual Warp Stall Sampling Metrics
All stall_* metrics show the information combined in Warp Stall Sampling individually. See Statistical Sampler for their descriptions.
See the Customization Guide on how to add additional metrics for this view and the Metrics Reference for further information on available metrics.
Register Dependencies
Dependencies between registers are displayed in the SASS view. When a register is read, all the potential addresses where it could have been written are found. The links between these lines are drawn in the view. All dependencies for registers, predicates, uniform registers and uniform predicates are shown in their respective columns.
The picture above shows some dependencies for a simple CUDA kernel. On the first row, which is line 9 of the SASS code, we can see writes on registers R2 and R3, represented by filled triangles pointing to the left. These registers are then read on lines 17, 20 and 23, and this is represented by regular triangles pointing to the right. There are also some lines where both types of triangles are on the same line, which means that a read and a write occured for the same register.
Dependencies across source files and functions are not tracked.
The Register Dependencies Tracking feature is enabled by default, but can be disabled completely in Tools > Options > Profile > Report Source Page > Enable Register Dependencies.
This metric was previously called Sampling Data (All).
This metric was previously called Sampling Data (Not Issued).
This button opens a dialog that shows you the list of saved source view profiles. Such profiles can be created using the Create button in the dialog. Profiles let you store the column properties of all views in the report to a file. Such properties include column visibility, freeze state, width, order and the selected navigation metric. A saved profile can be applied to any opened report using the Apply button. This updates the column properties mentioned above from the selected profile in all views.
Profiles are useful for configuring views to your preferences, or for a certain use case. Start by choosing metric columns from the Column Chooser. Next, configure other properties like freezing column, changing width or order and setting a heatmap metric in the Navigation dropdown before creating the profile. Once a profile is created, you can always use this profile on any opened report to hide all non-required columns or to restore your configured properties. Simply select the profile from the source view profiles dialog and click the Apply button.
Note that the column properties are stored separately for each View in the profile and when applied, only those views will be updated which are present in the selected profile. You will not see the metric columns that are not available in your report even if those were configured to be visible in the source profile you have applied.
Additional Tables
Inline Functions Table
This table shows all call sites where the selected function from the Source view is inlined.
In the Source view, correlated metric values of every inline function source line are an aggregation of metric values of related SASS lines from all call sites. The table provides such SASS lines info for each call site individually to help in identifying which call site contributes how much to the overall metric values.
Source Markers Table
The code in the different Views can also contain warnings, errors or just notifications that are displayed as Source Markers in the left header, as shown below. These can be generated from multiple sources, but as of now only NvRules are supported.
This table shows all lines containing Source Markers in the open View. The user can click on the line number to navigate to the corresponding line in the view.
Statistics Table
This table allows the user to select multiple lines in the source code above and calculate a variety of statistics. They can be used to check if anything seems unusual, like an unexpected gap between minimum and maximum.
Cells will be empty if the metric has either not been collected or its statistical value is not meaningful in the context of the calculation. For example, a memory access or any other non-numeric metric can’t have a minimum or maximum since there is no defined order.
Limitations
Range Replay
When using Range Replay mode, instruction-level source metrics are not available.
Graph Profiling
When profiling complete CUDA graphs, instruction-level source metrics are not available.
Context Page
The CPU Call Stack section of this report page shows the CPU call stack(s) for the executing CPU thread at the time the kernel was launched. For this information to show up in the profiler report, the option to collect CPU call stacks had to be enabled in the Connection Dialog or using the corresponding NVIDIA Nsight Compute CLI command line parameter.
NVIDIA Nsight Compute supports to collect native CPU call stacks as well as call stacks for Python applications.
Either or both types can be selected in the Activity menu of the Connection Dialog (via the “CPU Call Stack Types” option),
or using the NVIDIA Nsight Compute CLI command line parameter –call-stack-type.
In case both types are enabled, a dropdown menu will appear to select the desired call stack type.
Note that Python call stack collection requires CPython version 3.9 or later.
The NVTX State section of this report page shows the NVTX context when the kernel was launched. All thread-specific information is with respect to the thread of the kernel’s launch API call. Note that NVTX information is only collected if the profiler is started with NVTX support enabled, either in the Connection Dialog or using the NVIDIA Nsight Compute CLI command line parameter.
This page has been renamed from “Call Stack / NVTX Page”.
Comments Page
The Comments page aggregates all section comments in a single view and allows the user to edit those comments on any launch instance or section, as well as on the overall report. Comments are persisted with the report. If a section comment is added, the comment icon of the respective section in the Details Page will be highlighted.
Raw Page
The Raw page shows a list of all collected metrics with their units per profiled kernel launch. It can be exported, for example, to CSV format for further analysis. The page features a filter edit to quickly find specific metrics. You can transpose the table of kernels and metrics by using the Transpose button.
If a metric has multiple instance values, the number of instances is shown after the standard value. This metric for example has ten instance values: 35.48 {10}
. You can select in the Profile options dialog that all instance values should be shown individually or inspect the metric result in the Metric Details tool window.
Session Page
This Session page contains basic information about the report and the machine, as well as device attributes of all devices for which launches were profiled. When switching between launch instances, the respective device attributes are highlighted.
3.6.3. Metrics and Units
Numeric metric values are shown in various places in the report, including the header and tables and charts on most pages. NVIDIA Nsight Compute supports various ways to display those metrics and their values.
When available and applicable to the UI component, metrics are shown along with their unit. This is to make it apparent if a metric represents cycles, threads, bytes/s, and so on. The unit will normally be shown in rectangular brackets, e.g. Metric Name [bytes] 128
.
By default, units are scaled automatically so that metric values are shown with a reasonable order of magnitude. Units are scaled using their SI-factors, i.e. byte-based units are scaled using a factor of 1000 and the prefixes K, M, G, etc. Time-based units are also scaled using a factor of 1000, with the prefixes n, u and m. This scaling can be disabled in the Profile options.
Metrics which could not be collected are shown as n/a
and assigned a warning icon. If the metric floating point value is out of the regular range (i.e. nan
(Not a number) or inf
(infinite)), they are also assigned a warning icon. The exception are metrics for which these values are expected and which are allow-listed internally.
3.6.4. Filtered Profiler Report
A filtered profiler report is a subset of the full profiler report. You can save filtered or selected results from the full profiler report to a new file.
To save the filtered results, you can use the File > Save Filtered Results As menu option, which will save the Filtered Results to a new file.
Alternatively, you can select one or more results on the Summary Page or Raw Page and save them to a new file using the File > Save Selected Results As menu option or the right-click Save Result(s) menu option.
3.7. Baselines
NVIDIA Nsight Compute supports diffing collected results across one or multiple reports using Baselines. Each result in any report can be promoted to a baseline. This causes metric values from all results in all reports to show the difference to the baseline. If multiple baselines are selected simultaneously, metric values are compared to the average across all current baselines. Baselines are not stored with a report and are only available as long as the same NVIDIA Nsight Compute instance is open, unless they are saved to a ncu-bln
file from the Baselines tool window.
Select Add Baseline to promote the current result in focus to become a baseline. If a baseline is set, most metrics on the Details Page, Raw Page and Summary Page show two values: the current value of the result in focus, and the corresponding value of the baseline or the percentage of change from the corresponding baseline value. (Note that an infinite percentage gain, inf%, may be displayed when the baseline value for the metric is zero, while the focus value is not.)
If multiple baselines are selected, each metric will show the following notation:
<focus value> (<difference to baselines average [%]>, z=<standard score>@<number of values>)
The standard score is the difference between the current value and the average across all baselines, normalized by the standard deviation. If the number of metric values contributing to the standard score equals the number of results (current and all baselines), the @<number of values> notation is omitted.
Baseline added for the current result in focus is always shown on the top. However, the actual order of added baselines is shown in the Baselines tool window.
Double-clicking on a baseline name allows the user to edit the displayed name. Edits are committed by pressing Enter/Return
or upon loss of focus, and abandoned by pressing Esc
. Hovering over the baseline color icon allows the user to remove this specific baseline from the list.
Use the Clear Baselines entry from the dropdown button, the Profile menu, or the corresponding toolbar button to remove all baselines.
Baseline changes can also be made in the Baselines tool window.
3.8. Standalone Source Viewer
NVIDIA Nsight Compute includes a standalone source viewer for cubin files. This view is identical to the Source Page, except that it won’t include any performance metrics.
Cubin files can be opened from the File > Open main menu command. The SM Selection dialog will be shown before opening the standalone source view. If available, the SM version present in the file name is pre-selected. For example, if your file name is mergeSort.sm_80.cubin
then SM 8.0 will be pre-selected in the dialog. Choose the appropriate SM version from the drop down menu if it’s not included in the file name.
Click Ok button to open Standalone Source Viewer.
3.9. Source Comparison
Source comparison provides a way to see the source files of two profile results side by side. It enables to quickly identify source differences and understand changes in metric values.
To compare two results side by side add one result as a baseline, navigate to the other result, and then click the Source Comparison button located in the report header.
For example, if you want to compare kernel XYZ from report R1 with kernel XYZ from report R2, first open report R1, add the profile result for kernel XYZ as baseline, open report R2, choose kernel XYZ, and then click the Source Comparison button.
Source comparison will be shown only with first added baseline result.
Currently only high-level Source (CUDA-C) view and SASS view are supported for comparison.
Navigation to the previous or next difference is supported using the navigation buttons or the keyboard shortcuts Ctrl + 1 and Ctrl + 2.
3.10. Occupancy Calculator
NVIDIA Nsight Compute provides an Occupancy Calculator that allows you to compute the multiprocessor occupancy of a GPU for a given CUDA kernel. It offers feature parity to the CUDA Occupancy Calculator spreadsheet.
The Occupancy Calculator can be opened directly from a profile report or as a new activity. The occupancy calculator data can be saved to a file using File > Save. By default, the file uses the .ncu-occ
extension. The occupancy calculator file can be opened using File > Open File
Launching from the Connection Dialog
Select the Occupancy Calculator activity from the connection dialog. You can optionally specify an occupancy calculator data file, which is used to initialize the calculator with the data from the saved file. Click the Launch button to open the Occupancy Calculator.
The user interface consists of an input section as well as tables and graphs that display information about GPU occupancy. To use the calculator, change the input values in the input section, click the Apply button and examine the tables and graphs.
3.10.1. Tables
The tables show the occupancy, as well as the number of active threads, warps, and thread blocks per multiprocessor, and the maximum number of active blocks on the GPU.
3.11. Green Contexts support
NVIDIA Nsight Compute provides a number of features to ease the analysis of applications using
CUDA Green Contexts, see
the Kernel Profiling Guide for an overview about
profiling such applications.
To identify Green Context results quickly, the icon of the Attributes button in the Profiler Report Header
turns into a leaf for Green Context results. Clicking on this button navigates to the Launch Statistics section of the result (if available),
which contains additional metrics related to the Green Context (e.g., its ID and the number of SMs used).
Some of this information is also available from the tooltip when hovering over the Attributes button.
To distinguish between Green-Contexts attributable and non-attributable metrics when navigating the Details page, Green Context markers
may be used. These can be toggled via the View menu of the Report Header and adds a leaf icon
behind attributable metrics in all section headers. When hovering over attributable metrics, the tooltip also shows a hint about the
scaling of such metrics.
Additionally, the Metrics Details tool window shows an Attribution row for metrics of Green Context results.
When profiling Green Context applications interactively, the Resources tool window also allows to track
the state of all Green Contexts, and their associated streams.
3.12. Acceleration Structure Viewer
The Acceleration Structure Viewer allows inspection of acceleration structures built using the OptiX API. In modern ray tracing APIs like OptiX, acceleration structures are data structures describing the rendered scene’s geometries that will be intersected when performing ray tracing operations. More information concerning acceleration structures can be found in the OptiX programming guide.
It is the responsibility of the user to set these up and pass them to the OptiX API which translates them to internal data structures that perform well on modern GPUs. The description created by the user can be very error-prone and it is sometimes hard to understand why the rendered result is not as expected. The Acceleration Structure Viewer is a component allowing OptiX users to inspect the acceleration structures they build before launching a ray tracing pipeline.
The Acceleration Structure Viewer is opened through a button in the Resources window. The button will only be available when the currently viewed resource is OptiX: TraversableHandles. It opens the currently selected handle.
The viewer is multi-paned: it shows a hierarchical view of the acceleration structure on the left, a graphical view of the acceleration structure in the middle, and controls and options on the right. In the hierarchical tree view on the left of the viewer the instance acceleration structures (IAS), geometry acceleration structures (GAS), child instances and child geometries are shown. In addition to this, some general properties for each of them is shown such as their primitive count, surface area and size on the device.
In the hierarchical view on the left of the Acceleration Structure Viewer, the following information is displayed where applicable.
An identifier for each row in the hierarchy. Click on the check box next to the name to show or hide the selected geometry or hierarchy. Double-click on this entry to jump to the item in the rendering view.
# Prims
The number of primitives that make up this acceleration structure.
Surface Area
A calculation of the total surface area for the AABB that bounds the particular entry.
The size of the output buffer on the device holding this acceleration structure.
Performance analysis tools are accessible in the bottom left corner on the main view. These tools help identify potential performance problems that are outlined in the RTX Ray Tracing Best Practices Guide. These analysis tools aim to give a broad picture of acceleration structures that may exhibit sub-optimal performance. To find the most optimal solution, profiling and experimentation is recommended but these tools may paint a better picture as to why one structure performs poorly compared to another.
Instance Overlaps
Identifies instance AABBs that overlap with other instances in 3D. Consider merging GASes when instance world-space AABBs overlap significantly to potentially increase performance.
Instance Heatmap
This allows you to set the threshold used by the AABB heatmap rendered in the visualizer.
3.12.1. Navigation
The Acceleration Structure Viewer supports multiple navigation modes. The navigation mode can be changed using the combo box in the camera controls pane, to the right of the rendering pane. The keyboard and mouse bindings for each mode are as follows:
Move forward, backward, left, right
Move forward, backward, left, right
Track (Move up, down, left, right)
Move up/down
Move up/down
Increase/decrease field of view
Increase/decrease field of view
Increase/decrease field of view
Shift/Ctrl
Move faster/slower
Move faster/slower
Move faster/slower
Mousewheel
Zoom in/out
Zoom in/out
Zoom in/out
LMB + Drag
Rotate in place
Rotate left/right, move forward/backward
Rotate around the geometry
RMB + Drag
Zoom in/out
Rotate in place
Zoom in/out
MMB + Drag
Track (Move up, down, left, right)
Track (Move up, down, left, right)
Track (Move up, down, left, right)
Temporarily switch to Orbit Camera
Temporarily switch to Orbit Camera
F/Double Click
Focus on the selected geometry
Focus on the selected geometry
Focus on the selected geometry
Based on the coordinate system of the input geometry, you may need to change the Up Direction setting to Z-Axis or the Coordinates setting to RHS. To reset the camera to its original location, click Reset Camera.
There are also a selection of Camera Controls for fast and precise navigation. To save a position, use the bookmarks controls. Each node within the acceleration structure hierarchy can also be double-clicked to quickly navigate to that location.
3.12.2. Filtering and Highlighting
The acceleration structure view supports acceleration structure filtering as well as highlighting of data matching particular characteristics. The checkboxes next to each geometry allow users to toggle the rendering of each traversable.
Geometry instances can also be selected by clicking on them in the main graphical view. Additionally, right clicking in the main graphical view gives options to hide or show all geometry, hide the selected geometry, or hide all but the selected geometry.
Beyond filtering, the view also supports highlight-based identification of geometry specified with particular flags. Checking each highlight option will identify those resources matching that flag, colorizing for easy identification. Clicking an entry in this section will dim all geometry that does not meet the filter criteria allowing items that match the filter to standout. Selecting multiple filters requires the passing geometry to meet all selected filters (e.g., AND logic). Additionally, the heading text will be updated to reflect the number of items that meet this filter criteria.
3.12.4. Exporting
The data displayed in the acceleration structure viewer document can be saved to file. Exporting an Acceleration Structure Viewer document allows for persisting the data you have collected beyond the immediate analysis session. This capability is particularly valuable for comparing different revisions of your geometry or sharing with others. Bookmarks are persisted as well.
3.13. Options
NVIDIA Nsight Compute options can be accessed via the main menu under Tools > Options. All options are persisted on disk and available the next time NVIDIA Nsight Compute is launched. When an option is changed from its default setting, its label will become bold. You can use the Restore Defaults button to restore all options to their default values.
3.13.1. Profile
Sections Directory
Directory from which to import section files and rules. Relative paths are with respect to the NVIDIA Nsight Compute installation directory.
Include Sub-Directories
Recursively include section files and rules from sub-directories.
Yes (Default)/No
Apply Applicable Rules Automatically
Automatically apply active and applicable rules.
Yes (Default)/No
Reload Rules Before Applying
Force a rule reload before applying the rule to ensure changes in the rule script are recognized.
Yes/No (Default)
Default Report Page
The report page to show when a report is generated or opened. Auto lets the tool decide the best page to show when opening a report.
Session
Summary
Details
Source
Comments
Call Stack/NVTX
Auto (default)
Function Name Mode
Determines how function/kernel names are shown.
Auto (default): each component uses its preferred mode
Demangled: kernel names are shown demangled with all parameters
Function: kernel names are shown with their demangled function name without parameters
Mangled: kernel names are shown with their mangled name, if applicable
NVTX Rename Mode
Determines how NVTX information is used for renaming. Range replay results are always renamed when possible.
None: no renaming
Kernel: kernel names are renamed using the most recent enclosing push/pop range
Resources (default): resources like CPU threads or CUDA contexts and streams are renamed
All: Kernel and Resources
Show Metrics Aggregation
Show aggregate of all results per each counter metric in the table header. Also, show aggregated value of all randomly selected metrics from summary or raw page table in the bottom-right label.
Yes (Default)/No
Rename Demangled Names
Perform auto-simplification on kernel demangled names or import renamed names from a configuration file. See more unique keywords in kernel demangled names first while expanding the column.
Yes (Default)/No
Rename Kernels Config Path
Use a configuration file to rename multiple demangled names. Also, export demangled names from the report to the specified file with mappings for renaming them.
ncu-kernel-renames.yaml
Maximum Baseline Name Length
The maximum length of baseline names.
1..N (Default: 40)
Number of Full Baselines to Display
Number of baselines to display in the report header with all details in addition to the current result or the baseline added for the current result
0..N (Default: 2)
Auto-Convert Metric Units
Auto-adjust displayed metric units and values (e.g. Bytes to KBytes).
Yes (Default)/No
Show Instanced Metric Values
Show the individual values of instanced metrics in tables.
Yes/No (Default)
Show Metrics As Floating Point
Show all numeric metrics as floating-point numbers.
Yes/No (Default)
Show Knowledge Base Information
Show information from the knowledge base in (metric) tooltips to explain terminology. Note: Nsight Compute needs to be restarted for this option to take effect.
Yes (Default)/No
Metrics/Properties
List of metrics and properties to show on the summary page. Comma-separated list of metric entries. Each entry has the format {Label:MetricName}.
Delay Load ‘Source’ Page
Delays loading the content of the report page until the page becomes visible. Avoids processing costs and memory overhead until the report page is opened.
Yes/No (Default)
Show Single File For Multi-File Sources
Shows a single file in each Source page view, even for multi-file sources.
Yes/No (Default)
Show Only Executed Functions
Shows only executed functions in the source page views. Disabling this can impact performance.
Yes (Default)/No
Auto-Resolve Remote Source Files
Automatically try to resolve remote source files on the source page (e.g. via SSH) if the connection is still registered.
Yes/No (Default)
Enable Register Dependencies
Track dependencies between SASS registers/predicates and display them in the SASS view.
Yes (Default)/No
Kernel Analysis Size Threshold (KB)
Enable SASS flow graph analysis for functions below this threshold. SASS analysis is required for Live Register and Register Dependency information. Set to -1 to enable analysis for all functions.
-1..N (Default: 1024)
Enable ELF Verification
Enable ELF (cubin) verification to run every time before SASS analysis. This should only be enabled when working with applications compiled before CUDA 11.0 or when encountering source page issues.
Yes/No (Default)
API Call History
Number of recent API calls shown in API Stream View.
1..N (Default: 100)
Mixed DPI Scaling
Disable Mixed DPI Scaling if unwanted artifacts are detected when using monitors with different DPIs.
Auto (Default)
Default Document Folder
Directory where documents unassociated with a project will be saved.
At Startup
What to do when NVIDIA Nsight Compute is launched.
Show welcome page (Default)
Show quick launch dialog
Load last project
Show empty environment
Show version update notifications
Show notifications when a new version of this product is available.
Yes (Default)
3.13.3. Connection
Connection properties are grouped into Target Connection Options and Host Connection Properties.
Target Connection Properties
The Target Connection Properties determine how the host connects to the target application during an Interactive Profile Activity. This connection is used to transfer profile information to the host during the profile session.
Base Port
Base port used to establish a connection from the host to the target application during an Interactive Profile activity (both local and remote).
1-65535 (Default: 49152)
Maximum Ports
Maximum number of ports to try (starting from Base Port) when attempting to connect to the target application.
2-65534 (Default: 64)
Host Connection Properties
The Host Connection Properties determine how the command line profiler will connect to the host application during a Profile Activity. This connection is used to transfer profile information to the host during the profile session.
Base Port
Base port used to establish a connection from the command line profiler to the host application during a Profile activity (both local and remote).
1-65535 (Default: 50152)
Maximum Ports
Maximum number of ports to try (starting from Base Port) when attempting to connect to the host application.
1-100 (Default: 10)
Program Source Locations
Set program source search paths. These paths are used to resolve CUDA-C source files on the Source page if the respective file cannot be found in its original location. Files which cannot be found are marked with a File Not Found error. See the Ignore File Properties option for files that are found but don’t match.
Ignore File Properties
Ignore file properties (e.g. timestamp, size) for source resolution. If this is disabled, all file properties like modification timestamp and file size are checked against the information stored by the compiler in the application during compilation. If a file with the same name exists on a source lookup path, but not all properties match, it won’t be used for resolution (and a File Mismatch error will be shown).
Yes/No (Default)
Collect Usage and Platform Data
Choose whether or not you wish to allow NVIDIA Nsight Compute to collect usage and platform data.
No (Default)
3.14. Projects
NVIDIA Nsight Compute uses Project Files to group and organize profiling reports. At any given time, only one project can be open in NVIDIA Nsight Compute. Collected reports are automatically assigned to the current project. Reports stored on disk can be assigned to a project at any time. In addition to profiling reports, related files such as notes or source code can be associated with the project for future reference.
Note that only references to reports or other files are saved in the project file. Those references can become invalid, for example when associated files are deleted, removed or not available on the current system, in case the project file was moved itself.
NVIDIA Nsight Compute uses the ncu-proj
file extension for project files.
When no custom project is current, a default project is used to store e.g. the current Start Activity Dialog entries. To remove all information from the default project, you must close NVIDIA Nsight Compute and then delete the file from disk.
On Windows, the file is located at <USER>\AppData\Local\NVIDIA Corporation\NVIDIA Nsight Compute\
On Linux, the file is located at <USER>/.local/share/NVIDIA Corporation/NVIDIA Nsight Compute/
On MacOSX, the file is located at <USER>/Library/Application Support/NVIDIA Corporation/NVIDIA Nsight Compute/
3.14.1. Project Dialogs
New Project
Creates a new project. The project must be given a name, which will also be used for the project file. You can select the location where the project file should be saved on disk. Select whether a new directory with the project name should be created in that location.
3.14.2. Project Explorer
The Project Explorer window allows you to inspect and manage the current project. It shows the project name as well as all Items (profile reports and other files) associated with it. Right-click on any entry to see further actions, such as adding, removing or grouping items. Type in the Search project toolbar at the top to filter the currently shown entries.
3.15. Visual Profiler Transition Guide
This guide provides tips for moving from Visual Profiler to NVIDIA Nsight Compute. NVIDIA Nsight Compute tries to provide as much parity as possible with Visual Profiler’s kernel profiling features, but some functionality is now covered by different tools.
3.15.1. Trace
NVIDIA Nsight Compute does not support tracing GPU or API activities on an accurate timeline. This functionality is covered by NVIDIA Nsight Systems. In the Interactive Profile Activity, the API Stream tool window provides a stream of recent API calls on each thread. However, since all tracked API calls are serialized by default, it does not collect accurate timestamps.
3.15.2. Sessions
Instead of sessions, NVIDIA Nsight Compute uses Projects to launch and gather connection details and collected reports.
Executable and Import Sessions
Use the Project Explorer or the Main Menu to create a new project. Reports collected from the command line, i.e. using NVIDIA Nsight Compute CLI, can be opened directly using the main menu. In addition, you can use the Project Explorer to associate existing reports as well as any other artifacts such as executables, notes, etc., with the project. Note that those associations are only references; in other words, moving or deleting the project file on disk will not update its artifacts.
nvprof or command-line profiler output files, as well as Visual Profiler sessions, cannot be imported into NVIDIA Nsight Compute.
3.15.3. Timeline
Since trace analysis is now covered by Nsight Systems, NVIDIA Nsight Compute does not provide views of the application timeline. The API Stream tool window does show a per-thread stream of the last captured CUDA API calls. However, those are serialized and do not maintain runtime concurrency or provide accurate timing information.
3.15.4. Analysis
Guided Analysis
All trace-based analysis is now covered by NVIDIA Nsight Systems. This means that NVIDIA Nsight Compute does not include analysis regarding concurrent CUDA streams or (for example) UVM events. For per-kernel analysis, NVIDIA Nsight Compute provides recommendations based on collected performance data on the Details Page. These rules currently require you to collect the required metrics via their sections up front, and do not support partial on-demand profiling.
To use the rule-based recommendations, enable the respective rules in the Metric Selection. Before profiling, enable Apply Rules in the Profile Options, or click the Apply Rules button in the report afterward.
Unguided Analysis
All trace-based analysis is now covered by Nsight Systems. For per-kernel analysis, Python-based rules provide analysis and recommendations. See Guided Analysis above for more details.
PC Sampling View
Source-correlated PC sampling information can now be viewed in the Source Page. Aggregated warp states are shown on the Details Page in the Warp State Statistics section.
Memory Statistics
Memory Statistics are located on the Details Page. Enable the Memory Workload Analysis sections to collect the respective information.
NVLink View
NVLink topology diagram and NVLink property table are located on the Details Page. Enable the NVLink Topology and NVLink Table sections to collect the respective information.
Refer to the Known Issues section for the limitations related to NVLink.
Source-Disassembly View
Source correlated with PTX and SASS disassembly is shown on the Source Page. Which information is available depends on your application’s compilation/JIT flags.
GPU Details View
NVIDIA Nsight Compute does not automatically collect data for each executed kernel, and it does not collect any data for device-side memory copies. Summary information for all profiled kernel launches is shown on the Summary Page. Comprehensive information on all collected metrics for all profiled kernel launches is shown on the Raw Page.
CPU Details View
CPU callstack sampling is now covered by NVIDIA Nsight Systems.
OpenACC Details View
OpenACC performance analysis with NVIDIA Nsight Compute is available to limited extent. OpenACC parallel regions are not explicitly recognized, but CUDA kernels generated by the OpenACC compiler can be profiled as regular CUDA kernels. See the NVIDIA Nsight Systems release notes to check its latest support status.
OpenMP Details View
OpenMP performance analysis is not supported by NVIDIA Nsight Compute. See the NVIDIA Nsight Systems release notes to check its latest support status.
Properties View
NVIDIA Nsight Compute does not collect CUDA API and GPU activities and their properties. Performance data for profiled kernel launches is reported (for example) on the Details Page.
Console View
NVIDIA Nsight Compute does not currently collect stdout/stderr application output.
Settings View
Application launch settings are specified in the Start Activity Dialog. For reports collected from the UI, launch settings can be inspected on the Session Page after profiling.
CPU Source View
Source for CPU-only APIs is not available. Source for profiled GPU kernel launches is shown on the Source Page.
3.15.5. Command Line Arguments
Please execute ncu-ui with the -h
parameter within a shell window to see the currently supported command line arguments for the NVIDIA Nsight Compute UI.
To open a collected profile report with ncu-ui, simply pass the path to the report file as a parameter to the shell command.
3.16. Visual Studio Integration Guide
This guide provides information on using NVIDIA Nsight Compute within Microsoft Visual Studio, using the NVIDIA Nsight Integration Visual Studio extension, allowing for a seamless development workflow.
3.16.1. Visual Studio Integration Overview
NVIDIA Nsight Integration is a Visual Studio extension that allows you to access the power of NVIDIA Nsight Compute from within Visual Studio.
When NVIDIA Nsight Compute is installed along with NVIDIA Nsight Integration, NVIDIA Nsight Compute activities will appear under the NVIDIA ‘Nsight’ menu in the Visual Studio menu bar. These activities launch NVIDIA Nsight Compute with the current project settings and executable.
For more information about using NVIDIA Nsight Compute from within Visual Studio, please visit
Notices
Notices
ALL NVIDIA DESIGN SPECIFICATIONS, REFERENCE BOARDS, FILES, DRAWINGS, DIAGNOSTICS, LISTS, AND OTHER DOCUMENTS (TOGETHER AND SEPARATELY, “MATERIALS”) ARE BEING PROVIDED “AS IS.” NVIDIA MAKES NO WARRANTIES, EXPRESSED, IMPLIED, STATUTORY, OR OTHERWISE WITH RESPECT TO THE MATERIALS, AND EXPRESSLY DISCLAIMS ALL IMPLIED WARRANTIES OF NONINFRINGEMENT, MERCHANTABILITY, AND FITNESS FOR A PARTICULAR PURPOSE.
Information furnished is believed to be accurate and reliable. However, NVIDIA Corporation assumes no responsibility for the consequences of use of such information or for any infringement of patents or other rights of third parties that may result from its use. No license is granted by implication of otherwise under any patent rights of NVIDIA Corporation. Specifications mentioned in this publication are subject to change without notice. This publication supersedes and replaces all other information previously supplied. NVIDIA Corporation products are not authorized as critical components in life support devices or systems without express written approval of NVIDIA Corporation.
Trademarks
NVIDIA and the NVIDIA logo are trademarks or registered trademarks of NVIDIA Corporation in the U.S. and other countries. Other company and product names may be trademarks of the respective companies with which they are associated.
大气的木耳 · Querydsl 参考指南 - 南石 - 博客园 6 天前 |
儒雅的手电筒 · 什么是:局部异常因子 (LOF) - 轻松学习统计学 2 周前 |
细心的毛巾 · Windows中从Git Bash运行jq 1 月前 |