Results for the Photo Handling Assessment
Applies To: Windows 8.1
This topic provides guidance for understanding the results of the Photo handling assessment in addition to guidance on how to use those results to identify and resolve several common issues that negatively affect the customer experience during photo handling operations.
The Photo handling assessment measures the time that it takes to complete common photo-related tasks, such as viewing, searching and manipulating a set of photos. The purpose of this assessment is to enable comparison of photo handling experiences across computers and to diagnose performance issues that affect the photo handling experience. In interactive tasks such as photo viewing and manipulation, response times are very important for user satisfaction. Therefore, this assessment uses response times to measure the user experience.
The metrics generated by the assessment can be grouped as follows:
Top-level metrics, which provide an aggregation of various metrics related to responsiveness under the “Explorer Experience” group.
Resource consumption, which shows metrics that are related to consuming resources on the computer during the assessment.
Explorer Experience, which shows metrics that are related to actions in File Explorer.
Photo Manipulation, which shows metrics that are related to photo manipulation actions.
In this topic:
Goals File
Metrics
Issues
For more information about the system requirements, workloads and assessment settings, see Photo Handling.
Goals File
You can create custom goals to measure your improvements in the Results View. Goals files are a triage tool that can help you understand how a PC is performing and to compare PCs in your business.
For example, goals for a basic laptop might be different than the goals you set for a high end desktop computer, or market expectations might change in such a way that you want the flexibility to define different goals and key requirements as time passes and technology improves.
When a metric value is compared to the goal for that metric, the status is color coded in the Result View as follows:
Light purple means that the system has a great user experience and that there are no perceived problems.
Medium purple means that the user experience is tolerable and you can optimize the system. Review the recommendations and analysis to see what improvements can be made to the system. These can be software changes, configuration changes or hardware changes.
Dark purple means that the system has a poor user experience and that there is significant room for improvements. Review the recommendations and analysis to see the improvements that can be made to the system. These can be software changes, configuration changes or hardware changes. You might have to consider making tradeoffs to deliver a high quality Windows experience.
No color means that there are no goals defined for the metric.
Note
In the Windows Assessment Toolkit for Windows 8, some assessments include default goals files. The first time you view results using this version of the tools, the default goals file is used. However, you can also define custom goals for Windows 8 the same way that you can for Windows 8.1.
You can set the goals file location and add a goals file to that location before you can use the UI to apply the custom goals. Once a goals file is selected it will continue to be the goals file that is used for any results that are opened.
Only one goals file can be used at a time. Goals for all assessments are set in a single goals file. The assessment tools will search for goals in the following order:
A custom goals file
Goals that are defined in the results file
Goals that are defined in the assessment manifest
You can use the sample goals file that is provided at %PROGRAMFILES%\Windows Kits\8.1\Assessment and Deployment Kit\Windows Assessment Toolkit\SDK\Samples\Goals to create your own goals file.
Note
You cannot package a goals file with a job, but you can store it on a share for others to use.
Metrics
This section describes the key metrics reported by the Photo handling assessment, common causes of poor results for these metrics, and common remediation for these issues. This section also attempts to identify the audience that has the most influence on each of these metrics.
This section includes:
Responses without Delays
Noticeable Delays
Unsatisfactory Delays
Resource Consumption Metrics
Change to Large Icons View
Display Context Menu
Photo Manipulation Metrics
Responses without Delays
Most applicable to: OEM/ODM, IHV, ISV, IT Professionals, Enthusiasts
This metric gives the percentage of operational responses which are meeting or exceeding the default goal. This metric indicates the overall responsiveness of interactive actions. With the exception of the Display Search Results metric, all responsiveness metrics in the “Explorer Experience” group are used to calculate this metric.
Analysis and Remediation Steps
This metric is an aggregation of several sub-metrics. Analyze the individual sub-metrics which are not meeting the goals and those which are close to the goal. Because these sub-metric values are averages, there may be individual instances of not meeting the goal, but when the value is averaged the goal is met.
Noticeable Delays
Most applicable to: OEM/ODM, IHV, ISV, IT Professionals, Enthusiasts
This metric gives the percentage of operational responses with moderate delays based on the predefined goals. This indicates the responsiveness of interactive actions. With the exception of Display Search Result, all responsiveness metrics in the “Explorer Experience” group are used to calculate this metric.
Analysis and Remediation Steps
Because this metric is an aggregation of several sub-metrics, analyze the individual sub-metrics that have noticeable delays, and are marked as yellow in the results.
Unsatisfactory Delays
Most applicable to: OEM/ODM, IHV, ISV, IT Professionals, Enthusiasts
This metric gives the percentage of operational responses with long delays based on the predefined goals. This metric indicates the responsiveness of interactive actions. With the exception of Display Search Result, all responsiveness metrics in the “Explorer Experience” group are used to calculate this metric.
Analysis and Remediation Steps
Because this metric is an aggregation of several sub-metrics, analyze the individual sub-metrics that have unsatisfactory delays, and are marked as red in the results.
Resource Consumption Metrics
Most applicable to: OEM/ODM, IHV, ISV, IT Professionals, Enthusiasts
This metric provides information about the consumption of resources on the computer during the assessment and can provide useful pointers for possible issues in the system.
The sub-metrics and remediation steps are outlined in the following table.
Metric | Description | Recommendation |
---|---|---|
Average CPU Utilization |
Shows average CPU use, as a percentage, across all processors (cores or CPUs) during the assessment. |
Choose the WPA in-depth analysis link to use the WPA Computation graphs: CPU Usage (Sampled) or CPU Usage (Precise). |
Hard Fault Count |
Shows the total number of hard faults. Hard faults occur when the operating system retrieves memory pages from disk instead of from the in-memory pages that the memory manager maintains. |
Choose the WPA in-depth analysis link to use the WPA Hard Faults Memory graph. |
Average Disk Utilization |
Shows average disk use, as a percentage, during the assessment. The disk utilization percentage is computed by calculating the amount of time that a disk is servicing I/O requests vs. the time that is spent idling. |
Choose the WPA in-depth analysis link to use the WPA Disk Usage graphs. |
Long running DPC/ISR |
Shows the total time, in microseconds, for deferred procedure calls (DPC) and interrupt service requests (ISR) that continue longer than 3 microseconds. |
If this metric value is larger than 0, the assessment generates an issues which can be analyzed further in WPA. For more information, see Common In-Depth Analysis Issues. |
Change to Large Icons View
Most applicable to: OEM, ISV
This metric measures the average time that it takes to change the view to large icons. This is a responsiveness metric that is measured against specific goals.
Typical Influencing Factors
Changing the icon view on a folder that contains photos involves opening the image files, decoding the images and rendering the icons. Therefore, factors influencing I/O time and graphics rendering speed are relevant. For example, anti-virus software might influence the I/O times and affect this operation.
Analysis and Remediation Steps
The Photo Handling assessment analyzes the trace file to generate issues affecting the assessment activities. Open the assessment trace file in WPA and analyze the activity corresponding to this metric. In addition to reporting issues for long-running ISRs and DPCs, the assessment will also report issues caused by delays (pre-emptions by other threads) which increase the duration of the activity. For more information, see Common In-Depth Analysis Issues.
Similar Metrics
Similar analysis and remediation steps can be applied to the following related metrics.
Metric | Description |
---|---|
Change to Medium Icons View |
Shows the average time, in milliseconds, to change the view to medium icons. |
Change to Extra Large Icons View |
Shows the average time, in milliseconds, to change the view to extra-large icons. |
Change to Details View |
Shows the average time, in milliseconds, to change the view to details. |
Change to Sort by Date |
Shows the average time, in milliseconds, to change the sort order to sort by date. |
Change Sort Order |
Shows the average time, in milliseconds, to change the sorting order from ascending to descending. |
Scroll Page |
Shows the average time for scrolling a page in File Explorer (showing a folder that has photos), by pressing the Page Up and Page Down keys. |
Navigate |
Shows the total time, in milliseconds, for folder navigation, including choosing the Back button. |
Display Context Menu
Most applicable to: OEM, ISV, IT Professionals, Enthusiasts
This metric measures the average time, in milliseconds, to display the shortcut menu after a right-click inside a folder in File Explorer. This is a responsiveness metric that is measured against specific goals.
Typical Influencing Factors
The action of displaying a context menu should be responsive (<250 milliseconds) even on low end computer. The main factor that could negatively affect this metric is resource consumption, typically CPU, resulting in File Explorer being unresponsive.
Analysis and Remediation Steps
The Photo Handling assessment analyzes the trace file to generate issues that affect the assessment activities. The name of the activity corresponding to this metric are Display View Context Sub-Menu and Display Sort Context Sub-Menu. Open the assessment trace file in WPA and analyze the activity corresponding to this metric.
In addition to analyzing generated issues, you can also analyze the CPU usage graphs in WPA to find processes with high CPU utilization. If these processes are found to affect the photo usage activities on the system, consider fixing these programs or (if non critical) stopping them during photo usage activities.
Similar Metrics
Similar analysis and remediation steps can be applied to the following related metrics.
Metric | Description |
---|---|
Display View Context Sub-Menu |
Shows the average time, in milliseconds, to display the View context submenu. This submenu appears in File Explorer after you right-click and then choose View. |
Display Sort Context Sub-Menu |
Shows the average time, in milliseconds, to display a Sort context submenu. This submenu appears in File Explorer after you right-click and then choose Sort. |
Acquire Search Box Focus |
Shows the average time, in milliseconds, to acquire focus on the search box after you choose the search box. |
Photo Manipulation Metrics
Most applicable to: OEM, ODM, IHV, ISV
This group of metrics reports time for performing common photo manipulation actions such as zooming, cropping and rotating a picture using Windows Imaging Component (WIC) API.
Sub-metrics are described in the following table.
Metric name | Description |
---|---|
Photo Opening Time |
Shows the average time, in milliseconds, to open a single photo. |
Photo Saving Time |
Shows the average time, in milliseconds, to save changes that were made to a single photo. |
Photo Zooming Time |
Shows the average time, in milliseconds, to scale a photo to one-quarter of its original size. This is the equivalent of zooming in on a photo. |
Photo Cropping Time |
Shows the average time, in milliseconds, to crop a photo to one-quarter of its original size. |
Photo Rotating Time |
Shows the average time, in milliseconds, to rotate a photo 90 degrees clockwise. |
Typical Influencing Factors
Unlike the Responsiveness metrics, these photo manipulation metrics are expected to scale with the hardware configuration of the system. So the assessment does not apply responsiveness metric goals to these metric values. The durations of these operations may be affected by bottlenecks in CPU and Disk.
Analysis and Remediation Steps
To analyze issues generated for the corresponding activity of this metric, choose the WPA in-depth analysis link in the issue description. Analyze any delays affecting these metrics. In addition to the delay analysis, CPU and Disk Usage summary tables can be analyzed in WPA looking for high resource consumption processes. If these processes are found to affect the photo usage activities on the system, consider fixing these programs or (if non critical) stopping them during photo usage activities.
Issues
This assessment performs advanced issue analysis and provides links to Windows® Performance Analyzer (WPA) to troubleshoot the issues that are identified. In most cases, you can choose the WPA Analysis link to troubleshoot the issues that appear. When WPA opens additional details about disk activity or CPU activity might be available depending on the type of issue identified. For more information about in-depth analysis issues and recommendations, see Common In-Depth Analysis Issues.
The assessment reports an exit code of 0x80050006
This error occurs when maintenance tasks have been registered on the PC but have not completed before the assessment run. This prevents the assessment from running, as maintenance tasks often impact assessment metrics.
To resolve this issue, do one of the following:
Ensure that the computer is connected to a network and is running on AC power. Manually initiate pending maintenance tasks with the following command from an elevated prompt:
rundll32.exe advapi32.dll,ProcessIdleTasks
Disable regular and idle maintenance tasks, and stop all maintenance tasks before running the assessment.
See Also
Tasks
Create and Run an Energy Efficiency Job
Concepts
Photo Handling
Assessments
File Handling
Other Resources
Windows Assessment Toolkit Technical Reference
Windows Assessments Console Step-by-Step Guide
Windows Performance Toolkit Technical Reference