CHARTING & REPORTING KS RESULTS DATA
There are a range of charting and management reporting options available in your KS dashboard.
Search & group results
The first task is to filter the raw results data into more manageable blocks. There are a range of searching and grouping options available, including; score ranges, date ranges, elapsed time, test topic, user names, training keywords, plus 5 additional custom datafields.
This chart displays group results in performance quartiles. The upper left quadrant (Q1) contains results, where the users completed their assessment accurately and fast.
The upper right quadrant (Q2) shows results which have high accuracy, but slower completion times. Bottom left (Q3) represents lower scores, but in a fast time.
Lastly, the bottom right quartile (Q4) shows test scores which are inaccurate and slow.
This chart highlights training topics for a given group of test results. The logic analyzes all of the results for a group, references the training tags assigned to questions presented during a test and lists those tags in priority order.
Red indicates the tasks which have been answered incorrectly by most people in a given group. Orange is the next highest priority, followed by Yellow; green training tags are the topics which have been answered correctly by most of the group, so represent the least urgent issues.
For example; 10 people are presented with one or more questions, which include the tag, ‘Lines’. If 7 of the 10 people answer one or more ‘Lines’ questions incorrectly, then the keyword, ‘Lines’, will be flagged at 70% and appear in Red on the chart.
This chart looks at how each individual question in a library has performed, in any given test.
The logic analyzes all of the results for each test and presents an aggregate percentage score and time for each question, divided by the total number of results for that test on the account.
For example; 10 people answer a question called ‘Lines’. If 7 of the 10 people answer the question 100% correctly, 1 person scores 50% and the other 2 score 0%, then the question will score an aggregate of 75% and appear in Yellow on the chart.
This chart displays user performance for any given group, in descending order. The X-axis shows the % score attained and the Y-axis displays user names.
This chart allows firms to compare performance, from one group to another, across (up to) 9 sets of data at a time. Group vs group comparisons can be used to compare a range of results data.
For example; pre and post-training performance, different project teams, offices from different geographic locations, data representing different job titles or industry disciplines, in-house data vs interview candidates, and so on.
This chart allows AEC firms to compare in-house performance against a wider set of anonymous industry aggregate results data, for selected tests.
Results are currently available for: AutoCAD 2D, MicroStation 2D, Revit Architecture, Revit MEP and Revit Structure. KS plans to publish additional benchmark figures in future.
Use the csv export tools, to capture KS results for further analysis in Excel.