This article will provide an overview of the components found in the DataCamp Course Maintenance Dashboard, and what they can tell you about how your course is performing. 

Accessing the dashboard

  1. Log into DataCamp.
  2. Open the Content Dashboard.

Locating your course 

Use the left-hand navigation bar to select Quality > Courses.

Part of the Content Dashboard left-hand navigation pane, showing the location of the

The dashboard should open to the "View Course" pane. If not, navigate to it now.

The pane navigation element of the Content Dashboard.

Use the “Select Course” dropdown to get the course you are interested in. You can filter the results by typing in the textbox underneath the dropdown.

The

By default, course metrics are calculated on the last 8 weeks of data. You can change this via the "Select date range" dropdown in the left-hand navigation bar.

The

Dashboard overview

An overview of the top of the

1. Pane selector

Choose the tool to view. See the end of this article for descriptions of panes other than "View Course".

2. Course Selector

Choose the course to view.

Go to the course on GitHub, Teach, or Campus.

4. Learner categories

Filter results by the various categories of learner.

5. Course-level metrics

Performance metrics for the whole course. To interpret these, see Course metric expectations.

6. Metric time series

A times series of the selected metric.

Chapter-level metrics

Performance metrics are calculated by chapter. This is particularly useful to see if overall course ratings are affected by one chapter with problems. To interpret these, see Chapter-level quality metrics.

The chapter-level metrics table in the

Exercise-level metrics

Next, performance metrics are calculated by exercise. This let's you hone in on specific problems. To interpret these, see Exercise-level quality metrics.

The exercise-level metrics table in the

Issues / Feedback / Incorrect Attempts 

The next table shows further detail about possible problems in the exercise that was selected in the exercise-level metrics table. To interpret this table, see Interpreting Issues/Feedback and Interpreting Code Diffs.

The Issues metrics table in the


Exercise contents

The final table shows the exercise that was selected in the exercise-level metrics table, broken down into components. This can be useful for reference.

The exercise contents panel in the

Other dashboard panes

Browse Courses

This pane lets you sort and filter to compare between all courses you have access to.

The

Browse Exercises

This pane lets you sort and filter to compare between exercises in all courses you have access to.

The

Browse Issues

This pane lets you view all the feedback for courses you have access to in one place. If you are doing regular maintenance, it is useful when sorted by issue creation date.

The

Browse feedback

This pane lets you view hint and SCT related feedback for all courses you have access to. Click the plus icon on the left to see the code the learner submitted before leaving feedback.

The

Browse Incorrect Attempts

This pane lets you view incorrect attempts by all users for all exercises in courses you have access to. This can help identify where learners are struggling to solve your exercises. Click the plus icon on the left to see the code the learner submitted.

The

Did this answer your question?