This article provides an overview of the components of the DataCamp Course Maintenance Dashboard and what each can tell you about how your course is performing. 

Accessing the dashboard

  1. Log into DataCamp.

  2. Open the Content Dashboard.

Locating your course 

Use the left-hand navigation pane to select Quality > Courses.

Part of the Content Dashboard left-hand navigation pane, showing the location of the

The dashboard should open to the "View Course" pane. If not, navigate to it now.

The pane navigation element of the Content Dashboard.

If you've written more than one course, use the “Select Course” dropdown to get the course you are interested in.

By default, course metrics are calculated on the last 8 weeks of data. You can change this via the "Select date range" dropdown in the left-hand navigation bar.

The

Dashboard overview

An overview of the top of the

1. Pane selector

Choose the tool to view. See the end of this article for descriptions of panes other than "View Course".

2. Course selector

Choose the course to view.

Go to the course on GitHub, Teach, or Campus.

4. Learner categories

Filter results by the various categories of learners.

5. Course-level metrics

Performance metrics for the whole course. To interpret these, see Course metric expectations.

6. Metric time series

A times series of the selected metric.

Chapter-level metrics

Below the time series visualization, there are performance metrics calculated by chapter. This is particularly useful to see if overall course ratings are affected by one chapter with problems. To interpret these, see Chapter-level quality metrics.

The chapter-level metrics table in the

Exercise-level metrics

Below the chapter-level metrics, there are performance metrics are calculated by exercise. This table lets you hone in on specific problems. To interpret these, see Exercise-level quality metrics.

The exercise-level metrics table in the

Issues / Feedback / Incorrect Attempts 

After selecting an exercise in the exercise-level metrics table, scroll down to get more information about possible problems in that exercise. To interpret this table, see Interpreting Issues/Feedback and Interpreting Code Diffs.

The Issues metrics table in the


Exercise contents

The final table shows the exercise that was selected in the exercise-level metrics table, broken down into components. This can be useful for reference.

The exercise contents panel in the

Other dashboard panes

Browse Courses

This pane lets you sort and filter to compare between all courses you've created.

The

Browse Exercises

This pane lets you sort and filter to compare between exercises in all courses you've created.

The

Browse Issues

This pane lets you view all the feedback for all courses you've created in one place. When performing regular maintenance, it is useful to sort them by the "Created" date.

The

Browse feedback

This pane lets you view hint and SCT-related feedback for all courses you've created. Click the plus icon on the left to see the code the learner submitted before leaving feedback.

The

Browse Incorrect Attempts

This pane lets you view incorrect attempts by all users for all exercises all courses you've created. This can help identify where learners are struggling to solve your exercises. Click the plus icon on the left to see the code the learner submitted.

The

Did this answer your question?