Versions Compared
Key
- This line was added.
- This line was removed.
- Formatting was changed.
Overview
The following components are available in the 1.0 release:
- Canvas Extractor
- Google Classroom Extractor
- Schoology Extractor
- LMS Data Store Loader
Please see LMS Toolkit for more information about the purpose of these tools.
Note |
---|
The LMS Data Store Loader pushes CSV files, created by the extractors, into a SQL Server database. That database can be the same as an Ed-Fi ODS. However, all of the data are loaded into tables in the |
Pre-Requisites
Warning |
---|
Python 3.9.5 has a bug that causes the extractors to crash, and thus should not be used. The Alliance's testing has used 3.9.4. |
Note | ||
---|---|---|
| ||
In practice, these tools have only been tested on Windows 10; however, these tools should work from any operating system that supports Python 3.9. |
Running the Tools
The LMS Toolkit components can be installed into other Python scripts as dependencies, or they can run as stand-alone command line scripts from the source code.
Deck | |||||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| |||||||||||||||||||||
|
Runtime Arguments and Options
Whether you run the extractors by incorporating into an existing Python package, or by using the stand-alone command line utility from the source repository, there are a number of required and optional arguments. When running with the command line tool, simply provide the --help
option for the full set of options for each extractor.
Argument | Required? | Applies To | Purpose |
---|---|---|---|
Feature | No | All | Define which optional features are to be retrieved from the upstream system. Default: none. Available features:
Note: Sections, Section Associations, and Users are always pulled from the Source System. |
Log Level | No | All | Valid options are: DEBUG, INFO (default), WARNING, ERROR, CRITICAL |
Output Directory | No | All | The output directory for the generated CSV files. Defaults to: ./data . |
Sync database directory | No | All | Directory for storing a SQLite database that is used in support of synchronizing the data between successive executions of the tool. Defaults to: ./data . |
Classroom account | Yes, for Google | Google Classroom | The email address of the Google Classroom admin account. |
Usage start date | No | Google Classroom | Start date for usage data pull in YYYY-MM-DD format. |
Usage end date | No | Google Classroom | End date for usage data pull in YYYY-MM-DD format. |
Tip | ||
---|---|---|
To retrieve multiple features with one call to the command line interface, list them out with spaces separating the values or commas. Examples:
|
Using Extractor Output
The LMS Data Store Loader pushes the extractor-created CSV files into a SQL Server database, where the data are available for use via standard SQL Server interfaces and tools. However, the CSV files can also be consumed directly to perform many interesting analyses. We have a developed a set of Jupyter notebooks that demonstrate analytics tasks that can be performed in Python using the Pandas framework, reading raw CSV files. Sample output from these notebooks is visible directly in GitHub, without needing to run the code locally:
- Filesystem Tutorial / In Danger of Failing / Missing Assignment Submissions: how to use the LMS Toolkit scripts to understand and access output files created by the extractors. Also includes two analysis scenarios - looking for students who are in danger of failing, and looking for missing assignment submissions.
- Record Counts: simply accesses all of the extracted files and provides summary count of records downloaded.
- Student Logins: simple visualization showing frequency of student logins to the LMS.
- Student Submissions: shows the count of assignments submitted per student, by status.
Operational Concerns
Logging
Deck | ||||||||||||||||||||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| ||||||||||||||||||||||||||||||||||||
|
Security
Upstream APIs
Each API has its own process for securing access. Please see the respective readme files for more information:
Data Storage
Given the LMS Toolkit deals with student data, both the filesystem and database (if uploading to SQL Server) are subject to all of the same access restrictions as the Ed-Fi ODS database.
Database Permissions
The LMS Data Store Loader tool manages its own database tables. Thus the first time you run the tool, the credentials used to connect to SQL Server need to have the db_ddladmin
permission in order to create the necessary tables. Subsequent executions can use an account with more restrictive permissions, i.e. the db_datawriter
role.
Scheduling
The API's provided by these three learning management systems are well defined at a granular level. From a performance perspective, this means that the process of getting a complete set of data is very chatty and may take a long time to process. It is difficult to predict the exact impact, although generally the time will scale proportional to the number of course sections. Some of the API's also do not have any mechanism for restricting the date range or looking for changed data, resulting in each execution of the extractor re-pulling the entire data set.
If running on a daily basis, then we recommend running after normal school hours to minimize contention with network traffic to the source system. If running weekly, then it may be best to run over the weekend.
It should be trivial to call these programs from Windows Task Scheduler, Linux chron, or a workflow engine such as Apache Airflow.
Contents
Table of Contents |
---|