Compare commits

...

3 Commits

Author SHA1 Message Date
3e677d9ddd added typing & more code cleanup 2023-03-03 13:13:28 +00:00
2b6fe45b42 READMEs update 2023-03-03 13:11:17 +00:00
964a1f6abb added TODO to .gitignore 2023-03-03 13:09:49 +00:00
7 changed files with 50 additions and 36 deletions

2
.gitignore vendored
View File

@@ -129,7 +129,7 @@ dmypy.json
.pyre/
# BBGradebookOrganiser
TODO
BB_gradebooks/
BB_submissions/
csv/

View File

@@ -1,6 +1,7 @@
# BBGradebookOrganiser
Blackboard Gradebook Organiser
### Blackboard gradebooks directory: *BB_gradebooks*
## Blackboard gradebooks directory: *BB_gradebooks*
Create a directory with an appropriate name for the gradebook / assignment in this directory, and extract the downloaded gradebook .zip file in it.

View File

@@ -1,9 +1,14 @@
# BBGradebookOrganiser
Blackboard Gradebook Organiser
### Blackboard submissions directory: *BB_submissions*
## Blackboard submissions directory: *BB_submissions*
Gradebooks will be organised into this directory, in a subdirectory with the same name
- Gradebook directory *AssignmentX* in *BB_gradebooks* will be organised into directory *AssignmentX* in *BB_submissions*
Also, a text file with all submission comments will be created in this directory, with the gradebook name as prefix
- e.g. *AssignmentX_comments.txt* will be created for gradebook *AssignmentX*
- Gradebooks from directory *BB_gradebooks* will be organised into this directory, in a subdirectory with the same name
- e.g. gradebook directory *AssignmentX* in *BB_gradebooks* will be organised into directory *AssignmentX* in *BB_submissions*
- Also, a text file with all submission comments will be created in this directory, with the gradebook name as prefix
- e.g. *AssignmentX_comments.txt* will be created for gradebook *AssignmentX*

View File

@@ -4,9 +4,9 @@ Blackboard Gradebook Organiser
## Description
**Blackboard Gradebook Organiser** is a tool for organising a downloaded gradebook with assignment submissions from [Blackboard Learn](https://en.wikipedia.org/wiki/Blackboard_Learn).
**Blackboard Gradebook Organiser** is a tool for organising a downloaded gradebook with assignment submissions from [Blackboard Learn ⧉](https://en.wikipedia.org/wiki/Blackboard_Learn).
The submission files are organised per student, by extracting the student number from the submission file names and creating a directory per student. Any compressed files (.zip, .rar, .7z) are extracted into the student's directory, with any remaining files submitted individually also moved into the student's directory. Student comments from submissions are also extracted into a single text file for convenient access and review.
Additionally, after organising submissions, you can inspect all submitted files to detect duplicated files from different submissions/students by generating and comparing SHA256 hashes. See section [Inspect submissions](#inspect-submissions-mag) for details.
Optionally, after organising a gradebook, you can inspect submissions to detect duplicated files from different submissions/students by generating and comparing SHA256 hashes. See section [Inspect submissions](#inspect-submissions-mag) for details.
### Features
@@ -18,7 +18,7 @@ Additionally, after organising submissions, you can inspect all submitted files
- Checks and extracts any comments from the student submission generated text files
- Checks if any compressed files (from the contents of the submitted compressed files) have been extracted and organised per student
- The path of any extracted and organised compressed files will be displayed on the terminal - they need to be extracted manually
- [Inspect submissions](#inspect-submissions-mag) by SHA256 hash :new:
- [Inspect submissions](#inspect-submissions-mag) by generating and comparing SHA256 hashes of submitted files :new:
## Instructions
@@ -63,7 +63,7 @@ Additionally, after organising submissions, you can inspect all submitted files
- Can exclude files from hashing, if provided with a CSV file listing the file names
- Compares the generated hashes and finds any duplicate hashes - ignores duplicates if they are by the same student/submission
- Finds all files with a duplicated hash and outputs them to CSV file with the following information: Student ID, file path, file name (without path), SHA256 hash
- Further inspection and filtering needs to be done manually, depending on the submission files
- Further inspection and filtering needs to be done manually
### Usage
@@ -71,13 +71,13 @@ Additionally, after organising submissions, you can inspect all submitted files
- `python -m pip install pandas`
- Usage: `python inspect_submissions.py GRADEBOOK_DIR_NAME`
- Note: run **after** organising a gradebook with `organise_gradebook.py`
- In order to exclude files from hashing, create a CSV file in directory *csv* to provide the file names to be excluded
- (Optional) In order to exclude files from hashing, create a CSV file in directory *csv* to provide the file names to be excluded
- e.g. for AssignmentX: create *AssignmentX_excluded.csv*, with a column named "exclude_filename" and list the file names
- Note: the directory *csv* is automatically created when you run `inspect_submissions.py` - you need to create it manually if you want to exclude files before the first run
- Generated CSV files can be found in directory *csv*, with *GRADEBOOK_DIR_NAME* as file name prefix
- e.g. inspecting submissions for *AssignmentX* will create 2 csv files:
- AssignmentX_file_hashes_[datetime].csv
- AssignmentX_suspicious_[datetime].csv
- *AssignmentX_file_hashes_[datetime].csv* - all files and their hashes
- *AssignmentX_suspicious_[datetime].csv* - files with duplicate hashes
## Notes

View File

@@ -4,7 +4,7 @@ from py7zr import SevenZipFile, exceptions
BAD_DIR_NAME = '__BAD__'
def mark_file_as_BAD(file, bad_exception):
def mark_file_as_BAD(file: str, bad_exception: Exception) -> None:
try:
filename = os.path.basename(file)
bad_dir = os.path.join(os.path.dirname(file), BAD_DIR_NAME)
@@ -16,7 +16,7 @@ def mark_file_as_BAD(file, bad_exception):
print(f'[Error] {e}')
def extract_zip(zip_file, target_dir):
def extract_zip(zip_file: str, target_dir: str) -> None:
try:
with zipfile.ZipFile(zip_file, 'r') as zip_ref:
members = [ m for m in zip_ref.infolist() if "__MACOSX" not in m.filename ]
@@ -25,10 +25,10 @@ def extract_zip(zip_file, target_dir):
except zipfile.BadZipfile as e:
mark_file_as_BAD(zip_file, e)
except Exception as e:
print(f'[ERROR] Something went wrong while extracting zip contents. Check the error message, get student id and download / organise manually\nError message: {e}')
print(f'[ERROR] Something went wrong while extracting the contents of a submitted zip file. Check the error message, get student id and download / organise manually\nError message: {e}')
def extract_rar(rar_file, target_dir):
def extract_rar(rar_file: str, target_dir: str) -> None:
try:
with rarfile.RarFile(rar_file, 'r') as rar_ref:
if platform.system() == 'Windows':
@@ -48,7 +48,7 @@ def extract_rar(rar_file, target_dir):
exit()
def extract_7z(seven_zip_file, target_dir):
def extract_7z(seven_zip_file: str, target_dir: str) -> None:
try: # extract the 7z file using py7zr
with open(seven_zip_file, 'rb') as f:
seven_zip = SevenZipFile(seven_zip_file, mode='r')
@@ -64,7 +64,7 @@ def extract_7z(seven_zip_file, target_dir):
mark_file_as_BAD(seven_zip_file, e)
def extract_file_to_dir(file_path, student_dir):
def extract_file_to_dir(file_path: str, student_dir: str) -> None:
os.makedirs(student_dir, exist_ok=True) # create the subdirectory for student
if file_path.lower().endswith('.zip'):

View File

@@ -16,6 +16,7 @@ def load_excluded_filenames(submissions_dir_name: str) -> list[str]: # helper f
try:
df = pd.read_csv(csv_file_path)
filename_list = df['exclude_filename'].tolist() # get the values of the 'filename' column as a list
filename_list = [ f.lower() for f in filename_list ] # convert to lowercase for comparison with submission files
print(f'[INFO] Using CSV file with list of excluded file names: {csv_file_path}')
return filename_list
except Exception as e: # any exception, print error and return empty list to continue without any excluded file names
@@ -28,7 +29,7 @@ def get_hashes_in_dir(dir_path: str, excluded_filenames: list = []) -> list: #
hash_list = []
for subdir, dirs, files in os.walk(dir_path): # loop through all files in the directory and generate hashes
for filename in files:
if filename not in excluded_filenames: # do not hash for inspection file names in the excluded list
if filename.lower() not in excluded_filenames: # convert to lowercase for comparison with excluded files & do not hash if in the excluded list
filepath = os.path.join(subdir, filename)
with open(filepath, 'rb') as f:
filehash = hashlib.sha256(f.read()).hexdigest()
@@ -62,10 +63,16 @@ def inspect_for_duplicate_hashes(hashes_csv_file_path: str): # main function fo
csv = pd.read_csv(hashes_csv_file_path)
df = pd.DataFrame(csv) # df with all files and their hashes
drop_columns = ['filepath', 'filename'] # only need to keep 'student id' and 'sha256 hash' for groupby later
df = df.drop(columns=drop_columns) # clear not needed columns
duplicate_hash = df.loc[df.duplicated(subset=['sha256 hash'], keep=False), :] # all files with duplicate hash - incl. files from the same student id
hash_with_multiple_student_ids = duplicate_hash.groupby('sha256 hash').agg(lambda x: len(x.unique())>1) # true if more than 1 unique student ids (= files with the same hash by multiple student ids), false if unique student id (= files from the same student id with the same hash)
suspicious_hashes_list = hash_with_multiple_student_ids[hash_with_multiple_student_ids['Student ID']==True].index.to_list() # list with duplicate hashes - only if different student id (doesn't include files from same student id)
df_clean = df.drop(columns=drop_columns) # clear not needed columns
duplicate_hash = df_clean.loc[df_clean.duplicated(subset=['sha256 hash'], keep=False), :] # all files with duplicate hash - incl. files from the same student id
# agg() for 'Student ID' True if more than 1 in groupby (= files with the same hash by multiple student ids)
# False if unique (= files from the same student id with the same hash)
hash_with_multiple_student_ids = duplicate_hash.groupby('sha256 hash').agg(lambda x: len(x.unique())>1)
# list with duplicate hashes - only if different student id (doesn't include files from same student id)
suspicious_hashes_list = hash_with_multiple_student_ids[hash_with_multiple_student_ids['Student ID']==True].index.to_list()
files_with_suspicious_hash = df[df['sha256 hash'].isin(suspicious_hashes_list)] # df with all files with duplicate/suspicious hash, excludes files from the same student id
df_suspicious = files_with_suspicious_hash.sort_values(['sha256 hash', 'Student ID']) # sort before output to csv

View File

@@ -4,7 +4,7 @@ from utils.extractor import extract_file_to_dir
BAD_DIR_NAME = '__BAD__'
def validate_gradebook_dir_name(src_dir):
def validate_gradebook_dir_name(src_dir: str) -> None:
if not os.path.isdir(src_dir): # check if it exists and is a directory
print(f"\n[Error] Incorrect directory: {src_dir}\n[Info] Make sure the directory exists in 'BB_gradebooks'")
exit()
@@ -16,7 +16,7 @@ def validate_gradebook_dir_name(src_dir):
exit()
def get_comment_from_submission_txt(file_path):
def get_comment_from_submission_txt(file_path: str) -> str | None:
no_comment_text = f'Comments:\nThere are no student comments for this assignment.'
no_comment_text_regex = no_comment_text
no_comment_regex_compile = re.compile(no_comment_text_regex)
@@ -30,9 +30,10 @@ def get_comment_from_submission_txt(file_path):
match = str(match).replace('\\n', '').replace('[','').replace(']','').replace('"','')
match = str(match).split('Comments:')[-1]
return match
return None
def get_gradebook_stats(src_dir):
def get_gradebook_stats(src_dir: str) -> dict[str, int]:
all_files = [ os.path.join(src_dir, f) for f in os.listdir(src_dir) if BAD_DIR_NAME not in f ]
dirs = [ f for f in all_files if os.path.isdir(f) and BAD_DIR_NAME not in f ]
normal_files = [ f for f in all_files if os.path.isfile(f) ]
@@ -57,7 +58,7 @@ def get_gradebook_stats(src_dir):
return files_counter
def organise_file_per_student(src_dir, dest_dir, file_name, student_no):
def organise_file_per_student(src_dir: str, dest_dir: str, file_name: str, student_no: str) -> None:
student_dir = os.path.join(dest_dir, student_no)
os.makedirs(student_dir, exist_ok=True) # create student directory if it doesn't exist
file_path = os.path.join(src_dir, file_name)
@@ -80,7 +81,7 @@ def organise_file_per_student(src_dir, dest_dir, file_name, student_no):
shutil.move(file_path, new_file_path) # move the file to student directory
def organise_gradebook(src_dir, dest_dir):
def organise_gradebook(src_dir: str, dest_dir: str) -> None:
"""1) extracts .zip, .rar, .7z files, organises contents into directories per student number, and deletes compressed files after successful extraction
2) organises all other files in gradebook into directories per student number
3) checks if there are any comments in submission text files and extracts them into a file
@@ -88,7 +89,7 @@ def organise_gradebook(src_dir, dest_dir):
validate_gradebook_dir_name(src_dir) # check if dir exists, and has files in it - exits if not
os.makedirs(dest_dir, exist_ok=True) # create the destination directory if it doesn't exist
files_counter = get_gradebook_stats(src_dir) # print stats about the files in gradebook and get files_counter dict to use later
students_numbers = [] # list to add and count unique student numbers from all files in gradebook
students_numbers: list[str] = [] # list to add and count unique student numbers from all files in gradebook
print('\nStart organising...\n')
for file_name in os.listdir(src_dir): # iterate through all files in the directory
if BAD_DIR_NAME not in file_name: # ignore dir BAD_DIR_NAME (created after first run if corrupt compressed files found)
@@ -107,11 +108,11 @@ def organise_gradebook(src_dir, dest_dir):
print(f'[Note] Compressed files (.zip, .rar, .7z) are automatically deleted from the gradebook directory after successful extraction')
def check_submissions_dir_for_compressed(submissions_dir):
def check_submissions_dir_for_compressed(submissions_dir: str) -> None:
"""checks if any submitted compressed files contain more compressed files inside (they are not recursively extracted)
\nprints any compressed files location that need to be extracted manually
"""
compressed_files = []
compressed_files: list[str] = []
abs_path = os.getcwd()
for the_path, dirc, files in os.walk(submissions_dir):
for fname in files: