Compare commits
3 Commits
7ae6929ffe
...
3e677d9ddd
| Author | SHA1 | Date | |
|---|---|---|---|
| 3e677d9ddd | |||
| 2b6fe45b42 | |||
| 964a1f6abb |
2
.gitignore
vendored
2
.gitignore
vendored
@@ -129,7 +129,7 @@ dmypy.json
|
|||||||
.pyre/
|
.pyre/
|
||||||
|
|
||||||
# BBGradebookOrganiser
|
# BBGradebookOrganiser
|
||||||
|
TODO
|
||||||
BB_gradebooks/
|
BB_gradebooks/
|
||||||
BB_submissions/
|
BB_submissions/
|
||||||
csv/
|
csv/
|
||||||
|
|||||||
@@ -1,6 +1,7 @@
|
|||||||
# BBGradebookOrganiser
|
# BBGradebookOrganiser
|
||||||
|
|
||||||
Blackboard Gradebook Organiser
|
Blackboard Gradebook Organiser
|
||||||
|
|
||||||
### Blackboard gradebooks directory: *BB_gradebooks*
|
## Blackboard gradebooks directory: *BB_gradebooks*
|
||||||
|
|
||||||
Create a directory with an appropriate name for the gradebook / assignment in this directory, and extract the downloaded gradebook .zip file in it.
|
Create a directory with an appropriate name for the gradebook / assignment in this directory, and extract the downloaded gradebook .zip file in it.
|
||||||
|
|||||||
@@ -1,9 +1,14 @@
|
|||||||
# BBGradebookOrganiser
|
# BBGradebookOrganiser
|
||||||
|
|
||||||
Blackboard Gradebook Organiser
|
Blackboard Gradebook Organiser
|
||||||
|
|
||||||
### Blackboard submissions directory: *BB_submissions*
|
## Blackboard submissions directory: *BB_submissions*
|
||||||
|
|
||||||
|
Gradebooks will be organised into this directory, in a subdirectory with the same name
|
||||||
|
|
||||||
|
- Gradebook directory *AssignmentX* in *BB_gradebooks* will be organised into directory *AssignmentX* in *BB_submissions*
|
||||||
|
|
||||||
|
Also, a text file with all submission comments will be created in this directory, with the gradebook name as prefix
|
||||||
|
|
||||||
- Gradebooks from directory *BB_gradebooks* will be organised into this directory, in a subdirectory with the same name
|
|
||||||
- e.g. gradebook directory *AssignmentX* in *BB_gradebooks* will be organised into directory *AssignmentX* in *BB_submissions*
|
|
||||||
- Also, a text file with all submission comments will be created in this directory, with the gradebook name as prefix
|
|
||||||
- e.g. *AssignmentX_comments.txt* will be created for gradebook *AssignmentX*
|
- e.g. *AssignmentX_comments.txt* will be created for gradebook *AssignmentX*
|
||||||
|
|
||||||
14
README.md
14
README.md
@@ -4,9 +4,9 @@ Blackboard Gradebook Organiser
|
|||||||
|
|
||||||
## Description
|
## Description
|
||||||
|
|
||||||
**Blackboard Gradebook Organiser** is a tool for organising a downloaded gradebook with assignment submissions from [Blackboard Learn](https://en.wikipedia.org/wiki/Blackboard_Learn).
|
**Blackboard Gradebook Organiser** is a tool for organising a downloaded gradebook with assignment submissions from [Blackboard Learn ⧉](https://en.wikipedia.org/wiki/Blackboard_Learn).
|
||||||
The submission files are organised per student, by extracting the student number from the submission file names and creating a directory per student. Any compressed files (.zip, .rar, .7z) are extracted into the student's directory, with any remaining files submitted individually also moved into the student's directory. Student comments from submissions are also extracted into a single text file for convenient access and review.
|
The submission files are organised per student, by extracting the student number from the submission file names and creating a directory per student. Any compressed files (.zip, .rar, .7z) are extracted into the student's directory, with any remaining files submitted individually also moved into the student's directory. Student comments from submissions are also extracted into a single text file for convenient access and review.
|
||||||
Additionally, after organising submissions, you can inspect all submitted files to detect duplicated files from different submissions/students by generating and comparing SHA256 hashes. See section [Inspect submissions](#inspect-submissions-mag) for details.
|
Optionally, after organising a gradebook, you can inspect submissions to detect duplicated files from different submissions/students by generating and comparing SHA256 hashes. See section [Inspect submissions](#inspect-submissions-mag) for details.
|
||||||
|
|
||||||
### Features
|
### Features
|
||||||
|
|
||||||
@@ -18,7 +18,7 @@ Additionally, after organising submissions, you can inspect all submitted files
|
|||||||
- Checks and extracts any comments from the student submission generated text files
|
- Checks and extracts any comments from the student submission generated text files
|
||||||
- Checks if any compressed files (from the contents of the submitted compressed files) have been extracted and organised per student
|
- Checks if any compressed files (from the contents of the submitted compressed files) have been extracted and organised per student
|
||||||
- The path of any extracted and organised compressed files will be displayed on the terminal - they need to be extracted manually
|
- The path of any extracted and organised compressed files will be displayed on the terminal - they need to be extracted manually
|
||||||
- [Inspect submissions](#inspect-submissions-mag) by SHA256 hash :new:
|
- [Inspect submissions](#inspect-submissions-mag) by generating and comparing SHA256 hashes of submitted files :new:
|
||||||
|
|
||||||
## Instructions
|
## Instructions
|
||||||
|
|
||||||
@@ -63,7 +63,7 @@ Additionally, after organising submissions, you can inspect all submitted files
|
|||||||
- Can exclude files from hashing, if provided with a CSV file listing the file names
|
- Can exclude files from hashing, if provided with a CSV file listing the file names
|
||||||
- Compares the generated hashes and finds any duplicate hashes - ignores duplicates if they are by the same student/submission
|
- Compares the generated hashes and finds any duplicate hashes - ignores duplicates if they are by the same student/submission
|
||||||
- Finds all files with a duplicated hash and outputs them to CSV file with the following information: Student ID, file path, file name (without path), SHA256 hash
|
- Finds all files with a duplicated hash and outputs them to CSV file with the following information: Student ID, file path, file name (without path), SHA256 hash
|
||||||
- Further inspection and filtering needs to be done manually, depending on the submission files
|
- Further inspection and filtering needs to be done manually
|
||||||
|
|
||||||
### Usage
|
### Usage
|
||||||
|
|
||||||
@@ -71,13 +71,13 @@ Additionally, after organising submissions, you can inspect all submitted files
|
|||||||
- `python -m pip install pandas`
|
- `python -m pip install pandas`
|
||||||
- Usage: `python inspect_submissions.py GRADEBOOK_DIR_NAME`
|
- Usage: `python inspect_submissions.py GRADEBOOK_DIR_NAME`
|
||||||
- Note: run **after** organising a gradebook with `organise_gradebook.py`
|
- Note: run **after** organising a gradebook with `organise_gradebook.py`
|
||||||
- In order to exclude files from hashing, create a CSV file in directory *csv* to provide the file names to be excluded
|
- (Optional) In order to exclude files from hashing, create a CSV file in directory *csv* to provide the file names to be excluded
|
||||||
- e.g. for AssignmentX: create *AssignmentX_excluded.csv*, with a column named "exclude_filename" and list the file names
|
- e.g. for AssignmentX: create *AssignmentX_excluded.csv*, with a column named "exclude_filename" and list the file names
|
||||||
- Note: the directory *csv* is automatically created when you run `inspect_submissions.py` - you need to create it manually if you want to exclude files before the first run
|
- Note: the directory *csv* is automatically created when you run `inspect_submissions.py` - you need to create it manually if you want to exclude files before the first run
|
||||||
- Generated CSV files can be found in directory *csv*, with *GRADEBOOK_DIR_NAME* as file name prefix
|
- Generated CSV files can be found in directory *csv*, with *GRADEBOOK_DIR_NAME* as file name prefix
|
||||||
- e.g. inspecting submissions for *AssignmentX* will create 2 csv files:
|
- e.g. inspecting submissions for *AssignmentX* will create 2 csv files:
|
||||||
- AssignmentX_file_hashes_[datetime].csv
|
- *AssignmentX_file_hashes_[datetime].csv* - all files and their hashes
|
||||||
- AssignmentX_suspicious_[datetime].csv
|
- *AssignmentX_suspicious_[datetime].csv* - files with duplicate hashes
|
||||||
|
|
||||||
## Notes
|
## Notes
|
||||||
|
|
||||||
|
|||||||
@@ -4,7 +4,7 @@ from py7zr import SevenZipFile, exceptions
|
|||||||
|
|
||||||
BAD_DIR_NAME = '__BAD__'
|
BAD_DIR_NAME = '__BAD__'
|
||||||
|
|
||||||
def mark_file_as_BAD(file, bad_exception):
|
def mark_file_as_BAD(file: str, bad_exception: Exception) -> None:
|
||||||
try:
|
try:
|
||||||
filename = os.path.basename(file)
|
filename = os.path.basename(file)
|
||||||
bad_dir = os.path.join(os.path.dirname(file), BAD_DIR_NAME)
|
bad_dir = os.path.join(os.path.dirname(file), BAD_DIR_NAME)
|
||||||
@@ -16,7 +16,7 @@ def mark_file_as_BAD(file, bad_exception):
|
|||||||
print(f'[Error] {e}')
|
print(f'[Error] {e}')
|
||||||
|
|
||||||
|
|
||||||
def extract_zip(zip_file, target_dir):
|
def extract_zip(zip_file: str, target_dir: str) -> None:
|
||||||
try:
|
try:
|
||||||
with zipfile.ZipFile(zip_file, 'r') as zip_ref:
|
with zipfile.ZipFile(zip_file, 'r') as zip_ref:
|
||||||
members = [ m for m in zip_ref.infolist() if "__MACOSX" not in m.filename ]
|
members = [ m for m in zip_ref.infolist() if "__MACOSX" not in m.filename ]
|
||||||
@@ -25,10 +25,10 @@ def extract_zip(zip_file, target_dir):
|
|||||||
except zipfile.BadZipfile as e:
|
except zipfile.BadZipfile as e:
|
||||||
mark_file_as_BAD(zip_file, e)
|
mark_file_as_BAD(zip_file, e)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
print(f'[ERROR] Something went wrong while extracting zip contents. Check the error message, get student id and download / organise manually\nError message: {e}')
|
print(f'[ERROR] Something went wrong while extracting the contents of a submitted zip file. Check the error message, get student id and download / organise manually\nError message: {e}')
|
||||||
|
|
||||||
|
|
||||||
def extract_rar(rar_file, target_dir):
|
def extract_rar(rar_file: str, target_dir: str) -> None:
|
||||||
try:
|
try:
|
||||||
with rarfile.RarFile(rar_file, 'r') as rar_ref:
|
with rarfile.RarFile(rar_file, 'r') as rar_ref:
|
||||||
if platform.system() == 'Windows':
|
if platform.system() == 'Windows':
|
||||||
@@ -48,7 +48,7 @@ def extract_rar(rar_file, target_dir):
|
|||||||
exit()
|
exit()
|
||||||
|
|
||||||
|
|
||||||
def extract_7z(seven_zip_file, target_dir):
|
def extract_7z(seven_zip_file: str, target_dir: str) -> None:
|
||||||
try: # extract the 7z file using py7zr
|
try: # extract the 7z file using py7zr
|
||||||
with open(seven_zip_file, 'rb') as f:
|
with open(seven_zip_file, 'rb') as f:
|
||||||
seven_zip = SevenZipFile(seven_zip_file, mode='r')
|
seven_zip = SevenZipFile(seven_zip_file, mode='r')
|
||||||
@@ -64,7 +64,7 @@ def extract_7z(seven_zip_file, target_dir):
|
|||||||
mark_file_as_BAD(seven_zip_file, e)
|
mark_file_as_BAD(seven_zip_file, e)
|
||||||
|
|
||||||
|
|
||||||
def extract_file_to_dir(file_path, student_dir):
|
def extract_file_to_dir(file_path: str, student_dir: str) -> None:
|
||||||
os.makedirs(student_dir, exist_ok=True) # create the subdirectory for student
|
os.makedirs(student_dir, exist_ok=True) # create the subdirectory for student
|
||||||
|
|
||||||
if file_path.lower().endswith('.zip'):
|
if file_path.lower().endswith('.zip'):
|
||||||
|
|||||||
@@ -16,6 +16,7 @@ def load_excluded_filenames(submissions_dir_name: str) -> list[str]: # helper f
|
|||||||
try:
|
try:
|
||||||
df = pd.read_csv(csv_file_path)
|
df = pd.read_csv(csv_file_path)
|
||||||
filename_list = df['exclude_filename'].tolist() # get the values of the 'filename' column as a list
|
filename_list = df['exclude_filename'].tolist() # get the values of the 'filename' column as a list
|
||||||
|
filename_list = [ f.lower() for f in filename_list ] # convert to lowercase for comparison with submission files
|
||||||
print(f'[INFO] Using CSV file with list of excluded file names: {csv_file_path}')
|
print(f'[INFO] Using CSV file with list of excluded file names: {csv_file_path}')
|
||||||
return filename_list
|
return filename_list
|
||||||
except Exception as e: # any exception, print error and return empty list to continue without any excluded file names
|
except Exception as e: # any exception, print error and return empty list to continue without any excluded file names
|
||||||
@@ -28,7 +29,7 @@ def get_hashes_in_dir(dir_path: str, excluded_filenames: list = []) -> list: #
|
|||||||
hash_list = []
|
hash_list = []
|
||||||
for subdir, dirs, files in os.walk(dir_path): # loop through all files in the directory and generate hashes
|
for subdir, dirs, files in os.walk(dir_path): # loop through all files in the directory and generate hashes
|
||||||
for filename in files:
|
for filename in files:
|
||||||
if filename not in excluded_filenames: # do not hash for inspection file names in the excluded list
|
if filename.lower() not in excluded_filenames: # convert to lowercase for comparison with excluded files & do not hash if in the excluded list
|
||||||
filepath = os.path.join(subdir, filename)
|
filepath = os.path.join(subdir, filename)
|
||||||
with open(filepath, 'rb') as f:
|
with open(filepath, 'rb') as f:
|
||||||
filehash = hashlib.sha256(f.read()).hexdigest()
|
filehash = hashlib.sha256(f.read()).hexdigest()
|
||||||
@@ -62,10 +63,16 @@ def inspect_for_duplicate_hashes(hashes_csv_file_path: str): # main function fo
|
|||||||
csv = pd.read_csv(hashes_csv_file_path)
|
csv = pd.read_csv(hashes_csv_file_path)
|
||||||
df = pd.DataFrame(csv) # df with all files and their hashes
|
df = pd.DataFrame(csv) # df with all files and their hashes
|
||||||
drop_columns = ['filepath', 'filename'] # only need to keep 'student id' and 'sha256 hash' for groupby later
|
drop_columns = ['filepath', 'filename'] # only need to keep 'student id' and 'sha256 hash' for groupby later
|
||||||
df = df.drop(columns=drop_columns) # clear not needed columns
|
df_clean = df.drop(columns=drop_columns) # clear not needed columns
|
||||||
duplicate_hash = df.loc[df.duplicated(subset=['sha256 hash'], keep=False), :] # all files with duplicate hash - incl. files from the same student id
|
duplicate_hash = df_clean.loc[df_clean.duplicated(subset=['sha256 hash'], keep=False), :] # all files with duplicate hash - incl. files from the same student id
|
||||||
hash_with_multiple_student_ids = duplicate_hash.groupby('sha256 hash').agg(lambda x: len(x.unique())>1) # true if more than 1 unique student ids (= files with the same hash by multiple student ids), false if unique student id (= files from the same student id with the same hash)
|
|
||||||
suspicious_hashes_list = hash_with_multiple_student_ids[hash_with_multiple_student_ids['Student ID']==True].index.to_list() # list with duplicate hashes - only if different student id (doesn't include files from same student id)
|
# agg() for 'Student ID' True if more than 1 in groupby (= files with the same hash by multiple student ids)
|
||||||
|
# False if unique (= files from the same student id with the same hash)
|
||||||
|
hash_with_multiple_student_ids = duplicate_hash.groupby('sha256 hash').agg(lambda x: len(x.unique())>1)
|
||||||
|
|
||||||
|
# list with duplicate hashes - only if different student id (doesn't include files from same student id)
|
||||||
|
suspicious_hashes_list = hash_with_multiple_student_ids[hash_with_multiple_student_ids['Student ID']==True].index.to_list()
|
||||||
|
|
||||||
files_with_suspicious_hash = df[df['sha256 hash'].isin(suspicious_hashes_list)] # df with all files with duplicate/suspicious hash, excludes files from the same student id
|
files_with_suspicious_hash = df[df['sha256 hash'].isin(suspicious_hashes_list)] # df with all files with duplicate/suspicious hash, excludes files from the same student id
|
||||||
df_suspicious = files_with_suspicious_hash.sort_values(['sha256 hash', 'Student ID']) # sort before output to csv
|
df_suspicious = files_with_suspicious_hash.sort_values(['sha256 hash', 'Student ID']) # sort before output to csv
|
||||||
|
|
||||||
|
|||||||
@@ -4,7 +4,7 @@ from utils.extractor import extract_file_to_dir
|
|||||||
BAD_DIR_NAME = '__BAD__'
|
BAD_DIR_NAME = '__BAD__'
|
||||||
|
|
||||||
|
|
||||||
def validate_gradebook_dir_name(src_dir):
|
def validate_gradebook_dir_name(src_dir: str) -> None:
|
||||||
if not os.path.isdir(src_dir): # check if it exists and is a directory
|
if not os.path.isdir(src_dir): # check if it exists and is a directory
|
||||||
print(f"\n[Error] Incorrect directory: {src_dir}\n[Info] Make sure the directory exists in 'BB_gradebooks'")
|
print(f"\n[Error] Incorrect directory: {src_dir}\n[Info] Make sure the directory exists in 'BB_gradebooks'")
|
||||||
exit()
|
exit()
|
||||||
@@ -16,7 +16,7 @@ def validate_gradebook_dir_name(src_dir):
|
|||||||
exit()
|
exit()
|
||||||
|
|
||||||
|
|
||||||
def get_comment_from_submission_txt(file_path):
|
def get_comment_from_submission_txt(file_path: str) -> str | None:
|
||||||
no_comment_text = f'Comments:\nThere are no student comments for this assignment.'
|
no_comment_text = f'Comments:\nThere are no student comments for this assignment.'
|
||||||
no_comment_text_regex = no_comment_text
|
no_comment_text_regex = no_comment_text
|
||||||
no_comment_regex_compile = re.compile(no_comment_text_regex)
|
no_comment_regex_compile = re.compile(no_comment_text_regex)
|
||||||
@@ -30,9 +30,10 @@ def get_comment_from_submission_txt(file_path):
|
|||||||
match = str(match).replace('\\n', '').replace('[','').replace(']','').replace('"','')
|
match = str(match).replace('\\n', '').replace('[','').replace(']','').replace('"','')
|
||||||
match = str(match).split('Comments:')[-1]
|
match = str(match).split('Comments:')[-1]
|
||||||
return match
|
return match
|
||||||
|
return None
|
||||||
|
|
||||||
|
|
||||||
def get_gradebook_stats(src_dir):
|
def get_gradebook_stats(src_dir: str) -> dict[str, int]:
|
||||||
all_files = [ os.path.join(src_dir, f) for f in os.listdir(src_dir) if BAD_DIR_NAME not in f ]
|
all_files = [ os.path.join(src_dir, f) for f in os.listdir(src_dir) if BAD_DIR_NAME not in f ]
|
||||||
dirs = [ f for f in all_files if os.path.isdir(f) and BAD_DIR_NAME not in f ]
|
dirs = [ f for f in all_files if os.path.isdir(f) and BAD_DIR_NAME not in f ]
|
||||||
normal_files = [ f for f in all_files if os.path.isfile(f) ]
|
normal_files = [ f for f in all_files if os.path.isfile(f) ]
|
||||||
@@ -57,7 +58,7 @@ def get_gradebook_stats(src_dir):
|
|||||||
return files_counter
|
return files_counter
|
||||||
|
|
||||||
|
|
||||||
def organise_file_per_student(src_dir, dest_dir, file_name, student_no):
|
def organise_file_per_student(src_dir: str, dest_dir: str, file_name: str, student_no: str) -> None:
|
||||||
student_dir = os.path.join(dest_dir, student_no)
|
student_dir = os.path.join(dest_dir, student_no)
|
||||||
os.makedirs(student_dir, exist_ok=True) # create student directory if it doesn't exist
|
os.makedirs(student_dir, exist_ok=True) # create student directory if it doesn't exist
|
||||||
file_path = os.path.join(src_dir, file_name)
|
file_path = os.path.join(src_dir, file_name)
|
||||||
@@ -80,7 +81,7 @@ def organise_file_per_student(src_dir, dest_dir, file_name, student_no):
|
|||||||
shutil.move(file_path, new_file_path) # move the file to student directory
|
shutil.move(file_path, new_file_path) # move the file to student directory
|
||||||
|
|
||||||
|
|
||||||
def organise_gradebook(src_dir, dest_dir):
|
def organise_gradebook(src_dir: str, dest_dir: str) -> None:
|
||||||
"""1) extracts .zip, .rar, .7z files, organises contents into directories per student number, and deletes compressed files after successful extraction
|
"""1) extracts .zip, .rar, .7z files, organises contents into directories per student number, and deletes compressed files after successful extraction
|
||||||
2) organises all other files in gradebook into directories per student number
|
2) organises all other files in gradebook into directories per student number
|
||||||
3) checks if there are any comments in submission text files and extracts them into a file
|
3) checks if there are any comments in submission text files and extracts them into a file
|
||||||
@@ -88,7 +89,7 @@ def organise_gradebook(src_dir, dest_dir):
|
|||||||
validate_gradebook_dir_name(src_dir) # check if dir exists, and has files in it - exits if not
|
validate_gradebook_dir_name(src_dir) # check if dir exists, and has files in it - exits if not
|
||||||
os.makedirs(dest_dir, exist_ok=True) # create the destination directory if it doesn't exist
|
os.makedirs(dest_dir, exist_ok=True) # create the destination directory if it doesn't exist
|
||||||
files_counter = get_gradebook_stats(src_dir) # print stats about the files in gradebook and get files_counter dict to use later
|
files_counter = get_gradebook_stats(src_dir) # print stats about the files in gradebook and get files_counter dict to use later
|
||||||
students_numbers = [] # list to add and count unique student numbers from all files in gradebook
|
students_numbers: list[str] = [] # list to add and count unique student numbers from all files in gradebook
|
||||||
print('\nStart organising...\n')
|
print('\nStart organising...\n')
|
||||||
for file_name in os.listdir(src_dir): # iterate through all files in the directory
|
for file_name in os.listdir(src_dir): # iterate through all files in the directory
|
||||||
if BAD_DIR_NAME not in file_name: # ignore dir BAD_DIR_NAME (created after first run if corrupt compressed files found)
|
if BAD_DIR_NAME not in file_name: # ignore dir BAD_DIR_NAME (created after first run if corrupt compressed files found)
|
||||||
@@ -107,11 +108,11 @@ def organise_gradebook(src_dir, dest_dir):
|
|||||||
print(f'[Note] Compressed files (.zip, .rar, .7z) are automatically deleted from the gradebook directory after successful extraction')
|
print(f'[Note] Compressed files (.zip, .rar, .7z) are automatically deleted from the gradebook directory after successful extraction')
|
||||||
|
|
||||||
|
|
||||||
def check_submissions_dir_for_compressed(submissions_dir):
|
def check_submissions_dir_for_compressed(submissions_dir: str) -> None:
|
||||||
"""checks if any submitted compressed files contain more compressed files inside (they are not recursively extracted)
|
"""checks if any submitted compressed files contain more compressed files inside (they are not recursively extracted)
|
||||||
\nprints any compressed files location that need to be extracted manually
|
\nprints any compressed files location that need to be extracted manually
|
||||||
"""
|
"""
|
||||||
compressed_files = []
|
compressed_files: list[str] = []
|
||||||
abs_path = os.getcwd()
|
abs_path = os.getcwd()
|
||||||
for the_path, dirc, files in os.walk(submissions_dir):
|
for the_path, dirc, files in os.walk(submissions_dir):
|
||||||
for fname in files:
|
for fname in files:
|
||||||
|
|||||||
Reference in New Issue
Block a user