Skip to content

grascii/tokenizers

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 
 
 

Repository files navigation

tokenizers

Training scripts for Grascii tokenizers typically used for machine learning models.

v1

This tokenizer operates on normalized Grascii and is intended for use with a Roberta model. It is trained on the gregg-preanniversary-words dataset.

The X and XS strokes are encoded as S and SS respectively due to their high visual similarity.

Releases

No releases published

Packages

No packages published

Languages