Tokenize is a Julia package that serves a similar purpose and API as the tokenize module in Python but for Julia. This is to take a string or buffer containing Julia code, perform lexical analysis and return a stream of tokens.
A tool written for the Blender API in the Python Language for Second Life primitives. All code is 100% original. Second Life and LSL are the property of Linden Lab. Blender is the property of the Blender Foundation. No infringement is intended.