Delimiter Tokenizer

class py_stringmatching.tokenizer.delimiter_tokenizer.DelimiterTokenizer(delim_set={' '}, return_set=False)[source]

Uses delimiters to find tokens, as apposed to using definitions.

Examples of delimiters include white space and punctuations. Examples of definitions include alphabetical and qgram tokens.

Parameters:
  • delim_set (set) – A set of delimiter strings (defaults to space delimiter).
  • return_set (boolean) – A flag to indicate whether to return a set of tokens instead of a bag of tokens (defaults to False).
return_set

An attribute to store the value of the flag return_set.

Type:boolean
get_delim_set()[source]

Gets the current set of delimiters.

Returns:A Python set which is the current set of delimiters.
get_return_set()

Gets the value of the return_set flag.

Returns:The boolean value of the return_set flag.
set_delim_set(delim_set)[source]

Sets the current set of delimiters.

Parameters:delim_set (set) – A set of delimiter strings.
set_return_set(return_set)

Sets the value of the return_set flag.

Parameters:return_set (boolean) – a flag to indicate whether to return a set of tokens instead of a bag of tokens.
tokenize(input_string)[source]

Tokenizes input string based on the set of delimiters.

Parameters:input_string (str) – The string to be tokenized.
Returns:A Python list which is a set or a bag of tokens, depending on whether return_set flag is set to True or False.
Raises:TypeError – If the input is not a string.

Examples

>>> delim_tok = DelimiterTokenizer()
>>> delim_tok.tokenize('data science')
['data', 'science']
>>> delim_tok = DelimiterTokenizer(['$#$'])
>>> delim_tok.tokenize('data$#$science')
['data', 'science']
>>> delim_tok = DelimiterTokenizer([',', '.'])
>>> delim_tok.tokenize('data,science.data,integration.')
['data', 'science', 'data', 'integration']
>>> delim_tok = DelimiterTokenizer([',', '.'], return_set=True)
>>> delim_tok.tokenize('data,science.data,integration.')
['data', 'science', 'integration']