Method List
-
#[] Document::Corpus
-
#[] Segment::RangeIndex
-
#__define_regexp_hook RegExpNER
-
#_parts AnnotID
-
#_parts SegID
-
#add Dictionary
-
#add Dictionary::TF_IDF
-
#add Dictionary::KL
-
#add_document Document::Corpus
-
#add_instance Finder
-
#add_pmid Document::Corpus
-
#add_pubmed_query Document::Corpus
-
#add_regexp RegExpNER
-
#advance TokenTrieNER::EnumeratedArray
-
align Segment
-
#appearence_order Normalizer
-
#arabic String
-
ascii Segment
-
#back TokenTrieNER::EnumeratedArray
-
bad_chars Segment
-
#best Dictionary::KL
-
#best Dictionary::TF_IDF
-
#bigrams String
-
bigrams BagOfWords
-
#case_insensitive NGramPrefixDictionary
-
chunk_dep_graph SpaCy
-
chunk_segments SpaCy
-
chunks SpaCy
-
claim Document::Corpus
-
claims Document::Corpus
-
#clean CueIndex
-
clean TokenTrieNER
-
clean_sort Segment
-
#code TokenTrieNER::Code
-
collisions SegmentRanges
-
#config CueIndex
-
config SpaCy
-
#config NERFeatures
-
#context NERFeatures
-
#corpus DocID
-
#corpus Segment::RangeIndex
-
count BagOfWords
-
#cues CueIndex
-
default_corpus DocID
-
#define NERFeatures
-
#define CueIndex
-
define Document
-
#define_comparisons Tokenizer
-
define_multiple Document
-
#define_regexp RegExpNER
-
#define_tokens Tokenizer
-
dep_graph SpaCy
-
#df Dictionary::TF_IDF
-
#direction NERFeatures
-
#docid SegID
-
#docids Document::Corpus
-
#docs Dictionary::TF_IDF
-
#documents Document::Corpus
-
#downcase_first String
-
#eend Segment
-
#eend SegID
-
#entities NER
-
entities GNormPlus
-
#entity NamedEntity
-
#entity_type NamedEntity
-
#entrez_score Normalizer
-
#eval Tokenizer::Custom
-
#eval Tokenizer::Operation
-
#evaluate Tokenizer
-
#evaluate_tokens Tokenizer
-
event_extraction NLP
-
extended TokenTrieNER::EnumeratedArray
-
#extract NER
-
#features NERFeatures
-
features BagOfWords
-
find TokenTrieNER
-
#find Finder
-
#find Finder::Instance
-
find_fail TokenTrieNER
-
#fix_segment Transformed
-
follow TokenTrieNER
-
#format Finder::Instance
-
gdep_chunks NLP
-
get_best Normalizer
-
#html NamedEntity
-
#html Relationship
-
#id DocID
-
#idf Dictionary::TF_IDF
-
#ignore_case Tokenizer::Operation
-
ignore_case Tokenizer
-
#ignore_case Tokenizer::Custom
-
#includes? SegmentRanges
-
#index TokenTrieNER
-
#index NGramPrefixDictionary
-
index Segment
-
index Segment::RangeIndex
-
index_for_tokens TokenTrieNER
-
init Abner
-
init OSCAR4
-
init OSCAR3
-
init Banner
-
init Linnaeus
-
init ChemicalTagger
-
#initialize PatternRelExt
-
#initialize Tokenizer::Transform
-
#initialize NGramPrefixDictionary
-
#initialize Dictionary::KL
-
#initialize Finder
-
#initialize Finder::Instance
-
#initialize Banner
-
#initialize Tokenizer::Custom
-
#initialize Tokenizer::Operation
-
#initialize Dictionary::TF_IDF
-
#initialize Normalizer
-
#initialize Abner
-
#initialize NER
-
#initialize Tokenizer
-
#initialize CueIndex
-
#initialize TokenTrieNER
-
#initialize TokenTrieNER::Code
-
#initialize NERFeatures
-
#initialize Dictionary
-
#initialize RegExpNER
-
#instances Finder
-
#is_special? String
-
#kl Dictionary::KL
-
#last? TokenTrieNER::EnumeratedArray
-
#left? TokenTrieNER::EnumeratedArray
-
#load CueIndex
-
load Brat
-
#longest_match TokenTrieNER
-
#main Tokenizer
-
make_match TokenTrieNER
-
#make_relative SegmentRanges
-
#match ChemicalTagger
-
match NGramPrefixDictionary
-
match Linnaeus
-
#match NGramPrefixDictionary
-
#match OSCAR3
-
match OSCAR4
-
match OSCAR3
-
#match OSCAR4
-
#match TokenTrieNER
-
#match Banner
-
#match Normalizer
-
#match Abner
-
#match RegExpNER
-
#match CueIndex
-
match ChemicalTagger
-
#match_chunks PatternRelExt
-
match_regexp RegExpNER
-
match_regexp_hash RegExpNER
-
match_regexp_list RegExpNER
-
#match_sentences PatternRelExt
-
#merge TokenTrieNER
-
merge TokenTrieNER
-
merge_vp_chunks NLP
-
#method_missing Tokenizer::Custom
-
#method_missing Tokenizer::Transform
-
#method_missing Tokenizer::Operation
-
#namespace Finder::Instance
-
#neg_dict Dictionary::KL
-
#new_token_trie PatternRelExt
-
#next TokenTrieNER::EnumeratedArray
-
nlp SpaCy
-
#no_clean TokenTrieNER
-
#normalizer Finder::Instance
-
#num_docs Dictionary::TF_IDF
-
#offset SegID
-
overlaps Segment
-
#overlaps SegmentRanges
-
#overlaps? SegmentRanges
-
paths SpaCy
-
#peek TokenTrieNER::EnumeratedArray
-
#pos TokenTrieNER::EnumeratedArray
-
#pos_dict Dictionary::KL
-
prepare_chunk_patterns PatternRelExt
-
prepare_token TokenTrieNER
-
process GNormPlus
-
process TokenTrieNER
-
process_hash NGramPrefixDictionary
-
process_labels NLP
-
process_stream NGramPrefixDictionary
-
#pull SegmentRanges
-
#push SegmentRanges
-
#range Segment
-
#range SegID
-
#range_in SegmentRanges
-
#regexps RegExpNER
-
#replace_segments Transformed
-
#report NamedEntity
-
#resolve Normalizer
-
#restore Transformed
-
returnFeatures NLP
-
reverse NERFeatures
-
#reverse NERFeatures
-
#segid AnnotID
-
#segment_length Segment
-
#segment_length SegID
-
segments SpaCy
-
#select Normalizer
-
sentence_split_detector OpenNLP
-
sentence_splitter OpenNLP
-
setup Document::Corpus
-
#shift Transformed
-
simple_pattern PatternRelExt
-
#slack PatternRelExt
-
#slack TokenTrieNER
-
sort Segment
-
split Segment
-
#split_at TokenTrieNER
-
#split_on_matches RegExpNER
-
#split_segments Segmented
-
#stem TokenTrieNER
-
#tagged_features NERFeatures
-
tagger OSCAR4
-
#template NERFeatures
-
#terms Dictionary::KL
-
#terms Dictionary
-
#terms Dictionary::TF_IDF
-
terms BagOfWords
-
#text Relationship
-
#text_features NERFeatures
-
#tf Dictionary::TF_IDF
-
#tf_idf Dictionary::TF_IDF
-
#to_s TokenTrieNER::Code
-
#token_evaluate Normalizer
-
#token_score RegExpNER
-
#token_score Normalizer
-
#token_trie PatternRelExt
-
#token_types Tokenizer
-
#tokenize Tokenizer
-
tokenize Token
-
tokenize TokenTrieNER
-
tokens SpaCy
-
tokens NERFeatures
-
#total_terms Dictionary::TF_IDF
-
#train NERFeatures
-
#transform Tokenizer::Transform
-
transform Transformed
-
transform_index PatternRelExt
-
transform_key PatternRelExt
-
#transformation_stack Transformed
-
#transformed_segments Transformed
-
#type Tokenizer
-
#type PatternRelExt
-
#type NGramPrefixDictionary
-
#type TokenTrieNER
-
#type TokenTrieNER::Code
-
#type AnnotID
-
weighted_features BagOfWords
-
#weights Dictionary::KL
-
#weights Dictionary::TF_IDF
-
#window NERFeatures
-
with_transform Transformed
-
#words String
-
words BagOfWords