本文整理汇总了Python中nltk.tokenize.WordPunctTokenizer类的典型用法代码示例。如果您正苦于以下问题:Python WordPunctTokenizer类的具体用法?Python WordPunctTokenizer怎么用?Python WordPunctTokenizer使用的例子?那么恭喜您, 这里精选的类代码示例或许可以为您提供帮助。
在下文中一共展示了WordPunctTokenizer类的20个代码示例,这些例子默认根据受欢迎程度排序。您可以为喜欢或者感觉有用的代码点赞,您的评价将有助于我们的系统推荐出更棒的Python代码示例。
示例1: number_of_different_words
def number_of_different_words(self):
# TODO: Stemming, then move to language specific classes
tokenizer = WordPunctTokenizer()
words = tokenizer.tokenize(self.text.strip())
only_textual_words = filter(unicode.isalpha, words)
return len(set(only_textual_words))
开发者ID:aufziehvogel,项目名称:sprakit,代码行数:7,代码来源:text_statistics.py
示例2: message_to_wordlist
def message_to_wordlist(message, lemmas_bool, remove_stopwords=False):
# Function to convert a document to a sequence of words,
# optionally removing stop words. Returns a list of words.
#
# 1. Remove HTML
#review_text = BeautifulSoup(review).get_text()
#
# 2. Remove messages numbers
message_text = re.sub(">>\d+","", message)
message_text = message_text.lower()
message_text = re.sub(u"ё", 'e', message_text, re.UNICODE)
message_text = clean_str(message_text)
tokenizer = WordPunctTokenizer()
# 3. Convert words to lower case and split them
words = tokenizer.tokenize(message_text)
lemmas = []
# 4. Optionally remove stop words (false by default)
if remove_stopwords:
stops = set(stopwords.words("english"))
words = [w for w in words if not w in stops]
if lemmas_bool == 'l':
for word in words:
word_parsed = morph.parse(word)
if len(word_parsed) > 0:
lemmas.append(word_parsed[0].normal_form)
elif lemmas_bool == 's':
for word in words:
word = stemmer.stem(word)
if len(word) > 0:
lemmas.append(word)
else:
lemmas = words
# 5. Return a list of words
return(lemmas)
开发者ID:denis-gordeev,项目名称:CNN-aggression-RU,代码行数:34,代码来源:train_tensorflow.py
示例3: tfIdf
def tfIdf():
TFIDF_MIN_SCORE = 100
import nltk
from nltk.tokenize import WordPunctTokenizer
tokenizer = WordPunctTokenizer()
collection = initialize_collection('documents')
docs = collection.find()
tfidf = []
idfMap = create_idf_map()
docs = collection.find()
for d in docs:
tfMap = {}
for word in set(tokenizer.tokenize(d['content'].lower())):
if word not in tfMap:
tfMap[word] = 1
else:
tfMap[word] += 1
tfIdfValues = []
for word in set(tokenizer.tokenize(d['content'].lower())):
if (tfMap[word] * 1000 / idfMap[word]) > TFIDF_MIN_SCORE:
tfIdfValues.append((word, tfMap[word] * 1000 / idfMap[word]))
tfIdfValues = sorted(tfIdfValues, key = lambda x : x[1], reverse = True)
d['tfidf'] = tfIdfValues
tfidf.append({'d' : d,
'tfidf' : tfIdfValues})
collection.save(d)
genFreq = generaral_frequency(idfMap)
return render_template("tfidf.html", documents = tfidf)
开发者ID:tempflip,项目名称:szakdoga,代码行数:31,代码来源:flask1.py
示例4: extract_nl_text
def extract_nl_text(ms):
"""
Extracts and tokenizes text from malware sample object
:param ms: MalwareSample object
:return: list of tokenized strings found in malware sample object's internal strings list
"""
wpt = WordPunctTokenizer()
all_tokenized_strings_in_ms = []
inside_xml_privileges = False
for s in ms.strings:
if 'requestedPrivileges' in s or 'This program cannot be run in DOS mode' in s:
continue
elif inside_xml_privileges:
continue
elif '<assembly xmlns' in s:
inside_xml_privileges = True
continue
elif '</assembly>' in s:
inside_xml_privileges = False
continue
tokenized_string = []
tokens = wpt.tokenize(s)
if tokens:
for t in tokens:
if wordnet.synsets(t) and len(t) > 3: # had to use length to eliminate false positives
tokenized_string.extend(tokens)
break
if tokenized_string:
all_tokenized_strings_in_ms.append(tokenized_string)
return all_tokenized_strings_in_ms
开发者ID:danzek,项目名称:nlhbi-malware-extractor,代码行数:32,代码来源:getNLindicators.py
示例5: tokens
def tokens(self):
"""Tokenize the text.
"""
tokenizer = WordPunctTokenizer()
# Get token character spans.
spans = list(tokenizer.span_tokenize(self.text))
# Materialize the token stream.
tokens = [self.text[c1:c2] for c1, c2 in spans]
tags = pos_tag(tokens)
return [
Token(
token=token.lower(),
char1=c1,
char2=c2,
pos=pos,
)
for (c1, c2), token, (_, pos) in
zip(spans, tokens, tags)
]
开发者ID:davidmcclure,项目名称:literary-interior,代码行数:26,代码来源:text.py
示例6: tokenize_words
def tokenize_words(sentence):
"""
:param sentence:
:return: list of words in sentence
"""
tokenizer = WordPunctTokenizer()
return tokenizer.tokenize(sentence)
开发者ID:paulzin,项目名称:NltkTokenizerDemo,代码行数:7,代码来源:nltk_tokenizer.py
示例7: class1
def class1():
import nltk
from nltk.tokenize import WordPunctTokenizer
docId = request.args.get('d')
tokenizer = WordPunctTokenizer()
collection = initialize_collection('documents')
featuresets = []
tagSet = set()
for d in collection.find():
bagOfWords = bag_of_words(tokenizer.tokenize(d['content']))
if 'tags' not in d: continue
for tag in d['tags']:
featuresets.append((bagOfWords, tag))
tagSet.add(tag)
classifier = nltk.NaiveBayesClassifier.train(featuresets)
d = collection.find_one({'_id' : ObjectId(docId)})
#classifier.show_most_informative_features(100)
cl = classifier.prob_classify(bag_of_words(tokenizer.tokenize(d['content'])))
probs = []
for tag in tagSet:
probs.append((tag, round(cl.prob(tag)*100) ))
classifier.show_most_informative_features(n=20)
probs = sorted(probs, key = lambda x : x[1], reverse = True)
return render_template('class1.html', probs = probs, d=d)
开发者ID:tempflip,项目名称:szakdoga,代码行数:27,代码来源:flask1.py
示例8: tokenize
def tokenize(text):
"""Tokenize a raw text.
Args:
text (str)
Returns: list of {token, char1, char2, pos}
"""
tokenizer = WordPunctTokenizer()
# Get token character spans.
spans = list(tokenizer.span_tokenize(text))
# Materialize the token stream.
tokens = [text[c1:c2] for c1, c2 in spans]
# Tag parts-of-speech.
tags = pos_tag(tokens)
return [
dict(
token=token.lower(),
char1=c1,
char2=c2,
pos=pos,
)
for (c1, c2), token, (_, pos) in
zip(spans, tokens, tags)
]
开发者ID:davidmcclure,项目名称:stacks,代码行数:32,代码来源:utils.py
示例9: TextProcessor
def TextProcessor(src, tgt, low=True, num=True):
print "processing "+src
if low==True:
print "lowercasing.."
if num==True:
print "removing numeric.."
srcfile = codecs.open(src,"r","utf-8")
tgtfile = codecs.open(tgt,"w","utf-8")
word_punct_tokenizer = WordPunctTokenizer()
linecount=0
for line in srcfile:
linecount+=1
line = word_punct_tokenizer.tokenize(line)
if low==True:
for i in range(0,len(line)):
line[i] = line[i].lower()
if num==True:
for i in range(0,len(line)):
if line[i].isnumeric()==True:
line[i] = "<number>"
tgtfile.write(listtostring(line))
srcfile.close()
tgtfile.close()
print "done processing "+str(linecount)+" lines!!"
开发者ID:apsarath,项目名称:pyNN,代码行数:30,代码来源:TextProcessor.py
示例10: clean_data
def clean_data(input_file_name, output_file_name):
def clean_word(word):
word = word.encode('ascii', 'ignore')
word = word.lower()
word = re.sub(r'(\S)\1+', r'\1\1', word) # normalize repeated characters to two
word = re.sub(r'(\S\S)\1+', r'\1\1', word)
if re.search(r'((([A-Za-z]{3,9}:(?:\/\/)?)(?:[-;:&=\+\$,\w][email protected])?[A-Za-z0-9.-]+|(?:www.|[-;:&=\+\$,\w][email protected])[A-Za-z0-9.-]+)((?:\/[\+~%\/.\w-]*)?\??(?:[-\+=&;%@.\w]*)#?(?:[\w]*))?)',word) is not None:
word = 'GENERIC_HTTP'
return word
tokenizer = WordPunctTokenizer()
data = []
with open(input_file_name) as input_file:
for sentences, label in json.load(input_file):
cleaned_sentences = []
for sentence in sentences:
cleaned_sentence = " ".join(map(clean_word, sentence.split()))
cleaned_sentence = tokenizer.tokenize(cleaned_sentence)
cleaned_sentences.append(cleaned_sentence)
data.append([cleaned_sentences, label])
with codecs.open(output_file_name, 'w', encoding='utf-8') as output_file:
json.dump(data, output_file)
开发者ID:BKJackson,项目名称:txtnets,代码行数:26,代码来源:prepare_amazon_sentiment.py
示例11: clean_data
def clean_data(input_file_name, output_file_name):
def clean_word(word):
word = word.lower()
word = word.replace('&','&').replace('<','<').replace('>','>').replace('"','"').replace(''',"'")
word = re.sub(r'(\S)\1+', r'\1\1', word) # normalize repeated characters to two
word = re.sub(r'(\S\S)\1+', r'\1\1', word)
word = word.encode('ascii', 'ignore')
if re.search(r'((([A-Za-z]{3,9}:(?:\/\/)?)(?:[-;:&=\+\$,\w][email protected])?[A-Za-z0-9.-]+|(?:www.|[-;:&=\+\$,\w][email protected])[A-Za-z0-9.-]+)((?:\/[\+~%\/.\w-]*)?\??(?:[-\+=&;%@.\w]*)#?(?:[\w]*))?)',word) is not None:
word = 'GENERIC_HTTP'
return word.encode('ascii', 'ignore')
tokenizer = WordPunctTokenizer()
with gzip.open(input_file_name) as input_file:
with gzip.open(output_file_name, 'w') as output_file:
for line in input_file:
sentences, score = json.loads(line)
cleaned_sentences = []
for sentence in sentences:
cleaned_sentence = " ".join(map(clean_word, sentence.split()))
cleaned_sentences.append(tokenizer.tokenize(cleaned_sentence))
json.dump([cleaned_sentences, score], output_file)
output_file.write("\n")
开发者ID:BKJackson,项目名称:txtnets,代码行数:27,代码来源:prepare_amazon_reviews.py
示例12: get_words_without_stopwords
def get_words_without_stopwords(self, text):
stopwords = nltk.corpus.stopwords.words('english')
stopwords.extend(string.punctuation)
stopwords.append('')
tokenizer = WordPunctTokenizer()
tokens = [token.lower().strip(string.punctuation) for token in tokenizer.tokenize(text) \
if token.lower().strip(string.punctuation) not in stopwords]
return tokens
开发者ID:elangovana,项目名称:Aristo,代码行数:8,代码来源:text_analyser.py
示例13: extract_words
def extract_words(text):
stemmer = PorterStemmer()
tokenizer = WordPunctTokenizer()
tokens = tokenizer.tokenize(text)
result = [stemmer.stem(x.lower()) for x in tokens if x not in stopwords.words('english') and len(x) > 1]
return result
开发者ID:teonghan,项目名称:crimefeeder,代码行数:8,代码来源:crimeclassifier_v2.py
示例14: get_tokens
def get_tokens(sentence):
"""
Tokenizes a list of sentences
:param sentence: list of sentences
:return: list of tokenized sentences
"""
tokenizer = WordPunctTokenizer()
return tokenizer.tokenize(sentence)
开发者ID:zweiss,项目名称:RC_Readability_Calculator,代码行数:9,代码来源:nlp.py
示例15: getBigram
def getBigram(haystack):
tokenizer = WordPunctTokenizer()
words = tokenizer.tokenize(haystack)
bcf = BigramCollocationFinder.from_words(words)
stopset = set(stopwords.words('english'))
filter_stops = lambda w: len(w) < 3 or w in stopset
bcf.apply_word_filter(filter_stops)
return bcf.nbest(BigramAssocMeasures.likelihood_ratio, 4)
开发者ID:blorenz,项目名称:cms,代码行数:9,代码来源:seo.py
示例16: change_db2
def change_db2(text, origin_dict, id):
print origin_dict
tokens_ar = []
word_punct_tokenizer = WordPunctTokenizer()
for token in word_punct_tokenizer.span_tokenize(origin_dict):
tokens_ar.append(token)
for line in text.split("\n"):
markup_error_line = line.split(';')
print "MARKUP", markup_error_line
convert_coord_2dbformat(markup_error_line, tokens_ar, id)
开发者ID:Alenush,项目名称:zhivoeslovo,代码行数:10,代码来源:download_data2db.py
示例17: tokenize
def tokenize(text):
tokens = tokenizer.tokenize(text)
wordtokenizer = WordPunctTokenizer()
wlist =[]
for token in tokens:
wtoken = wordtokenizer.tokenize(token)
wlist = wlist+wtoken
stems = stem_tokens(wlist, stemmer)
return stems
开发者ID:Derenas,项目名称:Master-Thesis,代码行数:10,代码来源:gensimtest.py
示例18: extract_words
def extract_words(text):
stemmer = PorterStemmer()
tokenizer = WordPunctTokenizer()
tokens = tokenizer.tokenize(text)
bigram_finder = BigramCollocationFinder.from_words(tokens)
bigrams = bigram_finder.nbest(BigramAssocMeasures.chi_sq, 500)
for bigram_tuple in bigrams:
x = "%s %s" % bigram_tuple
tokens.append(x)
result = [stemmer.stem(x.lower()) for x in tokens if x not in stopwords.words('english') and len(x) > 1]
return result
开发者ID:beeblook,项目名称:hotel-reviews-sentiment-ranking-analysis,代码行数:11,代码来源:classifier_nb.py
示例19: extract_bigrams
def extract_bigrams(text):
text = remove_stopwords(text)
tokenizer = WordPunctTokenizer()
tokens = [token for token in set(tokenizer.tokenize(text)) if
not is_number(token) and (is_valid_token(token) or is_name(token))]
bigram_finder = BigramCollocationFinder.from_words(tokens)
bigrams = bigram_finder.nbest(BigramAssocMeasures.dice, 500)
for bigram_tuple in bigrams:
x = "%s %s" % bigram_tuple
tokens.append(x)
result = [x.lower() for x in tokens if x not in stopwords.words("english") and len(x) > 3]
return result
开发者ID:ajaybhat,项目名称:email-categorization,代码行数:12,代码来源:ml_util.py
示例20: you_collocations
def you_collocations(raw):
tokenizer = WordPunctTokenizer()
tokens = tokenizer.tokenize(raw)
bigrams = [(tokens[i], tokens[i +1]) for i in range(len(tokens)-1)]
collocations = [(t1, t2) for (t1, t2) in bigrams if t1 == "you" or t1 == 'your']
trigrams = [(tokens[i], tokens[i +1], tokens[i+2]) for i in range(len(tokens)-2)]
trilocations = [(t1, t2, t3) for (t1, t2, t3) in trigrams if t1 == "you" or t1 == 'your']
return collocations, trilocations
开发者ID:rmomizo,项目名称:RePort_Bot,代码行数:12,代码来源:ePortfolio.py
注:本文中的nltk.tokenize.WordPunctTokenizer类示例由纯净天空整理自Github/MSDocs等源码及文档管理平台,相关代码片段筛选自各路编程大神贡献的开源项目,源码版权归原作者所有,传播和使用请参考对应项目的License;未经允许,请勿转载。 |
请发表评论