• 设为首页
  • 点击收藏
  • 手机版
    手机扫一扫访问
    迪恩网络手机版
  • 关注官方公众号
    微信扫一扫关注
    迪恩网络公众号

Python re.purge函数代码示例

原作者: [db:作者] 来自: [db:来源] 收藏 邀请

本文整理汇总了Python中re.purge函数的典型用法代码示例。如果您正苦于以下问题:Python purge函数的具体用法?Python purge怎么用?Python purge使用的例子?那么恭喜您, 这里精选的函数代码示例或许可以为您提供帮助。



在下文中一共展示了purge函数的20个代码示例,这些例子默认根据受欢迎程度排序。您可以为喜欢或者感觉有用的代码点赞,您的评价将有助于我们的系统推荐出更棒的Python代码示例。

示例1: finditer

def finditer(content, encodings, charset, min_size):
    '''Generator function that iterates over all string matches inside the given content which are at least
    min_size characters long.

    @param    content    Binary content to search in
    @param    encodings  Dictionary of encoding functions
    @param    charset    An interable object containing the characters to consider as part of a string
    @param    min_size   Minimal string size to consider as a string match

    @return A tuple containing the match offset in content, encoding name, encoding key and the deobfuscated
            string reconstructed from the blob found
    '''

    # iterate over available encoding fucntions
    for encoding_name, (encoding_function, encoding_range) in encodings.items():

        # iterate over all keys in range for that encoding function
        for key in encoding_range:
            encoded_charset = encoding_function(charset, key)

            pattern = '[%s]{%d,}' % (re.escape(encoded_charset), min_size)

            for match in re.finditer(pattern, content):
                # deobfuscation: reconstruct the original string
                deobf = ''.join(charset[encoded_charset.index(c)] for c in match.group(0))

                yield (match.start(0), encoding_name, key, deobf)

        # cleanup regex cache once in a while
        re.purge()
开发者ID:shrek0,项目名称:xstrings,代码行数:30,代码来源:xstrings.py


示例2: color

def color ( adjoining_words_i, data, balises ):
	"""Colorie les groupes de mots contigus dans une page web"""
	n = len(adjoining_words_i) + 1
	
	# on commence par les groupes les plus longs
	for i in xrange( n, 1, -1 ):
		
		# pour chaque groupe de mots
		for j in adjoining_words_i[i]:
			
			text = u'(\A|\W)(%s)(\W|\Z)'%( string.join([j[0][k] for k in range(0,i)] ,'(?:(?:</span>\W?)|\W)') )
			pattern1 = re.compile(text, re.I|re.U|re.S)		
			replace = u'\g<1><span class="%s" style="color:blue; background-color:grey;">\g<2></span>\g<3>'%(string.join(j[0],""))
			data = pattern1.sub(replace, data)
	
	re.purge()
	# recherche les emplacements sauvegardés des balises
	data_color = u''	
	flag3 = re.compile( u'#([0-9]+?)#', re.I|re.U|re.S )
	m = flag3.finditer( data )
	k = 0
	
	# remet les balises dans la chaine
	for j in m:
		data_color += data[k:j.start()] + balises[j.group(1)]
		k = j.end()
		
	data_color += data[k:]
	
	return data_color
开发者ID:Samael609,项目名称:PompOView,代码行数:30,代码来源:tp.py


示例3: get_info

def get_info(Term, Subject):
	url = "https://ssbp.mycampus.ca/prod/bwckschd.p_get_crse_unsec?TRM=U&term_in=" + Term + "&sel_subj=dummy&sel_day=dummy&sel_schd=dummy&sel_insm=dummy&sel_camp=dummy&sel_levl=dummy&sel_sess=dummy&sel_instr=dummy&sel_ptrm=dummy&sel_attr=dummy&sel_subj=" + Subject + "&sel_crse=&sel_title=&sel_from_cred=&sel_to_cred=&sel_camp=UON&begin_hh=0&begin_mi=0&begin_ap=a&end_hh=0&end_mi=0&end_ap=a"
	htmltext = urllib.urlopen(url).read();
	regex = '<TH CLASS="ddheader" scope="col" >(.+?)<BR><BR></TH>'
	pattern = re.compile(regex)
	courses = re.split(pattern, htmltext)
	re.purge()
	for course in courses:
		regex = '<TD CLASS="dbdefault">(.+?)</TD>\n<TD CLASS="dbdefault">(.+?)</TD>\n<TD CLASS="dbdefault">(.+?) - (.+?)</TD>\n<TD CLASS="dbdefault">(.+?)</TD>\n<TD CLASS="dbdefault">(.+?)</TD>\n<TD CLASS="dbdefault">(.+?) - (.+?)</TD>\n<TD CLASS="dbdefault">(.+?)</TD>\n<TD CLASS="dbdefault">(.+?) \(<ABBR title= "Primary">P</ABBR>\)</TD>'
		regex2 = '<TD CLASS="dbdefault">(.+?)</TD>\n<TD CLASS="dbdefault">(.+?)</TD>\n<TD CLASS="dbdefault">(.+?) - (.+?)</TD>\n<TD CLASS="dbdefault">(.+?)</TD>\n<TD CLASS="dbdefault">(.+?)</TD>\n<TD CLASS="dbdefault">(.+?) - (.+?)</TD>\n<TD CLASS="dbdefault">(.+?)</TD>\n<TD CLASS="dbdefault"><ABBR title = "To Be Announced">(.+?)</ABBR></TD>'
		regex3 = '<TD CLASS="dbdefault">(.+?)</TD>\n<TD CLASS="dbdefault">(.+?)</TD>\n<TD CLASS="dbdefault"><ABBR title = "To Be Announced">(.+?)</ABBR></TD>\n<TD CLASS="dbdefault">(.+?)</TD>\n<TD CLASS="dbdefault"><ABBR title = "To Be Announced">(.+?)</ABBR></TD>\n<TD CLASS="dbdefault">(.+?) - (.+?)</TD>\n<TD CLASS="dbdefault">(.+?)</TD>\n<TD CLASS="dbdefault"><ABBR title = "To Be Announced">(.+?)</ABBR></TD>'
		regex4 = '<TD CLASS="dbdefault">(.+?)</TD>\n<TD CLASS="dbdefault">(.+?)</TD>\n<TD CLASS="dbdefault">(.+?) - (.+?)</TD>\n<TD CLASS="dbdefault">(.+?)</TD>\n<TD CLASS="dbdefault">(.+?)</TD>\n<TD CLASS="dbdefault">(.+?) - (.+?)</TD>\n<TD CLASS="dbdefault">(.+?)</TD>\n<TD CLASS="dbdefault">(.+?) \(<ABBR title= "Primary">P</ABBR>\)(.+)?</TD>'
		pattern = re.compile(regex)
		pattern2 = re.compile(regex2)
		pattern3 = re.compile(regex3)
		pattern4 = re.compile(regex4)
		entries = re.findall(pattern3, course)				#this pattern is for courses that do not have a start time or class assigned
		if entries:											 
			print entries
		else:
			entries = re.findall(pattern2, course)			#this pattern is for instructor TBA
			if entries:
				print entries
			else:
				entries = re.findall(pattern, course)		#this pattern is for default structure of courses
				if entries:
					print entries
				else:
					entries = re.findall(pattern4, course)	#this pattern returns two values for instructor
					print entries
开发者ID:Adge2k,项目名称:MyCampus-WebScraper,代码行数:30,代码来源:WebScraper.py


示例4: dash_R_cleanup

def dash_R_cleanup(fs, ps, pic):
    import gc, copy_reg
    import _strptime, linecache, dircache
    import urlparse, urllib, urllib2, mimetypes, doctest
    import struct, filecmp
    from distutils.dir_util import _path_created

    # Restore some original values.
    warnings.filters[:] = fs
    copy_reg.dispatch_table.clear()
    copy_reg.dispatch_table.update(ps)
    sys.path_importer_cache.clear()
    sys.path_importer_cache.update(pic)

    # Clear assorted module caches.
    _path_created.clear()
    re.purge()
    _strptime._regex_cache.clear()
    urlparse.clear_cache()
    urllib.urlcleanup()
    urllib2.install_opener(None)
    dircache.reset()
    linecache.clearcache()
    mimetypes._default_mime_types()
    struct._cache.clear()
    filecmp._cache.clear()
    doctest.master = None

    # Collect cyclic trash.
    gc.collect()
开发者ID:alkorzt,项目名称:pypy,代码行数:30,代码来源:regrtest.py


示例5: main

def main():
    times = {}
    html = urllib2.urlopen('http://example.webscraping.com/places/default/view/United-Kingdom-239').read()
    NUM_ITERATIONS = 1000 # number of times to test each scraper
    for name, scraper in ('Regular expressions', regex_scraper), ('Beautiful Soup', beautiful_soup_scraper), ('Lxml', lxml_scraper):
        times[name] = []
        # record start time of scrape
        start = time.time()
        for i in range(NUM_ITERATIONS):
            if scraper == regex_scraper:
                # the regular expression module will cache results
                # so need to purge this cache for meaningful timings
                re.purge() 
            result = scraper(html)

            # check scraped result is as expected
            assert(result['area'] == '244,820 square kilometres')
            times[name].append(time.time() - start)
        # record end time of scrape and output the total
        end = time.time()
        print('{}: {:.2f} seconds'.format(name, end - start))

    writer = csv.writer(open('times.csv', 'w'))
    header = sorted(times.keys())
    writer.writerow(header)
    for row in zip(*[times[scraper] for scraper in header]):
        writer.writerow(row)
开发者ID:Rockyzsu,项目名称:base_function,代码行数:27,代码来源:bs_lxml_reg_perf.py


示例6: clear_cache

 def clear_cache(self):
     try:
         re.purge()
         dircache.reset()
         tiedobj.reset()
     except Exception, err:
         sys.stderr.write('Crond.clear_cache(): %s\n' % err)
开发者ID:Hiroyuki-Nagata,项目名称:saku,代码行数:7,代码来源:crond.py


示例7: retrieve_devpaths

def retrieve_devpaths():
    pipe = Popen('si projectinfo --devpaths --noacl --noattributes --noshowCheckpointDescription --noassociatedIssues --project="%s"' % sys.argv[1], shell=True, bufsize=1024, stdout=PIPE)
    devpaths = pipe.stdout.read()
    devpaths = devpaths [1:]
    devpaths_re = re.compile('    (.+) \(([0-9][\.0-9]+)\)\n')
    devpath_col = devpaths_re.findall(devpaths)
    re.purge()
    devpath_col.sort(key=lambda x: map(int, x[1].split('.'))) #order development paths by version
    return devpath_col
开发者ID:XenuIsWatching,项目名称:integrity2git,代码行数:9,代码来源:mks_checkpoints_to_git.py


示例8: purge

def purge():
    """re.purge: Purge internal regular expressions cache."""

    def _cache_empty():
        return not getattr(re, '_cache')

    re.match('', '')
    cache_created = not _cache_empty()
    re.purge()
    return cache_created and _cache_empty() and "empty cache"
开发者ID:yoeo,项目名称:pyhow,代码行数:10,代码来源:re.py


示例9: test_regex_equality_nocache

    def test_regex_equality_nocache(self):
        pattern = r'^(?:[a-z0-9\.\-]*)://'
        left = RegexValidator(pattern)
        re.purge()
        right = RegexValidator(pattern)

        self.assertEqual(
            left,
            right,
        )
开发者ID:01-,项目名称:django,代码行数:10,代码来源:tests.py


示例10: get_skips

    def get_skips(self, line):
        skip_points = []
        for r in self.skip_rules:
            pattern = '('+r[0]+')('+r[1]+')'
            matchobjs = re.finditer(pattern, line)
            for i in matchobjs:
                skip_points.append(i.end() )

        re.purge()
        return skip_points
开发者ID:FrITA,项目名称:frita,代码行数:10,代码来源:segment.py


示例11: get_breaks

    def get_breaks(self, line):
        break_points = []
        for r in self.break_rules:
            pattern = '('+r[0]+')('+r[1]+')'
            matchobjs = re.finditer(pattern, line)
            for i in matchobjs:
                break_points.append(i.end() )

        re.purge()
        return break_points
开发者ID:FrITA,项目名称:frita,代码行数:10,代码来源:segment.py


示例12: check

 def check(self, pattern):
     self.model.clear()
     if not pattern: return False
     try:
         re.compile(pattern, self.insertFlags())
         re.purge()
         return True
     except re.error as rerr:
         self.model.showError(str(rerr))
         return False
开发者ID:Stamoulohta,项目名称:PyRex,代码行数:10,代码来源:stdRe.py


示例13: getRegexpFeatures

def getRegexpFeatures(dct, number_of_words_per_type, number_of_words, select = None):
    it = list()
    for (mt, sen) in dct.iteritems():
        it.append((len(sen), mt, sen))
    it.sort(reverse=False)
    itt = list()
    for (l, mt, sen) in it:
        random.shuffle(sen)
        itt.append((l, mt, sen[0:1000]))
    regexps = dict()
    ret = list()
    types = list()
    for (_, meme, _sentences) in itt:
        types.extend([meme for _ in _sentences])
    types = [types]
    #glob = regExpChooser()
    #glob.add_types(types)
    for (_, meme_type, sentences) in it:
        if select != None and meme_type != select:
            continue
        regexps[meme_type] = cluster(sentences, meme_type)
        N = len(regexps[meme_type])
        n = 0
        start = time.time()
        loc = regExpChooser()
        loc.add_types(replaceNotEqual(types, meme_type))
        for regexp in regexps[meme_type]:
            re.purge()
            n += 1
            sys.stdout.write(
                "\r[{0}] {1}/{2} RE in {3} s. ({4})".format(
                    meme_type,
                    n,
                    N,
                    round(time.time() - start),
                    regexp
                ))
            sys.stdout.flush()
            compiled = re.compile(regexp)
            search_result = list()
            for (_, meme, _sentences) in itt: 
                for sent in _sentences:
                    search_result.append(
                        1 if compiled.search(sent.lower()) != None else 0)
            loc.add_regexp(regexp, search_result)
            #glob.add_regexp(regexp, search_result)
        selection = loc.getBest(number_of_words_per_type)
        ret.extend(selection)
        print("\r[{0}] Regular expressions selected in {1} seconds. (best: {2})".format(
            meme_type,
            time.time() - start,
            selection[0])
        )
    #ret.extend(glob.getBest(number_of_words))
    return ret
开发者ID:mic47,项目名称:Generate-me-2,代码行数:55,代码来源:regexpfeatures.py


示例14: markdownify_content

 def markdownify_content(self):
     self.content = re.sub(r'({{% question) "(.*)"(\s*%}})',r'### \2', self.content)
     re.purge()
     self.content = re.sub(r'{{< relref "(\w*)\.md[#\w\-éèà]*"\s*>}}\s*',r'\1', self.content)
     re.purge()
     self.content = re.sub(r'{{% (\w*) "(.*)" *%}}([\s\S]*?){{% \/\1 %}}',r'*\2*\3', self.content)
     re.purge()
     self.content = re.sub(r'\* Exemple : <.*\)',r'', self.content)
     re.purge()
     self.content = re.sub(r'(#+)\s',r'\1# ', self.content)
     re.purge()
开发者ID:numerique-gouv,项目名称:numerique.gouv.fr,代码行数:11,代码来源:import-politique-de-contribution-open-source.py


示例15: remove_links

def remove_links(s, replace_by):
	#quita url www.algo.com/djj
    re.purge()
    temp = re.compile(r"\s*www\.\. \w+\.(com|net|me|org)?(\s|/*[-\w+&@#/%!?=~_:.\[\]()0-9]*)")
    s = temp.sub(replace_by, s)
    #quita http://
    temp = re.compile(r"(((http|ftp|https)://\. |(http|ftp|https)://\.)[-/\w.]*)")
    s = temp.sub(replace_by, s)
    temp = re.compile(r"\w+/\w")
    s = temp.sub(replace_by, s)
    return s
开发者ID:alexgalo,项目名称:Scripts_Terminal,代码行数:11,代码来源:utility_functions.py


示例16: bench_regex_compile

def bench_regex_compile(loops, regexes):
    range_it = xrange(loops)
    t0 = perf.perf_counter()

    for _ in range_it:
        for regex, flags in regexes:
            re.purge()
            # ignore result (compiled regex)
            re.compile(regex, flags)

    return perf.perf_counter() - t0
开发者ID:Yaspee,项目名称:performance,代码行数:11,代码来源:bm_regex_compile.py


示例17: test_regex_compile

def test_regex_compile(count, timer):
    regexes = capture_regexes()
    times = []

    for _ in xrange(count):
        t0 = timer()
        for regex, flags in regexes:
            re.purge()
            re.compile(regex, flags)
        t1 = timer()
        times.append(t1 - t0)
    return times
开发者ID:bennn,项目名称:retic_performance,代码行数:12,代码来源:bm_regex_compile.py


示例18: process

def process(self, context, collection):
    '''
        Process collection, send names to rename and shared sort.
    '''

    # compare
    compare = []

    # clean
    clean = []

    # clean duplicates
    for name in collection:

        # remove duplicates
        if name[3][0] not in compare:

            # append
            compare.append(name[3][0])
            clean.append(name)

    # done with collection
    collection.clear()

    # name
    for i, name in enumerate(clean):
        rename(self, context, name, i)

    # randomize names (prevents conflicts)
    for name in clean:

        # randomize name
        name[3][0].name = str(random())

    # is shared sort or shared count
    if context.window_manager.BatchShared.sort or context.window_manager.BatchShared.count:

        # sort
        shared.main(self, context, clean, context.window_manager.BatchShared)

    # isnt shared sort or shared count
    else:

        # apply names
        for name in clean:
            name[3][0].name = name[1]

            # count
            if name[1] != name[2]:
                self.count += 1

    # purge re
    re.purge()
开发者ID:trentinfrederick,项目名称:name-panel,代码行数:53,代码来源:auto.py


示例19: dash_R_cleanup

def dash_R_cleanup(fs, ps, pic, zdc, abcs):
    import gc, copy_reg
    import _strptime, linecache
    dircache = test_support.import_module('dircache', deprecated=True)
    import urlparse, urllib, urllib2, mimetypes, doctest
    import struct, filecmp
    from distutils.dir_util import _path_created

    # Clear the warnings registry, so they can be displayed again
    for mod in sys.modules.values():
        if hasattr(mod, '__warningregistry__'):
            del mod.__warningregistry__

    # Restore some original values.
    warnings.filters[:] = fs
    copy_reg.dispatch_table.clear()
    copy_reg.dispatch_table.update(ps)
    sys.path_importer_cache.clear()
    sys.path_importer_cache.update(pic)
    try:
        import zipimport
    except ImportError:
        pass # Run unmodified on platforms without zipimport support
    else:
        zipimport._zip_directory_cache.clear()
        zipimport._zip_directory_cache.update(zdc)

    # clear type cache
    sys._clear_type_cache()

    # Clear ABC registries, restoring previously saved ABC registries.
    for abc, registry in abcs.items():
        abc._abc_registry = registry.copy()
        abc._abc_cache.clear()
        abc._abc_negative_cache.clear()

    # Clear assorted module caches.
    _path_created.clear()
    re.purge()
    _strptime._regex_cache.clear()
    urlparse.clear_cache()
    urllib.urlcleanup()
    urllib2.install_opener(None)
    dircache.reset()
    linecache.clearcache()
    mimetypes._default_mime_types()
    filecmp._cache.clear()
    struct._clearcache()
    doctest.master = None

    # Collect cyclic trash.
    gc.collect()
开发者ID:carlosrcjunior,项目名称:BCC-2s13-PI4-WebCrawler,代码行数:52,代码来源:regrtest.py


示例20: dash_R_cleanup

def dash_R_cleanup(fs, ps, pic, abcs):
    import gc, copy_reg
    import _strptime, linecache
    dircache = test_support.import_module('dircache', deprecated=True)
    import urlparse, urllib, urllib2, mimetypes, doctest
    import struct, filecmp
    from distutils.dir_util import _path_created

    # Clear the warnings registry, so they can be displayed again
    for mod in sys.modules.values():
        if hasattr(mod, '__warningregistry__'):
            del mod.__warningregistry__

    # Restore some original values.
    warnings.filters[:] = fs
    copy_reg.dispatch_table.clear()
    copy_reg.dispatch_table.update(ps)
    sys.path_importer_cache.clear()
    sys.path_importer_cache.update(pic)

    # clear type cache
    sys._clear_type_cache()

    # Clear ABC registries, restoring previously saved ABC registries.
    for abc, registry in abcs.items():
        abc._abc_registry = registry.copy()
        abc._abc_cache.clear()
        abc._abc_negative_cache.clear()

    # Clear assorted module caches.
    _path_created.clear()
    re.purge()
    _strptime._regex_cache.clear()
    urlparse.clear_cache()
    urllib.urlcleanup()
    urllib2.install_opener(None)
    dircache.reset()
    linecache.clearcache()
    mimetypes._default_mime_types()
    filecmp._cache.clear()
    struct._clearcache()
    doctest.master = None

    if _llvm:
        code_types = (types.CodeType, types.FunctionType, types.MethodType)
        for obj in gc.get_objects():
            if isinstance(obj, code_types):
                _llvm.clear_feedback(obj)

    # Collect cyclic trash.
    gc.collect()
开发者ID:ianloic,项目名称:unladen-swallow,代码行数:51,代码来源:regrtest.py



注:本文中的re.purge函数示例由纯净天空整理自Github/MSDocs等源码及文档管理平台,相关代码片段筛选自各路编程大神贡献的开源项目,源码版权归原作者所有,传播和使用请参考对应项目的License;未经允许,请勿转载。


鲜花

握手

雷人

路过

鸡蛋
该文章已有0人参与评论

请发表评论

全部评论

专题导读
上一篇:
Python re.re函数代码示例发布时间:2022-05-26
下一篇:
Python re.match函数代码示例发布时间:2022-05-26
热门推荐
阅读排行榜

扫描微信二维码

查看手机版网站

随时了解更新最新资讯

139-2527-9053

在线客服(服务时间 9:00~18:00)

在线QQ客服
地址:深圳市南山区西丽大学城创智工业园
电邮:jeky_zhao#qq.com
移动电话:139-2527-9053

Powered by 互联科技 X3.4© 2001-2213 极客世界.|Sitemap