当前位置: 首页>>代码示例>>Python>>正文


Python string.maketrans函数代码示例

本文整理汇总了Python中string.maketrans函数的典型用法代码示例。如果您正苦于以下问题:Python maketrans函数的具体用法?Python maketrans怎么用?Python maketrans使用的例子?那么恭喜您, 这里精选的函数代码示例或许可以为您提供帮助。


在下文中一共展示了maketrans函数的15个代码示例,这些例子默认根据受欢迎程度排序。您可以为喜欢或者感觉有用的代码点赞,您的评价将有助于系统推荐出更棒的Python代码示例。

示例1: get_kmer_densities

def get_kmer_densities(path, kmer):
    mC_trans = string.maketrans("C", "E")
    hmC_trans = string.maketrans("C", "O")
    c_density = KmerHdpDistribution(path, kmer)
    mc_density = KmerHdpDistribution(path, kmer.translate(mC_trans))
    hmc_density = KmerHdpDistribution(path, kmer.translate(hmC_trans))
    return c_density, mc_density, hmc_density
开发者ID:ArtRand,项目名称:signalAlign,代码行数:7,代码来源:vis_kmer_distributions.py

示例2: build_sample_ids_transtable

def build_sample_ids_transtable():
    """Build translation table for sample ids being MIENS compliant"""
    all_chars = "".join([chr(i) for i in range(128)])
    valid_sample_id_chars = letters + digits + "."
    non_valid_sample_id_chars = all_chars.translate(maketrans("", ""), valid_sample_id_chars)
    trans_table = maketrans(non_valid_sample_id_chars, "." * len(non_valid_sample_id_chars))
    return trans_table
开发者ID:josenavas,项目名称:FastUnifrac,代码行数:7,代码来源:sample_id_map_otu_table_conversion.py

示例3: solve_partial

def solve_partial(w, h, pattern, answer, fix_num, solve_num, reset_loop_count=RESET_LOOP_COUNT):
    trans_str_wall = answer[:fix_num]
    trans_table_wall = string.maketrans(trans_str_wall, 
                                        '=' * len(trans_str_wall))
    trans_str_asta = answer[fix_num + solve_num:-1].replace('=', '')
    trans_table_asta = string.maketrans(trans_str_asta,
                                        '*' * len(trans_str_asta))
    pattern_rep = pattern.translate(trans_table_wall)
    pattern_rep = pattern_rep.translate(trans_table_asta)
    answer_rep = answer.translate(trans_table_wall)
    answer_rep = answer_rep.translate(trans_table_asta)

    ####### debug #######
    print '--------- pattern_rep'
    print_pattern(w, h, pattern_rep)
    print '--------- answer_rep'
    print_pattern(w, h, answer_rep)
    ####### debug #######

    move = solve_all(w, h, pattern_rep, answer_rep, reset_loop_count)

    ####### debug #######
    if move:
        pattern_work = create_pattern(w, h, pattern, move)
        print '--------- succeeded'
        print_pattern(w, h, pattern_work)
    else:
        print '--------- not succeeded'
    ####### debug #######
    return move
开发者ID:yaeda,项目名称:gdd11,代码行数:30,代码来源:solve.py

示例4: build_anagrams

def build_anagrams():
    '''
    build a python dict of sorted-word and the set of words
    Using brown corpus and a file containing words from multiple corpora
    strip all punctuations, retain hyphens, replace underscores with space
    '''
    punctuations_replace = '#"(){}[]<>.+,/:;[email protected]_|~-'
    punctuations_remove = '!$\'%&\\*^`'
    # 1. Brown Corpus
    for word in nltk.corpus.brown.words():
        w = str(word)
        # polish just the way tokens were
        w_list = w.translate(string.maketrans(punctuations_replace,' '*len(punctuations_replace)), punctuations_remove).strip().lower().split()
        for each_w in w_list:
            # add the word to redis with key as a sorted word
            wam[''.join(sorted(each_w))].add(each_w)
    # 2. Wordnet
    for word in nltk.wordnet.wordnet.words():
        w = str(word)
        # polish just the way tokens were
        w_list = w.translate(string.maketrans(punctuations_replace,' '*len(punctuations_replace)), punctuations_remove).strip().lower().split()
        for each_w in w_list:
            # add the word to redis with key as a sorted word
            wam[''.join(sorted(each_w))].add(each_w)
    # 3. Other corpora
    with open(BIG_WL, 'r') as f:
        for line in f:
            w = str(line).strip()
            # polish just the way tokens were
            w_list = w.translate(string.maketrans(punctuations_replace,' '*len(punctuations_replace)), punctuations_remove).strip().lower().split()
            for each_w in w_list:
                # add the word to redis with key as a sorted word
                wam[''.join(sorted(each_w))].add(each_w)
开发者ID:karthikprasad,项目名称:krisp-search,代码行数:33,代码来源:utils.py

示例5: test_the_reverse_complement

 def test_the_reverse_complement(self):
     """Check obj.reverse_complement() method."""
     mapping = ""
     for example1 in self._examples:
         if isinstance(example1, MutableSeq):
             continue
         try:
             comp = example1.reverse_complement()
         except ValueError as e:
             self.assertEqual(str(e), "Proteins do not have complements!")
             continue
         str1 = str(example1)
         # This only does the unambiguous cases
         if any(("U" in str1, "u" in str1, example1.alphabet == generic_rna)):
             mapping = maketrans("ACGUacgu", "UGCAugca")
         elif any(("T" in str1, "t" in str1, example1.alphabet == generic_dna,
                  example1.alphabet == generic_nucleotide)):
             mapping = maketrans("ACGTacgt", "TGCAtgca")
         elif "A" not in str1 and "a" not in str1:
             mapping = maketrans("CGcg", "GCgc")
         else:
             # TODO - look at alphabet?
             continue
         self.assertEqual(str1.translate(mapping)[::-1], str(comp))
         self.assertEqual(comp.alphabet, example1.alphabet)
开发者ID:Benjamin-Lee,项目名称:biopython,代码行数:25,代码来源:test_Seq_objs.py

示例6: reverseComplement

def reverseComplement(seq, rna=False):
    if rna:
        complements = string.maketrans("AUCGN", "UAGCN")
        return convertToRNA(seq).translate(complements)[::-1]
    else:
        complements = string.maketrans("ATCGN", "TAGCN")
        return convertToDNA(seq).translate(complements)[::-1]
开发者ID:anthonyho,项目名称:arrayAnalysisTools,代码行数:7,代码来源:seqlib.py

示例7: html_for_url_node

def html_for_url_node(node):
    if not re.match("javascript:", node["url"]):
        linkURL = sanitize(node["url"])
        keysURL = linkURL
        ktrspaces = "                            "
        ktrtable = string.maketrans("[email protected]#$%^&*()_+-=`~;:'\",<.>/?\\|", ktrspaces)
        keysURL = str(keysURL).translate(ktrtable, "").lower()
        keysURL = str(keysURL).translate(None, "[email protected]#$%^&*()_+-=`~;:'\",<.>/?\\|").lower()
        #
        tags = sanitize(node["name"])
        # tags= node['name'] Check for UTF-8 etc...
        # print "TAGS: ",tags
        # tags = tags.translate(None,'[email protected]#$%^&*()_+-=`~;:\'",<.>/?\\|')
        # tags.translate(None,'[email protected]#$%^&*()_+-=`~;:\'",<.>/?\\|')
        # trtable =                          '                              '
        trspaces = "                            "
        trtable = string.maketrans("[email protected]#$%^&*()_+-=`~;:'\",<.>/?\\|", trspaces)
        # tags = str(tags).translate(trtable,'[email protected]#$%^&*()_+-=`~;:\'",<.>/?\\|')
        tags = str(tags).translate(trtable, "").lower()
        tags = str(tags).translate(None, "[email protected]#$%^&*()_+-=`~;:'\",<.>/?\\|").lower()
        #
        allTags = tags + " " + keysURL
        print "# '", sanitize(node["url"]), "'", allTags
        # print '# \'',sanitize(node['url']),'\'', tags
        return '<dt><a href="%s">%s</a>\n' % (sanitize(node["url"]), sanitize(node["name"]))
    else:
        return ""
开发者ID:mariotti,项目名称:abtag,代码行数:27,代码来源:py-chrome-bookmarks.py

示例8: scrape_links_and_wordlistify

def scrape_links_and_wordlistify(links, lower=False, verbose=1):
    import nltk
    import requests
    import string
    raw = ''
    wordlist = {}
    for site in links:
        try:
            if verbose == 1:
                print '[+] fetching data from: ', site
            if site.find('http://pastebin.com/') == 0:
                raw = requests.get(site.replace('http://pastebin.com/', 'http://pastebin.com/raw.php?i=')).content
            else:
                raw = requests.get(site).content
            if lower == False:
                l = string.translate(nltk.clean_html(raw), string.maketrans(string.punctuation, ' ' * 32)).split()
                freq_an(l, wordlist)
            else:
                l = string.lower(nltk.clean_html(raw))
                l = string.translate(l, string.maketrans(string.punctuation, ' ' * 32)).split()
                freq_an(l, wordlist)
        except:
            if verbose == 1:
                print '[-] Skipping url: ', site
    return wordlist
开发者ID:tkisason,项目名称:unhash,代码行数:25,代码来源:gwordlist.py

示例9: generateCounts

	def generateCounts(self):
		wordCounts = {}
		hashtagCounts = {}

		for tweet in self.trainSet:
			hashtags = []
			for word in tweet.split():
				if word.startswith('#') and len(word) > 2:
					word = word.lower().translate(string.maketrans("",""), string.punctuation) # remove punctuation
					hashtags.append(word)
					if word not in wordCounts:
						wordCounts[word] = 1
					else:
						wordCounts[word] += 1
				else:
					if '@' in word:
						continue
					if word in self.stopWords:
						continue
					word = word.lower().translate(string.maketrans("",""), string.punctuation) # remove punctuation
					if word not in wordCounts:
						wordCounts[word] = 1
					else:
						wordCounts[word] += 1

			for hashtag in hashtags:
				if hashtag not in hashtagCounts:
					hashtagCounts[hashtag] = 1.0
				else:
					hashtagCounts[hashtag] += 1.0

		return wordCounts, hashtagCounts
开发者ID:RamaneekGill,项目名称:Twitter-Hashtag-Prediction,代码行数:32,代码来源:NaiveBayesv2.py

示例10: ROT13

def ROT13(test_string):
    """Return a encrypted string offset by 13


    Encrytion does not impact spaces or special characters"""


    shifted_cypherlower = string.ascii_lowercase[13:] + string.ascii_lowercase[:13]
    shifted_cyperupper = string.ascii_uppercase[13:] + string.ascii_uppercase[:13]
    transtable_lower = string.maketrans(string.ascii_lowercase, shifted_cypherlower)
    #print transtable_lower 
    transtable_upper = string.maketrans(string.ascii_uppercase, shifted_cyperupper)
    #print transtable_upper

    
    encrypted_text = []

    for letter in test_string:
        if letter.islower():
            encrypted_text.append(letter.translate(transtable_lower))
        else:
            encrypted_text.append(letter.translate(transtable_upper))
        final_encryption =  "".join(encrypted_text)

    return final_encryption
开发者ID:carlyslater,项目名称:sea-c28-students,代码行数:25,代码来源:rot13.py

示例11: __init__

 def __init__(self): 
     self.currentTroll = random.randint(0, 10)
     self.trollCounter = 0
     self.trolls = [
         #aradia
         { 'prefix': '', 'replace': lambda x: x.translate(string.maketrans('oo', '00')) },
         #terezi
         { 'prefix': '', 'replace': lambda x: x.upper().translate(string.maketrans('AIE', '413')) },
         #tavros
         { 'prefix': '', 'replace': lambda x: x.title().swapcase() },
         #sollux
         { 'prefix': '', 'replace': lambda x: x.replace('s', '2').replace('S', '2').replace('i', 'ii').replace('I', 'II') },
         #karkat
         { 'prefix': '', 'replace': lambda x: x.upper() },
         #nepeta
         { 'prefix': ':33 <', 'replace': lambda x: x.replace('ee', '33').replace('EE', '33') },
         #kanaya
         { 'prefix': '', 'replace': lambda x: x.capitalize() },
         #vriska
         { 'prefix': '', 'replace': lambda x: x.translate(string.maketrans('bB', '88')).replace('ate', '8') },
         #equius
         { 'prefix': 'D -->', 'replace': lambda x: x.translate(string.maketrans('xX', '%%')) },
         #gamzee TODO need a full func
         #eridan
         { 'prefix': '', 'replace': lambda x: x.replace('w', 'ww').replace('v', 'vv').replace('W', 'WW').replace('V', 'VV') },
         #feferi
         { 'prefix': '', 'replace': lambda x: x.replace('h', ')(').replace('H', ')(').replace('E', '-E') },
     ]
开发者ID:Cheeseum,项目名称:suikabot-py,代码行数:28,代码来源:filters.py

示例12: __flatten

 def __flatten(dic, prefix):
     '''
     recursively pass through a dict and flatten it\'s "internal" dicts
     '''
     results = {}
     if dic is not None:
         try:
             for key in dic.keys():
                 if type(dic[key]) in [float, int]:
                     results["%s.%s" % (
                         prefix,
                         str(key).translate(string.maketrans(".", "_"))
                     )] = dic[key]
                 elif type(dic[key] in [dict]):
                     results.update(
                         GraphiteUploaderPlugin.__flatten(
                             dic[key],
                             "%s.%s" % (
                                 prefix,
                                 key.translate(string.maketrans(".", "_"))
                             )
                         )
                     )
         except AttributeError:
             pass
     return results
开发者ID:Blazemeter,项目名称:yandex-tank,代码行数:26,代码来源:plugin.py

示例13: tm

def tm():

    import sys
    import nltk
    import string

    input_file_name = raw_input("Please enter the input file name: ")
    input_path = raw_input("Please enter the input path: ")
    output_file_name = raw_input("Please enter the output file name: ")
    print '\nPlease note that the above entered filename would be used as',
    print 'a prefix for the entire set of documents to be generated.\n'
    output_path = raw_input("Please enter the output path: ")

    with open (input_path + '\\' + input_file_name + '.txt','r') as f:

        para = []
        data = f.read()
        selected=0
        notselect=0
        sentences = data.split("\n\n")

        print "Total # of paragraphs",len(sentences)

        for x in xrange(len(sentences)):
            cond = sentences[x].endswith(".")
            if cond:
                cnt = sentences[x].count(".")
            else:
                cnt= sentences[x].count(".")+1

            if cnt >5:
                #print "paragraph ",x+1,"is selected"
                selected+=1
                sentences[x] = '@'+sentences[x].lower()
                sentences[x] = sentences[x].translate(string.maketrans("",""),string.digits)
                sentences[x] = sentences[x].translate(string.maketrans("",""),string.punctuation)
                tokens = nltk.word_tokenize(sentences[x])
                lemma  = nltk.WordNetLemmatizer()
                porter = nltk.PorterStemmer()

                afewwords = [lemma.lemmatize(i) for i in tokens]
                afewwords = [porter.stem(i) for i in tokens]

                sentences[x] = ' '.join(afewwords)
                para.append(sentences[x])

                filename = output_path + '\\' + output_file_name + str(selected) + '.txt'
                w = open(filename,'w')
                w.write(''.join(para))
                w.close()
                para = []
            else:
                #print "paragraph ",x+1,"is not selected"
                notselect+=1
            #print "cnt - ", cnt
        #print"\n"

        print "# of paragraphs selected",selected
        print "# of paragraphs not selected",notselect
    f.close()
开发者ID:Sra1Phani,项目名称:Authorship-Ascription,代码行数:60,代码来源:doc2corpus.py

示例14: checkData

 def checkData(self,row):
     jobTime = row[4]
     jobTerm = row[5]
     w1={'full time':1,'part time':-1,'':0}    #this dictionary corresponds to time feature
     w2={'permanent':0,'contract':1,'':-1}     #this dictionary corresponds to term feature
     
     if jobTime == '' or jobTerm == '':
         s=row[2].lower()          
         s=s.translate(string.maketrans("‘’‚“”„†‡‰‹›!“#$%&‘()™*+,-�./0123456789:;<=>[email protected][\]_`{|}~–—΅Ά£¤¥¦§¨©�«¬®―°±²³΄µ¶·ΈΉΊ»Ό½ΎΏήίΰαβγδεζηθικλμνξοπρςστυφχψωϊϋόύώ�€Άβ—�ο�'","                                                                                                                                "))
         if jobTime=='':
             if ('full time' in s and 'part time' in s) or ('full time' not in s and 'part time' not in s):
                 word1=''
             else:
                 if 'full time' in s:      #searching full time in description
                     word1='full time'
                 else:
                     word1='part time'
         else:
             word1=jobTime.translate(string.maketrans("_"," ")) #removing underscore from time feature value
             
         if jobTerm=='':
             if ('permanent' in s and 'contract' in s) or ('permanent' not in s and 'contract' not in s):
                 word2=''
             else:
                 if 'permanent' in s:      #searching permanent in description
                     word2='permanent'
                 else:
                     word2='contract'
         else: word2=jobTerm.translate(string.maketrans("_"," "))   #removing underscore from term feature value
     
     else:
         word1=jobTime.translate(string.maketrans("_"," "))
         word2=jobTerm.translate(string.maketrans("_"," "))
         
     return [word1,w1[word1],word2,w2[word2]]
开发者ID:hshed,项目名称:salary-prediction,代码行数:35,代码来源:Dictionary.py

示例15: url_sign

def url_sign( uri_path, params, client_id, signing_key ):
    signing_key = signing_key.translate(string.maketrans('-_', '+/'))
    padding_factor = ( 4 - len( signing_key ) % 4 ) % 4
    signing_key += "=" * padding_factor
    binary_key = base64.b64decode(unicode(signing_key).translate(dict(zip(map(ord, u'-_'), u'+/'))))

    # construct URI for signing
    uri_path_params = uri_path + '?'
    first = True
    for k in params.keys():
        if not first:
            uri_path_params += '&'
        else:
            first = False
        uri_path_params = "%(base)s%(key)s=%(value)s" % {
                                                        'base':uri_path_params,
                                                        'key':k,
                                                        'value':urllib.quote_plus(str(params[k]))
                                                        }
    uri_path_params += '&client=' + client_id

    # Sign
    digest = hmac.new(binary_key, uri_path_params, hashlib.sha1).digest()
    digest = base64.b64encode( digest )
    digest = digest.translate(string.maketrans('+/', '-_'))
    return "%s&sig=%s" % ( uri_path_params, digest.rstrip('=') )
开发者ID:vdarmadi,项目名称:dishart,代码行数:26,代码来源:v2_updated_since.py


注:本文中的string.maketrans函数示例由纯净天空整理自Github/MSDocs等开源代码及文档管理平台,相关代码片段筛选自各路编程大神贡献的开源项目,源码版权归原作者所有,传播和使用请参考对应项目的License;未经允许,请勿转载。