# -*- coding: utf-8 -*- """ --- Corpus Clean - Stage 1 Copyright (c) 2008-2010 Gisle Ytrestol (gisley@ifi.uio.no) This program is free software; you can redistribute it and/or modify it under the terms of the GNU Lesser General Public License as published by the Free Software Foundation; either version 2.1 of the License, or (at your option) any later version. This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public License for more details. The first stage of this script takes a folder of Wikipedia articles in the original Wikipedia markup, and strips the article for unwanted markup. The script is intended to be used together with Tokenizer v1.0, http://www.cis.uni-muenchen.de/~wastl/misc/. The output of this script is a single file which should be used as input for Tokenizer v1.0 . The output from Tokenizer should be used by stage 2, cccp.py. This second stage inserts proper sentence boundaries and allows the user to specify whether the output should be one single file for each corresponding input file, or whether the entire corpus should be dumped into on single file/corpus. Run python ccp.py -h for help! """ import os, re, regex, urllib,cccp,langregex import sys,string import codecs from optparse import OptionParser class AdjustName: def addSlash(self,name): if name[-1] != '/': name = name+'/' return name else: return name def removeSlash(self,name): if name[-1] == '/': name = name[:-1] return name else: return name class WikiReader: def readFile(self,file): wholeFile = file.read() return wholeFile def listFiles(self,inputFolder): fileList = os.listdir(inputFolder) #print fileList newFileList = [] for file in fileList: if not file[0]== '.': newFileList.append(file) newFileList.sort() #print newFileList return newFileList def readFirstLine(self,file): firstLine = file.readline() return firstLine class WikiProcessor: def __init__(self,redirect): self.dictChecker = WikiDict() self.redirect = redirect if self.redirect == None: self.redirect = 1 def tableCleanerOld(self,input): input = regex.regtableConvertStart.sub(r'Ӂ',input) input = regex.regtableConvertEnd.sub(r'\1ጣ',input) while True: if regex.regtableConverted.search(input): input = regex.regtableConverted.sub(r'',input) else: break input = regex.regtableConvertRevertStart.sub('{|',input) input = regex.regtableConvertRevertEnd.sub('|}',input) return input def tableCleaner(self,input): while True: if regex.regtableClean.search(input): input = regex.regtableClean.sub(r'',input) else: break return input def tableCleaner2(self,input): splitInput = input.split("\n") keepInput = [] inTable = False for inputLine in splitInput: if inputLine.startswith("|") or inputLine.startswith("!") : if inTable == True: continue regtablestart if regex.regtablestart.search(inputLine): inTable = True continue if inputLine.startswith("|}") and inTable == True: inTable = False continue if inTable == True: if not inputLine.startswith("|"): if not inputLine.startswith("!"): inTable = False keepInput.append(inputLine) return "\n".join(keepInput) def tableCleaner2Reverse(self,input): splitInput = input.split("\n") splitInput.reverse() keepInput = [] inTable = False for inputLine in splitInput: if inputLine.startswith("|") or inputLine.startswith("!") : if inTable == True: continue if inputLine.startswith("|}"): inTable = True continue if inputLine.startswith("{|") and inTable == True: inTable = False continue if inTable == True: if not inputLine.startswith("|"): if not inputLine.startswith("!"): inTable = False keepInput.append(inputLine) keepInput.reverse() return "\n".join(keepInput) def processFile(self,firstLine,wholeFile): if self.redirect == 1: try: firstLine, wholeFile = self.redirectCheck(firstLine,wholeFile) except: sys.stdout.write("\nERROR WITH THE REDIRECT PROCESSING\nARE YOU SURE AN OFFLINE WIKIPEDIA READER IS ENABLED?\n\nTo run the script without redirect processing, use the -n option\n") #sys.exit(0) if self.dictChecker.checkIfIn(firstLine): firstLine,wholeFile = self.cleanArticle(firstLine,wholeFile) firstLine = self.addTitle(firstLine) return firstLine,wholeFile else: return False, False def cleanArticle(self,firstLine,wholeFile): firstline = self.regCleanFile(firstLine) wholeFile = self.regCleanFile(wholeFile) wholeFile = self.removeEnd(wholeFile) return firstLine,wholeFile def addTitle(self,firstLine): firstLine = '
'+firstLine.rstrip()+'
' return firstLine def redirectCheck(self,firstLine,wholeFile): # if the article contains a redirect link, the redirect URL will be used to if regex.regredirect.search(wholeFile): #retrive the correct article title = regex.regredirect.sub(r'\1',wholeFile) url = 'http://127.0.0.1:8000/article/'+title #address must correspond with local Wikipedia url = re.sub(r' ','_',url) #print url page = urllib.urlopen(url) page.close() thisfile = '/var/tmp/result' #the wiki article is stored here when it is accessed file = open(thisfile,'r') firstLine = file.readline() wholeFile = file.read() return firstLine,wholeFile else: return firstLine,wholeFile def addNewline(self,input): input = regex.regeos.sub('\n',input) return input def removeEnd(self,input): input = regex.regsourcelookahead.sub(r'___',input) input = regex.regsourcelookahead.sub(r'___',input) input = regex.regsourcelookahead.sub(r'___',input) input = regex.regseealsolookahead.sub(r'___',input) input = regex.regseealsolookahead.sub(r'___',input) input = regex.regseealsolookahead.sub(r'___',input) input = regex.regnoteslookahead.sub(r'___',input) input = regex.regnoteslookahead.sub(r'___',input) input = regex.regnoteslookahead.sub(r'___',input) input = regex.regreflookahead.sub(r'___',input) input = regex.regreflookahead.sub(r'___',input) input = regex.regreflookahead.sub(r'___',input) input = regex.regsealso.sub('',input) input = regex.regnotes.sub('',input) input = regex.regreferences.sub('',input) input = regex.regsources.sub('',input) input = regex.regsourcelookaheadrestore.sub(r'\2\1\2',input) input = regex.regseealsolookaheadrestore.sub(r'\2\1\2',input) input = regex.regnoteslookaheadrestore.sub(r'\2\1\2',input) input = regex.regreflookaheadrestore.sub(r'\2\1\2',input) input = regex.regbibliography.sub('',input) input = regex.regfootnotes.sub('',input) input = regex.regrelated.sub('',input) input = regex.regexternal.sub('',input) return input def regCleanFile(self,input): input = regex.regipa.sub(r'<___\1___>',input) input = regex.regjava.sub(r'<___\1___>',input) input = regex.regiast.sub(r'<___\1___>',input) ## japanese article templates #input = regex.regtransjap.sub(r'\1',input) #input = regex.reglangjap.sub(r'\1',input) input = regex.reglanggeneralpreserve.sub(r'<___\1___>',input) input = regex.regtransgeneralpreserve.sub(r'<___\1___>',input) input = regex.regnihongopreservere.sub(r'<___\1___>',input) input = regex.regharv_general.sub(r'<___\1___>',input) input = regex.regaudio_general.sub(r'<___\1___>',input) input = regex.regflagtemplate.sub(r'<___\1___>',input) """ IF WE WANT TO EXPAND THE TEMPLATES, USE THESE! input = regex.regharv_aut_aut_year_page.sub(r'(\1 & \2 \3, \4)',input) input = regex.regharv_aut_year_page.sub(r'(\1 2, \3)',input) input = regex.regharvtxt_aut_year_page.sub(r'\1 (\2, \3)',input) input = regex.regharvtxt_aut_year.sub(r'\1 (\2)',input) input = regex.regharvtxt_aut_aut_year_page.sub(r'\1 & \2 (\3, \4)',input) input = regex.regharvtxt_aut_aut_year.sub(r'\1 & \2 (\3)',input) input = regex.regharvnb_aut_year_page_nb.sub(r'\1 \2, \3',input) input = regex.regharvnb_aut_year_nb.sub(r'\1 \2',input) input = regex.regharvnb_aut_aut_year_page_nb.sub(r'\1 & \2 \3, \4',input) #input = regex.regharvnb_aut_aut_year_page_nb.sub(r'\1 & \2 \3, \4',input) input = regex.regharvnb_aut_aut_aut_year_page_nb.sub(r'\1, \2 & \3 \4, \5',input) input = regex.regharvnb_aut_aut_year_nb.sub(r'\1 & \2 \3',input) input = regex.regharvnb_aut_aut_aut_year_nb.sub(r'\1, \2 & \3 \4',input) input = regex.regharvcoltxt_aut_year_page.sub(r'\1 (\2:\3)',input) """ """ TO EXPAND TEMPLATES, USE THESE! #input = regex.regtransgeneral.sub(r'\1',input) #input = regex.reglanggeneral.sub(r'\1',input) input = regex.regnihongohardcode.sub(r'\1)',input) input = regex.regnihongojap5.sub(r'\1 (\2 \3 \4 \5)',input) input = regex.regnihongojap4.sub(r'\1 (\2 \3 \4)',input) input = regex.regnihongojap3.sub(r'\1 (\2 \3)',input) input = regex.regnihongojap2.sub(r'\1 (\2)',input) """ while regex.regcurly1.search(input): input = regex.regcurly1.sub('',input) input = regex.regblockquote.sub(r'
\1
',input) input = regex.regdiv2.sub('',input) input = regex.reggallery.sub('',input) input = regex.regimage.sub(r'',input) input = regex.regimage.sub(r'',input) input = regex.regref2.sub('',input) input = regex.regref.sub('',input) input = regex.regcomment.sub('',input) input = regex.regsingleast.sub('',input) input = regex.regdeflist.sub(r'\1',input) input = self.tableCleaner(input) input = self.tableCleaner2(input) input = self.tableCleaner2Reverse(input) input = regex.regwikitable3.sub('',input) input = regex.regtable.sub('',input) input = regex.regtableborder.sub('',input) input = regex.regtablehardcode.sub('',input) input = regex.regcategory.sub('',input) input = langregex.purgeLang(input) input = regex.regbacktocurly1.sub('{{',input) input = regex.regbacktocurly2.sub('}}',input) input = regex.regsentinitialbracket.sub(r'\1\2',input) input = regex.regbracket.sub(r'\1', input) input = regex.regbullets.sub(r'\1', input) input = regex.regbullets2.sub(r'\1', input) input = regex.regindentcolon.sub(r'\1', input) input = regex.regbulletscolon.sub(r'\2', input) input = regex.regbr.sub(r'',input) input = regex.regtitle.sub(r'\1',input) input = regex.regparagraph.sub('',input) input = regex.regyeareos.sub(r'\1',input) input = regex.regorg.sub(r'\1',input) input = regex.reghyphen.sub('',input) #remove no wiki #input = regex.regremovenowiki.sub('',input) input = self.removeLines(input) input = regex.regremovenewline.sub(' ',input) return input # def removeTableLeftover(self,input): # splitInput = input.split('\n') # newArticle = '' # for line in splitInput: # if not regex.regletternumber.search(line): # continue # if not "ARTICLE>" in line: # if regex.regonlyXML.match(line): # continue # if line.startswith("|") or line.startswith("!") or line.startswith("{|") or line.startswith("|}"): # continue # else: # newArticle = newArticle+str(line)+'\n' # return newArticle.rstrip() # def removeLines(self,line): splitLine = line.split("\n") keepLine = [] for line in splitLine: if regex.regletternumber.search(line): keepLine.append(line) return "\n".join(keepLine) class WikiDict: def __init__(self,wikiDict=None): self.wikiDict = {} def checkIfIn(self,firstLine): if firstLine in self.wikiDict: return False else: self.wikiDict[firstLine] = "" return True class WikiWriter: def __init__(self,output): self.outFile = open(output,'w') def writeFile(self,firstLine,wholeFile): if wholeFile: #print "Writing "+ firstLine self.outFile.write(''+firstLine+'') self.outFile.write(wholeFile) self.outFile.write('\n
\n') def closeFile(self): self.outFile.close() def checkSyntax(input,output): if output == None or input == None: return False if os.path.isdir(input) and not os.path.isdir(output): return True else: return False def main(): parser = OptionParser() parser.add_option("-i", "--input", dest="input", help="Input folder where Wikipedia Source files are stored", metavar="Input Folder") parser.add_option("-o", "--output", dest="output", help="Output file where cleansed Wikipedia Source files will be stored", metavar="Output File") parser.add_option("-n", "--noredirects", dest="redirects", action ="store_false", help="No connection to local Wikipedia Reader, therefore no redirect processing.", metavar="redirects") options, args = parser.parse_args() input = options.input output = options.output redirect = options.redirects if checkSyntax(input,output): pass else: sys.stdout.write("\nERROR WITH THE INPUT/OUTPUT FILES\n") sys.exit(0) wikiReader = WikiReader() wikiProcess = WikiProcessor(redirect) wikiWriter = WikiWriter(output) adjustName = AdjustName() input = adjustName.removeSlash(input) fileList = wikiReader.listFiles(input) for file in fileList: fileName = input+'/'+file if os.path.isfile(fileName): #test utf-8 #codecs.open( "someFile", "r", "utf-8" ) #fileObject = open(input+'/'+file,'r','utf-8') fileObject = open(input+'/'+file,'r') firstLine = wikiReader.readFirstLine(fileObject) wholeFile = wikiReader.readFile(fileObject) #print fileName #print len(wholeFile) firstLine, wholeFile = wikiProcess.processFile(firstLine,wholeFile) if firstLine: #returns false if article already in wikiWriter.writeFile(firstLine,wholeFile) fileObject.close() wikiWriter.closeFile() if __name__ == '__main__': main()