Uzanto:Maksim/sozd sta

De Wikipedio
  • cht_x_sta.py
# -*- coding: UTF-8 -*-

__version__='$Id:'

import sys, codecs, re, os
import wikipedia, date, xmlreader
from bib_tekst import *

def provstop(fstop):
    return os.access(fstop,os.F_OK)

def main():
    fstop='cht_x_sta_stop.txt'
    try:
        os.unlink(fstop)
    except:
        pass

    if provstop(fstop):
        wikipedia.output(u'stop0!\n')
        return

#    fkat = wikipedia.argHandler(sys.argv[1], 'cht_cat')
#    fzap = wikipedia.argHandler(sys.argv[2], 'cht_cat')
#    fotch = wikipedia.argHandler(sys.argv[3], 'cht_cat')
    fi = wikipedia.argHandler(sys.argv[1], 'cht_cat')
    fkat = fi+'.txt'
    fzap = fi
    fotch = fi+'_ot.txt'

    flotd=0
    if len(sys.argv)>=3 and sys.argv[2]==u'2':
        flotd=1

    mysite = wikipedia.getSite()

    f=codecs.open(fkat,'r',encoding='utf-8')

    skon=u''
    i=0

    if os.access(fotch,os.F_OK):
        otch = codecs.open(fotch, 'rb', 'utf-8')

        for s in otch.readlines():
            if ord(s[0]) == 65279:
                s=s[1:]
            s=s.replace(u'\r',u'')
            while len(s)>0 and s[len(s)-1]==u'\n':
                s=s[:len(s)-1]
            if (s[0]==u'+' or s[0].isdigit()) and len(s)>=9:
                s1=s[8:]
                skon=s1
                if s[0].isdigit():
                    i=0
                    j=0
                    while j<=5:
                        i=i*10+ord(s[j])-ord(u'0')
                        j+=1
                    i+=1
        otch.close()

    wikipedia.output(u'skon=%s\n i=%d' % (skon,i))

    otch = codecs.open(fotch, 'a', 'utf-8')

    flrab=0

    t1=f.readlines()
    if not flotd:
        if skon==u'':
            vertfile = codecs.open(u'%s_rez.txt' % fzap, 'w', 'utf-8')
        else:
            vertfile = codecs.open(u'%s_rez.txt' % fzap, 'a', 'utf-8')

    j0=0
    while j0<len(t1):
        plm=[]

        j=j0
        while j<len(t1) and j<j0+64:
            n=t1[j]
            if ord(n[0]) == 65279:
                n=n[1:]

            n=ubr_nk_prob(n)
 
            if n.startswith(u'[['):
                n=n[2:]
            if len(n)>2 and n[len(n)-2:]==u']]':
                n=n[:len(n)-2]

            wikipedia.output(u'... "%s" "%s"\n' % (n,skon))
            if skon==u'' or flrab:
                pl = wikipedia.Page(mysite,n)
                plm.append(pl)
            elif n==skon:
                flrab=1

            j+=1

        try:
            wikipedia.getall(mysite, plm)
        except xmlreader.xml.sax._exceptions.SAXParseException:
            wikipedia.getall(mysite, plm)

        for pl in plm:

            try:
                t = pl.get()
                if flotd:
                    vertfile = codecs.open(u'%s%06d.txt' % (fzap,i), 'w', 'utf-8')
                vertfile.write(pl.title() + '\n')
                vertfile.write(t + '\n')
                if flotd:
                    vertfile.close()
                else:
                    vertfile.write('================\n')
                    vertfile.flush()
                b = u'%06d  %s\n' % (i,pl.title())
                wikipedia.output(b)
                otch.write(b)
                otch.flush()
                i=i+1
            except wikipedia.IsRedirectPage:
                b = u'+       %s\n-> %s\n' % (pl.title(),pl.getRedirectTarget())
                wikipedia.output(b)
                otch.write(b)
                otch.flush()
            except wikipedia.NoPage:
                b = u'-       %s\n' % pl.title()
                wikipedia.output(b)
                otch.write(b)
                otch.flush()
        j0=j
        if provstop(fstop):
            wikipedia.output(u'stop!\n')
            return

try:
    main()
finally:
    wikipedia.stopme()
    try:
        excludefile.close()
    except:
        pass

* cht_mn_sta.py

# -*- coding: UTF-8 -*-

__version__='$Id:'

import sys, codecs, re
import wikipedia, date, catlib



try:

#    fkat = wikipedia.argHandler(sys.argv[1], 'cht_cat')
#    fzap = wikipedia.argHandler(sys.argv[2], 'cht_cat')
#    fotch = wikipedia.argHandler(sys.argv[3], 'cht_cat')
    fi = wikipedia.argHandler(sys.argv[1], 'cht_cat')
    fkat = fi+'.txt'
    fzap = fi
    fotch = fi+'_ot.txt'

    mysite = wikipedia.getSite()

    f=codecs.open(fkat,'r',encoding='utf-8')

    otch = codecs.open(fotch, 'w', 'utf-8')


    t1=f.readlines()
    i=0

    for line in t1:
        if ord(line[0]) == 65279:
            line=line[1:]

        workingcatname = line

        wikipedia.output(workingcatname)

        pl = wikipedia.Page(mysite,wikipedia.UnicodeToAsciiHtml(workingcatname))
        ple=pl.exists()

        if ple: 
            try:
                t = pl.get()
                vertfile = codecs.open(u'%s%06d.txt' % (fzap,i), 'w', 'utf-8')
                vertfile.write(pl.title() + '\n')
                vertfile.write(t + '\n')
                vertfile.flush()
                vertfile.close()
                b = u'%06d  %s\n' % (i,pl.title())
                wikipedia.output(b)
                otch.write(b)
                otch.flush()
                i=i+1
            except wikipedia.IsRedirectPage:
                b = u'+       %s\n-> %s\n' % (pl.title(),pl.getRedirectTarget())
                wikipedia.output(b)
                otch.write(b)
                otch.flush()
        else:
            b = u'-       %s\n' % pl.title()
            wikipedia.output(b)
            otch.write(b)
            otch.flush()

finally:
    wikipedia.stopme()
    try:
        excludefile.close()
    except:
        pass

* cht_mn_cat.py

# -*- coding: UTF-8 -*-

__version__='$Id:'

import sys, codecs, re
import wikipedia, date, catlib


try:
    fi = wikipedia.argHandler(sys.argv[1], 'cht_cat')
    fkat = fi+u'.txt'
    fzap = fi+u'_rez.txt'


#    fkat = wikipedia.argHandler(sys.argv[1], 'cht_cat')
#    fzap = wikipedia.argHandler(sys.argv[2], 'cht_cat')


    f=codecs.open(fkat,'r',encoding='utf-8')
#    f=open(fkat,'r') #,encoding='utf-8')

    vertfile = codecs.open(fzap, 'w', 'utf-8')

    mysite = wikipedia.getSite()

    for s in f.readlines():
        if ord(s[0]) == 65279:
            s=s[1:]
        s=s.replace(u'\r',u'')
        while len(s)>0 and s[len(s)-1]==u'\n':
            s=s[:len(s)-1]

        wikipedia.output(s)

        workingcat = catlib.Category(mysite,s)

        vertfile.write( u'# %s\n' % s )
        vertfile.flush()

        list = workingcat.articles()
        if list:
            for pl in list:
                vertfile.write( u'%s\n' % pl.title() )
                vertfile.flush()
        list = workingcat.subcategories()
        if list:
            for ka in list:
                vertfile.write( u'%s\n' % ka.title() )
                vertfile.flush()

finally:
    wikipedia.stopme()
    try:
        excludefile.close()
    except:
        pass

* sozd_sta.py

# -*- coding: UTF-8 -*-

__version__='$Id:'

import wikipedia
import re, sys, os
import codecs

def ubr_nk_prob(t):
    i=0
    while i<len(t) and ((t[i]==u' ') or (t[i]==u'\n') or (t[i]==u'\t') or
                        (t[i]==u'\r')):
        i+=1
    j=len(t)-1
    while j>i and ((t[j]==u' ') or (t[j]==u'\n') or (t[j]==u'\t') or
                        (t[j]==u'\r')):
        j-=1
    return t[i:j+1]


def sozd(otch,n,t,fldob):

    n=ubr_nk_prob(n)

    if n.startswith(u'[['):
        n=n[2:]

    if len(n)>2 and n[len(n)-2:]==u']]':
        n=n[:len(n)-2]

    pl = wikipedia.Page(mysite,n)
    wikipedia.output(pl.title())

    try:
        st=pl.get()
    except wikipedia.NoPage:
        st=u''
    except wikipedia.IsRedirectPage:
        st=u'--> [['+pl.getRedirectTarget()+u']]\n----\n'

    try:
        if st!=u'' and not fldob:
            wikipedia.output(u"Page %s already exists, not adding!"%pl.title())
            otch.write(u'-  %s\n' % pl.title())
            otch.flush()
        elif st!=u'' and fldob:
            pl.put(st+u'\n\n'+t, comment = u"", minorEdit = False)
            otch.write(u'++ %s\n' % pl.title())
            otch.flush()
        else:
            pl.put(t, comment = u"", minorEdit = False)
            otch.write(u'+  %s\n' % pl.title())
            otch.flush()
        return
    except wikipedia.EditConflict:
        wikipedia.output(u'!!!EditConflict!!!\n')
        otch.write(u'!  %s\n' % pl.title())
        otch.flush()


def provstop(fstop):
    return os.access(fstop,os.F_OK)


def main(filename,fotch,fldob):

    fstop='sozd_sta_stop.txt'
    try:
        os.unlink(fstop)
    except:
        pass

    if provstop(fstop):
        wikipedia.output(u'stop0!\n')
        return

    t=u''
    n=u''
    f=codecs.open(filename,'rb',encoding='utf-8')

#    skon=u''
    nkon=0

    if os.access(fotch,os.F_OK):
        otch = codecs.open(fotch, 'rb', 'utf-8')

        for s in otch.readlines():
            if ord(s[0]) == 65279:
                s=s[1:]
            s=s.replace(u'\r',u'')
            while len(s)>0 and s[len(s)-1]==u'\n':
                s=s[:len(s)-1]
                wikipedia.output(u'nnn\n')
            s1=s[3:]
            if len(s1)>=1:
#                skon=s1
                nkon+=1
        otch.close()

#    wikipedia.output(u'skon=%s\n' % skon)

    otch = codecs.open(fotch, 'a', 'utf-8')

#    flrab=0
    i=0
    sch=0
    for s in f.readlines():
#        wikipedia.output(u'%d\n' % ord(s[0]))
        if ord(s[0]) == 65279:
            s=s[1:]
#        wikipedia.output(u'%d   %s   %s\n' % (i,n,s))
        if s[0:7] == u'=======':
            if i==1:
#                wikipedia.output(u'... "%s" "%s"\n' % (n,skon))
                wikipedia.output(u'... %d %d "%s"\n' % (sch,nkon,n))
#                if skon==u'' or flrab:
#                    sozd(otch,n,t,fldob)
#                elif n==skon:
#                    flrab=1
                if sch>=nkon:
                    sozd(otch,n,t,fldob)
                sch+=1
                if provstop(fstop):
                    wikipedia.output(u'stop!\n')
                    return
            i=0    
            t=u''
        elif i==0:
            n=ubr_nk_prob(s)
#            while len(n)>0 and n[len(n)-1]==u'\n':
#                n=n[:len(n)-1]
            i=1
        else:
            t=t+s
#            t.append(s)



#        wikipedia.output(u"No title found - skipping a page.")


#    text=''.join(text)

filename = wikipedia.argHandler(sys.argv[1], 'cht_cat')
fotch = filename+u'_ot.txt'
filename+=u'.txt'

fldob=0
if len(sys.argv)>=4 and sys.argv[2]==u'2':
    fldob=1
mysite = wikipedia.getSite()

try:
    main(filename,fotch,fldob)
except:
    wikipedia.stopme()
    raise
else:
    wikipedia.stopme()

* sozd_alidir.py

# -*- coding: UTF-8 -*-

__version__='$Id:'

import wikipedia
import re, sys
import codecs


def sozd(n1,n2):

    pl1 = wikipedia.Page(mysite,wikipedia.UnicodeToAsciiHtml(n1))
    ple1= pl1.exists()
    ba=u''
    if ple1:
        si1=u'+'
        try:
            pl1.get()
        except wikipedia.IsRedirectPage:
            ba = u'-> %s\n' % pl1.getRedirectTarget()
    else:
        si1=u'-'
    ti1=pl1.title()
    b=u'%s %s\n%s' % (si1,ti1,ba)
    wikipedia.output(b)
    otch.write(b)
    otch.flush()

    pl2 = wikipedia.Page(mysite,wikipedia.UnicodeToAsciiHtml(n2))
    ple2= pl2.exists()
    ba=u''
    if ple2:
        si2=u'+'
        try:
            pl2.get()
        except wikipedia.IsRedirectPage:
            ba = u'-> %s\n' % pl2.getRedirectTarget()
    else:
        si2=u'-'
    ti2=pl2.title()
    b=u'%s %s\n%s' % (si2,ti2,ba)
    wikipedia.output(b)
    otch.write(b)
    otch.flush()

    if ple1 & (ple2==0):
        t=u'#REDIRECT [[%s]]' % ti1
        pl2.put(t, comment = t, minorEdit = True)
        wikipedia.output(t)
        otch.write(u'%s\n' % t)
        otch.flush()
    elif (ple1==0) & ple2:
        t=u'#REDIRECT [[%s]]' % ti2
        pl1.put(t, comment = t, minorEdit = True)
        wikipedia.output(t)
        otch.write(u'%s\n' % t)
        otch.flush()

    return


def main():
    n=u''
    f=codecs.open(filename,'rb',encoding='utf-8')

    i=0
    for s in f.readlines():
#        wikipedia.output(u'%d\n' % ord(s[0]))
        if ord(s[0]) == 65279:
            s=s[1:]
#        wikipedia.output(u'%d   %s   %s\n' % (i,n,s))
        if i==0:
            n=s
            i=1
        else:
            sozd(n,s)
            i=0    
#            t.append(s)



#        wikipedia.output(u"No title found - skipping a page.")


#    text=''.join(text)

filename = wikipedia.argHandler(sys.argv[1], 'cht_cat')
fotch = wikipedia.argHandler(sys.argv[2], 'cht_cat')

for arg in sys.argv[3:]:
    arg = wikipedia.argHandler(arg, 'pagefromfile')
    if arg:
        if arg=="-log":
            import logger
            sys.stdout = logger.Logger(sys.stdout, filename = 'pagefromfile.log')
mysite = wikipedia.getSite()
otch = codecs.open(fotch, 'w', 'utf-8')

try:
    main()
except:
    wikipedia.stopme()
    raise
else:
    wikipedia.stopme()

* sozd_ali1.py

# -*- coding: UTF-8 -*-

__version__='$Id:'

import wikipedia
import re, sys
import codecs


def prov(n1):

    pl1 = wikipedia.Page(mysite,wikipedia.UnicodeToAsciiHtml(n1))
    ple1= pl1.exists()
    ba=u''
    if ple1:
        si1=u'+'
        try:
            pl1.get()
        except wikipedia.IsRedirectPage:
            ba = u'-> %s\n' % pl1.getRedirectTarget()
    else:
        si1=u'-'
    ti1=pl1.title()
    b=u'%s %s\n%s' % (si1,ti1,ba)
    wikipedia.output(b)
    otch.write(b)
    otch.flush()

    return (pl1,ple1,ti1)

def sozd(pl1,ple1,ti1,pl2,ple2,ti2):

    if (ple1==0) & ple2:
        t=u'#REDIRECT [[%s]]' % ti2
        pl1.put(t, comment = t, minorEdit = True)
        b=u'++ %s\n%s\n' % (ti1,t)
        wikipedia.output(b)
        otch.write(b)
        otch.flush()

    otch.write(u'\n')
    otch.flush()

    return


def main():
    n=u''
    pl2pr=None
    ple2pr=0
    ti2pr=u''
    spr=u''
    f=codecs.open(filename,'rb',encoding='utf-8')

    i=0
    for s in f.readlines():
#        wikipedia.output(u'%d\n' % ord(s[0]))
        if ord(s[0]) == 65279:
            s=s[1:]
#        wikipedia.output(u'%d   %s   %s\n' % (i,n,s))
        if i==0:
            n=s
            i=1
        else:
            (pl1,ple1,ti1)=prov(n)
            if spr==s:
                pl2=pl2pr
                ple2=ple2pr
                ti2=ti2pr
            else:
                (pl2,ple2,ti2)=prov(s)
            sozd(pl1,ple1,ti1,pl2,ple2,ti2)
            spr=s
            pl2pr=pl2
            ple2pr=ple2
            ti2pr=ti2
            i=0    
#            t.append(s)



#        wikipedia.output(u"No title found - skipping a page.")


#    text=''.join(text)

filename = wikipedia.argHandler(sys.argv[1], 'cht_cat')
fotch = wikipedia.argHandler(sys.argv[2], 'cht_cat')

for arg in sys.argv[3:]:
    arg = wikipedia.argHandler(arg, 'pagefromfile')
    if arg:
        if arg=="-log":
            import logger
            sys.stdout = logger.Logger(sys.stdout, filename = 'pagefromfile.log')
mysite = wikipedia.getSite()
otch = codecs.open(fotch, 'w', 'utf-8')

try:
    main()
except:
    wikipedia.stopme()
    raise
else:
    wikipedia.stopme()

* saveHTML1.py

# -*- coding: utf-8  -*-
"""
(C) 2004 Thomas R. Koll, <tomk32@tomk32.de>
 Distribute under the terms of the PSF license.

This bot downloads the HTML-pages of articles and images
and saves the interesting parts, i.e. the article-text
and the footer to a file like Hauptseite.txt.

TODO:
   change the paths in the HTML-file


Options:

      -o:                Specifies the output-directory where to save the files   

      -images:           Downlaod all images
      -overwrite:[I|A|B] Ignore existing Images|Article|Both and
                         download them even if the exist


Features, not bugs:
* Won't d/l images of an article if you set -overwrite:A

"""
__version__ = '$Id: saveHTML.py,v 1.10 2005/10/13 20:10:03 leogregianin Exp $'


import wikipedia,httplib,StringIO,re,sys,md5,os, string,codecs,urllib
from htmlentitydefs import *

def extractArticle(data):
    """ takes a string with the complete HTML-file
    and returns the article which is contained in
    <div id='article'> and  the pagestats which
    contain information on last change """

    images = []
    s = StringIO.StringIO(data)
    rPagestats = re.compile('.*(\<span id\=(\"|\')pagestats(\"|\')\>.*\<\/span\>).*')
    rBody = re.compile('.*<div id\=\"content\">.*')
    rFooter = re.compile('.*<div id\=\"footer\">.*')
    rDivOpen = re.compile('.*<div ')
    rDivClose = re.compile('.*<\/div>.*')
    divLevel = 1
    divLast = -1
    inArticle = 0
    inFooter  = 0
    result = {'article':"",
              'footer':""}
    for line in s:
        if line == "<p><br /></p>":
            continue
        line = line.replace(" ", " ")
        line = line.replace(" ", " ")

        if rDivOpen.match(line):
            divLevel = divLevel + 1
        if rBody.match(line):
            inArticle = 1
            divLast = divLevel-2
        elif rFooter.match(line):
            divLast = divLevel-1
            inFooter  = 1
        if inArticle:
            result['article'] += line
        elif inFooter:
            result['footer'] += line
        if rDivClose.match(line):
            divLevel = divLevel - 1
            if divLevel == divLast:
                inArticle = 0
                inFooter = 0
                divLast = -1


    return result

def html2txt(str):
    dict = {"%C3%A4": "ä",
            "%C3%B6": "ö",
            "%C3%BC": "ü",
            "%C3%84": "Ä",
            "%C3%96": "Ö",
            "%C3%9C": "Ü",
            "%C3%9F": "ß",
            "%27": "'",
            "%28": "(",
            "%29": ")",
            "%2C": ","
            }
            
    for entry in dict:
        str = re.sub(entry, dict[entry], str)
    return str

def extractImages(data):
    """ takes a string with the complete HTML-file
    and returns the article which is contained in
    <div id='article'> and  the pagestats which
    contain information on last change """

    images = []
    rImage = re.compile('<a href=[\r\n]*?"/wiki/.*?:(.*?)".*?[\r\n]*?.*?class=[\r\n]*?"image"', re.MULTILINE)
    rThumb = re.compile('<a href=[\r\n]*?"/wiki/.*?:(.*?)".*?[\r\n]*?.*?class=[\r\n]*?"internal".*?[\r\n]*?.*?<img', re.MULTILINE or re.DOTALL)
    last = ""
    img = rImage.findall(data)
    timg = rThumb.findall(data)
    for i in timg:
        try:
            img.index(i)
        except:
            img.append(i)
    print "Bilder: ", img

    for image in img:
        path = md5.new(html2txt(image)).hexdigest()
        images.append( {'image': image,
                        'path' : str(path[0])+"/"+str(path[0:2])+"/"})
    images.sort()
    return images


def main():
    mysite = wikipedia.getSite()
    save_images = 0

    fi = wikipedia.argHandler(sys.argv[1], 'cht_cat')
    fkat = fi+'.txt'
    fzap = fi
    fotch = fi+'_ot.txt'

    if len(sys.argv)>=3:
        save_images = 1

    headers = {"Content-type": "application/x-www-form-urlencoded", 
               "User-agent": "RobHooftWikiRobot/1.0"}
    print "opening connection to", mysite.hostname(),
    conn = httplib.HTTPConnection(mysite.hostname())
    print " done"

    R = re.compile('.*/wiki/(.*)')

    f=codecs.open(fkat,'r',encoding='utf-8')
    otch = codecs.open(fotch, 'wb', 'utf-8')

    t1=f.readlines()
    nomerf=0
    for s in t1:
        if ord(s[0]) == 65279:
            s=s[1:]
        s=s.replace(u'\r',u'')
        if s[len(s)-1]==u'\n':
            s=s[:len(s)-1]
        if len(s)<1:
            continue

#    for article in sa:
#        filename = article.replace("/", "_")
#        filename = filename.replace(":", "_")
#        if os.path.isfile(output_directory + filename + ".txt") and overwrite_articles == False:
#            print "skipping " + article
#            continue
        data = ""
        ua = s
        enua = ua.replace(u' ', u'_').encode('utf-8')
        enua = urllib.quote(enua)

        url = '/wiki/'+ enua
        conn.request("GET", url, "", headers)
        response = conn.getresponse()
        data = response.read()
        if len(data) < 2:
            b = u'-       %s\n' % ua
            wikipedia.output(b)
            otch.write(b)
            otch.flush()

#        data = extractArticle(data)

        f = open(u'%s%06d.htm' % (fzap,nomerf), 'wb')
#        f.write(data['article'] + '\n' + data['footer'])
        f.write(data)
        f.flush()
        f.close()
        b = u'%06d  %s\n' % (nomerf,ua)
        wikipedia.output(b)
        otch.write(b)
        otch.flush()
        nomerf+=1

        if save_images:
#            images = extractImages(data['article'])
            images = extractImages(data)
            for i in images:
#                if overwrite_images == False and os.path.isfile(output_directory + i['image']):
#                    print "skipping existing " + i['image']
#                    continue
                print 'downloading ' + i['image'],
                uo = wikipedia.MyURLopener()
                file = uo.open( "http://upload.wikimedia.org/wikipedia/"
                                +mysite.lang + '/' + i['path'] + i['image'])
                content = file.read()
                if (len(content) < 500):
                    uo.close()
                    print "downloading from commons",
                    uo = wikipedia.MyURLopener()
                    file = uo.open( "http://commons.wikimedia.org/upload/"
                                    + i['path'] + i['image'])
                    #print "http://commons.wikimedia.org/upload/", i['path'] , i['image'], file
                    content = file.read()
#                f = open(output_directory + i['image'], "wb")
                f = open(u'%s%06d_%s' % (fzap,nomerf,i['image']), 'wb')
                f.write(content)
                f.close()
                b = u' +i %06d  %db  %s\n' % (nomerf,len(content),i['image'])
                wikipedia.output(b)
                otch.write(b)
                otch.flush()
#                print "\t\t", (len(content)/1024), "KB done"
    conn.close()

if __name__ == "__main__":
    main()

* podg_v_wiki.py

# -*- coding: UTF-8 -*-

__version__='$Id:'

import wikipedia
import re, sys
import codecs


def iskat(t,s):
    lt=len(t)
    ls=len(s)
    i=0
    for a in t[:lt-ls]:
        if t[i:i+ls]==s:
            return i
        i+=1
    return -1

def ubr_nk_prob(t):
    if len(t)<1:
        return t
    while (t[0]==u' ') or (t[0]==u'\n') or (t[0]==u'\t') or (t[0]==u'\r'):
        t=t[1:]
        if len(t)<1:
            return t
    while (t[len(t)-1]==u' ') or (t[len(t)-1]==u'\n') or (t[len(t)-1]==u'\t') or (t[len(t)-1]==u'\r'):
        t=t[:len(t)-1]
        if len(t)<1:
            return t
    return t


def main(ft,fvih):
    f=open(ft,'rb')
    fperv=1
#    rez=''
    for s in f.readlines():
#        if ord(s[0]) == 65279:
#            s=s[1:]
        s=s.replace('\r','')
        if s[len(s)-1]=='\n':
            s=s[:len(s)-1]
        if len(s)<1:
            continue
        if s[0]=='p':
            fi=ubr_nk_prob(s[2:])
            f1=open(fi,'rb')
            rez1=f1.read()
            if fperv:
                rez=rez1
                rez=rez[:0]
            rez+='* %s\n<code><pre><nowiki>\n'%fi
            print type(rez)
            print type(rez1)
            rez+=rez1
            print type(rez)
            print type('

') rez+='</nowiki>'

  1. rez1='* %s\n
    \n'%fi+rez1+'
    '
           if fperv:
  1. rez=rez1
               fperv=0
  1. else:
  2. rez+=rez1
           f1.close()
   fr=open(fvih, 'wb')
   fr.write(rez)
   fr.flush()
   fr.close()


if __name__ == "__main__":

   try:
       ft = wikipedia.argHandler(sys.argv[1], 'perevod')
       fvih = wikipedia.argHandler(sys.argv[2], 'perevod')
       main(ft,fvih)
   finally:
       wikipedia.stopme()

</nowiki>* zamen.py

# -*- coding: UTF-8 -*-

__version__='$Id:'

import wikipedia
import re, sys
import codecs
from bib_tekst import *

def vivod(b):
    wikipedia.output(b)
    otch.write(b)
    otch.flush()

def sravn_opc(t,i,s,opc):
    ls=len(s)
    if i+ls>len(t):
        return 0
    k=0
    while k<ls:
        a=t[i+k]
        b=s[k]
        if (u'k' in opc) and ( (k>0 and (s[k-1] in u'[:')) or 
                               (k==0 and (u'l' in opc)) ):
            a=a.upper()
            b=b.upper()
        if a!=b:
            return 0
        k+=1
    return 1

def iskat_opc(t,i,s0,opc):
    ss=[s0]
    if (u's' in opc) and len(s0)>2 and s0[len(s0)-2:]==u']]':
        ss.append(s0[:len(s0)-2]+u'|')

    lt=len(t)
    while i<lt:
        if (u'l' in opc) and (t[i:i+2]==u'[[' or t[i:i+2]==u'{{'):
            pn=i+2
            (pk,ik)=iskats_mn(t,pn,[u'[[',u']]',u'{{',u'}}',u'|',u'#'])
            if pk==-1:
                pk=lt
            while pn<pk and (t[pn] in [u' ',u'\n',u'\t',u'\r']):
                pn+=1
            if pn<pk and t[pn]==u':' and s0[0]!=u':':
                pn+=1
            while pn<pk and (t[pk-1] in [u' ',u'\n',u'\t',u'\r']):
                pk-=1

            ssilk=t[pn:pk].replace(u'_',u' ').replace(u'  ',u' ')
#            ssilk=perv_upper(ubr_nk_prob(ssilk))
#            vivod(u'%d %d %d %d %s %s\n'%(pn,pk,len(ssilk),len(s0),ssilk,s0))
            if len(ssilk)==len(s0) and sravn_opc(ssilk,0,s0,opc):
                return (pn,pk)
            i=pk
        for j in range(len(ss)):
            s=ss[j]
            ls=len(s)
            ravn=sravn_opc(t,i,s,opc)
            if ravn:
                if j==1:
                    m=i+ls
                    m1=iskkonpodp(t,m,u'[[',u']]')
                    if m1!=-1:
                        return (i,m1+2)
                return (i,i+ls)
        i+=1
    return (-1,-1)


def zam(n,tzfl,tzst,tzno,kol,komm):


    pl = wikipedia.Page(mysite,wikipedia.UnicodeToAsciiHtml(n))
    wikipedia.output(pl.title())
#    pl._contents=u'qwe [[qwe]] [[qwer|qwe]] [[fqwe]] qwe :qwe {{wqwe}} [[qwe|qwer]] [qwe] sqwr dwer [[:qwe]] qwet dqwe gqwej [[enn:qwe]] [[qwey]]\n'
#    pl._contents+=u'Qwe [[Qwe]] [[Qwer|Qwe]] [[fQwe]] Qwe :Qwe {{wQwe}} [[Qwe|Qwer]] [Qwe] sQwr dwer [[:Qwe]] Qwet dQwe gQwej [[enn:Qwe]] [[Qwey]]'


    try:
        nt = pl.get()
    except wikipedia.NoPage:
        b = u'- %s\n\n' % pl.title()
        vivod(b)
        return
    except wikipedia.IsRedirectPage:
        b = u'+ %s\n-> %s\n\n' % (pl.title(),pl.getRedirectTarget())
        vivod(b)
        return

    otch.write(u'+ %s\n' % pl.title())
    otch.flush()

    otch2.write(u'=========================\n%s\n-=======\n%s\n' % (pl.title(),nt))
    otch2.flush()

    j=0
    fliwi=1
    while j<kol:
        if (u'w' in tzfl[j]):
            fliwi=0
        j+=1

    if fliwi:
        oiwi = pl.interwiki()
        niwi = {}
        for pl2 in oiwi:
            if pl.site()!=pl2.site():
                niwi[pl2.site()] = pl2
        nt = wikipedia.replaceLanguageLinks(nt, {})

    nt = nt.replace(u'\r',u'')
#    while u' \n' in nt:
#        nt = nt.replace(u' \n',u'\n')

    fzam=0

    j=0
    while j<kol:
        zst=tzst[j]
        zst=zst[:len(zst)-1]
        zno=tzno[j]
        zno=zno[:len(zno)-1]
#        if mysite.lang==u'eo':
#            zst=zst.replace(u'Ĉ',u'Cx').replace(u'ĉ',u'cx')
#            zst=zst.replace(u'Ĝ',u'Gx').replace(u'ĝ',u'gx')
#            zst=zst.replace(u'Ĵ',u'Jx').replace(u'ĵ',u'jx')
#            zst=zst.replace(u'Ĥ',u'Hx').replace(u'ĥ',u'hx')
#            zst=zst.replace(u'Ŝ',u'Sx').replace(u'ŝ',u'sx')
#            zst=zst.replace(u'Ŭ',u'Ux').replace(u'ŭ',u'ux')
#            zno=zno.replace(u'Ĉ',u'Cx').replace(u'ĉ',u'cx')
#            zno=zno.replace(u'Ĝ',u'Gx').replace(u'ĝ',u'gx')
#            zno=zno.replace(u'Ĵ',u'Jx').replace(u'ĵ',u'jx')
#            zno=zno.replace(u'Ĥ',u'Hx').replace(u'ĥ',u'hx')
#            zno=zno.replace(u'Ŝ',u'Sx').replace(u'ŝ',u'sx')
#            zno=zno.replace(u'Ŭ',u'Ux').replace(u'ŭ',u'ux')
#        if (u'u' in tzfl[j]) and (zno.lower() in nt.lower()):
        if (u'u' in tzfl[j]):
            (p,pk)=iskat_opc(nt,0,zno,tzfl[j])
            if p!=-1:
                b=u'uzxe\n'
                wikipedia.output(b)
                otch.write(b)
                otch.flush()
                j+=1
                continue

#        nt = nt.replace(zst,zno)
        nz=0
        p0=0
        while (len(zno)>0 or len(zst)>0) and p0<len(nt):
            if zst==u'':
                if (u'n' in tzfl[j]):
                    p=0
                else:
                    p=len(nt)
                pk=p
            else:
#                p=iskats(nt,p0,zst)
                (p,pk)=iskat_opc(nt,p0,zst,tzfl[j])
            if p==-1:
                break
            nt=nt[:p]+zno+nt[pk:]
            p0=p+len(zno)
            nz+=1
            fzam=1
            if zst==u'':
                break

        b=u'%d\n' % nz
        wikipedia.output(b)
        otch.write(b)
        otch.flush()
        j+=1

    if not fliwi:
        pl1=wikipedia.Page(pl.site(),pl.title())
        pl1._contents=nt
        oiwi = pl1.interwiki()
        niwi = {}
        for pl2 in oiwi:
            if pl.site()!=pl2.site():
                niwi[pl2.site()] = pl2
        nt = wikipedia.replaceLanguageLinks(nt, {})

    if fzam:
        while nt[len(nt)-1:]==u'\n':
            nt=nt[:len(nt)-1]
#        if fliwi:
        nt+=wikipedia.replaceLanguageLinks(u'', niwi)

        otch2.write(u'+=======\n%s\n' % nt)
        otch2.flush()
        pl.put(nt, comment = komm, minorEdit = False)
    otch.write(u'\n')
    otch.flush()
    return


def main():
    zfl=0
    tzfl={}
    tzst={}
    tzno={}
    komm=u''
    n=[]
    f=codecs.open(filename,'rb',encoding='utf-8')

    i=0
    j=0
    for s in f.readlines():
#        wikipedia.output(u'%d\n' % ord(s[0]))
        if ord(s[0]) == 65279:
            s=s[1:]
#        wikipedia.output(u'%d %d  %s   %s\n' % (i,j,n,s))
        s=s.replace(u'\r',u'')
        if s[len(s)-1]!=u'\n':
            s=s+u'\n'

        if s[0:12] == u'============':
#            wikipedia.output(u'k%d %d  %s   %s\n' % (i,j,n,s))
            if i==100:
                b=u'!er %s\n' % n
                wikipedia.output(b)
                otch.write(b)
                otch.flush()
            if (i!=100) and (j>0):
#                wikipedia.output(u'kl%d %d  %s   %s\n' % (i,j,n,s))
                for n1 in n:
                    zam(n1,tzfl,tzst,tzno,j,komm)
            n=[]
            komm=u''
            i=0
            j=0
        elif i==0:
            if s[0:8]==u'?=======':
                i=2
            elif s[0]==u'|':
                komm=s[1:]
            elif s[0]!=u'#':
                n.append(s)
        elif i==2:
            zfl= (s[0]!=u'0')
            if zfl:
                tzfl[j]=s
                tzst[j]=u''
                tzno[j]=u''
                j+=1
            i=3
        elif i==3:
            if s[0:8]==u'-=======':
                i=4
            else:
                i=100
                wikipedia.output(u"er i==3\n")
        elif i==4:
            if s[0:8]==u'+=======':
                i=5
            else:
                if zfl:
                    tzst[j-1]+=s
        else: #if i==5:
            if s[0:8]==u'?=======':
                i=2
            else:
                if zfl:
                    tzno[j-1]+=s



#        wikipedia.output(u"No title found - skipping a page.")


#    text=''.join(text)

fi = wikipedia.argHandler(sys.argv[1], 'cht_cat')
filename = fi+'.txt'
fotch = fi+'_ot.txt'
fotch2 = fi+'_op.txt'

#filename = wikipedia.argHandler(sys.argv[1], 'cht_cat')
#fotch = wikipedia.argHandler(sys.argv[2], 'cht_cat')
#fotch2 = wikipedia.argHandler(sys.argv[3], 'cht_cat')

#fliwi=1
#if len(sys.argv)>=3 and sys.argv[2]==u'w':
#    fliwi=0

otch = codecs.open(fotch, 'w', 'utf-8')
otch2 = codecs.open(fotch2, 'w', 'utf-8')
mysite = wikipedia.getSite()

try:
    main()
except:
    wikipedia.stopme()
    raise
else:
    wikipedia.stopme()

* perev_kat.py

# -*- coding: UTF-8 -*-

__version__='$Id:'

import wikipedia
import re, sys, os
import codecs
import perevod
from bib_tekst import *
from bib_kateg import *
from get_buf import *

def vivod(b):
    wikipedia.output(b)
    otch.write(b)
    otch.flush()


def main(slov,vhjaz,vihjaz,fivh,firez):

    f=codecs.open(fivh,'rb',encoding='utf-8')
    frez=codecs.open(firez, 'w', 'utf-8')
    ish=[]
    for s in f.readlines():
        if ord(s[0]) == 65279:
            s=s[1:]
        s=ubr_nk_prob(s)
        ish.append(s)
    f.close

    ish.sort()
    vihsi=wikipedia.getSite(code = vihjaz,fam = u'wikipedia')

    spred=u''
    for s in ish:
        if s==spred:
            continue
        spred=s
        st=s.split(u':',2)
        if len(st)!=3:
            vivod(u'!!! len(st)!=3 s=%s\n'%s)
            continue
        kato=[Kateg(st[0],st[2])]
        nt_kat_sp=perev_kateg(slov,nssvoj,vihsi,kato,1,otch)
        otch.write(u'\n\n===============\n\n')
        otch.flush()
        for t in nt_kat_sp:
            frez.write(u'%s:%s\n%s\n\n'%(vihjaz,t,s))
            frez.flush()


if __name__ == "__main__":
    try:
        vhjaz = wikipedia.argHandler(sys.argv[1], 'perevod')
        vihjaz = wikipedia.argHandler(sys.argv[2], 'perevod')
        fslov = wikipedia.argHandler(sys.argv[3], 'perevod')
        fvh = wikipedia.argHandler(sys.argv[4], 'perevod')
        fi = wikipedia.argHandler(sys.argv[5], 'perevod')
        fvih= fi+u'.txt'
        fotch = fi+u'-ot.txt'
#        fotch2 = fi+u'-op.txt'
#        fipoln = fi+u'-pp.txt'
#        finesl = fi+u'-ns.txt'
#        fink = fi+u'-nk.txt'

        if vhjaz==u'-':
            vhjaz=u''

        otch = codecs.open(fotch, 'w', 'utf-8')
#        otch2 = codecs.open(fotch2, 'w', 'utf-8')
        slov={}
        nssvoj={}
        slovdop=[]
        perevod.zagruzslov(slov,slovdop,nssvoj,vhjaz,vihjaz,fslov)
        main(slov,vhjaz,vihjaz,fvh,fvih)
    except:
        wikipedia.stopme()
        raise
    else:
        wikipedia.stopme()

* perev_oivi.py

# -*- coding: UTF-8 -*-

__version__='$Id:'

import wikipedia
import re, sys, os
import codecs
import perevod
from bib_tekst import *
from bib_kateg import *

def vivod(b):
    wikipedia.output(b)
    otch.write(b)
    otch.flush()


def sozd(vhjaz,vihjaz,n,t):

    vhsi=wikipedia.getSite(code = vhjaz,fam = u'wikipedia')
    vihsi=wikipedia.getSite(code = vihjaz,fam = u'wikipedia')

    pl=wikipedia.Page(vhsi,n)
    pl._contents=t

    oiwi = pl.interwiki()
    s=u''
    for pl2 in oiwi:
        if pl2.site()==vihsi:
            s=pl2.title()

    if s==u'':
        return u''

    b=((u'|interwiki\n%s\n'+
            u'?=======\n1wu\n-=======\n+=======\n\n[[%s:%s]]\n'+
            u'============\n')%(s,vhjaz,n))

    return b

def main(vhjaz,vihjaz,filename,firez):

    t=u''
    n=u''
    f=codecs.open(filename,'rb',encoding='utf-8')

    frez=codecs.open(firez, 'w', 'utf-8')
    i=0
    for s in f.readlines():
        if ord(s[0]) == 65279:
            s=s[1:]
        if s[0:7] == u'=======':
            if i==1:
                rez=sozd(vhjaz,vihjaz,n,t)
                if rez!=None:
                    frez.write(rez)
                    frez.flush()
            i=0    
            t=u''
        elif i==0:
            n=ubr_nk_prob(s)
            i=1
        else:
            t=t+s


if __name__ == "__main__":
    try:
        vhjaz = wikipedia.argHandler(sys.argv[1], 'perevod')
        vihjaz = wikipedia.argHandler(sys.argv[2], 'perevod')
        fvh = wikipedia.argHandler(sys.argv[3], 'perevod')
        fi = wikipedia.argHandler(sys.argv[4], 'perevod')
        fvih= fi+u'.txt'
#        fotch = fi+u'-ot.txt'
#        otch = codecs.open(fotch, 'w', 'utf-8')
        main(vhjaz,vihjaz,fvh,fvih)
    except:
        wikipedia.stopme()
        raise
    else:
        wikipedia.stopme()

* perev_sta.py

# -*- coding: UTF-8 -*-

__version__='$Id:'

import wikipedia
import re, sys, os
import codecs
import perevod
from bib_tekst import *
from bib_kateg import *

def vivod(b):
    wikipedia.output(b)
    otch.write(b)
    otch.flush()

def prov_dobav_katcom1(nscom,zapriz,katcomsp,n):
    n=perv_upper(ubr_nk_prob(n))
    vivod(u' prov_dobav_katcom1 %s\n'%n)
    p=iskats(n,0,u'#')
    if p!=-1:
        n=n[:p]
        vivod(u' prov_dobav_katcom1# %s\n'%n)
    t=u'Category:'+n
    if nscom.has_key(t) and (not zapriz.has_key(n)):
        t1=u'[['+t+u']]\n'
        if not t1 in katcomsp:
            katcomsp.append(t1)
            vivod(u' <-- ++\n')
            return 1
    return 0

def prov_dobav_katcom(nscom,zapriz,katcomsp,vhs,jaz):
    prov_dobav_katcom1(nscom,zapriz,katcomsp,vhs)
    if jaz==u'en':
        vhok=vhs[len(vhs)-1]
        if vhok==u'y':
            vhs1=vhs[:len(vhs)-1]+u'ies'
            prov_dobav_katcom1(nscom,zapriz,katcomsp,vhs1)
        else: 
            vhs1=vhs+u's'
            prov_dobav_katcom1(nscom,zapriz,katcomsp,vhs1)
            vhs1=vhs+u'es'
            prov_dobav_katcom1(nscom,zapriz,katcomsp,vhs1)

def sozd_katcom(nscom,zapriz,nomio,kat):

    katcomsp=[]
    katprob=[]

    for (n,jaz) in nomio:
        katprob.append( (n,jaz) )
        prov_dobav_katcom(nscom,zapriz,katcomsp,n,jaz)

#    if len(katcomsp)<1:
#        shcom_v_katcom(nscom,zapriz,katcomsp,shcomi+shcomo)

    if len(katcomsp)<1:
        vivod(u'- katcom nomio\n')
        for ka in kat:
            katprob.append( (ka.n,ka.jaz) )
            prov_dobav_katcom(nscom,zapriz,katcomsp,ka.n,ka.jaz)

    if len(katcomsp)<1:
        vivod(u'- katcom\n')
        for (n1,jaz) in katprob:
            n2=n1.replace(u'(',u' ').replace(u')',u' ')
            for n in n2.split(u' '):
                if len(n)>=4:
                    prov_dobav_katcom(nscom,zapriz,katcomsp,n,jaz)

    if len(katcomsp)<1:
        vivod(u'- katcom2\n')

    katcom=u''
    for t in katcomsp:
        katcom+=t

    vivod(u'========katcom=\n%s==========\n' % katcom)
    return katcom

def prov_te_format_kart(t):
    minrzm=50
    t=ubr_nk_prob(t)
    if (t==u'thumb') or (t==u'thumbnail') or (t==u'frame') or (t==u'framed'):
        return 1
    if (t==u'left') or (t==u'center') or (t==u'right'):
        return 1
    if len(t)<3:
        return 0
    x=0
    i=0
    while i<6 and i<len(t) and t[i].isdigit():
        x=x*10+ord(t[i])-ord(u'0')
        i+=1
    if (i>=1) and (t[i:]==u'px'):
        if x<minrzm: 
            return 2
        return 1
    return 0

class Pereved_sta:
    def __init__(self, vhpn,n,t,statn,stat):
        self.vhpn = vhpn
        self.n = n
        self.n2 = n
        self.t = t
        self.nkat = 0
        self.statn = statn
        self.stat = stat
        self.prioritet = 0
        self.npredl = 0
    def vz_prioritet(self):
        return self.prioritet

def sum_per_stat(s,s1):
    if s1.flp:
        s.flp = 1
    s.nup += s1.nup
    s.nnp += s1.nnp
    s.nvap += s1.nvap
    s.snp += s1.snp
    s.svap += s1.svap


def sozd(slov,nscom,zapriz,sp_obr_st,vhjaz,vihjaz,n,t,snka,sniz,snizkat,
                              schss,tpref,
                              fl_d_iwi_n,fl_b_tekst):

    vivod(u'\n\n\n%s\n'%n)

    vhpn=vhjaz+u':'+n
    if fl_d_iwi_n and not sp_obr_st.has_key(vhpn):
        vivod(u'<- net v tbl\n')
        return
    if slov.has_key((vhjaz,n)) and slov[(vhjaz,n)].riwi:
        vivod(u'<- uzxe v real ivi\n')
        return

    vhsi=wikipedia.getSite(code = vhjaz,fam = u'wikipedia')
    vihsi=wikipedia.getSite(code = vihjaz,fam = u'wikipedia')
    vhprefi=vhsi.family.image_namespace(vhjaz)
    vihprefi=vihsi.family.image_namespace(vihjaz)

    pl=wikipedia.Page(vhsi,n)
    pl._contents=t

    nomio=[(n,vhjaz)]

    oiwi = pl.interwiki()
    niwi = {}
    for pl2 in oiwi:
        if pl2.site()==vihsi:
            vivod(u'uzxe vihjaz %s %s\n'%(n,pl2.title()))
            return
        if pl2.site()!=vhsi:
            niwi[pl2.site()] = pl2
            nomio.append( (pl2.title(),pl2.site().lang) )
    niwi[vhsi]=pl

    plkat = pl.categories()
    kato=[]
    for pka in plkat:
        kato.append(Kateg(vhjaz,pka.title(),1))

    t = wikipedia.removeLanguageLinks(t)
    t = wikipedia.removeCategoryLinks(t,vhsi)

    rezn_t=perevod.perevod_iwi_spis(slov,vhjaz,vihjaz,n)
    if len(rezn_t)>0:
        fl_iwi_n=1
        rezn=rezn_t[0]
        statn=perevod.Perev_stat()
    else:
        fl_iwi_n=0
        (rezn,statn)=perevod.perevod_stat(slov,vhjaz,vihjaz,n)

    vihpl=wikipedia.Page(vihsi,rezn)

    t = t.replace(u'\r',u' ').replace(u'\t',u' ')
    while u'  ' in t:
        t = t.replace(u'  ',u' ')
    t = t.replace(u' \n',u'\n')

    tks=[u'. ',u'.\n',u'\n\n',u'\n ',u'\n:',u'\n*',u'[[',u'{{',u'\n=',u'=\n',
           u'<gallery>',u'<math>',u']]',]

    tkzs={u'[[':u']]',u'{{':u'}}',u'<gallery>':u'</gallery>',
               u'<math>':u'</math>'}

    st=[]
    rt=u''
    sst=[]
    uktbl=[]
    npredl=0
    snizdl0=len(sniz)

    p=0
    while (not fl_b_tekst) and p<len(t):
        (p1,i)=iskats_mn(t,p,tks)

#        vivod(u'nach predl %d %d\n'%(p1,i))

        if p1==-1:
            p1=len(t)  
#        elif tks[i][0]==u'.' or tks[i][0]==u'=':
#        else:
        elif not tkzs.has_key(tks[i]):
            p1+=1

        rt+=t[p:p1].replace(u'[[',u' ').replace(u']]',u' ')

        kon_predl=1
        ss_vih=u''

        if tks[i]==u'[[':
            os=u'[['
            zs=u']]'
            p2=iskkonpodp(t,p1+len(os),os,zs)
            if p2==-1:
                p2=len(t)
            (pr,ir)=iskats_mn(t,p1+len(os),[u'|',os,zs])
            pnt=p1+len(os)
            pks=p2
            if pr!=-1 and ir==0:
                pnt=pr+1
                pks=pr
            ss=ubr_nk_prob(t[p1+len(os):pks])
            ss_pdt=iskats(ss,1,u':')

            if ss_pdt==-1:
                kon_predl=0

                perss=perevod.perevod_iwi_spis(slov,vhjaz,vihjaz,ss)
                p3=p2+len(zs)
                p4=p3
                if len(perss)>=1:
                    while p4<len(t) and t[p4].isalpha():
                        p4+=1

                    uktbl.append(len(rt))
                    rt+=( (t[pnt:p2]+t[p3:p4]).
                                 replace(u'[[',u' ').
                                 replace(u']]',u' ')
                          )
                    uktbl.append(len(rt))
                    sst.append(perss[0])
                    if schss.has_key(perss[0]):
                        schss[perss[0]]+=1
                    else:
                        schss[perss[0]]=1

                else:
                    rt+=t[pnt:p2].replace(u'[[',u' ').replace(u']]',u' ')
                p=p4
            elif RBR(vhprefi,ss[:ss_pdt]) or RBR(u'Image',ss[:ss_pdt]):
                ss=ss[ss_pdt+1:]
                ss=perv_upper(ubr_nk_prob(ss.replace(u'_',u' ')))
                if nscom.has_key(u'Image:'+ss):
                    ss_vih=vihprefi+u':'+ss
                else:
                    ss_vh=vhprefi+u':'+ss
                    perss=perevod.perevod_iwi_spis(slov,vhjaz,vihjaz,ss_vh)
                    if len(perss)>=1:
                        ss_vih=perss[0]
                    else:
                        b=u'[['+vhjaz+u':'+ss_vh+u']]'
                        if not b in sniz:
                            sniz.append(b)
                            snizkat.append(u'')

        if kon_predl:
#            st.append(rt)
            while u'  ' in rt:
                rt = rt.replace(u'  ',u' ')
            if len(rt)>=20:
                npredl+=1
            (r1,s1)=perevod.perevod_stat(slov,vhjaz,vihjaz,rt,uktbl=uktbl)
            qi=len(sst)-1
            while qi>=0:
                ss=sst[qi]
                q1=uktbl[qi*2]
                q2=uktbl[qi*2+1]
                if perv_upper(ss)==perv_upper(r1[q1:q2]):
                    r1=r1[:q1]+u'[['+r1[q1:q2]+u']]'+r1[q2:]
                else:
                    r1=r1[:q1]+u'[['+ss+u'|'+r1[q1:q2]+u']]'+r1[q2:]
                qi-=1

            st.append((r1,s1))
            rt=u''
            sst=[]
            uktbl=[]

            if tkzs.has_key(tks[i]):
                os=tks[i]
                zs=tkzs[tks[i]]
                p2=iskkonpodp(t,p1+len(os),os,zs)
                if p2==-1:
                    p=len(t)
                else:
                    p=p2+len(zs)  
                if tks[i]==u'<math>':
                    st.append((t[p1:p],perevod.Perev_stat()))
                elif ss_vih!=u'':
                    pa=p1+len(os)
                    while 1:
                        (pr,ir)=iskats_mn(t,pa,[u'|',zs])
                        if pr==-1:
                            pr=len(t)
                        if pa==p1+len(os) or prov_te_format_kart(t[pa:pr]):
                            pa=pr
                            if pa<len(t) and t[pa]==u'|':
                                pa+=1
                            continue
                        break
                    st.append((u'[['+ss_vih+u'|'+t[pnt:pa],
                                              perevod.Perev_stat()))
                    p=pa
            else:
                p=p1


    rez=u''
    stat=perevod.Perev_stat()
    for (r1,s1) in st:
        rez+=r1
        sum_per_stat(stat,s1)

    while u'  ' in rez:
        rez = rez.replace(u'  ',u' ')
    while u'\n\n\n' in rez:
        rez = rez.replace(u'\n\n\n',u'\n\n')
    rez=ubr_nk_prob(rez)

    nt_kat_sp=perev_kateg(slov,nssvoj,vihsi,kato,0,otch)

    if len(nt_kat_sp)<1:
        for ka in kato:
            jaz=ka.jaz
            issite=wikipedia.getSite(jaz,fam = pl.site().family)
            iskatprefi=issite.family.category_namespace(jaz)
            b=jaz+u':'+iskatprefi+u':'+ka.n
            if not b in snka:
                snka.append(b)

    nt_kat=kateg_v_tekst(nt_kat_sp)

    rez+=u'\n\n'+wikipedia.replaceLanguageLinks(nt_kat, niwi, site=vihsi)

    ps=Pereved_sta(vhpn,rezn,rez,statn,stat)
    ps.nkat=len(nt_kat_sp)

    osnp={}
    for a in statn.snp:
        osnp[a]=1
    for a in stat.snp:
#        vivod(u'%s\n'%a)
#        if not osnp.has_key(a):
        osnp[a]=1


    if snizdl0<len(sniz):
        katcom=sozd_katcom(nscom,zapriz,nomio,kato)
        katcom=katcom.replace(u'\n',u' ')
        i=snizdl0
        while i<len(sniz):
            snizkat[i]=katcom
            i+=1


    ps.npredl=npredl
    prioritet = (len(osnp)+statn.nvap)*10000+stat.nvap*10

    if not fl_b_tekst:
        if npredl<3:
            prioritet+=100000000
        else:
            prioritet/=npredl
        if tpref!=u'':
            ps.t=u'{{polurinda movu|%s}}\n'%rezn+ps.t
            ps.n2=tpref+ps.n
        else:
            ps.t=u'{{polurinda}}\n'+ps.t

    if len(nt_kat_sp)<1:
        prioritet+=10000000

    if fl_d_iwi_n and not fl_iwi_n:
        prioritet+=200000000

    ps.prioritet=prioritet

    vivod(u'%d  %d  %s\n\n'%(ps.prioritet,ps.npredl,n))

    return ps


def main(slov,nscom,zapriz,sp_obr_st,vhjaz,vihjaz,
                            filename,firez,finesl,finka,finiz,fiuzxe,fipref,
                            fl_d_iwi_n,fl_b_tekst,fl_schss):

    t=u''
    n=u''
    f=codecs.open(filename,'rb',encoding='utf-8')
    vivod(u'fipref=%s\n'%fipref)
    tpref=u''
    if fipref!=u'.':
        fpref=codecs.open(fipref, 'rb', 'utf-8')
        s=fpref.readlines()[0]
        fpref.close()
        if len(s)>0 and ord(s[0]) == 65279:
            s=s[1:]
        tpref=ubr_nk_prob(s)
    vivod(u'tpref=%s\n'%tpref)

    frez=codecs.open(firez, 'w', 'utf-8')
    fnesl=codecs.open(finesl, 'w', 'utf-8')
    fnka=codecs.open(finka, 'w', 'utf-8')
    fniz=codecs.open(finiz, 'w', 'utf-8')
    fuzxe=codecs.open(fiuzxe, 'w', 'utf-8')
    tps=[]
    snka=[]
    sniz=[]
    snizkat=[]
    schss={}
    i=0
    for s in f.readlines():
#        wikipedia.output(u'%d\n' % ord(s[0]))
        if ord(s[0]) == 65279:
            s=s[1:]
#        wikipedia.output(u'%d   %s   %s\n' % (i,n,s))
        if s[0:7] == u'=======':
            if i==1:
                rez=sozd(slov,nscom,zapriz,sp_obr_st,vhjaz,vihjaz,
                                   n,t,snka,sniz,snizkat,
                                   schss,tpref,fl_d_iwi_n,fl_b_tekst)
                if rez!=None:
                    tps.append(rez)
            i=0    
            t=u''
        elif i==0:
            n=s
            while len(n)>0 and n[len(n)-1]==u'\n':
                n=n[:len(n)-1]
            i=1
        else:
            t=t+s

    for t in snka:
        fnka.write(u'%s\n'%t)
        fnka.flush()
    for i in range(len(sniz)):
        fniz.write(u'%s %s\n'%(sniz[i],snizkat[i]))
        fniz.flush()

    if fl_schss==1:
        for tp in tps:
            if schss.has_key(tp.n):
                tp.prioritet=-schss[tp.n]
            else:
                tp.prioritet=1000
    elif fl_schss==2:
        for tp in tps:
            tp.prioritet=-tp.nkat

#    tps.sort(key=Pereved_sta.vz_prioritet,reverse=True)
    tps.sort(key=Pereved_sta.vz_prioritet)

    osnpo={}
    osvapo={}
    kolotch=50
    i=0
    while i<len(tps):

        pr0=tps[i].prioritet
        osnp={}
        osvap={}
        osnp1={}
        osvap1={}
        j=i
        while j<len(tps) and pr0==tps[j].prioritet:
            frez.write(u'%s\n%s\n========\n' % (tps[j].n2,tps[j].t))
            frez.flush()
            fuzxe.write(u'%s\n' % tps[j].vhpn)
            fuzxe.flush()

            perevod.perev_uch_nezsl(osnp,osvap,tps[j].statn)
            perevod.perev_uch_nezsl(osnp,osvap,tps[j].stat)

            for sl in osnp.iterkeys():
                if not osnpo.has_key(sl):
                    osnp1[sl]=osnp[sl]

            for sl in osvap.iterkeys():
                if not osvapo.has_key(sl):
                    osvap1[sl]=osvap[sl]

            perevod.perev_uch_nezsl(osnpo,osvapo,tps[j].statn)
            perevod.perev_uch_nezsl(osnpo,osvapo,tps[j].stat)

            j+=1

        fnesl.write(u'prior=%d  n=%d\n========\n' % (pr0,j-i))
        fnesl.flush()
        perevod.perev_pech_nezsl(slov,vhjaz,vihjaz,osnp1,osvap1,fnesl,kolotch)
        fnesl.write(u'=========================\n')
        fnesl.flush()
        i=j  

    fnesl.write(u'=========================*=*=*=*=\n')
    fnesl.flush()
    kolotch2=1000
    perevod.perev_pech_nezsl(slov,vhjaz,vihjaz,osnpo,osvapo,fnesl,kolotch2)

    fnesl.write(u'\n\n')
    fnesl.flush()

    ssnp=[]
    for t, n in osnpo.iteritems():
        ssnp.append((n,t))

    ssnp.sort(reverse=True)
    i=0
    for n, t in ssnp:
        if len(t)>=4 and t.startswith(u'Un'):
            t1=t[2:]
#            prist=u'Mal'
            prist=u'Ne'
        elif len(t)>=4 and (t.startswith(u'Ill') or
                          t.startswith(u'Imb') or 
                          t.startswith(u'Imm') or 
                          t.startswith(u'Imp') or 
                          t.startswith(u'Irr') or 
                          t.startswith(u'In') ):
            t1=t[2:]
            prist=u'Ne'
        else:
            continue

        (t2,stat2)=perevod.perevod_stat(slov,vhjaz,vihjaz,t1)
        if len(t2)>=6 and t2[0]==u'(' and t2[len(t2)-1]==u')':
            t2=t2[1:len(t2)-1]
        t2t=t2.split(u', ')
        t2r=u''
        i=0
        while i<len(t2t):
            t2r+=prist+t2t[i]
            i+=1
            if i>=len(t2t):
                break
            t2r+=u', '

        fnesl.write(u'%s : %s\n'%(t,t2r))
        fnesl.flush()



if __name__ == "__main__":
    try:
        vhjaz = wikipedia.argHandler(sys.argv[1], 'perevod')
        vihjaz = wikipedia.argHandler(sys.argv[2], 'perevod')
        fslov = wikipedia.argHandler(sys.argv[3], 'perevod')
        fvh = wikipedia.argHandler(sys.argv[4], 'perevod')
        fipref = wikipedia.argHandler(sys.argv[5], 'perevod')
        fi = wikipedia.argHandler(sys.argv[6], 'perevod')
        fvih= fi+u'.txt'
        fotch = fi+u'-ot.txt'
        fipoln = fi+u'-pp.txt'
        finesl = fi+u'-ns.txt'
        finka = fi+u'-nk.txt'
        finiz = fi+u'-iz.txt'
        fiuzxe = fi+u'-uz.txt'

        fl_d_iwi_n=0
        if len(sys.argv)>=8 and (u'w' in sys.argv[7]):
            fl_d_iwi_n=1
        fl_b_tekst=0
        if len(sys.argv)>=8 and (u'k' in sys.argv[7]):
            fl_b_tekst=1
        fl_schss=0
        if len(sys.argv)>=8 and (u's' in sys.argv[7]):
            fl_schss=1
        elif len(sys.argv)>=8 and (u'c' in sys.argv[7]):
            fl_schss=2

        otch = codecs.open(fotch, 'w', 'utf-8')
        #otch2 = codecs.open(fotch2, 'w', 'utf-8')
        slov={}
        nssvoj={}
        nscom={}
        zapriz={}
        slovdop={}
        perevod.zagruzslov(slov,slovdop,nssvoj,vhjaz,vihjaz,fslov)

        if slovdop.has_key(u'c'):
            nscom=slovdop[u'c']
        if slovdop.has_key(u'z'):
            zapriz=slovdop[u'z']

        sp_obr_st={}
        if slovdop.has_key(u'spst'):
            sp_obr_st=slovdop[u'spst']
        sp_obr_uzxe={}
        if slovdop.has_key(u'pe_uzxe'):
            sp_obr_uzxe=slovdop[u'pe_uzxe']
        for t in sp_obr_uzxe:
            if sp_obr_st.has_key(t):
                del sp_obr_st[t]
        main(slov,nscom,zapriz,sp_obr_st,vhjaz,vihjaz,
                            fvh,fvih,finesl,finka,finiz,fiuzxe,fipref,
                            fl_d_iwi_n,fl_b_tekst,fl_schss)
    except:
        wikipedia.stopme()
        raise
    else:
        wikipedia.stopme()

* perev_tit.py

# -*- coding: UTF-8 -*-

__version__='$Id:'

import wikipedia
import re, sys
import codecs
import perevod

def iskat(t,s):
    lt=len(t)
    ls=len(s)
    i=0
    while i<=lt-ls:
        if t[i:i+ls]==s:
            return i
        i+=1
    return -1

def iskats_mn(t,p0,ss):
    i=p0
    while i<len(t):
        for j in range(len(ss)):
            if t[i].startswith(ss[j]):
                return (i,j)
        i+=1
    return (-1,0)

def perv_upper(t):
    if len(t)<1:
        return u''
    return t[0].upper()+t[1:]

def ubr_nk_prob(t):
    i=0
    while i<len(t) and ((t[i]==u' ') or (t[i]==u'\n') or (t[i]==u'\t') or
                        (t[i]==u'\r')):
        i+=1
    j=len(t)-1
    while j>i and ((t[j]==u' ') or (t[j]==u'\n') or (t[j]==u'\t') or
                        (t[j]==u'\r')):
        j-=1
    return t[i:j+1]


def main(slov,nssvoj,vhjaz,vihjaz,fvh,fvih,fipoln,finesl,fispit,flvsjo):

    otch.write(u'len(nssvoj)=%d\n'%len(nssvoj))
    otch.flush()

    kolotch=1500
    f0=codecs.open(fvh,'rb',encoding='utf-8')
    f1=codecs.open(fvih, 'w', 'utf-8')
    fpoln=codecs.open(fipoln, 'w', 'utf-8')
    fnesl=codecs.open(finesl, 'w', 'utf-8')
    fspit=codecs.open(fispit, 'w', 'utf-8')
    osnp={}
    osvap={}
    for s in f0.readlines():
#        wikipedia.output(u'%d\n' % ord(s[0]))
        if ord(s[0]) == 65279:
            s=s[1:]
#        wikipedia.output(u'%d %d  %s   %s\n' % (i,j,n,s))
        s=s.replace(u'\r',u'').replace(u'_',u' ')
#        if s[len(s)-1]!=u'\n':
#            s=s+u'\n'

        s=ubr_nk_prob(s)

#        (rez,fl_perev)=perevod(slov,vhjaz,vihjaz,s)
        (rez,stat)=perevod.perevod_stat(slov,vhjaz,vihjaz,s)


        if flvsjo or (stat.nnp==0 and stat.nvap==0):
            f1.write(s+u'|'+rez+u'\n')
            f1.flush()
            if nssvoj.has_key(rez):
                otch.write(rez+u'\n')
                otch.write(u' <- uzxe\n')
                otch.flush()
            else:
                fspit.write(s+u'\n')
                fspit.flush()

        fpoln.write(rez+u'\n')
        fpoln.flush()

        perevod.perev_uch_nezsl(osnp,osvap,stat)

    perevod.perev_pech_nezsl(slov,vhjaz,vihjaz,osnp,osvap,fnesl,kolotch)


if __name__ == "__main__":
    try:
        vhjaz = wikipedia.argHandler(sys.argv[1], 'perevod')
        vihjaz = wikipedia.argHandler(sys.argv[2], 'perevod')
        fslov = wikipedia.argHandler(sys.argv[3], 'perevod')
        fvh = wikipedia.argHandler(sys.argv[4], 'perevod')
        fi = wikipedia.argHandler(sys.argv[5], 'perevod')
        fvih= fi+u'.txt'
        fotch = fi+u'-ot.txt'
        fipoln = fi+u'-pp.txt'
        finesl = fi+u'-ns.txt'
        fispit = fi+u'-st.txt'

        flvsjo=0
        if len(sys.argv)>=7 and sys.argv[6]==u'2':
            flvsjo=1

        otch = codecs.open(fotch, 'w', 'utf-8')
        #otch2 = codecs.open(fotch2, 'w', 'utf-8')
        slov={}
        nssvoj={}
        slovdop={}
        perevod.zagruzslov(slov,slovdop,nssvoj,vhjaz,vihjaz,fslov)
        main(slov,nssvoj,vhjaz,vihjaz,fvh,fvih,fipoln,finesl,fispit,flvsjo)
    finally:
        wikipedia.stopme()

* vich_slivi.py

# -*- coding: UTF-8 -*-

__version__='$Id:'

import wikipedia
import re, sys
import codecs
from bib_tekst import *

def vivod(b):
    wikipedia.output(b)
    otch.write(b)
    otch.flush()


def main():
    n=[]
    f=codecs.open(filename,'rb',encoding='utf-8')
    f1=codecs.open(fi1,'rb',encoding='utf-8')
    global otch
    otch = codecs.open(fotch, 'w', 'utf-8')
    frez = codecs.open(firez, 'w', 'utf-8')

    ns={}
    for s in f1.readlines():
        if ord(s[0]) == 65279:
            s=s[1:]
        s=ubr_nk_prob(s)
        if len(s)>0 and not s.startswith(vhjaz+u':'):
            ns[s]=1

    sch={}
    for s in f.readlines():
        if ord(s[0]) == 65279:
            s=s[1:]
        s=ubr_nk_prob(s)
        if ns.has_key(s):
            vivod(u'- %s\n'%s)
        else:
            frez.write(s+u'\n')
            frez.flush()

#fi = wikipedia.argHandler(sys.argv[1], 'cht_cat')
#filename = fi+'.txt'
#fotch2 = fi+'_op.txt'

vhjaz = wikipedia.argHandler(sys.argv[1], 'cht_cat')
fi1 = wikipedia.argHandler(sys.argv[2], 'cht_cat')
filename = wikipedia.argHandler(sys.argv[3], 'cht_cat')
fi = wikipedia.argHandler(sys.argv[4], 'cht_cat')
firez = fi+'.txt'
fotch = fi+'_ot.txt'

#otch2 = codecs.open(fotch2, 'w', 'utf-8')
mysite = wikipedia.getSite()

try:
    main()
except:
    wikipedia.stopme()
    raise
else:
    wikipedia.stopme()

* zamen_slivi.py

# -*- coding: UTF-8 -*-

__version__='$Id:'

import wikipedia
import re, sys
import codecs
from bib_tekst import *

def vivod(b):
    wikipedia.output(b)
    otch.write(b)
    otch.flush()


def main():
    n=[]
    f=codecs.open(filename,'rb',encoding='utf-8')
    f1=codecs.open(fi1,'rb',encoding='utf-8')
    global otch
    otch = codecs.open(fotch, 'w', 'utf-8')
    frez = codecs.open(firez, 'w', 'utf-8')

    ns={}
    for s in f1.readlines():
        if ord(s[0]) == 65279:
            s=s[1:]
        s=ubr_nk_prob(s)
        st=s.split(u'|')
        if len(st)==2:
            ns[vhjaz+u':'+ubr_nk_prob(st[0])]=vhjaz+u':'+ubr_nk_prob(st[1])

    sch={}
    for s in f.readlines():
        if ord(s[0]) == 65279:
            s=s[1:]
        s=ubr_nk_prob(s)
        i=0
        while ns.has_key(s):
            if sch.has_key(s):
                sch[s]+=1
            else:
                sch[s]=1
            s=ns[s]
            i=i+1
            if i>=10:
                vivod(u'!cikl %s\n'%s)
                break

        frez.write(s+u'\n')
        frez.flush()
    for ss,sn in ns.iteritems():
        n=0
        if sch.has_key(ss):
            n=sch[ss]
        vivod(u'%d %s | %s\n'%(n,ss,sn))


#fi = wikipedia.argHandler(sys.argv[1], 'cht_cat')
#filename = fi+'.txt'
#fotch2 = fi+'_op.txt'

vhjaz = wikipedia.argHandler(sys.argv[1], 'cht_cat')
filename = wikipedia.argHandler(sys.argv[2], 'cht_cat')
fi1 = wikipedia.argHandler(sys.argv[3], 'cht_cat')
fi = wikipedia.argHandler(sys.argv[4], 'cht_cat')
firez = fi+'.txt'
fotch = fi+'_ot.txt'

#otch2 = codecs.open(fotch2, 'w', 'utf-8')
mysite = wikipedia.getSite()

try:
    main()
except:
    wikipedia.stopme()
    raise
else:
    wikipedia.stopme()

* podg_mma_f0_v2.py

# -*- coding: UTF-8 -*-

__version__='$Id:'

import wikipedia
import re, sys
import codecs
import xmlreader
from bib_tekst import *


def ubrkoment (text):
    # Ignore tekst within nowiki tags and HTML comments
    nowikiOrHtmlCommentR = re.compile(r'<nowiki>.*?|<!--.*?-->', re.IGNORECASE | re.DOTALL)
    match = nowikiOrHtmlCommentR.search(text)
    while match:
        text = text[:match.start()] + text[match.end():]    
        match = nowikiOrHtmlCommentR.search(text)
    return text

def korr_im (jaz,t,pref):

    for p in pref:
        if t.startswith(p):
            t=t[len(p):]
            break

    t=perv_upper(t)

    if u':' in t:
        if ( (not t.startswith(u'Category:')) and 
                     (not t.startswith(u'Image:')) ):
            katprefi=mysite.family.category_namespace(jaz)
            if not t.startswith(katprefi+u':'):
                izprefi=mysite.family.image_namespace(jaz)
                if not t.startswith(izprefi+u':'):
                    return u''
    return t

iskkat=[
    u'[[Kategorio:Naskiĝ',
#    u'[[kategorio:Naskiĝ',
    u'[[Kategorio:naskiĝ',
#    u'[[kategorio:naskiĝ',
    u'[[Kategorio:Mort',
#    u'[[kategorio:Mort',
    u'[[Kategorio:mort',
#    u'[[kategorio:mort',
       ]

def main(vhjaz,fvh,fvih,fipref):
    tzfl=0
    tzst={}
    tzno={}


    pref=[]
    if fipref!=u'':
        fpref=codecs.open(fipref,'rb',encoding='utf-8')
        for s in fpref.readlines():
            if ord(s[0]) == 65279:
                s=s[1:]
            s=s.replace(u'\r',u'')
            if s[len(s)-1]==u'\n':
                s=s[:len(s)-1]
            pref.append(s)
        fpref.close()


    n=u''
#    f0=codecs.open(fvh,'rb',encoding='utf-8')
    f1=codecs.open(fvih, 'w', 'utf-8')

    insite=wikipedia.getSite(vhjaz,fam = u'wikipedia')

    tblredir = {}
    # open xml dump and read page titles out of it
    dump = xmlreader.XmlDump(fvh)
    redirR = wikipedia.getSite().redirectRegex()
    readPagesCount = 0

    for entry in dump.parse():
        readPagesCount += 1
        # always print status message after 1000 pages
        if readPagesCount % 1000 == 0:
            print '%i pages read...' % readPagesCount
        m = redirR.search(entry.text)
        if (not m) and (not u':' in entry.title):
            t=entry.text.replace(u'[[ ',u'[[').replace(u': ',u':')
            t=t.replace(u'[[k',u'[[K')
            dop=0
            for isk in iskkat:
                if isk in t:
                    dop=1
                    break
            if dop:
                b=u'%s\n' % entry.title
                f1.write(b)
                f1.flush()

vhjaz = u'eo'
#vhjaz = wikipedia.argHandler(sys.argv[1], 'slov_iz_xml')
fvh = wikipedia.argHandler(sys.argv[1], 'slov_iz_xml')
fvih = wikipedia.argHandler(sys.argv[2], 'slov_iz_xml')

fpref=u''

#fotch = wikipedia.argHandler(sys.argv[4], 'slov_iz_xml')

#mysite = wikipedia.getSite()
#otch = codecs.open(fotch, 'w', 'utf-8')
mysite = wikipedia.getSite()

try:
    main(vhjaz,fvh,fvih,fpref)
except:
    wikipedia.stopme()
    raise
else:
    wikipedia.stopme()

</nowiki>

* podg_mma_f1.py

# -*- coding: UTF-8 -*-

__version__='$Id:'

import wikipedia
import re, sys
import codecs
import xmlreader
from bib_tekst import *


def ubrkoment (text):
    # Ignore tekst within nowiki tags and HTML comments
    nowikiOrHtmlCommentR = re.compile(r'<nowiki>.*?|<!--.*?-->', re.IGNORECASE | re.DOTALL)
    match = nowikiOrHtmlCommentR.search(text)
    while match:
        text = text[:match.start()] + text[match.end():]    
        match = nowikiOrHtmlCommentR.search(text)
    return text

def korr_im (jaz,t,pref):

    for p in pref:
        if t.startswith(p):
            t=t[len(p):]
            break

    t=perv_upper(t)

    if u':' in t:
        if ( (not t.startswith(u'Category:')) and 
                     (not t.startswith(u'Image:')) ):
            katprefi=mysite.family.category_namespace(jaz)
            if not t.startswith(katprefi+u':'):
                izprefi=mysite.family.image_namespace(jaz)
                if not t.startswith(izprefi+u':'):
                    return u''
    return t


def main(vhjaz,fvh,fvih,fipref):
    tzfl=0
    tzst={}
    tzno={}


    pref=[]
    if fipref!=u'':
        fpref=codecs.open(fipref,'rb',encoding='utf-8')
        for s in fpref.readlines():
            if ord(s[0]) == 65279:
                s=s[1:]
            s=s.replace(u'\r',u'')
            if s[len(s)-1]==u'\n':
                s=s[:len(s)-1]
            pref.append(s)
        fpref.close()


    n=u''
#    f0=codecs.open(fvh,'rb',encoding='utf-8')
    f1=codecs.open(fvih, 'w', 'utf-8')

    insite=wikipedia.getSite(vhjaz,fam = u'wikipedia')

    tblredir = {}
    # open xml dump and read page titles out of it
    dump = xmlreader.XmlDump(fvh)
    redirR = wikipedia.getSite().redirectRegex()
    readPagesCount = 0

    for entry in dump.parse():
        readPagesCount += 1
        # always print status message after 10000 pages
        if readPagesCount % 10000 == 0:
            print '%i pages read...' % readPagesCount
        m = redirR.search(entry.text)
        if m:
            target = m.group(1)
            # There might be redirects to another wiki. Ignore these.
            for code in wikipedia.getSite().family.langs.keys():
                if target.startswith('%s:' % code) or target.startswith(':%s:' % code):
                    wikipedia.output(u'NOTE: Ignoring %s which is a redirect to %s:' % (entry.title, code))
                    target = None
                    break
            # if the redirect does not link to another wiki
            if target:
                target = target.replace('_', ' ')
                # remove leading and trailing whitespace
                target = target.strip()
                # capitalize the first letter
                if not wikipedia.getSite().nocapitalize:
                    target = target[0].upper() + target[1:]
                if '#' in target:
                    target = target[:target.index('#')]
                if '|' in target:
                    wikipedia.output(u'HINT: %s is a redirect with a pipelink.' % entry.title)  
                    target = target[:target.index('|')]
#                tblredir[entry.title] = target
#                b=u'%s|%s\n' % (entry.title, target)
                tt=entry.title
                tc=target
                tts=tt.split(u' ')
                tcs=tc.split(u' ')
                ltt=len(tts)
                ltc=len(tcs)
                if ltt==ltc and ltt>=2 and tt.lower()==tc.lower():
                    dop=1
                    for i in range(ltt):
                        if tts[i]!=tcs[i]:
                            if len(tts[i])!=len(tcs[i]):
                                dop=0
                            for j in range(len(tts[i])):
                                if (tts[i][j]==tts[i][j].lower() and
                                      tcs[i][j]!=tcs[i][j].lower()):
                                    dop=0
#                        for j in range(len(tcs[i])):
#                            if j>0 and tcs[i][j]!=tcs[i][j].lower():
#                                dop=0
                    b=u'%d|%s|%s\n' % (dop,tt, tc)
                    f1.write(b)
                    f1.flush()

vhjaz = u'eo'
#vhjaz = wikipedia.argHandler(sys.argv[1], 'slov_iz_xml')
fvh = wikipedia.argHandler(sys.argv[1], 'slov_iz_xml')
fvih = wikipedia.argHandler(sys.argv[2], 'slov_iz_xml')

fpref=u''
if len(sys.argv)>=5:
    fpref = wikipedia.argHandler(sys.argv[3], 'slov_iz_xml')


#fotch = wikipedia.argHandler(sys.argv[4], 'slov_iz_xml')

#mysite = wikipedia.getSite()
#otch = codecs.open(fotch, 'w', 'utf-8')
mysite = wikipedia.getSite()

try:
    main(vhjaz,fvh,fvih,fpref)
except:
    wikipedia.stopme()
    raise
else:
    wikipedia.stopme()

</nowiki>

* podg_mma_f1_v2.py

# -*- coding: UTF-8 -*-

__version__='$Id:'

import wikipedia
import re, sys
import codecs
import xmlreader
from bib_tekst import *


def ubrkoment (text):
    # Ignore tekst within nowiki tags and HTML comments
    nowikiOrHtmlCommentR = re.compile(r'<nowiki>.*?|<!--.*?-->', re.IGNORECASE | re.DOTALL)
    match = nowikiOrHtmlCommentR.search(text)
    while match:
        text = text[:match.start()] + text[match.end():]    
        match = nowikiOrHtmlCommentR.search(text)
    return text

def korr_im (jaz,t,pref):

    for p in pref:
        if t.startswith(p):
            t=t[len(p):]
            break

    t=perv_upper(t)

    if u':' in t:
        if ( (not t.startswith(u'Category:')) and 
                     (not t.startswith(u'Image:')) ):
            katprefi=mysite.family.category_namespace(jaz)
            if not t.startswith(katprefi+u':'):
                izprefi=mysite.family.image_namespace(jaz)
                if not t.startswith(izprefi+u':'):
                    return u''
    return t


def main(vhjaz,fvh,fvih,fipref):
    tzfl=0
    tzst={}
    tzno={}

    tblz={}

    pref=[]
    if 1:
        fpref=codecs.open(fipref,'rb',encoding='utf-8')
        for s in fpref.readlines():
            if ord(s[0]) == 65279:
                s=s[1:]
            s=ubr_nk_prob(s)
            tblz[s]=1
        fpref.close()


    n=u''
#    f0=codecs.open(fvh,'rb',encoding='utf-8')
    f1=codecs.open(fvih, 'w', 'utf-8')

    insite=wikipedia.getSite(vhjaz,fam = u'wikipedia')

    tblredir = {}
    # open xml dump and read page titles out of it
    dump = xmlreader.XmlDump(fvh)
    redirR = wikipedia.getSite().redirectRegex()
    readPagesCount = 0

    for entry in dump.parse():
        readPagesCount += 1
        # always print status message after 1000 pages
        if readPagesCount % 1000 == 0:
            print '%i pages read...' % readPagesCount
        m = redirR.search(entry.text)
        if m:
            target = m.group(1)
            # There might be redirects to another wiki. Ignore these.
            for code in wikipedia.getSite().family.langs.keys():
                if target.startswith('%s:' % code) or target.startswith(':%s:' % code):
                    wikipedia.output(u'NOTE: Ignoring %s which is a redirect to %s:' % (entry.title, code))
                    target = None
                    break
            # if the redirect does not link to another wiki
            if target:
                target = target.replace('_', ' ')
                # remove leading and trailing whitespace
                target = target.strip()
                # capitalize the first letter
                if not wikipedia.getSite().nocapitalize:
                    target = target[0].upper() + target[1:]
                if '#' in target:
                    target = target[:target.index('#')]
                if '|' in target:
                    wikipedia.output(u'HINT: %s is a redirect with a pipelink.' % entry.title)  
                    target = target[:target.index('|')]
#                tblredir[entry.title] = target
#                b=u'%s|%s\n' % (entry.title, target)
                tt=entry.title
                tc=target
                ttk=tt.replace(u'-',u' ').replace(u'.',u' ').replace(u'(',u' ')
                tck=tc.replace(u'-',u' ').replace(u'.',u' ').replace(u'(',u' ')
                tts=ttk.split(u' ')
                tcs=tck.split(u' ')
                ltt=len(tts)
                ltc=len(tcs)
                if tblz.has_key(tc) and ltt>=2:
                    dop=1
                    dop2=0
                    for i in range(ltt):
                        if len(tts[i])>=2 and tts[i][1:]!=tts[i][1:].lower():
                            dop2=1
                    for i in range(ltc):
#                        if (  ( len(tcs[i])>2 or 
#                               (len(tcs[i])==2 and tcs[i][1]!=u'.') ) and 
#                                     tcs[i]==tcs[i].upper()  ):
                        if len(tcs[i])>=2 and tcs[i][1:]!=tcs[i][1:].lower():
                            dop=0
                    if dop2==0:
                        dop=0
                    b=u'%d|%s|%s\n' % (dop,tt,tc)
                    f1.write(b)
                    f1.flush()

vhjaz = u'eo'
#vhjaz = wikipedia.argHandler(sys.argv[1], 'slov_iz_xml')
fvh = wikipedia.argHandler(sys.argv[1], 'slov_iz_xml')
fvih = wikipedia.argHandler(sys.argv[2], 'slov_iz_xml')
fpref = wikipedia.argHandler(sys.argv[3], 'slov_iz_xml')


#fotch = wikipedia.argHandler(sys.argv[4], 'slov_iz_xml')

#mysite = wikipedia.getSite()
#otch = codecs.open(fotch, 'w', 'utf-8')
mysite = wikipedia.getSite()

try:
    main(vhjaz,fvh,fvih,fpref)
except:
    wikipedia.stopme()
    raise
else:
    wikipedia.stopme()

</nowiki>

* podg_mma_f2.py

# -*- coding: UTF-8 -*-

__version__='$Id:'

import wikipedia
import re, sys
import codecs
import xmlreader
from bib_tekst import *


def ubrkoment (text):
    # Ignore tekst within nowiki tags and HTML comments
    nowikiOrHtmlCommentR = re.compile(r'<nowiki>.*?|<!--.*?-->', re.IGNORECASE | re.DOTALL)
    match = nowikiOrHtmlCommentR.search(text)
    while match:
        text = text[:match.start()] + text[match.end():]    
        match = nowikiOrHtmlCommentR.search(text)
    return text

def korr_im (jaz,t,pref):

    for p in pref:
        if t.startswith(p):
            t=t[len(p):]
            break

    t=perv_upper(t)

    if u':' in t:
        if ( (not t.startswith(u'Category:')) and 
                     (not t.startswith(u'Image:')) ):
            katprefi=mysite.family.category_namespace(jaz)
            if not t.startswith(katprefi+u':'):
                izprefi=mysite.family.image_namespace(jaz)
                if not t.startswith(izprefi+u':'):
                    return u''
    return t


def main(vhjaz,fvh,fvih,fipref):
    tzfl=0
    tzst={}
    tzno={}

    tblz={}

    pref=[]
    if 1:
        fpref=codecs.open(fipref,'rb',encoding='utf-8')
        for s in fpref.readlines():
            if ord(s[0]) == 65279:
                s=s[1:]
            s=s.replace(u'\r',u'')
            if s[len(s)-1]==u'\n':
                s=s[:len(s)-1]
            st=s.split(u'|')
            if len(st)==3 and st[0]==u'1':
                tblz[st[1]]=st[2]
        fpref.close()


    n=u''
#    f0=codecs.open(fvh,'rb',encoding='utf-8')
    f1=codecs.open(fvih, 'w', 'utf-8')

    insite=wikipedia.getSite(vhjaz,fam = u'wikipedia')

    tblredir = {}
    # open xml dump and read page titles out of it
    dump = xmlreader.XmlDump(fvh)
    redirR = wikipedia.getSite().redirectRegex()
    readPagesCount = 0

    sch_str=0
    sch_zam=0

    for entry in dump.parse():
        readPagesCount += 1
        # always print status message after 1000 pages
        if readPagesCount % 1000 == 0:
            print '%i pages read...' % readPagesCount
        m = redirR.search(entry.text)
        if m:
            pass
        else:
            if entry.title.startswith(u'Vikipedio:'):
                continue
            fperv=1
            for tt, tc in tblz.iteritems():
                if ( (not entry.title.startswith(tt)) and 
                     (not entry.title.startswith(tc)) and (tt in entry.text) ):

                    if fperv:
                        b=u'|malmajuskligo\n%s\n'%entry.title
                        f1.write(b)
                        f1.flush()
                        sch_str+=1
                    b=((u'?=======\n1l\n-=======\n%s\n+=======\n%s\n')%(tt,tc))
                    f1.write(b)
                    f1.flush()
                    fperv=0
                    sch_zam+=1

            if fperv==0:
                f1.write(u'========================\n')
                f1.flush()

    wikipedia.output(u'sch_str=%d sch_zam=%d'%(sch_str,sch_zam))


vhjaz = u'eo'
#vhjaz = wikipedia.argHandler(sys.argv[1], 'slov_iz_xml')
fvh = wikipedia.argHandler(sys.argv[1], 'slov_iz_xml')
fvih = wikipedia.argHandler(sys.argv[2], 'slov_iz_xml')

fpref = wikipedia.argHandler(sys.argv[3], 'slov_iz_xml')

#fotch = wikipedia.argHandler(sys.argv[4], 'slov_iz_xml')

#mysite = wikipedia.getSite()
#otch = codecs.open(fotch, 'w', 'utf-8')
mysite = wikipedia.getSite()

try:
    main(vhjaz,fvh,fvih,fpref)
except:
    wikipedia.stopme()
    raise
else:
    wikipedia.stopme()

</nowiki>

* svoj-ka.py

# -*- coding: UTF-8 -*-

__version__='$Id:'

import wikipedia, pagegenerators
import re, sys
import codecs
import perevod,imagetransfer1

def iskat(t,s):
    lt=len(t)
    ls=len(s)
    i=0
    while i<=lt-ls:
        if t[i:i+ls]==s:
            return i
        i+=1
    return -1

def iskats(t,i,s):
    lt=len(t)
    ls=len(s)
    while i<=lt-ls:
        if t[i:i+ls]==s:
            return i
        i+=1
    return -1

def iskato(t,i,s):
    ls=len(s)
    while i>=0:
        if t[i:i+ls]==s:
            return i
        i-=1
    return -1

def vivod(b):
    wikipedia.output(b)
    otch.write(b)
    otch.flush()

def ubr_nk_prob(t):
    if len(t)<1:
        return t
    while (t[0]==u' ') or (t[0]==u'\n') or (t[0]==u'\t') or (t[0]==u'\r'):
        t=t[1:]
        if len(t)<1:
            return t
    while (t[len(t)-1]==u' ') or (t[len(t)-1]==u'\n') or (t[len(t)-1]==u'\t') or (t[len(t)-1]==u'\r'):
        t=t[:len(t)-1]
        if len(t)<1:
            return t
    return t

    otch.write(u'\n\n\n')
    otch.flush()
    zapis_fimen(fimen,nomerf,pl)
    return


def main():

    mysite = wikipedia.getSite()
#    sico=wikipedia.getSite(code = u'commons',fam = u'commons')

    global otch
    otch = codecs.open(u'%s' % filename, 'w', 'utf-8')

#    for page in mysite.allpages(start = u'', namespace = 6):
#        vivod(page.title()+u'\n')
    for page in mysite.allpages(start = u'', namespace = 14):
        vivod(page.title()+u'\n')

filename = wikipedia.argHandler(sys.argv[1], 'cht_cat')

for arg in sys.argv[2:]:
    arg = wikipedia.argHandler(arg, 'ivsen')
    if arg:
        if arg=="-log":
            import logger
            sys.stdout = logger.Logger(sys.stdout, filename = 'com-izobr.log')

try:
    main()
except:
    wikipedia.stopme()
    raise
else:
    wikipedia.stopme()

* isk_ssilk.py

# -*- coding: UTF-8 -*-

__version__='$Id:'

import wikipedia
import re, sys
import codecs
from bib_tekst import *

def vivod(b):
    wikipedia.output(b)
    otch.write(b)
    otch.flush()

def isk_kateg(n):
    pl = wikipedia.Page(mysite,n)
    wikipedia.output(pl.title())

    try:
        ot = pl.get()
    except wikipedia.NoPage:
        b = u'- %s\n\n' % pl.title()
        vivod(b)
        return
    except wikipedia.IsRedirectPage:
        b = u'+ %s\n-> %s\n\n' % (pl.title(),pl.getRedirectTarget())
        vivod(b)
        return

    otch.write(u'+ %s\n' % pl.title())
    otch.flush()

    otch2.write(u'=========================\n%s\n-=======\n%s\n' % (pl.title(),ot))
    otch2.flush()

    katprefi=mysite.family.category_namespace(mysite.lang)

    lip=pl.linkedPages()
    for pl2 in lip:
        t=pl2.title()
#        if t.startwith(katprefi+u':') or t.startwith(u':'+katprefi+u':'):
        frez.write(t+u'\n')

def main():
    n=[]
    f=codecs.open(filename,'rb',encoding='utf-8')

    i=0
    j=0
    for s in f.readlines():
        if ord(s[0]) == 65279:
            s=s[1:]
        s=s.replace(u'\r',u'')
        if s[len(s)-1]==u'\n':
            s=s[:len(s)-1]
        isk_kateg(s)


fi = wikipedia.argHandler(sys.argv[1], 'cht_cat')
filename = fi+'.txt'
fotch = fi+'_ot.txt'
fotch2 = fi+'_op.txt'
firez = fi+'_sk.txt'

#filename = wikipedia.argHandler(sys.argv[1], 'cht_cat')
#fotch = wikipedia.argHandler(sys.argv[2], 'cht_cat')
#fotch2 = wikipedia.argHandler(sys.argv[3], 'cht_cat')

fliwi=1
if len(sys.argv)>=3 and sys.argv[2]==u'w':
    fliwi=0

otch = codecs.open(fotch, 'w', 'utf-8')
otch2 = codecs.open(fotch2, 'w', 'utf-8')
frez = codecs.open(firez, 'w', 'utf-8')
mysite = wikipedia.getSite()

try:
    main()
except:
    wikipedia.stopme()
    raise
else:
    wikipedia.stopme()