如何解决Tesseract ocr Google search+Wiki search indexError
我不是 Python 编码员。这就是为什么我需要帮助。任何人都可以帮我解决这个错误。当我运行此脚本时,它开始搜索 Google + wiki 搜索,然后显示 indexError 列表索引超出范围。当我在命令框中运行代码时,它使用 tesseract-ocr 并将图像转换为文本并搜索答案。正在运行,但是当搜索 google+ wiki 搜索时出错
'''
Todo:
* Implement normalize func
* Attempt to google wiki \"...\" part of question
* Rid of common appearances in 3 options
* Automate screenshot process
* Implement Asynchio for concurrency
//Script is in working condition at all times
//Todo is for improving accuracy
'''
# answering bot
import json
import urllib.request as urllib2
from bs4 import BeautifulSoup
from googleapi import google
from PIL import Image
import PyTesseract
import argparse
import cv2
import os
import pyscreenshot as Imagegrab
import sys
import wx
from halo import Halo
# for terminal colors
class bcolors:
HEADER = '\033[95m'
OKBLUE = '\033[94m'
OKGREEN = '\033[92m'
WARNING = '\033[93m'
FAIL = '\033[91m'
ENDC = '\033[0m'
BOLD = '\033[1m'
UNDERLINE = '\033[4m'
# sample questions from prevIoUs games
sample_questions = {}
# list of words to clean from the question during google search
remove_words = []
# negative words
negative_words= []
# GUI interface
def gui_interface():
app = wx.App()
frame = wx.Frame(None,-1,'win.py')
frame.SetDimensions(0,640,480)
frame.Show()
app.MainLoop()
return None
# load sample questions
def load_json():
global remove_words,sample_questions,negative_words
remove_words = json.loads(open("Data/settings.json").read())["remove_words"]
negative_words = json.loads(open("Data/settings.json").read())["negative_words"]
sample_questions = json.loads(open("Data/questions.json",encoding="utf8").read())
# take screenshot of question
def screen_grab(to_save):
# 31,228 485,620 co-ords of screenshot// left side of screen
im = Imagegrab.grab(bBox=(31,228,485,640))
im.save(to_save)
# get OCR text //questions and options
def read_screen():
spinner = Halo(text='Reading screen',spinner='bouncingBar')
spinner.start()
screenshot_file="Screens/to_ocr.png"
screen_grab(screenshot_file)
#prepare argparse
ap = argparse.ArgumentParser(description='HQ_Bot')
ap.add_argument("-i","--image",required=False,default=screenshot_file,help="path to input image to be OCR'd")
ap.add_argument("-p","--preprocess",type=str,default="thresh",help="type of preprocessing to be done")
args = vars(ap.parse_args())
# load the image
image = cv2.imread(args["image"])
gray = cv2.cvtColor(image,cv2.COLOR_BGR2GRAY)
if args["preprocess"] == "thresh":
gray = cv2.threshold(gray,255,cv2.THRESH_BINARY | cv2.THRESH_OTSU)[1]
elif args["preprocess"] == "blur":
gray = cv2.medianBlur(gray,3)
# store grayscale image as a temp file to apply OCR
filename = "Screens/{}.png".format(os.getpid())
cv2.imwrite(filename,gray)
# load the image as a PIL/Pillow image,apply OCR,and then delete the temporary file
text = PyTesseract.image_to_string(Image.open(filename))
os.remove(filename)
os.remove(screenshot_file)
# show the output images
'''cv2.imshow("Image",image)
cv2.imshow("Output",gray)
os.remove(screenshot_file)
if cv2.waitKey(0):
cv2.destroyAllWindows()
print(text)
'''
spinner.succeed()
spinner.stop()
return text
# get questions and options from OCR text
def parse_question():
text = read_screen()
lines = text.splitlines()
question = ""
options = list()
flag=False
for line in lines :
if not flag :
question=question+" "+line
if '?' in line :
flag=True
continue
if flag :
if line != '' :
options.append(line)
return question,options
# simplify question and remove which,what....etc //question is string
def simplify_ques(question):
neg=False
qwords = question.lower().split()
if [i for i in qwords if i in negative_words]:
neg=True
cleanwords = [word for word in qwords if word.lower() not in remove_words]
temp = ' '.join(cleanwords)
clean_question=""
#remove ?
for ch in temp:
if ch!="?" or ch!="\"" or ch!="\'":
clean_question=clean_question+ch
return clean_question.lower(),neg
# get web page
def get_page(link):
try:
if link.find('mailto') != -1:
return ''
req = urllib2.Request(link,headers={'User-Agent': 'Mozilla/5.0 (Windows NT 6.1; Win64; x64)'})
html = urllib2.urlopen(req).read()
return html
except (urllib2.URLError,urllib2.HTTPError,ValueError) as e:
return ''
# split the string
def split_string(source):
splitlist = ",!-.;/?@ #"
output = []
atsplit = True
for char in source:
if char in splitlist:
atsplit = True
else:
if atsplit:
output.append(char)
atsplit = False
else:
output[-1] = output[-1] + char
return output
# normalize points // get rid of common appearances // "quote" wiki option + ques
def normalize():
return None
# take screen shot of screen every 2 seconds and check for question
def check_screen():
return None
# wait for certain milli seconds
def wait(msec):
return None
# answer by combining two words
def smart_answer(content,qwords):
zipped= zip(qwords,qwords[1:])
points=0
for el in zipped :
if content.count(el[0]+" "+el[1])!=0 :
points+=1000
return points
# use google to get wiki page
def google_wiki(sim_ques,options,neg):
spinner = Halo(text='Googling and searching Wikipedia',spinner='dots2')
spinner.start()
num_pages = 1
points = list()
content = ""
maxo=""
maxp=-sys.maxsize
words = split_string(sim_ques)
for o in options:
o = o.lower()
original=o
o += ' wiki'
# get google search results for option + 'wiki'
search_wiki = google.search(o,num_pages)
link = search_wiki[0].link
content = get_page(link)
soup = BeautifulSoup(content,"lxml")
page = soup.get_text().lower()
temp=0
for word in words:
temp = temp + page.count(word)
temp+=smart_answer(page,words)
if neg:
temP*=-1
points.append(temp)
if temp>maxp:
maxp=temp
maxo=original
spinner.succeed()
spinner.stop()
return points,maxo
# return points for sample_questions
def get_points_sample():
simq = ""
x = 0
for key in sample_questions:
x = x + 1
points = []
simq,neg = simplify_ques(key)
options = sample_questions[key]
simq = simq.lower()
maxo=""
points,maxo = google_wiki(simq,neg)
print("\n" + str(x) + ". " + bcolors.UNDERLINE + key + bcolors.ENDC + "\n")
for point,option in zip(points,options):
if maxo == option.lower():
option=bcolors.OKGREEN+option+bcolors.ENDC
print(option + " { points: " + bcolors.BOLD + str(point) + bcolors.ENDC + " }\n")
# return points for live game // by screenshot
def get_points_live():
neg= False
question,options=parse_question()
simq = ""
points = []
simq,neg = simplify_ques(question)
maxo=""
m=1
if neg:
m=-1
points,neg)
print("\n" + bcolors.UNDERLINE + question + bcolors.ENDC + "\n")
for point,options):
if maxo == option.lower():
option=bcolors.OKGREEN+option+bcolors.ENDC
print(option + " { points: " + bcolors.BOLD + str(point*m) + bcolors.ENDC + " }\n")
# menu// main func
if __name__ == "__main__":
load_json()
while(1):
keypressed = input(bcolors.WARNING +'\nPress s to screenshot live game,sampq to run against sample questions or q to quit:\n' + bcolors.ENDC)
if keypressed == 's':
get_points_live()
elif keypressed == 'sampq':
get_points_sample()
elif keypressed == 'q':
break
else:
print(bcolors.FAIL + "\nUnkNown input" + bcolors.ENDC)
版权声明:本文内容由互联网用户自发贡献,该文观点与技术仅代表作者本人。本站仅提供信息存储空间服务,不拥有所有权,不承担相关法律责任。如发现本站有涉嫌侵权/违法违规的内容, 请发送邮件至 dio@foxmail.com 举报,一经查实,本站将立刻删除。