Files
aikaterna-cogs/dictionary/dictionary.py
aikaterna 2c086e7dff [Dictionary] Change up some things in the search
Let's see how long this one works
2021-04-07 16:02:47 -07:00

151 lines
5.5 KiB
Python

import aiohttp
from bs4 import BeautifulSoup
import json
import logging
import re
from redbot.core import commands
from redbot.core.utils.chat_formatting import pagify
log = logging.getLogger("red.aikaterna.dictionary")
class Dictionary(commands.Cog):
"""
Word, yo
Parts of this cog are adapted from the PyDictionary library.
"""
async def red_delete_data_for_user(self, **kwargs):
"""Nothing to delete"""
return
def __init__(self, bot):
self.bot = bot
self.session = aiohttp.ClientSession()
def cog_unload(self):
self.bot.loop.create_task(self.session.close())
@commands.command()
async def define(self, ctx, *, word: str):
"""Displays definitions of a given word."""
search_msg = await ctx.send("Searching...")
search_term = word.split(" ", 1)[0]
result = await self._definition(ctx, search_term)
str_buffer = ""
if not result:
await search_msg.delete()
await ctx.send("This word is not in the dictionary.")
return
for key in result:
str_buffer += f"\n**{key}**: \n"
counter = 1
j = False
for val in result[key]:
if val.startswith("("):
str_buffer += f"{str(counter)}. *{val})* "
counter += 1
j = True
else:
if j:
str_buffer += f"{val}\n"
j = False
else:
str_buffer += f"{str(counter)}. {val}\n"
counter += 1
await search_msg.delete()
for page in pagify(str_buffer, delims=["\n"]):
await ctx.send(page)
async def _definition(self, ctx, word):
data = await self._get_soup_object(f"http://wordnetweb.princeton.edu/perl/webwn?s={word}")
if not data:
return await ctx.send("Error fetching data.")
types = data.findAll("h3")
length = len(types)
lists = data.findAll("ul")
out = {}
if not lists:
return
for a in types:
reg = str(lists[types.index(a)])
meanings = []
for x in re.findall(r">\s\((.*?)\)\s<", reg):
if "often followed by" in x:
pass
elif len(x) > 5 or " " in str(x):
meanings.append(x)
name = a.text
out[name] = meanings
return out
@commands.command()
async def antonym(self, ctx, *, word: str):
"""Displays antonyms for a given word."""
search_term = word.split(" ", 1)[0]
result = await self._antonym_or_synonym(ctx, "antonyms", search_term)
if not result:
await ctx.send("This word is not in the dictionary or nothing was found.")
return
result_text = "*, *".join(result)
msg = f"Antonyms for **{search_term}**: *{result_text}*"
for page in pagify(msg, delims=["\n"]):
await ctx.send(page)
@commands.command()
async def synonym(self, ctx, *, word: str):
"""Displays synonyms for a given word."""
search_term = word.split(" ", 1)[0]
result = await self._antonym_or_synonym(ctx, "synonyms", search_term)
if not result:
await ctx.send("This word is not in the dictionary or nothing was found.")
return
result_text = "*, *".join(result)
msg = f"Synonyms for **{search_term}**: *{result_text}*"
for page in pagify(msg, delims=["\n"]):
await ctx.send(page)
async def _antonym_or_synonym(self, ctx, lookup_type, word):
if lookup_type not in ["antonyms", "synonyms"]:
return None
data = await self._get_soup_object(f"http://www.thesaurus.com/browse/{word}")
if not data:
await ctx.send("Error getting information from the website.")
return
website_data = None
script = data.find_all("script")
for item in script:
if item.string:
if "window.INITIAL_STATE" in item.string:
content = item.string
content = content.lstrip("window.INITIAL_STATE =").rstrip(";")
content = content.replace("undefined", '"None"').replace("true", '"True"').replace("false", '"False"')
try:
website_data = json.loads(content)
except json.decoder.JSONDecodeError:
return None
except Exception as e:
log.exception(e, exc_info=e)
await ctx.send("Something broke. Check your console for more information.")
return None
final = []
if website_data:
syn_list = website_data["searchData"]["tunaApiData"]["posTabs"][0][lookup_type]
for syn in syn_list:
final.append(syn["term"])
return final
async def _get_soup_object(self, url):
try:
async with self.session.request("GET", url) as response:
return BeautifulSoup(await response.text(), "html.parser")
except Exception:
log.error("Error fetching dictionary.py related webpage", exc_info=True)
return None