data = [] for source in sources: response = requests.get(source) soup = BeautifulSoup(response.content, 'html.parser') # Extract relevant data data.append({ "title": soup.find("title").text, "description": soup.find("description").text })
import unittest from data_collector import collect_data from indexing_engine import create_index, update_index index of megamind updated
def create_index(): es = Elasticsearch() es.indices.create(index="megamind-index", body={ "mappings": { "properties": { "title": {"type": "text"}, "description": {"type": "text"} } } }) data = [] for source in sources: response = requests