aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rw-r--r--content/pages/buscar.php.md12
-rw-r--r--plugins/tipue-search/tipue_search.py107
2 files changed, 80 insertions, 39 deletions
diff --git a/content/pages/buscar.php.md b/content/pages/buscar.php.md
index accc6d8..b06a021 100644
--- a/content/pages/buscar.php.md
+++ b/content/pages/buscar.php.md
@@ -62,9 +62,9 @@ if (isset($_GET['q'])) {
$keywords = explode(' ', $keywords);
$found_results = [];
- foreach ($web_content["pages"] as $page) {
+ foreach ($web_content["videos"] as $page) {
$score = 0;
- $page['text'] = htmlentities($page['text']);
+ $page['description'] = htmlentities($page['description']);
foreach ($keywords as $word) {
if (preg_match("/$word/i", $page['url'])) {
@@ -78,16 +78,16 @@ if (isset($_GET['q'])) {
}
// It replaces uppercase matches with lowercase matches, but it's fine for now.
if ($stop_words_ignored == 1) {
- $page['text'] = preg_replace("/$word/i", $word, $page['text'], -1, $match_count);
+ $page['description'] = preg_replace("/$word/i", $word, $page['description'], -1, $match_count);
} else {
- $page['text'] = preg_replace("/$word/i", '<span class="tipue_search_content_bold highlighted">' . $word . '</span>', $page['text'], -1, $match_count);
+ $page['description'] = preg_replace("/$word/i", '<span class="tipue_search_content_bold highlighted">' . $word . '</span>', $page['description'], -1, $match_count);
}
if ($match_count > 0) {
$score += 10 * $match_count;
}
}
if ($score != 0) {
- $found_results[] = ['score' => $score, 'title' => $page['title'], 'url' => $page['url'], 'text' => $page['text']];
+ $found_results[] = ['score' => $score, 'title' => $page['title'], 'url' => $page['url'], 'description' => $page['description']];
}
}
@@ -123,7 +123,7 @@ if (isset($_GET['q'])) {
printf('<div class="tipue_search_content_title"><a class="is-cyan" href="%s">%s</a></div>', $found_result['url'], $found_result['title']);
printf('<div class="tipue_search_content_url"><a href="%s">%s</a></div>', $found_result['url'], $found_result['url']);
- $description_words = explode(' ', $found_result['text']);
+ $description_words = explode(' ', $found_result['description']);
$description_words_count = count($description_words);
$first_match = false;
for ($i = 0; $i < $description_words_count; $i++) {
diff --git a/plugins/tipue-search/tipue_search.py b/plugins/tipue-search/tipue_search.py
index 2880850..8a62a68 100644
--- a/plugins/tipue-search/tipue_search.py
+++ b/plugins/tipue-search/tipue_search.py
@@ -32,66 +32,107 @@ class Tipue_Search_JSON_Generator(object):
self.siteurl = settings.get('SITEURL')
self.relative_urls = settings.get('RELATIVE_URLS')
self.tpages = settings.get('TEMPLATE_PAGES')
+ self.tstatic = settings.get('THEME_STATIC_DIR')
self.output_path = output_path
self.json_nodes = []
- def create_json_node(self, page):
+ def create_json_node(self, article):
- if getattr(page, 'status', 'published') != 'published':
+ if getattr(article, 'status', 'published') != 'published':
return
- soup_title = BeautifulSoup(page.title.replace('&nbsp;', ' '), 'html.parser')
- page_title = soup_title.get_text(' ', strip=True).replace('“', '"').replace('”', '"').replace('’', "'").replace('^', '&#94;')
-
- soup_text = BeautifulSoup(page.content, 'html.parser')
- page_text = soup_text.get_text(' ', strip=True).replace('“', '"').replace('”', '"').replace('’', "'").replace('¶', ' ').replace('^', '&#94;')
- page_text = ' '.join(page_text.split())
-
- page_category = page.category.name if getattr(page, 'category', 'None') != 'None' else ''
-
- page_url = '.'
- if page.url:
- page_url = page.url if self.relative_urls else (self.siteurl + '/' + page.url)
-
- node = {'title': page_title,
- 'text': page_text,
- 'tags': page_category,
- 'url': page_url}
+ soup_title = BeautifulSoup(article.title.replace('&nbsp;', ' '), 'html.parser')
+ video_title = soup_title.get_text(' ', strip=True).replace('“', '"').replace('”', '"').replace('’', "'").replace('^', '&#94;')
+
+ soup_text = BeautifulSoup(article.content, 'html.parser')
+ video_text = soup_text.get_text(' ', strip=True).replace('“', '"').replace('”', '"').replace('’', "'").replace('¶', ' ').replace('^', '&#94;')
+ video_text = ' '.join(video_text.split())
+
+ if self.relative_urls:
+ image_url = '.'
+ else:
+ image_url = self.siteurl
+
+ # thumbnail
+ video_image = article.image if getattr(
+ article, 'image', 'None') != 'None' else ''
+
+ url_image = "%s/%s/../wp-content/uploads/article/poster/%s" % (
+ image_url, self.tstatic, video_image
+ )
+
+ # publish
+ video_publish = article.date.strftime("%a, %d %B, %Y") if getattr(
+ article, 'date', 'None') != 'None' else ''
+
+ # author
+ video_author = str(article.author) if getattr(
+ article, 'author', 'None') != 'None' else ''
+
+ # time
+ video_time = article.time if getattr(
+ article, 'time', 'None') != 'None' else ''
+
+ video_url = '.'
+ if article.url:
+ video_url = article.url if self.relative_urls else (
+ self.siteurl + '/' + article.url)
+
+ video_src = article.og_video if getattr(
+ article, 'og_video', 'None') != 'None' else ''
+
+ video_category = article.category.name if getattr(
+ article, 'category', 'None') != 'None' else ''
+
+ node = {'title': video_title,
+ 'description': video_text,
+ 'videoThumbnail': url_image,
+ 'formatStreams': {
+ 'url': video_src,
+ },
+ 'author': video_author,
+ 'publishedText': video_publish,
+ 'time': video_time,
+ 'tags': video_category,
+ 'url': video_url}
self.json_nodes.append(node)
def create_tpage_node(self, srclink):
- srcfile = open(os.path.join(self.output_path, self.tpages[srclink]), encoding='utf-8')
+ srcfile = open(os.path.join(self.output_path,
+ self.tpages[srclink]),
+ encoding='utf-8')
soup = BeautifulSoup(srcfile, 'html.parser')
- page_title = soup.title.string if soup.title is not None else ''
- page_text = soup.get_text()
+ video_title = soup.title.string if soup.title is not None else ''
+ video_text = soup.get_text()
# Should set default category?
- page_category = ''
- page_url = urljoin(self.siteurl, self.tpages[srclink])
+ video_category = ''
+ video_url = urljoin(self.siteurl, self.tpages[srclink])
- node = {'title': page_title,
- 'text': page_text,
- 'tags': page_category,
- 'url': page_url}
+ node = {'title': video_title,
+ 'text': video_text,
+ 'tags': video_category,
+ 'url': video_url}
self.json_nodes.append(node)
def generate_output(self, writer):
path = os.path.join(self.output_path, 'tipuesearch_content.json')
- pages = self.context['pages'] + self.context['articles']
+ articles = self.context['articles']
for article in self.context['articles']:
- pages += article.translations
+ articles += article.translations
for srclink in self.tpages:
self.create_tpage_node(srclink)
- for page in pages:
- self.create_json_node(page)
- root_node = {'pages': self.json_nodes}
+ for article in articles:
+ self.create_json_node(article)
+
+ root_node = {'videos': self.json_nodes}
with open(path, 'w', encoding='utf-8') as fd:
json.dump(root_node, fd, separators=(',', ':'), ensure_ascii=False)