Compare commits
No commits in common. "7ae80ab600a012a3d874ac00b99c6dcdda74e70a" and "8c242d1ff934fa6f3b6f85fe4d3a8d6754d6e063" have entirely different histories.
7ae80ab600
...
8c242d1ff9
@ -7,5 +7,3 @@ aiohttp==3.9.1
|
|||||||
feedparser==6.0.10
|
feedparser==6.0.10
|
||||||
websockets==12.0
|
websockets==12.0
|
||||||
trafilatura==1.6.2
|
trafilatura==1.6.2
|
||||||
vaderSentiment
|
|
||||||
|
|
||||||
|
|||||||
@ -4,7 +4,6 @@ from fastapi.templating import Jinja2Templates
|
|||||||
import dataset
|
import dataset
|
||||||
import json
|
import json
|
||||||
import aiohttp
|
import aiohttp
|
||||||
import sentiment
|
|
||||||
import feedparser
|
import feedparser
|
||||||
import asyncio
|
import asyncio
|
||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
@ -357,17 +356,15 @@ async def websocket_sync(websocket: WebSocket):
|
|||||||
'last_synchronized': datetime.now().isoformat()
|
'last_synchronized': datetime.now().isoformat()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
existing = articles_table.find_one(guid=article_data['guid'])
|
existing = articles_table.find_one(guid=article_data['guid'])
|
||||||
if not existing:
|
if not existing:
|
||||||
new_articles.append(article_data)
|
new_articles.append(article_data)
|
||||||
articles_count += 1
|
articles_count += 1
|
||||||
article_data['sentiment'] = json.dumps(sentiment.analyze(entry.get('description', '') or entry.get('summary', '')))
|
|
||||||
articles_table.upsert(article_data, ['guid'])
|
articles_table.upsert(article_data, ['guid'])
|
||||||
|
|
||||||
# Index the article to ChromaDB
|
# Index the article to ChromaDB
|
||||||
doc_content = f"{article_data.get('title', '')}\n{article_data.get('description', '')}"
|
doc_content = f"{article_data.get('title', '')}\n{article_data.get('description', '')}"
|
||||||
|
|
||||||
metadata = {key: str(value) for key, value in article_data.items() if key != 'content'} # Exclude large content from metadata
|
metadata = {key: str(value) for key, value in article_data.items() if key != 'content'} # Exclude large content from metadata
|
||||||
chroma_collection.upsert(
|
chroma_collection.upsert(
|
||||||
documents=[doc_content],
|
documents=[doc_content],
|
||||||
@ -493,7 +490,6 @@ async def search_articles(
|
|||||||
for i, doc_id in enumerate(results['ids'][0]):
|
for i, doc_id in enumerate(results['ids'][0]):
|
||||||
res = results['metadatas'][0][i]
|
res = results['metadatas'][0][i]
|
||||||
res['distance'] = results['distances'][0][i]
|
res['distance'] = results['distances'][0][i]
|
||||||
res['sentiment'] = sentiment.analyze(res.get('description', '') or res.get('content', '') or res.get('title', ''))
|
|
||||||
formatted_results.append(res)
|
formatted_results.append(res)
|
||||||
|
|
||||||
return JSONResponse(content={"results": formatted_results})
|
return JSONResponse(content={"results": formatted_results})
|
||||||
@ -569,8 +565,6 @@ async def newspaper_latest(request: Request):
|
|||||||
for article in articles:
|
for article in articles:
|
||||||
for key, value in article.items():
|
for key, value in article.items():
|
||||||
article[key] = str(value).strip().replace(' ', '')
|
article[key] = str(value).strip().replace(' ', '')
|
||||||
|
|
||||||
article['sentiment'] = sentiment.analyze(article.get('description', '') or article.get('content', '') or res.get('title', ''))
|
|
||||||
return templates.TemplateResponse("newspaper_view.html", {
|
return templates.TemplateResponse("newspaper_view.html", {
|
||||||
"request": request,
|
"request": request,
|
||||||
"newspaper": first_newspaper,
|
"newspaper": first_newspaper,
|
||||||
|
|||||||
35
sentiment.py
35
sentiment.py
@ -1,35 +0,0 @@
|
|||||||
import json
|
|
||||||
from vaderSentiment.vaderSentiment import SentimentIntensityAnalyzer
|
|
||||||
|
|
||||||
def analyze_sentiment_vader(text, analyzer):
|
|
||||||
"""
|
|
||||||
Analyzes text using VADER and returns a dictionary with the results.
|
|
||||||
|
|
||||||
Args:
|
|
||||||
text (str): The text content to analyze.
|
|
||||||
analyzer (SentimentIntensityAnalyzer): An instantiated VADER analyzer.
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
dict: A dictionary containing the sentiment classification, compound score,
|
|
||||||
and detailed scores (positive, neutral, negative).
|
|
||||||
"""
|
|
||||||
scores = analyzer.polarity_scores(text)
|
|
||||||
compound_score = scores['compound']
|
|
||||||
|
|
||||||
if compound_score >= 0.05:
|
|
||||||
sentiment = 'Positive'
|
|
||||||
elif compound_score <= -0.05:
|
|
||||||
sentiment = 'Negative'
|
|
||||||
else:
|
|
||||||
sentiment = 'Neutral'
|
|
||||||
|
|
||||||
return {
|
|
||||||
'sentiment': sentiment,
|
|
||||||
'score': compound_score,
|
|
||||||
'details': scores
|
|
||||||
}
|
|
||||||
|
|
||||||
vader_analyzer = SentimentIntensityAnalyzer()
|
|
||||||
|
|
||||||
def analyze(content):
|
|
||||||
return analyze_sentiment_vader(content, vader_analyzer)
|
|
||||||
@ -164,7 +164,6 @@
|
|||||||
<h2 class="article-title">
|
<h2 class="article-title">
|
||||||
<a href="{{ article.link }}" target="_blank">{{ article.title }}</a>
|
<a href="{{ article.link }}" target="_blank">{{ article.title }}</a>
|
||||||
</h2>
|
</h2>
|
||||||
<input type="hidden" name="sentiment" value="{{ article.sentiment }}">
|
|
||||||
<div class="article-meta">
|
<div class="article-meta">
|
||||||
<span class="article-source">{{ article.feed_name }}</span>
|
<span class="article-source">{{ article.feed_name }}</span>
|
||||||
{% if article.author %}
|
{% if article.author %}
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user