restructure and implement feed.ParseContent
This commit is contained in:
parent
794e1ef668
commit
0b92829707
|
@ -1,11 +1,12 @@
|
|||
package main
|
||||
|
||||
import (
|
||||
"fmt"
|
||||
"html"
|
||||
"net/http"
|
||||
"strings"
|
||||
|
||||
"git.32bit.cafe/yequari/rss-gen/rss"
|
||||
"git.32bit.cafe/yequari/rss-gen/feed"
|
||||
)
|
||||
|
||||
func (app *application) home(w http.ResponseWriter, r *http.Request) {
|
||||
|
@ -28,8 +29,9 @@ func (app *application) generateRss(w http.ResponseWriter, r *http.Request) {
|
|||
pages[i] = strings.TrimSpace(pages[i])
|
||||
}
|
||||
|
||||
feed, err := rss.GenerateRss(siteUrl, siteName, siteDesc, pages...)
|
||||
feed, err := feed.GenerateRss(siteUrl, siteName, siteDesc, pages...)
|
||||
if err != nil {
|
||||
w.Write([]byte(fmt.Sprintf("<p class=\"error\">Error generating feed: %s</p>", err.Error())))
|
||||
app.infoLog.Printf("Error generating feed: %s\n", err.Error())
|
||||
}
|
||||
for _, line := range strings.Split(feed, "\n") {
|
||||
|
|
|
@ -1,4 +1,5 @@
|
|||
package rss
|
||||
package feed
|
||||
|
||||
var ParseArticle = parseArticle
|
||||
var FetchPage = fetchPage
|
||||
var ParseTime = parseTime
|
|
@ -0,0 +1,169 @@
|
|||
package feed
|
||||
|
||||
import (
|
||||
"fmt"
|
||||
"io"
|
||||
"net/http"
|
||||
"strings"
|
||||
"time"
|
||||
|
||||
"golang.org/x/net/html"
|
||||
)
|
||||
|
||||
type FeedBuilder interface {
|
||||
GenerateFeed() string
|
||||
}
|
||||
|
||||
type FeedInfo struct {
|
||||
SiteName string
|
||||
SiteUrl string
|
||||
SiteDesc string
|
||||
PageUrls []string
|
||||
errors map[string]error
|
||||
}
|
||||
|
||||
type FeedItem struct {
|
||||
Url string
|
||||
Title string
|
||||
Author string
|
||||
EscapedText string
|
||||
PubTime time.Time
|
||||
RawText string
|
||||
}
|
||||
|
||||
func fetchPage(url string) (string, error) {
|
||||
resp, err := http.Get(url)
|
||||
if err != nil {
|
||||
return "", fmt.Errorf("Error sending Get request: %w", err)
|
||||
}
|
||||
defer resp.Body.Close()
|
||||
body, err := io.ReadAll(resp.Body)
|
||||
|
||||
return string(body), nil
|
||||
}
|
||||
|
||||
func parseTime(timestr string) (time.Time, error) {
|
||||
var formats = []string {
|
||||
time.ANSIC,
|
||||
time.UnixDate,
|
||||
time.RubyDate,
|
||||
time.RFC822,
|
||||
time.RFC822Z,
|
||||
time.RFC850,
|
||||
time.RFC1123,
|
||||
time.RFC1123Z,
|
||||
"2006-01-02 15:04:05 -0700 -0700",
|
||||
time.RFC3339,
|
||||
time.RFC3339Nano,
|
||||
time.DateTime,
|
||||
time.DateOnly,
|
||||
}
|
||||
var pagetime time.Time
|
||||
var err error
|
||||
for _, f := range formats {
|
||||
pagetime, err = time.Parse(f, timestr)
|
||||
if err == nil {
|
||||
return pagetime, err
|
||||
}
|
||||
}
|
||||
return pagetime, fmt.Errorf("Error parsing time: invalid format")
|
||||
}
|
||||
|
||||
func getHtmlElement(doc *html.Node, tag string) (*html.Node, error) {
|
||||
var f func(*html.Node, string)
|
||||
var element *html.Node
|
||||
f = func(n *html.Node, s string) {
|
||||
if n.Type == html.ElementNode && n.Data == s{
|
||||
element = n
|
||||
return
|
||||
}
|
||||
for c := n.FirstChild; c != nil; c = c.NextSibling {
|
||||
f(c, tag)
|
||||
}
|
||||
}
|
||||
f(doc, tag)
|
||||
if element == nil {
|
||||
return nil, fmt.Errorf("no <%s> element found", tag)
|
||||
}
|
||||
return element, nil
|
||||
}
|
||||
|
||||
func (f *FeedItem) ParseContent(content string) error {
|
||||
doc, err := html.Parse(strings.NewReader(content))
|
||||
if err != nil {
|
||||
return fmt.Errorf("Error parsing HTML: %w", err)
|
||||
}
|
||||
earticle, err := getHtmlElement(doc, "article")
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
var builder strings.Builder
|
||||
html.Render(&builder, earticle)
|
||||
f.RawText = builder.String()
|
||||
|
||||
etime, err := getHtmlElement(earticle, "time")
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
var pubTime time.Time
|
||||
for _, d := range etime.Attr {
|
||||
if d.Key == "datetime" {
|
||||
pubTime, err = parseTime(d.Val)
|
||||
}
|
||||
if err != nil {
|
||||
return fmt.Errorf("Error parsing time: %w", err)
|
||||
}
|
||||
f.PubTime = pubTime
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
func NewFeedItem(url string) (*FeedItem, error) {
|
||||
rawhtml, err := fetchPage(url)
|
||||
if err != nil {
|
||||
return nil, fmt.Errorf("Could not fetch page '%s': %w", url, err)
|
||||
}
|
||||
item := FeedItem{
|
||||
Url: url,
|
||||
}
|
||||
item.ParseContent(rawhtml);
|
||||
if err != nil {
|
||||
return nil, fmt.Errorf("Could not parse feed item: %w", err)
|
||||
}
|
||||
return &item, nil
|
||||
}
|
||||
|
||||
// parseArticle returns an error if it could not parse the HTML or if it could not parse a time
|
||||
// if a time could not be parsed, the parsed html article will still be returned
|
||||
func parseArticle(content string) (string, *time.Time, error) {
|
||||
doc, err := html.Parse(strings.NewReader(content))
|
||||
if err != nil {
|
||||
return "", nil, fmt.Errorf("Error parsing HTML: %w", err)
|
||||
}
|
||||
var f func(*html.Node, string)
|
||||
var element *html.Node
|
||||
var pagetime time.Time
|
||||
f = func(n *html.Node, tag string) {
|
||||
if n.Type == html.ElementNode && n.Data == tag {
|
||||
element = n
|
||||
return
|
||||
}
|
||||
for c := n.FirstChild; c != nil; c = c.NextSibling {
|
||||
f(c, tag)
|
||||
}
|
||||
}
|
||||
|
||||
f(doc, "article")
|
||||
var builder strings.Builder
|
||||
html.Render(&builder, element)
|
||||
|
||||
f(element, "time")
|
||||
for _, d := range element.Attr {
|
||||
if d.Key == "datetime" {
|
||||
pagetime, err = parseTime(d.Val)
|
||||
}
|
||||
}
|
||||
|
||||
return builder.String(), &pagetime, nil
|
||||
}
|
||||
|
|
@ -1,8 +1,8 @@
|
|||
package rss_test
|
||||
package feed_test
|
||||
|
||||
import "testing"
|
||||
import "time"
|
||||
import "git.32bit.cafe/yequari/rss-gen/rss"
|
||||
import "git.32bit.cafe/yequari/rss-gen/feed"
|
||||
|
||||
func TestArticleParse(t *testing.T) {
|
||||
testDate, err := time.Parse("2006-Jan-02", "2004-May-14")
|
||||
|
@ -31,7 +31,7 @@ func TestArticleParse(t *testing.T) {
|
|||
|
||||
for _, tt := range tests {
|
||||
t.Run(tt.name, func (t *testing.T) {
|
||||
article, articleTime, err := rss.ParseArticle(tt.input)
|
||||
article, articleTime, err := feed.ParseArticle(tt.input)
|
||||
if err != nil {
|
||||
t.Errorf("error: %s", err)
|
||||
}
|
||||
|
@ -44,3 +44,39 @@ func TestArticleParse(t *testing.T) {
|
|||
})
|
||||
}
|
||||
}
|
||||
|
||||
func TestParseFeedItem(t *testing.T) {
|
||||
testDate, err := time.Parse("2006-Jan-02", "2004-May-14")
|
||||
if err != nil {
|
||||
t.Errorf("creating test date failed: %s", err)
|
||||
}
|
||||
var tests = []struct {
|
||||
name string
|
||||
input string
|
||||
want_time *time.Time
|
||||
want_article string
|
||||
}{
|
||||
{
|
||||
"article and time stripped out of basic HTML",
|
||||
"<html><head></head><body><article><time datetime=\"2004-05-14\">May 14 2004</time>hello world</article></body></html>",
|
||||
&testDate,
|
||||
"<article><time datetime=\"2004-05-14\">May 14 2004</time>hello world</article>",
|
||||
},
|
||||
}
|
||||
|
||||
for _, tt := range tests {
|
||||
t.Run(tt.name, func (t *testing.T) {
|
||||
item := feed.FeedItem{}
|
||||
err := item.ParseContent(tt.input)
|
||||
if err != nil {
|
||||
t.Errorf("error: %s", err)
|
||||
}
|
||||
if item.RawText != tt.want_article {
|
||||
t.Errorf("got %s, want %s", item.RawText, tt.want_article)
|
||||
}
|
||||
if tt.want_time != nil && !item.PubTime.Equal(*tt.want_time) {
|
||||
t.Errorf("got %s, want %s", item.PubTime, *tt.want_time)
|
||||
}
|
||||
})
|
||||
}
|
||||
}
|
|
@ -0,0 +1,64 @@
|
|||
package feed
|
||||
|
||||
import (
|
||||
"fmt"
|
||||
"strings"
|
||||
"time"
|
||||
|
||||
"golang.org/x/net/html"
|
||||
)
|
||||
|
||||
const feedfmt = `<?xml version="1.0" encoding="utf-8"?>
|
||||
<rss version="2.0">
|
||||
<channel>
|
||||
<title>%s</title>
|
||||
<link>%s</link>
|
||||
<description>%s</description>%s
|
||||
</channel>
|
||||
</rss>`
|
||||
|
||||
const itemfmt = `
|
||||
<item>
|
||||
<title>Content Title</title>
|
||||
<link>%s</link>
|
||||
<guid>%s</guid>
|
||||
<pubDate>%s</pubDate>
|
||||
<description>
|
||||
%s
|
||||
</description>
|
||||
</item>`
|
||||
|
||||
type RSSBuilder struct {
|
||||
Info FeedInfo
|
||||
Items []FeedItem
|
||||
}
|
||||
|
||||
func GenerateRss(siteUrl, siteTitle, siteDesc string, pageUrls ...string) (string, error) {
|
||||
var items strings.Builder
|
||||
var errs strings.Builder
|
||||
var err error
|
||||
|
||||
for _, u := range pageUrls {
|
||||
var formattedArticle strings.Builder
|
||||
var err error
|
||||
page, err := fetchPage(u)
|
||||
if err != nil {
|
||||
continue
|
||||
}
|
||||
article, atime, err := parseArticle(page)
|
||||
if err != nil && article == "" {
|
||||
errs.WriteString(fmt.Sprintf("error parsing article %s: %s", u, err.Error()))
|
||||
continue
|
||||
}
|
||||
for _, line := range strings.Split(article, "\n") {
|
||||
formattedArticle.WriteString(fmt.Sprintf("\t\t%s\n", html.EscapeString(line)))
|
||||
}
|
||||
if atime != nil {
|
||||
items.WriteString(fmt.Sprintf(itemfmt, u, u, atime.Format("Mon, 2 Jan 2006 15:04:05 MST"), formattedArticle.String()))
|
||||
} else {
|
||||
items.WriteString(fmt.Sprintf(itemfmt, u, u, time.Now().Format("Mon, 2 Jan 2006 15:04:05 MST"), formattedArticle.String()))
|
||||
}
|
||||
}
|
||||
|
||||
return fmt.Sprintf(feedfmt, siteTitle, siteUrl, siteDesc, items.String()), err
|
||||
}
|
121
rss/rss.go
121
rss/rss.go
|
@ -1,121 +0,0 @@
|
|||
package rss
|
||||
|
||||
import (
|
||||
"fmt"
|
||||
"io"
|
||||
"net/http"
|
||||
"strings"
|
||||
"time"
|
||||
|
||||
"golang.org/x/net/html"
|
||||
)
|
||||
const feedfmt = `<?xml version="1.0" encoding="utf-8"?>
|
||||
<rss version="2.0">
|
||||
<channel>
|
||||
<title>%s</title>
|
||||
<link>%s</link>
|
||||
<description>%s</description>
|
||||
%s
|
||||
</channel>
|
||||
</rss>`
|
||||
|
||||
const itemfmt = `<item>
|
||||
<title>Content Title</title>
|
||||
<link>%s</link>
|
||||
<guid>%s</guid>
|
||||
<pubDate>%s</pubDate>
|
||||
<description><![CDATA[%s]]></description>
|
||||
</item>`
|
||||
|
||||
func fetchPage(url string) (string, error) {
|
||||
resp, err := http.Get(url)
|
||||
if err != nil {
|
||||
return "", fmt.Errorf("Error sending Get request: %w", err)
|
||||
}
|
||||
defer resp.Body.Close()
|
||||
body, err := io.ReadAll(resp.Body)
|
||||
|
||||
return string(body), nil
|
||||
}
|
||||
|
||||
func parseTime(timestr string) (*time.Time, error) {
|
||||
var formats = []string {
|
||||
time.ANSIC,
|
||||
time.UnixDate,
|
||||
time.RubyDate,
|
||||
time.RFC822,
|
||||
time.RFC822Z,
|
||||
time.RFC850,
|
||||
time.RFC1123,
|
||||
time.RFC1123Z,
|
||||
time.RFC3339,
|
||||
time.RFC3339Nano,
|
||||
time.DateTime,
|
||||
time.DateOnly,
|
||||
}
|
||||
for _, f := range formats {
|
||||
pagetime, err := time.Parse(f, timestr)
|
||||
if err == nil {
|
||||
return &pagetime, nil
|
||||
}
|
||||
}
|
||||
return nil, fmt.Errorf("Error parsing time: invalid format")
|
||||
}
|
||||
|
||||
// parseArticle returns an error if it could not parse the HTML or if it could not parse a time
|
||||
// if a time could not be parsed, the parsed html article will still be returned
|
||||
func parseArticle(content string) (string, *time.Time, error) {
|
||||
doc, err := html.Parse(strings.NewReader(content))
|
||||
if err != nil {
|
||||
return "", nil, fmt.Errorf("Error parsing HTML: %w", err)
|
||||
}
|
||||
var f func(*html.Node, string)
|
||||
var element *html.Node
|
||||
var pagetime *time.Time
|
||||
f = func(n *html.Node, tag string) {
|
||||
if n.Type == html.ElementNode && n.Data == tag {
|
||||
element = n
|
||||
return
|
||||
}
|
||||
for c := n.FirstChild; c != nil; c = c.NextSibling {
|
||||
f(c, tag)
|
||||
}
|
||||
}
|
||||
|
||||
f(doc, "article")
|
||||
var builder strings.Builder
|
||||
html.Render(&builder, element)
|
||||
|
||||
f(element, "time")
|
||||
for _, d := range element.Attr {
|
||||
if d.Key == "datetime" {
|
||||
pagetime, err = parseTime(d.Val)
|
||||
}
|
||||
}
|
||||
|
||||
return builder.String(), pagetime, nil
|
||||
}
|
||||
|
||||
func GenerateRss(siteUrl, siteTitle, siteDesc string, pageUrls ...string) (string, error) {
|
||||
var items strings.Builder
|
||||
var err error
|
||||
|
||||
for _, u := range pageUrls {
|
||||
var err error
|
||||
page, err := fetchPage(u)
|
||||
if err != nil {
|
||||
continue
|
||||
}
|
||||
article, atime, err := parseArticle(page)
|
||||
if err != nil && article == "" {
|
||||
continue
|
||||
}
|
||||
if atime != nil {
|
||||
items.WriteString(fmt.Sprintf(itemfmt, u, u, atime.Format("Mon, 2 Jan 2006 15:04:05 MST"), article))
|
||||
} else {
|
||||
items.WriteString(fmt.Sprintf(itemfmt, u, u, time.Now().Format("Mon, 2 Jan 2006 15:04:05 MST"), article))
|
||||
}
|
||||
}
|
||||
|
||||
return fmt.Sprintf(feedfmt, siteTitle, siteUrl, siteDesc, items.String()), err
|
||||
}
|
Loading…
Reference in New Issue