| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212 |
- // Copyright 2017 Frédéric Guillot. All rights reserved.
- // Use of this source code is governed by the Apache 2.0
- // license that can be found in the LICENSE file.
- package handler // import "miniflux.app/reader/handler"
- import (
- "fmt"
- "time"
- "miniflux.app/config"
- "miniflux.app/errors"
- "miniflux.app/http/client"
- "miniflux.app/locale"
- "miniflux.app/logger"
- "miniflux.app/model"
- "miniflux.app/reader/browser"
- "miniflux.app/reader/icon"
- "miniflux.app/reader/parser"
- "miniflux.app/reader/processor"
- "miniflux.app/storage"
- "miniflux.app/timer"
- )
- var (
- errDuplicate = "This feed already exists (%s)"
- errNotFound = "Feed %d not found"
- errCategoryNotFound = "Category not found for this user"
- )
- // FeedCreationArgs represents the arguments required to create a new feed.
- type FeedCreationArgs struct {
- UserID int64
- CategoryID int64
- FeedURL string
- UserAgent string
- Username string
- Password string
- Crawler bool
- Disabled bool
- IgnoreHTTPCache bool
- FetchViaProxy bool
- ScraperRules string
- RewriteRules string
- BlocklistRules string
- KeeplistRules string
- }
- // CreateFeed fetch, parse and store a new feed.
- func CreateFeed(store *storage.Storage, args *FeedCreationArgs) (*model.Feed, error) {
- defer timer.ExecutionTime(time.Now(), fmt.Sprintf("[CreateFeed] FeedURL=%s", args.FeedURL))
- if !store.CategoryExists(args.UserID, args.CategoryID) {
- return nil, errors.NewLocalizedError(errCategoryNotFound)
- }
- request := client.NewClientWithConfig(args.FeedURL, config.Opts)
- request.WithCredentials(args.Username, args.Password)
- request.WithUserAgent(args.UserAgent)
- if args.FetchViaProxy {
- request.WithProxy()
- }
- response, requestErr := browser.Exec(request)
- if requestErr != nil {
- return nil, requestErr
- }
- if store.FeedURLExists(args.UserID, response.EffectiveURL) {
- return nil, errors.NewLocalizedError(errDuplicate, response.EffectiveURL)
- }
- subscription, parseErr := parser.ParseFeed(response.EffectiveURL, response.BodyAsString())
- if parseErr != nil {
- return nil, parseErr
- }
- subscription.UserID = args.UserID
- subscription.UserAgent = args.UserAgent
- subscription.Username = args.Username
- subscription.Password = args.Password
- subscription.Crawler = args.Crawler
- subscription.Disabled = args.Disabled
- subscription.IgnoreHTTPCache = args.IgnoreHTTPCache
- subscription.FetchViaProxy = args.FetchViaProxy
- subscription.ScraperRules = args.ScraperRules
- subscription.RewriteRules = args.RewriteRules
- subscription.BlocklistRules = args.BlocklistRules
- subscription.KeeplistRules = args.KeeplistRules
- subscription.WithCategoryID(args.CategoryID)
- subscription.WithClientResponse(response)
- subscription.CheckedNow()
- processor.ProcessFeedEntries(store, subscription)
- if storeErr := store.CreateFeed(subscription); storeErr != nil {
- return nil, storeErr
- }
- logger.Debug("[CreateFeed] Feed saved with ID: %d", subscription.ID)
- checkFeedIcon(store, subscription.ID, subscription.SiteURL, args.FetchViaProxy)
- return subscription, nil
- }
- // RefreshFeed refreshes a feed.
- func RefreshFeed(store *storage.Storage, userID, feedID int64) error {
- defer timer.ExecutionTime(time.Now(), fmt.Sprintf("[RefreshFeed] feedID=%d", feedID))
- userLanguage := store.UserLanguage(userID)
- printer := locale.NewPrinter(userLanguage)
- originalFeed, storeErr := store.FeedByID(userID, feedID)
- if storeErr != nil {
- return storeErr
- }
- if originalFeed == nil {
- return errors.NewLocalizedError(errNotFound, feedID)
- }
- weeklyEntryCount := 0
- if config.Opts.PollingScheduler() == model.SchedulerEntryFrequency {
- var weeklyCountErr error
- weeklyEntryCount, weeklyCountErr = store.WeeklyFeedEntryCount(userID, feedID)
- if weeklyCountErr != nil {
- return weeklyCountErr
- }
- }
- originalFeed.CheckedNow()
- originalFeed.ScheduleNextCheck(weeklyEntryCount)
- request := client.NewClientWithConfig(originalFeed.FeedURL, config.Opts)
- request.WithCredentials(originalFeed.Username, originalFeed.Password)
- request.WithUserAgent(originalFeed.UserAgent)
- if !originalFeed.IgnoreHTTPCache {
- request.WithCacheHeaders(originalFeed.EtagHeader, originalFeed.LastModifiedHeader)
- }
- if originalFeed.FetchViaProxy {
- request.WithProxy()
- }
- response, requestErr := browser.Exec(request)
- if requestErr != nil {
- originalFeed.WithError(requestErr.Localize(printer))
- store.UpdateFeedError(originalFeed)
- return requestErr
- }
- if store.AnotherFeedURLExists(userID, originalFeed.ID, response.EffectiveURL) {
- storeErr := errors.NewLocalizedError(errDuplicate, response.EffectiveURL)
- originalFeed.WithError(storeErr.Error())
- store.UpdateFeedError(originalFeed)
- return storeErr
- }
- if originalFeed.IgnoreHTTPCache || response.IsModified(originalFeed.EtagHeader, originalFeed.LastModifiedHeader) {
- logger.Debug("[RefreshFeed] Feed #%d has been modified", feedID)
- updatedFeed, parseErr := parser.ParseFeed(response.EffectiveURL, response.BodyAsString())
- if parseErr != nil {
- originalFeed.WithError(parseErr.Localize(printer))
- store.UpdateFeedError(originalFeed)
- return parseErr
- }
- originalFeed.Entries = updatedFeed.Entries
- processor.ProcessFeedEntries(store, originalFeed)
- // We don't update existing entries when the crawler is enabled (we crawl only inexisting entries).
- if storeErr := store.RefreshFeedEntries(originalFeed.UserID, originalFeed.ID, originalFeed.Entries, !originalFeed.Crawler); storeErr != nil {
- originalFeed.WithError(storeErr.Error())
- store.UpdateFeedError(originalFeed)
- return storeErr
- }
- // We update caching headers only if the feed has been modified,
- // because some websites don't return the same headers when replying with a 304.
- originalFeed.WithClientResponse(response)
- checkFeedIcon(store, originalFeed.ID, originalFeed.SiteURL, originalFeed.FetchViaProxy)
- } else {
- logger.Debug("[RefreshFeed] Feed #%d not modified", feedID)
- }
- originalFeed.ResetErrorCounter()
- if storeErr := store.UpdateFeed(originalFeed); storeErr != nil {
- originalFeed.WithError(storeErr.Error())
- store.UpdateFeedError(originalFeed)
- return storeErr
- }
- return nil
- }
- func checkFeedIcon(store *storage.Storage, feedID int64, websiteURL string, fetchViaProxy bool) {
- if !store.HasIcon(feedID) {
- icon, err := icon.FindIcon(websiteURL, fetchViaProxy)
- if err != nil {
- logger.Debug(`[CheckFeedIcon] %v (feedID=%d websiteURL=%s)`, err, feedID, websiteURL)
- } else if icon == nil {
- logger.Debug(`[CheckFeedIcon] No icon found (feedID=%d websiteURL=%s)`, feedID, websiteURL)
- } else {
- if err := store.CreateFeedIcon(feedID, icon); err != nil {
- logger.Debug(`[CheckFeedIcon] %v (feedID=%d websiteURL=%s)`, err, feedID, websiteURL)
- }
- }
- }
- }
|