72 lines
4.0 KiB
Python
72 lines
4.0 KiB
Python
# This file is part of morss
|
|
#
|
|
# Copyright (C) 2013-2020 pictuga <contact@pictuga.com>
|
|
#
|
|
# This program is free software: you can redistribute it and/or modify it under
|
|
# the terms of the GNU Affero General Public License as published by the Free
|
|
# Software Foundation, either version 3 of the License, or (at your option) any
|
|
# later version.
|
|
#
|
|
# This program is distributed in the hope that it will be useful, but WITHOUT
|
|
# ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
|
|
# FOR A PARTICULAR PURPOSE. See the GNU Affero General Public License for more
|
|
# details.
|
|
#
|
|
# You should have received a copy of the GNU Affero General Public License along
|
|
# with this program. If not, see <https://www.gnu.org/licenses/>.
|
|
|
|
import argparse
|
|
import os.path
|
|
import sys
|
|
|
|
from .morss import FeedFetch, FeedFormat, FeedGather, Options
|
|
|
|
|
|
def cli_app():
|
|
parser = argparse.ArgumentParser(
|
|
prog='morss',
|
|
description='Get full-text RSS feeds',
|
|
epilog='GNU AGPLv3 code'
|
|
)
|
|
|
|
parser.add_argument('url', help='feed url')
|
|
|
|
parser.add_argument('--post', action='store', type=str, metavar='STRING', help='POST request')
|
|
parser.add_argument('--xpath', action='store', type=str, metavar='XPATH', help='xpath rule to manually detect the article')
|
|
|
|
group = parser.add_argument_group('output')
|
|
group.add_argument('--format', default='rss', choices=('rss', 'json', 'html', 'csv'), help='output format')
|
|
group.add_argument('--search', action='store', type=str, metavar='STRING', help='does a basic case-sensitive search in the feed')
|
|
group.add_argument('--clip', action='store_true', help='stick the full article content under the original feed content (useful for twitter)')
|
|
group.add_argument('--indent', action='store_true', help='returns indented XML or JSON, takes more place, but human-readable')
|
|
|
|
group = parser.add_argument_group('action')
|
|
group.add_argument('--cache', action='store_true', help='only take articles from the cache (ie. don\'t grab new articles\' content), so as to save time')
|
|
group.add_argument('--force', action='store_true', help='force refetch the rss feed and articles')
|
|
group.add_argument('--proxy', action='store_true', help='doesn\'t fill the articles')
|
|
group.add_argument('--order', default='first', choices=('first', 'last', 'newest', 'oldest'), help='order in which to process items (which are however NOT sorted in the output)')
|
|
group.add_argument('--firstlink', action='store_true', help='pull the first article mentioned in the description instead of the default link')
|
|
group.add_argument('--resolve', action='store_true', help='replace tracking links with direct links to articles (not compatible with --proxy)')
|
|
|
|
group = parser.add_argument_group('custom feeds')
|
|
group.add_argument('--items', action='store', type=str, metavar='XPATH', help='(mandatory to activate the custom feeds function) xpath rule to match all the RSS entries')
|
|
group.add_argument('--item_link', action='store', type=str, metavar='XPATH', help='xpath rule relative to items to point to the entry\'s link')
|
|
group.add_argument('--item_title', action='store', type=str, metavar='XPATH', help='entry\'s title')
|
|
group.add_argument('--item_content', action='store', type=str, metavar='XPATH', help='entry\'s content')
|
|
group.add_argument('--item_time', action='store', type=str, metavar='XPATH', help='entry\'s date & time (accepts a wide range of time formats)')
|
|
|
|
group = parser.add_argument_group('misc')
|
|
group.add_argument('--nolink', action='store_true', help='drop links, but keeps links\' inner text')
|
|
group.add_argument('--noref', action='store_true', help='drop items\' link')
|
|
group.add_argument('--silent', action='store_true', help='don\'t output the final RSS (useless on its own, but can be nice when debugging)')
|
|
|
|
options = Options(vars(parser.parse_args()))
|
|
url = options.url
|
|
|
|
url, rss = FeedFetch(url, options)
|
|
rss = FeedGather(rss, url, options)
|
|
out = FeedFormat(rss, options, 'unicode')
|
|
|
|
if not options.silent:
|
|
print(out)
|