Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 7 years ago.
Improve this question
I'm building a database + tool that scours news feeds for a certain term. For example "food poisoning from nuts". I want to scour social media sites, news sites, major news aggregators, etc... for that term.
Question 1: What are some of the news
aggregator APIs out there?
Question 2: How Would you go about coding and
receiving only the latest news from
the API?
Edit Added schematic:
alt text http://koopics.com/news_parser.jpg
Do you know Yahoo! Pipes? It's a very flexible feed aggregator, and you can manipulate it using YQL, which is quite powerful and has a Python librabry, python-yql :). YQL also has specific "tables" for Twitter and other services and news sources, so, depending on what you want, you might not even need Pipes.
There's a quick example of Pipes + YQL usage at http://blog.ouseful.info/2009/04/27/using-yql-with-yahoo-pipes/ , but you can play around with them at the YQL Console.
So, that's your "Question 1". But, using YQL, "Question 2" (if I understood it correctly) comes naturally, as you use it almost like regular SQL, imposing limits, ordering etc.
Have you tried Universtal Feed Parser ??
Google has a news feeds api but I think it is only for non commercial use
http://www.google.com/support/news/bin/answer.py?answer=59255&hl=en
If you've got monetary support backing your project, Reuters has news data feeds that you can subscribe to on a monthly basis.
Related
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 3 years ago.
Improve this question
I am trying to get full scorecard and statistics of the cricket world cup 2019 and matches of the past 5 years. I have gone through many but none of those are good enough.
I have gone with these while looking for any good source of free APIs (at least a few initial hit):
Pycricbuzz: But this does not give many choices. It is just for few and live matches.
Python PyCricket: This does not work nowadays.
Cricket API from Roanuz: Upfront payment and too costly. Though, I haven't tested it yet.
Cric API: It does not provide a full scorecard of previous matches.
Cricsheet: It do provide a lot of information but I need to download a lot of files on upfront.
What I want?
full scorecard of last five years matches and other matches related information as we get on the cricbuzz website.
Is there any?
There is not any such available yet. CricinfoESPN do have a plan to bring soon. Till then you can use web scraping techniques on crickbuzz or Cricinfo sites to get scorecards otherwise you can also look howstats website too for more statistics.
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 6 years ago.
Improve this question
I'm searching for an API or a program (preferably Python and open-source) which lets me download the first n pictures of a Google Image Search for let's say bicycles. It would also be helpful if it could download the first n .pdf files from a normal search. Since not all pictures and .pdf files are found on Google and since there are many other search engines, a program which could also scrape results from Yahoo or Bing would be very convenient. Are there any such programs or is there an API from Google which lets me do more than 100 searches a day?
edit: People passing by may want to look at my attempt of programming such a scraper here
According to this post, all Google Search APIs have been deprecated.
However, GoogleScraper, an open-source library can help you achieve what you intend achieving.
If want to go barebones, and implement this yourself, BeautifulSoup is a very nice library to work with.
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 8 years ago.
Improve this question
I stumbled upon the wikidump python library, which I think suits me just fine.
I could get by by looking at the source code, but I'm new at python and I don't want to write BS code as the project I need it for is kind of important to me.
I got the 'wiki-SPECIFICDATE-pages-articles.xml.bz2' file and I would need to use that as my source for single article fetching. Can anyone give me some pointers as to properly achieve this or, even better, point at some documentation? I couldn't find any!
(p.s. if you got any better and properly doc'd lib, please tell me)
Not sure if I understand the question, but if you have the Wikipedia dump and you need to parse the wikicode, I would suggest mwparserfromhell lib.
Another powerful framework is Pywikibot, that is the historic framework for bot users on Wikipedia (thus, it has many scripts dedicated to writing pages, instead of reading and parsing articles). It has a lot of documentation (though, sometimes obsolete) and it uses MediaWiki API.
You can use them both, of course: PWB for fetching articles and mwparserfromhell for parsing.
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 3 years ago.
Improve this question
I've been searching for a library to use to access the eBay Dev API (using Python 3) for quite some time, but the sources I've found either don't specify the use of Python 3 (and I've learned to assume that this means they use 2.x) or haven't been updated for years.
Can somebody point me in the right direction? Is there a library I'm missing? Or should I try to implement it myself? If I have to implement it myself...any words/tutorials to help? I'm really confused here....thanks in advance.
eBay API can be found in this general link: https://www.x.com/developers/ebay/documentation-tools/tutorials
And 2.7x isn't obsolete? Pray tell why it won't be in a couple years? Or...just...what?
sigh due to the words spoken here and read elsewhere...I'll just bite the bullet and re-learn Python in 2.7.x....whatever. Thanks anyways.
python-ebay is a python wrapper for eBay APIs. It covers most of the eBay APIs.
PS: I wrote it.
I tried python-ebay but it doesn't implement every function of the Finding API (for example, there is no findCompletedItems function).
The alternative ebaysdk-python seems to be more actively maintained, is a little more flexible and does everything I need in a more simple way (IMO).
For example, getting completed items for a category is as simple as:
api = finding(appid='APP_ID')
api.execute('findCompletedItems', {'categoryId': '619'})
print api.response_dict()
The dictionary that's passed to the execute function exactly mirrors the eBay Finding API documentation, so if you can read that, you can basically do anything you need with the above three line of code.
The official eBay page https://go.developer.ebay.com/ebay-sdks recommends https://github.com/timotheus/ebaysdk-python . Currently ebaysdk-python has been updated 9 days ago and regularly since 2011, while python-ebay hasn't been updated for 20 months.
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 4 years ago.
Improve this question
I'm just starting with mwclient. I'm going to create bots to query our MediaWiki database and make small revisions.
But I cannot find anywhere a simple list of python commands like how to get ages of pages, contents of categories, contents of pages, etc.
Does anyone know a good starters resource?
The official docs at https://github.com/mwclient/mwclient/wiki have some introductory tutorials. I'm in charge for documentation for mwclient but haven't had enough time to really expand them - could use help from anyone who is willing.
One of my colleagues just sent me a link to the MediaWiki API wiki page.
I currently use python+urllib for API queries, and mwclient whenever I need to edit/create a page.
An useful place to get started with mwclient (read/edit/create a page):
http://brianna.modernthings.org/article/134/write-api-enabled-on-wikimedia-sites
The Bot Manual also has tons of good info and links, e.g. creating a bot.