Wikipedia
A Pythonic wrapper for the Wikipedia API
Wikipedia — wikipedia 0.9 documentation
I'm looking to enhance several "objects" in my application with human-readable data. To that end, I've seen Freebase, DBpedia and wikidata.org, and am currently working with Freebase. I can't help but wonder, though, what I am missing.
So: what's the difference? Specifically, what is the coverage difference, and what is the difference in terms of data provided for the most commonly-viewed Wikipedia articles types (such as "Person", "Place", "Artist", "Album", etc.).
Source: (StackOverflow)
If I have the url to a page, how would I obtain the infobox information on the right using MediaWiki webservices?
Source: (StackOverflow)
I'm trying to find out if there's a wikipedia api (I Think it is related to the mediawiki?).
If so, I would like to know how I would tell wikipedia to give me an article about the new york yankees for example.
What would the REST url be for this example?
All the docs on this subject seem fairly complicated.
Source: (StackOverflow)
I need just to retrieve first paragraph of a Wikipedia page.
Content must be html formated, ready to be displayed on my websites (so NO BBCODE, or WIKIPEDIA special CODE!)
Source: (StackOverflow)
I try to fetch a Wikipedia article with Python's urllib:
f = urllib.urlopen("http://en.wikipedia.org/w/index.php?title=Albert_Einstein&printable=yes")
s = f.read()
f.close()
However instead of the html page I get the following response: Error - Wikimedia Foundation:
Request: GET http://en.wikipedia.org/w/index.php?title=Albert_Einstein&printable=yes, from 192.35.17.11 via knsq1.knams.wikimedia.org (squid/2.6.STABLE21) to ()
Error: ERR_ACCESS_DENIED, errno [No Error] at Tue, 23 Sep 2008 09:09:08 GMT
Wikipedia seems to block request which are not from a standard browser.
Anybody know how to work around this?
Source: (StackOverflow)
I'm studying the wikipedia API,
some demo api call
What is the pageid? How do I change it into a real page url?
I mean <page pageid="18630637" ns="0" title="Translation" />
, how to change 18630637
into http://en.wikipedia.org/wiki/Translation
?
Source: (StackOverflow)
How can I extract the first paragraph from a Wikipedia article, using Python?
For example, for Albert Einstein, that would be:
Albert Einstein (pronounced /ˈælbərt
ˈaɪnstaɪn/; German: [ˈalbɐt ˈaɪnʃtaɪn]
( listen); 14 March 1879 – 18 April
1955) was a theoretical physicist,
philosopher and author who is widely
regarded as one of the most
influential and iconic scientists and
intellectuals of all time. A
German-Swiss Nobel laureate, Einstein
is often regarded as the father of
modern physics.[2] He received the
1921 Nobel Prize in Physics "for his
services to theoretical physics, and
especially for his discovery of the
law of the photoelectric effect".[3]
Source: (StackOverflow)
Is it possible, using the existing Wikipedia API's to get a list of articles around a Geo-location? Sort of like how Google maps does it?
I would like to say that I am "here" and find out what is around me on Wikipedia.
I can see on articles like this you can see the "Coordinates" on the right hand side, so I would like to do a query on these coordinates...
any thoughts?
Source: (StackOverflow)
I`m trying to insert a custom link to a special page in VisualEditor toolbar. See the image below.
See Image
I googled a lot but without success. Someone please give a path...
Source: (StackOverflow)
Wikipedia's Wavelet article contains this text:
The discrete wavelet transform is also less computationally complex, taking O(N) time as compared to O(N log N) for the fast Fourier transform. This computational advantage is not inherent to the transform, but reflects the choice of a logarithmic division of frequency, in contrast to the equally spaced frequency divisions of the FFT.
Does this imply that there's also an FFT-like algorithm that uses a logarithmic division of frequency instead of linear? Is it also O(N)? This would obviously be preferable for a lot of applications.
Source: (StackOverflow)
I downloaded a Wikipedia dump and I want to convert the wiki format into my object format. Is there a wiki parser available that converts the object into XML?
Source: (StackOverflow)
I recently used Wikipedia's function "What links here" (which is found under the "Toolbox" element in any entry's left menu) and it got me started wondering how this function actually works.
I'm guessing that searching through all the article entries after links isn't very effective, so are all the links stored in a separate database? If so, is this updated when an article is edited or another time?
Thanks.
Source: (StackOverflow)
I'm trying to fetch Wikipedia pages using LWP::Simple, but they're not coming back. This code:
#!/usr/bin/perl
use strict;
use LWP::Simple;
print get("http://en.wikipedia.org/wiki/Stack_overflow");
doesn't print anything. But if I use some other webpage, say http://www.google.com
, it works fine.
Is there some other name that I should be using to refer to Wikipedia pages?
What could be going on here?
Source: (StackOverflow)
I am looking to download full Wikipedia text for my college project. Do I have to write my own spider to download this or is there a public dataset of Wikipedia available online?
To just give you some overview of my project, I want to find out the interesting words of few articles I am interested in. But to find these interesting words, I am planning to apply tf/idf to calculate term frequency for each word and pick the ones with high frequency. But to calculate the tf, I need to know the total occurrences in whole of Wikipedia.
Your help would be greatly appreciated.
Thank you
Bala
Source: (StackOverflow)
Is anyone aware of a way to programatically download images from Wikimedia Commons without registering for a Bot account? It seems like the only way to get approval for a Bot account is if it adds to or edits information already on Wikimedia. If you try to download any images, without a bot account, using some of the api libraries out there you get error messages instead of the images. Seems like they block anyone not coming in from a browser? Anyone else have any experience with this? Am I missing something here?
Source: (StackOverflow)