1. Advertising
    y u no do it?

    Advertising (learn more)

    Advertise virtually anything here, with CPM banner ads, CPM email ads and CPC contextual links. You can target relevant areas of the site and show ads based on geographical location of the user if you wish.

    Starts at just $1 per CPM or $0.10 per CPC.

Wiki grabber?

Discussion in 'Content Management' started by Toopac, May 27, 2006.

  1. #1
    Is it possible to download Wikipedia, but just one section? like actors for example...
     
    Toopac, May 27, 2006 IP
  2. webmasterlabor.com

    webmasterlabor.com Peon

    Messages:
    2,889
    Likes Received:
    76
    Best Answers:
    0
    Trophy Points:
    0
    #2
    You probably will need a custom spider for this. Contact me for custom content extraction. Make sure you comply with Wikipedia's content license rules though.
     
    webmasterlabor.com, May 27, 2006 IP
  3. iconv

    iconv Well-Known Member

    Messages:
    189
    Likes Received:
    3
    Best Answers:
    0
    Trophy Points:
    108
    #3
    You can download a dump in xml, and then filter out the categorie you are interested in. This way you can avoid spidering wikipedia.org.
     
    iconv, May 28, 2006 IP
  4. pwaring

    pwaring Well-Known Member

    Messages:
    846
    Likes Received:
    25
    Best Answers:
    0
    Trophy Points:
    135
    #4
    You can download dumps of the databases from:

    http://download.wikimedia.org/

    I don't think it's possible to download only specific pages though, but you could quite easily setup some sort of cron job to download the dump, extract it and parse just the XML files you wanted.
     
    pwaring, May 28, 2006 IP
  5. webmasterlabor.com

    webmasterlabor.com Peon

    Messages:
    2,889
    Likes Received:
    76
    Best Answers:
    0
    Trophy Points:
    0
    #5
    If you use their download dump and need a PARSER, hit me up :)
     
    webmasterlabor.com, May 28, 2006 IP