维基百科

找到一个可以访问wikipedia的方式

被刻印的时光 ゝ 提交于 2019-12-07 17:53:01
http://zh.wikipedia.org.nyud.net:8090 把zh改了,也可以访问其他语言的wikipedia。 http://en.wikipedia.org.nyud.net:8090 下面这个地址也可以访问 http://www.jaylee.cn/ 下面这是一个国内的wiki,做的还可以,顺便记以下 http://www.wikilib.com 来源: CSDN 作者: jolestar 链接: https://blog.csdn.net/Jolestar/article/details/83138515

访问wikipedia的方法

允我心安 提交于 2019-12-07 17:52:42
在Windows下,修改 \WINDOWS\system32\drivers\etc\hosts配置文件 如下添加一行: 145.97.39.155 en.wikipedia.org 这一行的格式是:IP+空格+域名1+空格+域名2+空格+域名3 ... 这个IP可以通过http://www.ip-adress.com/查询得到 如果碰到其他被封了的网站,同理突破 在Linux下修改/etc/hosts文件即可 来源: CSDN 作者: xiongx_ 链接: https://blog.csdn.net/xiongx_/article/details/83196640

Wikipedia 语料库处理

混江龙づ霸主 提交于 2019-12-07 17:52:24
参考: http://licstar.net/archives/262 参考的博客略旧,根据博客内容调整了处理过程 第一步:下载语料 最新的: http://download.wikipedia.com/zhwiki/latest/zhwiki-latest-pages-articles.xml.bz2 第二步:抽取正文 下载 Wikipedia Extractor: https://github.com/attardi/wikiextractor 安装 Wikipedia Extractor: sudo python setup.py install 抽取正文: python WikiExractorPath/WikiExtractor.py -b 1000M -o extracted WikiDataPath/zhwiki-latest-pages-articles.xml.bz2 得到的文件默认在当前路径下的,./extracted/AA/下 第三步:繁简转换 下载 OpenCC: https://github.com/BYVoid/OpenCC 安装OpenCC: make sudo make install 繁简转换: opencc -i wiki_00 -o wiki_00_chs -c OpenCCPath/data/config/t2s.json opencc -i

进Wikipedia

蓝咒 提交于 2019-12-07 17:52:07
1.直接通过 http://gollum.easycp.de/en/ 访问Wikipedia。 2.使用Firefox浏览器,下载Great Ladder的插件, https://addons.mozilla.org/firefox/2864/ 突破GFW,即可访问被大陆封锁的wikipedia.org ,#blogspot.com,Google网页快照等。此方法详见 http://www.kuju.cn/post/257.html 3.使用Firefox浏览器,下载Access Flickr(为伊朗人所开发)这个插件, https://addons.mozilla.org/en-US/firefox/addon/4286 ,即可访问Flickr。此方法详见 http://news.17tech.com/news/07061041629.html 来源: CSDN 作者: orangespring 链接: https://blog.csdn.net/Moonlily_L/article/details/2589824

访问wikipedia方法

别来无恙 提交于 2019-12-07 17:51:44
中国的域名解析把WIKIPEDIA给干掉了 有两种方法可以访问WIKIPEDIA: 1、https://secure.wikimedia.org,用https访问,不会比较慢,而且老是提示安全站点确认 2、通过IP还是访问;可以改你的hosts文件(%windir%\system32\drivers\etc),加上这么几行 145.97.39.155 www.wikipedia.org 145.97.39.155 en.wikipedia.org 145.97.39.155 zh.wikipedia.org 145.97.39.155 upload.wikimedia.org 此方法来源于夏桅 来源: CSDN 作者: weixin_34074740 链接: https://blog.csdn.net/weixin_34074740/article/details/86141455

通过改hosts访问wikipedia

柔情痞子 提交于 2019-12-07 17:51:29
能访问 https://www.wikipedia.org 大部分时候我们是可以访问的wikipedia的主页或是 英文首页 的,但 中文页面 却经常被墙,这时候你可以通过在命令行下 ping www.wikipedia.org 得到返回的IP地址,然后改 hosts 文件,添加 198.35 .26 .96 zh .wikipedia .org 就能正常访问 中文维基百科 了,但是注意 一定要用https 不能访问 https://www.wikipedia.org 和上面的区别是你必须在hosts里为所有wikipedia的域名添加IP映射关系 可以通过 站长工具 查询wikipedia下的二级域名 或者直接用下面这些 2480 zh .wikipedia .org 2760 zh .wikipedia .org 3 dd7zh .wikipedia .org 5458 ja .wikipedia .org 5 a0en .wikipedia .org 5 a0zh .wikipedia .org 8111 zh .wikipedia .org ab .wikipedia .org ae58zh .wikipedia .org af .wikipedia .org ak .wikipedia .org als .wikipedia .org am .wikipedia .org

How to access wikipedia

匿名 (未验证) 提交于 2019-12-03 07:50:05
可以将文章内容翻译成中文,广告屏蔽插件可能会导致该功能失效(如失效,请关闭广告屏蔽插件后再试): 问题: I want to access HTML content from wikipedia .But it is showing access denied. How can i access Wiki. Please give some suggestion 回答1: Use HttpWebRequest Try the following: string Text = "http://www.wikipedia.org/"; HttpWebRequest request = (HttpWebRequest)HttpWebRequest.Create(Text); request.UserAgent = "Mozilla/4.0 (compatible; MSIE 7.0; Windows NT 6.0)"; HttpWebResponse respons; respons = (HttpWebResponse)request.GetResponse(); Encoding enc = Encoding.GetEncoding(respons.CharacterSet); StreamReader reader = new StreamReader(respons

Wikipedia Category Hierarchy from dumps

匿名 (未验证) 提交于 2019-12-03 02:51:02
可以将文章内容翻译成中文,广告屏蔽插件可能会导致该功能失效(如失效,请关闭广告屏蔽插件后再试): 问题: Using Wikipedia's dumps I want to build a hierarchy for its categories. I have downloaded the main dump (enwiki-latest-pages-articles) and the category SQL dump (enwiki-latest-category). But I can't find the hierarchy information. For example, the SQL categories' dump has entries for each category but I can't find anything about how they relate to each other. The other dump (latest-pages-articles) says the parent categories for each page but in an unordered way. It just states all the parents. I have seen wikiprep's category hierarchy ( http

How to get all Wikipedia article titles?

匿名 (未验证) 提交于 2019-12-03 02:44:02
可以将文章内容翻译成中文,广告屏蔽插件可能会导致该功能失效(如失效,请关闭广告屏蔽插件后再试): 问题: How to get all Wikipedia article titles in one place without extra characters and pageids. Just the article's title. Something like this: When I download wikipedia dump, I get this Maybe I know a movement that might get me all pages but I wanted to get all pages in one take. 回答1: You'll find it on https://dumps.wikimedia.org The latest List of page titles in main namespace for English Wikipedia as a database dump is here (69 MB). If you rather want it through the API you use query and list=allpages but that only give you maximum 500 (5k for

API for Wikipedia in Objective C?

匿名 (未验证) 提交于 2019-12-03 02:14:01
可以将文章内容翻译成中文,广告屏蔽插件可能会导致该功能失效(如失效,请关闭广告屏蔽插件后再试): 问题: Does anybody know if there is an implementation of the wikipedia api in objective c? which can be used to access and display the information a bit nicer then with the plain website. I'm thinking of something like the wikipanion app but not so sophisticated. Thanks Joerg 回答1: I know the question is very old. But just for other ppl who are dealing with the same issue. I wrote a very small library which provides basic access to the Wikipedia API. You can get it on github: https://github.com/prine/WikiApiObjectiveC 回答2: You can use the MediaWiki