Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

大家可以对此项目提出建议 #4

Open
Hatcat123 opened this issue Jul 3, 2019 · 12 comments
Open

大家可以对此项目提出建议 #4

Hatcat123 opened this issue Jul 3, 2019 · 12 comments
Labels
公告 公告

Comments

@Hatcat123
Copy link
Contributor

我们团队会积极回应大家的建议

@687766616e
Copy link

fofa導出建议:
導出格式: {http://|https://}(可選){ip|hostname}{:port}}(可選)

@Hatcat123 Hatcat123 pinned this issue Jul 3, 2019
@ghost ghost unpinned this issue Jul 6, 2019
@Hatcat123 Hatcat123 added the 公告 公告 label Jul 8, 2019
@Hatcat123 Hatcat123 pinned this issue Jul 8, 2019
@ggann9527
Copy link

DianpingCrawler/dazhong.py
51行:‘res = requests.get(self._cur_request_css_url, headers=self._default_headers, cookies=self._cookies)’
应该为‘res = requests.get(url, headers=self._default_headers, cookies=self._cookies)’

@Hatcat123
Copy link
Contributor Author

DianpingCrawler/dazhong.py
51行:‘res = requests.get(self._cur_request_css_url, headers=self._default_headers, cookies=self._cookies)’
应该为‘res = requests.get(url, headers=self._default_headers, cookies=self._cookies)’

好的,你的意见真的很有用,同时你可以进行pR提交

@687766616e
Copy link

DianpingCrawler/dazhong.py
51行:‘res = requests.get(self._cur_request_css_url, headers=self._default_headers, cookies=self._cookies)’
应该为‘res = requests.get(url, headers=self._default_headers, cookies=self._cookies)’

好的,你的意见真的很有用,同时你可以进行pR提交

google translate? 感覺怪怪的...😅

@Hatcat123
Copy link
Contributor Author

添加影评分析爬虫

@gaowei1012
Copy link

老哥,最近爬虫风声很近,慎搞慎搞

@TianmaBifang
Copy link

运行环境,用到的依赖,使用方法,写清楚

@Hatcat123
Copy link
Contributor Author

这个已经加入到projects计划。项目比较多正在整理,完善

@wangx404
Copy link

界面完全可以用PyQt来写,比TK好用。

@QuentinDan
Copy link

您好,目前有一个需求是像搜狗微信一样按关键词搜索公众号文章,获得文章标题、链接、公众号名称,请问贵团队知道有技术路径可以实现这个需求吗?为什么搜狗微信老是搜索不全呢?

@Hatcat123
Copy link
Contributor Author

您好,目前有一个需求是像搜狗微信一样按关键词搜索公众号文章,获得文章标题、链接、公众号名称,请问贵团队知道有技术路径可以实现这个需求吗?为什么搜狗微信老是搜索不全呢?

搜狗微信只能搜索部分数据,拿不到全部的数据,我们之前和您的想法一致,但是在实践中放弃了这种做法。我们的做法是:定时监控采集某个相关领域的所有公众号的文章,然后再做搜索。http://wechat.doonsec.com
当然这样也还是不能根据关键字得到100%的数据,只能保证在我监控公众号内的范围。除非你能拥有腾讯微信数据库的权限,这看起来似乎没有可能。

@siyu1992
Copy link

您好,安居客的经纪人信息爬虫目前好像网站上是用的二维码转到小程序去才能爬到电话号码,这个点能做吗?感谢感谢

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
公告 公告
Projects
None yet
Development

No branches or pull requests

8 participants