提交 bcf9dc87 作者: XveLingKun

0909-更新mysql和es的ip和mongo的ip

上级 9e335d9a
......@@ -550,7 +550,7 @@ class BaseCore:
def sendkafka(self, post_data, topic):
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], api_version=(2, 0, 2))
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], api_version=(2, 0, 2))
kafka_result = producer.send(topic, json.dumps(post_data, ensure_ascii=False).encode('utf8'))
print(kafka_result.get(timeout=10))
......
......@@ -35,7 +35,7 @@ thread_local = threading.local()
from tempfile import TemporaryFile
r = redis.StrictRedis(host='114.115.221.202', port=6379, db=1, decode_responses=True, password='clbzzsn')
db_storage = pymongo.MongoClient('mongodb://114.115.221.202:27017/', username='admin', password='ZZsn@9988').中科软[
db_storage = pymongo.MongoClient('mongodb://1.95.69.135:27017/', username='admin', password='ZZsn@9988').中科软[
'数据源_0504']
# path = r'D:\soft\msedgedriver.exe'
......
......@@ -61,7 +61,7 @@ aa_dict = {
'status': 0, # 状态
}
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], api_version=(2, 0, 2))
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], api_version=(2, 0, 2))
kafka_result = producer.send("regionInfo", json.dumps(aa_dict, ensure_ascii=False).encode('utf8'))
print(kafka_result.get(timeout=10))
except:
......
......@@ -419,7 +419,7 @@ if __name__ == '__main__':
baseCore.writerToExcel(name_list, file_name)
log.info(f'采集{com_name}成功=======耗时{baseCore.getTimeCost(start_time,time.time())}')
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], api_version=(2, 0, 2))
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], api_version=(2, 0, 2))
kafka_result = producer.send("regionInfo", json.dumps(post_data, ensure_ascii=False).encode('utf8'))
print(kafka_result.get(timeout=10))
except:
......
......@@ -44,7 +44,7 @@ aa_dict = {
}
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], api_version=(2, 0, 2))
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], api_version=(2, 0, 2))
kafka_result = producer.send("regionInfo", json.dumps(aa_dict, ensure_ascii=False).encode('utf8'))
print(kafka_result.get(timeout=10))
except:
......
......@@ -44,7 +44,7 @@ def create_driver():
# 发送数据
def sendkafka(post_data):
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], api_version=(2, 0, 2))
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], api_version=(2, 0, 2))
kafka_result = producer.send("enterpriseInfo", json.dumps(post_data, ensure_ascii=False).encode('utf8'))
print(kafka_result.get(timeout=10))
except:
......
......@@ -26,7 +26,7 @@ tag = Tag()
# 发送数据
def sendkafka(post_data):
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], api_version=(2, 0, 2))
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], api_version=(2, 0, 2))
kafka_result = producer.send("enterpriseInfo", json.dumps(post_data, ensure_ascii=False).encode('utf8'))
print(kafka_result.get(timeout=10))
except:
......
......@@ -26,7 +26,7 @@ tag = Tag()
# 发送数据
def sendkafka(post_data):
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], api_version=(2, 0, 2))
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], api_version=(2, 0, 2))
kafka_result = producer.send("enterpriseInfo", json.dumps(post_data, ensure_ascii=False).encode('utf8'))
print(kafka_result.get(timeout=10))
except:
......
......@@ -405,7 +405,7 @@ if __name__ == '__main__':
log.info(f'采集{com_name}成功=======耗时{baseCore.getTimeCost(start_time, time.time())}')
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], api_version=(2, 0, 2))
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], api_version=(2, 0, 2))
kafka_result = producer.send("regionInfo", json.dumps(post_data, ensure_ascii=False).encode('utf8'))
print(kafka_result.get(timeout=10))
except:
......
......@@ -94,7 +94,7 @@ def login():
import pymongo
# 连接数据库
db_storage = pymongo.MongoClient('mongodb://114.115.221.202:27017/', username='admin', password='ZZsn@9988').ZZSN[
db_storage = pymongo.MongoClient('mongodb://1.95.69.135:27017/', username='admin', password='ZZsn@9988').ZZSN[
'股东信息']
# 从数据库存储
......
......@@ -121,7 +121,7 @@ if __name__=='__main__':
}
print(com_dict)
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], api_version=(2, 0, 2))
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], api_version=(2, 0, 2))
kafka_result = producer.send("regionInfo", json.dumps(com_dict, ensure_ascii=False).encode('utf8'))
log.info(kafka_result.get(timeout=10))
log.info(f'{cik}---{cname}---基本信息采集成功')
......
......@@ -11,7 +11,7 @@ from kafka import KafkaProducer
import urllib3
from selenium.webdriver.support.wait import WebDriverWait
db_storage = pymongo.MongoClient('mongodb://114.115.221.202:27017/', username='admin', password='ZZsn@9988').ZZSN[
db_storage = pymongo.MongoClient('mongodb://1.95.69.135:27017/', username='admin', password='ZZsn@9988').ZZSN[
'天眼查登录信息']
urllib3.disable_warnings(urllib3.exceptions.InsecureRequestWarning)
from dateutil.relativedelta import relativedelta
......@@ -47,7 +47,7 @@ def create_driver():
# 发送数据
def sendkafka(post_data):
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], api_version=(2, 0, 2))
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], api_version=(2, 0, 2))
kafka_result = producer.send("enterpriseInfo", json.dumps(post_data, ensure_ascii=False).encode('utf8'))
print(kafka_result.get(timeout=10))
except:
......
......@@ -12,7 +12,7 @@ from selenium.webdriver.edge.service import Service
import urllib3
from selenium.webdriver.support.wait import WebDriverWait
db_storage = pymongo.MongoClient('mongodb://114.115.221.202:27017/', username='admin', password='ZZsn@9988').ZZSN[
db_storage = pymongo.MongoClient('mongodb://1.95.69.135:27017/', username='admin', password='ZZsn@9988').ZZSN[
'天眼查登录信息']
urllib3.disable_warnings(urllib3.exceptions.InsecureRequestWarning)
from dateutil.relativedelta import relativedelta
......@@ -54,7 +54,7 @@ def create_driver():
# 发送数据
def sendkafka(post_data):
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], api_version=(2, 0, 2))
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], api_version=(2, 0, 2))
kafka_result = producer.send("enterpriseInfo", json.dumps(post_data, ensure_ascii=False).encode('utf8'))
print(kafka_result.get(timeout=10))
except:
......
......@@ -13,7 +13,7 @@ from kafka import KafkaProducer
import urllib3
from retry import retry
from selenium.webdriver.support.wait import WebDriverWait
db_storage = pymongo.MongoClient('mongodb://114.115.221.202:27017/', username='admin', password='ZZsn@9988').ZZSN[
db_storage = pymongo.MongoClient('mongodb://1.95.69.135:27017/', username='admin', password='ZZsn@9988').ZZSN[
'天眼查登录信息']
urllib3.disable_warnings(urllib3.exceptions.InsecureRequestWarning)
......@@ -52,7 +52,7 @@ def create_driver():
# 发送数据
def sendkafka(post_data):
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], api_version=(2, 0, 2))
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], api_version=(2, 0, 2))
kafka_result = producer.send("enterpriseInfo", json.dumps(post_data, ensure_ascii=False).encode('utf8'))
print(kafka_result.get(timeout=10))
except:
......
......@@ -11,7 +11,7 @@ from kafka import KafkaProducer
import urllib3
from selenium.webdriver.support.wait import WebDriverWait
db_storage = pymongo.MongoClient('mongodb://114.115.221.202:27017/', username='admin', password='ZZsn@9988').ZZSN[
db_storage = pymongo.MongoClient('mongodb://1.95.69.135:27017/', username='admin', password='ZZsn@9988').ZZSN[
'天眼查登录信息']
urllib3.disable_warnings(urllib3.exceptions.InsecureRequestWarning)
......@@ -50,7 +50,7 @@ def create_driver():
# 发送数据
def sendkafka(post_data):
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], api_version=(2, 0, 2))
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], api_version=(2, 0, 2))
kafka_result = producer.send("enterpriseInfo", json.dumps(post_data, ensure_ascii=False).encode('utf8'))
print(kafka_result.get(timeout=10))
except:
......
......@@ -14,9 +14,9 @@ baseCore = BaseCore.BaseCore()
log = baseCore.getLogger()
cnx = baseCore.cnx
cursor = baseCore.cursor
db_storage = pymongo.MongoClient('mongodb://114.115.221.202:27017/', username='admin', password='ZZsn@9988').ZZSN[
db_storage = pymongo.MongoClient('mongodb://1.95.69.135:27017/', username='admin', password='ZZsn@9988').ZZSN[
'天眼查登录信息']
db_storage2 = pymongo.MongoClient('mongodb://114.115.221.202:27017/', username='admin', password='ZZsn@9988').ZZSN[
db_storage2 = pymongo.MongoClient('mongodb://1.95.69.135:27017/', username='admin', password='ZZsn@9988').ZZSN[
'最大股东信息0902']
......
......@@ -225,7 +225,7 @@ def beinWork(tyc_code, social_code):
baseCore.recordLog(social_code, taskType, state, takeTime, link, e)
continue
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
kafka_result = producer.send("researchReportDynamicTopic",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
......@@ -5,7 +5,7 @@ from selenium import webdriver
import pymongo
from selenium.webdriver.common.by import By
db_storage = pymongo.MongoClient('mongodb://114.115.221.202:27017/', username='admin', password='ZZsn@9988').ZZSN[
db_storage = pymongo.MongoClient('mongodb://1.95.69.135:27017/', username='admin', password='ZZsn@9988').ZZSN[
'天眼查登录信息']
url = 'https://www.tianyancha.com/'
......
......@@ -255,7 +255,7 @@ def beinWork(tyc_code, social_code,start_time):
# print(dic_news)
# 将相应字段通过kafka传输保存
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
kafka_result = producer.send("researchReportDynamicTopic",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
......@@ -265,7 +265,7 @@ def beinWork(tyc_code, social_code, start_time):
# print(dic_news)
# 将相应字段通过kafka传输保存
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
kafka_result = producer.send("researchReportDynamicTopic",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
......@@ -246,7 +246,7 @@ def beinWork(tyc_code, social_code,start_time):
# print(dic_news)
# 将相应字段通过kafka传输保存
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
kafka_result = producer.send("researchReportDynamicTopic",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
......@@ -241,7 +241,7 @@ def download(data, order_by,header):
'type': '0'
}
# 将相应字段通过kafka传输保存
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
try:
kafka_result = producer.send("researchReportStudyTopic", json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
log.info(kafka_result.get(timeout=10))
......@@ -321,7 +321,7 @@ def download(data, order_by,header):
# log.info(dic_news)
# 将相应字段通过kafka传输保存
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
kafka_result = producer.send("researchReportStudyTopic", json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
log.info(kafka_result.get(timeout=10))
......
......@@ -18,7 +18,7 @@ cursor = baseCore.cursor
def sendKafka(dic_news):
try: # 114.116.116.241
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'],max_request_size=1024*1024*20)
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'],max_request_size=1024*1024*20)
kafka_result = producer.send("researchReportYearTopic",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
......@@ -18,7 +18,7 @@ cursor = baseCore.cursor
def sendKafka(dic_news):
try: # 114.116.116.241
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'],max_request_size=1024*1024*20)
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'],max_request_size=1024*1024*20)
kafka_result = producer.send("researchReportYearTopic",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
import json
import json
......@@ -196,7 +196,7 @@ def SpiderByZJH(url, payload, dic_info, num, start_time):
# print(dic_news)
# 将相应字段通过kafka传输保存
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'],max_request_size=1024*1024*20)
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'],max_request_size=1024*1024*20)
kafka_result = producer.send("researchReportYearTopic",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
# -*- coding: utf-8 -*-
# -*- coding: utf-8 -*-
......@@ -225,7 +225,7 @@ def spider_annual_report(dict_info,num):
}
# 将相应字段通过kafka传输保存
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'],max_request_size=1024*1024*20)
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'],max_request_size=1024*1024*20)
kafka_result = producer.send("researchReportYearTopic",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
......@@ -35,7 +35,7 @@ log.info(f'=============当前pid为{baseCore.getPID()}==============')
def sendKafka(dic_news):
start_time = time.time()
try: # 114.116.116.241
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'],max_request_size=1024*1024*20)
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'],max_request_size=1024*1024*20)
kafka_result = producer.send("researchReportYearTopic",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
......@@ -40,7 +40,7 @@ log.info(f'=============当前pid为{baseCore.getPID()}==============')
def sendKafka(dic_news):
start_time = time.time()
try: # 114.116.116.241
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'],max_request_size=1024*1024*20)
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'],max_request_size=1024*1024*20)
kafka_result = producer.send("researchReportYearTopic",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
......@@ -35,7 +35,7 @@ taskType = '企业年报'
def sendKafka(dic_news):
start_time = time.time()
try: # 114.116.116.241
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
kafka_result = producer.send("researchReportYearTopic",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
......@@ -35,7 +35,7 @@ file_path = 'D:/kkwork/zzsn_spider/data/1_福布斯2000_PDF_50_郑'
def sendKafka(dic_news):
start_time = time.time()
try: # 114.116.116.241
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
kafka_result = producer.send("researchReportYearTopic",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
......@@ -44,7 +44,7 @@ taskType = '企业年报'
def sendKafka(dic_news,xydm):
start_time = time.time()
try: # 114.116.116.241
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], max_request_size=1024 * 1024 * 20)
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], max_request_size=1024 * 1024 * 20)
kafka_result = producer.send("researchReportYearTopic",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
......@@ -30,7 +30,7 @@ pathType = 'QYYearReport/'
def sendKafka(dic_news):
try: # 114.116.116.241
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'],max_request_size=1024*1024*20)
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'],max_request_size=1024*1024*20)
kafka_result = producer.send("researchReportYearTopic",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
......@@ -197,7 +197,7 @@ def spider(com_name,cik,up_okCount):
# print(dic_news)
# 将相应字段通过kafka传输保存
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'],compression_type='gzip',batch_size=1638400,linger_ms=1,buffer_memory=33445532*2,max_request_size=1024*1024*50) #,batch_size=20480000,buffer_memory=64000000)
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'],compression_type='gzip',batch_size=1638400,linger_ms=1,buffer_memory=33445532*2,max_request_size=1024*1024*50) #,batch_size=20480000,buffer_memory=64000000)
kafka_result = producer.send("researchReportYearTopic",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
......@@ -11,7 +11,7 @@ database=caiji
url=jdbc:mysql://114.115.159.144:3306/caiji?useUnicode=true&characterEncoding=utf-8&serverTimezone=Asia/Shanghai&useSSL=false
[kafka]
bootstrap_servers=114.115.159.144:9092
bootstrap_servers=1.95.3.121:9092
topic=keyWordsInfo
groupId=python_baidu
......
......@@ -167,7 +167,7 @@ def sendTokafka(ddata):
'type': ''
}
log.info(aa_dict)
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
try:
kafka_result = producer.send("crawlerInfo", json.dumps(aa_dict, ensure_ascii=False).encode('utf8'))
r.sadd(info_code+'-test',sourceAddress)
......
......@@ -127,7 +127,7 @@ def two_dfsm_mtgc():
}
log.info(f'{page}--{title}--{href}')
# info_list.append(result_dict)
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
try:
kafka_result = producer.send("crawlerInfo",
json.dumps(result_dict, ensure_ascii=False).encode('utf8'))
......
......@@ -21,7 +21,7 @@ db_storage = pymongo.MongoClient('mongodb://114.115.221.202:27017', username='ad
@retry(tries=2, delay=5)
def sendKafka(dic):
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], max_request_size=1024 * 1024 * 20)
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], max_request_size=1024 * 1024 * 20)
kafka_result = producer.send("research_center_fourth",
json.dumps(dic, ensure_ascii=False).encode('utf8'))
log.info(f'{dic["sourceAddress"]}传输成功')
......
......@@ -128,7 +128,7 @@ def spider_work(href):
'labels': [],
}
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
kafka_result = producer.send("crawlerInfo", json.dumps(aa_dict, ensure_ascii=False).encode('utf8'))
print(kafka_result.get(timeout=10))
#print(aa_dict)
......
......@@ -138,7 +138,7 @@ def gzyw():
}
log.info(f'{page}--{title}--{href}')
# info_list.append(result_dict)
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
try:
kafka_result = producer.send("crawlerInfo",
json.dumps(result_dict, ensure_ascii=False).encode('utf8'))
......
......@@ -142,7 +142,7 @@ def sendTokafka(ddata):
'source': 'python定制采集',
'type': ''
}
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
try:
kafka_result = producer.send("crawlerInfo", json.dumps(aa_dict, ensure_ascii=False).encode('utf8'))
r.sadd(info_code+'-test',sourceAddress)
......
......@@ -188,7 +188,7 @@ if __name__ == "__main__":
# 将相应字段通过kafka传输保存
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
kafka_result = producer.send("crawlerInfo",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
......@@ -149,7 +149,7 @@ def getData(key):
# 将相应字段通过kafka传输保存
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
kafka_result = producer.send("crawlerInfo",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
......@@ -230,7 +230,7 @@ def zzcx():
print(result_dict)
# break
# break
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
try:
kafka_result = producer.send("crawlerInfo",
json.dumps(result_dict, ensure_ascii=False).encode('utf8'))
......
......@@ -49,7 +49,7 @@ def is_member_containing_string(key, string):
def sendKafka(dic_info):
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
try:
kafka_result = producer.send("research_center_fourth",
json.dumps(dic_info, ensure_ascii=False).encode('utf8'))
......
......@@ -150,7 +150,7 @@ if __name__ == "__main__":
}
r.sadd(info_code, newsUrl)
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
try:
kafka_result = producer.send("research_center_fourth",
json.dumps(dic_info, ensure_ascii=False).encode('utf8'))
......
......@@ -16,7 +16,7 @@ headers = {
def sendKafka(dic_news):
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
kafka_result = producer.send("leadership",
json.dumps(dic_news, ensure_ascii=False).encode('utf-8'))
......
......@@ -186,7 +186,7 @@ def getDicB(data, soup):
# 数据发送至Kafka
@retry(tries=3, delay=1)
def sendKafka(dic_news, start_time):
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
kafka_result = producer.send("researchReportTopic",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
......@@ -57,7 +57,7 @@ def getNowDate():
def sendkafka(processitem):
try:
producer = KafkaProducer(bootstrap_servers=["114.115.159.144:9092"])
producer = KafkaProducer(bootstrap_servers=["1.95.3.121:9092"])
content=processitem['content']
publishDate=str(processitem['publishDate'])
title=processitem['title']
......
......@@ -44,7 +44,7 @@ def getNowDate():
def sendKafka(dic_news):
start_time = time.time()
try: # 114.116.116.241
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'],max_request_size=1024*1024*20)
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'],max_request_size=1024*1024*20)
kafka_result = producer.send("crawlerInfo",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
......@@ -408,7 +408,7 @@ if __name__ == '__main__':
log.info(f'采集{com_name}成功=======耗时{baseCore.getTimeCost(start_time,time.time())}')
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], api_version=(2, 0, 2))
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], api_version=(2, 0, 2))
kafka_result = producer.send("regionInfo", json.dumps(post_data, ensure_ascii=False).encode('utf8'))
print(kafka_result.get(timeout=10))
except:
......
......@@ -416,7 +416,7 @@ if __name__ == '__main__':
baseCore.writerToExcel(name_list,'中国100强企业.xlsx')
log.info(f'采集{com_name}成功=======耗时{baseCore.getTimeCost(start_time,time.time())}')
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], api_version=(2, 0, 2))
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], api_version=(2, 0, 2))
kafka_result = producer.send("regionInfo", json.dumps(post_data, ensure_ascii=False).encode('utf8'))
print(kafka_result.get(timeout=10))
except:
......
......@@ -416,7 +416,7 @@ if __name__ == '__main__':
baseCore.writerToExcel(name_list,'跨国公司100大.xlsx')
log.info(f'采集{com_name}成功=======耗时{baseCore.getTimeCost(start_time,time.time())}')
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], api_version=(2, 0, 2))
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], api_version=(2, 0, 2))
kafka_result = producer.send("regionInfo", json.dumps(post_data, ensure_ascii=False).encode('utf8'))
print(kafka_result.get(timeout=10))
except:
......
......@@ -408,7 +408,7 @@ if __name__ == '__main__':
log.info(f'采集{com_name}成功=======耗时{baseCore.getTimeCost(start_time,time.time())}')
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], api_version=(2, 0, 2))
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], api_version=(2, 0, 2))
kafka_result = producer.send("regionInfo", json.dumps(post_data, ensure_ascii=False).encode('utf8'))
print(kafka_result.get(timeout=10))
except:
......
......@@ -408,7 +408,7 @@ if __name__ == '__main__':
log.info(f'采集{com_name}成功=======耗时{baseCore.getTimeCost(start_time,time.time())}')
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], api_version=(2, 0, 2))
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], api_version=(2, 0, 2))
kafka_result = producer.send("regionInfo", json.dumps(post_data, ensure_ascii=False).encode('utf8'))
print(kafka_result.get(timeout=10))
except:
......
......@@ -217,7 +217,7 @@ def GetContent(pdf_url, pdf_name, social_code, year, pub_time, start_time):
# print(dic_news)
# 将相应字段通过kafka传输保存
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
kafka_result = producer.send("researchReportTopic", json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
print(kafka_result.get(timeout=10))
......
import os
import os
......@@ -250,7 +250,7 @@ def GetContent(pdf_url,info_url, pdf_name, social_code, year, pub_time, start_ti
# print(dic_news)
# 将相应字段通过kafka传输保存
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'],max_request_size=1024*1024*20)
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'],max_request_size=1024*1024*20)
kafka_result = producer.send("researchReportNoticeTopic",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
"""
"""
......@@ -263,7 +263,7 @@ def GetContent(pdf_url,info_url, pdf_name, social_code, year, pub_time, start_ti
# print(dic_news)
# 将相应字段通过kafka传输保存
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'],max_request_size=1024*1024*20)
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'],max_request_size=1024*1024*20)
kafka_result = producer.send("researchReportNoticeTopic",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
import os
import os
......@@ -218,7 +218,7 @@ def ifInstert(short_name, social_code, pdf_url):
def sendKafka(social_code,newsUrl,dic_news):
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], max_request_size=1024*1024*20)
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], max_request_size=1024*1024*20)
kafka_result = producer.send("researchReportNoticeTopic",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
......@@ -352,7 +352,7 @@ def GetContent(pdf_url, pdf_name, social_code, year, pub_time, start_time,com_na
# print(dic_news)
# 将相应字段通过kafka传输保存
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], max_request_size=1024*1024*20)
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], max_request_size=1024*1024*20)
kafka_result = producer.send("researchReportNoticeTopic", json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
print(kafka_result.get(timeout=10))
......
......@@ -93,7 +93,7 @@ class ClassTool():
def sendKafka(self, dic_news):
try: # 114.116.116.241
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], max_request_size=1024 * 1024 * 20)
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], max_request_size=1024 * 1024 * 20)
kafka_result = producer.send("policy",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
......@@ -208,7 +208,7 @@ def getDic(driver):
@retry(tries=2, delay=5)
def sendKafka(dic):
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], max_request_size=1024 * 1024 * 20)
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], max_request_size=1024 * 1024 * 20)
kafka_result = producer.send("research_center_fourth",
json.dumps(dic, ensure_ascii=False).encode('utf8'))
log.info(f'{dic["sourceAddress"]}传输成功')
......
......@@ -64,7 +64,7 @@ def save_data(dic_news):
def sendKafka(dic_news):
start_time = time.time()
try: # 114.116.116.241
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
kafka_result = producer.send("policy",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
......@@ -118,7 +118,7 @@ def getDic(social_code, li):
# 数据发送至Kafka
@retry(tries=3, delay=1)
def sendKafka(dic_news, start_time):
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
kafka_result = producer.send("researchReportTopic",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
......@@ -141,7 +141,7 @@ def getDic(social_code, title, href, pub_time):
# 数据发送至Kafka
@retry(tries=3, delay=1)
def sendKafka(dic_news, start_time):
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
kafka_result = producer.send("researchReportTopic",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
......@@ -149,7 +149,7 @@ def getDic(social_code, title, href, pub_time):
# 数据发送至Kafka
@retry(tries=3, delay=1)
def sendKafka(dic_news, start_time):
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
kafka_result = producer.send("researchReportTopic",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
......@@ -37,7 +37,7 @@ def ifInstert(social_code, pdf_url):
@retry(tries=3, delay=1)
def sendKafka(dic_news):
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
kafka_result = producer.send("researchReportTopic", json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
print(kafka_result.get(timeout=10))
......
......@@ -37,7 +37,7 @@ session.mount('http://', HTTPAdapter(max_retries=3))
# 发送kafka
def sendKafka(dic_news):
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
kafka_result = producer.send("tenderClusterData",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
......@@ -20,7 +20,7 @@ URL = 'http://www.ccgp.gov.cn/'
def sendKafka(dic_news):
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
kafka_result = producer.send("tenderClusterData",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
......@@ -37,7 +37,7 @@ session.get(URL, headers=headers)
def sendKafka(dic_news):
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
kafka_result = producer.send("tenderClusterData",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
......@@ -260,7 +260,7 @@ def sendToKafka(detailmsg):
'socialCreditCode': detailmsg['socialCreditCode'],
'year': detailmsg['year']
}
producer=KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer=KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
try:
kafka_result = producer.send("researchReportTopic",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
......@@ -276,7 +276,7 @@ def sendToKafka(detailmsg):
'socialCreditCode': detailmsg['socialCreditCode'],
'year': detailmsg['year']
}
producer=KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer=KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
try:
kafka_result = producer.send("researchReportTopic",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
......@@ -241,7 +241,7 @@ def beinWork(tyc_code, social_code,start_time):
# print(dic_news)
# 将相应字段通过kafka传输保存
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
kafka_result = producer.send("researchReportTopic",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
......@@ -325,7 +325,7 @@ class Tycdt(object):
# print(dic_news)
# 将相应字段通过kafka传输保存
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
kafka_result = producer.send("researchReportTopic",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
......@@ -11,7 +11,7 @@ database=caiji
url=jdbc:mysql://114.115.159.144:3306/caiji?useUnicode=true&characterEncoding=utf-8&serverTimezone=Asia/Shanghai&useSSL=false
[kafka]
bootstrap_servers=114.115.159.144:9092
bootstrap_servers=1.95.3.121:9092
topic=keyWordsInfo
groupId=python_baidu
......
......@@ -74,8 +74,8 @@ if __name__ == "__main__":
# loadinfo = [token,cookies]
# 保存到数据库中
# insert = f"insert into weixin_tokenCookies_person (token,cookies,create_time,fenghao_time,user_name,update_time) values ('{token}','{escape_string(cookies)}',now(),DATE_SUB(NOW(), INTERVAL 1 DAY),'{user_name}',now())"
insert = f"insert into weixin_tokenCookies (token,cookies,create_time,fenghao_time,user_name,update_time) values ('{token}','{escape_string(cookies)}',now(),DATE_SUB(NOW(), INTERVAL 1 DAY),'{user_name}',now())"
insert = f"insert into weixin_tokenCookies_person (token,cookies,create_time,fenghao_time,user_name,update_time) values ('{token}','{escape_string(cookies)}',now(),DATE_SUB(NOW(), INTERVAL 1 DAY),'{user_name}',now())"
# insert = f"insert into weixin_tokenCookies (token,cookies,create_time,fenghao_time,user_name,update_time) values ('{token}','{escape_string(cookies)}',now(),DATE_SUB(NOW(), INTERVAL 1 DAY),'{user_name}',now())"
cursor_.execute(insert)
cnx_.commit()
browser.close()
......
......@@ -177,7 +177,7 @@ def get_info(sid,json_search,origin,url_,info_source_code,page):
'createDate': time_now
}
# for nnn in range(0, 3):
# producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
# producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
# try:
# kafka_result = producer.send("crawlerInfo", json.dumps(dic_info, ensure_ascii=False).encode('utf8'))
# kafka_time_out = kafka_result.get(timeout=10)
......@@ -202,9 +202,9 @@ def get_info(sid,json_search,origin,url_,info_source_code,page):
# 'source': '1',
# }
# for nnn2 in range(0, 3):
# producer2 = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
# producer2 = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
# try:
# # producer2 = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
# # producer2 = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
# kafka_result2 = producer2.send("collectionAndDispatcherInfo",
# json.dumps(dic_info2, ensure_ascii=False).encode('utf8'))
# break
......
......@@ -243,7 +243,7 @@ def get_info(dict_json, linkid):
}
for nnn in range(0, 3):
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], api_version=(2, 7, 0))
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], api_version=(2, 7, 0))
kafka_result = producer.send("crawlerInfo", json.dumps(dic_info, ensure_ascii=False).encode('utf8'))
kafka_time_out = kafka_result.get(timeout=10)
# add_url(sid, url_news)
......@@ -267,7 +267,7 @@ def get_info(dict_json, linkid):
}
for nnn2 in range(0, 3):
try:
producer2 = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], api_version=(2,7,0))
producer2 = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], api_version=(2,7,0))
kafka_result2 = producer2.send("collectionAndDispatcherInfo",
json.dumps(dic_info2, ensure_ascii=False).encode('utf8'))
break
......
......@@ -246,7 +246,7 @@ def get_info(dict_json, linkid):
log.info(dic_info)
# for nnn in range(0, 3):
# try:
# producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], api_version=(2, 7, 0))
# producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], api_version=(2, 7, 0))
# kafka_result = producer.send("crawlerInfo_test", json.dumps(dic_info, ensure_ascii=False).encode('utf8'))
# kafka_time_out = kafka_result.get(timeout=10)
# # add_url(sid, url_news)
......@@ -270,7 +270,7 @@ def get_info(dict_json, linkid):
# }
# for nnn2 in range(0, 3):
# try:
# producer2 = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], api_version=(2,7,0))
# producer2 = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], api_version=(2,7,0))
# kafka_result2 = producer2.send("collectionAndDispatcherInfo",
# json.dumps(dic_info2, ensure_ascii=False).encode('utf8'))
# break
......
# -*- coding: utf-8 -*-
# -*- coding: utf-8 -*-
......@@ -19,7 +19,7 @@ sys.path.append('D:\\zzsn_spider\\base')
import BaseCore
# todo 连接mongo
db_storage = pymongo.MongoClient('mongodb://114.115.221.202:27017/', username='admin', password='ZZsn@9988').ZZSN[
db_storage = pymongo.MongoClient('mongodb://1.95.69.135:27017/', username='admin', password='ZZsn@9988').ZZSN[
'机器人分会']
......@@ -37,7 +37,7 @@ rMonitor = redis.Redis(host='114.116.90.53', port=6380, password='clbzzsn', db=1
@retry(tries=2, delay=5)
def sendMonitorKafka(dic_news):
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], api_version=(2, 7, 0))
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], api_version=(2, 7, 0))
kafka_result = producer.send("data_lifecycle_log_data_crawler",
json.dumps(dic_news, ensure_ascii=False).encode('utf-8'))
log.info('数据监控Kafka发送成功')
......@@ -45,7 +45,7 @@ def sendMonitorKafka(dic_news):
@retry(tries=3, delay=5)
def sendKafka(result_dict):
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], api_version=(2, 7, 0))
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], api_version=(2, 7, 0))
kafka_result = producer.send("crawlerInfo",
json.dumps(result_dict, ensure_ascii=False).encode('utf8'))
log.info(f"{result_dict['title']}===发送kafka成功!")
......
......@@ -40,7 +40,7 @@ headers = {
# 发送kafka
@retry(delay=5)
def sendKafka(company_dict):
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], api_version=(2, 0, 2))
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], api_version=(2, 0, 2))
# kafka_result = producer.send("regionInfo", json.dumps(company_dict, ensure_ascii=False).encode('utf8'))
kafka_result = producer.send("enterpriseInfo", json.dumps(company_dict, ensure_ascii=False).encode('utf8'))
kafka_result.get(timeout=10)
......
# 雅虎财经企业动态获取
# 雅虎财经企业动态获取
......@@ -30,7 +30,7 @@ last_url = ''
# 发送kafka
@retry(tries=3, delay=5)
def sendKafka(dic_news):
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
kafka_result = producer.send("researchReportTopic",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
from kafka import KafkaConsumer
from kafka import KafkaConsumer
......@@ -270,7 +270,7 @@ def get_content_by_user_uid(url, sid):
# for one_news_info in list_all_info: # 将每一个文章数据转换为json格式,把json文件用kafka发送出去
# for num_pro in range(0, 3):
# try:
# producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'])
# producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'])
# kafka_result = producer.send("crawlerInfo",
# json.dumps(one_news_info, ensure_ascii=False).encode('utf8'))
# print(kafka_result.get(timeout=10))
......@@ -288,8 +288,8 @@ def get_content_by_user_uid(url, sid):
def consume():
"""auto_commit_enable=True, auto_commit_interval_ms=3000"""
consumer = KafkaConsumer("weiBoCrawl", auto_offset_reset='earliest', group_id="python_weibo",
bootstrap_servers=['114.115.159.144:9092'])
# consumer = KafkaConsumer("pythonInfo", auto_offset_reset='earliest', bootstrap_servers=['114.115.159.144:9092'])
bootstrap_servers=['1.95.3.121:9092'])
# consumer = KafkaConsumer("pythonInfo", auto_offset_reset='earliest', bootstrap_servers=['1.95.3.121:9092'])
for message in consumer:
mes_dict = json.loads(message.value.decode('utf-8'))
# print(message.value.decode('utf-8'))
......
import json
import json
......@@ -6,7 +6,7 @@ import pymongo
# url = "https://web.archive.org/web/20230702131549/https://www.forbes.com/lists/global2000/"
url = "https://web.archive.org/web/20220929184024/https://www.forbes.com/lists/global2000/"
db_storage = pymongo.MongoClient('mongodb://114.115.221.202:27017/', username='admin', password='ZZsn@9988').ZZSN[
db_storage = pymongo.MongoClient('mongodb://1.95.69.135:27017/', username='admin', password='ZZsn@9988').ZZSN[
'福布斯企业人数']
headers = {
'Accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,image/apng,*/*;q=0.8,application/signed-exchange;v=b3;q=0.7',
......
import json
import json
......@@ -6,7 +6,7 @@ import requests
from bs4 import BeautifulSoup
from retry import retry
db_storage = pymongo.MongoClient('mongodb://114.115.221.202:27017/', username='admin', password='ZZsn@9988').ZZSN[
db_storage = pymongo.MongoClient('mongodb://1.95.69.135:27017/', username='admin', password='ZZsn@9988').ZZSN[
'2022年福布斯企业人数']
url = 'https://web.archive.org/web/20220929184024/https://www.forbes.com/lists/global2000/'
headers = {
......
import base64
import base64
......@@ -228,7 +228,7 @@ def getProcessitem(bdetail):
def sendkafka(processitem):
try:
producer = KafkaProducer(bootstrap_servers="114.115.159.144:9092")
producer = KafkaProducer(bootstrap_servers="1.95.3.121:9092")
content=processitem['content']
publishDate=str(processitem['publishDate'])
title=processitem['title']
......
#coding=utf-8
#coding=utf-8
......@@ -257,7 +257,7 @@ def getProcessitem(bdetail):
def sendkafka(processitem):
try:
producer = KafkaProducer(bootstrap_servers="114.115.159.144:9092")
producer = KafkaProducer(bootstrap_servers="1.95.3.121:9092")
content=processitem['content']
publishDate=str(processitem['publishDate'])
title=processitem['title']
......
......@@ -17,7 +17,7 @@ baseCore = BaseCore()
log = baseCore.getLogger()
cnx = baseCore.cnx
cursor = baseCore.cursor
db_storage = pymongo.MongoClient('mongodb://114.115.221.202:27017/', username='admin', password='ZZsn@9988').ZZSN[
db_storage = pymongo.MongoClient('mongodb://1.95.69.135:27017/', username='admin', password='ZZsn@9988').ZZSN[
'天眼查登录信息']
......
import json
import json
......@@ -18,7 +18,7 @@ from urllib.parse import unquote
baseCore = BaseCore.BaseCore(sqlFlg=False)
log = baseCore.getLogger()
db_storage = pymongo.MongoClient('mongodb://114.115.221.202:27017/', username='admin', password='ZZsn@9988').ZZSN[
db_storage = pymongo.MongoClient('mongodb://1.95.69.135:27017/', username='admin', password='ZZsn@9988').ZZSN[
'新华丝路-丝路商机100+']
lock = threading.Lock()
......
import json
import json
......@@ -17,7 +17,7 @@ from urllib.parse import unquote
baseCore = BaseCore.BaseCore(sqlFlg=False)
log = baseCore.getLogger()
db_storage = pymongo.MongoClient('mongodb://114.115.221.202:27017/', username='admin', password='ZZsn@9988').ZZSN[
db_storage = pymongo.MongoClient('mongodb://1.95.69.135:27017/', username='admin', password='ZZsn@9988').ZZSN[
'7-17全球企业资讯删除数据']
lock = threading.Lock()
......
import json
import json
......@@ -17,7 +17,7 @@ from urllib.parse import unquote
baseCore = BaseCore.BaseCore(sqlFlg=False)
log = baseCore.getLogger()
db_storage = pymongo.MongoClient('mongodb://114.115.221.202:27017/', username='admin', password='ZZsn@9988').ZZSN[
db_storage = pymongo.MongoClient('mongodb://1.95.69.135:27017/', username='admin', password='ZZsn@9988').ZZSN[
'新华丝路-丝路商机']
lock = threading.Lock()
......
......@@ -11,7 +11,7 @@ database=caiji
url=jdbc:mysql://114.115.159.144:3306/caiji?useUnicode=true&characterEncoding=utf-8&serverTimezone=Asia/Shanghai&useSSL=false
[kafka]
bootstrap_servers=114.115.159.144:9092
bootstrap_servers=1.95.3.121:9092
topic=keyWordsInfo
groupId=python_google
......
......@@ -21,7 +21,7 @@ db_storage = pymongo.MongoClient('mongodb://114.115.221.202:27017', username='ad
@retry(tries=2, delay=5)
def sendKafka(dic):
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], max_request_size=1024 * 1024 * 20)
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], max_request_size=1024 * 1024 * 20)
kafka_result = producer.send("research_center_fourth",
json.dumps(dic, ensure_ascii=False).encode('utf8'))
log.info(f'{dic["sourceAddress"]}传输成功')
......
......@@ -40,7 +40,7 @@ headers = {
@retry(tries=2, delay=5)
def sendKafka(dic):
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], max_request_size=1024 * 1024 * 20)
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], max_request_size=1024 * 1024 * 20)
kafka_result = producer.send("research_center_fourth",
json.dumps(dic, ensure_ascii=False).encode('utf8'))
log.info(f'{dic["sourceAddress"]}传输成功')
......
......@@ -37,7 +37,7 @@ headers = {
@retry(tries=2, delay=5)
def sendKafka(dic):
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], max_request_size=1024 * 1024 * 20)
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], max_request_size=1024 * 1024 * 20)
kafka_result = producer.send("research_center_fourth",
json.dumps(dic, ensure_ascii=False).encode('utf8'))
log.info(f'{dic["sourceAddress"]}传输成功')
......
......@@ -45,7 +45,7 @@ headers = {
@retry(tries=2, delay=5)
def sendKafka(dic):
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'], max_request_size=1024 * 1024 * 20)
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'], max_request_size=1024 * 1024 * 20)
kafka_result = producer.send("research_center_fourth",
json.dumps(dic, ensure_ascii=False).encode('utf8'))
log.info(f'{dic["sourceAddress"]}传输成功')
......
......@@ -11,7 +11,7 @@ database=caiji
url=jdbc:mysql://114.115.159.144:3306/caiji?useUnicode=true&characterEncoding=utf-8&serverTimezone=Asia/Shanghai&useSSL=false
[kafka]
bootstrap_servers=114.115.159.144:9092
bootstrap_servers=1.95.3.121:9092
topic=keyWordsInfo
groupId=python_sougou
......
......@@ -16,7 +16,7 @@ log = baseCore.getLogger()
r = redis.Redis(host="114.115.236.206", port=6379, password='clbzzsn', db=0)
def sendKafka(dic_news):
try:
producer = KafkaProducer(bootstrap_servers=['114.115.159.144:9092'],max_request_size=1024*1024*20)
producer = KafkaProducer(bootstrap_servers=['1.95.3.121:9092'],max_request_size=1024*1024*20)
kafka_result = producer.send("crawlerInfo",
json.dumps(dic_news, ensure_ascii=False).encode('utf8'))
......
......@@ -11,7 +11,7 @@ database=caiji
url=jdbc:mysql://114.115.159.144:3306/caiji?useUnicode=true&characterEncoding=utf-8&serverTimezone=Asia/Shanghai&useSSL=false
[kafka]
bootstrap_servers=114.115.159.144:9092
bootstrap_servers=1.95.3.121:9092
topic=keyWordsInfo
groupId=python_sougou
......
......@@ -11,7 +11,7 @@ database=caiji
url=jdbc:mysql://114.115.159.144:3306/caiji?useUnicode=true&characterEncoding=utf-8&serverTimezone=Asia/Shanghai&useSSL=false
[kafka]
bootstrap_servers=114.115.159.144:9092
bootstrap_servers=1.95.3.121:9092
topic=keyWordsInfo
groupId=python_sougou
......
......@@ -11,7 +11,7 @@ database=caiji
url=jdbc:mysql://114.115.159.144:3306/caiji?useUnicode=true&characterEncoding=utf-8&serverTimezone=Asia/Shanghai&useSSL=false
[kafka]
bootstrap_servers=114.115.159.144:9092
bootstrap_servers=1.95.3.121:9092
topic=keyWordsInfo
groupId=python_sougou
......
Markdown 格式
0%
您添加了 0 到此讨论。请谨慎行事。
请先完成此评论的编辑!
注册 或者 后发表评论