neglect
1月前来过
全职 · 300/日  ·  6525/月
工作时间: 工作日07:00-22:00、周末07:00-23:00工作地点: 远程
服务企业: 1家累计提交: 1工时
联系方式:
********
********
********
聊一聊

使用APP扫码聊一聊

个人介绍

一名大三考研学生学习计算机科学与技术专业正在努力中想要优秀就要拼命已经学习了java python c语言 c++等多种语言对于这方面甚是有兴趣,希望能在工作中学习到更多有用的知识完善自己 让自己变得优秀

工作经历

  • 2020-07-23 -2020-08-23喜士客服务员

    兢兢业业,从不迟到,服务态度很好。能够及时处理好特殊情况。在这份工作中,锻炼了自己的心性格,能吃苦耐闹,服从安排,会从不同方向努力去完成上级分布的任务

教育经历

  • 2020-09-01 - 2022-01-11安徽三联学院计算机科学与技术本科

    在这个学校中,我十分遵守学校规章制度,成绩在班级中较为优越,曾经差点拿到奖学金,惜败惜败

技能

C++
Android
MySQL
0
1
2
3
4
5
0
1
2
3
4
5
作品
逻辑与

.text .global _start _start: mov r3,r3 ldr r0,=myarray mov r4,#0 loop: ldr r1,[r0],#4 add r4,r4,r1 cmp r1,#0 bne loop stop: b stop myarray: .word 0x20 .word 0x01 .word 0x01 .word 0x01 .word 0x76 .word 0x00 .end

0
2022-01-11 11:23
Python爬虫评论

import csv from lxml import etree import requests import json import time #输入头文件 headers = { 'User-Agent': 'Mozilla/5.0 (Windows NT 6.1; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/64.0.3282.140 Safari/537.36', } #将获得的评论写入一个csv文件 f1 = open('知乎评论.csv', 'w', encoding='utf-8') writer = csv.writer(f1) writer.writerow(['id', 'created_time', 'author', 'content']) def A(): i = 0 while True: #通过f12找到网址 url = 'https://www.zhihu.com/api/v4/questions/20270540/answers' \ '?include=data[*].is_normal,admin_closed_comment,reward_info,is_collapsed,annotation_action,' \ 'annotation_detail,collapse_reason,is_sticky,collapsed_by,suggest_edit,comment_count,can_comment,' \ 'content,editable_content,attachment,voteup_count,reshipment_settings,comment_permission,created_time,' \ 'updated_time,review_info,relevant_info,question,excerpt,is_labeled,paid_info,paid_info_content,relationship.is_authorized,is_author,' \ 'voting,is_thanked,is_nothelp,is_recognized;data[*].mark_infos[*].url;data[*].author.follower_count,vip_info,badge[*].topics;data[*].settings.table_of_content.enabled&limit=5&offset=5&platform=desktop' 'sort_by=default'.format(i)#format(i)用作换页 state=1 while state: try: res = requests.get(url, headers=headers) state=0 except: continue res.encoding = 'utf-8' #用json解析text js = json.loads(res.text) is_end = js['paging']['is_end']#筛选信息 for i in js['data']: l = list() id_1 = str(i['id']) l.append(id_1) l.append(time.strftime("%Y-%m-%d %H:%M:%S", time.localtime(i['created_time']))) l.append(i['author']['name']) l.append(''.join(etree.HTML(i['content']).xpath('//p//text()'))) writer.writerow(l) print(l) if i['admin_closed_comment'] == False and i['can_comment']['status'] and i['comment_count'] > 0: B(id_1) i += 5 print('打印到第{0}页'.format(int(i / 5))) if is_end: break time.sleep(2) def B(id_1): j = 0 while 1: url = 'https://www.zhihu.com/api/v4/answers/{0}/root_comments?order=normal&limit=20&offset={1}&status=open'.format(id_1, j) a=1 while a: try: res = requests.get(url, headers=headers) a=0 except: continue res.encoding = 'utf-8' jsC = json.loads(res.text) is_end = jsC['paging']['is_end'] for data in jsC['data']: l = list() ID_2 = str(id_1) + "_" + str(data['id']) l.append(ID_2) l.append(time.strftime("%Y-%m-%d %H:%M:%S", time.localtime(data['created_time']))) l.append(data['author']['member']['name']) l.append(''.join(etree.HTML(data['content']).xpath('//p//text()'))) writer.writerow(l) print(l) for J in data['child_comments']: l.clear() l.append(str(ID_2) + "_" + str(J['id'])) l.append(time.strftime("%Y-%m-%d %H:%M:%S", time.localtime(J['created_time']))) l.append(J['author']['member']['name']) l.append(''.join(etree.HTML(J['content']).xpath('//p//text()'))) writer.writerow(l) print(l) j += 20 if is_end: break time.sleep(2) A() f1.close()

0
2023-02-24 08:32
Python实现各种算法

for i in range(1, 10): for j in range(1, i+1): print('{}x{}={}\t'.format(j, i, i*j), end='') print()

0
2023-02-24 08:32
更新于: 2021-08-23 浏览: 163