ͻ񻣼
Ò».»ù±¾º¯Êýµ÷ÓÃÓ붨Òå 1 package main 2 3 import ( 4 "encoding/json" 5 "errors" 6 "fmt" 7 "math/rand" 8 "mylib/pkg/student" 9 "mylib/pkg/utils" 10 "sort" 11 "str ÔĶÁÈ«ÎÄ
ͻ񻣼
Àý1 1 #! /bin/bash 2 read_dir(){ 3 a=0 4 declare -a arr 5 for file in `ls $1` 6 do 7 a=$a+1 8 if [ -d $1"/"$file ] 9 then 10 arr[$a]=$1"/"$file 11 echo ÔĶÁÈ«ÎÄ
ͻ񻣼
1.°²×°VS Code https://code.visualstudio.com/2.°²×°go https://golang.google.cn/3.ÅäÖÃgo»·¾³±äÁ¿ export PATH=/home/chad/golang/go/bin:$PATH # GO»·¾³µÄBinĿ¼ export GO ÔĶÁÈ«ÎÄ
ͻ񻣼
1 # °²×° 2 pip3 install scrapy_redis 3 # Ô´Âë 4 https://github.com/rmax/scrapy-redis.git 5 # Îĵµ 6 https://github.com/rmax/scrapy-redis 7 8 # ÅäÖÃ˵Ã÷: https:/ ÔĶÁÈ«ÎÄ
ͻ񻣼
SplashµÄ³õʶ 1 # SplashÀàËÆSeleniumÄ£Äâä¯ÀÀÆ÷ä¯ÀÀÍøÒ³À´¶¯Ì¬ÅÀÈ¡ÍøÕ¾ 2 # Îĵµ:https://splash.readthedocs.io/en/stable/ 3 # Splash ÔÚdockerÖа²×°: https://splash.readthedocs.io/en ÔĶÁÈ«ÎÄ
ͻ񻣼
Scrapy ³õʶ 1 Scrapy ±àÂëÁ÷³Ì 2 1.´´½¨ÅÀ³æ,Ò²¾ÍÊǼ̳Ðscrapy.Spider»òscrapy.CrawlSpider 3 2.½«ÒªÅÀÈ¡µÄÍøÒ³ÌîдÔÚstart_urlsÖÐ 4 3.ʵÏÖparse½âÎöÍøÕ¾µÄ»Øµ÷ 5 4.ÔÚItemsÖбàÂë¹æ¶¨Êý¾Ý¶ÔÏó 6 5.middleware ÔĶÁÈ«ÎÄ
ͻ񻣼
urllibÀ´ÊµÏÖcookieºÍip´úÀí 1 from urllib.request import Request, build_opener, urlopen 2 from fake_useragent import UserAgent 3 from urllib.parse import url ÔĶÁÈ«ÎÄ
ͻ񻣼
1 import requests 2 from time import sleep 3 from lxml import etree 4 from fake_useragent import UserAgent 5 from random import randint 6 import re 7 ÔĶÁÈ«ÎÄ
ͻ񻣼
ÏêϸʹÓð¸Àý XPathÓëjsonpath 1 import json 2 from lxml import etree 3 from jsonpath import jsonpath 4 5 def json_test(): 6 str1 = '{"name":"°£Àï¿ËÉ"}' 7 # ½«×Ö·û´®×ª ÔĶÁÈ«ÎÄ
ͻ񻣼
·þÎñ¶Ë 1 import logging 2 import json 3 from base64 import b64encode 4 from odoo import http, SUPERUSER_ID, _ 5 from odoo.http import request 6 7 _logger ÔĶÁÈ«ÎÄ