python利用xpath爬取网上数据并存储到django模型中

作者:cll_869241 时间:2022-01-19 03:46:51 

帮朋友制作一个网站,需要一些产品数据信息,因为是代理其他公司产品,直接爬取代理公司产品数据

1.设计数据库


from django.db import models
from uuslug import slugify
import uuid
import os

def products_directory_path(instance, filename):
 ext = filename.split('.')[-1]
 filename = '{}.{}'.format(uuid.uuid4().hex[:8], ext)
 # return the whole path to the file
 return os.path.join('images', "products", instance.title, filename)

def product_relatedimage_directory_path(instance, filename):
 ext = filename.split('.')[-1]
 filename = '{}.{}'.format(uuid.uuid4().hex[:8], ext)
 # return the whole path to the file
 return os.path.join('images', "product_relatedimage", instance.product.title, filename)

class ProductsCategory(models.Model):
 """产品分类"""
 name = models.CharField('产品分类名', max_length=80, unique=True)
 description = models.TextField('产品分类描述', blank=True, null=True)
 slug = models.SlugField('slug', max_length=80, blank=True, null=True)
 parent_category = models.ForeignKey('self', verbose_name="父级分类", blank=True, null=True, on_delete=models.CASCADE)

def save(self, *args, **kwargs):
   if not self.id or not self.slug:
     self.slug = slugify(self.name)
   super().save(*args, **kwargs)

def __str__(self):
   return self.name

class Meta:
   ordering = ['name']
   verbose_name = "产品分类"
   verbose_name_plural = verbose_name

class ProductsTag(models.Model):
 """产品标签"""
 name = models.CharField('产品标签名', max_length=30, unique=True)
 slug = models.SlugField('slug', max_length=40)

def __str__(self):
   return self.name

def save(self, *args, **kwargs):
   if not self.id or not self.slug:
     self.slug = slugify(self.name)
   super().save(*args, **kwargs)

class Meta:
   ordering = ['name']
   verbose_name = "产品标签"
   verbose_name_plural = verbose_name

class Product(models.Model):
 title = models.CharField('标题', max_length=255, unique=True)
 slug = models.SlugField('slug', max_length=255, blank=True, null=True)
 jscs = models.TextField('技术参数', blank=True, null=True)
 image = models.ImageField(upload_to=products_directory_path, verbose_name="产品图片")
 views = models.PositiveIntegerField('浏览量', default=0)
 category = models.ForeignKey('ProductsCategory', verbose_name='分类', on_delete=models.CASCADE, blank=True, null=True)
 tags = models.ManyToManyField('ProductsTag', verbose_name='标签集合', blank=True)

def save(self, *args, **kwargs):
   if not self.id or not self.slug:
     self.slug = slugify(self.title)
   super().save(*args, **kwargs)

def update_views(self):
   self.views += 1
   self.save(update_fields=['views'])

def get_pre(self):
   return Product.objects.filter(id__lt=self.id).order_by('-id').first()

def get_next(self):
   return Product.objects.filter(id__gt=self.id).order_by('id').first()

def __str__(self):
   return self.title

class Meta:
   verbose_name = "产品"
   verbose_name_plural = verbose_name

class ProductAdvantage(models.Model):
 content = models.TextField('产品优势', blank=True, null=True)
 product = models.ForeignKey(Product, on_delete=models.CASCADE, blank=True, null=True)

def __str__(self):
   return self.content

class Meta:
   verbose_name = "产品优势"
   verbose_name_plural = verbose_name

class ProductBody(models.Model):
 body = models.CharField('产品内容', max_length=256, blank=True, null=True)
 product = models.ForeignKey(Product, on_delete=models.CASCADE, blank=True, null=True)

def __str__(self):
   return self.product.title

class Meta:
   verbose_name = "产品内容"
   verbose_name_plural = verbose_name

2.脚本编写

2.1编写获取网页源代码函数


def get_one_page(url):
 try:
   headers = {
     "User-Agent": "Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/51.0.2704.103 Safari/537.36"}
   res = requests.get(url=url, headers=headers)
   res.encoding = 'utf-8'
   if res.status_code == 200:
     return res.text
   else:
     return None
 except Exception:
   return None

2.2根据base页面获取所有产品分类页面链接


if __name__ == '__main__':
 content = get_one_page(url)
 tree = etree.HTML(content)
 # 产品分类url
 catgory_urls = tree.xpath('//div[@class="fdh-01-nav"]/div/h3/a/@href')
 # 处理catgory_urls
 for url in catgory_urls:
   url = 'http://www.kexinjianji.com' + url
   print(url)

2.3根据产品分类页面链接获取对应所有产品链接


if __name__ == '__main__':
 content = get_one_page(url)
 tree = etree.HTML(content)
 # 产品分类
 catgory = tree.xpath('//div[@class="cplb-3n-ts-03 b"]/h3/span/text()')
 print("产品分类:" + catgory[0])
 # 该分类下产品url
 urls = tree.xpath('//div[@class="cplb-3n-ts-03-list"]/dl/dt/a/@href')
 # 处理url
 for url in urls:
   url = 'http://www.kexinjianji.com' + url
   print(url)
 print("=====================================================")

两者结合起来就可以打印出所有产品链接


if __name__ == '__main__':
 content = get_one_page(url)
 tree = etree.HTML(content)
 # 产品分类url
 catgory_urls = tree.xpath('//div[@class="fdh-01-nav"]/div/h3/a/@href')
 # 处理catgory_urls
 for url in catgory_urls:
   url = 'http://www.kexinjianji.com' + url
   content = get_one_page(url)
   tree = etree.HTML(content)
   # 产品分类
   catgory = tree.xpath('//div[@class="cplb-3n-ts-03 b"]/h3/span/text()')
   print("产品分类:" + catgory[0])
   # 该分类下产品url
   urls = tree.xpath('//div[@class="cplb-3n-ts-03-list"]/dl/dt/a/@href')
   # 处理url
   for url in urls:
     url = 'http://www.kexinjianji.com' + url
     print(url)
   print("=====================================================")

2.2使用xpath解析函数返回产品链接的内容


if __name__ == '__main__':
 content = get_one_page(url)
 tree = etree.HTML(content)
 # 产品名称
 title = tree.xpath('//*[@id="wrap"]//h1/text()')
 images = tree.xpath('//div[@class="sol_tj_left"]/a/img/@src')
 # 产品图片
 images_url = 'http://www.kexinjianji.com/' + images[0]
 # 性能特点
 xntd = tree.xpath('//div[@class="w"]//div/span/text()|//div[@class="w"]//div/text()')
 # 技术参数
 jscs = tree.xpath('//table')[0]
 jscs_str = etree.tostring(jscs, encoding='utf-8').decode('utf-8')
 # 产品内容
 cpnr = tree.xpath('//div[@class="describe"]/p')
 print('产品名称:' + title[0])
 print('产品图片:' + images_url)
 for td in xntd:
   print('性能特点:' + td)
 print('技术参数:' + jscs_str)
 for cp in cpnr:
   # string(.) 获取当前标签下所有文本内容
   cp = cp.xpath('string(.)')
   print('产品内容:' + cp)
 print('============================================')

将三者结合在一起就可以获取所有产品信息


if __name__ == '__main__':
 content = get_one_page(url)
 tree = etree.HTML(content)
 # 产品分类url
 catgory_urls = tree.xpath('//div[@class="fdh-01-nav"]/div/h3/a/@href')
 # 处理catgory_urls
 for url in catgory_urls:
   url = 'http://www.kexinjianji.com' + url
   content = get_one_page(url)
   tree = etree.HTML(content)
   # 产品分类
   catgory = tree.xpath('//div[@class="cplb-3n-ts-03 b"]/h3/span/text()')
   # 该分类下产品url
   urls = tree.xpath('//div[@class="cplb-3n-ts-03-list"]/dl/dt/a/@href')
   # 处理url
   for url in urls:
     url = 'http://www.kexinjianji.com' + url
     content = get_one_page(url)
     try:
       tree = etree.HTML(content)
       # 产品名称
       title = tree.xpath('//*[@id="wrap"]//h1/text()')
       images = tree.xpath('//div[@class="sol_tj_left"]/a/img/@src')
       # 产品图片
       images_url = 'http://www.kexinjianji.com' + images[0]
       # 性能特点
       xntd = tree.xpath('//div[@class="w"]//div/span/text()|//div[@class="w"]//div/text()')
       # 技术参数
       jscs = tree.xpath('//table')[0]
       jscs_str = etree.tostring(jscs, encoding='utf-8').decode('utf-8')
       # 产品内容
       cpnr = tree.xpath('//div[@class="describe"]/p')
       print("产品分类:" + catgory[0])
       print('产品链接:' + url)
       print('产品名称:' + title[0])
       print('产品图片:' + images_url)
       for td in xntd:
         print('性能特点:' + td.strip())
       # print('技术参数:' + jscs_str)
       for cp in cpnr:
         # string(.) 获取当前标签下所有文本内容
         cp = cp.xpath('string(.)')
         print('产品内容:' + cp)
       print('============================================')
     except Exception as e:
       print(e)
       print('出错url:' + url)
       pass

3.存储到django模型


import requests
from lxml.html import etree
import os
import django
import uuid
from django.core.files.base import ContentFile

os.environ.setdefault("DJANGO_SETTINGS_MODULE", "jiaobanzhan.settings")
django.setup()

from products.models import ProductBody, ProductsCategory, Product, ProductAdvantage

url = 'http://www.kexinjianji.com/product/hzshntjbz_1/'

def get_one_page(url):
 try:
   headers = {
     "User-Agent": "Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/51.0.2704.103 Safari/537.36"}
   res = requests.get(url=url, headers=headers, timeout=10)
   res.encoding = 'utf-8'
   if res.status_code == 200:
     return res.text
   else:
     return None
 except Exception:
   print('aa')
   return None

if __name__ == '__main__':
 content = get_one_page(url)
 tree = etree.HTML(content)
 # 产品分类url
 catgory_urls = tree.xpath('//div[@class="fdh-01-nav"]/div/h3/a/@href')
 # 处理catgory_urls
 for url in catgory_urls:
   url = 'http://www.kexinjianji.com' + url
   content = get_one_page(url)
   tree = etree.HTML(content)
   # 产品分类
   p_catgory = tree.xpath('//div[@class="cplb-3n-ts-03 b"]/h3/span/text()')
   # 该分类下产品url
   urls = tree.xpath('//div[@class="cplb-3n-ts-03-list"]/dl/dt/a/@href')
   # 处理url
   for url in urls:
     url = 'http://www.kexinjianji.com' + url
     content = get_one_page(url)
     try:
       tree = etree.HTML(content)
       # 产品名称
       title = tree.xpath('//*[@id="wrap"]//h1/text()')
       images = tree.xpath('//div[@class="sol_tj_left"]/a/img/@src')
       # 产品图片
       images_url = 'http://www.kexinjianji.com' + images[0]
       # 性能特点
       xntd = tree.xpath('//div[@class="w"]//div/span/text()|//div[@class="w"]//div/text()')
       # 技术参数
       jscs = tree.xpath('//table')[0]
       jscs_str = etree.tostring(jscs, encoding='utf-8').decode('utf-8')
       # 产品内容
       cpnr = tree.xpath('//div[@class="describe"]/p')
       # 判断是否有这分类,没有则新建
       catgory = p_catgory[0]
       products_catgory = ProductsCategory.objects.filter(name=catgory).exists()
       if products_catgory:
         products_catgory = ProductsCategory.objects.get(name=catgory)
       else:
         products_catgory = ProductsCategory(name=catgory)
         products_catgory.save()
       print(products_catgory)

# 保存产品图片
       image_content = requests.get(url=images_url)
       ext = images_url.split('.')[-1] # 获取图片类型
       filename = '{}.{}'.format(uuid.uuid4().hex[:8], ext) # 随机生成图片名字
       upload_image_file = ContentFile(image_content.content, name=filename) # 将图片保存为django类型
       product = Product(title=title[0], jscs=jscs_str, image=upload_image_file, category=products_catgory)
       product.save()
       for td in xntd:
         product_advantage = ProductAdvantage()
         product_advantage.content = td
         product_advantage.product = product
         product_advantage.save()
       for cp in cpnr:
         cp = cp.xpath('string(.)')
         product_body = ProductBody()
         product_body.body = cp
         product_body.product = product
         product_body.save()
     except Exception as e:
       print(e)
       print('出错url:' + url)

最后自己手动处理出错url(页面没有获取到技术参数,技术参数是一张图片)

4.总结

1.xpath 获取标签内容时,p标签中嵌套span标签,源码如下


<div class="describe" style="position: relative;">
  <p><span>板  宽:</span>1500mm</p>
  <p><span>板  厚:</span>4.5 mm</p>
  <p><span>出料口:</span>6口</p>
  <p><span>重  量:</span>6000 kg</p>
</div>

使用xpath获取p标签内容
我想得到的效果如下
板 宽:1500mm
板 厚:4.5 mm
出料口:6口
重 量:6000 kg
使用以下xpath 只能分开获取,不是想要的效果


//div[@class="describe"]/p/span/text()|//div[@class="describe"]/p/text()

百度之后找到的解决办法,使用xpath(‘string(.)')
1.先获取所有p标签


cpnr = tree.xpath('//div[@class="describe"]/p')

2.使用**string(.)**获取所有标签所有文本


cp = cp.xpath('string(.)')

循环遍历所有p标签即可

来源:https://blog.csdn.net/cll_869241/article/details/114005783

标签:xpath,爬取,django,模型
0
投稿

猜你喜欢

  • Python numpy.power()函数使用说明

    2021-11-23 02:05:15
  • 用Python 执行cmd命令

    2022-06-08 15:29:17
  • Tensorflow--取tensorf指定列的操作方式

    2021-06-25 21:21:02
  • vue前端开发keepAlive使用详解

    2024-05-02 16:33:52
  • Python dict的常用方法示例代码

    2023-05-17 09:58:15
  • 透明数据加密(TDE)库的备份和还原

    2024-01-14 04:51:57
  • 详解Python:面向对象编程

    2023-10-26 16:17:53
  • Python中反转二维数组的行和列问题

    2021-06-24 13:48:27
  • Pytorch 图像变换函数集合小结

    2022-06-14 08:52:09
  • Python 数据分析之Beautiful Soup 提取页面信息

    2022-04-30 04:34:10
  • Python入门教程3. 列表基本操作【定义、运算、常用函数】 <font color=red>原创</font>

    2023-07-15 13:09:19
  • Django如何使用asyncio协程和ThreadPoolExecutor多线程

    2022-06-22 11:43:34
  • python列表数据增加和删除的具体实例

    2021-08-05 15:11:09
  • Python内建函数之raw_input()与input()代码解析

    2021-01-12 01:37:36
  • 产品设计与用户体验

    2009-02-02 10:15:00
  • 简单了解Python多态与属性运行原理

    2021-03-13 21:29:42
  • 在Python中操作列表之list.extend()方法的使用

    2023-07-02 20:43:16
  • 数据库性能优化之冗余字段的作用

    2011-03-03 19:21:00
  • 自然语言处理之文本热词提取(含有《源码》和《数据》)

    2021-11-26 11:14:58
  • Python爬取阿拉丁统计信息过程图解

    2022-04-02 06:23:22
  • asp之家 网络编程 m.aspxhome.com