当前位置: 首页 > article >正文

Python定时任务管理器

工作中需要写很多数据采集器以不同的时间间隔运行。首先想到的是通过cron或jenkins等工具来进行管理,由于各种原因不太方便使用这些工具。

问了一下 deepseek,发现可以使用 croniter 实现类似 cron 的功能,完美满足了我的需求。

使用方法

1. 首先安装  croniter

pip install croniter

2. 创建任务配置文件

tasks:
  - name: "Task name"
    schedule: "*/10 * * * *"
    command: "..."
    workdir: "..."

配置文件可以采用yaml格式,其中name参数表示任务的名称, schedule 参数指定任务的运行时间,格式与 cron 保持一致; command 参数指定要运行的命令,最好使用全路径;workdir 参数命令运行目录。

在配置文件中可以定义多个任务。

3. 定时任务管理程序如下

在 deekseek 给出的代码基础上做了部分优化。

import os
import sys
import time
import signal
import logging
import subprocess
from datetime import datetime
from croniter import croniter
import pickle
import yaml
import threading
import json
from pathlib import Path
from logging.handlers import RotatingFileHandler

# 日志配置
logging.basicConfig(
    level=logging.INFO,
    format='%(asctime)s - %(name)s - %(levelname)s - %(message)s',
    handlers=[
        RotatingFileHandler(
            'taskscheduler.log',
            maxBytes=50*1024*1024,  # 50MB大小限制
            backupCount=2,          # 保留2个备份
            encoding='utf-8'        # 明确指定编码
        ),
        logging.StreamHandler()
    ]
)
logger = logging.getLogger('taskscheduler')

def readConf(cfname):
    cfpath = os.path.join(os.path.dirname(__file__), f"{cfname}")
    if not os.path.exists(cfpath):
        logging.warning(f"Missing configuration file : {cfpath}")
        sys.exit(1)

    with open(cfpath, 'r') as fd:
        return yaml.load(fd, Loader=yaml.FullLoader)

def writeDataFile(path, data, type='json'):
    fname = os.path.join(os.path.dirname(__file__), f"{path}")
    with open(fname, 'w') as fd:
        if type == 'json':
            return json.dump(data, fd)
        elif type == 'pickle':
            return pickle.dump(data, fd)
        elif type == 'yaml':
            return yaml.dump(data, fd)

def readDataFile(path, type='json'):
    fname = os.path.join(os.path.dirname(__file__), f"{path}")
    if not os.path.exists(fname):
        logging.warning(f"Missing data file : {fname}")
        raise

    with open(fname, 'r') as fd:
        if type == 'json':
            return json.load(fd)
        elif type == 'pickle':
            return pickle.load(fd)
        elif type == 'yaml':
            return yaml.load(fd, Loader=yaml.FullLoader)
        
class TaskScheduler:
    def __init__(self, config_file="taskscheduler.yaml", state_file='.taskscheduler.states.json'):
        self.config_file = config_file
        self.state_file = state_file
        self.lock = threading.Lock()
        self.tasks = {}
        self.shutdown_flag = False
        self._load_config()
        self._load_states()
        self._setup_signal_handlers()

    def _setup_signal_handlers(self):
        """注册信号处理"""
        signal.signal(signal.SIGINT, self._handle_signal)
        signal.signal(signal.SIGTERM, self._handle_signal)

    def _handle_signal(self, signum, frame):
        """处理中断信号"""
        logger.warning(f"接收到终止信号 {signum},开始优雅关闭...")
        self.shutdown_flag = True

    def _load_config(self):
        """加载YAML配置文件"""
        try:
            config = readConf(self.config_file)
            
            for task_cfg in config['tasks']:
                self._validate_task_config(task_cfg)
                task_id = task_cfg['name']
                
                def make_task_closure(cmd, workdir, task_id):
                    def task_function():
                        try:
                            Path(workdir).mkdir(parents=True, exist_ok=True)
                            logger.info(f"开始执行命令: {cmd}", extra={'task_id': task_id})
                            
                            start_time = time.time()
                            process = subprocess.run(
                                cmd,
                                cwd=workdir,
                                shell=True,
                                check=True,
                                stdout=subprocess.PIPE,
                                stderr=subprocess.STDOUT,
                                text=True
                            )
                            
                            duration = time.time() - start_time
                            logger.info(
                                f"命令执行成功\n"
                                f"• 耗时: {duration:.2f}s\n"
                                f"• 输出: {process.stdout.strip() or '无输出'}",
                                extra={'task_id': task_id}
                            )
                        except subprocess.CalledProcessError as e:
                            logger.error(
                                f"命令执行失败(退出码 {e.returncode})\n"
                                f"• 错误输出: {e.stdout.strip()}",
                                extra={'task_id': task_id}
                            )
                        except Exception as e:
                            logger.error(
                                f"执行异常: {str(e)}",
                                exc_info=True,
                                extra={'task_id': task_id}
                            )
                    return task_function

                self.tasks[task_id] = {
                    'cron': task_cfg['schedule'],
                    'func': make_task_closure(
                        task_cfg['command'],
                        task_cfg.get('workdir', '.'),
                        task_id
                    ),
                    'next_run': None
                }
            logger.info("配置文件加载完成")
            
        except Exception as e:
            logger.critical(f"配置文件加载失败: {str(e)}", exc_info=True)
            raise

    def _validate_task_config(self, config):
        """验证任务配置"""
        required_fields = ['name', 'schedule', 'command']
        for field in required_fields:
            if field not in config:
                raise ValueError(f"任务配置缺少必要字段: {field}")
        
        # 验证cron表达式有效性
        try:
            croniter(config['schedule'])
        except Exception as e:
            raise ValueError(f"无效的cron表达式: {config['schedule']}") from e

    def _load_states(self):
        """加载任务状态"""
        try:
            states = readDataFile(self.state_file)
            if states:
                for task_id, timestamp in states.items():
                    if timestamp and task_id in self.tasks:
                        states[task_id] = datetime.fromisoformat(timestamp)
                self.states = states
            else:
                self.states = {}
        except Exception as e:
            logger.warning(f"加载任务状态失败: {str(e)}", exc_info=True)
            self.states = {}

        # 初始化未记录的任务
        for task_id in self.tasks:
            if task_id not in self.states:
                now = datetime.now()
                prev_time = croniter(self.tasks[task_id]['cron'], now).get_prev(datetime)
                self.states[task_id] = prev_time

    def _save_states(self):
        """保存任务状态"""
        logger.info(f"states: {self.states}")
        serialized = {k: v.isoformat() if v else None 
                        for k, v in self.states.items()}
        writeDataFile(self.state_file, serialized)

    def _update_next_run(self, task_id):
        """更新任务的下次执行时间"""
        task = self.tasks[task_id]
        cron_gen = croniter(task['cron'], self.states[task_id])
        task['next_run'] = cron_gen.get_next(datetime)

    def run(self):
        """启动调度器主循环"""
        logger.info("调度器启动")
        with self.lock:
            for task_id in self.tasks:
                self._update_next_run(task_id)

        logger.info(f"已加载 {len(self.tasks)} 个任务")
        for task_id, task in self.tasks.items():
            logger.info(
                f"任务注册 - ID: {task_id}",
                extra={'task_id': task_id}
            )
            logger.info(
                f"下次执行时间: {task['next_run']}",
                extra={'task_id': task_id}
            )

        try:
            while not self.shutdown_flag:
                now = datetime.now()
                earliest_run = None

                with self.lock:
                    # 执行到期任务
                    for task_id, task in self.tasks.items():
                        # logger.info(f"任务 {task_id} 下次执行时间: {task['next_run']} 当前时间: {now}")
                        # 跳过未设置下次执行时间的任务
                        if task['next_run'] is None:
                            continue
                        if task['next_run'] <= now:
                            self._execute_task(task_id)
                            self._update_next_run(task_id)

                    # 查找最近的下次执行时间
                    for task in self.tasks.values():
                        if (earliest_run is None) or (task['next_run'] < earliest_run):
                            earliest_run = task['next_run']

                if earliest_run:
                    sleep_time = (earliest_run - datetime.now()).total_seconds()
                    sleep_time = max(sleep_time, 0)
                    timeout = min(sleep_time, 60)
                    
                    # 分段休眠以快速响应关闭信号
                    start = time.time()
                    while (time.time() - start < timeout) and not self.shutdown_flag:
                        time.sleep(1)
                else:
                    time.sleep(1)

        finally:
            self._shutdown()

    def _execute_task(self, task_id):
        """执行单个任务"""
        task = self.tasks[task_id]
        logger.info(
            "触发执行",
            extra={'task_id': task_id}
        )
        try:
            task['func']()
            # 更新状态
            self.states[task_id] = task['next_run']
            self._save_states()
        except Exception as e:
            logger.error(
                f"任务执行异常: {str(e)}",
                exc_info=True,
                extra={'task_id': task_id}
            )

    def _shutdown(self):
        """关闭时的清理操作"""
        logger.info("正在关闭调度器...")
        # 确保保存最终状态
        with self.lock:
            self._save_states()
        logger.info("所有任务状态已保存")
        logging.shutdown()

if __name__ == "__main__":
    scheduler = TaskScheduler()
    try:
        scheduler.run()
    except Exception as e:
        logger.critical(f"未捕获的异常: {str(e)}", exc_info=True)
        raise


http://www.kler.cn/a/577969.html

相关文章:

  • 微博热点信息爬虫
  • Java链接redis
  • tomcat的安装与配置(包含在idea中配置tomcat)
  • 景联文科技:以精准数据标注赋能AI进化,构筑智能时代数据基石
  • RAG技术的PDF智能问答系统
  • 【图像阈值分割、区域分割、边缘分割】
  • 计算机毕业设计SpringBoot+Vue.js高校专业实习管理系统(源码+文档+PPT+讲解)
  • CentOS Docker 安装指南
  • mounted() 钩子函数
  • 大数据学习(58)-DolphinScheduler使用DataX实现数据同步
  • 如何在DigitalOcean的H100 GPU服务器上运行DeepSeek R1 模型
  • leetcode日记(85)验证二叉搜索树
  • 深度学习驱动的跨行业智能化革命:技术突破与实践创新
  • 鸿蒙Next-应用检测、安装以及企业内部商店的实现
  • 常见的算法题python
  • SecureCRT 文件上传下载操作指南
  • PyTorch系列教程:Tensor.view() 方法详解
  • <script setup>和export default { setup() { ... } }区别
  • 计算机网络-服务器模型
  • pom.xml配置(mybatisplus增删改查实现;PageInfo分页实现;JSONObject实现)