Ipyparallel在大量工作上失败

如何解决Ipyparallel在大量工作上失败

我不熟悉Ipyparallel,并尝试寻求帮助,但是找不到我能理解的有助于我的情况的东西。如果答案在那里-我道歉!

我正在创建一个Jupyter Notebook,可用于将二进制文件转换为ASCII文件。我将分享尽可能多的代码,但我无法提供一个可行的示例。

我感觉笔记本正在运行。转换几个文件(少于1000个)的工作符合我的预期,但是,当我增加要转换的文件的数量时,就好像什么都没有发生了,即使达到5000个也常常无法完成。

我的代码如下:

def convert_bhb_tff_parallel(p):
    from pathlib import Path
    from pyrosap import WLSFile,BhbFile
    def read_wfh_params(wfh):
        params = {}
        with open(wfh) as f:
            for line in f.readlines():
                if line.startswith('!SHEILAPM'):
                    _,param,value = line.split()
                    params[param] = value
        return params

    def convert_bhb_to_tff(bhb):
        """Convert the bhb file to tff.

        Ensure WLS comparison is run and a log written.

        Args:
            bhb (Path): Pathlib path of the bhb file.

        Returns:
            str: The log line
            bool: if an error occured in the WLS comparison (True/False)

        """
        wls = None
        sup_file = sup_files.get(d[bhb.stem],None)
        if not sup_file:
            line = '{:<40}    {:<20}'
            line += '    {:<25.5}    {:<20.5}'.format(0.0,0.0)
            line += ' || WARNING,Support file missing.'
            error = 3
        else:
            if sup_file_type == 'WLS':
                wls = WLSFile(sup_file)
                wls.read_file()
                params = wls.parameters
            elif sup_file_type == 'WFH':
                params = read_wfh_params(sup_file)
            bhb_f = BhbFile(bhb,select=select_dict)
            bhb_f.read_file()

            tff_path = Path(out_dir) / (sup_file.stem + '.tff')

            line,error = bhb_f.convert_to_tff(tff_path,params,file_format=tff_format,map_data=column_mapping,wls=wls)
        if not line:
            line = '{:<40}    {:<20}'
            line += '    {:<25}    {:<20} || ACCEPTED. BHB file converted.'.format(0.0,0.0)
        line = line.format(bhb.stem,d[bhb.stem])
        return line,error
    errors = 0
    
    bhb_name,sup_file_type,d,select_dict = p[:4]
    bhb_files,sup_files,out_dir,tff_format,column_mapping = p[4:]
    bhb = bhb_files.get(bhb_name,None)
    if bhb:
        line,error = convert_bhb_to_tff(bhb)
    else:
        line = '{:<40}    {:<20}'.format(bhb_name,d[bhb_name])
        line += '    {:<25.5}    {:<20.5}'.format(0.0,0.0)
        line += ' || WARNING,BHB file missing.'
        error = 4
    if error > 0:
        errors = True
    return line,errors

subprocess.Popen(["ipcluster","start","-n={:d}".format(num_engines)])
print('Notebook sleeping for 5 seconds in order to allow the parallel computing to start')
time.sleep(5)

log = []
full_log = []
error_log = []
error = False
rc = Client()
view = rc.load_balanced_view()

try:
    for bhb_name in d.keys():
        args = (bhb_name,select_dict,bhb_files,column_mapping)
        out = view.apply_async(convert_bhb_tff_parallel,(args))
        log.append(out)

    rc.wait_interactive(log)
    error = any([ar.get()[1] > 0 for ar in log])
    full_log.extend([ar.get()[0] for ar in log])
    error_log.extend([ar.get()[0] for ar in log if ar.get()[1] > 0])
    
    flog_fn = Path(out_dir) / 'log.log'
    elog_fn = Path(out_dir) / 'errorlog.log'
    with open(flog_fn,'w') as f:
        f.write('{:<40}    {:<20}    {:<25}    {:<20} || ASSESSMENT\n\n'.format(
            'BHB file name','Support file name','Wave Elevation Tolerance','Correlation factor'
        ))
        f.write('\n'.join(full_log))

    with open(elog_fn,'Correlation factor'
        ))
        f.write('\n'.join(error_log))

    if error is True:
        print('Errors/Warnings was found when comparing wave data,please check the log file.')
    else:
        print('No errors was found during wave comparison.\n\n')
finally:
    close_parallel = subprocess.Popen(["ipcluster","stop"])

是否有什么地方我做错了,因为当d的长度增加时,它开始失败了?

版权声明:本文内容由互联网用户自发贡献,该文观点与技术仅代表作者本人。本站仅提供信息存储空间服务,不拥有所有权,不承担相关法律责任。如发现本站有涉嫌侵权/违法违规的内容, 请发送邮件至 dio@foxmail.com 举报,一经查实,本站将立刻删除。

相关推荐


依赖报错 idea导入项目后依赖报错,解决方案:https://blog.csdn.net/weixin_42420249/article/details/81191861 依赖版本报错:更换其他版本 无法下载依赖可参考:https://blog.csdn.net/weixin_42628809/a
错误1:代码生成器依赖和mybatis依赖冲突 启动项目时报错如下 2021-12-03 13:33:33.927 ERROR 7228 [ main] o.s.b.d.LoggingFailureAnalysisReporter : *************************** APPL
错误1:gradle项目控制台输出为乱码 # 解决方案:https://blog.csdn.net/weixin_43501566/article/details/112482302 # 在gradle-wrapper.properties 添加以下内容 org.gradle.jvmargs=-Df
错误还原:在查询的过程中,传入的workType为0时,该条件不起作用 &lt;select id=&quot;xxx&quot;&gt; SELECT di.id, di.name, di.work_type, di.updated... &lt;where&gt; &lt;if test=&qu
报错如下,gcc版本太低 ^ server.c:5346:31: 错误:‘struct redisServer’没有名为‘server_cpulist’的成员 redisSetCpuAffinity(server.server_cpulist); ^ server.c: 在函数‘hasActiveC
解决方案1 1、改项目中.idea/workspace.xml配置文件,增加dynamic.classpath参数 2、搜索PropertiesComponent,添加如下 &lt;property name=&quot;dynamic.classpath&quot; value=&quot;tru
删除根组件app.vue中的默认代码后报错:Module Error (from ./node_modules/eslint-loader/index.js): 解决方案:关闭ESlint代码检测,在项目根目录创建vue.config.js,在文件中添加 module.exports = { lin
查看spark默认的python版本 [root@master day27]# pyspark /home/software/spark-2.3.4-bin-hadoop2.7/conf/spark-env.sh: line 2: /usr/local/hadoop/bin/hadoop: No s
使用本地python环境可以成功执行 import pandas as pd import matplotlib.pyplot as plt # 设置字体 plt.rcParams[&#39;font.sans-serif&#39;] = [&#39;SimHei&#39;] # 能正确显示负号 p
错误1:Request method ‘DELETE‘ not supported 错误还原:controller层有一个接口,访问该接口时报错:Request method ‘DELETE‘ not supported 错误原因:没有接收到前端传入的参数,修改为如下 参考 错误2:cannot r
错误1:启动docker镜像时报错:Error response from daemon: driver failed programming external connectivity on endpoint quirky_allen 解决方法:重启docker -&gt; systemctl r
错误1:private field ‘xxx‘ is never assigned 按Altʾnter快捷键,选择第2项 参考:https://blog.csdn.net/shi_hong_fei_hei/article/details/88814070 错误2:启动时报错,不能找到主启动类 #
报错如下,通过源不能下载,最后警告pip需升级版本 Requirement already satisfied: pip in c:\users\ychen\appdata\local\programs\python\python310\lib\site-packages (22.0.4) Coll
错误1:maven打包报错 错误还原:使用maven打包项目时报错如下 [ERROR] Failed to execute goal org.apache.maven.plugins:maven-resources-plugin:3.2.0:resources (default-resources)
错误1:服务调用时报错 服务消费者模块assess通过openFeign调用服务提供者模块hires 如下为服务提供者模块hires的控制层接口 @RestController @RequestMapping(&quot;/hires&quot;) public class FeignControl
错误1:运行项目后报如下错误 解决方案 报错2:Failed to execute goal org.apache.maven.plugins:maven-compiler-plugin:3.8.1:compile (default-compile) on project sb 解决方案:在pom.
参考 错误原因 过滤器或拦截器在生效时,redisTemplate还没有注入 解决方案:在注入容器时就生效 @Component //项目运行时就注入Spring容器 public class RedisBean { @Resource private RedisTemplate&lt;String
使用vite构建项目报错 C:\Users\ychen\work&gt;npm init @vitejs/app @vitejs/create-app is deprecated, use npm init vite instead C:\Users\ychen\AppData\Local\npm-