为什么我在PyTorch中用于降噪的自动编码器会学习将所有内容归零?

如何解决为什么我在PyTorch中用于降噪的自动编码器会学习将所有内容归零?

我需要使用pytorch构建去噪自动编码器,以清除信号。

例如,我可以使用cosine函数并对其进行间隔采样(其中我有两个参数-BKB是间隔数我以每个示例为例,K是每个间隔中有多少个采样点(等距)),例如,我可以采用B = 5个间隔并测量每个间隔中的K = 8个点。因此,每个点之间的距离为2pi / 8,而我总共有40个点。我尝试概括的功能数量为L,我将其视为不同的渠道。然后,我为每个示例添加一个随机的起始位置(使其稍有不同),然后添加随机噪声,并将其发送到自动编码器进行训练。

问题是,无论架构或学习速度如何,它都会逐渐学习输出零。自动编码器非常简单,因此我不认为这有问题,而是我如何生成数据的问题。

我仍然附上了两个代码:

class ConvAutoencoder(nn.Module):
def __init__(self,enc_channels,dec_channels):
    super(ConvAutoencoder,self).__init__()
    ## encoder layers ##

    encoder_layers = []
    decoder_layers = []

    in_channels = enc_channels[0]

    for i in range(1,len(enc_channels)):
        out_channels = enc_channels[i]

        encoder_layers += [nn.ConvTranspose2d(in_channels,out_channels,kernel_size=1,bias=True),nn.ReLU()]
        in_channels = out_channels

    in_channels = dec_channels[0]

    for i in range(1,len(dec_channels)):
        out_channels = dec_channels[i]

        decoder_layers += [nn.ConvTranspose2d(in_channels,nn.ReLU()]
        in_channels = out_channels


    self.encoder = nn.Sequential(*encoder_layers)
    self.decoder = nn.Sequential(*decoder_layers)


def forward(self,x):

    if len(x.shape) == 3:
        x = x.unsqueeze(dim=-1)

    res = self.decoder(self.encoder(x)).squeeze(-1)

    return res

数据生成如下:

def generate_data(batch_size: int,intervals: int,sample_length: int,channels_functions,noise_scale=1)->torch.tensor:
    channels = len(channels_functions)
    mul_term = 2 * np.pi / sample_length # each sample is 2pi and equally distance
    
    # each example is K * B long
    positions = np.arange(0,sample_length * intervals)
    x = positions * mul_term

    # creating random start points (from negative to positive)
    random_starting_pos = (np.random.rand(batch_size) - 0.5) * 10000
    start_pos_mat = np.tile(random_starting_pos,(sample_length * intervals,1))
    start_pos_mat = np.tile(start_pos_mat,(channels,1)).T
    start_pos_mat = np.reshape(start_pos_mat,(batch_size,channels,sample_length * intervals))

    X = np.tile(x,1))
    X = np.repeat(X[np.newaxis,:,:],batch_size,axis=0)

    X += start_pos_mat #adding the random starting position

    # apply each function to a different channel
    for i,function in enumerate(channels_functions):
        X[:,i,:] = function(X[:,:])

    clean = X
    noise = np.random.normal(scale=noise_scale,size=clean.shape)
    noisy = clean + noise

    # normalizing each sample
    row_sums = np.linalg.norm(clean,axis=2)
    clean = clean / row_sums[:,np.newaxis]
    row_sums = np.linalg.norm(noisy,axis=2)
    noisy = noisy / row_sums[:,np.newaxis]

    clean = torch.from_numpy(clean)
    noisy = torch.from_numpy(noisy)

    return clean,noisy

编辑-添加了整个训练循环:

if __name__ == '__main__':

func_list = [lambda x: np.cos(x),lambda x: np.cos((x**4) / 10),lambda x: np.sin(x**3 * np.cos(x**2)),lambda x: 0.25*np.cos(x**2) - 10*np.sin(0.25*x)]

L = len(func_list)
K = 3
B = 4

enc_channels = [L,64,128,256]


num_epochs = 100

model = models.ConvAutoencoder(enc_channels,enc_channels[::-1])

criterion = torch.nn.BCELoss()

optimizer = torch.optim.Adam(model.parameters(),lr=0.005,weight_decay=1e-5)

for epoch in range(num_epochs):
    clean,noisy = util.generate_data(128,K,B,func_list)

    # ===================forward=====================
    output = model(noisy.float())
    loss = criterion(output.float(),clean.float())
    # ===================backward====================
    optimizer.zero_grad()
    loss.backward()
    optimizer.step()
    # ===================log========================
    print('epoch [{}/{}],loss:{:.4f}'.format(epoch + 1,num_epochs,loss.data))

    if epoch % 10 == 0:
        show_clean,show_noisy = util.generate_data(1,func_list)
        print("clean\n{}".format(show_clean))
        print("noisy\n{}".format(show_noisy))
        print("denoised\n{}".format(model(show_noisy.float())))

在大约10个时期后,请确保模型输出足够:

clean vector
tensor([[[ 0.3611,-0.1905,-0.3611,0.1905,0.3611,0.1905],[ 0.3387,-0.0575,-0.2506,-0.3531,-0.3035,0.3451,0.3537,-0.2416,0.2652,-0.3126,-0.3203,-0.1707],[-0.0369,0.4412,-0.1323,0.1802,-0.2943,0.3590,0.4549,0.0827,-0.0164,0.4350,-0.1413,-0.3395],[ 0.3997,0.3516,0.2451,0.1136,-0.0458,-0.1944,-0.3225,-0.3925,-0.3971,-0.3382,-0.2457,-0.1153]]],dtype=torch.float64)
noisy vector
tensor([[[-0.1071,-0.0671,0.0993,-0.2029,0.1587,-0.4407,-0.0867,-0.2598,0.2426,-0.6939,-0.3011,-0.0870],[ 0.0889,-0.3415,-0.1434,-0.2393,-0.4708,0.0144,0.2620,-0.1186,0.6424,0.0886,-0.2192,-0.1562],[ 0.1989,0.2794,0.0848,-0.2729,-0.2168,0.1475,0.5294,0.4788,0.1754,0.2333,-0.0549,-0.3665],[ 0.3611,0.3535,0.1957,0.1980,-0.1115,-0.1912,-0.2713,-0.4087,-0.3669,-0.3675,-0.2991,-0.1356]]],dtype=torch.float64)
denoised vector
tensor([[[0.,0.,0.],[0.,0.]]],grad_fn=<SqueezeBackward1>)

谢谢

解决方法

问题在于您在最后一层中使用ReLU,但是目标(干净)包含负值。在最后一层中使用ReLU时,无法获得负值。

只需用以下内容替换您的解码器:

    for i in range(1,len(dec_channels)):
        out_channels = dec_channels[i]
        
        if i == len(dec_channels) - 1:
            # last layer
            decoder_layers += [nn.ConvTranspose2d(in_channels,out_channels,kernel_size=1,bias=True)]
        else:
            decoder_layers += [nn.ConvTranspose2d(in_channels,bias=True),nn.ReLU()]
        in_channels = out_channels

然后我建议使用L2损失。

版权声明:本文内容由互联网用户自发贡献,该文观点与技术仅代表作者本人。本站仅提供信息存储空间服务,不拥有所有权,不承担相关法律责任。如发现本站有涉嫌侵权/违法违规的内容, 请发送邮件至 dio@foxmail.com 举报,一经查实,本站将立刻删除。

相关推荐


依赖报错 idea导入项目后依赖报错,解决方案:https://blog.csdn.net/weixin_42420249/article/details/81191861 依赖版本报错:更换其他版本 无法下载依赖可参考:https://blog.csdn.net/weixin_42628809/a
错误1:代码生成器依赖和mybatis依赖冲突 启动项目时报错如下 2021-12-03 13:33:33.927 ERROR 7228 [ main] o.s.b.d.LoggingFailureAnalysisReporter : *************************** APPL
错误1:gradle项目控制台输出为乱码 # 解决方案:https://blog.csdn.net/weixin_43501566/article/details/112482302 # 在gradle-wrapper.properties 添加以下内容 org.gradle.jvmargs=-Df
错误还原:在查询的过程中,传入的workType为0时,该条件不起作用 &lt;select id=&quot;xxx&quot;&gt; SELECT di.id, di.name, di.work_type, di.updated... &lt;where&gt; &lt;if test=&qu
报错如下,gcc版本太低 ^ server.c:5346:31: 错误:‘struct redisServer’没有名为‘server_cpulist’的成员 redisSetCpuAffinity(server.server_cpulist); ^ server.c: 在函数‘hasActiveC
解决方案1 1、改项目中.idea/workspace.xml配置文件,增加dynamic.classpath参数 2、搜索PropertiesComponent,添加如下 &lt;property name=&quot;dynamic.classpath&quot; value=&quot;tru
删除根组件app.vue中的默认代码后报错:Module Error (from ./node_modules/eslint-loader/index.js): 解决方案:关闭ESlint代码检测,在项目根目录创建vue.config.js,在文件中添加 module.exports = { lin
查看spark默认的python版本 [root@master day27]# pyspark /home/software/spark-2.3.4-bin-hadoop2.7/conf/spark-env.sh: line 2: /usr/local/hadoop/bin/hadoop: No s
使用本地python环境可以成功执行 import pandas as pd import matplotlib.pyplot as plt # 设置字体 plt.rcParams[&#39;font.sans-serif&#39;] = [&#39;SimHei&#39;] # 能正确显示负号 p
错误1:Request method ‘DELETE‘ not supported 错误还原:controller层有一个接口,访问该接口时报错:Request method ‘DELETE‘ not supported 错误原因:没有接收到前端传入的参数,修改为如下 参考 错误2:cannot r
错误1:启动docker镜像时报错:Error response from daemon: driver failed programming external connectivity on endpoint quirky_allen 解决方法:重启docker -&gt; systemctl r
错误1:private field ‘xxx‘ is never assigned 按Altʾnter快捷键,选择第2项 参考:https://blog.csdn.net/shi_hong_fei_hei/article/details/88814070 错误2:启动时报错,不能找到主启动类 #
报错如下,通过源不能下载,最后警告pip需升级版本 Requirement already satisfied: pip in c:\users\ychen\appdata\local\programs\python\python310\lib\site-packages (22.0.4) Coll
错误1:maven打包报错 错误还原:使用maven打包项目时报错如下 [ERROR] Failed to execute goal org.apache.maven.plugins:maven-resources-plugin:3.2.0:resources (default-resources)
错误1:服务调用时报错 服务消费者模块assess通过openFeign调用服务提供者模块hires 如下为服务提供者模块hires的控制层接口 @RestController @RequestMapping(&quot;/hires&quot;) public class FeignControl
错误1:运行项目后报如下错误 解决方案 报错2:Failed to execute goal org.apache.maven.plugins:maven-compiler-plugin:3.8.1:compile (default-compile) on project sb 解决方案:在pom.
参考 错误原因 过滤器或拦截器在生效时,redisTemplate还没有注入 解决方案:在注入容器时就生效 @Component //项目运行时就注入Spring容器 public class RedisBean { @Resource private RedisTemplate&lt;String
使用vite构建项目报错 C:\Users\ychen\work&gt;npm init @vitejs/app @vitejs/create-app is deprecated, use npm init vite instead C:\Users\ychen\AppData\Local\npm-