NO.8_485. Max Consecutive Ones

485. Max Consecutive Ones

Given a binary array, find the maximum number of consecutive 1s in this array.

Example 1:

Input: [1,1,0,1,1,1]
Output: 3
Explanation: The first two digits or the last three digits are consecutive 1s.
    The maximum number of consecutive 1s is 3.

Note:

  • The input array will only contain 0 and 1.
  • The length of input array is a positive integer and will not exceed 10,000

//参考解

//运用乘法,效率一般

class Solution {
public:
int findMaxConsecutiveOnes(vector<int>& nums) 
    {
        int max = 0;
        int numsSize=nums.size();
 	int sum = 0;
	 for (int i=0; i<numsSize; i++)
 	 {
   		  sum = (sum+nums[i])*nums[i];
    		  if(max<sum){max=sum;}
	 }
	return max;
    }
};

//解1

//分为在0列or1列两种情况

class Solution {
public:
    int findMaxConsecutiveOnes(vector<int>& nums) 
    {
        int res=0;
        int length=nums.size();
        bool flag=0;//标志位
        int num=0;//当前列连续1的个数
        for(int i=0;i<length ;i++)
        {
            if(flag)//在1队列中
            {
                if(nums[i])
                {
                    num++;
                    res=num>res?num:res;
                }
                else
                {
                    flag=0;
                    num=0;
                }
            }
            else//在1队列外
            {
                if(nums[i])
                {
                    flag=1;
                    num++;
                    res=1>res?1:res;
                }
            }
        }
        return res;
    }
};


  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
import torch import torch.nn as nn import torch.optim as optim import numpy as np 定义基本循环神经网络模型 class RNNModel(nn.Module): def init(self, rnn_type, input_size, hidden_size, output_size, num_layers=1): super(RNNModel, self).init() self.rnn_type = rnn_type self.input_size = input_size self.hidden_size = hidden_size self.output_size = output_size self.num_layers = num_layers self.encoder = nn.Embedding(input_size, hidden_size) if rnn_type == 'RNN': self.rnn = nn.RNN(hidden_size, hidden_size, num_layers) elif rnn_type == 'GRU': self.rnn = nn.GRU(hidden_size, hidden_size, num_layers) self.decoder = nn.Linear(hidden_size, output_size) def forward(self, input, hidden): input = self.encoder(input) output, hidden = self.rnn(input, hidden) output = output.view(-1, self.hidden_size) output = self.decoder(output) return output, hidden def init_hidden(self, batch_size): if self.rnn_type == 'RNN': return torch.zeros(self.num_layers, batch_size, self.hidden_size) elif self.rnn_type == 'GRU': return torch.zeros(self.num_layers, batch_size, self.hidden_size) 定义数据集 with open('汉语音节表.txt', encoding='utf-8') as f: chars = f.readline() chars = list(chars) idx_to_char = list(set(chars)) char_to_idx = dict([(char, i) for i, char in enumerate(idx_to_char)]) corpus_indices = [char_to_idx[char] for char in chars] 定义超参数 input_size = len(idx_to_char) hidden_size = 256 output_size = len(idx_to_char) num_layers = 1 batch_size = 32 num_steps = 5 learning_rate = 0.01 num_epochs = 100 定义模型、损失函数和优化器 model = RNNModel('RNN', input_size, hidden_size, output_size, num_layers) criterion = nn.CrossEntropyLoss() optimizer = optim.Adam(model.parameters(), lr=learning_rate) 训练模型 for epoch in range(num_epochs): model.train() hidden = model.init_hidden(batch_size) loss = 0 for X, Y in data_iter_consecutive(corpus_indices, batch_size, num_steps): optimizer.zero_grad() hidden = hidden.detach() output, hidden = model(X, hidden) loss = criterion(output, Y.view(-1)) loss.backward() torch.nn.utils.clip_grad_norm_(model.parameters(), max_norm=1.0) optimizer.step() if epoch % 10 == 0: print(f"Epoch {epoch}, Loss: {loss.item()}")请正确缩进代码
06-09
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值