Detailed understanding of input, output and parameters of RNN, LSTM and GRU, the basic units of cyclic neural network in pytorch

This article is for those who have deeply understood the mathematical principles and operation process of RNN, LSTM and GRU. If they do not understand its basic idea and process, it may not be very simple to understand. 1, Let's start with an example This is an example on the official website. Taking LSTM as an example this tim ...

Posted by oocuz on Wed, 10 Nov 2021 21:46:30 -0800

ResNet18 is introduced and used to classify CIFAR-10 data sets

ResNet, an article published by he Kaiming on CVPR in 2015, uses the concept of residual connection. As soon as the paper was published, it directly detonated the whole cv world. And ResNet won the first place on ImageNet in 2016. ResNet has been used in cutting-edge technologies in various fields of AI. I would be satisfied if I cited one ten ...

Posted by timtom3 on Wed, 10 Nov 2021 10:11:05 -0800

non_max_suppression code analysis

non_max_suppression code analysis NMS was performed simply according to confidence def non_max_suppression(boxes, conf_thres=0.5, nms_thres=0.3): detection = boxes # 1. Find the box in the picture whose score is greater than the threshold function. The number of coincident boxes can be greatly reduced by filtering scores before screen ...

Posted by dloeppky on Mon, 08 Nov 2021 03:17:38 -0800

2021-11-05 creation and common methods of pytorch Tensor

Creation and common methods of pytorch Tensor Import pytorch package import torch View version number torch.__version__ '1.7.1' 1, Basic creation and types of tensors 1. Tensor creation method Use the pytorch tensor creation function: torch.tensor(): # Create tensor from list t = torch.tensor([1, 2]) t tensor([1,2]) # Creating tenso ...

Posted by benrussell on Sat, 06 Nov 2021 11:45:54 -0700

Message Passing parsing in Python geometric

MessagePassing in Python geometric Convolution computation in the graph is usually called neighborhood aggregation or message passing. Definition x i ( k − ...

Posted by pbarney on Sat, 06 Nov 2021 04:25:16 -0700

Ubuntu 18.04 + Anaconda + CUDA + cudnn + Python environment configuration (updated on 2021.10, available for pro testing)

This set of environment is really too complex. There are too many pits. After more than half a year, it finally succeeded today. Graphics card driver Direct installation system recommends graphics card driver, with the lowest error probability. sudo ubuntu-drivers autoinstall View installation status nvidia-smi The graphics card version ...

Posted by Coronet on Fri, 29 Oct 2021 23:07:37 -0700

[Linux] reproduction of fast r-cnn

0. Download Code using Git Installation and use of Git: https://blog.csdn.net/qq_44747572/article/details/121006841Clone Code: Pytorch version 1.0.0 source code: https://github.com/jwyang/faster-rcnn.pytorch/tree/pytorch-1.0 Download the zip and unzip it in the specified folder: 1. Data preparation Switch the path to fast-rcnn.pytorch ...

Posted by devarticles on Fri, 29 Oct 2021 18:46:39 -0700

[pytorch learning notes] Chapter 4 - neural network

Previous chapter We have learned about automatic gradient autograd. torch.nn can be used to build neural network in pytorch. NN depends on autograd to define the model and differentiate it. nn.Module contains the layer and the method forward(input) that returns output. Artificial Neural Networks (abbreviated as ANNs), also referred to as ne ...

Posted by tmyonline on Sat, 23 Oct 2021 18:31:19 -0700

Train your dataset with yolov5 and deploy yolov5 through flash

Use yolov5 to train your own dataset (detailed process) and deploy yolov5 through flash github project address Use yolov5 to train your own data set (detailed process) and deploy it through flash 1. Prepare data set PASCAL VOC In this paper PASCAL VOC extraction code: 07wp Take the dataset as an example. Put the dataset under the project d ...

Posted by sentback on Sat, 23 Oct 2021 02:48:39 -0700

torch.nn neural network -- use of pooling layer + nonlinear activation function (ReLU and sigmoid) + Sequential() function

nn.Module neural network 4. Pool layer Pooling layer: the pooling function uses the overall statistical characteristics of adjacent outputs at a location to replace the network output at that location. The essence is downsampling to reduce the amount of network parameters Still the old rule, import module and dataset, CIFAR10,batchsize=64: ...

Posted by Zoofu on Fri, 22 Oct 2021 08:17:06 -0700