Detailed understanding of input, output and parameters of RNN, LSTM and GRU, the basic units of cyclic neural network in pytorch
This article is for those who have deeply understood the mathematical principles and operation process of RNN, LSTM and GRU. If they do not understand its basic idea and process, it may not be very simple to understand.
1, Let's start with an example
This is an example on the official website. Taking LSTM as an example this tim ...
Posted by oocuz on Wed, 10 Nov 2021 21:46:30 -0800
ResNet18 is introduced and used to classify CIFAR-10 data sets
ResNet, an article published by he Kaiming on CVPR in 2015, uses the concept of residual connection. As soon as the paper was published, it directly detonated the whole cv world. And ResNet won the first place on ImageNet in 2016. ResNet has been used in cutting-edge technologies in various fields of AI.
I would be satisfied if I cited one ten ...
Posted by timtom3 on Wed, 10 Nov 2021 10:11:05 -0800
non_max_suppression code analysis
non_max_suppression code analysis
NMS was performed simply according to confidence
def non_max_suppression(boxes, conf_thres=0.5, nms_thres=0.3):
detection = boxes
# 1. Find the box in the picture whose score is greater than the threshold function. The number of coincident boxes can be greatly reduced by filtering scores before screen ...
Posted by dloeppky on Mon, 08 Nov 2021 03:17:38 -0800
2021-11-05 creation and common methods of pytorch Tensor
Creation and common methods of pytorch Tensor
Import pytorch package
import torch
View version number
torch.__version__
'1.7.1'
1, Basic creation and types of tensors
1. Tensor creation method
Use the pytorch tensor creation function: torch.tensor():
# Create tensor from list
t = torch.tensor([1, 2])
t
tensor([1,2])
# Creating tenso ...
Posted by benrussell on Sat, 06 Nov 2021 11:45:54 -0700
Message Passing parsing in Python geometric
MessagePassing in Python geometric
Convolution computation in the graph is usually called neighborhood aggregation or message passing. Definition
x
i
(
k
−
...
Posted by pbarney on Sat, 06 Nov 2021 04:25:16 -0700
Ubuntu 18.04 + Anaconda + CUDA + cudnn + Python environment configuration (updated on 2021.10, available for pro testing)
This set of environment is really too complex. There are too many pits. After more than half a year, it finally succeeded today.
Graphics card driver
Direct installation system recommends graphics card driver, with the lowest error probability.
sudo ubuntu-drivers autoinstall
View installation status
nvidia-smi
The graphics card version ...
Posted by Coronet on Fri, 29 Oct 2021 23:07:37 -0700
[Linux] reproduction of fast r-cnn
0. Download Code using Git
Installation and use of Git: https://blog.csdn.net/qq_44747572/article/details/121006841Clone Code: Pytorch version 1.0.0 source code: https://github.com/jwyang/faster-rcnn.pytorch/tree/pytorch-1.0 Download the zip and unzip it in the specified folder:
1. Data preparation
Switch the path to fast-rcnn.pytorch ...
Posted by devarticles on Fri, 29 Oct 2021 18:46:39 -0700
[pytorch learning notes] Chapter 4 - neural network
Previous chapter We have learned about automatic gradient autograd. torch.nn can be used to build neural network in pytorch. NN depends on autograd to define the model and differentiate it. nn.Module contains the layer and the method forward(input) that returns output.
Artificial Neural Networks (abbreviated as ANNs), also referred to as ne ...
Posted by tmyonline on Sat, 23 Oct 2021 18:31:19 -0700
Train your dataset with yolov5 and deploy yolov5 through flash
Use yolov5 to train your own dataset (detailed process) and deploy yolov5 through flash
github project address
Use yolov5 to train your own data set (detailed process) and deploy it through flash
1. Prepare data set
PASCAL VOC
In this paper PASCAL VOC extraction code: 07wp Take the dataset as an example. Put the dataset under the project d ...
Posted by sentback on Sat, 23 Oct 2021 02:48:39 -0700
torch.nn neural network -- use of pooling layer + nonlinear activation function (ReLU and sigmoid) + Sequential() function
nn.Module neural network
4. Pool layer
Pooling layer: the pooling function uses the overall statistical characteristics of adjacent outputs at a location to replace the network output at that location. The essence is downsampling to reduce the amount of network parameters
Still the old rule, import module and dataset, CIFAR10,batchsize=64:
...
Posted by Zoofu on Fri, 22 Oct 2021 08:17:06 -0700