The largest database of trusted experimental protocols

Tesla p100 gpu

Manufactured by NVIDIA
Sourced in United States

The Tesla P100 GPU is a high-performance computing solution designed for data centers and scientific research. It features 3,584 CUDA cores, 16GB of HBM2 memory, and a maximum power consumption of 250W. The Tesla P100 GPU is capable of delivering up to 10.6 teraflops of peak single-precision performance and is optimized for scientific computing, deep learning, and other GPU-accelerated applications.

Automatically generated - may contain errors

43 protocols using tesla p100 gpu

1

Deep Learning-based Hologram Reconstruction

Check if the same lab product or an alternative is used in the 5 most similar protocols
The hologram reconstruction, synthetic phase-gradient image calculation, and all of the digital preprocessing procedures performed on the images were implemented with MATLAB R2016b. All of the abovementioned processes were done on a desktop computer with an Intel Core i7-2600 CPU @ 3.40 GHz and 8.00 GB RAM, running on a Windows 10 operating system (Microsoft). The deep learning architecture and training/testing procedures were implemented in Python version 3.6.4 using the TensorFlow library version 1.10.0. The training and testing of the network were performed on a Tesla P100 GPU (NVIDIA) using the Google Cloud Platform. The framework was trained for 120 epochs, which lasted 31.5 h. Each image generation lasts ∼0.08 s on a NVIDIA Tesla P100 GPU.
+ Open protocol
+ Expand
2

Comparative Evaluation of Image Analysis Techniques

Check if the same lab product or an alternative is used in the 5 most similar protocols
We evaluated the local, global and global + local approaches, respectively. Particularly, the two global + local techniques14 (link), 15 we examined mainly differ in how image patches are obtained from the whole image (i.e. manually or automatically) and how global features (from the whole image) and local features (from the image patch) are utilised. We also assessed data processing strategies in terms of data quality enhancement and data volume expansion. Our experiments were performed on Swinburne supercomputer OzSTAR* with a cluster of NVIDIA Tesla P100 GPUs.
+ Open protocol
+ Expand
3

Hyperparameter Tuning of TacticAI Models

Check if the same lab product or an alternative is used in the 5 most similar protocols
We train each of TacticAI’s models in isolation, using NVIDIA Tesla P100 GPUs. To minimise overfitting, each model’s learning objective is regularised with an L2 norm penalty with respect to the network parameters. During training, we use the Adam stochastic gradient descent optimiser33 over the regularised loss.
All models, including baselines, have been given an equal hyperparameter tuning budget, spanning the number of message passing steps ({1, 2, 4}), initial learning rate ({0.0001, 0.00005}), batch size ({128, 256}) and L2 regularisation coefficient ({0.01, 0.005, 0.001, 0.0001, 0}). We summarise the chosen hyperparameters of each TacticAI model in Supplementary Table 1.
+ Open protocol
+ Expand
4

Iterative Reconstruction of Slice-Parallel MRI

Check if the same lab product or an alternative is used in the 5 most similar protocols
The acquired k-space data with slice parallel imaging was reconstructed using 1D-GRAPPA along the slice dimension to fill the missing kz partitions. The TE images were then reconstructed using the iterative reconstruction algorithm. The subspace constrained reconstruction problem was solved using the alternating direction method of multipliers (ADMM)44 (link) using 5 PCs. The T2, T1, and B1n ranges for the subspace basis were 20–350 ms, 500–2500 ms, and 0.5–1.6, respectively. For slab-selective excitation, the excitation RF slice profile was computed using the Shinnar-Le Roux algorithm45 (link) and used to generate a slice dependent basis. All reconstruction algorithms were implemented offline using MATLAB (MathWorks, MA) and the Gadgetron framework46 (link) in a hybrid central processing unit/graphics processing unit (CPU-GPU) approach. The ADMM algorithm was implemented on the CPU with the nonuniform fast Fourier transform computed on a GPU at each iteration and the slices were processed in parallel. This resulted in a reconstruction time of ~15 min/slice on a server with 44 cores (2.4 GHz Intel Xeon processor E5-2699A), 440GB RAM, and 8-NVIDIA Tesla P100 GPUs.
+ Open protocol
+ Expand
5

Optimized Deep Learning Model Training

Check if the same lab product or an alternative is used in the 5 most similar protocols
Since we are able to represent every task as a classification problem, in each task we minimize the categorical cross-entropy loss. All models are trained with Adam38 optimizer that is initialized with the default hyperparameters. Furthermore, we use different batch sizes depending on the number of trainable parameters and the available memory on the GPU cards. As a rule of thumb, we try to pick a batch size that allows us to use as much of the GPU memory as possible. All models are trained on Nvidia Tesla P100 GPUs.
+ Open protocol
+ Expand
6

Detecting Plant Leaf Diseases using DCNN

Check if the same lab product or an alternative is used in the 5 most similar protocols
This section includes a comprehensive overview of the proposed DCNN model’s architecture and training method, including the preparation of the dataset and experimental procedures. The suggested model for detecting plant leaf diseases begins with dataset preparation and concludes with model prediction. Python 3.8, TensorFlow Library version 2.10.0, NumPy 1.23.4, matplotlib 3.6.1, and OpenCV 4.6.0 are used to prepare the training dataset and implement the proposed DCNN model, respectively. The simulations, i.e., model development, training, validation, etc., are performed on an HP Z440 workstation consisting of core i7 12 cores of CPU and a DDR4 ram of 48 GB. The proposed scheme also utilized NVidia RTX-3090 Graphical Processor Unit (GPU), which uses the CUDA framework to allow the parallel processing speeds up the proposed model training and testing procedure. The workstation for implementing the proposed DCNN is equipped with a dual Intel Xeon Silver 4310 (12 cores, 24 threads, and 2.10Ghz) processor and six Nvidia Tesla P100 GPUs to expedite the training of deep neural networks. The following sections will explain all the important phases of the proposed plant disease detection framework in detail. The section that follows addresses the specifics of data set preparation and preprocessing.
+ Open protocol
+ Expand
7

High-Performance Computing Protocol

Check if the same lab product or an alternative is used in the 5 most similar protocols
Experiments were performed using a two-socket server with 2 × 16 Intel Xeon cores, 256 GB memory, 54 TB network mounted storage, and two NVIDIA Tesla P100 GPUs.
+ Open protocol
+ Expand
8

Evaluating Single and Dual-View Neural Architectures

Check if the same lab product or an alternative is used in the 5 most similar protocols
To determine the accuracy of both single-view and dual-view architectures, both neural networks are trained on the paired synthesized topogram and volumetric CT dataset. For both models, an initial learning rate of 0.0002 on the Adam optimizer is used to minimize the mean-squared-error (MSE) loss function L(Ypred, Ytruth) via stochastic gradient descent and backpropagation. All training was conducted on two SLI-connected NVIDIA Tesla P100 GPUs, each with 16 GB of VRAM. Model weights are saved locally every 10 epochs, and the training process automatically terminates after convergence. The weights with the lowest average loss are preserved and serialized.
+ Open protocol
+ Expand
9

Supervised Training of Neural Network Worm Matching

Check if the same lab product or an alternative is used in the 5 most similar protocols
The model was trained on 2.304 × 105 semi-synthetic animals derived from recordings of 12 individuals. The model was trained only once and the same trained model was used throughout this work.
Training is as follows. We performed supervised learning with ground truth matches provided by the semi-synthetically generated data. A cross-entropy loss function was used. If neuron i and neuron j were matched by human, the cross-entropy loss function favors the model to output pij=1 . If neuron i and neuron j were not matched, the loss function favors the model to output pij=0 . The model was trained for 12 hr on a 2.40 GHz Intel machine with NVIDIA Tesla P100 GPU.
We trained different models with different hyperparameters and chose the one with best performance. The training curve for each model we trained is shown in Figure 2—figure supplement 1. All the models converged after 12 hr of training. We show the performance of trained models on a held-out validation set consisting of 12,800 semi-synthetic worms in Table 7. We chose the model with 6 layers and 128 dimensional embedding space since it reaches the highest performance and increasing the complexity of the model did not appear to increase the performance dramatically.
+ Open protocol
+ Expand
10

Optimizing CNN for Imbalanced Genomic Data

Check if the same lab product or an alternative is used in the 5 most similar protocols
We trained a CNN using imbalanced and balanced datasets of 18,884 genes from 14,301 (imbalanced) and 35,362 (balanced) samples. A CNN requires input data in a two-dimensional matrix format, therefore a zero-padding to gene vectors was applied (18,884 increased to 19,044) to reshape them into 138 by 138 square matrices. The CNN included 4 back-to-back convolutional blocks, each consisting of a convolutional layer with a 3 × 3 convolutional kernel, followed by batch normalization and ReLU activation. Max pooling was applied after the second and fourth convolutional blocks to reduce the size of the features vector by half. After the last convolutional block, a 34 × 34 × 256 features vector was produced for each input and then flattened into a 295,936 × 1 embeddings vector. Two fully connected layers were then applied to this embeddings vector to gradually reduce the size to 47 × 1, finally arriving at classification softmax layer (Supplementary Fig. S1b). All ML analyses were performed on Nvidia Tesla P100 GPU with a memory capacity of 16 Gigabytes.
+ Open protocol
+ Expand

About PubCompare

Our mission is to provide scientists with the largest repository of trustworthy protocols and intelligent analytical tools, thereby offering them extensive information to design robust protocols aimed at minimizing the risk of failures.

We believe that the most crucial aspect is to grant scientists access to a wide range of reliable sources and new useful tools that surpass human capabilities.

However, we trust in allowing scientists to determine how to construct their own protocols based on this information, as they are the experts in their field.

Ready to get started?

Sign up for free.
Registration takes 20 seconds.
Available from any computer
No download required

Sign up now

Revolutionizing how scientists
search and build protocols!