Categories
Misc

Dream State: Cybersecurity Vendors Detect Breaches in an Instant with NVIDIA Morpheus

In the geography of data center security, efforts have long focused on protecting north-south traffic — the data that passes between the data center and the rest of the network. But one of the greatest risks has become east-west traffic — network packets passing between servers within a data center. That’s due to the growth Read article >

The post Dream State: Cybersecurity Vendors Detect Breaches in an Instant with NVIDIA Morpheus appeared first on The Official NVIDIA Blog.

Categories
Misc

NVIDIA Launches Morpheus to Bring AI-Driven Automation to Cybersecurity Industry

New Framework Powered by NVIDIA GPUs, BlueField DPUs Enables Cybersecurity Providers to Develop AI Solutions That Can Instantly Detect Cyber BreachesSANTA CLARA, Calif., April 12, 2021 (GLOBE …

Categories
Misc

Fast Track to Enterprise AI: New NVIDIA Workflow Lets Any User Choose, Adapt, Deploy Models Easily

AI is the most powerful new technology of our time, but it’s been a force that’s hard to harness for many enterprises — until now. Many companies lack the specialized skills, access to large datasets or accelerated computing that deep learning requires. Others are realizing the benefits of AI and want to spread them quickly Read article >

The post Fast Track to Enterprise AI: New NVIDIA Workflow Lets Any User Choose, Adapt, Deploy Models Easily appeared first on The Official NVIDIA Blog.

Categories
Misc

NVIDIA Announces Availability of Jarvis Interactive Conversational AI Framework

Pre-Trained Deep Learning Models and Software Tools Enable Developers to Adapt Jarvis for All Industries; Easily Deployed from Any Cloud to EdgeSANTA CLARA, Calif., April 12, 2021 (GLOBE …

Categories
Misc

NVIDIA Launches Omniverse Design Collaboration and Simulation Platform for Enterprises

Leading Computer Makers Launch Workstations and NVIDIA-Certified Systems for Omniverse; BMW Group, Ericsson, Foster + Partners, WPP Among Early AdoptersSANTA CLARA, Calif., April 12, 2021 …

Categories
Misc

Model was constructed with shape (1, 16, 1), but it was called on an input with incompatible shape (1, 1, 1)

I’m new to deep learning and I trying to model a univariate time series using the sliding window approach with a LSTM model. My training dataset takes 16 values to predict the next 16. My code is writing in R.

I am getting a warning and I cannot understand what I am doing wrong. I think that the problem is when specifying the model.

I am totally new to this. So if you could help me would be great.

Bellow is the whole code. I got the warning at the very end, after predicting

train_sliding = create_dataset(data = kt_train_male_scaled, n_input = 16, n_out = 16)

X_train = train_sliding[[1]] #97, 16

y_train = train_sliding[[2]] #97, 16

#Array transformation to Keras LSTM

dim(X_train) = c(dim(X_train), 1)

dim(X_train) # 97, 16, 1

I think the problem should be in this chunk of code. I think I am building the model wrong

#Model in Keras

X_shape2 = dim(X_train)[2] #16

X_shape3 = dim(X_train)[3] #1

batch_size = 1

model <- keras_model_sequential()

model%>%

layer_lstm(units = 64, activation = “relu”, batch_size = batch_size, input_shape = c(dim(X_train)[2], dim(X_train)[3]),stateful= TRUE)%>%

#layer_lstm(units = 5, activation = “relu”, stateful= TRUE) %>%

layer_dense(units = 1)

summary(model)

model %>% compile(

loss = ‘mse’,

optimizer = optimizer_adam(lr= 0.01, decay = 1e-6 ),

metrics = c(‘mae’)

)

Epochs = 100

for(i in 1:Epochs ){

model %>% fit(X_train, y_train, epochs=1, batch_size=batch_size, verbose=1, shuffle=FALSE)

model %>% reset_states()

}

L = length(kt_test_male_scaled)

scaler = Scaled$scaler

predictions = numeric(L)

I get the warning after running this part. Also, all my 16 predictions have the same value. I also tried to use dim(X) = c(1,16,1) but it did not work

for(i in 1:L){

X = kt_test_male_scaled[i]

dim(X) = c(1,1,1)

yhat = model %>% predict(X, batch_size=batch_size)

# invert scaling

yhat = invert_scaling(yhat, scaler, c(-1, 1))

# invert differencing

#yhat = yhat + kt_male[(n+i)]

# store

predictions[i] <- yhat

}

Model was constructed with shape (1, 16, 1) for input KerasTensor(type_spec=TensorSpec(shape=(1, 16, 1), dtype=tf.float32, name=’lstm_107_input’), name=’lstm_107_input’, description=”created by layer ‘lstm_107_input'”), but it was called on an input with incompatible shape (1, 1, 1)

submitted by /u/rods2292
[visit reddit] [comments]

Categories
Misc

Deep Learning with TensorFlow – Free course from udemy

Deep Learning with TensorFlow - Free course from udemy submitted by /u/Ordinary_Craft
[visit reddit] [comments]
Categories
Misc

AM I the only person deeply irritated by the logo of TF with the shadow?

It doesn’t make sense. the shadow of the corner T looks like it has 5 cubic lengths when the physical figure’s T has 1 cubic length on the top right.

submitted by /u/Silver4R4449
[visit reddit] [comments]

Categories
Misc

How many objects can be detected using Tensorflow?

submitted by /u/NickLRealtor
[visit reddit] [comments]

Categories
Misc

Inception Spotlight: Deepset collaborates with NVIDIA and AWS on BERT Optimization

Deepset bridges the gap between NLP research and industry – their core product, Haystack, is an open-source framework that enables developers to utilize the latest NLP models for semantic search and question answering at scale.

Language models are essential for modern NLP. Building a new language model from scratch can be beneficial for many domains. NVIDIA Inception member Deepset bridges the gap between NLP research and industry – their core product, Haystack, is an open-source framework that enables developers to utilize the latest NLP models for semantic search and question answering at scale. Haystack Hub, is their software as a service (SaaS) platform, used by developers from various industries, including finance, legal, and automotive, to find answers in all kinds of text documents. 

In a collaborative effort with NVIDIA and AWS, deepset used NVIDIA V100 GPUs for training their language model. The GPU performance profiles were captured by the NVIDIA Nsight Systems.

The collaboration was a product of the partnership between NVIDIA Inception and AWS Activate, an initiative to support AI startups by providing access to the benefits of both acceleration programs. The benefits for NVIDIA Inception startups joining AWS Activate include business and marketing support, as well as AWS Cloud credits, which can be used to access NVIDIA’s latest generation GPUs in Amazon EC2 – P3 Instances. AWS Activate members that are using AI and machine learning are referred to NVIDIA Inception and can benefit from immediate preferred pricing on NVIDIA GPUs and Deep Learning Institute credits.

“A considerable amount of manual development is required to create the training data and vocabulary, configure hyperparameters, start and monitor training jobs, and run periodical evaluation of different model checkpoints. In our first training runs, we also found several bugs only after multiple hours of training, resulting in a slow development cycle. In summary, language model training can be a painful job for a developer and easily consumes multiple days of work”.

“The increased efficiency of training jobs reduces our energy usage and lowers our carbon footprint. By tackling different areas of FARM’s training pipeline, we were able to significantly optimize the resource utilization. In the end, we were able to achieve a speedup in training time of 3.9 times faster, a 12.8 times reduction in training cost, and reduced the developer effort required from days to hours”.

Collaborating with NVIDIA and AWS, NVIDIA Inception partner deeepset achieves a 3.9x speedup and 12.8x cost reduction for training NLP models. As a result, the developer effort was significantly reduced.

Read more about technologies used in the training and their impact on improving BERT training performance.