Categories
Misc

Can’t use tensorflow 2 (I need tensorflow 2cant use 1) because of no protobuf version working for it.

If I have any other version of protobuf except for 3.6.0 I will
get “ImportError: DLL load failed: The specified procedure could
not be found” but if I use protobuf 3.6.0 I get
“AttributeError:
‘google.protobuf.pyext._message.RepeatedCompositeCo’ object has
no attribute ‘append’” this error occurs when I try to build
the model.

I have tried every 2.x version of tensorflow have reinstalled
python 3.6 I have made sure my path variables are correct. I can
find no useful information on the internet. I have tried countless
versions of protobuf. Please help! I have no clue what the hell is
going on.

Maybe upgrade python 3.6 to 3.7? as I have previously had
tensorflow 2.x working on python 3.7 but I don’t know.

submitted by /u/FunnyForWrongReason

[visit reddit]

[comments]

Categories
Misc

Fastest way to develop a custom translation model with RNN?

I’m a Python web developer, so I have some professional coding
experience, but I’m a complete novice when it comes to machine
learning.

In short, I have a dataset (csv form) with 65,000 sentences in
two languages. One of the languages is real, the other is not. I’d
live to quickly dive into an RNN example online so that I can train
a model based on this dataset, but all of the examples seem to
prefer that I use existing, binary datasets (that I can’t
read).

My laptop is relatively old, and processing a dataset properly
can take a week, so every example I’ve attempted to adapt to my
needs has cost lots of time and lots of heartache when I discover
that I can’t use it.

Is there an RNN translation tutorial that anyone would recommend
for the purpose of translating between an existing corpus and a
constructed language? I can do research on any terms listed below,
but the topic of machine learning has so regularly stumped me that,
even though I know easy examples for what I want to do probably
already exist, I don’t even know where to start.

Thank you for your time!

submitted by /u/ehowardhill

[visit reddit]

[comments]

Categories
Misc

Any tutorials that you can recommend?

So I understood the attention mechanism ( Bahdanau Attention,
2017 paper) and I was looking for the implementation of the paper
and then I landed on the tensorflow website which has a tutorial on
attention mechanism. Nut Frankly speaking, I found it very hard to
understand the code. Are there any tutorials that you can share
that will help me to understand the code of the attention
mechanism.

submitted by /u/Consistent_Ad767

[visit reddit]

[comments]

Categories
Offsites

RxR: A Multilingual Benchmark for Navigation Instruction Following

A core challenge in machine learning (ML) is to build agents that can navigate complex human environments in response to spoken or written commands. While today’s agents, including robots, can often navigate complicated environments, they cannot yet understand navigation goals expressed in natural language, such as, “Go past the brown double doors that are closed to your right and stand behind the chair at the head of the table.”

This challenge, referred to as vision-and-language navigation (VLN), demands a sophisticated understanding of spatial language. For example, the ability to identify the position “behind the chair at the head of the table requires finding the table, identifying which part of the table is considered to be the “head”, finding the chair closest to the head, identifying the area behind this chair and so on. While people can follow these instructions easily, these challenges cannot be easily solved with current ML-based methods, requiring systems that can better connect language to the physical world it describes.

To help spur progress in this area, we are excited to introduce Room-Across-Room (RxR), a new dataset for VLN. Described in “Room-Across-Room: Multilingual Vision-and-Language Navigation with Dense Spatiotemporal Grounding”, RxR is the first multilingual dataset for VLN, containing 126,069 human-annotated navigation instructions in three typologically diverse languages — English, Hindi and Telugu. Each instruction describes a path through a photorealistic simulator populated with indoor environments from the Matterport3D dataset, which includes 3D captures of homes, offices and public buildings. To track progress on VLN, we are also announcing the RxR Challenge, a competition that encourages the machine learning community to train and evaluate their own instruction following agents on RxR instructions.

Language Instruction
en-US Starting next to the long dining room table, turn so the table is to your right. Walk towards the glass double doors. When you reach the mat before the doors, turn immediately left and walk down the stairs. When you reach the bottom of the stairs, walk through the open doors to your left and continue through the art exhibit with the tub to your right hand side. Down the length of the table until you reach the small step at the end of the room before you reach the tub and stop.
   
hi-IN अभी हमारे बायीं ओर एक बड़ा मेज़ है कुछ कुर्सियाँ हैं और कुछ दीपक मेज़ के ऊपर रखे हैं। उलटी दिशा में घूम जाएँ और सिधा चलें। अभी हमारे दायीं ओर एक गोल मेज़ है वहां से सीधा बढ़ें और सामने एक शीशे का बंद दरवाज़ा है उससे पहले बायीं ओर एक सीढ़ी है उससे निचे उतरें। निचे उतरने के बाद दायीं ओर मुड़े और एक भूरे रंग के दरवाज़े से अंदर प्रवेश करें और सीधा चलें। अभी हमारे दायीं ओर एक बड़ा मेज़ है और दो कुर्सियां राखी हैं सीधा आगे बढ़ें। हमारे सामने एक पानी का कल है और सामने तीन कुर्सियां दिवार के पास रखी हैं यहीं पर ठहर जाएँ।
   
te-IN ఉన్న చోటు నుండి వెనకకు తిరిగి, నేరుగా వెళ్తే, మీ ముందర ఒక బల్ల ఉంటుంది. దాన్ని దాటుకొని ఎడమవైపుకి తిరిగితే, మీ ముందర మెట్లు ఉంటాయి. వాటిని పూర్తిగా దిగండి. ఇప్పుడు మీ ముందర రెండు తెరిచిన ద్వారాలు ఉంటాయి. ఎడమవైపు ఉన్న ద్వారం గుండా బయటకు వెళ్ళి, నేరుగా నడవండి. ఇప్పుడు మీ కుడివైపున పొడవైన బల్ల ఉంటుంది. దాన్ని దాటుకొని ముందరే ఉన్న మెట్ల వద్దకు వెళ్ళి ఆగండి.

Examples of English, Hindi and Telugu navigation instructions from the RxR dataset. Each navigation instruction describes the same path.

Pose Traces
In addition to navigation instructions and paths, RxR also includes a new, more detailed multimodal annotation called a pose trace. Inspired by the mouse traces captured in the Localized Narratives dataset, pose traces provide dense groundings between language, vision and movement in a rich 3D setting. To generate navigation instructions, we ask guide annotators to move along a path in the simulator while narrating the path based on the surroundings. The pose trace is a record of everything the guide sees along the path, time-aligned with the words in the navigation instructions. These traces are then paired with pose traces from follower annotators, who are tasked with following the intended path by listening to the guide’s audio, thereby validating the quality of the navigation instructions. Pose traces implicitly capture notions of landmark selection and visual saliency, and represent a play-by-play account of how to solve the navigation instruction generation task (for guides) and the navigation instruction following task (for followers).

Example English navigation instruction in the RxR dataset. Words in the instruction text (right) are color-coded to align with the pose trace (left) that illustrates the movements and visual percepts of the guide annotator as they move through the environment describing the path.
The same RxR example with words in the navigation instruction aligned to 360° images along the path. The parts of the scene the guide annotator observed are highlighted; parts of the scene ignored by the annotator are faded. Red and yellow boxes highlight some of the close alignments between the textual instructions and the annotator’s visual cues. The red cross indicates the next direction the annotator moved.

Scale
In total, RxR contains almost 10 million words, making it around 10 times larger than existing datasets, such as R2R and Touchdown/Retouchdown. This is important because, in comparison to tasks based on static image and text data, language tasks that require learning through movement or interaction with an environment typically suffer from a lack of large-scale training data. RxR also addresses known biases in the construction of the paths that have arisen in other datasets, such as R2R in which all paths have similar lengths and take the shortest route to the goal. In contrast, the paths in RxR are on average longer and less predictable, making them more challenging to follow and encouraging models trained on the dataset to place greater emphasis on the role of language in the task. The size, scope and detail of RxR will expand the frontier for research on grounded language learning while reducing the dominance of high resource languages such as English.

Left: RxR is an order of magnitude larger than similar existing datasets. Right: Compared to R2R, the paths in RxR are typically longer and less predictable, making them more challenging to follow.

Baselines
To better characterize and understand the RxR dataset, we trained a variety of agents on RxR using our open source framework VALAN, and language representations from the multilingual BERT model. We found that results were improved by including follower annotations as well as guide annotations during training, and that independently trained monolingual agents outperformed a single multilingual agent.

Conceptually, evaluation of these agents is straightforward — did the agent follow the intended path? Empirically, we measure the similarity between the path taken by the VLN agent and the reference path using NDTW, a normalized measure of path fidelity that ranges between 100 (perfect correspondence) and 0 (completely wrong). The average score for the follower annotators across all three languages is 79.5, due to natural variation between similar paths. In contrast, the best model (a composite of three independently trained monolingual agents, one for each language) achieved an NDTW score on the RxR test set of 41.5. While this is much better than random (15.4), it remains far below human performance. Although advances in language modeling continue to rapidly erode the headroom for improvement in text-only language understanding benchmarks such as GLUE and SuperGLUE, benchmarks like RxR that connect language to the physical world offer substantial room for improvement.

Results for our multilingual and monolingual instruction following agents on the RxR test-standard split. While performance is much better than a random walk, there remains considerable headroom to reach human performance on this task.

Competition
To encourage further research in this area, we are launching the RxR Challenge, an ongoing competition for the machine learning community to develop computational agents that can follow natural language navigation instructions. To take part, participants upload the navigation paths taken by their agent in response to the provided RxR test instructions. In the most difficult setting (reported here and in the paper), all the test environments are previously unseen. However, we also allow for settings in which the agent is either trained in or explores the test environments in advance. For more details and the latest results please visit the challenge website.

PanGEA
We are also releasing the custom web-based annotation tool that we developed to collect the RxR dataset. The Panoramic Graph Environment Annotation toolkit (PanGEA), is a lightweight and customizable codebase for collecting speech and text annotations in panoramic graph environments, such as Matterport3D and StreetLearn. It includes speech recording and virtual pose tracking, as well as tooling to align the resulting pose trace with a manual transcript. For more details please visit the PanGEA github page.

Acknowledgements
The authors would like to thank Roma Patel, Eugene Ie and Jason Baldridge for their contributions to this research. We would also like to thank all the annotators, Sneha Kudugunta for analyzing the Telugu annotations, and Igor Karpov, Ashwin Kakarla and Christina Liu for their tooling and annotation support for this project, Austin Waters and Su Wang for help with image features, and Daphne Luong for executive support for the data collection.

Categories
Misc

New DRIVE OS and DriveWorks Updates Enable Streamlined AV Software Development

DRIVE OS and DriveWorks releases are now available on NVIDIA DRIVE Developer, providing DRIVE OS users access to DriveWorks middleware and even more updates.

You asked, we listened: DRIVE OS and DriveWorks releases are now available on NVIDIA DRIVE Developer, providing DRIVE OS users access to DriveWorks middleware and even more updates.

With these releases, developers have access to the latest DRIVE OS and DriveWorks software for autonomous vehicle development, including new features, without having to wait for DRIVE Software updates.

The foundation of the NVIDIA DRIVE software stack, NVIDIA DRIVE OS is designed specifically for accelerated computing and artificial intelligence. It includes NvMedia for sensor input processing, NVIDIA CUDA for efficient parallel computing implementations, NVIDIA TensorRT™ for real-time AI inference, and specialized developer tools and modules that allow developers to access the accelerated hardware engines. 

The NVIDIA DriveWorks SDK provides functionality fundamental to autonomous vehicle development, consisting of a sensor abstraction layer (SAL), sensor plugins, data recorder, vehicle I/O support, and a deep neural network (DNN) framework. It’s modular, open and designed to be compliant with automotive industry software standards.

And now, these key components for autonomous vehicle software development are even more accessible to developers, with frequent updates to unlock performance on the NVIDIA DRIVE AGX platform for greater flexibility and capabilities.

Laying a Foundation with DRIVE OS

DRIVE OS is a robust operating system for autonomous vehicle development, providing access to the underlying compute accelerators in DRIVE AGX Xavier.

New for this release is NvMedia Sensor Input Processing Library (SIPL), an image processing API, targeted for safety. SIPL adds sensor device and query block sources, in addition to source files and libraries to an expanded range of sensor modules. It also delivers safety proxy support for Linux — a mechanism that makes it easier to develop safety applications on non-safety platforms. 

Updated for this release, NvStreams enables efficient allocation, sharing and synchronizing of data buffers across the SoC, dGPU and CPU engine APIs, making it easy for developers to move large data buffers for processing. 

Also included is the latest TensorRT with dynamic shape, reformat-free I/O, explicit precision, pointwise layer fusion and shuffle elimination, as well as new plugins and samples to help developers take advantage of the platform.

Going Further with DriveWorks

On top of DRIVE OS, DriveWorks enables applications to help incorporate software into the vehicle. These include integrating automotive sensors within the software stack, accelerating camera and lidar data processing, interfacing with the vehicle, accelerating inference for perception and calibrating multiple sensor modalities with precision.

Key DriveWorks highlights include the integration of the DriveWorks SAL with NvMedia SIPL, enabling recording of additional GMSL cameras such as the Sony IMX390 and the ON Semi AR0820. Additionally, DriveWorks SAL now supports even more sensors out-of-the-box, such as the Luminar H3 and Ouster OS2-128 lidars as well as the U-blox ZED-F9P GNSS module. 

As always, developers can integrate their own sensors into DriveWorks using the Sensor Plugin Framework. 

Finally, the DriveWorks SAL also now includes a new Time Sensor module for synchronizing timestamps of supported sensors. This module maintains time correlation information correspondence data and supports the conversion between the different clocks used to timestamp sensor data.  

For image processing, the release adds new algorithms to run on the NVIDIA DRIVE AGX Programmable Vision Accelerator (PVA) in addition to the GPU. A new DNN tensor module wraps raw tensor data into a structure, allowing the user to define dimensions and layouts. It also supports the traversal of complex layouts as well as the ability to lock/unlock the data to prevent simultaneous operations. 

By making DRIVE OS and DriveWorks releases available together, developers now have the latest and greatest DRIVE OS features and performance together with seamless autonomous vehicle integration and utilities provided by DriveWorks SDK.

DRIVE AGX developers may access the latest release here.

Categories
Misc

How to Optimize Self-Driving DNNs with TensorRT

Register for our upcoming webinar to learn how to use TensorRT to optimize autonomous driving DNNs for robust AV development.

When it comes to autonomous vehicle development, to ensure the highest level of safety, one of the most important areas of evaluation is performance.

High-performance, energy-efficient compute enables developers to balance the complexity, accuracy and resource consumption of the deep neural networks (DNN) that run in the vehicle. Getting the most out of hardware computing power requires optimized software.

NVIDIA TensorRT is an SDK for high-performance deep learning inference. It includes a deep learning inference optimizer and runtime that delivers low latency and high-throughput for deep learning inference applications, such as autonomous driving.

You can register for our upcoming webinar on Feb. 3 to learn how to use TensorRT to optimize autonomous driving DNNs for robust autonomous vehicle development.

Manage Massive Workloads

DNN-based workloads in autonomous driving are incredibly complex, with a variety of computation-intensive layer operations just to perform computer vision tasks. 

Managing these types of operations requires optimized compute performance, however, it isn’t always the case that the theoretical peak performance of hardware translates to any software achievable execution. TensorRT ensures developers can tackle these massive workloads without leaving any performance on the table.

By performing optimization at every stage of processing — from tooling, to ingesting DNNs, to inference — TensorRT ensures the most efficient operations possible.

The SDK is also seamless to use, allowing developers to toggle different settings depending on the platform. For example, lower precision, i.e., FP16 or INT8, is used to enable higher compute throughput and lower memory bandwidth on Tensor Core. In addition, workloads can be shifted from the GPU to the deep learning accelerator (DLA).

Master the Model Backbone

This webinar will show how TensorRT for AV development works in action, tackling one of the chunkiest portions in the inference pipeline — the model backbone.

Many developers use off-the-shelf model backbones (for example, ResNets or EfficientNets) to get started on solving computer vision tasks such as object detection or semantic segmentation. However, these backbones aren’t always performance-optimized, creating bottlenecks down the line. TensorRT addresses these problems by optimizing trained neural networks to generate deployment-ready inference engines that maximize GPU inference performance and power efficiency.

Learn from NVIDIA experts how to leverage these tools in autonomous vehicle development. Register today for the Feb 3rd webinar, plus catch up on past TensorRT and DriveWorks webinars.

Categories
Misc

On the Road Again: GeForce NOW Alliance Expanding to Turkey, Saudi Arabia and Australia

Bringing more games to more gamers, our GeForce NOW game-streaming service is coming soon to Turkey, Saudi Arabia and Australia. Turkcell, Zain KSA and Pentanet are the latest telcos to join the GeForce NOW Alliance. By placing NVIDIA RTX Servers on the edge, GeForce NOW Alliance partners deliver even lower latency gaming experiences. And this Read article >

The post On the Road Again: GeForce NOW Alliance Expanding to Turkey, Saudi Arabia and Australia appeared first on The Official NVIDIA Blog.

Categories
Misc

Take Note: Otter.ai CEO Sam Liang on Bringing Live Captions to a Meeting Near You

Sam Liang is making things easier for the creators of the NVIDIA AI Podcast — and just about every remote worker. He’s the CEO and co-founder of Otter.ai, which uses AI to produce speech-to-text transcriptions in real time or from recording uploads. The platform has a range of capabilities, from differentiating between multiple people, to Read article >

The post Take Note: Otter.ai CEO Sam Liang on Bringing Live Captions to a Meeting Near You appeared first on The Official NVIDIA Blog.

Categories
Misc

Batch training in tf 2.0

When performing custom batch training in the training loop,
which one should be used?

tf.gradient_tape or train_on_batch?

What is the difference?

submitted by /u/SuccMyStrangerThings

[visit reddit]

[comments]

Categories
Misc

I got this error while trying to run the webcam_demo.py example in Posenet library from tensorflow. how to resolve this? #46575

I got this error/warning while trying to run the webcam_demo.py
example in Posenet library from Tensorflow. how to resolve
this?

This is the Git Repo from where I forked this
code : posenet-python

and This is my Output Screen :

>>>

RESTART: A:PythonScriptsPosenet-Forked —
OGCodeposenet-python-masterwebcam_demo.py

Cannot find model file ./_modelsmodel-mobilenet_v1_101.pb,
converting from tfjs…

WARNING:tensorflow:From
A:Pythonlibsite-packagestensorflowpythontoolsfreeze_graph.py:127:
checkpoint_exists (from
tensorflow.python.training.checkpoint_management) is deprecated and
will be removed in a future version.

Instructions for updating:

Use standard file APIs to check for files with this prefix.

Traceback (most recent call last):

File “A:PythonScriptsPosenet-Forked —
OGCodeposenet-python-masterwebcam_demo.py”, line 66, in
<module>

main()

File “A:PythonScriptsPosenet-Forked —
OGCodeposenet-python-masterwebcam_demo.py”, line 20, in main

model_cfg, model_outputs = posenet.load_model(args.model,
sess)

File “A:PythonScriptsPosenet-Forked —
OGCodeposenet-python-masterposenetmodel.py“, line 42, in load_model

convert(model_ord, model_dir, check=False)

File “A:PythonScriptsPosenet-Forked —
OGCodeposenet-python-masterposenetconvertertfjs2python.py“, line 198, in
convert

initializer_nodes=””)

File
“A:Pythonlibsite-packagestensorflowpythontoolsfreeze_graph.py”,
line 361, in freeze_graph

checkpoint_version=checkpoint_version)

File
“A:Pythonlibsite-packagestensorflowpythontoolsfreeze_graph.py”,
line 190, in freeze_graph_with_def_protos

var_list=var_list, write_version=checkpoint_version)

File
“A:Pythonlibsite-packagestensorflowpythontrainingsaver.py“, line 835, in __init__

self.build()

File
“A:Pythonlibsite-packagestensorflowpythontrainingsaver.py“, line 847, in build

self._build(self._filename, build_save=True,
build_restore=True)

File
“A:Pythonlibsite-packagestensorflowpythontrainingsaver.py“, line 885, in _build

build_restore=build_restore)

File
“A:Pythonlibsite-packagestensorflowpythontrainingsaver.py“, line 489, in _build_internal

names_to_saveables)

File
“A:Pythonlibsite-packagestensorflowpythontrainingsavingsaveable_object_util.py”,
line 362, in validate_and_slice_inputs

for converted_saveable_object in saveable_objects_for_op(op,
name):

File
“A:Pythonlibsite-packagestensorflowpythontrainingsavingsaveable_object_util.py”,
line 223, in saveable_objects_for_op

yield ResourceVariableSaveable(variable, “”, name)

File
“A:Pythonlibsite-packagestensorflowpythontrainingsavingsaveable_object_util.py”,
line 95, in __init__

self.handle_op = var.op.inputs[0]

IndexError: tuple index out of range

>>>

My Git
Issue Link

submitted by /u/Section_Disastrous

[visit reddit]

[comments]