HYPE: A Benchmark for Human eYe Perceptual Evaluation of Generative Models
Human evaluation for generative models have been ad-hoc.
They propose a standard human benchmark for generative realism that is grounded in psychophysics research in perception.
https://arxiv.org/abs/1904.01121
https://hype.stanford.edu/
Human evaluation for generative models have been ad-hoc.
They propose a standard human benchmark for generative realism that is grounded in psychophysics research in perception.
https://arxiv.org/abs/1904.01121
https://hype.stanford.edu/
arXiv.org
HYPE: A Benchmark for Human eYe Perceptual Evaluation of Generative Models
Generative models often use human evaluations to measure the perceived quality of their outputs. Automated metrics are noisy indirect proxies, because they rely on heuristics or pretrained...
BagNet: Berkeley Analog Generator with Layout Optimizer Boosted with Deep Neural Networks
Hakhamaneshi et al.: https://arxiv.org/abs/1907.10515
#SignalProcessing #MachineLearning #NeuralComputing
Hakhamaneshi et al.: https://arxiv.org/abs/1907.10515
#SignalProcessing #MachineLearning #NeuralComputing
Equilibrium Propagation: Bridging the Gap between Energy-Based Models and Backpropagation
https://www.frontiersin.org/articles/10.3389/fncom.2017.00024/full
https://www.frontiersin.org/articles/10.3389/fncom.2017.00024/full
Frontiers
Equilibrium Propagation: Bridging the Gap between Energy-Based Models and Backpropagation
We introduce Equilibrium Propagation, a learning framework for energy-based models. It involves only one kind of neural computation, performed in both the first phase (when the prediction is made) and the second phase of training (after the target or prediction…
Intelligent artificial agents learning to play 'hide and seek'
https://www.profillic.com/paper/arxiv:1909.07528
https://www.profillic.com/paper/arxiv:1909.07528
Profillic
Profillic: AI models, code & research to supercharge your projects
Explore state-of-the-art in machine learning, AI, and robotics research. Browse models, source code, papers by topics and authors. Connect with researchers and engineers working on related problems in machine learning, deep learning, natural language processing…
rlpyt: A Research Code Base for Deep Reinforcement Learning in PyTorch
Blog by Adam Stookei : https://bair.berkeley.edu/blog/2019/09/24/rlpyt/
#DeepLearning #ReinforcementLearning #PyTorch
Blog by Adam Stookei : https://bair.berkeley.edu/blog/2019/09/24/rlpyt/
#DeepLearning #ReinforcementLearning #PyTorch
The Berkeley Artificial Intelligence Research Blog
rlpyt: A Research Code Base for Deep Reinforcement Learning in PyTorch
The BAIR Blog
Deep Dynamics Models for Learning Dexterous Manipulation
Nagabandi et al.: https://arxiv.org/abs/1909.11652
#Robotics #MachineLearning #ReinforcementLearning
Nagabandi et al.: https://arxiv.org/abs/1909.11652
#Robotics #MachineLearning #ReinforcementLearning
arXiv.org
Deep Dynamics Models for Learning Dexterous Manipulation
Dexterous multi-fingered hands can provide robots with the ability to flexibly perform a wide range of manipulation skills. However, many of the more complex behaviors are also notoriously...
Identifying and eliminating bugs in learned predictive models
https://deepmind.com/blog/article/robust-and-verified-ai
https://deepmind.com/blog/article/robust-and-verified-ai
Deepmind
Towards Robust and Verified AI: Specification Testing, Robust Training, and Formal Verification
One in a series of posts explaining the theories underpinning our research. Bugs and software have gone hand in hand since the beginning of computer programming. Over time, software developers have established a set of best practices for testing and debugging…
AI solution about RealEstate in Singapore
They built on open-source geospatial features and were able to predict Singapore real estate prices with 87% accuracy (i.e., within an error margin of S$100).
They used a) OpenStreetMap (https://download.geofabrik.de/asia/malaysia-singapore-brunei.html)
b) Geomancer for geospatial features (https://stories.thinkingmachin.es/geomancer/)
Link: https://download.geofabrik.de/asia/malaysia-singapore-brunei.html
They built on open-source geospatial features and were able to predict Singapore real estate prices with 87% accuracy (i.e., within an error margin of S$100).
They used a) OpenStreetMap (https://download.geofabrik.de/asia/malaysia-singapore-brunei.html)
b) Geomancer for geospatial features (https://stories.thinkingmachin.es/geomancer/)
Link: https://download.geofabrik.de/asia/malaysia-singapore-brunei.html
stories.thinkingmachin.es
Introducing Geomancer: an open-source library for geospatial feature engineering
Tired of doing all of your geospatial feature engineering from scratch? Don’t fret; we’ve got a magical tool for you!
Extreme Language Model Compression with Optimal Subwords and Shared Projections
Zhao et al.: https://arxiv.org/abs/1909.11687
#neuralnetwork #bert #nlp
Zhao et al.: https://arxiv.org/abs/1909.11687
#neuralnetwork #bert #nlp
arXiv.org
Extremely Small BERT Models from Mixed-Vocabulary Training
Pretrained language models like BERT have achieved good results on NLP tasks, but are impractical on resource-limited devices due to memory footprint. A large fraction of this footprint comes from...
Mathematical Reasoning in Latent Space
Lee et al.: https://arxiv.org/pdf/1909.11851v1.pdf
#Mathematics #Reasoning #LatentSpace
Lee et al.: https://arxiv.org/pdf/1909.11851v1.pdf
#Mathematics #Reasoning #LatentSpace
Learning Pixel Representations for Generic Segmentation. https://arxiv.org/abs/1909.11735
Revisit Knowledge Distillation: a Teacher-free Framework. https://arxiv.org/abs/1909.11723
Augmenting the Pathology Lab: An Intelligent Whole Slide Image Classification System for... https://arxiv.org/abs/1909.11212
arXiv.org
Augmenting the Pathology Lab: An Intelligent Whole Slide Image...
Standard of care diagnostic procedure for suspected skin cancer is
microscopic examination of hematoxylin \& eosin stained tissue by a
pathologist. Areas of high inter-pathologist discordance and...
microscopic examination of hematoxylin \& eosin stained tissue by a
pathologist. Areas of high inter-pathologist discordance and...
Scale-Equivariant Neural Networks with Decomposed Convolutional Filters. https://arxiv.org/abs/1909.11193
Neural reparameterization improves structural optimization
Hoyer et al.: https://arxiv.org/abs/1909.04240
#MachineLearning #NeuralNetworks #Optimization
Hoyer et al.: https://arxiv.org/abs/1909.04240
#MachineLearning #NeuralNetworks #Optimization
arXiv.org
Neural reparameterization improves structural optimization
Structural optimization is a popular method for designing objects such as bridge trusses, airplane wings, and optical devices. Unfortunately, the quality of solutions depends heavily on how the...
Infinite-resolution image generation in the browser
Made by Erlend Gjesteland Ekern using CPPNs, GANs and TensorFlow.js : https://thispicturedoesnotexist.com
#CPPN #GAN #TensorflowJS
Made by Erlend Gjesteland Ekern using CPPNs, GANs and TensorFlow.js : https://thispicturedoesnotexist.com
#CPPN #GAN #TensorflowJS
Thispicturedoesnotexist
This Picture Does Not Exist
Interactive, Infinite-Resolution Image and Video Generation using Generative Adversarial Networks (... in the Browser!)
Representation Wars: Enacting an Armistice through Active Inference
https://www.researchgate.net/publication/333969028_Representation_Wars_Enacting_an_Armistice_through_Active_Inference
https://www.researchgate.net/publication/333969028_Representation_Wars_Enacting_an_Armistice_through_Active_Inference
Artificial Design: Modeling Artificial Super Intelligence with Extended General Relativity and Universal Darwinism via Geometrization for Universal Design Automation
https://openreview.net/forum?id=SyxQ_TEFwS
https://openreview.net/forum?id=SyxQ_TEFwS
OpenReview
Artificial Design: Modeling Artificial Super Intelligence with...
GAMIN: An Adversarial Approach to Black-Box Model Inversion. https://arxiv.org/abs/1909.11835
#weekend_read
Check out this mind-blowing survey on HRL with some really-strong proven hypothesis. Maybe the best till date!
Paper-Title: WHY DOES HIERARCHY (SOMETIMES) WORK SO WELL IN REINFORCEMENT LEARNING?
Link to the paper: https://arxiv.org/pdf/1909.10618.pdf
#GoogleAI #UCB
Four Hypothesis: The four hypotheses may also be categorized as hierarchical training (H1 and H3) and hierarchical exploration (H2 and H4).
(H1) Temporally extended training. High-level actions correspond to multiple environment steps. To the high-level agent, episodes are effectively shorter. Thus, rewards are propagated faster and learning should improve.
(H2) Temporally extended exploration. Since high-level actions correspond to multiple environment steps, exploration in the high-level is mapped to environment exploration which is temporally correlated across steps. This way, an HRL agent explores the environment more efficiently. As a motivating example, the distribution associated with a random (Gaussian) walk is wider when the random noise is temporally correlated.
(H3) Semantic training. High-level actor and critic networks are trained with respect to semantically meaningful actions. These semantic actions are more correlated with future values, and thus easier to learn, compared to training with respect to the atomic actions of the environment. For example, in a robot navigation task it is easier to learn future values with respect to deltas in x-y coordinates rather than robot joint torques.
(H4) Semantic exploration. Exploration strategies (in the simplest case, random action noise) are applied to semantically meaningful actions and are thus more meaningful than the same strategies would be if applied to the atomic actions of the environment. For example, in a robot navigation task, it intuitively makes more sense to explore at the level of x-y coordinates rather than robot joint torques.
TL;DR: A large number of conclusions can be drawn based on empirical analysis. Here are few:-
In terms of the benefits of training, it is clear that training with respect to semantically meaningful abstract actions (H3) has a negligible effect on the success of HRL.
Moreover, temporally extended training (H1) is only important insofar as it enables the use of multi-step rewards, as opposed to training with respect to temporally extended actions.
The main and arguably most surprising, the benefit of the hierarchy is due to exploration. This is evidenced by the fact that temporally extended goal-reaching and agent-switching can enable non-hierarchical agents to solve tasks that otherwise can only be solved.
These results suggest that the empirical effectiveness of hierarchical agents simply reflects the improved exploration that these agents can attain.
Check out this mind-blowing survey on HRL with some really-strong proven hypothesis. Maybe the best till date!
Paper-Title: WHY DOES HIERARCHY (SOMETIMES) WORK SO WELL IN REINFORCEMENT LEARNING?
Link to the paper: https://arxiv.org/pdf/1909.10618.pdf
#GoogleAI #UCB
Four Hypothesis: The four hypotheses may also be categorized as hierarchical training (H1 and H3) and hierarchical exploration (H2 and H4).
(H1) Temporally extended training. High-level actions correspond to multiple environment steps. To the high-level agent, episodes are effectively shorter. Thus, rewards are propagated faster and learning should improve.
(H2) Temporally extended exploration. Since high-level actions correspond to multiple environment steps, exploration in the high-level is mapped to environment exploration which is temporally correlated across steps. This way, an HRL agent explores the environment more efficiently. As a motivating example, the distribution associated with a random (Gaussian) walk is wider when the random noise is temporally correlated.
(H3) Semantic training. High-level actor and critic networks are trained with respect to semantically meaningful actions. These semantic actions are more correlated with future values, and thus easier to learn, compared to training with respect to the atomic actions of the environment. For example, in a robot navigation task it is easier to learn future values with respect to deltas in x-y coordinates rather than robot joint torques.
(H4) Semantic exploration. Exploration strategies (in the simplest case, random action noise) are applied to semantically meaningful actions and are thus more meaningful than the same strategies would be if applied to the atomic actions of the environment. For example, in a robot navigation task, it intuitively makes more sense to explore at the level of x-y coordinates rather than robot joint torques.
TL;DR: A large number of conclusions can be drawn based on empirical analysis. Here are few:-
In terms of the benefits of training, it is clear that training with respect to semantically meaningful abstract actions (H3) has a negligible effect on the success of HRL.
Moreover, temporally extended training (H1) is only important insofar as it enables the use of multi-step rewards, as opposed to training with respect to temporally extended actions.
The main and arguably most surprising, the benefit of the hierarchy is due to exploration. This is evidenced by the fact that temporally extended goal-reaching and agent-switching can enable non-hierarchical agents to solve tasks that otherwise can only be solved.
These results suggest that the empirical effectiveness of hierarchical agents simply reflects the improved exploration that these agents can attain.