MONTRÉAL.AI | Montréal Artificial Intelligence

Montréal.AI Academy

ARTIFICIAL INTELLIGENCE 101

AI 101: The First World-Class Overview of AI for All

Location: NRH Prince Arthur - Ballroom B, 3625 Avenue du Parc, Montreal (Québec), Canada, H2X 3P8.

Tickets selling fast! THIS EVENT WILL SELL OUT.

Encompassing all facets of AI, the General Secretariat of MONTREAL.AI presents, with authority and from insider knowledge: “Artificial Intelligence 101: The First World-Class Overview of AI for the General Public“.

Artificial Intelligence 101: The First World-Class Overview of AI for the General Public

Eventbrite - Montréal.AI Academy: AI 101

AI opens up a world of new possibilities. This AI 101 tutorial harnesses the fundamentals of artificial intelligence for the purpose of providing participants with powerful AI tools to learn, deploy and scale AI.

Breakthrough in machine learning would be worth 10 Microsofts.“ — Bill Gates

A Well-Crafted Actionable 75 Minutes Tutorial

POWERFUL & USEFUL. This actionable tutorial is designed to entrust participants with the mindset, the skills and the tools to see artificial intelligence from an empowering new vantage point by :

Exalting state of the art discoveries and science ;
Curating the best open-source codes & implementations ; and
Embodying the impetus that drives today’s artificial intelligence.

Program overview (“Pioneering an Impactful Understanding of AI”)

Opening Address

Session 0 · Getting Started

  • In the Cloud
  • On a Local Machine

Session 1 · Deep Learning

  • Neural Networks
  • Convolution Neural Networks (CNNs)
  • Recurrent Neural Networks (RNNs)
  • Unsupervised Deep Learning (Self-Supervised Learning, Generative Adversarial Nets and Variational Autoencoders)

Session 2 · Autonomous Agents

  • Evolution Strategies
  • Deep Reinforcement Learning
  • Self Play: A Quantum Leap
  • Deep Meta-Learning

Session 3 · Environments

  • OpenAI Gym
  • DeepMind Lab
  • Unity ML-Agents

Special Address

Tickets: https://montrealai101.eventbrite.ca

Language: Tutorial given in English.

Location: NRH Prince Arthur - Ballroom B, 3625 Avenue du Parc, Montreal (Québec), Canada, H2X 3P8.

The best way to predict the future is to invent it.“ — Alan Kay

VIP AI 101 CheatSheet for All

For the purpose of entrusting all sentient beings with powerful AI tools to learn, deploy and scale AI in order to enhance their prosperity, to settle planetary-scale problems and to inspire those who, with AI, will shape the 21st Century, Montréal.AI introduces the “VIP AI 101 CheatSheet for All”.

Artificial Intelligence 101 International Webinar

Location: This is an online event. Attendees will receive a link to join the event two (2) weeks prior to the event.

ARTIFICIAL INTELLIGENCE 101 INTERNATIONAL WEBINAR

Eventbrite - ARTIFICIAL INTELLIGENCE 101 INTERNATIONAL WEBINAR

Program overview (“Pioneering an Impactful Understanding of AI”) and tickets: https://ai101webinar.eventbrite.ca

Language: Tutorial (webinar) given in English.

Curated Open-Source Codes, Implementations and Science

Montréal.AI is the largest artificial intelligence community in Canada. Join us and learn at https://www.facebook.com/groups/MontrealAI/ !

Curated Open-Source Codes, Implementations and Science

You are qualified for a career in machine learning!

0. Getting Started

AI opens up a world of new possibilities. Today’s artificial intelligence is powerful, useful and accessible to all.

Tinker with Neural Networks : Neural Network Playground — TensorFlow

In the Cloud

Free GPU compute via Colab.

Colaboratory is a hosted Jupyter notebook environment that is free to use and requires no setup.

On a Local Machine

Install Anaconda and Launch ‘Anaconda Navigator’.

JupyterLab is Ready for Users.

Update Jupyterlab and Launch the Application. Under Notebook, Click on ‘Python 3’.

In the Browser

TensorFlow.js: a library for developing and training ML models in JavaScript.

Datasets

Making it easier to discover datasets.

Google Dataset Search

Preliminary Readings

When you first study a field, it seems like you have to memorize a zillion things. You don’t. What you need is to identify the 3-5 core principles that govern the field. The million things you thought you had to memorize are various combinations of the core principles.“ — J. Reed

The Loss Landscape. Source: Visuals by Javier Ideami.

1. Multiply things together 2. Add them up 3. Replaces negatives with zeros 4. Return to step 1, a hundred times.“ — Jeremy Howard

Types of Learning, by Alex Graves at NeurIPS 2018

An embedding is a mapping from discrete objects, such as words, to vectors of real numbers.

1. Deep Learning

Deep learning allows computational models that are composed of multiple processing layers to learn REPRESENTATIONS of (raw) data with multiple levels of abstraction. At a high-level, neural networks are either encoders, decoders, or a combination of both.

DL is essentially a new style of programming–”differentiable programming”–and the field is trying to work out the reusable constructs in this style. We have some: convolution, pooling, LSTM, GAN, VAE, memory units, routing units, etc.“ — Thomas G. Dietterich

1.1 Neural Networks

Neural networks” are a sad misnomer. They’re neither neural nor even networks. They’re chains of differentiable, parameterized geometric functions, trained with gradient descent (with gradients obtained via the chain rule). A small set of highschool-level ideas put together.“ — François Chollet

Forward pass in a neural networks with two hidden layers

Backward pass in a neural networks with two hidden layers

1.1.1 Universal Approximation Theorem

The universal approximation theorem states that a feed-forward network with a single hidden layer containing a finite number of neurons can solve any given problem to arbitrarily close accuracy as long as you add enough parameters.

The chain rule applied to one single unit. Image source: Markus Völk.

Neural Networks + Gradient Descent + GPU:

When a choice must be made, just feed the (raw) data to a deep neural network (universal function approximator).

1.2 Convolution Neural Network

In images, local combinations of edges form motifs, motifs assemble into parts, and parts form objects.

The deep convolutional network, inspired by Hubel and Wiesel’s seminal work on early visual cortex, uses hierarchical layers of tiled convolutional filters to mimic the effects of receptive fields, thereby exploiting the local spatial correlations present in images.

Architecture of LeNet-5, a Convolutional Neural Network. LeCun et al., 1998

A ConvNet is made up of Layers. Every Layer has a simple API: It transforms an input 3D volume to an output 3D volume with some differentiable function that may or may not have parameters.

I admire the elegance of your method of computation; it must be nice to ride through these fields upon the horse of true mathematics while the like of us have to make our way laboriously on foot.“ — A. Einstein

Interactive Inceptionv3 created by TensorSpace https://tensorspace.org

1.3 Recurrent Neural Networks

For sequential inputs. Recurrent neural networks are networks with loops in them, allowing information to persist. RNNs process an input sequence one element at a time, maintaining in their hidden units a ‘state vector’ that implicitly contains information about the history of all the past elements of the sequence.

RNN Layers Reuse Weights for Multiple Timesteps

Google Smart Reply System is built on a pair of recurrent neural networks. Diagram by Chris Olah

  • The Unreasonable Effectiveness of Recurrent Neural Networks — Andrej Karpathy
  • Massive Exploration of Neural Machine Translation Architectures arXiv | Docs | Code — Denny Britz, Anna Goldie, Minh-Thang Luong, Quoc Le
  • A TensorFlow implementation of : “Hybrid computing using a neural network with dynamic external memory” GitHub — Alex Graves, Greg Wayne, Malcolm Reynolds, Tim Harley, Ivo Danihelka, Agnieszka Grabska-Barwińska, Sergio Gómez Colmenarejo, Edward Grefenstette, Tiago Ramalho, John Agapiou, Adrià Puigdomènech Badia, Karl Moritz Hermann, Yori Zwols, Georg Ostrovski, Adam Cain, Helen King, Christopher Summerfield, Phil Blunsom, Koray Kavukcuoglu & Demis Hassabis

I feel like a significant percentage of Deep Learning breakthroughs ask the question “how can I reuse weights in multiple places?”
– Recurrent (LSTM) layers reuse for multiple timesteps
– Convolutional layers reuse in multiple locations.
– Capsules reuse across orientation.
“ — Trask

1.4 Capsules

1.5 Unsupervised Learning

True intelligence will require independent learning strategies. Unsupervised learning is a paradigm for creating AI that learns without a particular task in mind: learning for the sake of learning. It captures some characteristics of the joint distribution of the observed random variables (learn the underlying structure).

Self-supervised learning is derived form unsupervised learning where the data provides the supervision.

1.5.1 Generative Adversarial Network

Simultaneously train two models: a generative model G that captures the data distribution, and a discriminative model D that estimates the probability that a sample came from the training data rather than G. The training procedure for G is to maximize the probability of D making a mistake.

$$\min_{\theta_g} \max_{\theta_d} [{\rm IE_{x\sim p_{data}(x)}} [log D_{\theta_d}(x)] + {\rm IE_{z\sim p_z(z)}} [log(1 - D_{\theta_d}(G_{\theta_g}(z)))]]$$

What I cannot create, I do not understand.“ — Richard Feynman

This framework corresponds to a minimax two-player game. Generative Adversarial Nets — Goodfellow et al.

GAN: Neural Networks Architecture Pioneered by Ian Goodfellow at University of Montreal (2014)

StyleGAN: A Style-Based Generator Architecture for Generative Adversarial Networks Paper | Code

Demo of BigGAN in an official Colaboratory notebook (backed by a GPU)

1.5.2 Variational Auto-Encoders (VAEs)

Variational Auto-Encoders (VAEs) are powerful models for learning low-dimensional representations.

Variational Autoencoders (VAEs): Powerful Generative Models.

I think transfer learning is the key to general intelligence. And I think the key to doing transfer learning will be the acquisition of conceptual knowledge that is abstracted away from perceptual details of where you learned it from.“ — Demis Hassabis

2. Autonomous Agents

Reinforcement learning (RL) studies how an agent can learn how to achieve goals in a complex, uncertain environment.

Artificial Intelligence 101: The First World-Class Overview of AI for the General Public | Autonomous Agents

An autonomous agent is any device that perceives its environment and takes actions that maximize its chance of success at some goal. At the bleeding edge of AI, autonomous agents can learn from experience, simulate worlds and orchestrate meta-solutions. Here’s an informal definition of the universal intelligence of agent `\pi`

$$\Upsilon(\pi) := \sum\limits_{\mu \in E} 2^{-K(\mu)} V^{\pi}_{\mu}$$

Intelligence measures an agent’s ability to achieve goals in a wide range of environments.“ — Shane Legg

2.1 Evolution Strategies

Evolution is a slow learning algorithm that with the sufficient amount of compute produces a human brain.“ — Wojciech Zaremba

Evolution and neural networks proved a potent combination in nature. Natural evolutionary strategy directly evolves the weights of a DNN and performs competitively with the best deep reinforcement learning algorithms. Neuroevolution enables capabilities that are typically unavailable to gradient-based approaches.

Exploiting Potential Energy to Locomote. Source: The Surprising Creativity of Digital Evolution: A Collection of Anecdotes from the Evolutionary Computation and Artificial Life Research Communities.

Flexible Muscle-Based Locomotion for Bipedal Creatures. Source: ACM Transactions on Graphics, Vol. 32, Nr. 6 (Proc. of SIGGRAPH Asia 2013).

Neural architecture search has advanced to the point where it can outperform human-designed models.

… evolution — whether biological or computational — is inherently creative, and should routinely be expected to surprise, delight, and even outwit us.“ — The Surprising Creativity of Digital Evolution, Lehman et al.

2.2 Deep Reinforcement Learning

In reinforcement learning, an agent interacts with an environment through a Markov decision process.

An Agent Interacts with an Environment

The goal in reinforcement learning is to train the agent to maximize the sum of future rewards, called the return.

Asynchronous Advantage Actor-Critic (A3C). Source: Petar Velickovic

Open-source RL

2.2.1 Model-Based RL

In Model-Based RL, the agent generates predictions about the next state and reward before choosing each action.

No superintelligent AI is going to bother with a task that is harder than hacking its reward function.“ — The Lebowski theorem

2.3 Self Play

Self-play mirrors similar insights from coevolution.

AlphaGo Zero showed that algorithms matter much more than big data and massive amounts of computation:

AlphaGo Zero : Algorithms matter much more than big data and massive amounts of computation

Self-Play is Automated Knowledge Creation.“ — Carlos E. Perez

Transfer learning is the key to go from self-play to the real world.

2.4 Multi-Agent Populations

2.5 Deep Meta-Learning

Learning to Learn. The goal of meta-learning is to train a model on a variety of learning tasks, such that it can solve new learning tasks using only a small number of training samples. A meta-learning algorithm takes in a distribution of tasks, where each task is a learning problem, and it produces a quick learner — a learner that can generalize from a small number of examples.

The notion of a neural “architecture” is going to disappear thanks to meta learning.“ — Andrew Trask

3. Environments

3.1 OpenAI Gym

3.2 Unity ML-Agents

3.3 DeepMind Control Suite

3.4 Brigham Young University | Holodeck

  • BYU Holodeck: A high-fidelity simulator for deep reinforcement learning Website | GitHub | Documentation — Brigham Young University

3.5 Facebook’s Horizon

  • Horizon: Facebook’s Open Source Applied Reinforcement Learning Platform Paper | GitHub | Blog — Jason Gauci, Edoardo Conti, Yitao Liang, Kittipat Virochsiri, Yuchen He, Zachary Kaden, Vivek Narayanan, Xiaohui Ye

3.6 PhysX

  • PhysX SDK, an Open-Source Physics Engine GitHub | Blog — NVIDIA

4. General Readings, Ressources and Tools

A Map of Ethical and Right-Based Approaches. Source: https://ai-hr.cyber.harvard.edu/primp-viz.html

(AI) will rank among our greatest technological achievements, and everyone deserves to play a role in shaping it.“ — Fei-Fei Li

ML paper writing pro-tip: you can download the raw source of any arxiv paper. Click on the “Other formats” link, then click “Download source”. This gets you a .tar.gz with all the .tex files, all the image files for the figures in their original resolution, etc.“ — Ian Goodfellow

Artificial Intelligence 101: The First World-Class Overview of AI for the General Public

Eventbrite - Montréal.AI Academy: AI 101


Group Reservation :

✉️ Email Us : info@montreal.ai
📞 Phone : +1.514.829.8269
🌐 Website : http://www.montreal.ai
📝 LinkedIn : https://www.linkedin.com/in/montrealai
🏛 Headquarters : 350, PRINCE-ARTHUR STREET W., SUITE #2105, MONTREAL [QC], CANADA, H2X 3R4 *Administrative Head Office

#AIFirst #MontrealAI #MontrealAIAcademy #MontrealArtificialIntelligence

Your browser is out-of-date!

Update your browser to view this website correctly. Update my browser now

×