Wykay / alpha_models

[ALPHA] Models for Runway

Home Page:https://runwayml.com/docs/install/

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Runway Models for the alpha version

This repository contains the current public models for Runway. This models are the ones you can use from the app or as a standalone application.

Structure

models.json is the file that Runway requires when opened. It has all the necessary information to create and install models when running the Runway app.

Each model lives in its own folder and has its own configurations. See below for more detailed information.

Status

Implemented

In Progress

How models work in Runway

Docker Containers

Each model is a docker container thats creates a public HTTP socket server. Containers must public in the Docker Hub. More on this soon.

Server

Models running inside Docker can work in any language or framework. They just need to be accessible through a public url. A server with socket connection must be created. The file server_template.py is a blueprint of the minimum requirements for a Runway server.

models.json

The file model.js is a collection of meta information about docker container. This are the final models Runway queries and installs. A valid model must have the following format in models.json

{
    containerId: String,
    tag: String,
    name: String,
    specs: {
      cpu: Boolean,
      gpu: Boolean,
    },
    port: String,
    query: {
      http: {
        route: String,
        method: String,
        type: String,
      },
      socket: {
        namespace: String,
        emit: String,
        event: String
      },
      format: {
        data: String
      }
    },
    size: [String, String],
    options: {
      image: String,
      Tty: Boolean,
      Detach: Boolean,
      ExposedPorts: { String: {} },
      HostConfig: {
        AutoRemove: true,
        PortBindings: {
          String: [{
            HostIP: String,
            HostPort: String
          }],
        },
      }
    },
    needsTraining: Boolean,
    description: {
      first: ,
      second: String,
      third: String,
      authors: String,
      image: String,
      organization: String,
      framework: String,
      github: String,
      paper: String,
      licence: String,
    },
    inputs: [String, String]
  }

For instance, this is the current implementation of im2txt:

{
    containerId: '84d72a2f82f6',
    tag: 'cvalenzuelab/im2text:latest',
    name: 'im2txt',
    specs: {
      cpu: true,
      gpu: true,
    },
    port: '33100',
    query: {
      http: {
        route: 'query_once',
        method: 'post',
        type: 'JSON',
      },
      socket: {
        namespace: 'query',
        emit: 'update_request',
        event: 'update_response'
      },
      format: {
        data: 'data:image/jpg;base64,'
      }
    },
    size: ['2.19', 'GB'],
    options: {
      image: '84d72a2f82f6',
      Tty: false,
      Detach: true,
      ExposedPorts: { '33000/tcp': {} },
      HostConfig: {
        AutoRemove: true,
        PortBindings: {
          '33000/tcp': [{
            HostIP: '0.0.0.0',
            HostPort: '33100'
          }],
        },
      }
    },
    needsTraining: false,
    description: {
      first: 'The im2txt model, also called the Show and Tell model, is a deep neural network that learns how to describe the content of images',
      second: 'More technically, it is an example of an encoder-decoder neural network. It works by first "encoding" an image into a fixed-length vector representation, and then "decoding" the representation into a natural language description.',
      third: 'The image encoder is a deep convolutional neural network. This type of network is widely used for image tasks and is currently state-of-the-art for object recognition and detection. Our particular choice of network is the Inception v3 image recognition model pretrained on the ILSVRC-2012-CLS image classification dataset. The decoder is a long short-term memory (LSTM) network. This type of network is commonly used for sequence modeling tasks such as language modeling and machine translation. In the Show and Tell model, the LSTM network is trained as a language model conditioned on the image encoding.',
      authors: 'Oriol Vinyals, Alexander Toshev, Samy Bengio, Dumitru Erhan.',
      image: 'https://raw.githubusercontent.com/cvalenzuela/runway_models/master/im2txt/imgs/demo.png?token=AKHU_R2F7AQImVwfzZ1RqeRfVKYWJYwhks5azcGywA%3D%3D',
      organization: 'Google',
      framework: 'Tensorflow',
      github: 'https://github.com/tensorflow/models/tree/master/research/im2txt',
      paper: 'https://arxiv.org/abs/1609.06647',
      licence: 'Apache License 2.0',
    },
    inputs: ['Image', 'Video']
  },

Deploying as stand-alone apps

More on this soon.

Contributing

SEE CONTRIBUTING.md

About

[ALPHA] Models for Runway

https://runwayml.com/docs/install/

License:GNU General Public License v3.0


Languages

Language:PureBasic 72.6%Language:Python 26.8%Language:Dockerfile 0.3%Language:Shell 0.2%Language:JavaScript 0.1%Language:HTML 0.0%Language:CMake 0.0%