Automated front-end using deep learning

Overview

Creating intuitive and engaging experiences for users is a critical goal for companies of all sizes, and it’s a process driven by quick cycles of prototyping, designing, and user testing. Large corporations like Facebook have the bandwidth to dedicate entire teams to the design process, which can take several weeks and involve multiple stakeholders; small businesses don’t have these resources, and their user interfaces may suffer as a result.

A typical design workflow might look like the following:

  • Product managers perform user research to generate a list of specifications
  • Designers take those requirements and explore low-fidelity prototypes, eventually creating high-fidelity mockups
  • Engineers implement those designs into code and finally ship the product to users

The problem I was solving falls under a broader umbrella of tasks known as program synthesis, the automated generation of working source code. Though much of program synthesis deals with code generated from natural language specifications or execution traces, in my case I could leverage the fact that I had a source image (the hand-drawn wireframe) to start with.

There’s a well-studied domain in machine learning called image captioning that seeks to learn models that tie together images and text, specifically to generate descriptions of the contents of a source image.

Project Link

Using an image captioning model architecture

Now that I had my data ready, I could finally feed it into the model!

I leveraged a model architecture used in image captioning that consists of three major parts:

  • A computer vision model that uses a Convolutional Neural Network (CNN) to extract image features from the source images.
  • A language model consisting of a Gated Recurrent Unit (GRU) that encodes sequences of source code tokens.
  • A decoder model (also a GRU), which takes in the output from the previous two steps as its input, and predicts the next token in the sequence.

Dependencies

  • Python
  • Keras
  • tensorflow
  • nltk
  • opencv
  • numpy
  • matplotlib
  • Pillow
  • scipy

 

How to run

Converting an example drawn image into HTML code, using pretrained weights:

python convert_single_image.py –png_path ../examples/drawn_example1.png \ –output_folder ./generated_html \ –model_json_file ../bin/model_json.json \ –model_weights_file ../bin/weights.h5

 

 

 

Free

xpertup

Let's Expert Up