Fast Neural Style Transfer by PyTorch (Mac OS)

2021-Jan-31: The git repo has been upgraded from PyTorch 0.3.0 to PyTorch 1.7.0 with Python 3.8.3.

C. Cui's Blog

2021-Jan-31: The git repo has been upgraded from PyTorch 0.3.0 to PyTorch 1.7.0.

Continue my last post Image Style Transfer Using ConvNets by TensorFlow (Windows), this article will introduce the Fast Neural Style Transfer by PyTorch on MacOS.

The original program is written in Python, and uses [PyTorch], [SciPy]. A GPU is not necessary but can provide a significant speedup especially for training a new model. Regular sized images can be styled on a laptop or desktop using saved models.

More details about the algorithm could be found in the following papers:

  1. Perceptual Losses for Real-Time Style Transfer and Super-Resolution ;
  2. Instance Normalization: The Missing Ingredient for Fast Stylization.

If you could not download the papers, here are the Papers.

You can find all the source code and images at my GitHub: fast_neural_style .

View original post 302 more words

Increasing Transparency into What It Takes to Achieve Performance Gains of Machine Learning Algorithms

The computations required for Deep Learning research have been doubling every few months, resulting in an estimated 300,000x increase from 2012 to 2018. AI could account for as much as one-tenth of the world’s electricity use by 2025 according to this article [1].

Continue reading “Increasing Transparency into What It Takes to Achieve Performance Gains of Machine Learning Algorithms”

Just Got a Reviewer Certificate from Data Mining and Knowledge Discovery (WIREs)

Thanks to the Editors and Board of WIREs for supporting me. As an independent reviewer, I will be fair to everyone and never give in to the “scientific mafia” and “citation cartels”.

Data Mining and Knowledge Discovery (WIREs) (Impact Factor: 2.541)


A Taste of TensorFlow on My Android Phone

If you like Google’s open-source machine learning framework, TensorFlow, do not miss this “TensorFlow For Poets“.  I went through the tutorial this afternoon and found it is super Awesome. See the photos below, I first tested it on the coffee mug from my Intern company, Aurecon Group. I used the virtual device, Nexus 5X, from Android Studio 3.0.1 on MacBook Air 11′  (Do not do this unless you have enough SSD 😛 ).


This slideshow requires JavaScript.

Then, I successfully installed the compiled app (TF_Classify) on my XIAO MI – 4C (MIUI 9.0 – Android 7.0) and tested it on my coffee mug at home.
You can download and install it on your own Android devices from the following link:

Continue reading “A Taste of TensorFlow on My Android Phone”

Starting My First Intern at Melbourne Australia Tomorrow

Dear All,

I am using the song above to thank you all for your help and support in the past. You know that I have spent the last three years (2014-2017) in pursuing my Ph.D. degree in Computer Science and got a plan to be graduated in 2018.

Continue reading “Starting My First Intern at Melbourne Australia Tomorrow”

Building ConvNets on MNIST dataset by TensorFlow with the new WIN10 GPU Monitor

A few days ago, I updated my  Windows 10 to version 1709 and found out that Microsoft added the GPU monitor in the Task Manager which I thought is awesome for ML developers and researchers.

Here is a screen capture of the official MNIST codes running Tensorflow-GPU on my Desktop.  It is clear to see that the GTX 960 uses about 3.5GB memory out of 4.0GB to train the ConvNets, which is much faster than the CPU computing.


You can find more models from the TensorFlow Models. This repository contains a number of different models implemented in TensorFlow.


What do I think about PyTorch and TensorFlow?

As we all know, the TensorFlow is very powerful and mature deep learning library with strong visualization capabilities and several options to use for high-level model development. PyTorch is still young framework which is getting momentum fast.

I strongly suggest CS and IT researchers/engineers learn both of them.

Tensorflow will be a good option if you are developing models for production or on mobile platforms, maybe in the future for large-scale distributed model training. Because it has good community support and comprehensive documentation, it is easier to find answers and get helps online.

Well, PyTorch is a good fit if you are doing research or your production are not very demanding.

Personly, I think Pytorch has better development and debugging experience.

Continue reading “What do I think about PyTorch and TensorFlow?”

TensorFlow Neural Network Playground in Matlab

The amazing website  can help you open a Neural Network on your Web Browser. The GUI is mind blowing, and you could download all the codes to study or to build your own project.

Now,   The Good News!  Amro and Ray Phan have created the MATLAB version of the NN playground, it looks just like the GUI of the Tensorflow version. However, it is not tensorflow-based, it is built on the Neural Networks Toolbox of Matlab (>R2009b). The authors said they are inspired by the [TensorFlow Neural Networks Playground] interface readily available online, so they created a MATLAB implementation of the same Neural Network interface for using Artificial Neural Networks for regression and classification of highly nonlinear data. Continue reading “TensorFlow Neural Network Playground in Matlab”

Python For Data Science (Cheat Sheets Collections)

PDF Download: Python For Data Science

Here is the list:

  1. Numpy
  2. Matplotlib
  3. Scikit-Learn
  4. SciPy-Linear Algebra
  5. Pandas

When you doing the programming this is very helpful, cheers 🙂

Python Cheat Sheets Conbination_Page_1

Continue reading “Python For Data Science (Cheat Sheets Collections)”