最近在看风格迁移的时候,看到了一个好玩的网站,叫Tensorfire
Tensorflow(https://github.com/tensorflow/tensorflow)
TensorFire是一个浏览器内部的神经网络库,它利用深度学习技术,在WebGL的加速下于浏览器中运行神经网络
TensorFire is a framework for running neural networks in the browser, accelerated by WebGL.
Applications powered by TensorFire can utilize deep learning in almost any modern web browser with no setup or installation.
TensorFire models run up to 100x faster than previous in-browser neural network libraries, at speeds comparable to highly-optimized native CPU code.
TensorFire has two parts: a low-level language based on GLSL for easily writing massively parallel WebGL shaders(着色器) that operate on 4D tensors, and a high-level library for importing models trained with Keras or TensorFlow.
It works on any GPU, whether or not it supports CUDA. That means on computers with AMD graphics, like the latest 2016 Retina Macbook Pro。
With TensorFire, you can build applications which leverage the power of deep learning without forcing people to install native apps, without having to pay for expensive compute farms, nor waiting for a server to respond. Rather than bringing the data to the model, you can deliver your model straight to your users, respecting their right to privacy(尊重他们的隐私权).
We’ve prototyped(原型、样板) some demos, but it barely scratches the surface of what’s possible. **TensorFire can
The low-level API can also be used to do arbitrary(任意的) parallel general purpose computation. We’ve used it to multiply matrices, solve linear systems of equations, and to compute PageRank, simulate cellular automata, transform and filter images, and more.
Modern desktops, laptops, and phones contain powerful GPUs optimized for highly-parallel computation.
By transforming neural network weights into WebGL textures and implementing common layers as fragment shaders, we can use the graphics capabilities of browsers designed for 3D games to speed up the execution(完成) of neural networks.
支持低精度的量化传感器,追求了faster
Unlike other WebGL compute frameworks, we support low-precision quantized tensors(website). This allows us to support browsers that don’t fully support the OES_texture_float extension, and run even faster with even smaller models.
Sign up for updates! We’re still frantically mashing on our keyboards to document our APIs, and you’ll be the first to hear once that’s ready. Like all good things, it’ll be open source, and we’ll be depending on people like you to make cool stuff.
We’re going to be launching more demos over the next couple weeks, and if there’s a bunch of people on that list we’ll feel really bad about disappointing all of you who have read this public promise to do so.
We’re a group of recent MIT graduates who all think this whole “deep learning” thing is pretty neat.
Kevin Kwok and Guillermo Webster have previously built things combining Javascript and Computer Vision, like Project Naptha— a browser extension that lets you seamlessly highlight, copy/paste, and translate text from within pictures, and Tesseract.js: a fully in-browser OCR library. Anish Athalye and Logan Engstrom have respectively built the first TensorFlow implementations of Gatys’ neural artistic style, and Johnson’s fast style transfer algorithms.
Emboldened by the AI Grant, we’ve spent some time putting together this framework, and we’ve had lots of fun building stuff on top of it. Soon it will be your turn!
Contact us at team@tenso.rs.
WebGL
GLSL
SqueezeNet
OCR