Sign In Start Free Trial
Account

Add to playlist

Create a Playlist

Modal Close icon
You need to login to use this feature.
  • Book Overview & Buying Neural Networks with Keras Cookbook
  • Table Of Contents Toc
  • Feedback & Rating feedback
Neural Networks with Keras Cookbook

Neural Networks with Keras Cookbook

By : V Kishore Ayyadevara
3.3 (8)
close
close
Neural Networks with Keras Cookbook

Neural Networks with Keras Cookbook

3.3 (8)
By: V Kishore Ayyadevara

Overview of this book

This book will take you from the basics of neural networks to advanced implementations of architectures using a recipe-based approach. We will learn about how neural networks work and the impact of various hyper parameters on a network's accuracy along with leveraging neural networks for structured and unstructured data. Later, we will learn how to classify and detect objects in images. We will also learn to use transfer learning for multiple applications, including a self-driving car using Convolutional Neural Networks. We will generate images while leveraging GANs and also by performing image encoding. Additionally, we will perform text analysis using word vector based techniques. Later, we will use Recurrent Neural Networks and LSTM to implement chatbot and Machine Translation systems. Finally, you will learn about transcribing images, audio, and generating captions and also use Deep Q-learning to build an agent that plays Space Invaders game. By the end of this book, you will have developed the skills to choose and customize multiple neural network architectures for various deep learning problems you might encounter.
Table of Contents (18 chapters)
close
close

Building word vectors using fastText

fastText is a library created by the Facebook Research Team for the efficient learning of word representations and sentence classification.

fastText differs from word2vec in the sense that word2vec treats every single word as the smallest unit whose vector representation is to be found, but fastText assumes a word to be formed by a n-grams of character; for example, sunny is composed of [sun, sunn, sunny],[sunny, unny, nny], and so on, where we see a subset of the original word of size n, where n could range from 1 to the length of the original word.

Another reason for the use of fastText would be that the words do not meet the minimum frequency cut-off in the skip-gram or CBOW models. For example, the word appended would not be very different than append. However, if append occurs frequently, and in the new sentence we have the word appended...

Unlock full access

Continue reading for free

A Packt free trial gives you instant online access to our library of over 7000 practical eBooks and videos, constantly updated with the latest in tech

Create a Note

Modal Close icon
You need to login to use this feature.
notes
bookmark search playlist download font-size

Change the font size

margin-width

Change margin width

day-mode

Change background colour

Close icon Search
Country selected

Close icon Your notes and bookmarks

Delete Bookmark

Modal Close icon
Are you sure you want to delete it?
Cancel
Yes, Delete

Delete Note

Modal Close icon
Are you sure you want to delete it?
Cancel
Yes, Delete

Edit Note

Modal Close icon
Write a note (max 255 characters)
Cancel
Update Note

Confirmation

Modal Close icon
claim successful

Buy this book with your credits?

Modal Close icon
Are you sure you want to buy this book with one of your credits?
Close
YES, BUY