Definitions: Phoneme – Distinct sounds in a language Grapheme – Distinct unit written in a language Lexicon – Dictionary that maps phoneme to graph(i.e. Arpabet) Architecture design choices of the above for ASR : 1. Target audio straight to Grapheme(without lexicon) – so basically extract features to Graphemes(words) – usedContinue Reading

CBOW  learns to predict the word by the context window(+nt -nt words) around it by taking the max probability of the word that fits this context(the most frequent word it learns).Therefore words with infrequent values will not work great with this technique.   Skip-Gram is inverted to CBOW, thus learns toContinue Reading

Start by building a very simple network: import numpy as np class NeuralNetwork: def __init__(self, x,y): self.input = x self.y = y self.Weights1 = np.random.randn(self.input.shape[1],5) self.Weights2 = np.random.randn(5,1) self.output = np.zeros(self.y.shape) def sigmoid_z(self,x): #create a sigmoid function z = 1/(1 + np.exp(-x)) return z def sigmoid_z_derivative(self,x): return self.sigmoid_z(x)*(1-self.sigmoid_z(x)) def forwardpropogation(self):Continue Reading

Tensorflow

After training we can optimize a frozen graph or even a dynamic graph by removes training-specific and debug-specific nodes, fusing common operations, and removes code that isn’t used/reached. Code Example from tensorflow.python.tools import optimize_for_inference_lib inputGraph = tf.GraphDef() #read in a frozen model with tf.gfile.Open(‘frozentensorflowModel.pb’, “rb”) as f: data2read = f.read() inputGraph.ParseFromString(data2read) outputGraph = optimize_for_inference_lib.optimize_for_inference(inputGraph, [“inputTensor”],        Continue Reading

Hidden states are the unknowns we try to detect or predict. The Hidden states have a relationship amongst themselves called the transition probabilities. Observations are the evidence variables that we have a priori. Observations and states have a relationship between them called the emission probabilities.Continue Reading