Deep Neural Network Based Speech Enhancement - ViVoLab

1y ago
35 Views
2 Downloads
1.95 MB
25 Pages
Last View : 2d ago
Last Download : 3m ago
Upload by : Milena Petrie
Transcription

5/26/2014Speech Enhancement Based onDeep Neural NetworksChin-Hui LeeSchool of ECE, Georgia Techchl@ece.gatech.eduJoint work with Yong Xu and Jun Du at USTC1Outline and Talk Agenda In Signal Processing Letter, Jan. 2014 Speech enhancement Background Conventional speech enhancement methods Speech enhancement based on deep neural network s SE-DNN: background DNN baseline and enhancement Noise-universal SE-DNNZaragoza, 27/05/1421

5/26/2014Outline Speech enhancement Background Conventional speech enhancement methods Speech enhancement based on deep neural network s SE-DNN: background DNN baseline and enhancement Noise-universal SE-DNNZaragoza, 27/05/143Speech Enhancement Speech enhancement aims at improving the intelligibility and/oroverall perceptual quality of degraded speech signals using audiosignal processing techniques One of the most addressed classical SP problems in recent yearsEnhancingNoisy speech,Exhibition, SNR 5dBClean speech, 8kHzZaragoza, 27/05/1442

5/26/2014Speech Enhancement ApplicationsMobile phone/communicationHearing aidsSecurity monitoring/intelligenceRobust speech/speaker/languagerecognition, etc.Zaragoza, 27/05/145Noise in Speech Enhancement1. Additive noise:๐‘ฆ ๐‘ก ๐‘ฅ ๐‘ก ๐‘› ๐‘กSTFT๐‘Œ ๐‘›, ๐‘‘ ๐‘‹ ๐‘›, ๐‘‘ ๐‘(๐‘›, ๐‘‘)Focused2. Convolutional noise:๐‘ฆ ๐‘ก ๐‘ฅ ๐‘ก โ„Ž ๐‘ก3. Mixed noise:๐‘ฆ ๐‘ก ๐‘ฅ ๐‘ก โ„Ž ๐‘ก ๐‘› ๐‘ก๐‘ฆ ๐‘ก [๐‘ฅ ๐‘ก ๐‘ฃ ๐‘ก ] โ„Ž ๐‘ก ๐‘› ๐‘กZaragoza, 27/05/1463

5/26/2014Outline Speech enhancement Background Conventional speech enhancement methods Speech enhancement based on deep neural network s SE-DNN: background DNN baseline and enhancement Noise-universal SE-DNNZaragoza, 27/05/147Conventional Speech Enhancement Classified by the number of microphones1. Single channel speech enhancement methods Time and frequency information2.FocusedMicrophone based speech enhancement methods Time and frequency information Spatial information Microphone arraysZaragoza, 27/05/1484

5/26/2014Conventional Single Channel SE1. Spectrum Subtraction, SS2. Wiener Filtering3. Minimum Mean Square Error Short-Time SpectralAmplitude, MMSE-STSA4. MMSE Log Spectral Amplitude, MMSE-LSA5. Optimally Modified LSA, OM-LSA6. Zaragoza, 27/05/149Conventional Single Channel SEy (t )xห†(t )STFTISTFTY (l , k ) AmplitudeSquaredG (l , k )Y (l , k )2GainFunctionCalculationNoiseEstimation ห†d (l , k )1. STFT on the noisy signal y, get the time-frequency signal ๐‘Œ2. estimate the variance of noise ๐œ†๐‘‘3. estimate all of the parameters (prior SNR ๐›พ, posterior SNR ๐œ‰ and thespeech presence probability, etc.) needed by the gain function4. calculate the gain function G5. multiply ๐‘Œ with G, then ISTFT to obtain the enhanced signal (usingthe phase of noisy speech)Zaragoza, 27/05/14105

5/26/2014Conventional Single Channel SE: Issues1. Musical noise:Enhanced by SSNoisy speech, exhibitionnoise, SNR 5dBZaragoza, 27/05/1411Conventional Single Channel SE: Issues2. Difficult to deal with the highly non-stationary noise:Enhanced by OM-LSANoisy, Machine Gun,SNR -5dBZaragoza, 27/05/14126

5/26/2014Conventional Single Channel SE: Issues3. Difficult to deal with the low SNR cases:Enhanced by OM-LSANoisy, AGWN, SNR -5dBEasily distorting thespeech components andhas much residual noiseZaragoza, 27/05/1413Conventional Single Channel SE: Issues4. Introducing some non-linear distortion which isfatal for the back-end recognition, coding, etc.5. Learning from human listening experience:many years of exposure to clean speech and noiseZaragoza, 27/05/14147

5/26/2014Conventional Single Channel SE: Issues Analysis of these disadvantagesSTFT๐‘ฆ ๐‘ก ๐‘ฅ ๐‘ก ๐‘› ๐‘ก๐‘Œ ๐‘›, ๐‘‘ ๐‘‹ ๐‘›, ๐‘‘ ๐‘(๐‘›, ๐‘‘)Gaussian assumptionsUn-correlated assumptionsE{ Y (n, d) } E{ X (n, d) } E{ N (n, d) }222 (n, d) x (n, d) d (n, d)Binary model assumptions:H 0 (n, d) : Y (n, d) N (n, d)H1 (n, d) : Y (n, d) X (n, d) N (n, d)With these inaccurate assumptions, it is hard for conventionalmethods to deliver a satisfactory performance!Zaragoza, 27/05/1415Outline Speech enhancement Background Conventional speech enhancement methods Speech enhancement based on deep neural networks 2.1 Background DNN baseline and enhancement Noise-universal SE-DNNZaragoza, 27/05/14168

5/26/2014DNN-Based Speech Enhancement The signal model of the additive noise:๐‘ฆ ๐‘ก ๐‘ฅ ๐‘ก ๐‘› ๐‘กSTFT๐‘Œ ๐‘›, ๐‘‘ ๐‘‹ ๐‘›, ๐‘‘ ๐‘(๐‘›, ๐‘‘) Many enhancement methods are derived from this signal model, however,most of them assume that ๐‘‹ n, ๐‘‘ is described by a Gaussian mixture model(GMM) and ๐‘(n, d) is a single Gaussian. The relationship between thespeech and noise is complicated in some non-linear fashion. DNN assumes a nonlinear mapping function F:X ๐น(๐‘Œ) Construct the stereo data based on the additive noise model No special assumptions were made in the DNN based SE methodZaragoza, 27/05/1417Deep Neural Network: Overview1. Hinton proposed the unsupervised Restricted BoltzmannMachine (RBM) based pre-training in 20062. In 2012, MSR, Google and IBM got a great success in largevocabulary continuous speech recognition using DNNs3. Later, DNNs were adopted in many speech-related tasksInputlayerHidden layerOutputlayerMathematicalapproximationZaragoza, 27/05/14189

5/26/2014DNN Based SE: Related Work4. In 2013, Xugang Luproposed deep denoising auto-encoderbased speechenhancementX.-G. Lu, Y. Tsao, S. Matsuda and C. Hori, โ€œSpeech enhancement based on deepdenoising Auto-Encoder,โ€ Proc. Interspeech, pp. 436-440, 2013.23DNN Based SE: Related Work6. In 2013, Deliang Wang proposed using DNN to classifythe time-Frequency bins into 1/0 units (ideal binary mask)IBM-DNN enhancedNoisyspeechCochlearFilteringT-F unitslevel featureextractionDNN-based ed IBMEnhancedspeechY. X. Wang and D. L. Wang, โ€œTowards scaling up classi๏ฌcation based speech separation,โ€ IEEETrans. on Audio, Speech and Language Processing, Vol. 21, No. 7, pp. 1381-1390, 2013.10

5/26/2014DNN Based SE: Issues Advantages of SE-DNN1. The complicated relationship between noisy andclean speech could be automatically learnt2. The deep architecture could well fit the non-linearrelationship for regression function approximation3. The highly non-stationary noise could be wellsuppressed in the off-line learning framework4. Nearly no Gaussian or independent assumptions5. Nearly no empirical thresholds to avoid the non-lineardistortion in SS-based speech enhancementZaragoza, 27/05/1426DNN Based SE: Issues Difficulties in using SE-DNN1. Which domain is suitable for DNN-based mapping?2. The generalization capacity to unknown environments,especially for unseen noise types?3. How to perform noise adaptation? โ€“ robustness issueZaragoza, 27/05/142711

5/26/2014Outline Speech enhancement Background Conventional speech enhancement methods Speech enhancement based on deep neural networks SE-DNN: background DNN baseline and enhancement Noise-universal SE-DNNZaragoza, 27/05/1428System OverviewTraining gEnhancement ฬ‚ lWaveformReconstructionXฬ‚ t Y f1. Feature extraction: log-power spectra2. Waveform reconstruction: overlap-add algorithm3. DNN Training: RBM pre-training back-propagation fine-tuningZaragoza, 27/05/142912

5/26/2014DNN Training(Output with a single frame of clean speech features)2. Fine-tuning ๐‘พ๐Ÿ’1.Pre-training h3๐‘พ๐Ÿ‘ ๐œบ ๐Ÿ‘h2 h1 ๐‘พ๐Ÿ ๐œบ ๐Ÿ๐‘พ๐Ÿ ๐œบ ๐Ÿ (Input with multiple frames of noisy speech features)1.MMSE-based object function: ๐ธ 1๐‘๐‘๐‘› 1๐ท๐‘‘๐‘‘ 1(๐‘‹๐‘›๐‘พ, ๐’ƒ ๐‘‹๐‘›๐‘‘ )2 ๐œ† ๐‘พ2 302Experimental Setup1. Clean speech set: TIMIT corpus, 8kHz2. Noise set: Additive Gaussian White Noise (AGWN), Babble,Restaurant, Street3. Signal to Noise ratios: Clean, 20dB, 15dB, 10dB, 5dB, 0dB, -5dB4. Construct 100 hours multi-condition training data5. Test set: 200 randomly selected utterances from TIMIT test set,and two unseen noise types: Car and Exhibition6. Three objective quality measures: segmental SNR (SegSNR indB), log-spectral distortion (LSD in dB), perceptual evaluation ofspeech quality (PESQ)7. Standard DNN configurations: 11 frames expansion, 3 hiddenlayers and 2048 hidden units for each8. Competing methods: improved version of the optimallymodi๏ฌed log-spectral amplitude (OM-LSA), denoted as logMMSE (L-MMSE)Zaragoza, 27/05/143113

5/26/2014Baseline Experimental Results: I1. Average LSD using input with different acoustic context on the testset at different SNRs across four noise types: A good choice: 11 framesZaragoza, 27/05/1432Baseline Experimental Results: II2. Average SegSNRs using different training set size on the test set atdifferent SNRs across four noise types: still improving with 100 hoursZaragoza, 27/05/143314

5/26/2014Baseline Experimental Results: III3. Average PESQs among methods on the test set at different SNRswith four noise types. The subscript of ๐ท๐‘๐‘๐‘™ represents ๐‘™ hidden -MMSE SNN**Shallow Neural Network (SNN) has the same computation complexity with ๐ท๐‘๐‘3 Deep structure can get better performance compared with SNN. ๐ท๐‘๐‘3 outperforms the L-MMSE method, especially at low SNRs.Zaragoza, 27/05/1434Baseline Experimental Results: IV4. PESQs among Noisy, L-MMSE, SNN and ๐ท๐‘๐‘3 at different SNRs inmismatch environments under Car (A) and Exhibition (B) 303.313.013.032.692.712.332.351.931.961.542.832.47 SE-DNN has a generalization capacity to unseen noise types. It canbe further strengthened by adding more noise types in trainingZaragoza, 27/05/143515

5/26/2014Baseline Experimental Results VDNN enhancedPESQ 3.39L-MMSE enhancedPESQ 2.64DNN can dealwith nonstationarynoise givensome noisecharacteristicsto learn.Clean, PESQ 4.5Noisy, street, SNR 10dB,PESQ 2.2Zaragoza, 27/05/1436More demos could be found at: http://home.ustc.edu.cn/ xuyong62/demo/SE DNN.htmlOver-smoothing with SE-DNN (1/2)1. The global variances of the training set were shown. ๐บ๐‘‰๐‘Ÿ๐‘’๐‘“ (๐‘‘) and ๐บ๐‘‰๐‘’๐‘ ๐‘ก (๐‘‘)represented the d-th dimension of the global variance of the reference featuresand the estimation features, respectively. And the corresponding dimension37independent variances were denoted as ๐บ๐‘‰๐‘Ÿ๐‘’๐‘“ and ๐บ๐‘‰๐‘’๐‘ ๐‘ก16

5/26/2014Over-smoothing with SE-DNN (2/2)DNN enhancedCleanAGWN, SNR 0dB2. The formant peaks were suppressed, especially in the highfrequency band which leads to muf๏ฌ‚ed speechZaragoza, 27/05/1438Methods to Address Over-smoothing The definition of global variance equalization factors:๐›ฝ ๏ฟฝ๏ฟฝ๐‘’๐‘“ (๐‘‘)๐บ๐‘‰๐‘’๐‘ ๐‘ก (๐‘‘)๐›ผ(๐‘‘) 1๐›ผ ๐ท๐ท๐›ผ(๐‘‘)๐‘‘ 1 Proposed method 1: post-processing๐‘‹ โ€ฒโ€ฒ ๐‘‘ ๐‘‹ ๐‘‘ ๐œ‚ ๐‘ฃ(๐‘‘) ๐‘š(๐‘‘)where ๐‘š(๐‘‘) and ๐‘ฃ(๐‘‘) are the d-th component of the mean and varianceof input noisy features, respectively. And ฮท could be ๐›ฝ , ๐›ผ(๐‘‘) or ๐›ผ Proposed method 2: post-training๐ธ 1๐‘๐‘๐ท(๐‘‹๐‘›๐‘‘ ๐‘พ, ๐’ƒ ฮท ๐‘‹๐‘›๐‘‘ )2 ๐œ† ๐‘พ22๐‘› 1 ๐‘‘ 1Zaragoza, 27/05/143917

5/26/2014GV Experimental Results (1/2)PESQ results of the L-MMSE method and DNN baseline, comparedwith different post-processing and post-training schemes using ๐›ฝ,๐›ผ(๐‘‘) and ๐›ผ on the test set at different SNRs across four noise 4๐œถ3.723.493.202.862.472.022.961. ๐›ผ is better than ๐›ฝ, and ๐›ผ(๐‘‘) is the worst, indicating that thedegree of over-smoothing on different dimensions was similar2. Equalization operations were much more beneficial for high SNRs3. Post-training was a little better than Post-processing40GV Experimental Results (2/2)PESQ results in unseen environments under Car and Exhibition noises,labeled as case A and B, respectively. The DNN baseline wascompared with the L-MMSE method and the proposed two globalvariance equalization approaches using the factor ๐›ผ 81. GV equalization is slightly more effective for unseen noise types2. Post-training was a little better than Post-processingZaragoza, 27/05/144118

5/26/2014Summary I: DNN-SE Properties1. SE-DNN achieves better performance than traditional singlechannel speech enhancement methods (e.g., OM-LSA),especially for low SNRs and non-stationary noise.2. A large training set is crucial to learn the rich structure of DNN3. Using more acoustic context information improves performanceand makes the enhanced speech less discontinuous4. Multi-condition training can deal with speech enhancement ofnew speakers, unseen noise types, various SNR levels underdifferent conditions, and even cross-language generalization.5. The over-smoothing problem in SE-DNN could be alleviatedusing two global variance equalization methods, and theequalization factor tends to be independent with the dimension6. The global variance equalization was much more helpful forunseen noise typesZaragoza, 27/05/1442Outline Speech enhancement task Backgrounds Conventional speech enhancement methods Speech enhancement based on deep neural networks SE-DNN: background DNN baseline and enhancement Noise-universal SE-DNNZaragoza, 27/05/144319

5/26/2014Noise Universal SE-DNNTraining N DropoutTrainingNoiseEstimationEnhancement PostprocessingWaveformXฬ‚l ReconstructionXฬ‚ t Y f*The global variance equalization was adopted in the post-processing.Zaragoza, 27/05/1444Noise Universal SE-DNN1. DNN to learn the characteristics of many noise types ionใ€animalใ€natureใ€human, etc.alarmZaragoza, 27/05/14cryG. Hu, 100 non-speech environmental sounds, pus.html. 4520

5/26/2014Noise Universal SE-DNN2. Noise aware training Using the average feature of the first T frames of thecurrent utterance to help DNN to learn a โ€œnoise codeโ€3. Dropout learning Randomly disable some units of the input layer and thehidden layers to improve generalization capacity (0.1 forthe input layer and 0.2 for the hidden layers) Regulation technology and avoid over-fittingZaragoza, 27/05/1446Experimental Setup1.2.3.4.5.Clean speech training set: TIMIT corpus, 8kHzNoise training set: 104 noise typesSignal to Noise ratios: Clean, 20dB, 15dB, 10dB, 5dB, 0dB, -5dBConstruct 100/625 hours multi-condition training dataTest set: 200 randomly selected utterances from the TIMIT testset corrupted by the noises from the NOISEX-92 database6. Three objective quality measures: segmental SNR (SegSNR indB), log-spectral distortion (LSD in dB), perceptual evaluation ofspeech quality (PESQ)7. Standard DNN configurations: 11 frames context expansion, 3hidden layers and 2048 hidden units for each hidden layer8. Competing state-of-the-art methods: improved version of theoptimally modi๏ฌed log-spectral amplitude (OM-LSA), denotedas log-MMSE (L-MMSE)Zaragoza, 27/05/144721

5/26/2014Enhanced Experimental Results: I LSD comparison between models trained with four noise types(4NT) and 104 noise types (104NT) on the test set at differentSNRs of three unseen noise environments :Abundance of noise types is important to predict unseen noise typesZaragoza, 27/05/1448Enhanced Experimental Results: II Spectrograms of an utterance tested with Exhibition noise at SNR 5dB. (a) noisy(PESQ 1.42), (b) LogMMSE (PESQ 1.83), (c) DNN baseline (PESQ 1.87), (d)improved by dropout (PESQ 2.06), (e) improved by GV equalization (PESQ 2.00),(f) improved by dropout and GV (PESQ 2.13), (g) jointly improved by dropout,NAT and GV equalization (PESQ 2.25), and the clean (PESQ 4.5):1.2.SE-DNN can suppress the highly non-stationary noise and get less residual noiseDropout and NAT can reduce noise while GV equalization can brighten speechZaragoza, 27/05/144922

5/26/2014Enhanced Experimental Results: III Spectrograms of an utterance with machine gun noise at SNR -5dB:with 104-noise DNN enhanced (upper left, PESQ 2.78), Log-MMSEenhanced (lower left, PESQ 1.86), 4-noise DNN enhanced (upperright, PESQ 2.14), and noisy speech (lower right, PESQ 1.85):104NT-DNNenhanced PESQ 2.784NT-DNN enhancedPESQ 2.14Log-MMSE enhancedPESQ 1.86noisy๏ผŒmachine gun๏ผŒSNR -5dB PESQ 1.85Even the 4NT-DNNis much better thanLogMMSE, SE-DNNcan suppress highlynon-stationary noiseZaragoza, 27/05/1450Enhanced Experimental Results: IV Average PESQ among LogMMSE, DNN baseline with 100 hoursdata, improved DNN with 100 hours data and improved DNN with625 hours data on the test set at different SNRs across the whole15 unseen noise types in the NOISEX-92 database:1. A good generalization capacity to unseen noise can be obtained.2. SE-DNN outperformed the Log-MMSE, especially at low SNRsZaragoza, 27/05/145123

5/26/2014Enhanced Experimental Results: V Spectrograms of a noisy utterance extracted from the movie ForrestGump with: improved DNN (upper left), Log-MMSE (upper right),and noisy speech (bottom left): with real-world noise never seenUniversal SE-DNN enhancedLog-MMSE enhanced NoisyGood generalization capacityto real-world noisy speechCould be further improved byadding more varieties of cleandata into the training setZaragoza, 27/05/1452Summary II: Noise-Universal DNN1. Noise aware training (NAT) and dropout learning could suppressmore residual noise2. GV equalization could highlight the speech spectrum to get abetter hearing perception3. The generalization capacity to unseen noise types could bestrengthened by adopting more noise types in the training set4. Noise-universal SE-DNN was also effective in dealing with noisyspeech recorded in the real world5. The generalization capacity could be further improved byadding clean speech data (encompassing different languages,various speaking styles, etc.) into the training set6. Future work: DNN adaption and other objective functionsZaragoza, 27/05/145324

5/26/2014Other Recent Efforts1. Demos: http://home.ustc.edu.cn/ xuyong62/demo/SE DNN.html2. Speech separation: DNN-based semi-supervised speech separationworks better than state-of-the-art supervised speech separation(paper submitted to Interspeech20143. Dual-Output DNN for separation (submitted to ISCSLP2014)4. Robust speech recognition: better results than state-of-the-artwith only DNN-based pre-processing, additional compensation canbe added later (paper submitted to Interspeech2014)5. Transfer language learning for DNN (submitted to ISCSLP2014)6. DNN-based bandwidth expansion works better than all other stateof-the-art techniques (submitted to publication)Zaragoza, 27/05/1454References[1] Y. Ephraim and D. Malah, โ€œSpeech enhancement using a minimum-mean square error short-timespectral amplitude estimator,โ€ IEEE Trans. on Acoustics, Speech and Signal Processing, Vol. 32, No.6,pp. 1109-1121, 1984.[2] I. Cohen, โ€œNoise spectrum estimation in adverse environments: improved minima controlledrecursive averaging,โ€IEEE Trans. on Speech and Audio Processing, Vol. 11, No. 5, pp. 466-475, 2003.[3] E. A. Wan and A. T. Nelson, โ€œNetworks for speech enhancement,โ€ in Handbook of Neural Networksfor Speech Processing, Edited by Shigeru Katagiri, Artech House, Boston, 1998.[4] Y. Xu, J. Du, L.-R. Dai and C.-H. Lee, โ€œAn experimental study on speech enhancement based ondeep neural net-works,โ€ IEEE Signal Processing Letters, Vol. 21, No. 1, pp. 65-68, 2014.[5] B.-Y. Xia and C.-C. Bao, โ€œSpeech enhancement with weighted denoising Auto-Encoder,โ€ Proc.Interspeech, pp. 3444-3448, 2013.[6] X.-G. Lu and Y. Tsao and S. Matsuda and C. Hori,โ€œSpeech enhancement based on deep denoisingAuto-Encoder,โ€ Proc. Interspeech, pp. 436-440, 2013.[7] Y. X. Wang and D. L. Wang, โ€œTowards scaling up classification based speech separation,โ€ IEEE Trans.on Audio, Speech and Language Processing, Vol. 21, No. 7, pp. 1381-1390, 2013.[8] G. Hu, 100 nonspeech environmental sounds, pus.html.[9] S. I. Tamura, โ€œAn analysis of a noise reduction neural network,โ€ Proc. ICASSP, pp. 2001-2004, 1989.[10] F. Xie and D. V. Compernolle, โ€œA family of MLP based nonlinear spectral estimators for noisereduction,โ€ Proc. ICASSP, pp. 53-56, 1994.[11] B.-Y. Xia and C.-C. Bao, โ€œWiener filtering based speech enhancement with Weighted Denoising55Auto-encoder and noise classification,โ€ Speech Communication, V. 60, P. 13โ€“29, 2014.25

Speech enhancement based on deep neural network s SE-DNN: background DNN baseline and enhancement Noise-universal SE-DNN Zaragoza, 27/05/14 3 Speech Enhancement Enhancing Speech enhancement aims at improving the intelligibility and/or overall perceptual quality of degraded speech signals using audio signal processing techniques

Related Documents:

Speech Enhancement Speech Recognition Speech UI Dialog 10s of 1000 hr speech 10s of 1,000 hr noise 10s of 1000 RIR NEVER TRAIN ON THE SAME DATA TWICE Massive . Spectral Subtraction: Waveforms. Deep Neural Networks for Speech Enhancement Direct Indirect Conventional Emulation Mirsamadi, Seyedmahdad, and Ivan Tashev. "Causal Speech

neural networks and substantial trials of experiments to design e ective neural network structures. Thus we believe that the design of neural network structure needs a uni ed guidance. This paper serves as a preliminary trial towards this goal. 1.1. Related Work There has been extensive work on the neural network structure design. Generic algorithm (Scha er et al.,1992;Lam et al.,2003) based .

Deep Convolutional Neural Network for Image . We note directly applying existing deep neural networks does not produce reasonable results. Our solution is to establish the connection between traditional optimization-based schemes and a neural network architecture where

14.4: A Scalable Speech Recognizer with Deep-Neural-Network Acoustic Models and Voice-Activated Power Gating 2017 IEEE International Solid-State Circuits Conference2of 31 Diversification of speech interfaces Today Personal assistant (smartphone) Search (smartphone, PC) Future Wearables Appliances Robots

Deep Neural Networks Convolutional Neural Networks (CNNs) Convolutional Neural Networks (CNN, ConvNet, DCN) CNN a multiโ€layer neural network with โ€“ Local connectivity: Neurons in a layer are only connected to a small region of the layer before it โ€“ Share weight parameters across spatial positions:

Neural Network, Power, Inference, Domain Specific Architecture ACM Reference Format: KiseokKwon,1,2 AlonAmid,1 AmirGholami,1 BichenWu,1 KrsteAsanovic,1 Kurt Keutzer1. 2018. Invited: Co-Design of Deep Neural Nets and Neural Net Accelerators f

Deep Learning 1 Introduction Deep learning is a set of learning methods attempting to model data with complex architectures combining different non-linear transformations. The el-ementary bricks of deep learning are the neural networks, that are combined to form the deep neural networks.

4 Shaft Capacity in Clay (Alpha Method) Soft-stiff clay Adhesion factors