Search

Search Funnelback University

Search powered by Funnelback
151 - 200 of 1,000 search results for TALK:Vb20 24 / |u:mi.eng.cam.ac.uk where 0 match all words and 1,000 match some words.
  1. Results that match 1 of 2 words

  2. 20 Feb 2018: The static feature set comprised 24 Mel-Cepstral coefficients,logarithm of F0 and aperiodic energy components in five frequency. ... 12 sentences werethen randomly selected to make up a testset for each listener, leadingto 24 wave files pairs (12 for
  3. 20 Feb 2018: Corpus Mean (SD) Grades Correlationn Human Auto R p. L 21 24.2 (3.1) 17.1 (1.9). ... 69C 50 24.0 (3.0) 15.6 (3.3). 59. 01. Table 2: Mean (standard deviation) of human andautomated grades, along with Pearson’s correla-tions between the human and
  4. 20 Feb 2018: 1.2% 2.0%Request 17.4% 24.5% 18.4% 24.4%. ... 24, no. 4, pp. 562–588, 2010. [22] J Peters and S Schaal, “Natural Actor-Critic,” Neurocomput-ing, vol.
  5. Mark John Francis Gales Thesis, Book Chapters and Magazines ...

    mi.eng.cam.ac.uk/~mjfg/publications_list.pdf
    11 Oct 2023: Audio Speech Lang.Process., vol. 24, no. 8, pp. 1438–1449, 2016. 9. ... Signal Processing,ICASSP 2015, South Brisbane, Queensland, Australia, April 19-24, 2015, pp.
  6. 20 Feb 2018: 0.24. 0.22. 0.20. 0.18. 0.16. Log-. likel. ihoo. dpe. rmic. ro-tu. ... 24, no. 4, pp. 562–588, 2010. [14] SpaceBook. EC FP7/2011-16, grant number 270019.
  7. 20 Feb 2018: on Mancorpora)a 91.40 90.17 90.24. 90.20Auto (Google MT) 90.81 90.77 87.72 89.223. ... 24, no. 2, pp. 150–174, April 2010. [8] P. Koehn, H.
  8. tech.dvi

    mi.eng.cam.ac.uk/~sjy/papers/bghk13.pdf
    20 Feb 2018: 24, pp. 562–588, 2010. [3] G. Aist, J. Allen, E. Campana, C. ... 24] M. Henderson, M. Gašić, B. Thomson, P. Tsiakoulis, K.Yu,and S.
  9. acl2010.dvi

    mi.eng.cam.ac.uk/~sjy/papers/gjkm10.pdf
    20 Feb 2018: Computer Speech and Language, 24(2):150–174.
  10. 8 Sep 2010: Speech Lang.,vol. 24, no. 4, pp. 648–662, 2010. [8] B. Taskar, “Learning structured prediction models: a large marginapproach,” Ph.D.
  11. sigdial11_sdc10-Feb27-V2

    mi.eng.cam.ac.uk/~sjy/papers/bbch11.pdf
    20 Feb 2018: 6% 24.6% 14.7% 9.6%. ... Length (s) Turns/call Words/turn. SYS1 control 155 18.29 2.87 (2.84) SYS1 live 111 16.24 2.15 (1.03) SYS2 control 147 17.57 1.63
  12. 20 Feb 2018: 24, no. 2, pp.150–174, 2010. [5] B. Thomson and S. Young, “Bayesian update of dialogue state:A POMDP framework for spoken dialogue systems,” ComputerSpeech and Language, vol. ... 24, no. 4, pp. 562–588, 2010. [6] M. Gašić, C. Breslin, M.
  13. 20 Feb 2018: In mostcases, a data-driven approach is followed, either by detect-ing/annotating emphasized words in existing corpora [23, 10] orby collecting speech corpora specifically designed for emphasismodeling [24]. ... Appointment Booking Task
  14. 20 Feb 2018: Computer Speech and Language,24:562–588. Jason D. Williams and Steve Young. 2007.
  15. Uncertainty management for on-line optimisation of a…

    mi.eng.cam.ac.uk/~sjy/papers/dgcg11.pdf
    20 Feb 2018: 24, no. 2, pp. 150–174,2010. [6] W. Eckert, E. Levin, and R. ... 9] O. Pietquin, M. Geist, S. Chandramohan, and H. Frezza-Buet,“Sample-Efficient Batch Reinforcement Learning for DialogueManagement Optimization,” ACM Transactions on Speech
  16. 5 Apr 2016: Speech andSignal Processing, ICASSP 2015, South Brisbane, Queens-land, Australia, April 19-24, 2015, 2015, pp. ... IEEE, 2015, pp. 4315–4319. [24] Mark JF Gales, “Cluster adaptive training of hidden markovmodels,” Speech and Audio Processing, IEEE
  17. 20 Feb 2018: 24, pp. 562–588, 2010. [20] M. Lukoeviius and H. Jaeger, “Reservoir computing approachesto recurrent neural network training,” Computer Science Review,vol. ... abs/1412.2306, 2014. [24] G. Mesnil, Y. Dauphin, K. Yao, Y. Bengio, L.
  18. Learning Domain-Independent Dialogue Policies via…

    mi.eng.cam.ac.uk/~sjy/papers/wsws15.pdf
    20 Feb 2018: Computer Speech andLanguage, 24(4):562–588. Zhuoran Wang and Oliver Lemon. 2013. A simpleand generic belief tracking mechanism for the Dia-log State Tracking Challenge: On the believabilityof observed information.
  19. 20 Feb 2018: When errors are correlated belief tracking is less accurate be-cause it tends to over-estimate alternatives in the N-best list[24]. ... 24, no. 4, pp. 562–588, 2010. 17. T. Minka, “Expectation Propagation for Approximate Bayesian Inference,” in
  20. LEARNING BETWEEN DIFFERENT TEACHER AND STUDENT MODELS IN ASR ...

    mi.eng.cam.ac.uk/~mjfg/ALTA/ASRU2019_TS.pdf
    20 Dec 2019: The derivatives of the per-frame. observation log-likelihoods with respects to the parameters are [24]. ... Work in [24] suggests several methods to improve gra-dient descent training of a GMM.
  21. is-05-hvs6_final

    mi.eng.cam.ac.uk/~sjy/papers/seyo05.pdf
    20 Feb 2018: 52 class n-gram 26.3 25.0 24.9 HVS_52 21.7 20.4 20.1. Table 1: Perplexity for models of varied stack depths trained for 250 iterations.
  22. 29 Sep 2016: 24 of 67. S2S: Generative Models [5, 6]. • Consider two sequences L T: input: x1:T = {x1, x2,. , ... 47 of 67. ASR: Sequence Training [24]. • Cross-Entropy using fixed alignment standard criterion (RNN).
  23. poyosp08

    mi.eng.cam.ac.uk/~sjy/papers/dpyo08.pdf
    20 Feb 2018: T - R 0.54 0.22 0.24. R - O 0.52 0.31 0.17.
  24. yokou

    mi.eng.cam.ac.uk/~sjy/papers/toyo09.pdf
    20 Feb 2018: where. x(d)q = 2. “c(d)q. Dc(d)q. E”(v(cq) v(c)) p(d)v. (24). 4.3.
  25. 12 Apr 2022: default:cout << "Trading error: trading system failure." << endl;exit(-1);. }}. 24.
  26. Online_ASRU11.dvi

    mi.eng.cam.ac.uk/~sjy/papers/gjty11.pdf
    20 Feb 2018: 24, no. 2, pp. 150–174, 2010. [9] B. Thomson and S. ... 24, no. 4, pp. 562–588, 2010. [10] M. Gǎsić, S. Keizer, F.
  27. 12 Jul 2016: When the segment level features are used, the log-linear modelparameters η̂ could be considered as phone dependent acousticmodel scales [24]. ... In joint decoding, 2% relative WER performance gain wasachieved over the hybrid system, from 11.24% to
  28. 20 Feb 2018: 24, no. 4, pp. 562–588, 2010. [13] M Gašić, C Breslin, M Henderson, D Kim, M Szummer,B Thomson, P Tsiakoulis, and S Young, “POMDP-based dia-logue manager adaptation
  29. JOINT MODELLING OF VOICING LABEL AND CONTINUOUS F0 FOR ...

    mi.eng.cam.ac.uk/~sjy/papers/yuyo11a.pdf
    20 Feb 2018: Mixed excitation using STRAIGHT was employed [12].The speech features used were 24 Mel-Cepstral spectral coefficients,the logarithm of F0, and aperiodic components in five frequencybands (0 to 1, 1
  30. System Combination with Log-linear Models

    mi.eng.cam.ac.uk/~mjfg/icassp16_yang.pdf
    5 Apr 2016: When the segment level features are used, the log-linear modelparameters η̂ could be considered as phone dependent acousticmodel scales [24]. ... In joint decoding, 2% relative WER performance gain wasachieved over the hybrid system, from 11.24% to
  31. 13 Jun 2013: Better performance couldbe achieved by gradually increasing C. Equation (16) is also known as the training criterion ofthe structural SVM [23, 24]. ... ACM, 2004. [24] Shi-Xiong Zhang and Mark Gales, “Structured SVMs for auto-matic speech
  32. 20 Feb 2018: Sec. 2. the reward [20, 21, 22, 23] by using the PARADISE frame-work [24]. ... 2356–2361. [24] M. Walker, D. J. Litman, C. A. Kamm, and A.
  33. Low-Resource Speech Recognition and Keyword-Spotting

    mi.eng.cam.ac.uk/~mjfg/SPECOM_2017.pdf
    29 Nov 2017: 23/63. Stimulated Systems. /ey//em/. /sil/. /sh/. /ow/ /ay/. 24/63. Stimulated Network Training. •
  34. slides.dvi

    mi.eng.cam.ac.uk/~mjfg/Bilbao14/talk.pdf
    25 Jun 2014: Cambridge University. Engineering DepartmenteNTERFACE June 2014 24. Controllable and Adaptable Statistical Parametric Speech Synthesis Systems. ... Integrated Expressive Speech Training [24]. Training. Expressive State. Prediction. ExtractionAcoustic
  35. Tsiakoulis_Pirros_1293

    mi.eng.cam.ac.uk/~sjy/papers/tghp12.pdf
    20 Feb 2018: Proc. ICASSP, Taipei, Taiwan. Thomson, B. & Young, S. (2010)“Bayesian Update of Dialogue State: A POMDP framework for spoken dialogue systems.” Computer Speech and Language 24(4):562-588.
  36. 20 Feb 2018: 24, no. 4, pp. 562 – 588, 2010. [2] R. Sutton and A.
  37. main.dvi

    mi.eng.cam.ac.uk/~sjy/papers/youn07
    20 Feb 2018: Hence, an itera-tive algorithm can be implemented which repeatedlyscans through the vocabulary, testing each word tosee if moving it to some other class would increasethe likelihood [24]. ... Thed p nuisance dimensions are modelled by a
  38. paper.dvi

    mi.eng.cam.ac.uk/~mjfg/yw293_ASRU11.pdf
    19 Jan 2012: and. J(m)xδ =. g. xtδ. µ(m)xe ,µl,µn. (24). Thus the model parameters are compensated by. ... Forexample, using the initial noise estimate, RVTSJ performancevaried from 27.5% to 31.7%, while the performance of MLestimated noise only varied from 24.3%
  39. 20 Feb 2018: 6. RELATED WORK. In machine learning in general much research has looked atadaptation of statistical models [21, 22, 23] however researchinto adaptation of SDS components to new domains [24, 25,26, ... 24, pp.562–588, 2010. [16] Milica Gašić and
  40. SSVM_LVCSR_ASRU11.dvi

    mi.eng.cam.ac.uk/~mjfg/sxz20_ASRU11.pdf
    19 Jan 2012: 23)), finds the most violated constraint (Eq. (24)), andadds it to the working set. ... Parallelingthe loop for Eq. (24) will lead to a substantial speed-up in thenumber of threads.
  41. 11 Mar 2016: In order to solve this problem, recentlythere has increasing research interest in deriving efficient paralleltraining algorithms for RNNLMs [22, 23, 24, 25]. ... 9.1 2.9 117.85GRNN 5472.1 170.0 24.3 7.9 2.4 117.6.
  42. 20 Feb 2018: F1 ICE. Slot4 95.29% 90.89% 95.72% 93.24% 0.478 89.92% 74.73% 61.56% 67.51% 0.743. ... Task - - - - - 97.12% 83.24% 64.93% 72.95% 0.175.
  43. 20 Feb 2018: scr-10% 2.24 2.03 2.00 1.92. p <0.05, p <0.005Table 2: Human evaluation for utterance quality intwo domains.
  44. draft21.dvi

    mi.eng.cam.ac.uk/~mjfg/ASRU13.pdf
    7 Nov 2013: procedure described in [24]. Unilingual and multilingual AMs were. each built from a flat start. ... 24] J. Park et al., “The Efficient Incorporation of MLP Fea-.
  45. 20 Feb 2018: 24,no. 4, pp. 562–588, Oct. 2010. [7] Jost Schatzmann, Statistical user and error modellingfor spoken dialogue systems, Ph.D.
  46. paper.dvi

    mi.eng.cam.ac.uk/~mjfg/sxz20_inter11.pdf
    19 Jan 2012: SpeechLang., vol. 24, no. 4, pp. 648–662, 2010. [6] B. Taskar, “Learning structured prediction models: a large marginapproach,” Ph.D.
  47. 20 Feb 2018: In this section, a Gaussianprocess-based reward estimator is described which uses active learning tolimit intrusive requests for feedback and a noise model to mitigate the effectsof inaccurate feedback [24]. ... 24. Figure 13: The number of times each
  48. 13 Jun 2013: Ney LM with optimized discounting parameters [24] usinga modified version of the RWTH open source decoder [25]; and (6)DBN, a deep belief network hybrid model [26, 27] with discrimi-native ... Ney, “Posterior-scaled mpe: Novel discriminative
  49. paper.dvi

    mi.eng.cam.ac.uk/reports/svr-ftp/liu_icassp2004.pdf
    29 May 2004: Gauss 24.0 20.7 20.7 17.7 16.0WER (%) 35.3 35.1 35.2 35.3 35.5. ... The 16 component systems was then iteratively split until thenumber of components was 24.
  50. lect1.dvi

    mi.eng.cam.ac.uk/~mjfg/local/4F10/lect1.pdf
    10 Nov 2015: for minimum error with generative models. 24 Engineering Part IIB: Module 4F10 Statistical Pattern Processing.
  51. eps.dis.dur.testa.eps

    mi.eng.cam.ac.uk/~mjfg/gales_ASRU09.pdf
    14 Sep 2010: Using 17 pairs, about 24% of thetotal number of pairs, 92% of the WER improvement usingthe 1-v-1 system over the VTS baseline was achieved.

Search history

Recently clicked results

Recently clicked results

Your click history is empty.

Recent searches

Recent searches

Your search history is empty.