Search

Search Funnelback University

Search powered by Funnelback
1 - 10 of 468 search results for KaKaoTalk:ZA31 24 24 |u:mi.eng.cam.ac.uk where 0 match all words and 468 match some words.
  1. Results that match 2 of 3 words

  2. 20 Feb 2018: Trial # users average # calls median # callsAMT 140 6.5 2Cambridge 17 24.4 20. ... vol. 24, no. 2, pp.150–174, 2010. [7] Amazon, “Amazon Mechanical Turk,” 2011.
  3. PowerPoint プレゼンテーション

    mi.eng.cam.ac.uk/UKSpeech2017/posters/e_tsunoo.pdf
    3 Jul 2018: 24,000 fishermen a year. Mostly in storms. And not every country keeps accurate records.
  4. 4 Nov 2018: 23] L. Breiman. Bagging predictors. Machine learning,24(2):123–140, 1996. [24] O. Siohan, B. ... IEEE/ACM Transactions on Audio, Speech,and Language Processing, 24(8):1438–1449, 2016. Introduction. Graphemic English systems.
  5. 20 Feb 2018: error rate of 33.2 %, and for the first and secondorder derivatives the error rates of the classifiers are 33.1 %and 24.2 %, respectively. ... 24.8 42.4par 22.7 13.1 21.7 32.4 32.7 25.2 27.4 45.1.
  6. Simplifying very deep convolutional neural network architectures for…

    mi.eng.cam.ac.uk/UKSpeech2017/posters/j_rownicka.pdf
    3 Jul 2018: training set of Aurora4. Model A B C D AVGDNN/clntr 2.71 43.00 24.06 58.66 45.48VDCNN-max-4FC/clntr 2.32 35.99 21.20 ... 24, no. 12, pp. 2263-2276, Dec. 2016. Contact: j.m.rownicka@sms.ed.ac.uk.
  7. 15 Jun 2018: Sig-nificant WER improvements were observed after interpolatingwith the n-gram LM for n-best rescoring – a common practicefor speech recognition [8, 24, 25]. ... 24] S. Kombrink, T. Mikolov, M. Karafiát, and L. Burget, “Recurrentneural network
  8. 20 Feb 2018: The Knowledge Engineering Review, Vol. 00:0, 1–24. c 2006, Cambridge University PressDOI: 10.1017/S000000000000000 Printed in the United Kingdom.
  9. Template.dvi

    mi.eng.cam.ac.uk/~ar527/chen_asru2017.pdf
    15 Jun 2018: LM rescoredev eval. Vit CN Vit CNng4 - 23.8 23.5 24.2 23.9. ... LM #succ words dev evalng4 23.8 24.2. uni-rnn - 21.7 22.1.
  10. ICSLPDataCollection-10

    mi.eng.cam.ac.uk/~sjy/papers/wiyo04b.pdf
    20 Feb 2018: Per-turn. WER. Per-dialog WER. None 2 6 24 83 % 0 % 0 % Low 4 12 48 83 % 32 % 28 % Med 4 12 48 77 % 46 % 41 % Hi 2 6 24 ... Dataset. Metrics (task & user sat). R2 Significant predictors. ALL User-S 52 % 1.03 Task ALL User-C 60 % 5.29 Task – 1.54
  11. 15 Jun 2018: In [24] phonetic pronunciation features consisting ofa set of phone-pair distances were proposed for vowels and ap-plied to read speech. ... 8, no. 4, pp. 369–394, 1994. [24] N. Minematsu, S. Asakawa, and K.

Refine your results

Search history

Recently clicked results

Recently clicked results

Your click history is empty.

Recent searches

Recent searches

Your search history is empty.