Search

Search Funnelback University

Search powered by Funnelback
1 - 15 of 15 search results for `broadcast news texts` |u:mi.eng.cam.ac.uk
  1. Fully-matching results

  2. The Cambridge University Spoken Document Retrieval System

    mi.eng.cam.ac.uk/reports/full_html/johnson_icassp99.html/
    8 Mar 2000: 132 million words of broadcast news texts, the LDC-distributed 1995 newswire texts, and the transcriptions of the acoustic training data. ... Furthermore new language models were estimated from an extended set of broadcast news transcripts and newspaper
  3. The 1998 HTK Broadcast News Transcription System: Development and…

    mi.eng.cam.ac.uk/reports/full_html/woodland_darpa99.html/
    2 Mar 2000: The 1997 system used N-gram language models trained on 132 million words of broadcast news texts, the LDC-distributed 1995 newswire texts, and the transcriptions from BNtrain97 (LMtrain97). ... The word N-grams were trained by interpolating (and merging)
  4. LARGE SCALE MMIE TRAINING FOR CONVERSATIONAL TELEPHONE SPEECH…

    mi.eng.cam.ac.uk/reports/full_html/woodland_stw00.html/
    5 Oct 2000: The system uses word-based N-gram LMs estimated from an interpolation of Hub5 acoustic training transcriptions and Broadcast News texts.
  5. General Query Expansion Techniques for Spoken Document Retrieval

    mi.eng.cam.ac.uk/reports/full_html/jourlin_esca99.html/
    2 Mar 2000: Our retrieval system was run on 7 different sets of automatically transcribed broadcast news texts with a WER varying from 24.8% to 66.2%. ... J. Young. Segment Generation and Clustering in the HTK Broadcast News Transcription System [in Proc.
  6. The 1997 HTK Broadcast News Transcription System

    mi.eng.cam.ac.uk/reports/full_html/woodland_darpa98.html/
    1 Mar 2000: The system uses a language model trained on 132 million words of broadcast news texts, the LDC-distributed 1995 newswire texts, and the transcriptions from BNtrain96. ... using the broadcast news training texts, the acoustic training data and 1995
  7. Experiments in Broadcast News Transcription

    mi.eng.cam.ac.uk/reports/full_html/woodland_icassp98.html/
    1 Mar 2000: Experiments in Broadcast News Transcription. P.C. Woodland, T. Hain, S.E. Johnson, T. ... Cross-word context dependent decision tree state clustered mixture Gaussian HMMs are used with a 65k word vocabulary and a language model trained on 132 million
  8. The 1998 HTK Broadcast News Transcription System:Development and…

    mi.eng.cam.ac.uk/reports/svr-ftp/woodland_darpa99.pdf
    8 Mar 2000: The1997 system used N-gram language models trained on 132million words of broadcast news texts, the LDC-distributed1995 newswire texts, and the transcriptions from BNtrain97(LMtrain97). ... 5.1. Language Models. The word N-grams were trained by
  9. THE CAMBRIDGE UNIVERSITY SPOKEN DOCUMENT RETRIEVAL SYSTEM S.E.…

    mi.eng.cam.ac.uk/reports/svr-ftp/johnson_icassp99.pdf
    8 Mar 2000: The paper is organised as follows. Firstly an overview ofthe HTK broadcast news transcription system used to generatehigh quality automatic transcriptions is given. ... of broadcast news texts, the LDC-distributed 1995newswire texts, and the
  10. EXPERIMENTS IN BROADCAST NEWS TRANSCRIPTION P.C. Woodland, T. Hain,…

    mi.eng.cam.ac.uk/reports/svr-ftp/woodland_icassp98.pdf
    10 Apr 2000: Cross-word context dependent deci-sion tree state clustered mixture Gaussian HMMs are used with a65k word vocabulary and a language model trained on 132 millionwords of broadcast news texts, along ... Young S.J. (1997) TheDevelopment of the 1996
  11. GENERAL QUERY EXPANSION TECHNIQUESFOR SPOKEN DOCUMENT RETRIEVAL…

    mi.eng.cam.ac.uk/reports/svr-ftp/jourlin_esca99.pdf
    10 Apr 2000: different sets ofautomatically transcribed broadcast news texts with aWER varying from24.8% to 66.2%. ... The input data is presented to the system as com-plete episodes of broadcast news shows and these are.
  12. Spoken Document Retrieval for TREC-8 at Cambridge University

    mi.eng.cam.ac.uk/reports/full_html/johnson_trec8.html/
    30 Mar 2000: Three fixed backoff word-based language models were trained, from broadcast news text, newspaper texts and acoustic transcriptions, which were all generated using data from before 31st January 1998. ... 1996 to Jan. 1998 (from the Primary Source Media
  13. THE DEVELOPMENT OF THE1996 HTK BROADCAST NEWS TRANSCRIPTION SYSTEM ...

    mi.eng.cam.ac.uk/reports/svr-ftp/woodland_darpa97.pdf
    8 Mar 2000: 4.6. Static Language Models. The evaluation system language model (LM) had a word listcontaining 65423 words chosen from the most frequent wordsin the broadcast news training texts, with the most ... 5. CONCLUSIONThis paper has described our initial
  14. The 1997 HTK Broadcast News Transcription SystemP.C. Woodland, T. ...

    mi.eng.cam.ac.uk/reports/svr-ftp/woodland_darpa98.pdf
    8 Mar 2000: The 1997 HTK Broadcast News Transcription SystemP.C. Woodland, T. Hain, S.E. ... These were trained on the 132million words of the LDC broadcast news training texts,the transcriptions of the BNtrain97 data (added twice), the1995 newswire texts (both
  15. LARGE SCALE DISCRIMINATIVE TRAINING FORSPEECH RECOGNITION P.C.…

    mi.eng.cam.ac.uk/reports/svr-ftp/woodland_asr00.pdf
    6 Nov 2000: It was reported in [32] that FD didn’t improveerror rates over MLE trained models for a broadcast newsrecognition task. ... The system uses word-based N-gramLMs estimated from an interpolation of Hub5 acoustic train-ing transcriptions and Broadcast
  16. SPOKEN DOCUMENT RETRIEVAL FOR TREC-8 AT CAMBRIDGE UNIVERSITY S.E. ...

    mi.eng.cam.ac.uk/reports/svr-ftp/johnson_trec8.pdf
    10 Apr 2000: Three fixed backoff word-based language models were trained,from broadcast news text, newspaper texts and acoustic tran-scriptions, which were all generated using data from before 31stJanuary 1998. ... The first model was built using 190 million wordsof

Refine your results

Related searches for `broadcast news texts` |u:mi.eng.cam.ac.uk

Search history

Recently clicked results

Recently clicked results

Your click history is empty.

Recent searches

Recent searches

Your search history is empty.